diff --git a/CHANGELOG.md b/CHANGELOG.md index 89bb5c302e..373185ff66 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,39 @@ # Changelog +## [2.176.0](https://github.com/googleapis/google-api-python-client/compare/v2.175.0...v2.176.0) (2025-07-08) + + +### Features + +* **adsenseplatform:** Update the api https://togithub.com/googleapis/google-api-python-client/commit/d4e048a17a01e4765f8b1cc06fbb483ee279b319 ([14b2f0d](https://github.com/googleapis/google-api-python-client/commit/14b2f0db55b909445b0e83cdb7ff079f9cda0d35)) +* **aiplatform:** Update the api https://togithub.com/googleapis/google-api-python-client/commit/8f166c35be80736a2e4962707b5306e1286106b7 ([14b2f0d](https://github.com/googleapis/google-api-python-client/commit/14b2f0db55b909445b0e83cdb7ff079f9cda0d35)) +* **androidpublisher:** Update the api https://togithub.com/googleapis/google-api-python-client/commit/497de380ebf932e0b39dc2ed3e74302bc62c9f4c ([14b2f0d](https://github.com/googleapis/google-api-python-client/commit/14b2f0db55b909445b0e83cdb7ff079f9cda0d35)) +* **apihub:** Update the api https://togithub.com/googleapis/google-api-python-client/commit/052ab9c0c87e629b60718442e554e026e2ad1d5c ([14b2f0d](https://github.com/googleapis/google-api-python-client/commit/14b2f0db55b909445b0e83cdb7ff079f9cda0d35)) +* **backupdr:** Update the api https://togithub.com/googleapis/google-api-python-client/commit/1a00f4eab55872f098655116e8ded67a5689f8c0 ([14b2f0d](https://github.com/googleapis/google-api-python-client/commit/14b2f0db55b909445b0e83cdb7ff079f9cda0d35)) +* **datamigration:** Update the api https://togithub.com/googleapis/google-api-python-client/commit/c8e508858fd205e9df2bb73433f807be6a4445f8 ([14b2f0d](https://github.com/googleapis/google-api-python-client/commit/14b2f0db55b909445b0e83cdb7ff079f9cda0d35)) +* **dataproc:** Update the api https://togithub.com/googleapis/google-api-python-client/commit/87901777ab66b96b26c305d0e34b35e8623e06c7 ([14b2f0d](https://github.com/googleapis/google-api-python-client/commit/14b2f0db55b909445b0e83cdb7ff079f9cda0d35)) +* **datastream:** Update the api https://togithub.com/googleapis/google-api-python-client/commit/5c59d65a79d2e5963d467dbdd82fe8ca346275de ([14b2f0d](https://github.com/googleapis/google-api-python-client/commit/14b2f0db55b909445b0e83cdb7ff079f9cda0d35)) +* **dialogflow:** Update the api https://togithub.com/googleapis/google-api-python-client/commit/5bd3e6d2ad8c7bc223b81164380d1eedc33e2308 ([14b2f0d](https://github.com/googleapis/google-api-python-client/commit/14b2f0db55b909445b0e83cdb7ff079f9cda0d35)) +* **discoveryengine:** Update the api https://togithub.com/googleapis/google-api-python-client/commit/b8b4f53ab1a4e2a5ad6bd620a8e3efd3c7b0ca48 ([14b2f0d](https://github.com/googleapis/google-api-python-client/commit/14b2f0db55b909445b0e83cdb7ff079f9cda0d35)) +* **dlp:** Update the api https://togithub.com/googleapis/google-api-python-client/commit/f6d0422efea1c3f08d8684fc52692a5dea441a52 ([14b2f0d](https://github.com/googleapis/google-api-python-client/commit/14b2f0db55b909445b0e83cdb7ff079f9cda0d35)) +* **drive:** Update the api https://togithub.com/googleapis/google-api-python-client/commit/7d17d49eaaed37fafeb1ae41f74348911831ad93 ([14b2f0d](https://github.com/googleapis/google-api-python-client/commit/14b2f0db55b909445b0e83cdb7ff079f9cda0d35)) +* **iam:** Update the api https://togithub.com/googleapis/google-api-python-client/commit/4bcf549bbfc3e57d6dacedcbb61acae9d3e59d79 ([14b2f0d](https://github.com/googleapis/google-api-python-client/commit/14b2f0db55b909445b0e83cdb7ff079f9cda0d35)) +* **merchantapi:** Update the api https://togithub.com/googleapis/google-api-python-client/commit/b33a19c205b0ade5524febe46e7355379f68d012 ([14b2f0d](https://github.com/googleapis/google-api-python-client/commit/14b2f0db55b909445b0e83cdb7ff079f9cda0d35)) +* **policysimulator:** Update the api https://togithub.com/googleapis/google-api-python-client/commit/3e66a59c7d81634671ac2649003efa6e60ff0aee ([14b2f0d](https://github.com/googleapis/google-api-python-client/commit/14b2f0db55b909445b0e83cdb7ff079f9cda0d35)) +* **redis:** Update the api https://togithub.com/googleapis/google-api-python-client/commit/c0f8e8db10c26521d6e3a5f09a6f25d02ae55d5e ([14b2f0d](https://github.com/googleapis/google-api-python-client/commit/14b2f0db55b909445b0e83cdb7ff079f9cda0d35)) +* **retail:** Update the api https://togithub.com/googleapis/google-api-python-client/commit/3e3c6956b068bd7053ab57e02e5171965e530dfe ([14b2f0d](https://github.com/googleapis/google-api-python-client/commit/14b2f0db55b909445b0e83cdb7ff079f9cda0d35)) +* **securitycenter:** Update the api https://togithub.com/googleapis/google-api-python-client/commit/da5d1e9c428f55f1a33ad0631ee54b2baace23a9 ([14b2f0d](https://github.com/googleapis/google-api-python-client/commit/14b2f0db55b909445b0e83cdb7ff079f9cda0d35)) +* **spanner:** Update the api https://togithub.com/googleapis/google-api-python-client/commit/ceb6aaefe7d46f763ec8d9c390814e52ec9b0a95 ([14b2f0d](https://github.com/googleapis/google-api-python-client/commit/14b2f0db55b909445b0e83cdb7ff079f9cda0d35)) +* **storage:** Update the api https://togithub.com/googleapis/google-api-python-client/commit/06056fcc8c4c9a00192472977740a048759f2fdd ([14b2f0d](https://github.com/googleapis/google-api-python-client/commit/14b2f0db55b909445b0e83cdb7ff079f9cda0d35)) +* **workstations:** Update the api https://togithub.com/googleapis/google-api-python-client/commit/a03d74072877a744151d0f501f688c017086e40e ([14b2f0d](https://github.com/googleapis/google-api-python-client/commit/14b2f0db55b909445b0e83cdb7ff079f9cda0d35)) + + +### Bug Fixes + +* **apigee:** Update the api https://togithub.com/googleapis/google-api-python-client/commit/48a853ece2a2d9155200045f643e295ad657fa16 ([14b2f0d](https://github.com/googleapis/google-api-python-client/commit/14b2f0db55b909445b0e83cdb7ff079f9cda0d35)) +* **composer:** Update the api https://togithub.com/googleapis/google-api-python-client/commit/e1a6cc9098c231fdb60870ddf58216fdf59ccaba ([14b2f0d](https://github.com/googleapis/google-api-python-client/commit/14b2f0db55b909445b0e83cdb7ff079f9cda0d35)) +* **secretmanager:** Update the api https://togithub.com/googleapis/google-api-python-client/commit/57727618a8e0fac0291823c52ce21e4fb8f82507 ([14b2f0d](https://github.com/googleapis/google-api-python-client/commit/14b2f0db55b909445b0e83cdb7ff079f9cda0d35)) + ## [2.175.0](https://github.com/googleapis/google-api-python-client/compare/v2.174.0...v2.175.0) (2025-07-04) diff --git a/docs/dyn/adsenseplatform_v1alpha.accounts.platforms.groups.html b/docs/dyn/adsenseplatform_v1alpha.accounts.platforms.groups.html index fc4614ffd4..000352b88b 100644 --- a/docs/dyn/adsenseplatform_v1alpha.accounts.platforms.groups.html +++ b/docs/dyn/adsenseplatform_v1alpha.accounts.platforms.groups.html @@ -86,6 +86,9 @@

Instance Methods

list_next()

Retrieves the next page of results.

+

+ patch(name, body=None, updateMask=None, x__xgafv=None)

+

Update a Platform Group.

Method Details

close() @@ -159,4 +162,39 @@

Method Details

+
+ patch(name, body=None, updateMask=None, x__xgafv=None) +
Update a Platform Group.
+
+Args:
+  name: string, Identifier. Format: accounts/{account}/platforms/{platform}/groups/{platform_group} (required)
+  body: object, The request body.
+    The object takes the form of:
+
+{ # Representation of a Transparent Platform Group.
+  "description": "A String", # Output only. Description of the PlatformGroup.
+  "name": "A String", # Identifier. Format: accounts/{account}/platforms/{platform}/groups/{platform_group}
+  "revshareMillipercent": { # A representation of a decimal value, such as 2.5. Clients may convert values into language-native decimal formats, such as Java's [BigDecimal](https://docs.oracle.com/en/java/javase/11/docs/api/java.base/java/math/BigDecimal.html) or Python's [decimal.Decimal](https://docs.python.org/3/library/decimal.html). # The revenue share of the PlatformGroup, in millipercent (e.g. 15000 = 15%).
+    "value": "A String", # The decimal value, as a string. The string representation consists of an optional sign, `+` (`U+002B`) or `-` (`U+002D`), followed by a sequence of zero or more decimal digits ("the integer"), optionally followed by a fraction, optionally followed by an exponent. An empty string **should** be interpreted as `0`. The fraction consists of a decimal point followed by zero or more decimal digits. The string must contain at least one digit in either the integer or the fraction. The number formed by the sign, the integer and the fraction is referred to as the significand. The exponent consists of the character `e` (`U+0065`) or `E` (`U+0045`) followed by one or more decimal digits. Services **should** normalize decimal values before storing them by: - Removing an explicitly-provided `+` sign (`+2.5` -> `2.5`). - Replacing a zero-length integer value with `0` (`.5` -> `0.5`). - Coercing the exponent character to upper-case, with explicit sign (`2.5e8` -> `2.5E+8`). - Removing an explicitly-provided zero exponent (`2.5E0` -> `2.5`). Services **may** perform additional normalization based on its own needs and the internal decimal implementation selected, such as shifting the decimal point and exponent value together (example: `2.5E-1` <-> `0.25`). Additionally, services **may** preserve trailing zeroes in the fraction to indicate increased precision, but are not required to do so. Note that only the `.` character is supported to divide the integer and the fraction; `,` **should not** be supported regardless of locale. Additionally, thousand separators **should not** be supported. If a service does support them, values **must** be normalized. The ENBF grammar is: DecimalString = '' | [Sign] Significand [Exponent]; Sign = '+' | '-'; Significand = Digits '.' | [Digits] '.' Digits; Exponent = ('e' | 'E') [Sign] Digits; Digits = { '0' | '1' | '2' | '3' | '4' | '5' | '6' | '7' | '8' | '9' }; Services **should** clearly document the range of supported values, the maximum supported precision (total number of digits), and, if applicable, the scale (number of digits after the decimal point), as well as how it behaves when receiving out-of-bounds values. Services **may** choose to accept values passed as input even when the value has a higher precision or scale than the service supports, and **should** round the value to fit the supported scale. Alternatively, the service **may** error with `400 Bad Request` (`INVALID_ARGUMENT` in gRPC) if precision would be lost. Services **should** error with `400 Bad Request` (`INVALID_ARGUMENT` in gRPC) if the service receives a value outside of the supported range.
+  },
+}
+
+  updateMask: string, Optional. The list of fields to update - currently only supports updating the `description` field.
+  x__xgafv: string, V1 error format.
+    Allowed values
+      1 - v1 error format
+      2 - v2 error format
+
+Returns:
+  An object of the form:
+
+    { # Representation of a Transparent Platform Group.
+  "description": "A String", # Output only. Description of the PlatformGroup.
+  "name": "A String", # Identifier. Format: accounts/{account}/platforms/{platform}/groups/{platform_group}
+  "revshareMillipercent": { # A representation of a decimal value, such as 2.5. Clients may convert values into language-native decimal formats, such as Java's [BigDecimal](https://docs.oracle.com/en/java/javase/11/docs/api/java.base/java/math/BigDecimal.html) or Python's [decimal.Decimal](https://docs.python.org/3/library/decimal.html). # The revenue share of the PlatformGroup, in millipercent (e.g. 15000 = 15%).
+    "value": "A String", # The decimal value, as a string. The string representation consists of an optional sign, `+` (`U+002B`) or `-` (`U+002D`), followed by a sequence of zero or more decimal digits ("the integer"), optionally followed by a fraction, optionally followed by an exponent. An empty string **should** be interpreted as `0`. The fraction consists of a decimal point followed by zero or more decimal digits. The string must contain at least one digit in either the integer or the fraction. The number formed by the sign, the integer and the fraction is referred to as the significand. The exponent consists of the character `e` (`U+0065`) or `E` (`U+0045`) followed by one or more decimal digits. Services **should** normalize decimal values before storing them by: - Removing an explicitly-provided `+` sign (`+2.5` -> `2.5`). - Replacing a zero-length integer value with `0` (`.5` -> `0.5`). - Coercing the exponent character to upper-case, with explicit sign (`2.5e8` -> `2.5E+8`). - Removing an explicitly-provided zero exponent (`2.5E0` -> `2.5`). Services **may** perform additional normalization based on its own needs and the internal decimal implementation selected, such as shifting the decimal point and exponent value together (example: `2.5E-1` <-> `0.25`). Additionally, services **may** preserve trailing zeroes in the fraction to indicate increased precision, but are not required to do so. Note that only the `.` character is supported to divide the integer and the fraction; `,` **should not** be supported regardless of locale. Additionally, thousand separators **should not** be supported. If a service does support them, values **must** be normalized. The ENBF grammar is: DecimalString = '' | [Sign] Significand [Exponent]; Sign = '+' | '-'; Significand = Digits '.' | [Digits] '.' Digits; Exponent = ('e' | 'E') [Sign] Digits; Digits = { '0' | '1' | '2' | '3' | '4' | '5' | '6' | '7' | '8' | '9' }; Services **should** clearly document the range of supported values, the maximum supported precision (total number of digits), and, if applicable, the scale (number of digits after the decimal point), as well as how it behaves when receiving out-of-bounds values. Services **may** choose to accept values passed as input even when the value has a higher precision or scale than the service supports, and **should** round the value to fit the supported scale. Alternatively, the service **may** error with `400 Bad Request` (`INVALID_ARGUMENT` in gRPC) if precision would be lost. Services **should** error with `400 Bad Request` (`INVALID_ARGUMENT` in gRPC) if the service receives a value outside of the supported range.
+  },
+}
+
+ \ No newline at end of file diff --git a/docs/dyn/aiplatform_v1.batchPredictionJobs.html b/docs/dyn/aiplatform_v1.batchPredictionJobs.html index bfc2c82911..2765e6ec54 100644 --- a/docs/dyn/aiplatform_v1.batchPredictionJobs.html +++ b/docs/dyn/aiplatform_v1.batchPredictionJobs.html @@ -371,7 +371,7 @@

Method Details

}, "healthRoute": "A String", # Immutable. HTTP path on the container to send health checks to. Vertex AI intermittently sends GET requests to this path on the container's IP address and port to check that the container is healthy. Read more about [health checks](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#health). For example, if you set this field to `/bar`, then Vertex AI intermittently sends a GET request to the `/bar` path on the port of your container specified by the first value of this `ModelContainerSpec`'s ports field. If you don't specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/ DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) "imageUri": "A String", # Required. Immutable. URI of the Docker image to be used as the custom container for serving predictions. This URI must identify an image in Artifact Registry or Container Registry. Learn more about the [container publishing requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#publishing), including permissions requirements for the Vertex AI Service Agent. The container image is ingested upon ModelService.UploadModel, stored internally, and this original path is afterwards not used. To learn about the requirements for the Docker image itself, see [Custom container requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#). You can use the URI to one of Vertex AI's [pre-built container images for prediction](https://cloud.google.com/vertex-ai/docs/predictions/pre-built-containers) in this field. - "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with [PredictionService.Invoke] eg: "/invoke/foo/bar". Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. + "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with invoke http call eg: "/invoke/foo/bar", however the [PredictionService.Invoke] RPC is not supported yet. Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. "livenessProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes liveness probe. "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. @@ -730,7 +730,7 @@

Method Details

}, "healthRoute": "A String", # Immutable. HTTP path on the container to send health checks to. Vertex AI intermittently sends GET requests to this path on the container's IP address and port to check that the container is healthy. Read more about [health checks](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#health). For example, if you set this field to `/bar`, then Vertex AI intermittently sends a GET request to the `/bar` path on the port of your container specified by the first value of this `ModelContainerSpec`'s ports field. If you don't specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/ DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) "imageUri": "A String", # Required. Immutable. URI of the Docker image to be used as the custom container for serving predictions. This URI must identify an image in Artifact Registry or Container Registry. Learn more about the [container publishing requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#publishing), including permissions requirements for the Vertex AI Service Agent. The container image is ingested upon ModelService.UploadModel, stored internally, and this original path is afterwards not used. To learn about the requirements for the Docker image itself, see [Custom container requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#). You can use the URI to one of Vertex AI's [pre-built container images for prediction](https://cloud.google.com/vertex-ai/docs/predictions/pre-built-containers) in this field. - "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with [PredictionService.Invoke] eg: "/invoke/foo/bar". Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. + "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with invoke http call eg: "/invoke/foo/bar", however the [PredictionService.Invoke] RPC is not supported yet. Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. "livenessProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes liveness probe. "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. @@ -1095,7 +1095,7 @@

Method Details

}, "healthRoute": "A String", # Immutable. HTTP path on the container to send health checks to. Vertex AI intermittently sends GET requests to this path on the container's IP address and port to check that the container is healthy. Read more about [health checks](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#health). For example, if you set this field to `/bar`, then Vertex AI intermittently sends a GET request to the `/bar` path on the port of your container specified by the first value of this `ModelContainerSpec`'s ports field. If you don't specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/ DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) "imageUri": "A String", # Required. Immutable. URI of the Docker image to be used as the custom container for serving predictions. This URI must identify an image in Artifact Registry or Container Registry. Learn more about the [container publishing requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#publishing), including permissions requirements for the Vertex AI Service Agent. The container image is ingested upon ModelService.UploadModel, stored internally, and this original path is afterwards not used. To learn about the requirements for the Docker image itself, see [Custom container requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#). You can use the URI to one of Vertex AI's [pre-built container images for prediction](https://cloud.google.com/vertex-ai/docs/predictions/pre-built-containers) in this field. - "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with [PredictionService.Invoke] eg: "/invoke/foo/bar". Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. + "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with invoke http call eg: "/invoke/foo/bar", however the [PredictionService.Invoke] RPC is not supported yet. Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. "livenessProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes liveness probe. "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. @@ -1466,7 +1466,7 @@

Method Details

}, "healthRoute": "A String", # Immutable. HTTP path on the container to send health checks to. Vertex AI intermittently sends GET requests to this path on the container's IP address and port to check that the container is healthy. Read more about [health checks](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#health). For example, if you set this field to `/bar`, then Vertex AI intermittently sends a GET request to the `/bar` path on the port of your container specified by the first value of this `ModelContainerSpec`'s ports field. If you don't specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/ DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) "imageUri": "A String", # Required. Immutable. URI of the Docker image to be used as the custom container for serving predictions. This URI must identify an image in Artifact Registry or Container Registry. Learn more about the [container publishing requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#publishing), including permissions requirements for the Vertex AI Service Agent. The container image is ingested upon ModelService.UploadModel, stored internally, and this original path is afterwards not used. To learn about the requirements for the Docker image itself, see [Custom container requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#). You can use the URI to one of Vertex AI's [pre-built container images for prediction](https://cloud.google.com/vertex-ai/docs/predictions/pre-built-containers) in this field. - "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with [PredictionService.Invoke] eg: "/invoke/foo/bar". Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. + "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with invoke http call eg: "/invoke/foo/bar", however the [PredictionService.Invoke] RPC is not supported yet. Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. "livenessProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes liveness probe. "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. diff --git a/docs/dyn/aiplatform_v1.endpoints.html b/docs/dyn/aiplatform_v1.endpoints.html index 3c5f3cdb6a..26d0e91ef3 100644 --- a/docs/dyn/aiplatform_v1.endpoints.html +++ b/docs/dyn/aiplatform_v1.endpoints.html @@ -327,7 +327,7 @@

Method Details

"temperature": 3.14, # Optional. Controls the randomness of predictions. "thinkingConfig": { # Config for thinking features. # Optional. Config for thinking features. An error will be returned if this field is set for models that don't support thinking. "includeThoughts": True or False, # Optional. Indicates whether to include thoughts in the response. If true, thoughts are returned only when available. - "thinkingBudget": 42, # Optional. Indicates the thinking budget in tokens. This is only applied when enable_thinking is true. + "thinkingBudget": 42, # Optional. Indicates the thinking budget in tokens. }, "topK": 3.14, # Optional. If specified, top-k sampling will be used. "topP": 3.14, # Optional. If specified, nucleus sampling will be used. @@ -768,7 +768,7 @@

Method Details

"temperature": 3.14, # Optional. Controls the randomness of predictions. "thinkingConfig": { # Config for thinking features. # Optional. Config for thinking features. An error will be returned if this field is set for models that don't support thinking. "includeThoughts": True or False, # Optional. Indicates whether to include thoughts in the response. If true, thoughts are returned only when available. - "thinkingBudget": 42, # Optional. Indicates the thinking budget in tokens. This is only applied when enable_thinking is true. + "thinkingBudget": 42, # Optional. Indicates the thinking budget in tokens. }, "topK": 3.14, # Optional. If specified, top-k sampling will be used. "topP": 3.14, # Optional. If specified, nucleus sampling will be used. @@ -1468,7 +1468,7 @@

Method Details

"temperature": 3.14, # Optional. Controls the randomness of predictions. "thinkingConfig": { # Config for thinking features. # Optional. Config for thinking features. An error will be returned if this field is set for models that don't support thinking. "includeThoughts": True or False, # Optional. Indicates whether to include thoughts in the response. If true, thoughts are returned only when available. - "thinkingBudget": 42, # Optional. Indicates the thinking budget in tokens. This is only applied when enable_thinking is true. + "thinkingBudget": 42, # Optional. Indicates the thinking budget in tokens. }, "topK": 3.14, # Optional. If specified, top-k sampling will be used. "topP": 3.14, # Optional. If specified, nucleus sampling will be used. diff --git a/docs/dyn/aiplatform_v1.projects.locations.batchPredictionJobs.html b/docs/dyn/aiplatform_v1.projects.locations.batchPredictionJobs.html index b858915e19..f5ecc701b4 100644 --- a/docs/dyn/aiplatform_v1.projects.locations.batchPredictionJobs.html +++ b/docs/dyn/aiplatform_v1.projects.locations.batchPredictionJobs.html @@ -402,7 +402,7 @@

Method Details

}, "healthRoute": "A String", # Immutable. HTTP path on the container to send health checks to. Vertex AI intermittently sends GET requests to this path on the container's IP address and port to check that the container is healthy. Read more about [health checks](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#health). For example, if you set this field to `/bar`, then Vertex AI intermittently sends a GET request to the `/bar` path on the port of your container specified by the first value of this `ModelContainerSpec`'s ports field. If you don't specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/ DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) "imageUri": "A String", # Required. Immutable. URI of the Docker image to be used as the custom container for serving predictions. This URI must identify an image in Artifact Registry or Container Registry. Learn more about the [container publishing requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#publishing), including permissions requirements for the Vertex AI Service Agent. The container image is ingested upon ModelService.UploadModel, stored internally, and this original path is afterwards not used. To learn about the requirements for the Docker image itself, see [Custom container requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#). You can use the URI to one of Vertex AI's [pre-built container images for prediction](https://cloud.google.com/vertex-ai/docs/predictions/pre-built-containers) in this field. - "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with [PredictionService.Invoke] eg: "/invoke/foo/bar". Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. + "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with invoke http call eg: "/invoke/foo/bar", however the [PredictionService.Invoke] RPC is not supported yet. Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. "livenessProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes liveness probe. "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. @@ -760,7 +760,7 @@

Method Details

}, "healthRoute": "A String", # Immutable. HTTP path on the container to send health checks to. Vertex AI intermittently sends GET requests to this path on the container's IP address and port to check that the container is healthy. Read more about [health checks](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#health). For example, if you set this field to `/bar`, then Vertex AI intermittently sends a GET request to the `/bar` path on the port of your container specified by the first value of this `ModelContainerSpec`'s ports field. If you don't specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/ DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) "imageUri": "A String", # Required. Immutable. URI of the Docker image to be used as the custom container for serving predictions. This URI must identify an image in Artifact Registry or Container Registry. Learn more about the [container publishing requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#publishing), including permissions requirements for the Vertex AI Service Agent. The container image is ingested upon ModelService.UploadModel, stored internally, and this original path is afterwards not used. To learn about the requirements for the Docker image itself, see [Custom container requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#). You can use the URI to one of Vertex AI's [pre-built container images for prediction](https://cloud.google.com/vertex-ai/docs/predictions/pre-built-containers) in this field. - "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with [PredictionService.Invoke] eg: "/invoke/foo/bar". Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. + "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with invoke http call eg: "/invoke/foo/bar", however the [PredictionService.Invoke] RPC is not supported yet. Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. "livenessProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes liveness probe. "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. @@ -1160,7 +1160,7 @@

Method Details

}, "healthRoute": "A String", # Immutable. HTTP path on the container to send health checks to. Vertex AI intermittently sends GET requests to this path on the container's IP address and port to check that the container is healthy. Read more about [health checks](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#health). For example, if you set this field to `/bar`, then Vertex AI intermittently sends a GET request to the `/bar` path on the port of your container specified by the first value of this `ModelContainerSpec`'s ports field. If you don't specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/ DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) "imageUri": "A String", # Required. Immutable. URI of the Docker image to be used as the custom container for serving predictions. This URI must identify an image in Artifact Registry or Container Registry. Learn more about the [container publishing requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#publishing), including permissions requirements for the Vertex AI Service Agent. The container image is ingested upon ModelService.UploadModel, stored internally, and this original path is afterwards not used. To learn about the requirements for the Docker image itself, see [Custom container requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#). You can use the URI to one of Vertex AI's [pre-built container images for prediction](https://cloud.google.com/vertex-ai/docs/predictions/pre-built-containers) in this field. - "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with [PredictionService.Invoke] eg: "/invoke/foo/bar". Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. + "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with invoke http call eg: "/invoke/foo/bar", however the [PredictionService.Invoke] RPC is not supported yet. Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. "livenessProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes liveness probe. "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. @@ -1531,7 +1531,7 @@

Method Details

}, "healthRoute": "A String", # Immutable. HTTP path on the container to send health checks to. Vertex AI intermittently sends GET requests to this path on the container's IP address and port to check that the container is healthy. Read more about [health checks](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#health). For example, if you set this field to `/bar`, then Vertex AI intermittently sends a GET request to the `/bar` path on the port of your container specified by the first value of this `ModelContainerSpec`'s ports field. If you don't specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/ DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) "imageUri": "A String", # Required. Immutable. URI of the Docker image to be used as the custom container for serving predictions. This URI must identify an image in Artifact Registry or Container Registry. Learn more about the [container publishing requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#publishing), including permissions requirements for the Vertex AI Service Agent. The container image is ingested upon ModelService.UploadModel, stored internally, and this original path is afterwards not used. To learn about the requirements for the Docker image itself, see [Custom container requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#). You can use the URI to one of Vertex AI's [pre-built container images for prediction](https://cloud.google.com/vertex-ai/docs/predictions/pre-built-containers) in this field. - "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with [PredictionService.Invoke] eg: "/invoke/foo/bar". Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. + "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with invoke http call eg: "/invoke/foo/bar", however the [PredictionService.Invoke] RPC is not supported yet. Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. "livenessProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes liveness probe. "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. diff --git a/docs/dyn/aiplatform_v1.projects.locations.endpoints.html b/docs/dyn/aiplatform_v1.projects.locations.endpoints.html index 7d4ef9f8f3..b528ce9fb4 100644 --- a/docs/dyn/aiplatform_v1.projects.locations.endpoints.html +++ b/docs/dyn/aiplatform_v1.projects.locations.endpoints.html @@ -380,7 +380,7 @@

Method Details

"temperature": 3.14, # Optional. Controls the randomness of predictions. "thinkingConfig": { # Config for thinking features. # Optional. Config for thinking features. An error will be returned if this field is set for models that don't support thinking. "includeThoughts": True or False, # Optional. Indicates whether to include thoughts in the response. If true, thoughts are returned only when available. - "thinkingBudget": 42, # Optional. Indicates the thinking budget in tokens. This is only applied when enable_thinking is true. + "thinkingBudget": 42, # Optional. Indicates the thinking budget in tokens. }, "topK": 3.14, # Optional. If specified, top-k sampling will be used. "topP": 3.14, # Optional. If specified, nucleus sampling will be used. @@ -869,6 +869,16 @@

Method Details

"projectAllowlist": [ # A list of Projects from which the forwarding rule will target the service attachment. "A String", ], + "pscAutomationConfigs": [ # Optional. List of projects and networks where the PSC endpoints will be created. This field is used by Online Inference(Prediction) only. + { # PSC config that is used to automatically create PSC endpoints in the user projects. + "errorMessage": "A String", # Output only. Error message if the PSC service automation failed. + "forwardingRule": "A String", # Output only. Forwarding rule created by the PSC service automation. + "ipAddress": "A String", # Output only. IP address rule created by the PSC service automation. + "network": "A String", # Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/get): `projects/{project}/global/networks/{network}`. + "projectId": "A String", # Required. Project id used to create forwarding rule. + "state": "A String", # Output only. The state of the PSC service automation. + }, + ], "serviceAttachment": "A String", # Output only. The name of the generated service attachment resource. This is only populated if the endpoint is deployed with PrivateServiceConnect. }, "satisfiesPzi": True or False, # Output only. Reserved for future use. @@ -1700,7 +1710,7 @@

Method Details

"temperature": 3.14, # Optional. Controls the randomness of predictions. "thinkingConfig": { # Config for thinking features. # Optional. Config for thinking features. An error will be returned if this field is set for models that don't support thinking. "includeThoughts": True or False, # Optional. Indicates whether to include thoughts in the response. If true, thoughts are returned only when available. - "thinkingBudget": 42, # Optional. Indicates the thinking budget in tokens. This is only applied when enable_thinking is true. + "thinkingBudget": 42, # Optional. Indicates the thinking budget in tokens. }, "topK": 3.14, # Optional. If specified, top-k sampling will be used. "topP": 3.14, # Optional. If specified, nucleus sampling will be used. @@ -2414,6 +2424,16 @@

Method Details

"projectAllowlist": [ # A list of Projects from which the forwarding rule will target the service attachment. "A String", ], + "pscAutomationConfigs": [ # Optional. List of projects and networks where the PSC endpoints will be created. This field is used by Online Inference(Prediction) only. + { # PSC config that is used to automatically create PSC endpoints in the user projects. + "errorMessage": "A String", # Output only. Error message if the PSC service automation failed. + "forwardingRule": "A String", # Output only. Forwarding rule created by the PSC service automation. + "ipAddress": "A String", # Output only. IP address rule created by the PSC service automation. + "network": "A String", # Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/get): `projects/{project}/global/networks/{network}`. + "projectId": "A String", # Required. Project id used to create forwarding rule. + "state": "A String", # Output only. The state of the PSC service automation. + }, + ], "serviceAttachment": "A String", # Output only. The name of the generated service attachment resource. This is only populated if the endpoint is deployed with PrivateServiceConnect. }, "satisfiesPzi": True or False, # Output only. Reserved for future use. @@ -2661,6 +2681,16 @@

Method Details

"projectAllowlist": [ # A list of Projects from which the forwarding rule will target the service attachment. "A String", ], + "pscAutomationConfigs": [ # Optional. List of projects and networks where the PSC endpoints will be created. This field is used by Online Inference(Prediction) only. + { # PSC config that is used to automatically create PSC endpoints in the user projects. + "errorMessage": "A String", # Output only. Error message if the PSC service automation failed. + "forwardingRule": "A String", # Output only. Forwarding rule created by the PSC service automation. + "ipAddress": "A String", # Output only. IP address rule created by the PSC service automation. + "network": "A String", # Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/get): `projects/{project}/global/networks/{network}`. + "projectId": "A String", # Required. Project id used to create forwarding rule. + "state": "A String", # Output only. The state of the PSC service automation. + }, + ], "serviceAttachment": "A String", # Output only. The name of the generated service attachment resource. This is only populated if the endpoint is deployed with PrivateServiceConnect. }, "satisfiesPzi": True or False, # Output only. Reserved for future use. @@ -3131,6 +3161,16 @@

Method Details

"projectAllowlist": [ # A list of Projects from which the forwarding rule will target the service attachment. "A String", ], + "pscAutomationConfigs": [ # Optional. List of projects and networks where the PSC endpoints will be created. This field is used by Online Inference(Prediction) only. + { # PSC config that is used to automatically create PSC endpoints in the user projects. + "errorMessage": "A String", # Output only. Error message if the PSC service automation failed. + "forwardingRule": "A String", # Output only. Forwarding rule created by the PSC service automation. + "ipAddress": "A String", # Output only. IP address rule created by the PSC service automation. + "network": "A String", # Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/get): `projects/{project}/global/networks/{network}`. + "projectId": "A String", # Required. Project id used to create forwarding rule. + "state": "A String", # Output only. The state of the PSC service automation. + }, + ], "serviceAttachment": "A String", # Output only. The name of the generated service attachment resource. This is only populated if the endpoint is deployed with PrivateServiceConnect. }, "satisfiesPzi": True or False, # Output only. Reserved for future use. @@ -3365,6 +3405,16 @@

Method Details

"projectAllowlist": [ # A list of Projects from which the forwarding rule will target the service attachment. "A String", ], + "pscAutomationConfigs": [ # Optional. List of projects and networks where the PSC endpoints will be created. This field is used by Online Inference(Prediction) only. + { # PSC config that is used to automatically create PSC endpoints in the user projects. + "errorMessage": "A String", # Output only. Error message if the PSC service automation failed. + "forwardingRule": "A String", # Output only. Forwarding rule created by the PSC service automation. + "ipAddress": "A String", # Output only. IP address rule created by the PSC service automation. + "network": "A String", # Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/get): `projects/{project}/global/networks/{network}`. + "projectId": "A String", # Required. Project id used to create forwarding rule. + "state": "A String", # Output only. The state of the PSC service automation. + }, + ], "serviceAttachment": "A String", # Output only. The name of the generated service attachment resource. This is only populated if the endpoint is deployed with PrivateServiceConnect. }, "satisfiesPzi": True or False, # Output only. Reserved for future use. @@ -3819,7 +3869,7 @@

Method Details

"temperature": 3.14, # Optional. Controls the randomness of predictions. "thinkingConfig": { # Config for thinking features. # Optional. Config for thinking features. An error will be returned if this field is set for models that don't support thinking. "includeThoughts": True or False, # Optional. Indicates whether to include thoughts in the response. If true, thoughts are returned only when available. - "thinkingBudget": 42, # Optional. Indicates the thinking budget in tokens. This is only applied when enable_thinking is true. + "thinkingBudget": 42, # Optional. Indicates the thinking budget in tokens. }, "topK": 3.14, # Optional. If specified, top-k sampling will be used. "topP": 3.14, # Optional. If specified, nucleus sampling will be used. @@ -4614,6 +4664,16 @@

Method Details

"projectAllowlist": [ # A list of Projects from which the forwarding rule will target the service attachment. "A String", ], + "pscAutomationConfigs": [ # Optional. List of projects and networks where the PSC endpoints will be created. This field is used by Online Inference(Prediction) only. + { # PSC config that is used to automatically create PSC endpoints in the user projects. + "errorMessage": "A String", # Output only. Error message if the PSC service automation failed. + "forwardingRule": "A String", # Output only. Forwarding rule created by the PSC service automation. + "ipAddress": "A String", # Output only. IP address rule created by the PSC service automation. + "network": "A String", # Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/get): `projects/{project}/global/networks/{network}`. + "projectId": "A String", # Required. Project id used to create forwarding rule. + "state": "A String", # Output only. The state of the PSC service automation. + }, + ], "serviceAttachment": "A String", # Output only. The name of the generated service attachment resource. This is only populated if the endpoint is deployed with PrivateServiceConnect. }, "satisfiesPzi": True or False, # Output only. Reserved for future use. diff --git a/docs/dyn/aiplatform_v1.projects.locations.featureOnlineStores.featureViews.html b/docs/dyn/aiplatform_v1.projects.locations.featureOnlineStores.featureViews.html index 24e9357288..91d1fdf5d6 100644 --- a/docs/dyn/aiplatform_v1.projects.locations.featureOnlineStores.featureViews.html +++ b/docs/dyn/aiplatform_v1.projects.locations.featureOnlineStores.featureViews.html @@ -326,7 +326,7 @@

Method Details

}, "int64Value": "A String", # Int64 feature value. "metadata": { # Metadata of feature value. # Metadata of feature value. - "generateTime": "A String", # Feature generation timestamp. Typically, it is provided by user at feature ingestion time. If not, feature store will use the system timestamp when the data is ingested into feature store. For streaming ingestion, the time, aligned by days, must be no older than five years (1825 days) and no later than one year (366 days) in the future. + "generateTime": "A String", # Feature generation timestamp. Typically, it is provided by user at feature ingestion time. If not, feature store will use the system timestamp when the data is ingested into feature store. Legacy Feature Store: For streaming ingestion, the time, aligned by days, must be no older than five years (1825 days) and no later than one year (366 days) in the future. }, "stringArrayValue": { # A list of string values. # A list of string type feature value. "values": [ # A list of string values. @@ -751,7 +751,7 @@

Method Details

}, "int64Value": "A String", # Int64 feature value. "metadata": { # Metadata of feature value. # Metadata of feature value. - "generateTime": "A String", # Feature generation timestamp. Typically, it is provided by user at feature ingestion time. If not, feature store will use the system timestamp when the data is ingested into feature store. For streaming ingestion, the time, aligned by days, must be no older than five years (1825 days) and no later than one year (366 days) in the future. + "generateTime": "A String", # Feature generation timestamp. Typically, it is provided by user at feature ingestion time. If not, feature store will use the system timestamp when the data is ingested into feature store. Legacy Feature Store: For streaming ingestion, the time, aligned by days, must be no older than five years (1825 days) and no later than one year (366 days) in the future. }, "stringArrayValue": { # A list of string values. # A list of string type feature value. "values": [ # A list of string values. diff --git a/docs/dyn/aiplatform_v1.projects.locations.featureOnlineStores.html b/docs/dyn/aiplatform_v1.projects.locations.featureOnlineStores.html index e33c1343f1..381d2e0c5d 100644 --- a/docs/dyn/aiplatform_v1.projects.locations.featureOnlineStores.html +++ b/docs/dyn/aiplatform_v1.projects.locations.featureOnlineStores.html @@ -144,6 +144,16 @@

Method Details

"projectAllowlist": [ # A list of Projects from which the forwarding rule will target the service attachment. "A String", ], + "pscAutomationConfigs": [ # Optional. List of projects and networks where the PSC endpoints will be created. This field is used by Online Inference(Prediction) only. + { # PSC config that is used to automatically create PSC endpoints in the user projects. + "errorMessage": "A String", # Output only. Error message if the PSC service automation failed. + "forwardingRule": "A String", # Output only. Forwarding rule created by the PSC service automation. + "ipAddress": "A String", # Output only. IP address rule created by the PSC service automation. + "network": "A String", # Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/get): `projects/{project}/global/networks/{network}`. + "projectId": "A String", # Required. Project id used to create forwarding rule. + "state": "A String", # Output only. The state of the PSC service automation. + }, + ], "serviceAttachment": "A String", # Output only. The name of the generated service attachment resource. This is only populated if the endpoint is deployed with PrivateServiceConnect. }, "publicEndpointDomainName": "A String", # Output only. This field will be populated with the domain name to use for this FeatureOnlineStore @@ -260,6 +270,16 @@

Method Details

"projectAllowlist": [ # A list of Projects from which the forwarding rule will target the service attachment. "A String", ], + "pscAutomationConfigs": [ # Optional. List of projects and networks where the PSC endpoints will be created. This field is used by Online Inference(Prediction) only. + { # PSC config that is used to automatically create PSC endpoints in the user projects. + "errorMessage": "A String", # Output only. Error message if the PSC service automation failed. + "forwardingRule": "A String", # Output only. Forwarding rule created by the PSC service automation. + "ipAddress": "A String", # Output only. IP address rule created by the PSC service automation. + "network": "A String", # Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/get): `projects/{project}/global/networks/{network}`. + "projectId": "A String", # Required. Project id used to create forwarding rule. + "state": "A String", # Output only. The state of the PSC service automation. + }, + ], "serviceAttachment": "A String", # Output only. The name of the generated service attachment resource. This is only populated if the endpoint is deployed with PrivateServiceConnect. }, "publicEndpointDomainName": "A String", # Output only. This field will be populated with the domain name to use for this FeatureOnlineStore @@ -352,6 +372,16 @@

Method Details

"projectAllowlist": [ # A list of Projects from which the forwarding rule will target the service attachment. "A String", ], + "pscAutomationConfigs": [ # Optional. List of projects and networks where the PSC endpoints will be created. This field is used by Online Inference(Prediction) only. + { # PSC config that is used to automatically create PSC endpoints in the user projects. + "errorMessage": "A String", # Output only. Error message if the PSC service automation failed. + "forwardingRule": "A String", # Output only. Forwarding rule created by the PSC service automation. + "ipAddress": "A String", # Output only. IP address rule created by the PSC service automation. + "network": "A String", # Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/get): `projects/{project}/global/networks/{network}`. + "projectId": "A String", # Required. Project id used to create forwarding rule. + "state": "A String", # Output only. The state of the PSC service automation. + }, + ], "serviceAttachment": "A String", # Output only. The name of the generated service attachment resource. This is only populated if the endpoint is deployed with PrivateServiceConnect. }, "publicEndpointDomainName": "A String", # Output only. This field will be populated with the domain name to use for this FeatureOnlineStore @@ -415,6 +445,16 @@

Method Details

"projectAllowlist": [ # A list of Projects from which the forwarding rule will target the service attachment. "A String", ], + "pscAutomationConfigs": [ # Optional. List of projects and networks where the PSC endpoints will be created. This field is used by Online Inference(Prediction) only. + { # PSC config that is used to automatically create PSC endpoints in the user projects. + "errorMessage": "A String", # Output only. Error message if the PSC service automation failed. + "forwardingRule": "A String", # Output only. Forwarding rule created by the PSC service automation. + "ipAddress": "A String", # Output only. IP address rule created by the PSC service automation. + "network": "A String", # Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/get): `projects/{project}/global/networks/{network}`. + "projectId": "A String", # Required. Project id used to create forwarding rule. + "state": "A String", # Output only. The state of the PSC service automation. + }, + ], "serviceAttachment": "A String", # Output only. The name of the generated service attachment resource. This is only populated if the endpoint is deployed with PrivateServiceConnect. }, "publicEndpointDomainName": "A String", # Output only. This field will be populated with the domain name to use for this FeatureOnlineStore diff --git a/docs/dyn/aiplatform_v1.projects.locations.featurestores.entityTypes.html b/docs/dyn/aiplatform_v1.projects.locations.featurestores.entityTypes.html index 142e441043..9d85a4e014 100644 --- a/docs/dyn/aiplatform_v1.projects.locations.featurestores.entityTypes.html +++ b/docs/dyn/aiplatform_v1.projects.locations.featurestores.entityTypes.html @@ -743,7 +743,7 @@

Method Details

}, "int64Value": "A String", # Int64 feature value. "metadata": { # Metadata of feature value. # Metadata of feature value. - "generateTime": "A String", # Feature generation timestamp. Typically, it is provided by user at feature ingestion time. If not, feature store will use the system timestamp when the data is ingested into feature store. For streaming ingestion, the time, aligned by days, must be no older than five years (1825 days) and no later than one year (366 days) in the future. + "generateTime": "A String", # Feature generation timestamp. Typically, it is provided by user at feature ingestion time. If not, feature store will use the system timestamp when the data is ingested into feature store. Legacy Feature Store: For streaming ingestion, the time, aligned by days, must be no older than five years (1825 days) and no later than one year (366 days) in the future. }, "stringArrayValue": { # A list of string values. # A list of string type feature value. "values": [ # A list of string values. @@ -783,7 +783,7 @@

Method Details

}, "int64Value": "A String", # Int64 feature value. "metadata": { # Metadata of feature value. # Metadata of feature value. - "generateTime": "A String", # Feature generation timestamp. Typically, it is provided by user at feature ingestion time. If not, feature store will use the system timestamp when the data is ingested into feature store. For streaming ingestion, the time, aligned by days, must be no older than five years (1825 days) and no later than one year (366 days) in the future. + "generateTime": "A String", # Feature generation timestamp. Typically, it is provided by user at feature ingestion time. If not, feature store will use the system timestamp when the data is ingested into feature store. Legacy Feature Store: For streaming ingestion, the time, aligned by days, must be no older than five years (1825 days) and no later than one year (366 days) in the future. }, "stringArrayValue": { # A list of string values. # A list of string type feature value. "values": [ # A list of string values. @@ -930,7 +930,7 @@

Method Details

}, "int64Value": "A String", # Int64 feature value. "metadata": { # Metadata of feature value. # Metadata of feature value. - "generateTime": "A String", # Feature generation timestamp. Typically, it is provided by user at feature ingestion time. If not, feature store will use the system timestamp when the data is ingested into feature store. For streaming ingestion, the time, aligned by days, must be no older than five years (1825 days) and no later than one year (366 days) in the future. + "generateTime": "A String", # Feature generation timestamp. Typically, it is provided by user at feature ingestion time. If not, feature store will use the system timestamp when the data is ingested into feature store. Legacy Feature Store: For streaming ingestion, the time, aligned by days, must be no older than five years (1825 days) and no later than one year (366 days) in the future. }, "stringArrayValue": { # A list of string values. # A list of string type feature value. "values": [ # A list of string values. @@ -970,7 +970,7 @@

Method Details

}, "int64Value": "A String", # Int64 feature value. "metadata": { # Metadata of feature value. # Metadata of feature value. - "generateTime": "A String", # Feature generation timestamp. Typically, it is provided by user at feature ingestion time. If not, feature store will use the system timestamp when the data is ingested into feature store. For streaming ingestion, the time, aligned by days, must be no older than five years (1825 days) and no later than one year (366 days) in the future. + "generateTime": "A String", # Feature generation timestamp. Typically, it is provided by user at feature ingestion time. If not, feature store will use the system timestamp when the data is ingested into feature store. Legacy Feature Store: For streaming ingestion, the time, aligned by days, must be no older than five years (1825 days) and no later than one year (366 days) in the future. }, "stringArrayValue": { # A list of string values. # A list of string type feature value. "values": [ # A list of string values. @@ -1061,7 +1061,7 @@

Method Details

}, "int64Value": "A String", # Int64 feature value. "metadata": { # Metadata of feature value. # Metadata of feature value. - "generateTime": "A String", # Feature generation timestamp. Typically, it is provided by user at feature ingestion time. If not, feature store will use the system timestamp when the data is ingested into feature store. For streaming ingestion, the time, aligned by days, must be no older than five years (1825 days) and no later than one year (366 days) in the future. + "generateTime": "A String", # Feature generation timestamp. Typically, it is provided by user at feature ingestion time. If not, feature store will use the system timestamp when the data is ingested into feature store. Legacy Feature Store: For streaming ingestion, the time, aligned by days, must be no older than five years (1825 days) and no later than one year (366 days) in the future. }, "stringArrayValue": { # A list of string values. # A list of string type feature value. "values": [ # A list of string values. diff --git a/docs/dyn/aiplatform_v1.projects.locations.html b/docs/dyn/aiplatform_v1.projects.locations.html index b6568b6a95..d389991e21 100644 --- a/docs/dyn/aiplatform_v1.projects.locations.html +++ b/docs/dyn/aiplatform_v1.projects.locations.html @@ -651,7 +651,7 @@

Method Details

}, "healthRoute": "A String", # Immutable. HTTP path on the container to send health checks to. Vertex AI intermittently sends GET requests to this path on the container's IP address and port to check that the container is healthy. Read more about [health checks](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#health). For example, if you set this field to `/bar`, then Vertex AI intermittently sends a GET request to the `/bar` path on the port of your container specified by the first value of this `ModelContainerSpec`'s ports field. If you don't specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/ DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) "imageUri": "A String", # Required. Immutable. URI of the Docker image to be used as the custom container for serving predictions. This URI must identify an image in Artifact Registry or Container Registry. Learn more about the [container publishing requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#publishing), including permissions requirements for the Vertex AI Service Agent. The container image is ingested upon ModelService.UploadModel, stored internally, and this original path is afterwards not used. To learn about the requirements for the Docker image itself, see [Custom container requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#). You can use the URI to one of Vertex AI's [pre-built container images for prediction](https://cloud.google.com/vertex-ai/docs/predictions/pre-built-containers) in this field. - "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with [PredictionService.Invoke] eg: "/invoke/foo/bar". Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. + "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with invoke http call eg: "/invoke/foo/bar", however the [PredictionService.Invoke] RPC is not supported yet. Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. "livenessProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes liveness probe. "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. @@ -786,7 +786,7 @@

Method Details

}, }, "metrics": [ # Required. The metrics used for evaluation. - { # The metric used for dataset level evaluation. + { # The metric used for running evaluations. "aggregationMetrics": [ # Optional. The aggregation metrics to use. "A String", ], diff --git a/docs/dyn/aiplatform_v1.projects.locations.indexEndpoints.html b/docs/dyn/aiplatform_v1.projects.locations.indexEndpoints.html index e1a4a54037..ae2cd17f74 100644 --- a/docs/dyn/aiplatform_v1.projects.locations.indexEndpoints.html +++ b/docs/dyn/aiplatform_v1.projects.locations.indexEndpoints.html @@ -192,11 +192,11 @@

Method Details

"serviceAttachment": "A String", # Output only. The name of the service attachment resource. Populated if private service connect is enabled. }, "pscAutomationConfigs": [ # Optional. If set for PSC deployed index, PSC connection will be automatically created after deployment is done and the endpoint information is populated in private_endpoints.psc_automated_endpoints. - { # PSC config that is used to automatically create forwarding rule via ServiceConnectionMap. + { # PSC config that is used to automatically create PSC endpoints in the user projects. "errorMessage": "A String", # Output only. Error message if the PSC service automation failed. "forwardingRule": "A String", # Output only. Forwarding rule created by the PSC service automation. "ipAddress": "A String", # Output only. IP address rule created by the PSC service automation. - "network": "A String", # Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/insert): `projects/{project}/global/networks/{network}`. Where {project} is a project number, as in '12345', and {network} is network name. + "network": "A String", # Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/get): `projects/{project}/global/networks/{network}`. "projectId": "A String", # Required. Project id used to create forwarding rule. "state": "A String", # Output only. The state of the PSC service automation. }, @@ -223,6 +223,16 @@

Method Details

"projectAllowlist": [ # A list of Projects from which the forwarding rule will target the service attachment. "A String", ], + "pscAutomationConfigs": [ # Optional. List of projects and networks where the PSC endpoints will be created. This field is used by Online Inference(Prediction) only. + { # PSC config that is used to automatically create PSC endpoints in the user projects. + "errorMessage": "A String", # Output only. Error message if the PSC service automation failed. + "forwardingRule": "A String", # Output only. Forwarding rule created by the PSC service automation. + "ipAddress": "A String", # Output only. IP address rule created by the PSC service automation. + "network": "A String", # Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/get): `projects/{project}/global/networks/{network}`. + "projectId": "A String", # Required. Project id used to create forwarding rule. + "state": "A String", # Output only. The state of the PSC service automation. + }, + ], "serviceAttachment": "A String", # Output only. The name of the generated service attachment resource. This is only populated if the endpoint is deployed with PrivateServiceConnect. }, "publicEndpointDomainName": "A String", # Output only. If public_endpoint_enabled is true, this field will be populated with the domain name to use for this index endpoint. @@ -365,11 +375,11 @@

Method Details

"serviceAttachment": "A String", # Output only. The name of the service attachment resource. Populated if private service connect is enabled. }, "pscAutomationConfigs": [ # Optional. If set for PSC deployed index, PSC connection will be automatically created after deployment is done and the endpoint information is populated in private_endpoints.psc_automated_endpoints. - { # PSC config that is used to automatically create forwarding rule via ServiceConnectionMap. + { # PSC config that is used to automatically create PSC endpoints in the user projects. "errorMessage": "A String", # Output only. Error message if the PSC service automation failed. "forwardingRule": "A String", # Output only. Forwarding rule created by the PSC service automation. "ipAddress": "A String", # Output only. IP address rule created by the PSC service automation. - "network": "A String", # Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/insert): `projects/{project}/global/networks/{network}`. Where {project} is a project number, as in '12345', and {network} is network name. + "network": "A String", # Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/get): `projects/{project}/global/networks/{network}`. "projectId": "A String", # Required. Project id used to create forwarding rule. "state": "A String", # Output only. The state of the PSC service automation. }, @@ -607,11 +617,11 @@

Method Details

"serviceAttachment": "A String", # Output only. The name of the service attachment resource. Populated if private service connect is enabled. }, "pscAutomationConfigs": [ # Optional. If set for PSC deployed index, PSC connection will be automatically created after deployment is done and the endpoint information is populated in private_endpoints.psc_automated_endpoints. - { # PSC config that is used to automatically create forwarding rule via ServiceConnectionMap. + { # PSC config that is used to automatically create PSC endpoints in the user projects. "errorMessage": "A String", # Output only. Error message if the PSC service automation failed. "forwardingRule": "A String", # Output only. Forwarding rule created by the PSC service automation. "ipAddress": "A String", # Output only. IP address rule created by the PSC service automation. - "network": "A String", # Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/insert): `projects/{project}/global/networks/{network}`. Where {project} is a project number, as in '12345', and {network} is network name. + "network": "A String", # Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/get): `projects/{project}/global/networks/{network}`. "projectId": "A String", # Required. Project id used to create forwarding rule. "state": "A String", # Output only. The state of the PSC service automation. }, @@ -638,6 +648,16 @@

Method Details

"projectAllowlist": [ # A list of Projects from which the forwarding rule will target the service attachment. "A String", ], + "pscAutomationConfigs": [ # Optional. List of projects and networks where the PSC endpoints will be created. This field is used by Online Inference(Prediction) only. + { # PSC config that is used to automatically create PSC endpoints in the user projects. + "errorMessage": "A String", # Output only. Error message if the PSC service automation failed. + "forwardingRule": "A String", # Output only. Forwarding rule created by the PSC service automation. + "ipAddress": "A String", # Output only. IP address rule created by the PSC service automation. + "network": "A String", # Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/get): `projects/{project}/global/networks/{network}`. + "projectId": "A String", # Required. Project id used to create forwarding rule. + "state": "A String", # Output only. The state of the PSC service automation. + }, + ], "serviceAttachment": "A String", # Output only. The name of the generated service attachment resource. This is only populated if the endpoint is deployed with PrivateServiceConnect. }, "publicEndpointDomainName": "A String", # Output only. If public_endpoint_enabled is true, this field will be populated with the domain name to use for this index endpoint. @@ -730,11 +750,11 @@

Method Details

"serviceAttachment": "A String", # Output only. The name of the service attachment resource. Populated if private service connect is enabled. }, "pscAutomationConfigs": [ # Optional. If set for PSC deployed index, PSC connection will be automatically created after deployment is done and the endpoint information is populated in private_endpoints.psc_automated_endpoints. - { # PSC config that is used to automatically create forwarding rule via ServiceConnectionMap. + { # PSC config that is used to automatically create PSC endpoints in the user projects. "errorMessage": "A String", # Output only. Error message if the PSC service automation failed. "forwardingRule": "A String", # Output only. Forwarding rule created by the PSC service automation. "ipAddress": "A String", # Output only. IP address rule created by the PSC service automation. - "network": "A String", # Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/insert): `projects/{project}/global/networks/{network}`. Where {project} is a project number, as in '12345', and {network} is network name. + "network": "A String", # Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/get): `projects/{project}/global/networks/{network}`. "projectId": "A String", # Required. Project id used to create forwarding rule. "state": "A String", # Output only. The state of the PSC service automation. }, @@ -761,6 +781,16 @@

Method Details

"projectAllowlist": [ # A list of Projects from which the forwarding rule will target the service attachment. "A String", ], + "pscAutomationConfigs": [ # Optional. List of projects and networks where the PSC endpoints will be created. This field is used by Online Inference(Prediction) only. + { # PSC config that is used to automatically create PSC endpoints in the user projects. + "errorMessage": "A String", # Output only. Error message if the PSC service automation failed. + "forwardingRule": "A String", # Output only. Forwarding rule created by the PSC service automation. + "ipAddress": "A String", # Output only. IP address rule created by the PSC service automation. + "network": "A String", # Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/get): `projects/{project}/global/networks/{network}`. + "projectId": "A String", # Required. Project id used to create forwarding rule. + "state": "A String", # Output only. The state of the PSC service automation. + }, + ], "serviceAttachment": "A String", # Output only. The name of the generated service attachment resource. This is only populated if the endpoint is deployed with PrivateServiceConnect. }, "publicEndpointDomainName": "A String", # Output only. If public_endpoint_enabled is true, this field will be populated with the domain name to use for this index endpoint. @@ -856,11 +886,11 @@

Method Details

"serviceAttachment": "A String", # Output only. The name of the service attachment resource. Populated if private service connect is enabled. }, "pscAutomationConfigs": [ # Optional. If set for PSC deployed index, PSC connection will be automatically created after deployment is done and the endpoint information is populated in private_endpoints.psc_automated_endpoints. - { # PSC config that is used to automatically create forwarding rule via ServiceConnectionMap. + { # PSC config that is used to automatically create PSC endpoints in the user projects. "errorMessage": "A String", # Output only. Error message if the PSC service automation failed. "forwardingRule": "A String", # Output only. Forwarding rule created by the PSC service automation. "ipAddress": "A String", # Output only. IP address rule created by the PSC service automation. - "network": "A String", # Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/insert): `projects/{project}/global/networks/{network}`. Where {project} is a project number, as in '12345', and {network} is network name. + "network": "A String", # Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/get): `projects/{project}/global/networks/{network}`. "projectId": "A String", # Required. Project id used to create forwarding rule. "state": "A String", # Output only. The state of the PSC service automation. }, @@ -970,11 +1000,11 @@

Method Details

"serviceAttachment": "A String", # Output only. The name of the service attachment resource. Populated if private service connect is enabled. }, "pscAutomationConfigs": [ # Optional. If set for PSC deployed index, PSC connection will be automatically created after deployment is done and the endpoint information is populated in private_endpoints.psc_automated_endpoints. - { # PSC config that is used to automatically create forwarding rule via ServiceConnectionMap. + { # PSC config that is used to automatically create PSC endpoints in the user projects. "errorMessage": "A String", # Output only. Error message if the PSC service automation failed. "forwardingRule": "A String", # Output only. Forwarding rule created by the PSC service automation. "ipAddress": "A String", # Output only. IP address rule created by the PSC service automation. - "network": "A String", # Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/insert): `projects/{project}/global/networks/{network}`. Where {project} is a project number, as in '12345', and {network} is network name. + "network": "A String", # Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/get): `projects/{project}/global/networks/{network}`. "projectId": "A String", # Required. Project id used to create forwarding rule. "state": "A String", # Output only. The state of the PSC service automation. }, @@ -1001,6 +1031,16 @@

Method Details

"projectAllowlist": [ # A list of Projects from which the forwarding rule will target the service attachment. "A String", ], + "pscAutomationConfigs": [ # Optional. List of projects and networks where the PSC endpoints will be created. This field is used by Online Inference(Prediction) only. + { # PSC config that is used to automatically create PSC endpoints in the user projects. + "errorMessage": "A String", # Output only. Error message if the PSC service automation failed. + "forwardingRule": "A String", # Output only. Forwarding rule created by the PSC service automation. + "ipAddress": "A String", # Output only. IP address rule created by the PSC service automation. + "network": "A String", # Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/get): `projects/{project}/global/networks/{network}`. + "projectId": "A String", # Required. Project id used to create forwarding rule. + "state": "A String", # Output only. The state of the PSC service automation. + }, + ], "serviceAttachment": "A String", # Output only. The name of the generated service attachment resource. This is only populated if the endpoint is deployed with PrivateServiceConnect. }, "publicEndpointDomainName": "A String", # Output only. If public_endpoint_enabled is true, this field will be populated with the domain name to use for this index endpoint. @@ -1081,11 +1121,11 @@

Method Details

"serviceAttachment": "A String", # Output only. The name of the service attachment resource. Populated if private service connect is enabled. }, "pscAutomationConfigs": [ # Optional. If set for PSC deployed index, PSC connection will be automatically created after deployment is done and the endpoint information is populated in private_endpoints.psc_automated_endpoints. - { # PSC config that is used to automatically create forwarding rule via ServiceConnectionMap. + { # PSC config that is used to automatically create PSC endpoints in the user projects. "errorMessage": "A String", # Output only. Error message if the PSC service automation failed. "forwardingRule": "A String", # Output only. Forwarding rule created by the PSC service automation. "ipAddress": "A String", # Output only. IP address rule created by the PSC service automation. - "network": "A String", # Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/insert): `projects/{project}/global/networks/{network}`. Where {project} is a project number, as in '12345', and {network} is network name. + "network": "A String", # Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/get): `projects/{project}/global/networks/{network}`. "projectId": "A String", # Required. Project id used to create forwarding rule. "state": "A String", # Output only. The state of the PSC service automation. }, @@ -1112,6 +1152,16 @@

Method Details

"projectAllowlist": [ # A list of Projects from which the forwarding rule will target the service attachment. "A String", ], + "pscAutomationConfigs": [ # Optional. List of projects and networks where the PSC endpoints will be created. This field is used by Online Inference(Prediction) only. + { # PSC config that is used to automatically create PSC endpoints in the user projects. + "errorMessage": "A String", # Output only. Error message if the PSC service automation failed. + "forwardingRule": "A String", # Output only. Forwarding rule created by the PSC service automation. + "ipAddress": "A String", # Output only. IP address rule created by the PSC service automation. + "network": "A String", # Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/get): `projects/{project}/global/networks/{network}`. + "projectId": "A String", # Required. Project id used to create forwarding rule. + "state": "A String", # Output only. The state of the PSC service automation. + }, + ], "serviceAttachment": "A String", # Output only. The name of the generated service attachment resource. This is only populated if the endpoint is deployed with PrivateServiceConnect. }, "publicEndpointDomainName": "A String", # Output only. If public_endpoint_enabled is true, this field will be populated with the domain name to use for this index endpoint. diff --git a/docs/dyn/aiplatform_v1.projects.locations.models.html b/docs/dyn/aiplatform_v1.projects.locations.models.html index 47d7d85c92..35f7a54a22 100644 --- a/docs/dyn/aiplatform_v1.projects.locations.models.html +++ b/docs/dyn/aiplatform_v1.projects.locations.models.html @@ -400,7 +400,7 @@

Method Details

}, "healthRoute": "A String", # Immutable. HTTP path on the container to send health checks to. Vertex AI intermittently sends GET requests to this path on the container's IP address and port to check that the container is healthy. Read more about [health checks](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#health). For example, if you set this field to `/bar`, then Vertex AI intermittently sends a GET request to the `/bar` path on the port of your container specified by the first value of this `ModelContainerSpec`'s ports field. If you don't specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/ DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) "imageUri": "A String", # Required. Immutable. URI of the Docker image to be used as the custom container for serving predictions. This URI must identify an image in Artifact Registry or Container Registry. Learn more about the [container publishing requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#publishing), including permissions requirements for the Vertex AI Service Agent. The container image is ingested upon ModelService.UploadModel, stored internally, and this original path is afterwards not used. To learn about the requirements for the Docker image itself, see [Custom container requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#). You can use the URI to one of Vertex AI's [pre-built container images for prediction](https://cloud.google.com/vertex-ai/docs/predictions/pre-built-containers) in this field. - "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with [PredictionService.Invoke] eg: "/invoke/foo/bar". Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. + "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with invoke http call eg: "/invoke/foo/bar", however the [PredictionService.Invoke] RPC is not supported yet. Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. "livenessProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes liveness probe. "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. @@ -783,7 +783,7 @@

Method Details

}, "healthRoute": "A String", # Immutable. HTTP path on the container to send health checks to. Vertex AI intermittently sends GET requests to this path on the container's IP address and port to check that the container is healthy. Read more about [health checks](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#health). For example, if you set this field to `/bar`, then Vertex AI intermittently sends a GET request to the `/bar` path on the port of your container specified by the first value of this `ModelContainerSpec`'s ports field. If you don't specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/ DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) "imageUri": "A String", # Required. Immutable. URI of the Docker image to be used as the custom container for serving predictions. This URI must identify an image in Artifact Registry or Container Registry. Learn more about the [container publishing requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#publishing), including permissions requirements for the Vertex AI Service Agent. The container image is ingested upon ModelService.UploadModel, stored internally, and this original path is afterwards not used. To learn about the requirements for the Docker image itself, see [Custom container requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#). You can use the URI to one of Vertex AI's [pre-built container images for prediction](https://cloud.google.com/vertex-ai/docs/predictions/pre-built-containers) in this field. - "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with [PredictionService.Invoke] eg: "/invoke/foo/bar". Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. + "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with invoke http call eg: "/invoke/foo/bar", however the [PredictionService.Invoke] RPC is not supported yet. Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. "livenessProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes liveness probe. "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. @@ -1176,7 +1176,7 @@

Method Details

}, "healthRoute": "A String", # Immutable. HTTP path on the container to send health checks to. Vertex AI intermittently sends GET requests to this path on the container's IP address and port to check that the container is healthy. Read more about [health checks](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#health). For example, if you set this field to `/bar`, then Vertex AI intermittently sends a GET request to the `/bar` path on the port of your container specified by the first value of this `ModelContainerSpec`'s ports field. If you don't specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/ DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) "imageUri": "A String", # Required. Immutable. URI of the Docker image to be used as the custom container for serving predictions. This URI must identify an image in Artifact Registry or Container Registry. Learn more about the [container publishing requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#publishing), including permissions requirements for the Vertex AI Service Agent. The container image is ingested upon ModelService.UploadModel, stored internally, and this original path is afterwards not used. To learn about the requirements for the Docker image itself, see [Custom container requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#). You can use the URI to one of Vertex AI's [pre-built container images for prediction](https://cloud.google.com/vertex-ai/docs/predictions/pre-built-containers) in this field. - "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with [PredictionService.Invoke] eg: "/invoke/foo/bar". Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. + "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with invoke http call eg: "/invoke/foo/bar", however the [PredictionService.Invoke] RPC is not supported yet. Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. "livenessProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes liveness probe. "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. @@ -1557,7 +1557,7 @@

Method Details

}, "healthRoute": "A String", # Immutable. HTTP path on the container to send health checks to. Vertex AI intermittently sends GET requests to this path on the container's IP address and port to check that the container is healthy. Read more about [health checks](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#health). For example, if you set this field to `/bar`, then Vertex AI intermittently sends a GET request to the `/bar` path on the port of your container specified by the first value of this `ModelContainerSpec`'s ports field. If you don't specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/ DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) "imageUri": "A String", # Required. Immutable. URI of the Docker image to be used as the custom container for serving predictions. This URI must identify an image in Artifact Registry or Container Registry. Learn more about the [container publishing requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#publishing), including permissions requirements for the Vertex AI Service Agent. The container image is ingested upon ModelService.UploadModel, stored internally, and this original path is afterwards not used. To learn about the requirements for the Docker image itself, see [Custom container requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#). You can use the URI to one of Vertex AI's [pre-built container images for prediction](https://cloud.google.com/vertex-ai/docs/predictions/pre-built-containers) in this field. - "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with [PredictionService.Invoke] eg: "/invoke/foo/bar". Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. + "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with invoke http call eg: "/invoke/foo/bar", however the [PredictionService.Invoke] RPC is not supported yet. Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. "livenessProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes liveness probe. "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. @@ -1893,7 +1893,7 @@

Method Details

}, "healthRoute": "A String", # Immutable. HTTP path on the container to send health checks to. Vertex AI intermittently sends GET requests to this path on the container's IP address and port to check that the container is healthy. Read more about [health checks](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#health). For example, if you set this field to `/bar`, then Vertex AI intermittently sends a GET request to the `/bar` path on the port of your container specified by the first value of this `ModelContainerSpec`'s ports field. If you don't specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/ DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) "imageUri": "A String", # Required. Immutable. URI of the Docker image to be used as the custom container for serving predictions. This URI must identify an image in Artifact Registry or Container Registry. Learn more about the [container publishing requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#publishing), including permissions requirements for the Vertex AI Service Agent. The container image is ingested upon ModelService.UploadModel, stored internally, and this original path is afterwards not used. To learn about the requirements for the Docker image itself, see [Custom container requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#). You can use the URI to one of Vertex AI's [pre-built container images for prediction](https://cloud.google.com/vertex-ai/docs/predictions/pre-built-containers) in this field. - "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with [PredictionService.Invoke] eg: "/invoke/foo/bar". Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. + "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with invoke http call eg: "/invoke/foo/bar", however the [PredictionService.Invoke] RPC is not supported yet. Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. "livenessProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes liveness probe. "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. @@ -2228,7 +2228,7 @@

Method Details

}, "healthRoute": "A String", # Immutable. HTTP path on the container to send health checks to. Vertex AI intermittently sends GET requests to this path on the container's IP address and port to check that the container is healthy. Read more about [health checks](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#health). For example, if you set this field to `/bar`, then Vertex AI intermittently sends a GET request to the `/bar` path on the port of your container specified by the first value of this `ModelContainerSpec`'s ports field. If you don't specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/ DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) "imageUri": "A String", # Required. Immutable. URI of the Docker image to be used as the custom container for serving predictions. This URI must identify an image in Artifact Registry or Container Registry. Learn more about the [container publishing requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#publishing), including permissions requirements for the Vertex AI Service Agent. The container image is ingested upon ModelService.UploadModel, stored internally, and this original path is afterwards not used. To learn about the requirements for the Docker image itself, see [Custom container requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#). You can use the URI to one of Vertex AI's [pre-built container images for prediction](https://cloud.google.com/vertex-ai/docs/predictions/pre-built-containers) in this field. - "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with [PredictionService.Invoke] eg: "/invoke/foo/bar". Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. + "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with invoke http call eg: "/invoke/foo/bar", however the [PredictionService.Invoke] RPC is not supported yet. Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. "livenessProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes liveness probe. "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. @@ -2702,7 +2702,7 @@

Method Details

}, "healthRoute": "A String", # Immutable. HTTP path on the container to send health checks to. Vertex AI intermittently sends GET requests to this path on the container's IP address and port to check that the container is healthy. Read more about [health checks](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#health). For example, if you set this field to `/bar`, then Vertex AI intermittently sends a GET request to the `/bar` path on the port of your container specified by the first value of this `ModelContainerSpec`'s ports field. If you don't specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/ DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) "imageUri": "A String", # Required. Immutable. URI of the Docker image to be used as the custom container for serving predictions. This URI must identify an image in Artifact Registry or Container Registry. Learn more about the [container publishing requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#publishing), including permissions requirements for the Vertex AI Service Agent. The container image is ingested upon ModelService.UploadModel, stored internally, and this original path is afterwards not used. To learn about the requirements for the Docker image itself, see [Custom container requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#). You can use the URI to one of Vertex AI's [pre-built container images for prediction](https://cloud.google.com/vertex-ai/docs/predictions/pre-built-containers) in this field. - "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with [PredictionService.Invoke] eg: "/invoke/foo/bar". Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. + "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with invoke http call eg: "/invoke/foo/bar", however the [PredictionService.Invoke] RPC is not supported yet. Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. "livenessProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes liveness probe. "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. diff --git a/docs/dyn/aiplatform_v1.projects.locations.pipelineJobs.html b/docs/dyn/aiplatform_v1.projects.locations.pipelineJobs.html index d1013892c2..bdb185354a 100644 --- a/docs/dyn/aiplatform_v1.projects.locations.pipelineJobs.html +++ b/docs/dyn/aiplatform_v1.projects.locations.pipelineJobs.html @@ -404,6 +404,7 @@

Method Details

"state": "A String", # Output only. State of the task. "taskId": "A String", # Output only. The system generated ID of the task. "taskName": "A String", # Output only. The user specified name of the task that is defined in pipeline_spec. + "taskUniqueName": "A String", # Output only. The unique name of a task. This field is used by rerun pipeline job. Console UI and Vertex AI SDK will support triggering pipeline job reruns. The name is constructed by concatenating all the parent tasks name with the task name. For example, if a task named "child_task" has a parent task named "parent_task_1" and parent task 1 has a parent task named "parent_task_2", the task unique name will be "parent_task_2.parent_task_1.child_task". }, ], }, @@ -639,6 +640,7 @@

Method Details

"state": "A String", # Output only. State of the task. "taskId": "A String", # Output only. The system generated ID of the task. "taskName": "A String", # Output only. The user specified name of the task that is defined in pipeline_spec. + "taskUniqueName": "A String", # Output only. The unique name of a task. This field is used by rerun pipeline job. Console UI and Vertex AI SDK will support triggering pipeline job reruns. The name is constructed by concatenating all the parent tasks name with the task name. For example, if a task named "child_task" has a parent task named "parent_task_1" and parent task 1 has a parent task named "parent_task_2", the task unique name will be "parent_task_2.parent_task_1.child_task". }, ], }, @@ -915,6 +917,7 @@

Method Details

"state": "A String", # Output only. State of the task. "taskId": "A String", # Output only. The system generated ID of the task. "taskName": "A String", # Output only. The user specified name of the task that is defined in pipeline_spec. + "taskUniqueName": "A String", # Output only. The unique name of a task. This field is used by rerun pipeline job. Console UI and Vertex AI SDK will support triggering pipeline job reruns. The name is constructed by concatenating all the parent tasks name with the task name. For example, if a task named "child_task" has a parent task named "parent_task_1" and parent task 1 has a parent task named "parent_task_2", the task unique name will be "parent_task_2.parent_task_1.child_task". }, ], }, @@ -1164,6 +1167,7 @@

Method Details

"state": "A String", # Output only. State of the task. "taskId": "A String", # Output only. The system generated ID of the task. "taskName": "A String", # Output only. The user specified name of the task that is defined in pipeline_spec. + "taskUniqueName": "A String", # Output only. The unique name of a task. This field is used by rerun pipeline job. Console UI and Vertex AI SDK will support triggering pipeline job reruns. The name is constructed by concatenating all the parent tasks name with the task name. For example, if a task named "child_task" has a parent task named "parent_task_1" and parent task 1 has a parent task named "parent_task_2", the task unique name will be "parent_task_2.parent_task_1.child_task". }, ], }, diff --git a/docs/dyn/aiplatform_v1.projects.locations.publishers.models.html b/docs/dyn/aiplatform_v1.projects.locations.publishers.models.html index 46c7ef4008..87c07e5753 100644 --- a/docs/dyn/aiplatform_v1.projects.locations.publishers.models.html +++ b/docs/dyn/aiplatform_v1.projects.locations.publishers.models.html @@ -331,7 +331,7 @@

Method Details

"temperature": 3.14, # Optional. Controls the randomness of predictions. "thinkingConfig": { # Config for thinking features. # Optional. Config for thinking features. An error will be returned if this field is set for models that don't support thinking. "includeThoughts": True or False, # Optional. Indicates whether to include thoughts in the response. If true, thoughts are returned only when available. - "thinkingBudget": 42, # Optional. Indicates the thinking budget in tokens. This is only applied when enable_thinking is true. + "thinkingBudget": 42, # Optional. Indicates the thinking budget in tokens. }, "topK": 3.14, # Optional. If specified, top-k sampling will be used. "topP": 3.14, # Optional. If specified, nucleus sampling will be used. @@ -772,7 +772,7 @@

Method Details

"temperature": 3.14, # Optional. Controls the randomness of predictions. "thinkingConfig": { # Config for thinking features. # Optional. Config for thinking features. An error will be returned if this field is set for models that don't support thinking. "includeThoughts": True or False, # Optional. Indicates whether to include thoughts in the response. If true, thoughts are returned only when available. - "thinkingBudget": 42, # Optional. Indicates the thinking budget in tokens. This is only applied when enable_thinking is true. + "thinkingBudget": 42, # Optional. Indicates the thinking budget in tokens. }, "topK": 3.14, # Optional. If specified, top-k sampling will be used. "topP": 3.14, # Optional. If specified, nucleus sampling will be used. @@ -1700,7 +1700,7 @@

Method Details

"temperature": 3.14, # Optional. Controls the randomness of predictions. "thinkingConfig": { # Config for thinking features. # Optional. Config for thinking features. An error will be returned if this field is set for models that don't support thinking. "includeThoughts": True or False, # Optional. Indicates whether to include thoughts in the response. If true, thoughts are returned only when available. - "thinkingBudget": 42, # Optional. Indicates the thinking budget in tokens. This is only applied when enable_thinking is true. + "thinkingBudget": 42, # Optional. Indicates the thinking budget in tokens. }, "topK": 3.14, # Optional. If specified, top-k sampling will be used. "topP": 3.14, # Optional. If specified, nucleus sampling will be used. diff --git a/docs/dyn/aiplatform_v1.projects.locations.schedules.html b/docs/dyn/aiplatform_v1.projects.locations.schedules.html index f6c693decc..ea0a04f253 100644 --- a/docs/dyn/aiplatform_v1.projects.locations.schedules.html +++ b/docs/dyn/aiplatform_v1.projects.locations.schedules.html @@ -367,6 +367,7 @@

Method Details

"state": "A String", # Output only. State of the task. "taskId": "A String", # Output only. The system generated ID of the task. "taskName": "A String", # Output only. The user specified name of the task that is defined in pipeline_spec. + "taskUniqueName": "A String", # Output only. The unique name of a task. This field is used by rerun pipeline job. Console UI and Vertex AI SDK will support triggering pipeline job reruns. The name is constructed by concatenating all the parent tasks name with the task name. For example, if a task named "child_task" has a parent task named "parent_task_1" and parent task 1 has a parent task named "parent_task_2", the task unique name will be "parent_task_2.parent_task_1.child_task". }, ], }, @@ -697,6 +698,7 @@

Method Details

"state": "A String", # Output only. State of the task. "taskId": "A String", # Output only. The system generated ID of the task. "taskName": "A String", # Output only. The user specified name of the task that is defined in pipeline_spec. + "taskUniqueName": "A String", # Output only. The unique name of a task. This field is used by rerun pipeline job. Console UI and Vertex AI SDK will support triggering pipeline job reruns. The name is constructed by concatenating all the parent tasks name with the task name. For example, if a task named "child_task" has a parent task named "parent_task_1" and parent task 1 has a parent task named "parent_task_2", the task unique name will be "parent_task_2.parent_task_1.child_task". }, ], }, @@ -1069,6 +1071,7 @@

Method Details

"state": "A String", # Output only. State of the task. "taskId": "A String", # Output only. The system generated ID of the task. "taskName": "A String", # Output only. The user specified name of the task that is defined in pipeline_spec. + "taskUniqueName": "A String", # Output only. The unique name of a task. This field is used by rerun pipeline job. Console UI and Vertex AI SDK will support triggering pipeline job reruns. The name is constructed by concatenating all the parent tasks name with the task name. For example, if a task named "child_task" has a parent task named "parent_task_1" and parent task 1 has a parent task named "parent_task_2", the task unique name will be "parent_task_2.parent_task_1.child_task". }, ], }, @@ -1413,6 +1416,7 @@

Method Details

"state": "A String", # Output only. State of the task. "taskId": "A String", # Output only. The system generated ID of the task. "taskName": "A String", # Output only. The user specified name of the task that is defined in pipeline_spec. + "taskUniqueName": "A String", # Output only. The unique name of a task. This field is used by rerun pipeline job. Console UI and Vertex AI SDK will support triggering pipeline job reruns. The name is constructed by concatenating all the parent tasks name with the task name. For example, if a task named "child_task" has a parent task named "parent_task_1" and parent task 1 has a parent task named "parent_task_2", the task unique name will be "parent_task_2.parent_task_1.child_task". }, ], }, @@ -1761,6 +1765,7 @@

Method Details

"state": "A String", # Output only. State of the task. "taskId": "A String", # Output only. The system generated ID of the task. "taskName": "A String", # Output only. The user specified name of the task that is defined in pipeline_spec. + "taskUniqueName": "A String", # Output only. The unique name of a task. This field is used by rerun pipeline job. Console UI and Vertex AI SDK will support triggering pipeline job reruns. The name is constructed by concatenating all the parent tasks name with the task name. For example, if a task named "child_task" has a parent task named "parent_task_1" and parent task 1 has a parent task named "parent_task_2", the task unique name will be "parent_task_2.parent_task_1.child_task". }, ], }, @@ -2092,6 +2097,7 @@

Method Details

"state": "A String", # Output only. State of the task. "taskId": "A String", # Output only. The system generated ID of the task. "taskName": "A String", # Output only. The user specified name of the task that is defined in pipeline_spec. + "taskUniqueName": "A String", # Output only. The unique name of a task. This field is used by rerun pipeline job. Console UI and Vertex AI SDK will support triggering pipeline job reruns. The name is constructed by concatenating all the parent tasks name with the task name. For example, if a task named "child_task" has a parent task named "parent_task_1" and parent task 1 has a parent task named "parent_task_2", the task unique name will be "parent_task_2.parent_task_1.child_task". }, ], }, diff --git a/docs/dyn/aiplatform_v1.projects.locations.trainingPipelines.html b/docs/dyn/aiplatform_v1.projects.locations.trainingPipelines.html index 15adb0e289..87fec36d00 100644 --- a/docs/dyn/aiplatform_v1.projects.locations.trainingPipelines.html +++ b/docs/dyn/aiplatform_v1.projects.locations.trainingPipelines.html @@ -269,7 +269,7 @@

Method Details

}, "healthRoute": "A String", # Immutable. HTTP path on the container to send health checks to. Vertex AI intermittently sends GET requests to this path on the container's IP address and port to check that the container is healthy. Read more about [health checks](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#health). For example, if you set this field to `/bar`, then Vertex AI intermittently sends a GET request to the `/bar` path on the port of your container specified by the first value of this `ModelContainerSpec`'s ports field. If you don't specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/ DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) "imageUri": "A String", # Required. Immutable. URI of the Docker image to be used as the custom container for serving predictions. This URI must identify an image in Artifact Registry or Container Registry. Learn more about the [container publishing requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#publishing), including permissions requirements for the Vertex AI Service Agent. The container image is ingested upon ModelService.UploadModel, stored internally, and this original path is afterwards not used. To learn about the requirements for the Docker image itself, see [Custom container requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#). You can use the URI to one of Vertex AI's [pre-built container images for prediction](https://cloud.google.com/vertex-ai/docs/predictions/pre-built-containers) in this field. - "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with [PredictionService.Invoke] eg: "/invoke/foo/bar". Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. + "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with invoke http call eg: "/invoke/foo/bar", however the [PredictionService.Invoke] RPC is not supported yet. Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. "livenessProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes liveness probe. "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. @@ -670,7 +670,7 @@

Method Details

}, "healthRoute": "A String", # Immutable. HTTP path on the container to send health checks to. Vertex AI intermittently sends GET requests to this path on the container's IP address and port to check that the container is healthy. Read more about [health checks](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#health). For example, if you set this field to `/bar`, then Vertex AI intermittently sends a GET request to the `/bar` path on the port of your container specified by the first value of this `ModelContainerSpec`'s ports field. If you don't specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/ DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) "imageUri": "A String", # Required. Immutable. URI of the Docker image to be used as the custom container for serving predictions. This URI must identify an image in Artifact Registry or Container Registry. Learn more about the [container publishing requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#publishing), including permissions requirements for the Vertex AI Service Agent. The container image is ingested upon ModelService.UploadModel, stored internally, and this original path is afterwards not used. To learn about the requirements for the Docker image itself, see [Custom container requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#). You can use the URI to one of Vertex AI's [pre-built container images for prediction](https://cloud.google.com/vertex-ai/docs/predictions/pre-built-containers) in this field. - "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with [PredictionService.Invoke] eg: "/invoke/foo/bar". Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. + "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with invoke http call eg: "/invoke/foo/bar", however the [PredictionService.Invoke] RPC is not supported yet. Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. "livenessProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes liveness probe. "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. @@ -1113,7 +1113,7 @@

Method Details

}, "healthRoute": "A String", # Immutable. HTTP path on the container to send health checks to. Vertex AI intermittently sends GET requests to this path on the container's IP address and port to check that the container is healthy. Read more about [health checks](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#health). For example, if you set this field to `/bar`, then Vertex AI intermittently sends a GET request to the `/bar` path on the port of your container specified by the first value of this `ModelContainerSpec`'s ports field. If you don't specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/ DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) "imageUri": "A String", # Required. Immutable. URI of the Docker image to be used as the custom container for serving predictions. This URI must identify an image in Artifact Registry or Container Registry. Learn more about the [container publishing requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#publishing), including permissions requirements for the Vertex AI Service Agent. The container image is ingested upon ModelService.UploadModel, stored internally, and this original path is afterwards not used. To learn about the requirements for the Docker image itself, see [Custom container requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#). You can use the URI to one of Vertex AI's [pre-built container images for prediction](https://cloud.google.com/vertex-ai/docs/predictions/pre-built-containers) in this field. - "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with [PredictionService.Invoke] eg: "/invoke/foo/bar". Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. + "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with invoke http call eg: "/invoke/foo/bar", however the [PredictionService.Invoke] RPC is not supported yet. Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. "livenessProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes liveness probe. "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. @@ -1528,7 +1528,7 @@

Method Details

}, "healthRoute": "A String", # Immutable. HTTP path on the container to send health checks to. Vertex AI intermittently sends GET requests to this path on the container's IP address and port to check that the container is healthy. Read more about [health checks](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#health). For example, if you set this field to `/bar`, then Vertex AI intermittently sends a GET request to the `/bar` path on the port of your container specified by the first value of this `ModelContainerSpec`'s ports field. If you don't specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/ DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) "imageUri": "A String", # Required. Immutable. URI of the Docker image to be used as the custom container for serving predictions. This URI must identify an image in Artifact Registry or Container Registry. Learn more about the [container publishing requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#publishing), including permissions requirements for the Vertex AI Service Agent. The container image is ingested upon ModelService.UploadModel, stored internally, and this original path is afterwards not used. To learn about the requirements for the Docker image itself, see [Custom container requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#). You can use the URI to one of Vertex AI's [pre-built container images for prediction](https://cloud.google.com/vertex-ai/docs/predictions/pre-built-containers) in this field. - "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with [PredictionService.Invoke] eg: "/invoke/foo/bar". Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. + "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with invoke http call eg: "/invoke/foo/bar", however the [PredictionService.Invoke] RPC is not supported yet. Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. "livenessProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes liveness probe. "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. diff --git a/docs/dyn/aiplatform_v1.projects.locations.tuningJobs.html b/docs/dyn/aiplatform_v1.projects.locations.tuningJobs.html index f1f3f919ba..0b134d870a 100644 --- a/docs/dyn/aiplatform_v1.projects.locations.tuningJobs.html +++ b/docs/dyn/aiplatform_v1.projects.locations.tuningJobs.html @@ -184,7 +184,7 @@

Method Details

}, ], "endpoint": "A String", # Output only. A resource name of an Endpoint. Format: `projects/{project}/locations/{location}/endpoints/{endpoint}`. - "model": "A String", # Output only. The resource name of the TunedModel. Format: `projects/{project}/locations/{location}/models/{model}`. + "model": "A String", # Output only. The resource name of the TunedModel. Format: `projects/{project}/locations/{location}/models/{model}@{version_id}` When tuning from a base model, the version_id will be 1. For continuous tuning, the version id will be incremented by 1 from the last version id in the parent model. E.g., `projects/{project}/locations/{location}/models/{model}@{last_version_id + 1}` }, "tunedModelDisplayName": "A String", # Optional. The display name of the TunedModel. The name can be up to 128 characters long and can consist of any UTF-8 characters. "tuningDataStats": { # The tuning data statistic values for TuningJob. # Output only. The tuning data statistics associated with this TuningJob. @@ -356,7 +356,7 @@

Method Details

}, ], "endpoint": "A String", # Output only. A resource name of an Endpoint. Format: `projects/{project}/locations/{location}/endpoints/{endpoint}`. - "model": "A String", # Output only. The resource name of the TunedModel. Format: `projects/{project}/locations/{location}/models/{model}`. + "model": "A String", # Output only. The resource name of the TunedModel. Format: `projects/{project}/locations/{location}/models/{model}@{version_id}` When tuning from a base model, the version_id will be 1. For continuous tuning, the version id will be incremented by 1 from the last version id in the parent model. E.g., `projects/{project}/locations/{location}/models/{model}@{last_version_id + 1}` }, "tunedModelDisplayName": "A String", # Optional. The display name of the TunedModel. The name can be up to 128 characters long and can consist of any UTF-8 characters. "tuningDataStats": { # The tuning data statistic values for TuningJob. # Output only. The tuning data statistics associated with this TuningJob. @@ -535,7 +535,7 @@

Method Details

}, ], "endpoint": "A String", # Output only. A resource name of an Endpoint. Format: `projects/{project}/locations/{location}/endpoints/{endpoint}`. - "model": "A String", # Output only. The resource name of the TunedModel. Format: `projects/{project}/locations/{location}/models/{model}`. + "model": "A String", # Output only. The resource name of the TunedModel. Format: `projects/{project}/locations/{location}/models/{model}@{version_id}` When tuning from a base model, the version_id will be 1. For continuous tuning, the version id will be incremented by 1 from the last version id in the parent model. E.g., `projects/{project}/locations/{location}/models/{model}@{last_version_id + 1}` }, "tunedModelDisplayName": "A String", # Optional. The display name of the TunedModel. The name can be up to 128 characters long and can consist of any UTF-8 characters. "tuningDataStats": { # The tuning data statistic values for TuningJob. # Output only. The tuning data statistics associated with this TuningJob. @@ -720,7 +720,7 @@

Method Details

}, ], "endpoint": "A String", # Output only. A resource name of an Endpoint. Format: `projects/{project}/locations/{location}/endpoints/{endpoint}`. - "model": "A String", # Output only. The resource name of the TunedModel. Format: `projects/{project}/locations/{location}/models/{model}`. + "model": "A String", # Output only. The resource name of the TunedModel. Format: `projects/{project}/locations/{location}/models/{model}@{version_id}` When tuning from a base model, the version_id will be 1. For continuous tuning, the version id will be incremented by 1 from the last version id in the parent model. E.g., `projects/{project}/locations/{location}/models/{model}@{last_version_id + 1}` }, "tunedModelDisplayName": "A String", # Optional. The display name of the TunedModel. The name can be up to 128 characters long and can consist of any UTF-8 characters. "tuningDataStats": { # The tuning data statistic values for TuningJob. # Output only. The tuning data statistics associated with this TuningJob. @@ -920,7 +920,7 @@

Method Details

}, ], "endpoint": "A String", # Output only. A resource name of an Endpoint. Format: `projects/{project}/locations/{location}/endpoints/{endpoint}`. - "model": "A String", # Output only. The resource name of the TunedModel. Format: `projects/{project}/locations/{location}/models/{model}`. + "model": "A String", # Output only. The resource name of the TunedModel. Format: `projects/{project}/locations/{location}/models/{model}@{version_id}` When tuning from a base model, the version_id will be 1. For continuous tuning, the version id will be incremented by 1 from the last version id in the parent model. E.g., `projects/{project}/locations/{location}/models/{model}@{last_version_id + 1}` }, "tunedModelDisplayName": "A String", # Optional. The display name of the TunedModel. The name can be up to 128 characters long and can consist of any UTF-8 characters. "tuningDataStats": { # The tuning data statistic values for TuningJob. # Output only. The tuning data statistics associated with this TuningJob. diff --git a/docs/dyn/aiplatform_v1.publishers.models.html b/docs/dyn/aiplatform_v1.publishers.models.html index cab9b45c4a..8707dcf179 100644 --- a/docs/dyn/aiplatform_v1.publishers.models.html +++ b/docs/dyn/aiplatform_v1.publishers.models.html @@ -325,7 +325,7 @@

Method Details

"temperature": 3.14, # Optional. Controls the randomness of predictions. "thinkingConfig": { # Config for thinking features. # Optional. Config for thinking features. An error will be returned if this field is set for models that don't support thinking. "includeThoughts": True or False, # Optional. Indicates whether to include thoughts in the response. If true, thoughts are returned only when available. - "thinkingBudget": 42, # Optional. Indicates the thinking budget in tokens. This is only applied when enable_thinking is true. + "thinkingBudget": 42, # Optional. Indicates the thinking budget in tokens. }, "topK": 3.14, # Optional. If specified, top-k sampling will be used. "topP": 3.14, # Optional. If specified, nucleus sampling will be used. @@ -766,7 +766,7 @@

Method Details

"temperature": 3.14, # Optional. Controls the randomness of predictions. "thinkingConfig": { # Config for thinking features. # Optional. Config for thinking features. An error will be returned if this field is set for models that don't support thinking. "includeThoughts": True or False, # Optional. Indicates whether to include thoughts in the response. If true, thoughts are returned only when available. - "thinkingBudget": 42, # Optional. Indicates the thinking budget in tokens. This is only applied when enable_thinking is true. + "thinkingBudget": 42, # Optional. Indicates the thinking budget in tokens. }, "topK": 3.14, # Optional. If specified, top-k sampling will be used. "topP": 3.14, # Optional. If specified, nucleus sampling will be used. @@ -1362,7 +1362,7 @@

Method Details

}, "healthRoute": "A String", # Immutable. HTTP path on the container to send health checks to. Vertex AI intermittently sends GET requests to this path on the container's IP address and port to check that the container is healthy. Read more about [health checks](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#health). For example, if you set this field to `/bar`, then Vertex AI intermittently sends a GET request to the `/bar` path on the port of your container specified by the first value of this `ModelContainerSpec`'s ports field. If you don't specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/ DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) "imageUri": "A String", # Required. Immutable. URI of the Docker image to be used as the custom container for serving predictions. This URI must identify an image in Artifact Registry or Container Registry. Learn more about the [container publishing requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#publishing), including permissions requirements for the Vertex AI Service Agent. The container image is ingested upon ModelService.UploadModel, stored internally, and this original path is afterwards not used. To learn about the requirements for the Docker image itself, see [Custom container requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#). You can use the URI to one of Vertex AI's [pre-built container images for prediction](https://cloud.google.com/vertex-ai/docs/predictions/pre-built-containers) in this field. - "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with [PredictionService.Invoke] eg: "/invoke/foo/bar". Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. + "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with invoke http call eg: "/invoke/foo/bar", however the [PredictionService.Invoke] RPC is not supported yet. Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. "livenessProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes liveness probe. "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. @@ -1541,7 +1541,7 @@

Method Details

}, "healthRoute": "A String", # Immutable. HTTP path on the container to send health checks to. Vertex AI intermittently sends GET requests to this path on the container's IP address and port to check that the container is healthy. Read more about [health checks](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#health). For example, if you set this field to `/bar`, then Vertex AI intermittently sends a GET request to the `/bar` path on the port of your container specified by the first value of this `ModelContainerSpec`'s ports field. If you don't specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/ DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) "imageUri": "A String", # Required. Immutable. URI of the Docker image to be used as the custom container for serving predictions. This URI must identify an image in Artifact Registry or Container Registry. Learn more about the [container publishing requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#publishing), including permissions requirements for the Vertex AI Service Agent. The container image is ingested upon ModelService.UploadModel, stored internally, and this original path is afterwards not used. To learn about the requirements for the Docker image itself, see [Custom container requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#). You can use the URI to one of Vertex AI's [pre-built container images for prediction](https://cloud.google.com/vertex-ai/docs/predictions/pre-built-containers) in this field. - "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with [PredictionService.Invoke] eg: "/invoke/foo/bar". Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. + "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with invoke http call eg: "/invoke/foo/bar", however the [PredictionService.Invoke] RPC is not supported yet. Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. "livenessProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes liveness probe. "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. @@ -2029,7 +2029,7 @@

Method Details

"temperature": 3.14, # Optional. Controls the randomness of predictions. "thinkingConfig": { # Config for thinking features. # Optional. Config for thinking features. An error will be returned if this field is set for models that don't support thinking. "includeThoughts": True or False, # Optional. Indicates whether to include thoughts in the response. If true, thoughts are returned only when available. - "thinkingBudget": 42, # Optional. Indicates the thinking budget in tokens. This is only applied when enable_thinking is true. + "thinkingBudget": 42, # Optional. Indicates the thinking budget in tokens. }, "topK": 3.14, # Optional. If specified, top-k sampling will be used. "topP": 3.14, # Optional. If specified, nucleus sampling will be used. diff --git a/docs/dyn/aiplatform_v1beta1.batchPredictionJobs.html b/docs/dyn/aiplatform_v1beta1.batchPredictionJobs.html index 608cf0daf6..80aa559939 100644 --- a/docs/dyn/aiplatform_v1beta1.batchPredictionJobs.html +++ b/docs/dyn/aiplatform_v1beta1.batchPredictionJobs.html @@ -502,7 +502,7 @@

Method Details

}, "healthRoute": "A String", # Immutable. HTTP path on the container to send health checks to. Vertex AI intermittently sends GET requests to this path on the container's IP address and port to check that the container is healthy. Read more about [health checks](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#health). For example, if you set this field to `/bar`, then Vertex AI intermittently sends a GET request to the `/bar` path on the port of your container specified by the first value of this `ModelContainerSpec`'s ports field. If you don't specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/ DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) "imageUri": "A String", # Required. Immutable. URI of the Docker image to be used as the custom container for serving predictions. This URI must identify an image in Artifact Registry or Container Registry. Learn more about the [container publishing requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#publishing), including permissions requirements for the Vertex AI Service Agent. The container image is ingested upon ModelService.UploadModel, stored internally, and this original path is afterwards not used. To learn about the requirements for the Docker image itself, see [Custom container requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#). You can use the URI to one of Vertex AI's [pre-built container images for prediction](https://cloud.google.com/vertex-ai/docs/predictions/pre-built-containers) in this field. - "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with [PredictionService.Invoke] eg: "/invoke/foo/bar". Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. + "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with invoke http call eg: "/invoke/foo/bar", however the [PredictionService.Invoke] RPC is not supported yet. Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. "livenessProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes liveness probe. "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. @@ -992,7 +992,7 @@

Method Details

}, "healthRoute": "A String", # Immutable. HTTP path on the container to send health checks to. Vertex AI intermittently sends GET requests to this path on the container's IP address and port to check that the container is healthy. Read more about [health checks](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#health). For example, if you set this field to `/bar`, then Vertex AI intermittently sends a GET request to the `/bar` path on the port of your container specified by the first value of this `ModelContainerSpec`'s ports field. If you don't specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/ DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) "imageUri": "A String", # Required. Immutable. URI of the Docker image to be used as the custom container for serving predictions. This URI must identify an image in Artifact Registry or Container Registry. Learn more about the [container publishing requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#publishing), including permissions requirements for the Vertex AI Service Agent. The container image is ingested upon ModelService.UploadModel, stored internally, and this original path is afterwards not used. To learn about the requirements for the Docker image itself, see [Custom container requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#). You can use the URI to one of Vertex AI's [pre-built container images for prediction](https://cloud.google.com/vertex-ai/docs/predictions/pre-built-containers) in this field. - "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with [PredictionService.Invoke] eg: "/invoke/foo/bar". Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. + "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with invoke http call eg: "/invoke/foo/bar", however the [PredictionService.Invoke] RPC is not supported yet. Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. "livenessProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes liveness probe. "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. @@ -1488,7 +1488,7 @@

Method Details

}, "healthRoute": "A String", # Immutable. HTTP path on the container to send health checks to. Vertex AI intermittently sends GET requests to this path on the container's IP address and port to check that the container is healthy. Read more about [health checks](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#health). For example, if you set this field to `/bar`, then Vertex AI intermittently sends a GET request to the `/bar` path on the port of your container specified by the first value of this `ModelContainerSpec`'s ports field. If you don't specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/ DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) "imageUri": "A String", # Required. Immutable. URI of the Docker image to be used as the custom container for serving predictions. This URI must identify an image in Artifact Registry or Container Registry. Learn more about the [container publishing requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#publishing), including permissions requirements for the Vertex AI Service Agent. The container image is ingested upon ModelService.UploadModel, stored internally, and this original path is afterwards not used. To learn about the requirements for the Docker image itself, see [Custom container requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#). You can use the URI to one of Vertex AI's [pre-built container images for prediction](https://cloud.google.com/vertex-ai/docs/predictions/pre-built-containers) in this field. - "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with [PredictionService.Invoke] eg: "/invoke/foo/bar". Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. + "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with invoke http call eg: "/invoke/foo/bar", however the [PredictionService.Invoke] RPC is not supported yet. Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. "livenessProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes liveness probe. "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. @@ -1990,7 +1990,7 @@

Method Details

}, "healthRoute": "A String", # Immutable. HTTP path on the container to send health checks to. Vertex AI intermittently sends GET requests to this path on the container's IP address and port to check that the container is healthy. Read more about [health checks](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#health). For example, if you set this field to `/bar`, then Vertex AI intermittently sends a GET request to the `/bar` path on the port of your container specified by the first value of this `ModelContainerSpec`'s ports field. If you don't specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/ DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) "imageUri": "A String", # Required. Immutable. URI of the Docker image to be used as the custom container for serving predictions. This URI must identify an image in Artifact Registry or Container Registry. Learn more about the [container publishing requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#publishing), including permissions requirements for the Vertex AI Service Agent. The container image is ingested upon ModelService.UploadModel, stored internally, and this original path is afterwards not used. To learn about the requirements for the Docker image itself, see [Custom container requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#). You can use the URI to one of Vertex AI's [pre-built container images for prediction](https://cloud.google.com/vertex-ai/docs/predictions/pre-built-containers) in this field. - "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with [PredictionService.Invoke] eg: "/invoke/foo/bar". Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. + "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with invoke http call eg: "/invoke/foo/bar", however the [PredictionService.Invoke] RPC is not supported yet. Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. "livenessProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes liveness probe. "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. diff --git a/docs/dyn/aiplatform_v1beta1.endpoints.html b/docs/dyn/aiplatform_v1beta1.endpoints.html index e6d0905ff8..247a3cc576 100644 --- a/docs/dyn/aiplatform_v1beta1.endpoints.html +++ b/docs/dyn/aiplatform_v1beta1.endpoints.html @@ -334,7 +334,7 @@

Method Details

"temperature": 3.14, # Optional. Controls the randomness of predictions. "thinkingConfig": { # Config for thinking features. # Optional. Config for thinking features. An error will be returned if this field is set for models that don't support thinking. "includeThoughts": True or False, # Optional. Indicates whether to include thoughts in the response. If true, thoughts are returned only when available. - "thinkingBudget": 42, # Optional. Indicates the thinking budget in tokens. This is only applied when enable_thinking is true. + "thinkingBudget": 42, # Optional. Indicates the thinking budget in tokens. }, "topK": 3.14, # Optional. If specified, top-k sampling will be used. "topP": 3.14, # Optional. If specified, nucleus sampling will be used. @@ -789,7 +789,7 @@

Method Details

"temperature": 3.14, # Optional. Controls the randomness of predictions. "thinkingConfig": { # Config for thinking features. # Optional. Config for thinking features. An error will be returned if this field is set for models that don't support thinking. "includeThoughts": True or False, # Optional. Indicates whether to include thoughts in the response. If true, thoughts are returned only when available. - "thinkingBudget": 42, # Optional. Indicates the thinking budget in tokens. This is only applied when enable_thinking is true. + "thinkingBudget": 42, # Optional. Indicates the thinking budget in tokens. }, "topK": 3.14, # Optional. If specified, top-k sampling will be used. "topP": 3.14, # Optional. If specified, nucleus sampling will be used. @@ -1508,7 +1508,7 @@

Method Details

"temperature": 3.14, # Optional. Controls the randomness of predictions. "thinkingConfig": { # Config for thinking features. # Optional. Config for thinking features. An error will be returned if this field is set for models that don't support thinking. "includeThoughts": True or False, # Optional. Indicates whether to include thoughts in the response. If true, thoughts are returned only when available. - "thinkingBudget": 42, # Optional. Indicates the thinking budget in tokens. This is only applied when enable_thinking is true. + "thinkingBudget": 42, # Optional. Indicates the thinking budget in tokens. }, "topK": 3.14, # Optional. If specified, top-k sampling will be used. "topP": 3.14, # Optional. If specified, nucleus sampling will be used. diff --git a/docs/dyn/aiplatform_v1beta1.projects.locations.batchPredictionJobs.html b/docs/dyn/aiplatform_v1beta1.projects.locations.batchPredictionJobs.html index 6f2a76f6a0..8f3529da83 100644 --- a/docs/dyn/aiplatform_v1beta1.projects.locations.batchPredictionJobs.html +++ b/docs/dyn/aiplatform_v1beta1.projects.locations.batchPredictionJobs.html @@ -533,7 +533,7 @@

Method Details

}, "healthRoute": "A String", # Immutable. HTTP path on the container to send health checks to. Vertex AI intermittently sends GET requests to this path on the container's IP address and port to check that the container is healthy. Read more about [health checks](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#health). For example, if you set this field to `/bar`, then Vertex AI intermittently sends a GET request to the `/bar` path on the port of your container specified by the first value of this `ModelContainerSpec`'s ports field. If you don't specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/ DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) "imageUri": "A String", # Required. Immutable. URI of the Docker image to be used as the custom container for serving predictions. This URI must identify an image in Artifact Registry or Container Registry. Learn more about the [container publishing requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#publishing), including permissions requirements for the Vertex AI Service Agent. The container image is ingested upon ModelService.UploadModel, stored internally, and this original path is afterwards not used. To learn about the requirements for the Docker image itself, see [Custom container requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#). You can use the URI to one of Vertex AI's [pre-built container images for prediction](https://cloud.google.com/vertex-ai/docs/predictions/pre-built-containers) in this field. - "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with [PredictionService.Invoke] eg: "/invoke/foo/bar". Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. + "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with invoke http call eg: "/invoke/foo/bar", however the [PredictionService.Invoke] RPC is not supported yet. Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. "livenessProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes liveness probe. "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. @@ -1022,7 +1022,7 @@

Method Details

}, "healthRoute": "A String", # Immutable. HTTP path on the container to send health checks to. Vertex AI intermittently sends GET requests to this path on the container's IP address and port to check that the container is healthy. Read more about [health checks](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#health). For example, if you set this field to `/bar`, then Vertex AI intermittently sends a GET request to the `/bar` path on the port of your container specified by the first value of this `ModelContainerSpec`'s ports field. If you don't specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/ DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) "imageUri": "A String", # Required. Immutable. URI of the Docker image to be used as the custom container for serving predictions. This URI must identify an image in Artifact Registry or Container Registry. Learn more about the [container publishing requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#publishing), including permissions requirements for the Vertex AI Service Agent. The container image is ingested upon ModelService.UploadModel, stored internally, and this original path is afterwards not used. To learn about the requirements for the Docker image itself, see [Custom container requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#). You can use the URI to one of Vertex AI's [pre-built container images for prediction](https://cloud.google.com/vertex-ai/docs/predictions/pre-built-containers) in this field. - "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with [PredictionService.Invoke] eg: "/invoke/foo/bar". Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. + "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with invoke http call eg: "/invoke/foo/bar", however the [PredictionService.Invoke] RPC is not supported yet. Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. "livenessProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes liveness probe. "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. @@ -1553,7 +1553,7 @@

Method Details

}, "healthRoute": "A String", # Immutable. HTTP path on the container to send health checks to. Vertex AI intermittently sends GET requests to this path on the container's IP address and port to check that the container is healthy. Read more about [health checks](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#health). For example, if you set this field to `/bar`, then Vertex AI intermittently sends a GET request to the `/bar` path on the port of your container specified by the first value of this `ModelContainerSpec`'s ports field. If you don't specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/ DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) "imageUri": "A String", # Required. Immutable. URI of the Docker image to be used as the custom container for serving predictions. This URI must identify an image in Artifact Registry or Container Registry. Learn more about the [container publishing requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#publishing), including permissions requirements for the Vertex AI Service Agent. The container image is ingested upon ModelService.UploadModel, stored internally, and this original path is afterwards not used. To learn about the requirements for the Docker image itself, see [Custom container requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#). You can use the URI to one of Vertex AI's [pre-built container images for prediction](https://cloud.google.com/vertex-ai/docs/predictions/pre-built-containers) in this field. - "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with [PredictionService.Invoke] eg: "/invoke/foo/bar". Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. + "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with invoke http call eg: "/invoke/foo/bar", however the [PredictionService.Invoke] RPC is not supported yet. Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. "livenessProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes liveness probe. "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. @@ -2055,7 +2055,7 @@

Method Details

}, "healthRoute": "A String", # Immutable. HTTP path on the container to send health checks to. Vertex AI intermittently sends GET requests to this path on the container's IP address and port to check that the container is healthy. Read more about [health checks](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#health). For example, if you set this field to `/bar`, then Vertex AI intermittently sends a GET request to the `/bar` path on the port of your container specified by the first value of this `ModelContainerSpec`'s ports field. If you don't specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/ DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) "imageUri": "A String", # Required. Immutable. URI of the Docker image to be used as the custom container for serving predictions. This URI must identify an image in Artifact Registry or Container Registry. Learn more about the [container publishing requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#publishing), including permissions requirements for the Vertex AI Service Agent. The container image is ingested upon ModelService.UploadModel, stored internally, and this original path is afterwards not used. To learn about the requirements for the Docker image itself, see [Custom container requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#). You can use the URI to one of Vertex AI's [pre-built container images for prediction](https://cloud.google.com/vertex-ai/docs/predictions/pre-built-containers) in this field. - "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with [PredictionService.Invoke] eg: "/invoke/foo/bar". Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. + "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with invoke http call eg: "/invoke/foo/bar", however the [PredictionService.Invoke] RPC is not supported yet. Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. "livenessProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes liveness probe. "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. diff --git a/docs/dyn/aiplatform_v1beta1.projects.locations.datasets.html b/docs/dyn/aiplatform_v1beta1.projects.locations.datasets.html index 3ddbc563ba..998b736936 100644 --- a/docs/dyn/aiplatform_v1beta1.projects.locations.datasets.html +++ b/docs/dyn/aiplatform_v1beta1.projects.locations.datasets.html @@ -285,7 +285,7 @@

Method Details

"temperature": 3.14, # Optional. Controls the randomness of predictions. "thinkingConfig": { # Config for thinking features. # Optional. Config for thinking features. An error will be returned if this field is set for models that don't support thinking. "includeThoughts": True or False, # Optional. Indicates whether to include thoughts in the response. If true, thoughts are returned only when available. - "thinkingBudget": 42, # Optional. Indicates the thinking budget in tokens. This is only applied when enable_thinking is true. + "thinkingBudget": 42, # Optional. Indicates the thinking budget in tokens. }, "topK": 3.14, # Optional. If specified, top-k sampling will be used. "topP": 3.14, # Optional. If specified, nucleus sampling will be used. @@ -696,7 +696,7 @@

Method Details

"temperature": 3.14, # Optional. Controls the randomness of predictions. "thinkingConfig": { # Config for thinking features. # Optional. Config for thinking features. An error will be returned if this field is set for models that don't support thinking. "includeThoughts": True or False, # Optional. Indicates whether to include thoughts in the response. If true, thoughts are returned only when available. - "thinkingBudget": 42, # Optional. Indicates the thinking budget in tokens. This is only applied when enable_thinking is true. + "thinkingBudget": 42, # Optional. Indicates the thinking budget in tokens. }, "topK": 3.14, # Optional. If specified, top-k sampling will be used. "topP": 3.14, # Optional. If specified, nucleus sampling will be used. @@ -1156,7 +1156,7 @@

Method Details

"temperature": 3.14, # Optional. Controls the randomness of predictions. "thinkingConfig": { # Config for thinking features. # Optional. Config for thinking features. An error will be returned if this field is set for models that don't support thinking. "includeThoughts": True or False, # Optional. Indicates whether to include thoughts in the response. If true, thoughts are returned only when available. - "thinkingBudget": 42, # Optional. Indicates the thinking budget in tokens. This is only applied when enable_thinking is true. + "thinkingBudget": 42, # Optional. Indicates the thinking budget in tokens. }, "topK": 3.14, # Optional. If specified, top-k sampling will be used. "topP": 3.14, # Optional. If specified, nucleus sampling will be used. @@ -1567,7 +1567,7 @@

Method Details

"temperature": 3.14, # Optional. Controls the randomness of predictions. "thinkingConfig": { # Config for thinking features. # Optional. Config for thinking features. An error will be returned if this field is set for models that don't support thinking. "includeThoughts": True or False, # Optional. Indicates whether to include thoughts in the response. If true, thoughts are returned only when available. - "thinkingBudget": 42, # Optional. Indicates the thinking budget in tokens. This is only applied when enable_thinking is true. + "thinkingBudget": 42, # Optional. Indicates the thinking budget in tokens. }, "topK": 3.14, # Optional. If specified, top-k sampling will be used. "topP": 3.14, # Optional. If specified, nucleus sampling will be used. diff --git a/docs/dyn/aiplatform_v1beta1.projects.locations.deploymentResourcePools.html b/docs/dyn/aiplatform_v1beta1.projects.locations.deploymentResourcePools.html index b54583d69c..ef25e8820c 100644 --- a/docs/dyn/aiplatform_v1beta1.projects.locations.deploymentResourcePools.html +++ b/docs/dyn/aiplatform_v1beta1.projects.locations.deploymentResourcePools.html @@ -131,6 +131,9 @@

Method Details

"target": 42, # The target resource utilization in percentage (1% - 100%) for the given metric; once the real usage deviates from the target by a certain percentage, the machine replicas change. The default value is 60 (representing 60%) if not provided. }, ], + "flexStart": { # FlexStart is used to schedule the deployment workload on DWS resource. It contains the max duration of the deployment. # Optional. Immutable. If set, use DWS resource to schedule the deployment workload. reference: (https://cloud.google.com/blog/products/compute/introducing-dynamic-workload-scheduler) + "maxRuntimeDuration": "A String", # The max duration of the deployment is max_runtime_duration. The deployment will be terminated after the duration. The max_runtime_duration can be set up to 7 days. + }, "machineSpec": { # Specification of a single machine. # Required. Immutable. The specification of a single machine being used. "acceleratorCount": 42, # The number of accelerators to attach to the machine. "acceleratorType": "A String", # Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count. @@ -249,6 +252,9 @@

Method Details

"target": 42, # The target resource utilization in percentage (1% - 100%) for the given metric; once the real usage deviates from the target by a certain percentage, the machine replicas change. The default value is 60 (representing 60%) if not provided. }, ], + "flexStart": { # FlexStart is used to schedule the deployment workload on DWS resource. It contains the max duration of the deployment. # Optional. Immutable. If set, use DWS resource to schedule the deployment workload. reference: (https://cloud.google.com/blog/products/compute/introducing-dynamic-workload-scheduler) + "maxRuntimeDuration": "A String", # The max duration of the deployment is max_runtime_duration. The deployment will be terminated after the duration. The max_runtime_duration can be set up to 7 days. + }, "machineSpec": { # Specification of a single machine. # Required. Immutable. The specification of a single machine being used. "acceleratorCount": 42, # The number of accelerators to attach to the machine. "acceleratorType": "A String", # Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count. @@ -306,6 +312,9 @@

Method Details

"target": 42, # The target resource utilization in percentage (1% - 100%) for the given metric; once the real usage deviates from the target by a certain percentage, the machine replicas change. The default value is 60 (representing 60%) if not provided. }, ], + "flexStart": { # FlexStart is used to schedule the deployment workload on DWS resource. It contains the max duration of the deployment. # Optional. Immutable. If set, use DWS resource to schedule the deployment workload. reference: (https://cloud.google.com/blog/products/compute/introducing-dynamic-workload-scheduler) + "maxRuntimeDuration": "A String", # The max duration of the deployment is max_runtime_duration. The deployment will be terminated after the duration. The max_runtime_duration can be set up to 7 days. + }, "machineSpec": { # Specification of a single machine. # Required. Immutable. The specification of a single machine being used. "acceleratorCount": 42, # The number of accelerators to attach to the machine. "acceleratorType": "A String", # Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count. @@ -371,6 +380,9 @@

Method Details

"target": 42, # The target resource utilization in percentage (1% - 100%) for the given metric; once the real usage deviates from the target by a certain percentage, the machine replicas change. The default value is 60 (representing 60%) if not provided. }, ], + "flexStart": { # FlexStart is used to schedule the deployment workload on DWS resource. It contains the max duration of the deployment. # Optional. Immutable. If set, use DWS resource to schedule the deployment workload. reference: (https://cloud.google.com/blog/products/compute/introducing-dynamic-workload-scheduler) + "maxRuntimeDuration": "A String", # The max duration of the deployment is max_runtime_duration. The deployment will be terminated after the duration. The max_runtime_duration can be set up to 7 days. + }, "machineSpec": { # Specification of a single machine. # Required. Immutable. The specification of a single machine being used. "acceleratorCount": 42, # The number of accelerators to attach to the machine. "acceleratorType": "A String", # Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count. @@ -469,6 +481,9 @@

Method Details

"target": 42, # The target resource utilization in percentage (1% - 100%) for the given metric; once the real usage deviates from the target by a certain percentage, the machine replicas change. The default value is 60 (representing 60%) if not provided. }, ], + "flexStart": { # FlexStart is used to schedule the deployment workload on DWS resource. It contains the max duration of the deployment. # Optional. Immutable. If set, use DWS resource to schedule the deployment workload. reference: (https://cloud.google.com/blog/products/compute/introducing-dynamic-workload-scheduler) + "maxRuntimeDuration": "A String", # The max duration of the deployment is max_runtime_duration. The deployment will be terminated after the duration. The max_runtime_duration can be set up to 7 days. + }, "machineSpec": { # Specification of a single machine. # Required. Immutable. The specification of a single machine being used. "acceleratorCount": 42, # The number of accelerators to attach to the machine. "acceleratorType": "A String", # Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count. diff --git a/docs/dyn/aiplatform_v1beta1.projects.locations.endpoints.html b/docs/dyn/aiplatform_v1beta1.projects.locations.endpoints.html index a9b306205d..f4e4cf0636 100644 --- a/docs/dyn/aiplatform_v1beta1.projects.locations.endpoints.html +++ b/docs/dyn/aiplatform_v1beta1.projects.locations.endpoints.html @@ -396,7 +396,7 @@

Method Details

"temperature": 3.14, # Optional. Controls the randomness of predictions. "thinkingConfig": { # Config for thinking features. # Optional. Config for thinking features. An error will be returned if this field is set for models that don't support thinking. "includeThoughts": True or False, # Optional. Indicates whether to include thoughts in the response. If true, thoughts are returned only when available. - "thinkingBudget": 42, # Optional. Indicates the thinking budget in tokens. This is only applied when enable_thinking is true. + "thinkingBudget": 42, # Optional. Indicates the thinking budget in tokens. }, "topK": 3.14, # Optional. If specified, top-k sampling will be used. "topP": 3.14, # Optional. If specified, nucleus sampling will be used. @@ -701,6 +701,9 @@

Method Details

"target": 42, # The target resource utilization in percentage (1% - 100%) for the given metric; once the real usage deviates from the target by a certain percentage, the machine replicas change. The default value is 60 (representing 60%) if not provided. }, ], + "flexStart": { # FlexStart is used to schedule the deployment workload on DWS resource. It contains the max duration of the deployment. # Optional. Immutable. If set, use DWS resource to schedule the deployment workload. reference: (https://cloud.google.com/blog/products/compute/introducing-dynamic-workload-scheduler) + "maxRuntimeDuration": "A String", # The max duration of the deployment is max_runtime_duration. The deployment will be terminated after the duration. The max_runtime_duration can be set up to 7 days. + }, "machineSpec": { # Specification of a single machine. # Required. Immutable. The specification of a single machine being used. "acceleratorCount": 42, # The number of accelerators to attach to the machine. "acceleratorType": "A String", # Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count. @@ -911,6 +914,16 @@

Method Details

"projectAllowlist": [ # A list of Projects from which the forwarding rule will target the service attachment. "A String", ], + "pscAutomationConfigs": [ # Optional. List of projects and networks where the PSC endpoints will be created. This field is used by Online Inference(Prediction) only. + { # PSC config that is used to automatically create PSC endpoints in the user projects. + "errorMessage": "A String", # Output only. Error message if the PSC service automation failed. + "forwardingRule": "A String", # Output only. Forwarding rule created by the PSC service automation. + "ipAddress": "A String", # Output only. IP address rule created by the PSC service automation. + "network": "A String", # Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/get): `projects/{project}/global/networks/{network}`. + "projectId": "A String", # Required. Project id used to create forwarding rule. + "state": "A String", # Output only. The state of the PSC service automation. + }, + ], "serviceAttachment": "A String", # Output only. The name of the generated service attachment resource. This is only populated if the endpoint is deployed with PrivateServiceConnect. }, "satisfiesPzi": True or False, # Output only. Reserved for future use. @@ -1010,6 +1023,9 @@

Method Details

"target": 42, # The target resource utilization in percentage (1% - 100%) for the given metric; once the real usage deviates from the target by a certain percentage, the machine replicas change. The default value is 60 (representing 60%) if not provided. }, ], + "flexStart": { # FlexStart is used to schedule the deployment workload on DWS resource. It contains the max duration of the deployment. # Optional. Immutable. If set, use DWS resource to schedule the deployment workload. reference: (https://cloud.google.com/blog/products/compute/introducing-dynamic-workload-scheduler) + "maxRuntimeDuration": "A String", # The max duration of the deployment is max_runtime_duration. The deployment will be terminated after the duration. The max_runtime_duration can be set up to 7 days. + }, "machineSpec": { # Specification of a single machine. # Required. Immutable. The specification of a single machine being used. "acceleratorCount": 42, # The number of accelerators to attach to the machine. "acceleratorType": "A String", # Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count. @@ -1889,7 +1905,7 @@

Method Details

"temperature": 3.14, # Optional. Controls the randomness of predictions. "thinkingConfig": { # Config for thinking features. # Optional. Config for thinking features. An error will be returned if this field is set for models that don't support thinking. "includeThoughts": True or False, # Optional. Indicates whether to include thoughts in the response. If true, thoughts are returned only when available. - "thinkingBudget": 42, # Optional. Indicates the thinking budget in tokens. This is only applied when enable_thinking is true. + "thinkingBudget": 42, # Optional. Indicates the thinking budget in tokens. }, "topK": 3.14, # Optional. If specified, top-k sampling will be used. "topP": 3.14, # Optional. If specified, nucleus sampling will be used. @@ -2424,6 +2440,9 @@

Method Details

"target": 42, # The target resource utilization in percentage (1% - 100%) for the given metric; once the real usage deviates from the target by a certain percentage, the machine replicas change. The default value is 60 (representing 60%) if not provided. }, ], + "flexStart": { # FlexStart is used to schedule the deployment workload on DWS resource. It contains the max duration of the deployment. # Optional. Immutable. If set, use DWS resource to schedule the deployment workload. reference: (https://cloud.google.com/blog/products/compute/introducing-dynamic-workload-scheduler) + "maxRuntimeDuration": "A String", # The max duration of the deployment is max_runtime_duration. The deployment will be terminated after the duration. The max_runtime_duration can be set up to 7 days. + }, "machineSpec": { # Specification of a single machine. # Required. Immutable. The specification of a single machine being used. "acceleratorCount": 42, # The number of accelerators to attach to the machine. "acceleratorType": "A String", # Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count. @@ -2634,6 +2653,16 @@

Method Details

"projectAllowlist": [ # A list of Projects from which the forwarding rule will target the service attachment. "A String", ], + "pscAutomationConfigs": [ # Optional. List of projects and networks where the PSC endpoints will be created. This field is used by Online Inference(Prediction) only. + { # PSC config that is used to automatically create PSC endpoints in the user projects. + "errorMessage": "A String", # Output only. Error message if the PSC service automation failed. + "forwardingRule": "A String", # Output only. Forwarding rule created by the PSC service automation. + "ipAddress": "A String", # Output only. IP address rule created by the PSC service automation. + "network": "A String", # Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/get): `projects/{project}/global/networks/{network}`. + "projectId": "A String", # Required. Project id used to create forwarding rule. + "state": "A String", # Output only. The state of the PSC service automation. + }, + ], "serviceAttachment": "A String", # Output only. The name of the generated service attachment resource. This is only populated if the endpoint is deployed with PrivateServiceConnect. }, "satisfiesPzi": True or False, # Output only. Reserved for future use. @@ -2722,6 +2751,9 @@

Method Details

"target": 42, # The target resource utilization in percentage (1% - 100%) for the given metric; once the real usage deviates from the target by a certain percentage, the machine replicas change. The default value is 60 (representing 60%) if not provided. }, ], + "flexStart": { # FlexStart is used to schedule the deployment workload on DWS resource. It contains the max duration of the deployment. # Optional. Immutable. If set, use DWS resource to schedule the deployment workload. reference: (https://cloud.google.com/blog/products/compute/introducing-dynamic-workload-scheduler) + "maxRuntimeDuration": "A String", # The max duration of the deployment is max_runtime_duration. The deployment will be terminated after the duration. The max_runtime_duration can be set up to 7 days. + }, "machineSpec": { # Specification of a single machine. # Required. Immutable. The specification of a single machine being used. "acceleratorCount": 42, # The number of accelerators to attach to the machine. "acceleratorType": "A String", # Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count. @@ -2932,6 +2964,16 @@

Method Details

"projectAllowlist": [ # A list of Projects from which the forwarding rule will target the service attachment. "A String", ], + "pscAutomationConfigs": [ # Optional. List of projects and networks where the PSC endpoints will be created. This field is used by Online Inference(Prediction) only. + { # PSC config that is used to automatically create PSC endpoints in the user projects. + "errorMessage": "A String", # Output only. Error message if the PSC service automation failed. + "forwardingRule": "A String", # Output only. Forwarding rule created by the PSC service automation. + "ipAddress": "A String", # Output only. IP address rule created by the PSC service automation. + "network": "A String", # Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/get): `projects/{project}/global/networks/{network}`. + "projectId": "A String", # Required. Project id used to create forwarding rule. + "state": "A String", # Output only. The state of the PSC service automation. + }, + ], "serviceAttachment": "A String", # Output only. The name of the generated service attachment resource. This is only populated if the endpoint is deployed with PrivateServiceConnect. }, "satisfiesPzi": True or False, # Output only. Reserved for future use. @@ -2984,6 +3026,9 @@

Method Details

"target": 42, # The target resource utilization in percentage (1% - 100%) for the given metric; once the real usage deviates from the target by a certain percentage, the machine replicas change. The default value is 60 (representing 60%) if not provided. }, ], + "flexStart": { # FlexStart is used to schedule the deployment workload on DWS resource. It contains the max duration of the deployment. # Optional. Immutable. If set, use DWS resource to schedule the deployment workload. reference: (https://cloud.google.com/blog/products/compute/introducing-dynamic-workload-scheduler) + "maxRuntimeDuration": "A String", # The max duration of the deployment is max_runtime_duration. The deployment will be terminated after the duration. The max_runtime_duration can be set up to 7 days. + }, "machineSpec": { # Specification of a single machine. # Required. Immutable. The specification of a single machine being used. "acceleratorCount": 42, # The number of accelerators to attach to the machine. "acceleratorType": "A String", # Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count. @@ -3223,6 +3268,9 @@

Method Details

"target": 42, # The target resource utilization in percentage (1% - 100%) for the given metric; once the real usage deviates from the target by a certain percentage, the machine replicas change. The default value is 60 (representing 60%) if not provided. }, ], + "flexStart": { # FlexStart is used to schedule the deployment workload on DWS resource. It contains the max duration of the deployment. # Optional. Immutable. If set, use DWS resource to schedule the deployment workload. reference: (https://cloud.google.com/blog/products/compute/introducing-dynamic-workload-scheduler) + "maxRuntimeDuration": "A String", # The max duration of the deployment is max_runtime_duration. The deployment will be terminated after the duration. The max_runtime_duration can be set up to 7 days. + }, "machineSpec": { # Specification of a single machine. # Required. Immutable. The specification of a single machine being used. "acceleratorCount": 42, # The number of accelerators to attach to the machine. "acceleratorType": "A String", # Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count. @@ -3433,6 +3481,16 @@

Method Details

"projectAllowlist": [ # A list of Projects from which the forwarding rule will target the service attachment. "A String", ], + "pscAutomationConfigs": [ # Optional. List of projects and networks where the PSC endpoints will be created. This field is used by Online Inference(Prediction) only. + { # PSC config that is used to automatically create PSC endpoints in the user projects. + "errorMessage": "A String", # Output only. Error message if the PSC service automation failed. + "forwardingRule": "A String", # Output only. Forwarding rule created by the PSC service automation. + "ipAddress": "A String", # Output only. IP address rule created by the PSC service automation. + "network": "A String", # Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/get): `projects/{project}/global/networks/{network}`. + "projectId": "A String", # Required. Project id used to create forwarding rule. + "state": "A String", # Output only. The state of the PSC service automation. + }, + ], "serviceAttachment": "A String", # Output only. The name of the generated service attachment resource. This is only populated if the endpoint is deployed with PrivateServiceConnect. }, "satisfiesPzi": True or False, # Output only. Reserved for future use. @@ -3474,6 +3532,9 @@

Method Details

"target": 42, # The target resource utilization in percentage (1% - 100%) for the given metric; once the real usage deviates from the target by a certain percentage, the machine replicas change. The default value is 60 (representing 60%) if not provided. }, ], + "flexStart": { # FlexStart is used to schedule the deployment workload on DWS resource. It contains the max duration of the deployment. # Optional. Immutable. If set, use DWS resource to schedule the deployment workload. reference: (https://cloud.google.com/blog/products/compute/introducing-dynamic-workload-scheduler) + "maxRuntimeDuration": "A String", # The max duration of the deployment is max_runtime_duration. The deployment will be terminated after the duration. The max_runtime_duration can be set up to 7 days. + }, "machineSpec": { # Specification of a single machine. # Required. Immutable. The specification of a single machine being used. "acceleratorCount": 42, # The number of accelerators to attach to the machine. "acceleratorType": "A String", # Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count. @@ -3684,6 +3745,16 @@

Method Details

"projectAllowlist": [ # A list of Projects from which the forwarding rule will target the service attachment. "A String", ], + "pscAutomationConfigs": [ # Optional. List of projects and networks where the PSC endpoints will be created. This field is used by Online Inference(Prediction) only. + { # PSC config that is used to automatically create PSC endpoints in the user projects. + "errorMessage": "A String", # Output only. Error message if the PSC service automation failed. + "forwardingRule": "A String", # Output only. Forwarding rule created by the PSC service automation. + "ipAddress": "A String", # Output only. IP address rule created by the PSC service automation. + "network": "A String", # Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/get): `projects/{project}/global/networks/{network}`. + "projectId": "A String", # Required. Project id used to create forwarding rule. + "state": "A String", # Output only. The state of the PSC service automation. + }, + ], "serviceAttachment": "A String", # Output only. The name of the generated service attachment resource. This is only populated if the endpoint is deployed with PrivateServiceConnect. }, "satisfiesPzi": True or False, # Output only. Reserved for future use. @@ -4201,7 +4272,7 @@

Method Details

"temperature": 3.14, # Optional. Controls the randomness of predictions. "thinkingConfig": { # Config for thinking features. # Optional. Config for thinking features. An error will be returned if this field is set for models that don't support thinking. "includeThoughts": True or False, # Optional. Indicates whether to include thoughts in the response. If true, thoughts are returned only when available. - "thinkingBudget": 42, # Optional. Indicates the thinking budget in tokens. This is only applied when enable_thinking is true. + "thinkingBudget": 42, # Optional. Indicates the thinking budget in tokens. }, "topK": 3.14, # Optional. If specified, top-k sampling will be used. "topP": 3.14, # Optional. If specified, nucleus sampling will be used. @@ -4839,6 +4910,9 @@

Method Details

"target": 42, # The target resource utilization in percentage (1% - 100%) for the given metric; once the real usage deviates from the target by a certain percentage, the machine replicas change. The default value is 60 (representing 60%) if not provided. }, ], + "flexStart": { # FlexStart is used to schedule the deployment workload on DWS resource. It contains the max duration of the deployment. # Optional. Immutable. If set, use DWS resource to schedule the deployment workload. reference: (https://cloud.google.com/blog/products/compute/introducing-dynamic-workload-scheduler) + "maxRuntimeDuration": "A String", # The max duration of the deployment is max_runtime_duration. The deployment will be terminated after the duration. The max_runtime_duration can be set up to 7 days. + }, "machineSpec": { # Specification of a single machine. # Required. Immutable. The specification of a single machine being used. "acceleratorCount": 42, # The number of accelerators to attach to the machine. "acceleratorType": "A String", # Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count. @@ -5049,6 +5123,16 @@

Method Details

"projectAllowlist": [ # A list of Projects from which the forwarding rule will target the service attachment. "A String", ], + "pscAutomationConfigs": [ # Optional. List of projects and networks where the PSC endpoints will be created. This field is used by Online Inference(Prediction) only. + { # PSC config that is used to automatically create PSC endpoints in the user projects. + "errorMessage": "A String", # Output only. Error message if the PSC service automation failed. + "forwardingRule": "A String", # Output only. Forwarding rule created by the PSC service automation. + "ipAddress": "A String", # Output only. IP address rule created by the PSC service automation. + "network": "A String", # Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/get): `projects/{project}/global/networks/{network}`. + "projectId": "A String", # Required. Project id used to create forwarding rule. + "state": "A String", # Output only. The state of the PSC service automation. + }, + ], "serviceAttachment": "A String", # Output only. The name of the generated service attachment resource. This is only populated if the endpoint is deployed with PrivateServiceConnect. }, "satisfiesPzi": True or False, # Output only. Reserved for future use. diff --git a/docs/dyn/aiplatform_v1beta1.projects.locations.featureOnlineStores.featureViews.html b/docs/dyn/aiplatform_v1beta1.projects.locations.featureOnlineStores.featureViews.html index fc92978e71..b90755c6e8 100644 --- a/docs/dyn/aiplatform_v1beta1.projects.locations.featureOnlineStores.featureViews.html +++ b/docs/dyn/aiplatform_v1beta1.projects.locations.featureOnlineStores.featureViews.html @@ -329,7 +329,7 @@

Method Details

}, "int64Value": "A String", # Int64 feature value. "metadata": { # Metadata of feature value. # Metadata of feature value. - "generateTime": "A String", # Feature generation timestamp. Typically, it is provided by user at feature ingestion time. If not, feature store will use the system timestamp when the data is ingested into feature store. For streaming ingestion, the time, aligned by days, must be no older than five years (1825 days) and no later than one year (366 days) in the future. + "generateTime": "A String", # Feature generation timestamp. Typically, it is provided by user at feature ingestion time. If not, feature store will use the system timestamp when the data is ingested into feature store. Legacy Feature Store: For streaming ingestion, the time, aligned by days, must be no older than five years (1825 days) and no later than one year (366 days) in the future. }, "stringArrayValue": { # A list of string values. # A list of string type feature value. "values": [ # A list of string values. @@ -452,7 +452,7 @@

Method Details

}, "int64Value": "A String", # Int64 feature value. "metadata": { # Metadata of feature value. # Metadata of feature value. - "generateTime": "A String", # Feature generation timestamp. Typically, it is provided by user at feature ingestion time. If not, feature store will use the system timestamp when the data is ingested into feature store. For streaming ingestion, the time, aligned by days, must be no older than five years (1825 days) and no later than one year (366 days) in the future. + "generateTime": "A String", # Feature generation timestamp. Typically, it is provided by user at feature ingestion time. If not, feature store will use the system timestamp when the data is ingested into feature store. Legacy Feature Store: For streaming ingestion, the time, aligned by days, must be no older than five years (1825 days) and no later than one year (366 days) in the future. }, "stringArrayValue": { # A list of string values. # A list of string type feature value. "values": [ # A list of string values. @@ -919,7 +919,7 @@

Method Details

}, "int64Value": "A String", # Int64 feature value. "metadata": { # Metadata of feature value. # Metadata of feature value. - "generateTime": "A String", # Feature generation timestamp. Typically, it is provided by user at feature ingestion time. If not, feature store will use the system timestamp when the data is ingested into feature store. For streaming ingestion, the time, aligned by days, must be no older than five years (1825 days) and no later than one year (366 days) in the future. + "generateTime": "A String", # Feature generation timestamp. Typically, it is provided by user at feature ingestion time. If not, feature store will use the system timestamp when the data is ingested into feature store. Legacy Feature Store: For streaming ingestion, the time, aligned by days, must be no older than five years (1825 days) and no later than one year (366 days) in the future. }, "stringArrayValue": { # A list of string values. # A list of string type feature value. "values": [ # A list of string values. @@ -1074,7 +1074,7 @@

Method Details

}, "int64Value": "A String", # Int64 feature value. "metadata": { # Metadata of feature value. # Metadata of feature value. - "generateTime": "A String", # Feature generation timestamp. Typically, it is provided by user at feature ingestion time. If not, feature store will use the system timestamp when the data is ingested into feature store. For streaming ingestion, the time, aligned by days, must be no older than five years (1825 days) and no later than one year (366 days) in the future. + "generateTime": "A String", # Feature generation timestamp. Typically, it is provided by user at feature ingestion time. If not, feature store will use the system timestamp when the data is ingested into feature store. Legacy Feature Store: For streaming ingestion, the time, aligned by days, must be no older than five years (1825 days) and no later than one year (366 days) in the future. }, "stringArrayValue": { # A list of string values. # A list of string type feature value. "values": [ # A list of string values. diff --git a/docs/dyn/aiplatform_v1beta1.projects.locations.featureOnlineStores.html b/docs/dyn/aiplatform_v1beta1.projects.locations.featureOnlineStores.html index 8d52626924..7554eca2ee 100644 --- a/docs/dyn/aiplatform_v1beta1.projects.locations.featureOnlineStores.html +++ b/docs/dyn/aiplatform_v1beta1.projects.locations.featureOnlineStores.html @@ -145,6 +145,16 @@

Method Details

"projectAllowlist": [ # A list of Projects from which the forwarding rule will target the service attachment. "A String", ], + "pscAutomationConfigs": [ # Optional. List of projects and networks where the PSC endpoints will be created. This field is used by Online Inference(Prediction) only. + { # PSC config that is used to automatically create PSC endpoints in the user projects. + "errorMessage": "A String", # Output only. Error message if the PSC service automation failed. + "forwardingRule": "A String", # Output only. Forwarding rule created by the PSC service automation. + "ipAddress": "A String", # Output only. IP address rule created by the PSC service automation. + "network": "A String", # Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/get): `projects/{project}/global/networks/{network}`. + "projectId": "A String", # Required. Project id used to create forwarding rule. + "state": "A String", # Output only. The state of the PSC service automation. + }, + ], "serviceAttachment": "A String", # Output only. The name of the generated service attachment resource. This is only populated if the endpoint is deployed with PrivateServiceConnect. }, "publicEndpointDomainName": "A String", # Output only. This field will be populated with the domain name to use for this FeatureOnlineStore @@ -265,6 +275,16 @@

Method Details

"projectAllowlist": [ # A list of Projects from which the forwarding rule will target the service attachment. "A String", ], + "pscAutomationConfigs": [ # Optional. List of projects and networks where the PSC endpoints will be created. This field is used by Online Inference(Prediction) only. + { # PSC config that is used to automatically create PSC endpoints in the user projects. + "errorMessage": "A String", # Output only. Error message if the PSC service automation failed. + "forwardingRule": "A String", # Output only. Forwarding rule created by the PSC service automation. + "ipAddress": "A String", # Output only. IP address rule created by the PSC service automation. + "network": "A String", # Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/get): `projects/{project}/global/networks/{network}`. + "projectId": "A String", # Required. Project id used to create forwarding rule. + "state": "A String", # Output only. The state of the PSC service automation. + }, + ], "serviceAttachment": "A String", # Output only. The name of the generated service attachment resource. This is only populated if the endpoint is deployed with PrivateServiceConnect. }, "publicEndpointDomainName": "A String", # Output only. This field will be populated with the domain name to use for this FeatureOnlineStore @@ -361,6 +381,16 @@

Method Details

"projectAllowlist": [ # A list of Projects from which the forwarding rule will target the service attachment. "A String", ], + "pscAutomationConfigs": [ # Optional. List of projects and networks where the PSC endpoints will be created. This field is used by Online Inference(Prediction) only. + { # PSC config that is used to automatically create PSC endpoints in the user projects. + "errorMessage": "A String", # Output only. Error message if the PSC service automation failed. + "forwardingRule": "A String", # Output only. Forwarding rule created by the PSC service automation. + "ipAddress": "A String", # Output only. IP address rule created by the PSC service automation. + "network": "A String", # Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/get): `projects/{project}/global/networks/{network}`. + "projectId": "A String", # Required. Project id used to create forwarding rule. + "state": "A String", # Output only. The state of the PSC service automation. + }, + ], "serviceAttachment": "A String", # Output only. The name of the generated service attachment resource. This is only populated if the endpoint is deployed with PrivateServiceConnect. }, "publicEndpointDomainName": "A String", # Output only. This field will be populated with the domain name to use for this FeatureOnlineStore @@ -428,6 +458,16 @@

Method Details

"projectAllowlist": [ # A list of Projects from which the forwarding rule will target the service attachment. "A String", ], + "pscAutomationConfigs": [ # Optional. List of projects and networks where the PSC endpoints will be created. This field is used by Online Inference(Prediction) only. + { # PSC config that is used to automatically create PSC endpoints in the user projects. + "errorMessage": "A String", # Output only. Error message if the PSC service automation failed. + "forwardingRule": "A String", # Output only. Forwarding rule created by the PSC service automation. + "ipAddress": "A String", # Output only. IP address rule created by the PSC service automation. + "network": "A String", # Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/get): `projects/{project}/global/networks/{network}`. + "projectId": "A String", # Required. Project id used to create forwarding rule. + "state": "A String", # Output only. The state of the PSC service automation. + }, + ], "serviceAttachment": "A String", # Output only. The name of the generated service attachment resource. This is only populated if the endpoint is deployed with PrivateServiceConnect. }, "publicEndpointDomainName": "A String", # Output only. This field will be populated with the domain name to use for this FeatureOnlineStore diff --git a/docs/dyn/aiplatform_v1beta1.projects.locations.featurestores.entityTypes.html b/docs/dyn/aiplatform_v1beta1.projects.locations.featurestores.entityTypes.html index f7663fb305..8e5cec4037 100644 --- a/docs/dyn/aiplatform_v1beta1.projects.locations.featurestores.entityTypes.html +++ b/docs/dyn/aiplatform_v1beta1.projects.locations.featurestores.entityTypes.html @@ -748,7 +748,7 @@

Method Details

}, "int64Value": "A String", # Int64 feature value. "metadata": { # Metadata of feature value. # Metadata of feature value. - "generateTime": "A String", # Feature generation timestamp. Typically, it is provided by user at feature ingestion time. If not, feature store will use the system timestamp when the data is ingested into feature store. For streaming ingestion, the time, aligned by days, must be no older than five years (1825 days) and no later than one year (366 days) in the future. + "generateTime": "A String", # Feature generation timestamp. Typically, it is provided by user at feature ingestion time. If not, feature store will use the system timestamp when the data is ingested into feature store. Legacy Feature Store: For streaming ingestion, the time, aligned by days, must be no older than five years (1825 days) and no later than one year (366 days) in the future. }, "stringArrayValue": { # A list of string values. # A list of string type feature value. "values": [ # A list of string values. @@ -788,7 +788,7 @@

Method Details

}, "int64Value": "A String", # Int64 feature value. "metadata": { # Metadata of feature value. # Metadata of feature value. - "generateTime": "A String", # Feature generation timestamp. Typically, it is provided by user at feature ingestion time. If not, feature store will use the system timestamp when the data is ingested into feature store. For streaming ingestion, the time, aligned by days, must be no older than five years (1825 days) and no later than one year (366 days) in the future. + "generateTime": "A String", # Feature generation timestamp. Typically, it is provided by user at feature ingestion time. If not, feature store will use the system timestamp when the data is ingested into feature store. Legacy Feature Store: For streaming ingestion, the time, aligned by days, must be no older than five years (1825 days) and no later than one year (366 days) in the future. }, "stringArrayValue": { # A list of string values. # A list of string type feature value. "values": [ # A list of string values. @@ -935,7 +935,7 @@

Method Details

}, "int64Value": "A String", # Int64 feature value. "metadata": { # Metadata of feature value. # Metadata of feature value. - "generateTime": "A String", # Feature generation timestamp. Typically, it is provided by user at feature ingestion time. If not, feature store will use the system timestamp when the data is ingested into feature store. For streaming ingestion, the time, aligned by days, must be no older than five years (1825 days) and no later than one year (366 days) in the future. + "generateTime": "A String", # Feature generation timestamp. Typically, it is provided by user at feature ingestion time. If not, feature store will use the system timestamp when the data is ingested into feature store. Legacy Feature Store: For streaming ingestion, the time, aligned by days, must be no older than five years (1825 days) and no later than one year (366 days) in the future. }, "stringArrayValue": { # A list of string values. # A list of string type feature value. "values": [ # A list of string values. @@ -975,7 +975,7 @@

Method Details

}, "int64Value": "A String", # Int64 feature value. "metadata": { # Metadata of feature value. # Metadata of feature value. - "generateTime": "A String", # Feature generation timestamp. Typically, it is provided by user at feature ingestion time. If not, feature store will use the system timestamp when the data is ingested into feature store. For streaming ingestion, the time, aligned by days, must be no older than five years (1825 days) and no later than one year (366 days) in the future. + "generateTime": "A String", # Feature generation timestamp. Typically, it is provided by user at feature ingestion time. If not, feature store will use the system timestamp when the data is ingested into feature store. Legacy Feature Store: For streaming ingestion, the time, aligned by days, must be no older than five years (1825 days) and no later than one year (366 days) in the future. }, "stringArrayValue": { # A list of string values. # A list of string type feature value. "values": [ # A list of string values. @@ -1066,7 +1066,7 @@

Method Details

}, "int64Value": "A String", # Int64 feature value. "metadata": { # Metadata of feature value. # Metadata of feature value. - "generateTime": "A String", # Feature generation timestamp. Typically, it is provided by user at feature ingestion time. If not, feature store will use the system timestamp when the data is ingested into feature store. For streaming ingestion, the time, aligned by days, must be no older than five years (1825 days) and no later than one year (366 days) in the future. + "generateTime": "A String", # Feature generation timestamp. Typically, it is provided by user at feature ingestion time. If not, feature store will use the system timestamp when the data is ingested into feature store. Legacy Feature Store: For streaming ingestion, the time, aligned by days, must be no older than five years (1825 days) and no later than one year (366 days) in the future. }, "stringArrayValue": { # A list of string values. # A list of string type feature value. "values": [ # A list of string values. diff --git a/docs/dyn/aiplatform_v1beta1.projects.locations.html b/docs/dyn/aiplatform_v1beta1.projects.locations.html index 15f097eab2..50d5f52bfb 100644 --- a/docs/dyn/aiplatform_v1beta1.projects.locations.html +++ b/docs/dyn/aiplatform_v1beta1.projects.locations.html @@ -644,6 +644,9 @@

Method Details

"target": 42, # The target resource utilization in percentage (1% - 100%) for the given metric; once the real usage deviates from the target by a certain percentage, the machine replicas change. The default value is 60 (representing 60%) if not provided. }, ], + "flexStart": { # FlexStart is used to schedule the deployment workload on DWS resource. It contains the max duration of the deployment. # Optional. Immutable. If set, use DWS resource to schedule the deployment workload. reference: (https://cloud.google.com/blog/products/compute/introducing-dynamic-workload-scheduler) + "maxRuntimeDuration": "A String", # The max duration of the deployment is max_runtime_duration. The deployment will be terminated after the duration. The max_runtime_duration can be set up to 7 days. + }, "machineSpec": { # Specification of a single machine. # Required. Immutable. The specification of a single machine being used. "acceleratorCount": 42, # The number of accelerators to attach to the machine. "acceleratorType": "A String", # Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count. @@ -728,7 +731,7 @@

Method Details

}, "healthRoute": "A String", # Immutable. HTTP path on the container to send health checks to. Vertex AI intermittently sends GET requests to this path on the container's IP address and port to check that the container is healthy. Read more about [health checks](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#health). For example, if you set this field to `/bar`, then Vertex AI intermittently sends a GET request to the `/bar` path on the port of your container specified by the first value of this `ModelContainerSpec`'s ports field. If you don't specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/ DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) "imageUri": "A String", # Required. Immutable. URI of the Docker image to be used as the custom container for serving predictions. This URI must identify an image in Artifact Registry or Container Registry. Learn more about the [container publishing requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#publishing), including permissions requirements for the Vertex AI Service Agent. The container image is ingested upon ModelService.UploadModel, stored internally, and this original path is afterwards not used. To learn about the requirements for the Docker image itself, see [Custom container requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#). You can use the URI to one of Vertex AI's [pre-built container images for prediction](https://cloud.google.com/vertex-ai/docs/predictions/pre-built-containers) in this field. - "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with [PredictionService.Invoke] eg: "/invoke/foo/bar". Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. + "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with invoke http call eg: "/invoke/foo/bar", however the [PredictionService.Invoke] RPC is not supported yet. Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. "livenessProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes liveness probe. "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. @@ -855,6 +858,9 @@

Method Details

"target": 42, # The target resource utilization in percentage (1% - 100%) for the given metric; once the real usage deviates from the target by a certain percentage, the machine replicas change. The default value is 60 (representing 60%) if not provided. }, ], + "flexStart": { # FlexStart is used to schedule the deployment workload on DWS resource. It contains the max duration of the deployment. # Optional. Immutable. If set, use DWS resource to schedule the deployment workload. reference: (https://cloud.google.com/blog/products/compute/introducing-dynamic-workload-scheduler) + "maxRuntimeDuration": "A String", # The max duration of the deployment is max_runtime_duration. The deployment will be terminated after the duration. The max_runtime_duration can be set up to 7 days. + }, "machineSpec": { # Specification of a single machine. # Required. Immutable. The specification of a single machine being used. "acceleratorCount": 42, # The number of accelerators to attach to the machine. "acceleratorType": "A String", # Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count. @@ -935,7 +941,7 @@

Method Details

}, }, "metrics": [ # Required. The metrics used for evaluation. - { # The metric used for dataset level evaluation. + { # The metric used for running evaluations. "aggregationMetrics": [ # Optional. The aggregation metrics to use. "A String", ], diff --git a/docs/dyn/aiplatform_v1beta1.projects.locations.indexEndpoints.html b/docs/dyn/aiplatform_v1beta1.projects.locations.indexEndpoints.html index 45a794eaa7..4b018d0d11 100644 --- a/docs/dyn/aiplatform_v1beta1.projects.locations.indexEndpoints.html +++ b/docs/dyn/aiplatform_v1beta1.projects.locations.indexEndpoints.html @@ -146,6 +146,9 @@

Method Details

"target": 42, # The target resource utilization in percentage (1% - 100%) for the given metric; once the real usage deviates from the target by a certain percentage, the machine replicas change. The default value is 60 (representing 60%) if not provided. }, ], + "flexStart": { # FlexStart is used to schedule the deployment workload on DWS resource. It contains the max duration of the deployment. # Optional. Immutable. If set, use DWS resource to schedule the deployment workload. reference: (https://cloud.google.com/blog/products/compute/introducing-dynamic-workload-scheduler) + "maxRuntimeDuration": "A String", # The max duration of the deployment is max_runtime_duration. The deployment will be terminated after the duration. The max_runtime_duration can be set up to 7 days. + }, "machineSpec": { # Specification of a single machine. # Required. Immutable. The specification of a single machine being used. "acceleratorCount": 42, # The number of accelerators to attach to the machine. "acceleratorType": "A String", # Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count. @@ -193,11 +196,11 @@

Method Details

"serviceAttachment": "A String", # Output only. The name of the service attachment resource. Populated if private service connect is enabled. }, "pscAutomationConfigs": [ # Optional. If set for PSC deployed index, PSC connection will be automatically created after deployment is done and the endpoint information is populated in private_endpoints.psc_automated_endpoints. - { # PSC config that is used to automatically create forwarding rule via ServiceConnectionMap. + { # PSC config that is used to automatically create PSC endpoints in the user projects. "errorMessage": "A String", # Output only. Error message if the PSC service automation failed. "forwardingRule": "A String", # Output only. Forwarding rule created by the PSC service automation. "ipAddress": "A String", # Output only. IP address rule created by the PSC service automation. - "network": "A String", # Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/insert): `projects/{project}/global/networks/{network}`. Where {project} is a project number, as in '12345', and {network} is network name. + "network": "A String", # Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/get): `projects/{project}/global/networks/{network}`. "projectId": "A String", # Required. Project id used to create forwarding rule. "state": "A String", # Output only. The state of the PSC service automation. }, @@ -225,6 +228,16 @@

Method Details

"projectAllowlist": [ # A list of Projects from which the forwarding rule will target the service attachment. "A String", ], + "pscAutomationConfigs": [ # Optional. List of projects and networks where the PSC endpoints will be created. This field is used by Online Inference(Prediction) only. + { # PSC config that is used to automatically create PSC endpoints in the user projects. + "errorMessage": "A String", # Output only. Error message if the PSC service automation failed. + "forwardingRule": "A String", # Output only. Forwarding rule created by the PSC service automation. + "ipAddress": "A String", # Output only. IP address rule created by the PSC service automation. + "network": "A String", # Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/get): `projects/{project}/global/networks/{network}`. + "projectId": "A String", # Required. Project id used to create forwarding rule. + "state": "A String", # Output only. The state of the PSC service automation. + }, + ], "serviceAttachment": "A String", # Output only. The name of the generated service attachment resource. This is only populated if the endpoint is deployed with PrivateServiceConnect. }, "publicEndpointDomainName": "A String", # Output only. If public_endpoint_enabled is true, this field will be populated with the domain name to use for this index endpoint. @@ -321,6 +334,9 @@

Method Details

"target": 42, # The target resource utilization in percentage (1% - 100%) for the given metric; once the real usage deviates from the target by a certain percentage, the machine replicas change. The default value is 60 (representing 60%) if not provided. }, ], + "flexStart": { # FlexStart is used to schedule the deployment workload on DWS resource. It contains the max duration of the deployment. # Optional. Immutable. If set, use DWS resource to schedule the deployment workload. reference: (https://cloud.google.com/blog/products/compute/introducing-dynamic-workload-scheduler) + "maxRuntimeDuration": "A String", # The max duration of the deployment is max_runtime_duration. The deployment will be terminated after the duration. The max_runtime_duration can be set up to 7 days. + }, "machineSpec": { # Specification of a single machine. # Required. Immutable. The specification of a single machine being used. "acceleratorCount": 42, # The number of accelerators to attach to the machine. "acceleratorType": "A String", # Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count. @@ -368,11 +384,11 @@

Method Details

"serviceAttachment": "A String", # Output only. The name of the service attachment resource. Populated if private service connect is enabled. }, "pscAutomationConfigs": [ # Optional. If set for PSC deployed index, PSC connection will be automatically created after deployment is done and the endpoint information is populated in private_endpoints.psc_automated_endpoints. - { # PSC config that is used to automatically create forwarding rule via ServiceConnectionMap. + { # PSC config that is used to automatically create PSC endpoints in the user projects. "errorMessage": "A String", # Output only. Error message if the PSC service automation failed. "forwardingRule": "A String", # Output only. Forwarding rule created by the PSC service automation. "ipAddress": "A String", # Output only. IP address rule created by the PSC service automation. - "network": "A String", # Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/insert): `projects/{project}/global/networks/{network}`. Where {project} is a project number, as in '12345', and {network} is network name. + "network": "A String", # Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/get): `projects/{project}/global/networks/{network}`. "projectId": "A String", # Required. Project id used to create forwarding rule. "state": "A String", # Output only. The state of the PSC service automation. }, @@ -564,6 +580,9 @@

Method Details

"target": 42, # The target resource utilization in percentage (1% - 100%) for the given metric; once the real usage deviates from the target by a certain percentage, the machine replicas change. The default value is 60 (representing 60%) if not provided. }, ], + "flexStart": { # FlexStart is used to schedule the deployment workload on DWS resource. It contains the max duration of the deployment. # Optional. Immutable. If set, use DWS resource to schedule the deployment workload. reference: (https://cloud.google.com/blog/products/compute/introducing-dynamic-workload-scheduler) + "maxRuntimeDuration": "A String", # The max duration of the deployment is max_runtime_duration. The deployment will be terminated after the duration. The max_runtime_duration can be set up to 7 days. + }, "machineSpec": { # Specification of a single machine. # Required. Immutable. The specification of a single machine being used. "acceleratorCount": 42, # The number of accelerators to attach to the machine. "acceleratorType": "A String", # Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count. @@ -611,11 +630,11 @@

Method Details

"serviceAttachment": "A String", # Output only. The name of the service attachment resource. Populated if private service connect is enabled. }, "pscAutomationConfigs": [ # Optional. If set for PSC deployed index, PSC connection will be automatically created after deployment is done and the endpoint information is populated in private_endpoints.psc_automated_endpoints. - { # PSC config that is used to automatically create forwarding rule via ServiceConnectionMap. + { # PSC config that is used to automatically create PSC endpoints in the user projects. "errorMessage": "A String", # Output only. Error message if the PSC service automation failed. "forwardingRule": "A String", # Output only. Forwarding rule created by the PSC service automation. "ipAddress": "A String", # Output only. IP address rule created by the PSC service automation. - "network": "A String", # Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/insert): `projects/{project}/global/networks/{network}`. Where {project} is a project number, as in '12345', and {network} is network name. + "network": "A String", # Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/get): `projects/{project}/global/networks/{network}`. "projectId": "A String", # Required. Project id used to create forwarding rule. "state": "A String", # Output only. The state of the PSC service automation. }, @@ -643,6 +662,16 @@

Method Details

"projectAllowlist": [ # A list of Projects from which the forwarding rule will target the service attachment. "A String", ], + "pscAutomationConfigs": [ # Optional. List of projects and networks where the PSC endpoints will be created. This field is used by Online Inference(Prediction) only. + { # PSC config that is used to automatically create PSC endpoints in the user projects. + "errorMessage": "A String", # Output only. Error message if the PSC service automation failed. + "forwardingRule": "A String", # Output only. Forwarding rule created by the PSC service automation. + "ipAddress": "A String", # Output only. IP address rule created by the PSC service automation. + "network": "A String", # Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/get): `projects/{project}/global/networks/{network}`. + "projectId": "A String", # Required. Project id used to create forwarding rule. + "state": "A String", # Output only. The state of the PSC service automation. + }, + ], "serviceAttachment": "A String", # Output only. The name of the generated service attachment resource. This is only populated if the endpoint is deployed with PrivateServiceConnect. }, "publicEndpointDomainName": "A String", # Output only. If public_endpoint_enabled is true, this field will be populated with the domain name to use for this index endpoint. @@ -689,6 +718,9 @@

Method Details

"target": 42, # The target resource utilization in percentage (1% - 100%) for the given metric; once the real usage deviates from the target by a certain percentage, the machine replicas change. The default value is 60 (representing 60%) if not provided. }, ], + "flexStart": { # FlexStart is used to schedule the deployment workload on DWS resource. It contains the max duration of the deployment. # Optional. Immutable. If set, use DWS resource to schedule the deployment workload. reference: (https://cloud.google.com/blog/products/compute/introducing-dynamic-workload-scheduler) + "maxRuntimeDuration": "A String", # The max duration of the deployment is max_runtime_duration. The deployment will be terminated after the duration. The max_runtime_duration can be set up to 7 days. + }, "machineSpec": { # Specification of a single machine. # Required. Immutable. The specification of a single machine being used. "acceleratorCount": 42, # The number of accelerators to attach to the machine. "acceleratorType": "A String", # Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count. @@ -736,11 +768,11 @@

Method Details

"serviceAttachment": "A String", # Output only. The name of the service attachment resource. Populated if private service connect is enabled. }, "pscAutomationConfigs": [ # Optional. If set for PSC deployed index, PSC connection will be automatically created after deployment is done and the endpoint information is populated in private_endpoints.psc_automated_endpoints. - { # PSC config that is used to automatically create forwarding rule via ServiceConnectionMap. + { # PSC config that is used to automatically create PSC endpoints in the user projects. "errorMessage": "A String", # Output only. Error message if the PSC service automation failed. "forwardingRule": "A String", # Output only. Forwarding rule created by the PSC service automation. "ipAddress": "A String", # Output only. IP address rule created by the PSC service automation. - "network": "A String", # Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/insert): `projects/{project}/global/networks/{network}`. Where {project} is a project number, as in '12345', and {network} is network name. + "network": "A String", # Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/get): `projects/{project}/global/networks/{network}`. "projectId": "A String", # Required. Project id used to create forwarding rule. "state": "A String", # Output only. The state of the PSC service automation. }, @@ -768,6 +800,16 @@

Method Details

"projectAllowlist": [ # A list of Projects from which the forwarding rule will target the service attachment. "A String", ], + "pscAutomationConfigs": [ # Optional. List of projects and networks where the PSC endpoints will be created. This field is used by Online Inference(Prediction) only. + { # PSC config that is used to automatically create PSC endpoints in the user projects. + "errorMessage": "A String", # Output only. Error message if the PSC service automation failed. + "forwardingRule": "A String", # Output only. Forwarding rule created by the PSC service automation. + "ipAddress": "A String", # Output only. IP address rule created by the PSC service automation. + "network": "A String", # Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/get): `projects/{project}/global/networks/{network}`. + "projectId": "A String", # Required. Project id used to create forwarding rule. + "state": "A String", # Output only. The state of the PSC service automation. + }, + ], "serviceAttachment": "A String", # Output only. The name of the generated service attachment resource. This is only populated if the endpoint is deployed with PrivateServiceConnect. }, "publicEndpointDomainName": "A String", # Output only. If public_endpoint_enabled is true, this field will be populated with the domain name to use for this index endpoint. @@ -817,6 +859,9 @@

Method Details

"target": 42, # The target resource utilization in percentage (1% - 100%) for the given metric; once the real usage deviates from the target by a certain percentage, the machine replicas change. The default value is 60 (representing 60%) if not provided. }, ], + "flexStart": { # FlexStart is used to schedule the deployment workload on DWS resource. It contains the max duration of the deployment. # Optional. Immutable. If set, use DWS resource to schedule the deployment workload. reference: (https://cloud.google.com/blog/products/compute/introducing-dynamic-workload-scheduler) + "maxRuntimeDuration": "A String", # The max duration of the deployment is max_runtime_duration. The deployment will be terminated after the duration. The max_runtime_duration can be set up to 7 days. + }, "machineSpec": { # Specification of a single machine. # Required. Immutable. The specification of a single machine being used. "acceleratorCount": 42, # The number of accelerators to attach to the machine. "acceleratorType": "A String", # Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count. @@ -864,11 +909,11 @@

Method Details

"serviceAttachment": "A String", # Output only. The name of the service attachment resource. Populated if private service connect is enabled. }, "pscAutomationConfigs": [ # Optional. If set for PSC deployed index, PSC connection will be automatically created after deployment is done and the endpoint information is populated in private_endpoints.psc_automated_endpoints. - { # PSC config that is used to automatically create forwarding rule via ServiceConnectionMap. + { # PSC config that is used to automatically create PSC endpoints in the user projects. "errorMessage": "A String", # Output only. Error message if the PSC service automation failed. "forwardingRule": "A String", # Output only. Forwarding rule created by the PSC service automation. "ipAddress": "A String", # Output only. IP address rule created by the PSC service automation. - "network": "A String", # Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/insert): `projects/{project}/global/networks/{network}`. Where {project} is a project number, as in '12345', and {network} is network name. + "network": "A String", # Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/get): `projects/{project}/global/networks/{network}`. "projectId": "A String", # Required. Project id used to create forwarding rule. "state": "A String", # Output only. The state of the PSC service automation. }, @@ -932,6 +977,9 @@

Method Details

"target": 42, # The target resource utilization in percentage (1% - 100%) for the given metric; once the real usage deviates from the target by a certain percentage, the machine replicas change. The default value is 60 (representing 60%) if not provided. }, ], + "flexStart": { # FlexStart is used to schedule the deployment workload on DWS resource. It contains the max duration of the deployment. # Optional. Immutable. If set, use DWS resource to schedule the deployment workload. reference: (https://cloud.google.com/blog/products/compute/introducing-dynamic-workload-scheduler) + "maxRuntimeDuration": "A String", # The max duration of the deployment is max_runtime_duration. The deployment will be terminated after the duration. The max_runtime_duration can be set up to 7 days. + }, "machineSpec": { # Specification of a single machine. # Required. Immutable. The specification of a single machine being used. "acceleratorCount": 42, # The number of accelerators to attach to the machine. "acceleratorType": "A String", # Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count. @@ -979,11 +1027,11 @@

Method Details

"serviceAttachment": "A String", # Output only. The name of the service attachment resource. Populated if private service connect is enabled. }, "pscAutomationConfigs": [ # Optional. If set for PSC deployed index, PSC connection will be automatically created after deployment is done and the endpoint information is populated in private_endpoints.psc_automated_endpoints. - { # PSC config that is used to automatically create forwarding rule via ServiceConnectionMap. + { # PSC config that is used to automatically create PSC endpoints in the user projects. "errorMessage": "A String", # Output only. Error message if the PSC service automation failed. "forwardingRule": "A String", # Output only. Forwarding rule created by the PSC service automation. "ipAddress": "A String", # Output only. IP address rule created by the PSC service automation. - "network": "A String", # Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/insert): `projects/{project}/global/networks/{network}`. Where {project} is a project number, as in '12345', and {network} is network name. + "network": "A String", # Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/get): `projects/{project}/global/networks/{network}`. "projectId": "A String", # Required. Project id used to create forwarding rule. "state": "A String", # Output only. The state of the PSC service automation. }, @@ -1011,6 +1059,16 @@

Method Details

"projectAllowlist": [ # A list of Projects from which the forwarding rule will target the service attachment. "A String", ], + "pscAutomationConfigs": [ # Optional. List of projects and networks where the PSC endpoints will be created. This field is used by Online Inference(Prediction) only. + { # PSC config that is used to automatically create PSC endpoints in the user projects. + "errorMessage": "A String", # Output only. Error message if the PSC service automation failed. + "forwardingRule": "A String", # Output only. Forwarding rule created by the PSC service automation. + "ipAddress": "A String", # Output only. IP address rule created by the PSC service automation. + "network": "A String", # Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/get): `projects/{project}/global/networks/{network}`. + "projectId": "A String", # Required. Project id used to create forwarding rule. + "state": "A String", # Output only. The state of the PSC service automation. + }, + ], "serviceAttachment": "A String", # Output only. The name of the generated service attachment resource. This is only populated if the endpoint is deployed with PrivateServiceConnect. }, "publicEndpointDomainName": "A String", # Output only. If public_endpoint_enabled is true, this field will be populated with the domain name to use for this index endpoint. @@ -1045,6 +1103,9 @@

Method Details

"target": 42, # The target resource utilization in percentage (1% - 100%) for the given metric; once the real usage deviates from the target by a certain percentage, the machine replicas change. The default value is 60 (representing 60%) if not provided. }, ], + "flexStart": { # FlexStart is used to schedule the deployment workload on DWS resource. It contains the max duration of the deployment. # Optional. Immutable. If set, use DWS resource to schedule the deployment workload. reference: (https://cloud.google.com/blog/products/compute/introducing-dynamic-workload-scheduler) + "maxRuntimeDuration": "A String", # The max duration of the deployment is max_runtime_duration. The deployment will be terminated after the duration. The max_runtime_duration can be set up to 7 days. + }, "machineSpec": { # Specification of a single machine. # Required. Immutable. The specification of a single machine being used. "acceleratorCount": 42, # The number of accelerators to attach to the machine. "acceleratorType": "A String", # Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count. @@ -1092,11 +1153,11 @@

Method Details

"serviceAttachment": "A String", # Output only. The name of the service attachment resource. Populated if private service connect is enabled. }, "pscAutomationConfigs": [ # Optional. If set for PSC deployed index, PSC connection will be automatically created after deployment is done and the endpoint information is populated in private_endpoints.psc_automated_endpoints. - { # PSC config that is used to automatically create forwarding rule via ServiceConnectionMap. + { # PSC config that is used to automatically create PSC endpoints in the user projects. "errorMessage": "A String", # Output only. Error message if the PSC service automation failed. "forwardingRule": "A String", # Output only. Forwarding rule created by the PSC service automation. "ipAddress": "A String", # Output only. IP address rule created by the PSC service automation. - "network": "A String", # Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/insert): `projects/{project}/global/networks/{network}`. Where {project} is a project number, as in '12345', and {network} is network name. + "network": "A String", # Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/get): `projects/{project}/global/networks/{network}`. "projectId": "A String", # Required. Project id used to create forwarding rule. "state": "A String", # Output only. The state of the PSC service automation. }, @@ -1124,6 +1185,16 @@

Method Details

"projectAllowlist": [ # A list of Projects from which the forwarding rule will target the service attachment. "A String", ], + "pscAutomationConfigs": [ # Optional. List of projects and networks where the PSC endpoints will be created. This field is used by Online Inference(Prediction) only. + { # PSC config that is used to automatically create PSC endpoints in the user projects. + "errorMessage": "A String", # Output only. Error message if the PSC service automation failed. + "forwardingRule": "A String", # Output only. Forwarding rule created by the PSC service automation. + "ipAddress": "A String", # Output only. IP address rule created by the PSC service automation. + "network": "A String", # Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/get): `projects/{project}/global/networks/{network}`. + "projectId": "A String", # Required. Project id used to create forwarding rule. + "state": "A String", # Output only. The state of the PSC service automation. + }, + ], "serviceAttachment": "A String", # Output only. The name of the generated service attachment resource. This is only populated if the endpoint is deployed with PrivateServiceConnect. }, "publicEndpointDomainName": "A String", # Output only. If public_endpoint_enabled is true, this field will be populated with the domain name to use for this index endpoint. diff --git a/docs/dyn/aiplatform_v1beta1.projects.locations.models.html b/docs/dyn/aiplatform_v1beta1.projects.locations.models.html index f37048db5c..e114779856 100644 --- a/docs/dyn/aiplatform_v1beta1.projects.locations.models.html +++ b/docs/dyn/aiplatform_v1beta1.projects.locations.models.html @@ -400,7 +400,7 @@

Method Details

}, "healthRoute": "A String", # Immutable. HTTP path on the container to send health checks to. Vertex AI intermittently sends GET requests to this path on the container's IP address and port to check that the container is healthy. Read more about [health checks](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#health). For example, if you set this field to `/bar`, then Vertex AI intermittently sends a GET request to the `/bar` path on the port of your container specified by the first value of this `ModelContainerSpec`'s ports field. If you don't specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/ DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) "imageUri": "A String", # Required. Immutable. URI of the Docker image to be used as the custom container for serving predictions. This URI must identify an image in Artifact Registry or Container Registry. Learn more about the [container publishing requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#publishing), including permissions requirements for the Vertex AI Service Agent. The container image is ingested upon ModelService.UploadModel, stored internally, and this original path is afterwards not used. To learn about the requirements for the Docker image itself, see [Custom container requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#). You can use the URI to one of Vertex AI's [pre-built container images for prediction](https://cloud.google.com/vertex-ai/docs/predictions/pre-built-containers) in this field. - "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with [PredictionService.Invoke] eg: "/invoke/foo/bar". Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. + "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with invoke http call eg: "/invoke/foo/bar", however the [PredictionService.Invoke] RPC is not supported yet. Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. "livenessProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes liveness probe. "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. @@ -778,7 +778,7 @@

Method Details

}, "healthRoute": "A String", # Immutable. HTTP path on the container to send health checks to. Vertex AI intermittently sends GET requests to this path on the container's IP address and port to check that the container is healthy. Read more about [health checks](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#health). For example, if you set this field to `/bar`, then Vertex AI intermittently sends a GET request to the `/bar` path on the port of your container specified by the first value of this `ModelContainerSpec`'s ports field. If you don't specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/ DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) "imageUri": "A String", # Required. Immutable. URI of the Docker image to be used as the custom container for serving predictions. This URI must identify an image in Artifact Registry or Container Registry. Learn more about the [container publishing requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#publishing), including permissions requirements for the Vertex AI Service Agent. The container image is ingested upon ModelService.UploadModel, stored internally, and this original path is afterwards not used. To learn about the requirements for the Docker image itself, see [Custom container requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#). You can use the URI to one of Vertex AI's [pre-built container images for prediction](https://cloud.google.com/vertex-ai/docs/predictions/pre-built-containers) in this field. - "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with [PredictionService.Invoke] eg: "/invoke/foo/bar". Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. + "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with invoke http call eg: "/invoke/foo/bar", however the [PredictionService.Invoke] RPC is not supported yet. Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. "livenessProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes liveness probe. "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. @@ -1167,7 +1167,7 @@

Method Details

}, "healthRoute": "A String", # Immutable. HTTP path on the container to send health checks to. Vertex AI intermittently sends GET requests to this path on the container's IP address and port to check that the container is healthy. Read more about [health checks](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#health). For example, if you set this field to `/bar`, then Vertex AI intermittently sends a GET request to the `/bar` path on the port of your container specified by the first value of this `ModelContainerSpec`'s ports field. If you don't specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/ DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) "imageUri": "A String", # Required. Immutable. URI of the Docker image to be used as the custom container for serving predictions. This URI must identify an image in Artifact Registry or Container Registry. Learn more about the [container publishing requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#publishing), including permissions requirements for the Vertex AI Service Agent. The container image is ingested upon ModelService.UploadModel, stored internally, and this original path is afterwards not used. To learn about the requirements for the Docker image itself, see [Custom container requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#). You can use the URI to one of Vertex AI's [pre-built container images for prediction](https://cloud.google.com/vertex-ai/docs/predictions/pre-built-containers) in this field. - "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with [PredictionService.Invoke] eg: "/invoke/foo/bar". Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. + "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with invoke http call eg: "/invoke/foo/bar", however the [PredictionService.Invoke] RPC is not supported yet. Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. "livenessProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes liveness probe. "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. @@ -1544,7 +1544,7 @@

Method Details

}, "healthRoute": "A String", # Immutable. HTTP path on the container to send health checks to. Vertex AI intermittently sends GET requests to this path on the container's IP address and port to check that the container is healthy. Read more about [health checks](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#health). For example, if you set this field to `/bar`, then Vertex AI intermittently sends a GET request to the `/bar` path on the port of your container specified by the first value of this `ModelContainerSpec`'s ports field. If you don't specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/ DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) "imageUri": "A String", # Required. Immutable. URI of the Docker image to be used as the custom container for serving predictions. This URI must identify an image in Artifact Registry or Container Registry. Learn more about the [container publishing requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#publishing), including permissions requirements for the Vertex AI Service Agent. The container image is ingested upon ModelService.UploadModel, stored internally, and this original path is afterwards not used. To learn about the requirements for the Docker image itself, see [Custom container requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#). You can use the URI to one of Vertex AI's [pre-built container images for prediction](https://cloud.google.com/vertex-ai/docs/predictions/pre-built-containers) in this field. - "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with [PredictionService.Invoke] eg: "/invoke/foo/bar". Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. + "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with invoke http call eg: "/invoke/foo/bar", however the [PredictionService.Invoke] RPC is not supported yet. Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. "livenessProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes liveness probe. "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. @@ -1876,7 +1876,7 @@

Method Details

}, "healthRoute": "A String", # Immutable. HTTP path on the container to send health checks to. Vertex AI intermittently sends GET requests to this path on the container's IP address and port to check that the container is healthy. Read more about [health checks](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#health). For example, if you set this field to `/bar`, then Vertex AI intermittently sends a GET request to the `/bar` path on the port of your container specified by the first value of this `ModelContainerSpec`'s ports field. If you don't specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/ DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) "imageUri": "A String", # Required. Immutable. URI of the Docker image to be used as the custom container for serving predictions. This URI must identify an image in Artifact Registry or Container Registry. Learn more about the [container publishing requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#publishing), including permissions requirements for the Vertex AI Service Agent. The container image is ingested upon ModelService.UploadModel, stored internally, and this original path is afterwards not used. To learn about the requirements for the Docker image itself, see [Custom container requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#). You can use the URI to one of Vertex AI's [pre-built container images for prediction](https://cloud.google.com/vertex-ai/docs/predictions/pre-built-containers) in this field. - "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with [PredictionService.Invoke] eg: "/invoke/foo/bar". Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. + "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with invoke http call eg: "/invoke/foo/bar", however the [PredictionService.Invoke] RPC is not supported yet. Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. "livenessProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes liveness probe. "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. @@ -2207,7 +2207,7 @@

Method Details

}, "healthRoute": "A String", # Immutable. HTTP path on the container to send health checks to. Vertex AI intermittently sends GET requests to this path on the container's IP address and port to check that the container is healthy. Read more about [health checks](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#health). For example, if you set this field to `/bar`, then Vertex AI intermittently sends a GET request to the `/bar` path on the port of your container specified by the first value of this `ModelContainerSpec`'s ports field. If you don't specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/ DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) "imageUri": "A String", # Required. Immutable. URI of the Docker image to be used as the custom container for serving predictions. This URI must identify an image in Artifact Registry or Container Registry. Learn more about the [container publishing requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#publishing), including permissions requirements for the Vertex AI Service Agent. The container image is ingested upon ModelService.UploadModel, stored internally, and this original path is afterwards not used. To learn about the requirements for the Docker image itself, see [Custom container requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#). You can use the URI to one of Vertex AI's [pre-built container images for prediction](https://cloud.google.com/vertex-ai/docs/predictions/pre-built-containers) in this field. - "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with [PredictionService.Invoke] eg: "/invoke/foo/bar". Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. + "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with invoke http call eg: "/invoke/foo/bar", however the [PredictionService.Invoke] RPC is not supported yet. Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. "livenessProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes liveness probe. "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. @@ -2682,7 +2682,7 @@

Method Details

}, "healthRoute": "A String", # Immutable. HTTP path on the container to send health checks to. Vertex AI intermittently sends GET requests to this path on the container's IP address and port to check that the container is healthy. Read more about [health checks](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#health). For example, if you set this field to `/bar`, then Vertex AI intermittently sends a GET request to the `/bar` path on the port of your container specified by the first value of this `ModelContainerSpec`'s ports field. If you don't specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/ DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) "imageUri": "A String", # Required. Immutable. URI of the Docker image to be used as the custom container for serving predictions. This URI must identify an image in Artifact Registry or Container Registry. Learn more about the [container publishing requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#publishing), including permissions requirements for the Vertex AI Service Agent. The container image is ingested upon ModelService.UploadModel, stored internally, and this original path is afterwards not used. To learn about the requirements for the Docker image itself, see [Custom container requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#). You can use the URI to one of Vertex AI's [pre-built container images for prediction](https://cloud.google.com/vertex-ai/docs/predictions/pre-built-containers) in this field. - "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with [PredictionService.Invoke] eg: "/invoke/foo/bar". Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. + "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with invoke http call eg: "/invoke/foo/bar", however the [PredictionService.Invoke] RPC is not supported yet. Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. "livenessProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes liveness probe. "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. diff --git a/docs/dyn/aiplatform_v1beta1.projects.locations.pipelineJobs.html b/docs/dyn/aiplatform_v1beta1.projects.locations.pipelineJobs.html index 04ea4afd8a..afc771480d 100644 --- a/docs/dyn/aiplatform_v1beta1.projects.locations.pipelineJobs.html +++ b/docs/dyn/aiplatform_v1beta1.projects.locations.pipelineJobs.html @@ -404,6 +404,7 @@

Method Details

"state": "A String", # Output only. State of the task. "taskId": "A String", # Output only. The system generated ID of the task. "taskName": "A String", # Output only. The user specified name of the task that is defined in pipeline_spec. + "taskUniqueName": "A String", # Output only. The unique name of a task. This field is used by rerun pipeline job. Console UI and Vertex AI SDK will support triggering pipeline job reruns. The name is constructed by concatenating all the parent tasks name with the task name. For example, if a task named "child_task" has a parent task named "parent_task_1" and parent task 1 has a parent task named "parent_task_2", the task unique name will be "parent_task_2.parent_task_1.child_task". }, ], }, @@ -695,6 +696,7 @@

Method Details

"state": "A String", # Output only. State of the task. "taskId": "A String", # Output only. The system generated ID of the task. "taskName": "A String", # Output only. The user specified name of the task that is defined in pipeline_spec. + "taskUniqueName": "A String", # Output only. The unique name of a task. This field is used by rerun pipeline job. Console UI and Vertex AI SDK will support triggering pipeline job reruns. The name is constructed by concatenating all the parent tasks name with the task name. For example, if a task named "child_task" has a parent task named "parent_task_1" and parent task 1 has a parent task named "parent_task_2", the task unique name will be "parent_task_2.parent_task_1.child_task". }, ], }, @@ -1027,6 +1029,7 @@

Method Details

"state": "A String", # Output only. State of the task. "taskId": "A String", # Output only. The system generated ID of the task. "taskName": "A String", # Output only. The user specified name of the task that is defined in pipeline_spec. + "taskUniqueName": "A String", # Output only. The unique name of a task. This field is used by rerun pipeline job. Console UI and Vertex AI SDK will support triggering pipeline job reruns. The name is constructed by concatenating all the parent tasks name with the task name. For example, if a task named "child_task" has a parent task named "parent_task_1" and parent task 1 has a parent task named "parent_task_2", the task unique name will be "parent_task_2.parent_task_1.child_task". }, ], }, @@ -1332,6 +1335,7 @@

Method Details

"state": "A String", # Output only. State of the task. "taskId": "A String", # Output only. The system generated ID of the task. "taskName": "A String", # Output only. The user specified name of the task that is defined in pipeline_spec. + "taskUniqueName": "A String", # Output only. The unique name of a task. This field is used by rerun pipeline job. Console UI and Vertex AI SDK will support triggering pipeline job reruns. The name is constructed by concatenating all the parent tasks name with the task name. For example, if a task named "child_task" has a parent task named "parent_task_1" and parent task 1 has a parent task named "parent_task_2", the task unique name will be "parent_task_2.parent_task_1.child_task". }, ], }, diff --git a/docs/dyn/aiplatform_v1beta1.projects.locations.publishers.models.html b/docs/dyn/aiplatform_v1beta1.projects.locations.publishers.models.html index 4720cc1e06..004ae13ede 100644 --- a/docs/dyn/aiplatform_v1beta1.projects.locations.publishers.models.html +++ b/docs/dyn/aiplatform_v1beta1.projects.locations.publishers.models.html @@ -350,7 +350,7 @@

Method Details

"temperature": 3.14, # Optional. Controls the randomness of predictions. "thinkingConfig": { # Config for thinking features. # Optional. Config for thinking features. An error will be returned if this field is set for models that don't support thinking. "includeThoughts": True or False, # Optional. Indicates whether to include thoughts in the response. If true, thoughts are returned only when available. - "thinkingBudget": 42, # Optional. Indicates the thinking budget in tokens. This is only applied when enable_thinking is true. + "thinkingBudget": 42, # Optional. Indicates the thinking budget in tokens. }, "topK": 3.14, # Optional. If specified, top-k sampling will be used. "topP": 3.14, # Optional. If specified, nucleus sampling will be used. @@ -877,7 +877,7 @@

Method Details

"temperature": 3.14, # Optional. Controls the randomness of predictions. "thinkingConfig": { # Config for thinking features. # Optional. Config for thinking features. An error will be returned if this field is set for models that don't support thinking. "includeThoughts": True or False, # Optional. Indicates whether to include thoughts in the response. If true, thoughts are returned only when available. - "thinkingBudget": 42, # Optional. Indicates the thinking budget in tokens. This is only applied when enable_thinking is true. + "thinkingBudget": 42, # Optional. Indicates the thinking budget in tokens. }, "topK": 3.14, # Optional. If specified, top-k sampling will be used. "topP": 3.14, # Optional. If specified, nucleus sampling will be used. @@ -1911,7 +1911,7 @@

Method Details

"temperature": 3.14, # Optional. Controls the randomness of predictions. "thinkingConfig": { # Config for thinking features. # Optional. Config for thinking features. An error will be returned if this field is set for models that don't support thinking. "includeThoughts": True or False, # Optional. Indicates whether to include thoughts in the response. If true, thoughts are returned only when available. - "thinkingBudget": 42, # Optional. Indicates the thinking budget in tokens. This is only applied when enable_thinking is true. + "thinkingBudget": 42, # Optional. Indicates the thinking budget in tokens. }, "topK": 3.14, # Optional. If specified, top-k sampling will be used. "topP": 3.14, # Optional. If specified, nucleus sampling will be used. diff --git a/docs/dyn/aiplatform_v1beta1.projects.locations.reasoningEngines.html b/docs/dyn/aiplatform_v1beta1.projects.locations.reasoningEngines.html index d3771f91bb..72eed5523c 100644 --- a/docs/dyn/aiplatform_v1beta1.projects.locations.reasoningEngines.html +++ b/docs/dyn/aiplatform_v1beta1.projects.locations.reasoningEngines.html @@ -145,10 +145,10 @@

Method Details

"contextSpec": { # Configuration for how Agent Engine sub-resources should manage context. # Optional. Configuration for how Agent Engine sub-resources should manage context. "memoryBankConfig": { # Specification for a Memory Bank. # Optional. Specification for a Memory Bank, which manages memories for the Agent Engine. "generationConfig": { # Configuration for how to generate memories. # Optional. Configuration for how to generate memories for the Memory Bank. - "model": "A String", # Required. The model used to generate memories. Format: `projects/{project}/locations/{location}/publishers/google/models/{model}` or `projects/{project}/locations/{location}/endpoints/{endpoint}`. + "model": "A String", # Required. The model used to generate memories. Format: `projects/{project}/locations/{location}/publishers/google/models/{model}`. }, "similaritySearchConfig": { # Configuration for how to perform similarity search on memories. # Optional. Configuration for how to perform similarity search on memories. If not set, the Memory Bank will use the default embedding model `text-embedding-005`. - "embeddingModel": "A String", # Required. The model used to generate embeddings to lookup similar memories. Format: `projects/{project}/locations/{location}/publishers/google/models/{model}` or `projects/{project}/locations/{location}/endpoints/{endpoint}`. + "embeddingModel": "A String", # Required. The model used to generate embeddings to lookup similar memories. Format: `projects/{project}/locations/{location}/publishers/google/models/{model}`. }, }, }, @@ -274,10 +274,10 @@

Method Details

"contextSpec": { # Configuration for how Agent Engine sub-resources should manage context. # Optional. Configuration for how Agent Engine sub-resources should manage context. "memoryBankConfig": { # Specification for a Memory Bank. # Optional. Specification for a Memory Bank, which manages memories for the Agent Engine. "generationConfig": { # Configuration for how to generate memories. # Optional. Configuration for how to generate memories for the Memory Bank. - "model": "A String", # Required. The model used to generate memories. Format: `projects/{project}/locations/{location}/publishers/google/models/{model}` or `projects/{project}/locations/{location}/endpoints/{endpoint}`. + "model": "A String", # Required. The model used to generate memories. Format: `projects/{project}/locations/{location}/publishers/google/models/{model}`. }, "similaritySearchConfig": { # Configuration for how to perform similarity search on memories. # Optional. Configuration for how to perform similarity search on memories. If not set, the Memory Bank will use the default embedding model `text-embedding-005`. - "embeddingModel": "A String", # Required. The model used to generate embeddings to lookup similar memories. Format: `projects/{project}/locations/{location}/publishers/google/models/{model}` or `projects/{project}/locations/{location}/endpoints/{endpoint}`. + "embeddingModel": "A String", # Required. The model used to generate embeddings to lookup similar memories. Format: `projects/{project}/locations/{location}/publishers/google/models/{model}`. }, }, }, @@ -345,10 +345,10 @@

Method Details

"contextSpec": { # Configuration for how Agent Engine sub-resources should manage context. # Optional. Configuration for how Agent Engine sub-resources should manage context. "memoryBankConfig": { # Specification for a Memory Bank. # Optional. Specification for a Memory Bank, which manages memories for the Agent Engine. "generationConfig": { # Configuration for how to generate memories. # Optional. Configuration for how to generate memories for the Memory Bank. - "model": "A String", # Required. The model used to generate memories. Format: `projects/{project}/locations/{location}/publishers/google/models/{model}` or `projects/{project}/locations/{location}/endpoints/{endpoint}`. + "model": "A String", # Required. The model used to generate memories. Format: `projects/{project}/locations/{location}/publishers/google/models/{model}`. }, "similaritySearchConfig": { # Configuration for how to perform similarity search on memories. # Optional. Configuration for how to perform similarity search on memories. If not set, the Memory Bank will use the default embedding model `text-embedding-005`. - "embeddingModel": "A String", # Required. The model used to generate embeddings to lookup similar memories. Format: `projects/{project}/locations/{location}/publishers/google/models/{model}` or `projects/{project}/locations/{location}/endpoints/{endpoint}`. + "embeddingModel": "A String", # Required. The model used to generate embeddings to lookup similar memories. Format: `projects/{project}/locations/{location}/publishers/google/models/{model}`. }, }, }, @@ -421,10 +421,10 @@

Method Details

"contextSpec": { # Configuration for how Agent Engine sub-resources should manage context. # Optional. Configuration for how Agent Engine sub-resources should manage context. "memoryBankConfig": { # Specification for a Memory Bank. # Optional. Specification for a Memory Bank, which manages memories for the Agent Engine. "generationConfig": { # Configuration for how to generate memories. # Optional. Configuration for how to generate memories for the Memory Bank. - "model": "A String", # Required. The model used to generate memories. Format: `projects/{project}/locations/{location}/publishers/google/models/{model}` or `projects/{project}/locations/{location}/endpoints/{endpoint}`. + "model": "A String", # Required. The model used to generate memories. Format: `projects/{project}/locations/{location}/publishers/google/models/{model}`. }, "similaritySearchConfig": { # Configuration for how to perform similarity search on memories. # Optional. Configuration for how to perform similarity search on memories. If not set, the Memory Bank will use the default embedding model `text-embedding-005`. - "embeddingModel": "A String", # Required. The model used to generate embeddings to lookup similar memories. Format: `projects/{project}/locations/{location}/publishers/google/models/{model}` or `projects/{project}/locations/{location}/endpoints/{endpoint}`. + "embeddingModel": "A String", # Required. The model used to generate embeddings to lookup similar memories. Format: `projects/{project}/locations/{location}/publishers/google/models/{model}`. }, }, }, diff --git a/docs/dyn/aiplatform_v1beta1.projects.locations.reasoningEngines.sessions.events.html b/docs/dyn/aiplatform_v1beta1.projects.locations.reasoningEngines.sessions.events.html index b78f7483e8..8d37212842 100644 --- a/docs/dyn/aiplatform_v1beta1.projects.locations.reasoningEngines.sessions.events.html +++ b/docs/dyn/aiplatform_v1beta1.projects.locations.reasoningEngines.sessions.events.html @@ -78,7 +78,7 @@

Instance Methods

close()

Close httplib2 connections.

- list(parent, pageSize=None, pageToken=None, x__xgafv=None)

+ list(parent, filter=None, pageSize=None, pageToken=None, x__xgafv=None)

Lists Events in a given session.

list_next()

@@ -90,11 +90,12 @@

Method Details

- list(parent, pageSize=None, pageToken=None, x__xgafv=None) + list(parent, filter=None, pageSize=None, pageToken=None, x__xgafv=None)
Lists Events in a given session.
 
 Args:
   parent: string, Required. The resource name of the session to list events from. Format: `projects/{project}/locations/{location}/reasoningEngines/{reasoning_engine}/sessions/{session}` (required)
+  filter: string, Optional. The standard list filter. Supported fields: * `timestamp` range (i.e. `timestamp>="2025-01-31T11:30:00-04:00"` where the timestamp is in RFC 3339 format) More detail in [AIP-160](https://google.aip.dev/160).
   pageSize: integer, Optional. The maximum number of events to return. The service may return fewer than this value. If unspecified, at most 100 events will be returned. These events are ordered by timestamp in ascending order.
   pageToken: string, Optional. The next_page_token value returned from a previous list SessionService.ListEvents call.
   x__xgafv: string, V1 error format.
@@ -175,6 +176,9 @@ 

Method Details

"errorMessage": "A String", # Optional. Error message if the response is an error. "eventMetadata": { # Metadata relating to a LLM response event. # Optional. Metadata relating to this event. "branch": "A String", # Optional. The branch of the event. The format is like agent_1.agent_2.agent_3, where agent_1 is the parent of agent_2, and agent_2 is the parent of agent_3. Branch is used when multiple child agents shouldn't see their siblings' conversation history. + "customMetadata": { # The custom metadata of the LlmResponse. + "a_key": "", # Properties of the object. + }, "groundingMetadata": { # Metadata returned to client when grounding is enabled. # Optional. Metadata returned to client when grounding is enabled. "groundingChunks": [ # List of supporting references retrieved from specified grounding source. { # Grounding chunk. diff --git a/docs/dyn/aiplatform_v1beta1.projects.locations.reasoningEngines.sessions.html b/docs/dyn/aiplatform_v1beta1.projects.locations.reasoningEngines.sessions.html index ec5f2c823a..8c59151bf5 100644 --- a/docs/dyn/aiplatform_v1beta1.projects.locations.reasoningEngines.sessions.html +++ b/docs/dyn/aiplatform_v1beta1.projects.locations.reasoningEngines.sessions.html @@ -185,6 +185,9 @@

Method Details

"errorMessage": "A String", # Optional. Error message if the response is an error. "eventMetadata": { # Metadata relating to a LLM response event. # Optional. Metadata relating to this event. "branch": "A String", # Optional. The branch of the event. The format is like agent_1.agent_2.agent_3, where agent_1 is the parent of agent_2, and agent_2 is the parent of agent_3. Branch is used when multiple child agents shouldn't see their siblings' conversation history. + "customMetadata": { # The custom metadata of the LlmResponse. + "a_key": "", # Properties of the object. + }, "groundingMetadata": { # Metadata returned to client when grounding is enabled. # Optional. Metadata returned to client when grounding is enabled. "groundingChunks": [ # List of supporting references retrieved from specified grounding source. { # Grounding chunk. diff --git a/docs/dyn/aiplatform_v1beta1.projects.locations.schedules.html b/docs/dyn/aiplatform_v1beta1.projects.locations.schedules.html index 8c46cac382..73c17d17af 100644 --- a/docs/dyn/aiplatform_v1beta1.projects.locations.schedules.html +++ b/docs/dyn/aiplatform_v1beta1.projects.locations.schedules.html @@ -685,6 +685,7 @@

Method Details

"state": "A String", # Output only. State of the task. "taskId": "A String", # Output only. The system generated ID of the task. "taskName": "A String", # Output only. The user specified name of the task that is defined in pipeline_spec. + "taskUniqueName": "A String", # Output only. The unique name of a task. This field is used by rerun pipeline job. Console UI and Vertex AI SDK will support triggering pipeline job reruns. The name is constructed by concatenating all the parent tasks name with the task name. For example, if a task named "child_task" has a parent task named "parent_task_1" and parent task 1 has a parent task named "parent_task_2", the task unique name will be "parent_task_2.parent_task_1.child_task". }, ], }, @@ -1389,6 +1390,7 @@

Method Details

"state": "A String", # Output only. State of the task. "taskId": "A String", # Output only. The system generated ID of the task. "taskName": "A String", # Output only. The user specified name of the task that is defined in pipeline_spec. + "taskUniqueName": "A String", # Output only. The unique name of a task. This field is used by rerun pipeline job. Console UI and Vertex AI SDK will support triggering pipeline job reruns. The name is constructed by concatenating all the parent tasks name with the task name. For example, if a task named "child_task" has a parent task named "parent_task_1" and parent task 1 has a parent task named "parent_task_2", the task unique name will be "parent_task_2.parent_task_1.child_task". }, ], }, @@ -2135,6 +2137,7 @@

Method Details

"state": "A String", # Output only. State of the task. "taskId": "A String", # Output only. The system generated ID of the task. "taskName": "A String", # Output only. The user specified name of the task that is defined in pipeline_spec. + "taskUniqueName": "A String", # Output only. The unique name of a task. This field is used by rerun pipeline job. Console UI and Vertex AI SDK will support triggering pipeline job reruns. The name is constructed by concatenating all the parent tasks name with the task name. For example, if a task named "child_task" has a parent task named "parent_task_1" and parent task 1 has a parent task named "parent_task_2", the task unique name will be "parent_task_2.parent_task_1.child_task". }, ], }, @@ -2853,6 +2856,7 @@

Method Details

"state": "A String", # Output only. State of the task. "taskId": "A String", # Output only. The system generated ID of the task. "taskName": "A String", # Output only. The user specified name of the task that is defined in pipeline_spec. + "taskUniqueName": "A String", # Output only. The unique name of a task. This field is used by rerun pipeline job. Console UI and Vertex AI SDK will support triggering pipeline job reruns. The name is constructed by concatenating all the parent tasks name with the task name. For example, if a task named "child_task" has a parent task named "parent_task_1" and parent task 1 has a parent task named "parent_task_2", the task unique name will be "parent_task_2.parent_task_1.child_task". }, ], }, @@ -3575,6 +3579,7 @@

Method Details

"state": "A String", # Output only. State of the task. "taskId": "A String", # Output only. The system generated ID of the task. "taskName": "A String", # Output only. The user specified name of the task that is defined in pipeline_spec. + "taskUniqueName": "A String", # Output only. The unique name of a task. This field is used by rerun pipeline job. Console UI and Vertex AI SDK will support triggering pipeline job reruns. The name is constructed by concatenating all the parent tasks name with the task name. For example, if a task named "child_task" has a parent task named "parent_task_1" and parent task 1 has a parent task named "parent_task_2", the task unique name will be "parent_task_2.parent_task_1.child_task". }, ], }, @@ -4280,6 +4285,7 @@

Method Details

"state": "A String", # Output only. State of the task. "taskId": "A String", # Output only. The system generated ID of the task. "taskName": "A String", # Output only. The user specified name of the task that is defined in pipeline_spec. + "taskUniqueName": "A String", # Output only. The unique name of a task. This field is used by rerun pipeline job. Console UI and Vertex AI SDK will support triggering pipeline job reruns. The name is constructed by concatenating all the parent tasks name with the task name. For example, if a task named "child_task" has a parent task named "parent_task_1" and parent task 1 has a parent task named "parent_task_2", the task unique name will be "parent_task_2.parent_task_1.child_task". }, ], }, diff --git a/docs/dyn/aiplatform_v1beta1.projects.locations.trainingPipelines.html b/docs/dyn/aiplatform_v1beta1.projects.locations.trainingPipelines.html index 80e144b2b5..5d0329e629 100644 --- a/docs/dyn/aiplatform_v1beta1.projects.locations.trainingPipelines.html +++ b/docs/dyn/aiplatform_v1beta1.projects.locations.trainingPipelines.html @@ -269,7 +269,7 @@

Method Details

}, "healthRoute": "A String", # Immutable. HTTP path on the container to send health checks to. Vertex AI intermittently sends GET requests to this path on the container's IP address and port to check that the container is healthy. Read more about [health checks](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#health). For example, if you set this field to `/bar`, then Vertex AI intermittently sends a GET request to the `/bar` path on the port of your container specified by the first value of this `ModelContainerSpec`'s ports field. If you don't specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/ DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) "imageUri": "A String", # Required. Immutable. URI of the Docker image to be used as the custom container for serving predictions. This URI must identify an image in Artifact Registry or Container Registry. Learn more about the [container publishing requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#publishing), including permissions requirements for the Vertex AI Service Agent. The container image is ingested upon ModelService.UploadModel, stored internally, and this original path is afterwards not used. To learn about the requirements for the Docker image itself, see [Custom container requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#). You can use the URI to one of Vertex AI's [pre-built container images for prediction](https://cloud.google.com/vertex-ai/docs/predictions/pre-built-containers) in this field. - "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with [PredictionService.Invoke] eg: "/invoke/foo/bar". Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. + "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with invoke http call eg: "/invoke/foo/bar", however the [PredictionService.Invoke] RPC is not supported yet. Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. "livenessProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes liveness probe. "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. @@ -666,7 +666,7 @@

Method Details

}, "healthRoute": "A String", # Immutable. HTTP path on the container to send health checks to. Vertex AI intermittently sends GET requests to this path on the container's IP address and port to check that the container is healthy. Read more about [health checks](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#health). For example, if you set this field to `/bar`, then Vertex AI intermittently sends a GET request to the `/bar` path on the port of your container specified by the first value of this `ModelContainerSpec`'s ports field. If you don't specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/ DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) "imageUri": "A String", # Required. Immutable. URI of the Docker image to be used as the custom container for serving predictions. This URI must identify an image in Artifact Registry or Container Registry. Learn more about the [container publishing requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#publishing), including permissions requirements for the Vertex AI Service Agent. The container image is ingested upon ModelService.UploadModel, stored internally, and this original path is afterwards not used. To learn about the requirements for the Docker image itself, see [Custom container requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#). You can use the URI to one of Vertex AI's [pre-built container images for prediction](https://cloud.google.com/vertex-ai/docs/predictions/pre-built-containers) in this field. - "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with [PredictionService.Invoke] eg: "/invoke/foo/bar". Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. + "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with invoke http call eg: "/invoke/foo/bar", however the [PredictionService.Invoke] RPC is not supported yet. Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. "livenessProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes liveness probe. "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. @@ -1105,7 +1105,7 @@

Method Details

}, "healthRoute": "A String", # Immutable. HTTP path on the container to send health checks to. Vertex AI intermittently sends GET requests to this path on the container's IP address and port to check that the container is healthy. Read more about [health checks](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#health). For example, if you set this field to `/bar`, then Vertex AI intermittently sends a GET request to the `/bar` path on the port of your container specified by the first value of this `ModelContainerSpec`'s ports field. If you don't specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/ DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) "imageUri": "A String", # Required. Immutable. URI of the Docker image to be used as the custom container for serving predictions. This URI must identify an image in Artifact Registry or Container Registry. Learn more about the [container publishing requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#publishing), including permissions requirements for the Vertex AI Service Agent. The container image is ingested upon ModelService.UploadModel, stored internally, and this original path is afterwards not used. To learn about the requirements for the Docker image itself, see [Custom container requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#). You can use the URI to one of Vertex AI's [pre-built container images for prediction](https://cloud.google.com/vertex-ai/docs/predictions/pre-built-containers) in this field. - "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with [PredictionService.Invoke] eg: "/invoke/foo/bar". Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. + "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with invoke http call eg: "/invoke/foo/bar", however the [PredictionService.Invoke] RPC is not supported yet. Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. "livenessProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes liveness probe. "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. @@ -1516,7 +1516,7 @@

Method Details

}, "healthRoute": "A String", # Immutable. HTTP path on the container to send health checks to. Vertex AI intermittently sends GET requests to this path on the container's IP address and port to check that the container is healthy. Read more about [health checks](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#health). For example, if you set this field to `/bar`, then Vertex AI intermittently sends a GET request to the `/bar` path on the port of your container specified by the first value of this `ModelContainerSpec`'s ports field. If you don't specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/ DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) "imageUri": "A String", # Required. Immutable. URI of the Docker image to be used as the custom container for serving predictions. This URI must identify an image in Artifact Registry or Container Registry. Learn more about the [container publishing requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#publishing), including permissions requirements for the Vertex AI Service Agent. The container image is ingested upon ModelService.UploadModel, stored internally, and this original path is afterwards not used. To learn about the requirements for the Docker image itself, see [Custom container requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#). You can use the URI to one of Vertex AI's [pre-built container images for prediction](https://cloud.google.com/vertex-ai/docs/predictions/pre-built-containers) in this field. - "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with [PredictionService.Invoke] eg: "/invoke/foo/bar". Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. + "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with invoke http call eg: "/invoke/foo/bar", however the [PredictionService.Invoke] RPC is not supported yet. Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. "livenessProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes liveness probe. "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. diff --git a/docs/dyn/aiplatform_v1beta1.projects.locations.tuningJobs.html b/docs/dyn/aiplatform_v1beta1.projects.locations.tuningJobs.html index ad7f6c6568..a1e55f551b 100644 --- a/docs/dyn/aiplatform_v1beta1.projects.locations.tuningJobs.html +++ b/docs/dyn/aiplatform_v1beta1.projects.locations.tuningJobs.html @@ -207,7 +207,7 @@

Method Details

}, ], "endpoint": "A String", # Output only. A resource name of an Endpoint. Format: `projects/{project}/locations/{location}/endpoints/{endpoint}`. - "model": "A String", # Output only. The resource name of the TunedModel. Format: `projects/{project}/locations/{location}/models/{model}`. + "model": "A String", # Output only. The resource name of the TunedModel. Format: `projects/{project}/locations/{location}/models/{model}@{version_id}` When tuning from a base model, the version_id will be 1. For continuous tuning, the version id will be incremented by 1 from the last version id in the parent model. E.g., `projects/{project}/locations/{location}/models/{model}@{last_version_id + 1}` }, "tunedModelDisplayName": "A String", # Optional. The display name of the TunedModel. The name can be up to 128 characters long and can consist of any UTF-8 characters. "tuningDataStats": { # The tuning data statistic values for TuningJob. # Output only. The tuning data statistics associated with this TuningJob. @@ -508,7 +508,7 @@

Method Details

}, ], "endpoint": "A String", # Output only. A resource name of an Endpoint. Format: `projects/{project}/locations/{location}/endpoints/{endpoint}`. - "model": "A String", # Output only. The resource name of the TunedModel. Format: `projects/{project}/locations/{location}/models/{model}`. + "model": "A String", # Output only. The resource name of the TunedModel. Format: `projects/{project}/locations/{location}/models/{model}@{version_id}` When tuning from a base model, the version_id will be 1. For continuous tuning, the version id will be incremented by 1 from the last version id in the parent model. E.g., `projects/{project}/locations/{location}/models/{model}@{last_version_id + 1}` }, "tunedModelDisplayName": "A String", # Optional. The display name of the TunedModel. The name can be up to 128 characters long and can consist of any UTF-8 characters. "tuningDataStats": { # The tuning data statistic values for TuningJob. # Output only. The tuning data statistics associated with this TuningJob. @@ -816,7 +816,7 @@

Method Details

}, ], "endpoint": "A String", # Output only. A resource name of an Endpoint. Format: `projects/{project}/locations/{location}/endpoints/{endpoint}`. - "model": "A String", # Output only. The resource name of the TunedModel. Format: `projects/{project}/locations/{location}/models/{model}`. + "model": "A String", # Output only. The resource name of the TunedModel. Format: `projects/{project}/locations/{location}/models/{model}@{version_id}` When tuning from a base model, the version_id will be 1. For continuous tuning, the version id will be incremented by 1 from the last version id in the parent model. E.g., `projects/{project}/locations/{location}/models/{model}@{last_version_id + 1}` }, "tunedModelDisplayName": "A String", # Optional. The display name of the TunedModel. The name can be up to 128 characters long and can consist of any UTF-8 characters. "tuningDataStats": { # The tuning data statistic values for TuningJob. # Output only. The tuning data statistics associated with this TuningJob. @@ -1130,7 +1130,7 @@

Method Details

}, ], "endpoint": "A String", # Output only. A resource name of an Endpoint. Format: `projects/{project}/locations/{location}/endpoints/{endpoint}`. - "model": "A String", # Output only. The resource name of the TunedModel. Format: `projects/{project}/locations/{location}/models/{model}`. + "model": "A String", # Output only. The resource name of the TunedModel. Format: `projects/{project}/locations/{location}/models/{model}@{version_id}` When tuning from a base model, the version_id will be 1. For continuous tuning, the version id will be incremented by 1 from the last version id in the parent model. E.g., `projects/{project}/locations/{location}/models/{model}@{last_version_id + 1}` }, "tunedModelDisplayName": "A String", # Optional. The display name of the TunedModel. The name can be up to 128 characters long and can consist of any UTF-8 characters. "tuningDataStats": { # The tuning data statistic values for TuningJob. # Output only. The tuning data statistics associated with this TuningJob. @@ -1459,7 +1459,7 @@

Method Details

}, ], "endpoint": "A String", # Output only. A resource name of an Endpoint. Format: `projects/{project}/locations/{location}/endpoints/{endpoint}`. - "model": "A String", # Output only. The resource name of the TunedModel. Format: `projects/{project}/locations/{location}/models/{model}`. + "model": "A String", # Output only. The resource name of the TunedModel. Format: `projects/{project}/locations/{location}/models/{model}@{version_id}` When tuning from a base model, the version_id will be 1. For continuous tuning, the version id will be incremented by 1 from the last version id in the parent model. E.g., `projects/{project}/locations/{location}/models/{model}@{last_version_id + 1}` }, "tunedModelDisplayName": "A String", # Optional. The display name of the TunedModel. The name can be up to 128 characters long and can consist of any UTF-8 characters. "tuningDataStats": { # The tuning data statistic values for TuningJob. # Output only. The tuning data statistics associated with this TuningJob. diff --git a/docs/dyn/aiplatform_v1beta1.publishers.models.html b/docs/dyn/aiplatform_v1beta1.publishers.models.html index 14923fee60..29f5a733f7 100644 --- a/docs/dyn/aiplatform_v1beta1.publishers.models.html +++ b/docs/dyn/aiplatform_v1beta1.publishers.models.html @@ -338,7 +338,7 @@

Method Details

"temperature": 3.14, # Optional. Controls the randomness of predictions. "thinkingConfig": { # Config for thinking features. # Optional. Config for thinking features. An error will be returned if this field is set for models that don't support thinking. "includeThoughts": True or False, # Optional. Indicates whether to include thoughts in the response. If true, thoughts are returned only when available. - "thinkingBudget": 42, # Optional. Indicates the thinking budget in tokens. This is only applied when enable_thinking is true. + "thinkingBudget": 42, # Optional. Indicates the thinking budget in tokens. }, "topK": 3.14, # Optional. If specified, top-k sampling will be used. "topP": 3.14, # Optional. If specified, nucleus sampling will be used. @@ -793,7 +793,7 @@

Method Details

"temperature": 3.14, # Optional. Controls the randomness of predictions. "thinkingConfig": { # Config for thinking features. # Optional. Config for thinking features. An error will be returned if this field is set for models that don't support thinking. "includeThoughts": True or False, # Optional. Indicates whether to include thoughts in the response. If true, thoughts are returned only when available. - "thinkingBudget": 42, # Optional. Indicates the thinking budget in tokens. This is only applied when enable_thinking is true. + "thinkingBudget": 42, # Optional. Indicates the thinking budget in tokens. }, "topK": 3.14, # Optional. If specified, top-k sampling will be used. "topP": 3.14, # Optional. If specified, nucleus sampling will be used. @@ -1413,7 +1413,7 @@

Method Details

}, "healthRoute": "A String", # Immutable. HTTP path on the container to send health checks to. Vertex AI intermittently sends GET requests to this path on the container's IP address and port to check that the container is healthy. Read more about [health checks](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#health). For example, if you set this field to `/bar`, then Vertex AI intermittently sends a GET request to the `/bar` path on the port of your container specified by the first value of this `ModelContainerSpec`'s ports field. If you don't specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/ DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) "imageUri": "A String", # Required. Immutable. URI of the Docker image to be used as the custom container for serving predictions. This URI must identify an image in Artifact Registry or Container Registry. Learn more about the [container publishing requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#publishing), including permissions requirements for the Vertex AI Service Agent. The container image is ingested upon ModelService.UploadModel, stored internally, and this original path is afterwards not used. To learn about the requirements for the Docker image itself, see [Custom container requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#). You can use the URI to one of Vertex AI's [pre-built container images for prediction](https://cloud.google.com/vertex-ai/docs/predictions/pre-built-containers) in this field. - "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with [PredictionService.Invoke] eg: "/invoke/foo/bar". Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. + "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with invoke http call eg: "/invoke/foo/bar", however the [PredictionService.Invoke] RPC is not supported yet. Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. "livenessProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes liveness probe. "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. @@ -1493,6 +1493,9 @@

Method Details

"target": 42, # The target resource utilization in percentage (1% - 100%) for the given metric; once the real usage deviates from the target by a certain percentage, the machine replicas change. The default value is 60 (representing 60%) if not provided. }, ], + "flexStart": { # FlexStart is used to schedule the deployment workload on DWS resource. It contains the max duration of the deployment. # Optional. Immutable. If set, use DWS resource to schedule the deployment workload. reference: (https://cloud.google.com/blog/products/compute/introducing-dynamic-workload-scheduler) + "maxRuntimeDuration": "A String", # The max duration of the deployment is max_runtime_duration. The deployment will be terminated after the duration. The max_runtime_duration can be set up to 7 days. + }, "machineSpec": { # Specification of a single machine. # Required. Immutable. The specification of a single machine being used. "acceleratorCount": 42, # The number of accelerators to attach to the machine. "acceleratorType": "A String", # Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count. @@ -1593,7 +1596,7 @@

Method Details

}, "healthRoute": "A String", # Immutable. HTTP path on the container to send health checks to. Vertex AI intermittently sends GET requests to this path on the container's IP address and port to check that the container is healthy. Read more about [health checks](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#health). For example, if you set this field to `/bar`, then Vertex AI intermittently sends a GET request to the `/bar` path on the port of your container specified by the first value of this `ModelContainerSpec`'s ports field. If you don't specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/ DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) "imageUri": "A String", # Required. Immutable. URI of the Docker image to be used as the custom container for serving predictions. This URI must identify an image in Artifact Registry or Container Registry. Learn more about the [container publishing requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#publishing), including permissions requirements for the Vertex AI Service Agent. The container image is ingested upon ModelService.UploadModel, stored internally, and this original path is afterwards not used. To learn about the requirements for the Docker image itself, see [Custom container requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#). You can use the URI to one of Vertex AI's [pre-built container images for prediction](https://cloud.google.com/vertex-ai/docs/predictions/pre-built-containers) in this field. - "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with [PredictionService.Invoke] eg: "/invoke/foo/bar". Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. + "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with invoke http call eg: "/invoke/foo/bar", however the [PredictionService.Invoke] RPC is not supported yet. Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. "livenessProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes liveness probe. "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. @@ -1673,6 +1676,9 @@

Method Details

"target": 42, # The target resource utilization in percentage (1% - 100%) for the given metric; once the real usage deviates from the target by a certain percentage, the machine replicas change. The default value is 60 (representing 60%) if not provided. }, ], + "flexStart": { # FlexStart is used to schedule the deployment workload on DWS resource. It contains the max duration of the deployment. # Optional. Immutable. If set, use DWS resource to schedule the deployment workload. reference: (https://cloud.google.com/blog/products/compute/introducing-dynamic-workload-scheduler) + "maxRuntimeDuration": "A String", # The max duration of the deployment is max_runtime_duration. The deployment will be terminated after the duration. The max_runtime_duration can be set up to 7 days. + }, "machineSpec": { # Specification of a single machine. # Required. Immutable. The specification of a single machine being used. "acceleratorCount": 42, # The number of accelerators to attach to the machine. "acceleratorType": "A String", # Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count. @@ -1993,7 +1999,7 @@

Method Details

}, "healthRoute": "A String", # Immutable. HTTP path on the container to send health checks to. Vertex AI intermittently sends GET requests to this path on the container's IP address and port to check that the container is healthy. Read more about [health checks](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#health). For example, if you set this field to `/bar`, then Vertex AI intermittently sends a GET request to the `/bar` path on the port of your container specified by the first value of this `ModelContainerSpec`'s ports field. If you don't specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/ DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) "imageUri": "A String", # Required. Immutable. URI of the Docker image to be used as the custom container for serving predictions. This URI must identify an image in Artifact Registry or Container Registry. Learn more about the [container publishing requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#publishing), including permissions requirements for the Vertex AI Service Agent. The container image is ingested upon ModelService.UploadModel, stored internally, and this original path is afterwards not used. To learn about the requirements for the Docker image itself, see [Custom container requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#). You can use the URI to one of Vertex AI's [pre-built container images for prediction](https://cloud.google.com/vertex-ai/docs/predictions/pre-built-containers) in this field. - "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with [PredictionService.Invoke] eg: "/invoke/foo/bar". Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. + "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with invoke http call eg: "/invoke/foo/bar", however the [PredictionService.Invoke] RPC is not supported yet. Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. "livenessProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes liveness probe. "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. @@ -2073,6 +2079,9 @@

Method Details

"target": 42, # The target resource utilization in percentage (1% - 100%) for the given metric; once the real usage deviates from the target by a certain percentage, the machine replicas change. The default value is 60 (representing 60%) if not provided. }, ], + "flexStart": { # FlexStart is used to schedule the deployment workload on DWS resource. It contains the max duration of the deployment. # Optional. Immutable. If set, use DWS resource to schedule the deployment workload. reference: (https://cloud.google.com/blog/products/compute/introducing-dynamic-workload-scheduler) + "maxRuntimeDuration": "A String", # The max duration of the deployment is max_runtime_duration. The deployment will be terminated after the duration. The max_runtime_duration can be set up to 7 days. + }, "machineSpec": { # Specification of a single machine. # Required. Immutable. The specification of a single machine being used. "acceleratorCount": 42, # The number of accelerators to attach to the machine. "acceleratorType": "A String", # Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count. @@ -2173,7 +2182,7 @@

Method Details

}, "healthRoute": "A String", # Immutable. HTTP path on the container to send health checks to. Vertex AI intermittently sends GET requests to this path on the container's IP address and port to check that the container is healthy. Read more about [health checks](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#health). For example, if you set this field to `/bar`, then Vertex AI intermittently sends a GET request to the `/bar` path on the port of your container specified by the first value of this `ModelContainerSpec`'s ports field. If you don't specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/ DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) "imageUri": "A String", # Required. Immutable. URI of the Docker image to be used as the custom container for serving predictions. This URI must identify an image in Artifact Registry or Container Registry. Learn more about the [container publishing requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#publishing), including permissions requirements for the Vertex AI Service Agent. The container image is ingested upon ModelService.UploadModel, stored internally, and this original path is afterwards not used. To learn about the requirements for the Docker image itself, see [Custom container requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#). You can use the URI to one of Vertex AI's [pre-built container images for prediction](https://cloud.google.com/vertex-ai/docs/predictions/pre-built-containers) in this field. - "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with [PredictionService.Invoke] eg: "/invoke/foo/bar". Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. + "invokeRoutePrefix": "A String", # Immutable. Invoke route prefix for the custom container. "/*" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with invoke http call eg: "/invoke/foo/bar", however the [PredictionService.Invoke] RPC is not supported yet. Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint. "livenessProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes liveness probe. "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. @@ -2253,6 +2262,9 @@

Method Details

"target": 42, # The target resource utilization in percentage (1% - 100%) for the given metric; once the real usage deviates from the target by a certain percentage, the machine replicas change. The default value is 60 (representing 60%) if not provided. }, ], + "flexStart": { # FlexStart is used to schedule the deployment workload on DWS resource. It contains the max duration of the deployment. # Optional. Immutable. If set, use DWS resource to schedule the deployment workload. reference: (https://cloud.google.com/blog/products/compute/introducing-dynamic-workload-scheduler) + "maxRuntimeDuration": "A String", # The max duration of the deployment is max_runtime_duration. The deployment will be terminated after the duration. The max_runtime_duration can be set up to 7 days. + }, "machineSpec": { # Specification of a single machine. # Required. Immutable. The specification of a single machine being used. "acceleratorCount": 42, # The number of accelerators to attach to the machine. "acceleratorType": "A String", # Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count. @@ -2683,7 +2695,7 @@

Method Details

"temperature": 3.14, # Optional. Controls the randomness of predictions. "thinkingConfig": { # Config for thinking features. # Optional. Config for thinking features. An error will be returned if this field is set for models that don't support thinking. "includeThoughts": True or False, # Optional. Indicates whether to include thoughts in the response. If true, thoughts are returned only when available. - "thinkingBudget": 42, # Optional. Indicates the thinking budget in tokens. This is only applied when enable_thinking is true. + "thinkingBudget": 42, # Optional. Indicates the thinking budget in tokens. }, "topK": 3.14, # Optional. If specified, top-k sampling will be used. "topP": 3.14, # Optional. If specified, nucleus sampling will be used. diff --git a/docs/dyn/aiplatform_v1beta1.reasoningEngines.html b/docs/dyn/aiplatform_v1beta1.reasoningEngines.html index 7df9230853..26de1b616e 100644 --- a/docs/dyn/aiplatform_v1beta1.reasoningEngines.html +++ b/docs/dyn/aiplatform_v1beta1.reasoningEngines.html @@ -129,10 +129,10 @@

Method Details

"contextSpec": { # Configuration for how Agent Engine sub-resources should manage context. # Optional. Configuration for how Agent Engine sub-resources should manage context. "memoryBankConfig": { # Specification for a Memory Bank. # Optional. Specification for a Memory Bank, which manages memories for the Agent Engine. "generationConfig": { # Configuration for how to generate memories. # Optional. Configuration for how to generate memories for the Memory Bank. - "model": "A String", # Required. The model used to generate memories. Format: `projects/{project}/locations/{location}/publishers/google/models/{model}` or `projects/{project}/locations/{location}/endpoints/{endpoint}`. + "model": "A String", # Required. The model used to generate memories. Format: `projects/{project}/locations/{location}/publishers/google/models/{model}`. }, "similaritySearchConfig": { # Configuration for how to perform similarity search on memories. # Optional. Configuration for how to perform similarity search on memories. If not set, the Memory Bank will use the default embedding model `text-embedding-005`. - "embeddingModel": "A String", # Required. The model used to generate embeddings to lookup similar memories. Format: `projects/{project}/locations/{location}/publishers/google/models/{model}` or `projects/{project}/locations/{location}/endpoints/{endpoint}`. + "embeddingModel": "A String", # Required. The model used to generate embeddings to lookup similar memories. Format: `projects/{project}/locations/{location}/publishers/google/models/{model}`. }, }, }, @@ -259,10 +259,10 @@

Method Details

"contextSpec": { # Configuration for how Agent Engine sub-resources should manage context. # Optional. Configuration for how Agent Engine sub-resources should manage context. "memoryBankConfig": { # Specification for a Memory Bank. # Optional. Specification for a Memory Bank, which manages memories for the Agent Engine. "generationConfig": { # Configuration for how to generate memories. # Optional. Configuration for how to generate memories for the Memory Bank. - "model": "A String", # Required. The model used to generate memories. Format: `projects/{project}/locations/{location}/publishers/google/models/{model}` or `projects/{project}/locations/{location}/endpoints/{endpoint}`. + "model": "A String", # Required. The model used to generate memories. Format: `projects/{project}/locations/{location}/publishers/google/models/{model}`. }, "similaritySearchConfig": { # Configuration for how to perform similarity search on memories. # Optional. Configuration for how to perform similarity search on memories. If not set, the Memory Bank will use the default embedding model `text-embedding-005`. - "embeddingModel": "A String", # Required. The model used to generate embeddings to lookup similar memories. Format: `projects/{project}/locations/{location}/publishers/google/models/{model}` or `projects/{project}/locations/{location}/endpoints/{endpoint}`. + "embeddingModel": "A String", # Required. The model used to generate embeddings to lookup similar memories. Format: `projects/{project}/locations/{location}/publishers/google/models/{model}`. }, }, }, @@ -330,10 +330,10 @@

Method Details

"contextSpec": { # Configuration for how Agent Engine sub-resources should manage context. # Optional. Configuration for how Agent Engine sub-resources should manage context. "memoryBankConfig": { # Specification for a Memory Bank. # Optional. Specification for a Memory Bank, which manages memories for the Agent Engine. "generationConfig": { # Configuration for how to generate memories. # Optional. Configuration for how to generate memories for the Memory Bank. - "model": "A String", # Required. The model used to generate memories. Format: `projects/{project}/locations/{location}/publishers/google/models/{model}` or `projects/{project}/locations/{location}/endpoints/{endpoint}`. + "model": "A String", # Required. The model used to generate memories. Format: `projects/{project}/locations/{location}/publishers/google/models/{model}`. }, "similaritySearchConfig": { # Configuration for how to perform similarity search on memories. # Optional. Configuration for how to perform similarity search on memories. If not set, the Memory Bank will use the default embedding model `text-embedding-005`. - "embeddingModel": "A String", # Required. The model used to generate embeddings to lookup similar memories. Format: `projects/{project}/locations/{location}/publishers/google/models/{model}` or `projects/{project}/locations/{location}/endpoints/{endpoint}`. + "embeddingModel": "A String", # Required. The model used to generate embeddings to lookup similar memories. Format: `projects/{project}/locations/{location}/publishers/google/models/{model}`. }, }, }, @@ -406,10 +406,10 @@

Method Details

"contextSpec": { # Configuration for how Agent Engine sub-resources should manage context. # Optional. Configuration for how Agent Engine sub-resources should manage context. "memoryBankConfig": { # Specification for a Memory Bank. # Optional. Specification for a Memory Bank, which manages memories for the Agent Engine. "generationConfig": { # Configuration for how to generate memories. # Optional. Configuration for how to generate memories for the Memory Bank. - "model": "A String", # Required. The model used to generate memories. Format: `projects/{project}/locations/{location}/publishers/google/models/{model}` or `projects/{project}/locations/{location}/endpoints/{endpoint}`. + "model": "A String", # Required. The model used to generate memories. Format: `projects/{project}/locations/{location}/publishers/google/models/{model}`. }, "similaritySearchConfig": { # Configuration for how to perform similarity search on memories. # Optional. Configuration for how to perform similarity search on memories. If not set, the Memory Bank will use the default embedding model `text-embedding-005`. - "embeddingModel": "A String", # Required. The model used to generate embeddings to lookup similar memories. Format: `projects/{project}/locations/{location}/publishers/google/models/{model}` or `projects/{project}/locations/{location}/endpoints/{endpoint}`. + "embeddingModel": "A String", # Required. The model used to generate embeddings to lookup similar memories. Format: `projects/{project}/locations/{location}/publishers/google/models/{model}`. }, }, }, diff --git a/docs/dyn/aiplatform_v1beta1.reasoningEngines.sessions.events.html b/docs/dyn/aiplatform_v1beta1.reasoningEngines.sessions.events.html index 2d3ad003ef..0e2745495b 100644 --- a/docs/dyn/aiplatform_v1beta1.reasoningEngines.sessions.events.html +++ b/docs/dyn/aiplatform_v1beta1.reasoningEngines.sessions.events.html @@ -78,7 +78,7 @@

Instance Methods

close()

Close httplib2 connections.

- list(parent, pageSize=None, pageToken=None, x__xgafv=None)

+ list(parent, filter=None, pageSize=None, pageToken=None, x__xgafv=None)

Lists Events in a given session.

list_next()

@@ -90,11 +90,12 @@

Method Details

- list(parent, pageSize=None, pageToken=None, x__xgafv=None) + list(parent, filter=None, pageSize=None, pageToken=None, x__xgafv=None)
Lists Events in a given session.
 
 Args:
   parent: string, Required. The resource name of the session to list events from. Format: `projects/{project}/locations/{location}/reasoningEngines/{reasoning_engine}/sessions/{session}` (required)
+  filter: string, Optional. The standard list filter. Supported fields: * `timestamp` range (i.e. `timestamp>="2025-01-31T11:30:00-04:00"` where the timestamp is in RFC 3339 format) More detail in [AIP-160](https://google.aip.dev/160).
   pageSize: integer, Optional. The maximum number of events to return. The service may return fewer than this value. If unspecified, at most 100 events will be returned. These events are ordered by timestamp in ascending order.
   pageToken: string, Optional. The next_page_token value returned from a previous list SessionService.ListEvents call.
   x__xgafv: string, V1 error format.
@@ -175,6 +176,9 @@ 

Method Details

"errorMessage": "A String", # Optional. Error message if the response is an error. "eventMetadata": { # Metadata relating to a LLM response event. # Optional. Metadata relating to this event. "branch": "A String", # Optional. The branch of the event. The format is like agent_1.agent_2.agent_3, where agent_1 is the parent of agent_2, and agent_2 is the parent of agent_3. Branch is used when multiple child agents shouldn't see their siblings' conversation history. + "customMetadata": { # The custom metadata of the LlmResponse. + "a_key": "", # Properties of the object. + }, "groundingMetadata": { # Metadata returned to client when grounding is enabled. # Optional. Metadata returned to client when grounding is enabled. "groundingChunks": [ # List of supporting references retrieved from specified grounding source. { # Grounding chunk. diff --git a/docs/dyn/aiplatform_v1beta1.reasoningEngines.sessions.html b/docs/dyn/aiplatform_v1beta1.reasoningEngines.sessions.html index 3d41739fb0..3aacaafb4b 100644 --- a/docs/dyn/aiplatform_v1beta1.reasoningEngines.sessions.html +++ b/docs/dyn/aiplatform_v1beta1.reasoningEngines.sessions.html @@ -180,6 +180,9 @@

Method Details

"errorMessage": "A String", # Optional. Error message if the response is an error. "eventMetadata": { # Metadata relating to a LLM response event. # Optional. Metadata relating to this event. "branch": "A String", # Optional. The branch of the event. The format is like agent_1.agent_2.agent_3, where agent_1 is the parent of agent_2, and agent_2 is the parent of agent_3. Branch is used when multiple child agents shouldn't see their siblings' conversation history. + "customMetadata": { # The custom metadata of the LlmResponse. + "a_key": "", # Properties of the object. + }, "groundingMetadata": { # Metadata returned to client when grounding is enabled. # Optional. Metadata returned to client when grounding is enabled. "groundingChunks": [ # List of supporting references retrieved from specified grounding source. { # Grounding chunk. diff --git a/docs/dyn/androidmanagement_v1.enterprises.policies.html b/docs/dyn/androidmanagement_v1.enterprises.policies.html index 724e9218e6..9b2f026fbf 100644 --- a/docs/dyn/androidmanagement_v1.enterprises.policies.html +++ b/docs/dyn/androidmanagement_v1.enterprises.policies.html @@ -177,7 +177,7 @@

Method Details

], "disabled": True or False, # Whether the app is disabled. When disabled, the app data is still preserved. "extensionConfig": { # Configuration to enable an app as an extension app, with the capability of interacting with Android Device Policy offline. For Android versions 11 and above, extension apps are exempt from battery restrictions so will not be placed into the restricted App Standby Bucket (https://developer.android.com/topic/performance/appstandby#restricted-bucket). Extensions apps are also protected against users clearing their data or force-closing the application, although admins can continue to use the clear app data command on extension apps if needed for Android 11 and above. # Configuration to enable this app as an extension app, with the capability of interacting with Android Device Policy offline.This field can be set for at most one app.The signing key certificate fingerprint of the app on the device must match one of the entries in signingKeyFingerprintsSha256 or the signing key certificate fingerprints obtained from Play Store for the app to be able to communicate with Android Device Policy. If the app is not on Play Store and signingKeyFingerprintsSha256 is not set, a NonComplianceDetail with INVALID_VALUE is reported. - "notificationReceiver": "A String", # Fully qualified class name of the receiver service class for Android Device Policy to notify the extension app of any local command status updates. + "notificationReceiver": "A String", # Fully qualified class name of the receiver service class for Android Device Policy to notify the extension app of any local command status updates. The service must be exported in the extension app's AndroidManifest.xml and extend NotificationReceiverService (https://developers.google.com/android/management/reference/amapi/com/google/android/managementapi/notification/NotificationReceiverService) (see Integrate with the AMAPI SDK (https://developers.google.com/android/management/sdk-integration) guide for more details). "signingKeyFingerprintsSha256": [ # Hex-encoded SHA-256 hashes of the signing key certificates of the extension app. Only hexadecimal string representations of 64 characters are valid.The signing key certificate fingerprints are always obtained from the Play Store and this field is used to provide additional signing key certificate fingerprints. However, if the application is not available on the Play Store, this field needs to be set. A NonComplianceDetail with INVALID_VALUE is reported if this field is not set when the application is not available on the Play Store.The signing key certificate fingerprint of the extension app on the device must match one of the signing key certificate fingerprints obtained from the Play Store or the ones provided in this field for the app to be able to communicate with Android Device Policy.In production use cases, it is recommended to leave this empty. "A String", ], @@ -672,7 +672,7 @@

Method Details

], "disabled": True or False, # Whether the app is disabled. When disabled, the app data is still preserved. "extensionConfig": { # Configuration to enable an app as an extension app, with the capability of interacting with Android Device Policy offline. For Android versions 11 and above, extension apps are exempt from battery restrictions so will not be placed into the restricted App Standby Bucket (https://developer.android.com/topic/performance/appstandby#restricted-bucket). Extensions apps are also protected against users clearing their data or force-closing the application, although admins can continue to use the clear app data command on extension apps if needed for Android 11 and above. # Configuration to enable this app as an extension app, with the capability of interacting with Android Device Policy offline.This field can be set for at most one app.The signing key certificate fingerprint of the app on the device must match one of the entries in signingKeyFingerprintsSha256 or the signing key certificate fingerprints obtained from Play Store for the app to be able to communicate with Android Device Policy. If the app is not on Play Store and signingKeyFingerprintsSha256 is not set, a NonComplianceDetail with INVALID_VALUE is reported. - "notificationReceiver": "A String", # Fully qualified class name of the receiver service class for Android Device Policy to notify the extension app of any local command status updates. + "notificationReceiver": "A String", # Fully qualified class name of the receiver service class for Android Device Policy to notify the extension app of any local command status updates. The service must be exported in the extension app's AndroidManifest.xml and extend NotificationReceiverService (https://developers.google.com/android/management/reference/amapi/com/google/android/managementapi/notification/NotificationReceiverService) (see Integrate with the AMAPI SDK (https://developers.google.com/android/management/sdk-integration) guide for more details). "signingKeyFingerprintsSha256": [ # Hex-encoded SHA-256 hashes of the signing key certificates of the extension app. Only hexadecimal string representations of 64 characters are valid.The signing key certificate fingerprints are always obtained from the Play Store and this field is used to provide additional signing key certificate fingerprints. However, if the application is not available on the Play Store, this field needs to be set. A NonComplianceDetail with INVALID_VALUE is reported if this field is not set when the application is not available on the Play Store.The signing key certificate fingerprint of the extension app on the device must match one of the signing key certificate fingerprints obtained from the Play Store or the ones provided in this field for the app to be able to communicate with Android Device Policy.In production use cases, it is recommended to leave this empty. "A String", ], @@ -1149,7 +1149,7 @@

Method Details

], "disabled": True or False, # Whether the app is disabled. When disabled, the app data is still preserved. "extensionConfig": { # Configuration to enable an app as an extension app, with the capability of interacting with Android Device Policy offline. For Android versions 11 and above, extension apps are exempt from battery restrictions so will not be placed into the restricted App Standby Bucket (https://developer.android.com/topic/performance/appstandby#restricted-bucket). Extensions apps are also protected against users clearing their data or force-closing the application, although admins can continue to use the clear app data command on extension apps if needed for Android 11 and above. # Configuration to enable this app as an extension app, with the capability of interacting with Android Device Policy offline.This field can be set for at most one app.The signing key certificate fingerprint of the app on the device must match one of the entries in signingKeyFingerprintsSha256 or the signing key certificate fingerprints obtained from Play Store for the app to be able to communicate with Android Device Policy. If the app is not on Play Store and signingKeyFingerprintsSha256 is not set, a NonComplianceDetail with INVALID_VALUE is reported. - "notificationReceiver": "A String", # Fully qualified class name of the receiver service class for Android Device Policy to notify the extension app of any local command status updates. + "notificationReceiver": "A String", # Fully qualified class name of the receiver service class for Android Device Policy to notify the extension app of any local command status updates. The service must be exported in the extension app's AndroidManifest.xml and extend NotificationReceiverService (https://developers.google.com/android/management/reference/amapi/com/google/android/managementapi/notification/NotificationReceiverService) (see Integrate with the AMAPI SDK (https://developers.google.com/android/management/sdk-integration) guide for more details). "signingKeyFingerprintsSha256": [ # Hex-encoded SHA-256 hashes of the signing key certificates of the extension app. Only hexadecimal string representations of 64 characters are valid.The signing key certificate fingerprints are always obtained from the Play Store and this field is used to provide additional signing key certificate fingerprints. However, if the application is not available on the Play Store, this field needs to be set. A NonComplianceDetail with INVALID_VALUE is reported if this field is not set when the application is not available on the Play Store.The signing key certificate fingerprint of the extension app on the device must match one of the signing key certificate fingerprints obtained from the Play Store or the ones provided in this field for the app to be able to communicate with Android Device Policy.In production use cases, it is recommended to leave this empty. "A String", ], @@ -1240,7 +1240,7 @@

Method Details

], "disabled": True or False, # Whether the app is disabled. When disabled, the app data is still preserved. "extensionConfig": { # Configuration to enable an app as an extension app, with the capability of interacting with Android Device Policy offline. For Android versions 11 and above, extension apps are exempt from battery restrictions so will not be placed into the restricted App Standby Bucket (https://developer.android.com/topic/performance/appstandby#restricted-bucket). Extensions apps are also protected against users clearing their data or force-closing the application, although admins can continue to use the clear app data command on extension apps if needed for Android 11 and above. # Configuration to enable this app as an extension app, with the capability of interacting with Android Device Policy offline.This field can be set for at most one app.The signing key certificate fingerprint of the app on the device must match one of the entries in signingKeyFingerprintsSha256 or the signing key certificate fingerprints obtained from Play Store for the app to be able to communicate with Android Device Policy. If the app is not on Play Store and signingKeyFingerprintsSha256 is not set, a NonComplianceDetail with INVALID_VALUE is reported. - "notificationReceiver": "A String", # Fully qualified class name of the receiver service class for Android Device Policy to notify the extension app of any local command status updates. + "notificationReceiver": "A String", # Fully qualified class name of the receiver service class for Android Device Policy to notify the extension app of any local command status updates. The service must be exported in the extension app's AndroidManifest.xml and extend NotificationReceiverService (https://developers.google.com/android/management/reference/amapi/com/google/android/managementapi/notification/NotificationReceiverService) (see Integrate with the AMAPI SDK (https://developers.google.com/android/management/sdk-integration) guide for more details). "signingKeyFingerprintsSha256": [ # Hex-encoded SHA-256 hashes of the signing key certificates of the extension app. Only hexadecimal string representations of 64 characters are valid.The signing key certificate fingerprints are always obtained from the Play Store and this field is used to provide additional signing key certificate fingerprints. However, if the application is not available on the Play Store, this field needs to be set. A NonComplianceDetail with INVALID_VALUE is reported if this field is not set when the application is not available on the Play Store.The signing key certificate fingerprint of the extension app on the device must match one of the signing key certificate fingerprints obtained from the Play Store or the ones provided in this field for the app to be able to communicate with Android Device Policy.In production use cases, it is recommended to leave this empty. "A String", ], @@ -1726,7 +1726,7 @@

Method Details

], "disabled": True or False, # Whether the app is disabled. When disabled, the app data is still preserved. "extensionConfig": { # Configuration to enable an app as an extension app, with the capability of interacting with Android Device Policy offline. For Android versions 11 and above, extension apps are exempt from battery restrictions so will not be placed into the restricted App Standby Bucket (https://developer.android.com/topic/performance/appstandby#restricted-bucket). Extensions apps are also protected against users clearing their data or force-closing the application, although admins can continue to use the clear app data command on extension apps if needed for Android 11 and above. # Configuration to enable this app as an extension app, with the capability of interacting with Android Device Policy offline.This field can be set for at most one app.The signing key certificate fingerprint of the app on the device must match one of the entries in signingKeyFingerprintsSha256 or the signing key certificate fingerprints obtained from Play Store for the app to be able to communicate with Android Device Policy. If the app is not on Play Store and signingKeyFingerprintsSha256 is not set, a NonComplianceDetail with INVALID_VALUE is reported. - "notificationReceiver": "A String", # Fully qualified class name of the receiver service class for Android Device Policy to notify the extension app of any local command status updates. + "notificationReceiver": "A String", # Fully qualified class name of the receiver service class for Android Device Policy to notify the extension app of any local command status updates. The service must be exported in the extension app's AndroidManifest.xml and extend NotificationReceiverService (https://developers.google.com/android/management/reference/amapi/com/google/android/managementapi/notification/NotificationReceiverService) (see Integrate with the AMAPI SDK (https://developers.google.com/android/management/sdk-integration) guide for more details). "signingKeyFingerprintsSha256": [ # Hex-encoded SHA-256 hashes of the signing key certificates of the extension app. Only hexadecimal string representations of 64 characters are valid.The signing key certificate fingerprints are always obtained from the Play Store and this field is used to provide additional signing key certificate fingerprints. However, if the application is not available on the Play Store, this field needs to be set. A NonComplianceDetail with INVALID_VALUE is reported if this field is not set when the application is not available on the Play Store.The signing key certificate fingerprint of the extension app on the device must match one of the signing key certificate fingerprints obtained from the Play Store or the ones provided in this field for the app to be able to communicate with Android Device Policy.In production use cases, it is recommended to leave this empty. "A String", ], @@ -2210,7 +2210,7 @@

Method Details

], "disabled": True or False, # Whether the app is disabled. When disabled, the app data is still preserved. "extensionConfig": { # Configuration to enable an app as an extension app, with the capability of interacting with Android Device Policy offline. For Android versions 11 and above, extension apps are exempt from battery restrictions so will not be placed into the restricted App Standby Bucket (https://developer.android.com/topic/performance/appstandby#restricted-bucket). Extensions apps are also protected against users clearing their data or force-closing the application, although admins can continue to use the clear app data command on extension apps if needed for Android 11 and above. # Configuration to enable this app as an extension app, with the capability of interacting with Android Device Policy offline.This field can be set for at most one app.The signing key certificate fingerprint of the app on the device must match one of the entries in signingKeyFingerprintsSha256 or the signing key certificate fingerprints obtained from Play Store for the app to be able to communicate with Android Device Policy. If the app is not on Play Store and signingKeyFingerprintsSha256 is not set, a NonComplianceDetail with INVALID_VALUE is reported. - "notificationReceiver": "A String", # Fully qualified class name of the receiver service class for Android Device Policy to notify the extension app of any local command status updates. + "notificationReceiver": "A String", # Fully qualified class name of the receiver service class for Android Device Policy to notify the extension app of any local command status updates. The service must be exported in the extension app's AndroidManifest.xml and extend NotificationReceiverService (https://developers.google.com/android/management/reference/amapi/com/google/android/managementapi/notification/NotificationReceiverService) (see Integrate with the AMAPI SDK (https://developers.google.com/android/management/sdk-integration) guide for more details). "signingKeyFingerprintsSha256": [ # Hex-encoded SHA-256 hashes of the signing key certificates of the extension app. Only hexadecimal string representations of 64 characters are valid.The signing key certificate fingerprints are always obtained from the Play Store and this field is used to provide additional signing key certificate fingerprints. However, if the application is not available on the Play Store, this field needs to be set. A NonComplianceDetail with INVALID_VALUE is reported if this field is not set when the application is not available on the Play Store.The signing key certificate fingerprint of the extension app on the device must match one of the signing key certificate fingerprints obtained from the Play Store or the ones provided in this field for the app to be able to communicate with Android Device Policy.In production use cases, it is recommended to leave this empty. "A String", ], @@ -2710,7 +2710,7 @@

Method Details

], "disabled": True or False, # Whether the app is disabled. When disabled, the app data is still preserved. "extensionConfig": { # Configuration to enable an app as an extension app, with the capability of interacting with Android Device Policy offline. For Android versions 11 and above, extension apps are exempt from battery restrictions so will not be placed into the restricted App Standby Bucket (https://developer.android.com/topic/performance/appstandby#restricted-bucket). Extensions apps are also protected against users clearing their data or force-closing the application, although admins can continue to use the clear app data command on extension apps if needed for Android 11 and above. # Configuration to enable this app as an extension app, with the capability of interacting with Android Device Policy offline.This field can be set for at most one app.The signing key certificate fingerprint of the app on the device must match one of the entries in signingKeyFingerprintsSha256 or the signing key certificate fingerprints obtained from Play Store for the app to be able to communicate with Android Device Policy. If the app is not on Play Store and signingKeyFingerprintsSha256 is not set, a NonComplianceDetail with INVALID_VALUE is reported. - "notificationReceiver": "A String", # Fully qualified class name of the receiver service class for Android Device Policy to notify the extension app of any local command status updates. + "notificationReceiver": "A String", # Fully qualified class name of the receiver service class for Android Device Policy to notify the extension app of any local command status updates. The service must be exported in the extension app's AndroidManifest.xml and extend NotificationReceiverService (https://developers.google.com/android/management/reference/amapi/com/google/android/managementapi/notification/NotificationReceiverService) (see Integrate with the AMAPI SDK (https://developers.google.com/android/management/sdk-integration) guide for more details). "signingKeyFingerprintsSha256": [ # Hex-encoded SHA-256 hashes of the signing key certificates of the extension app. Only hexadecimal string representations of 64 characters are valid.The signing key certificate fingerprints are always obtained from the Play Store and this field is used to provide additional signing key certificate fingerprints. However, if the application is not available on the Play Store, this field needs to be set. A NonComplianceDetail with INVALID_VALUE is reported if this field is not set when the application is not available on the Play Store.The signing key certificate fingerprint of the extension app on the device must match one of the signing key certificate fingerprints obtained from the Play Store or the ones provided in this field for the app to be able to communicate with Android Device Policy.In production use cases, it is recommended to leave this empty. "A String", ], diff --git a/docs/dyn/androidpublisher_v3.purchases.html b/docs/dyn/androidpublisher_v3.purchases.html index 2cef790fe7..e29f5ec38c 100644 --- a/docs/dyn/androidpublisher_v3.purchases.html +++ b/docs/dyn/androidpublisher_v3.purchases.html @@ -79,6 +79,11 @@

Instance Methods

Returns the products Resource.

+

+ productsv2() +

+

Returns the productsv2 Resource.

+

subscriptions()

diff --git a/docs/dyn/androidpublisher_v3.purchases.productsv2.html b/docs/dyn/androidpublisher_v3.purchases.productsv2.html new file mode 100644 index 0000000000..78b8b90b07 --- /dev/null +++ b/docs/dyn/androidpublisher_v3.purchases.productsv2.html @@ -0,0 +1,139 @@ + + + +

Google Play Android Developer API . purchases . productsv2

+

Instance Methods

+

+ close()

+

Close httplib2 connections.

+

+ getproductpurchasev2(packageName, token, x__xgafv=None)

+

Checks the purchase and consumption status of an inapp item.

+

Method Details

+
+ close() +
Close httplib2 connections.
+
+ +
+ getproductpurchasev2(packageName, token, x__xgafv=None) +
Checks the purchase and consumption status of an inapp item.
+
+Args:
+  packageName: string, The package name of the application the inapp product was sold in (for example, 'com.some.thing'). (required)
+  token: string, The token provided to the user's device when the inapp product was purchased. (required)
+  x__xgafv: string, V1 error format.
+    Allowed values
+      1 - v1 error format
+      2 - v2 error format
+
+Returns:
+  An object of the form:
+
+    { # A ProductPurchaseV2 resource indicates the status of a user's inapp product purchase.
+  "acknowledgementState": "A String", # Output only. The acknowledgement state of the purchase.
+  "kind": "A String", # This kind represents a ProductPurchaseV2 object in the androidpublisher service.
+  "obfuscatedExternalAccountId": "A String", # An obfuscated version of the id that is uniquely associated with the user's account in your app. Only present if specified using https://developer.android.com/reference/com/android/billingclient/api/BillingFlowParams.Builder#setobfuscatedaccountid when the purchase was made.
+  "obfuscatedExternalProfileId": "A String", # An obfuscated version of the id that is uniquely associated with the user's profile in your app. Only present if specified using https://developer.android.com/reference/com/android/billingclient/api/BillingFlowParams.Builder#setobfuscatedprofileid when the purchase was made.
+  "orderId": "A String", # The order id associated with the purchase of the inapp product. May not be set if there is no order associated with the purchase.
+  "productLineItem": [ # Contains item-level info for a ProductPurchaseV2.
+    { # Contains item-level info for a ProductPurchaseV2.
+      "productId": "A String", # The purchased product ID (for example, 'monthly001').
+      "productOfferDetails": { # Offer details information related to a purchase line item. # The offer details for this item.
+        "consumptionState": "A String", # Output only. The consumption state of the purchase.
+        "offerId": "A String", # The offer ID. Only present for offers.
+        "offerTags": [ # The latest offer tags associated with the offer. It includes tags inherited from the purchase option.
+          "A String",
+        ],
+        "offerToken": "A String", # The per-transaction offer token used to make this purchase line item.
+        "purchaseOptionId": "A String", # The purchase option ID.
+        "quantity": 42, # The quantity associated with the purchase of the inapp product.
+        "refundableQuantity": 42, # The quantity eligible for refund, i.e. quantity that hasn't been refunded. The value reflects quantity-based partial refunds and full refunds.
+        "rentOfferDetails": { # Offer details information related to a rental line item. # Offer details about rent offers. This will only be set for rental line items.
+        },
+      },
+    },
+  ],
+  "purchaseCompletionTime": "A String", # The time when the purchase was successful, i.e., when the PurchaseState has changed to PURCHASED. This field will not be present until the payment is complete. For example, if the user initiated a pending transaction (https://developer.android.com/google/play/billing/integrate#pending), this field will not be populated until the user successfully completes the steps required to complete the transaction.
+  "purchaseStateContext": { # Context about the purchase state. # Information about the purchase state of the purchase.
+    "purchaseState": "A String", # Output only. The purchase state of the purchase.
+  },
+  "regionCode": "A String", # ISO 3166-1 alpha-2 billing region code of the user at the time the product was granted.
+  "testPurchaseContext": { # Context about a test purchase. # Information related to test purchases. This will only be set for test purchases.
+    "fopType": "A String", # The fop type of the test purchase.
+  },
+}
+
+ + \ No newline at end of file diff --git a/docs/dyn/apihub_v1.projects.locations.deployments.html b/docs/dyn/apihub_v1.projects.locations.deployments.html index f7f1b95dd8..8763f1797f 100644 --- a/docs/dyn/apihub_v1.projects.locations.deployments.html +++ b/docs/dyn/apihub_v1.projects.locations.deployments.html @@ -209,6 +209,34 @@

Method Details

], }, }, + "managementUrl": { # The attribute values associated with resource. # Optional. The uri where users can navigate to for the management of the deployment. This maps to the following system defined attribute: `projects/{project}/locations/{location}/attributes/system-management-url` The number of values for this attribute will be based on the cardinality of the attribute. The same can be retrieved via GetAttribute API. + "attribute": "A String", # Output only. The name of the attribute. Format: projects/{project}/locations/{location}/attributes/{attribute} + "enumValues": { # The attribute values of data type enum. # The attribute values associated with a resource in case attribute data type is enum. + "values": [ # Required. The attribute values in case attribute data type is enum. + { # The value that can be assigned to the attribute when the data type is enum. + "description": "A String", # Optional. The detailed description of the allowed value. + "displayName": "A String", # Required. The display name of the allowed value. + "id": "A String", # Required. The ID of the allowed value. * If provided, the same will be used. The service will throw an error if the specified id is already used by another allowed value in the same attribute resource. * If not provided, a system generated id derived from the display name will be used. In this case, the service will handle conflict resolution by adding a system generated suffix in case of duplicates. This value should be 4-63 characters, and valid characters are /a-z-/. + "immutable": True or False, # Optional. When set to true, the allowed value cannot be updated or deleted by the user. It can only be true for System defined attributes. + }, + ], + }, + "jsonValues": { # The attribute values of data type string or JSON. # The attribute values associated with a resource in case attribute data type is JSON. + "values": [ # Required. The attribute values in case attribute data type is string or JSON. + "A String", + ], + }, + "stringValues": { # The attribute values of data type string or JSON. # The attribute values associated with a resource in case attribute data type is string. + "values": [ # Required. The attribute values in case attribute data type is string or JSON. + "A String", + ], + }, + "uriValues": { # The attribute values of data type string or JSON. # The attribute values associated with a resource in case attribute data type is URL, URI or IP, like gs://bucket-name/object-name. + "values": [ # Required. The attribute values in case attribute data type is string or JSON. + "A String", + ], + }, + }, "name": "A String", # Identifier. The name of the deployment. Format: `projects/{project}/locations/{location}/deployments/{deployment}` "resourceUri": "A String", # Required. A uri that uniquely identfies the deployment within a particular gateway. For example, if the runtime resource is of type APIGEE_PROXY, then this field will be a combination of org, proxy name and environment. "slo": { # The attribute values associated with resource. # Optional. The SLO for this deployment. This maps to the following system defined attribute: `projects/{project}/locations/{location}/attributes/system-slo` attribute. The number of values for this attribute will be based on the cardinality of the attribute. The same can be retrieved via GetAttribute API. All values should be from the list of allowed values defined for the attribute. @@ -239,6 +267,7 @@

Method Details

], }, }, + "sourceEnvironment": "A String", # Optional. The environment at source for the deployment. For example: prod, dev, staging, etc. "sourceMetadata": [ # Output only. The list of sources and metadata from the sources of the deployment. { # SourceMetadata represents the metadata for a resource at the source. "originalResourceCreateTime": "A String", # Output only. The time at which the resource was created at the source. @@ -251,6 +280,35 @@

Method Details

"sourceType": "A String", # Output only. The type of the source. }, ], + "sourceProject": "A String", # Optional. The project to which the deployment belongs. For GCP gateways, this will refer to the project identifier. For others like Edge/OPDK, this will refer to the org identifier. + "sourceUri": { # The attribute values associated with resource. # Optional. The uri where additional source specific information for this deployment can be found. This maps to the following system defined attribute: `projects/{project}/locations/{location}/attributes/system-source-uri` The number of values for this attribute will be based on the cardinality of the attribute. The same can be retrieved via GetAttribute API. + "attribute": "A String", # Output only. The name of the attribute. Format: projects/{project}/locations/{location}/attributes/{attribute} + "enumValues": { # The attribute values of data type enum. # The attribute values associated with a resource in case attribute data type is enum. + "values": [ # Required. The attribute values in case attribute data type is enum. + { # The value that can be assigned to the attribute when the data type is enum. + "description": "A String", # Optional. The detailed description of the allowed value. + "displayName": "A String", # Required. The display name of the allowed value. + "id": "A String", # Required. The ID of the allowed value. * If provided, the same will be used. The service will throw an error if the specified id is already used by another allowed value in the same attribute resource. * If not provided, a system generated id derived from the display name will be used. In this case, the service will handle conflict resolution by adding a system generated suffix in case of duplicates. This value should be 4-63 characters, and valid characters are /a-z-/. + "immutable": True or False, # Optional. When set to true, the allowed value cannot be updated or deleted by the user. It can only be true for System defined attributes. + }, + ], + }, + "jsonValues": { # The attribute values of data type string or JSON. # The attribute values associated with a resource in case attribute data type is JSON. + "values": [ # Required. The attribute values in case attribute data type is string or JSON. + "A String", + ], + }, + "stringValues": { # The attribute values of data type string or JSON. # The attribute values associated with a resource in case attribute data type is string. + "values": [ # Required. The attribute values in case attribute data type is string or JSON. + "A String", + ], + }, + "uriValues": { # The attribute values of data type string or JSON. # The attribute values associated with a resource in case attribute data type is URL, URI or IP, like gs://bucket-name/object-name. + "values": [ # Required. The attribute values in case attribute data type is string or JSON. + "A String", + ], + }, + }, "updateTime": "A String", # Output only. The time at which the deployment was last updated. } @@ -362,6 +420,34 @@

Method Details

], }, }, + "managementUrl": { # The attribute values associated with resource. # Optional. The uri where users can navigate to for the management of the deployment. This maps to the following system defined attribute: `projects/{project}/locations/{location}/attributes/system-management-url` The number of values for this attribute will be based on the cardinality of the attribute. The same can be retrieved via GetAttribute API. + "attribute": "A String", # Output only. The name of the attribute. Format: projects/{project}/locations/{location}/attributes/{attribute} + "enumValues": { # The attribute values of data type enum. # The attribute values associated with a resource in case attribute data type is enum. + "values": [ # Required. The attribute values in case attribute data type is enum. + { # The value that can be assigned to the attribute when the data type is enum. + "description": "A String", # Optional. The detailed description of the allowed value. + "displayName": "A String", # Required. The display name of the allowed value. + "id": "A String", # Required. The ID of the allowed value. * If provided, the same will be used. The service will throw an error if the specified id is already used by another allowed value in the same attribute resource. * If not provided, a system generated id derived from the display name will be used. In this case, the service will handle conflict resolution by adding a system generated suffix in case of duplicates. This value should be 4-63 characters, and valid characters are /a-z-/. + "immutable": True or False, # Optional. When set to true, the allowed value cannot be updated or deleted by the user. It can only be true for System defined attributes. + }, + ], + }, + "jsonValues": { # The attribute values of data type string or JSON. # The attribute values associated with a resource in case attribute data type is JSON. + "values": [ # Required. The attribute values in case attribute data type is string or JSON. + "A String", + ], + }, + "stringValues": { # The attribute values of data type string or JSON. # The attribute values associated with a resource in case attribute data type is string. + "values": [ # Required. The attribute values in case attribute data type is string or JSON. + "A String", + ], + }, + "uriValues": { # The attribute values of data type string or JSON. # The attribute values associated with a resource in case attribute data type is URL, URI or IP, like gs://bucket-name/object-name. + "values": [ # Required. The attribute values in case attribute data type is string or JSON. + "A String", + ], + }, + }, "name": "A String", # Identifier. The name of the deployment. Format: `projects/{project}/locations/{location}/deployments/{deployment}` "resourceUri": "A String", # Required. A uri that uniquely identfies the deployment within a particular gateway. For example, if the runtime resource is of type APIGEE_PROXY, then this field will be a combination of org, proxy name and environment. "slo": { # The attribute values associated with resource. # Optional. The SLO for this deployment. This maps to the following system defined attribute: `projects/{project}/locations/{location}/attributes/system-slo` attribute. The number of values for this attribute will be based on the cardinality of the attribute. The same can be retrieved via GetAttribute API. All values should be from the list of allowed values defined for the attribute. @@ -392,6 +478,7 @@

Method Details

], }, }, + "sourceEnvironment": "A String", # Optional. The environment at source for the deployment. For example: prod, dev, staging, etc. "sourceMetadata": [ # Output only. The list of sources and metadata from the sources of the deployment. { # SourceMetadata represents the metadata for a resource at the source. "originalResourceCreateTime": "A String", # Output only. The time at which the resource was created at the source. @@ -404,6 +491,35 @@

Method Details

"sourceType": "A String", # Output only. The type of the source. }, ], + "sourceProject": "A String", # Optional. The project to which the deployment belongs. For GCP gateways, this will refer to the project identifier. For others like Edge/OPDK, this will refer to the org identifier. + "sourceUri": { # The attribute values associated with resource. # Optional. The uri where additional source specific information for this deployment can be found. This maps to the following system defined attribute: `projects/{project}/locations/{location}/attributes/system-source-uri` The number of values for this attribute will be based on the cardinality of the attribute. The same can be retrieved via GetAttribute API. + "attribute": "A String", # Output only. The name of the attribute. Format: projects/{project}/locations/{location}/attributes/{attribute} + "enumValues": { # The attribute values of data type enum. # The attribute values associated with a resource in case attribute data type is enum. + "values": [ # Required. The attribute values in case attribute data type is enum. + { # The value that can be assigned to the attribute when the data type is enum. + "description": "A String", # Optional. The detailed description of the allowed value. + "displayName": "A String", # Required. The display name of the allowed value. + "id": "A String", # Required. The ID of the allowed value. * If provided, the same will be used. The service will throw an error if the specified id is already used by another allowed value in the same attribute resource. * If not provided, a system generated id derived from the display name will be used. In this case, the service will handle conflict resolution by adding a system generated suffix in case of duplicates. This value should be 4-63 characters, and valid characters are /a-z-/. + "immutable": True or False, # Optional. When set to true, the allowed value cannot be updated or deleted by the user. It can only be true for System defined attributes. + }, + ], + }, + "jsonValues": { # The attribute values of data type string or JSON. # The attribute values associated with a resource in case attribute data type is JSON. + "values": [ # Required. The attribute values in case attribute data type is string or JSON. + "A String", + ], + }, + "stringValues": { # The attribute values of data type string or JSON. # The attribute values associated with a resource in case attribute data type is string. + "values": [ # Required. The attribute values in case attribute data type is string or JSON. + "A String", + ], + }, + "uriValues": { # The attribute values of data type string or JSON. # The attribute values associated with a resource in case attribute data type is URL, URI or IP, like gs://bucket-name/object-name. + "values": [ # Required. The attribute values in case attribute data type is string or JSON. + "A String", + ], + }, + }, "updateTime": "A String", # Output only. The time at which the deployment was last updated. }
@@ -539,6 +655,34 @@

Method Details

], }, }, + "managementUrl": { # The attribute values associated with resource. # Optional. The uri where users can navigate to for the management of the deployment. This maps to the following system defined attribute: `projects/{project}/locations/{location}/attributes/system-management-url` The number of values for this attribute will be based on the cardinality of the attribute. The same can be retrieved via GetAttribute API. + "attribute": "A String", # Output only. The name of the attribute. Format: projects/{project}/locations/{location}/attributes/{attribute} + "enumValues": { # The attribute values of data type enum. # The attribute values associated with a resource in case attribute data type is enum. + "values": [ # Required. The attribute values in case attribute data type is enum. + { # The value that can be assigned to the attribute when the data type is enum. + "description": "A String", # Optional. The detailed description of the allowed value. + "displayName": "A String", # Required. The display name of the allowed value. + "id": "A String", # Required. The ID of the allowed value. * If provided, the same will be used. The service will throw an error if the specified id is already used by another allowed value in the same attribute resource. * If not provided, a system generated id derived from the display name will be used. In this case, the service will handle conflict resolution by adding a system generated suffix in case of duplicates. This value should be 4-63 characters, and valid characters are /a-z-/. + "immutable": True or False, # Optional. When set to true, the allowed value cannot be updated or deleted by the user. It can only be true for System defined attributes. + }, + ], + }, + "jsonValues": { # The attribute values of data type string or JSON. # The attribute values associated with a resource in case attribute data type is JSON. + "values": [ # Required. The attribute values in case attribute data type is string or JSON. + "A String", + ], + }, + "stringValues": { # The attribute values of data type string or JSON. # The attribute values associated with a resource in case attribute data type is string. + "values": [ # Required. The attribute values in case attribute data type is string or JSON. + "A String", + ], + }, + "uriValues": { # The attribute values of data type string or JSON. # The attribute values associated with a resource in case attribute data type is URL, URI or IP, like gs://bucket-name/object-name. + "values": [ # Required. The attribute values in case attribute data type is string or JSON. + "A String", + ], + }, + }, "name": "A String", # Identifier. The name of the deployment. Format: `projects/{project}/locations/{location}/deployments/{deployment}` "resourceUri": "A String", # Required. A uri that uniquely identfies the deployment within a particular gateway. For example, if the runtime resource is of type APIGEE_PROXY, then this field will be a combination of org, proxy name and environment. "slo": { # The attribute values associated with resource. # Optional. The SLO for this deployment. This maps to the following system defined attribute: `projects/{project}/locations/{location}/attributes/system-slo` attribute. The number of values for this attribute will be based on the cardinality of the attribute. The same can be retrieved via GetAttribute API. All values should be from the list of allowed values defined for the attribute. @@ -569,6 +713,7 @@

Method Details

], }, }, + "sourceEnvironment": "A String", # Optional. The environment at source for the deployment. For example: prod, dev, staging, etc. "sourceMetadata": [ # Output only. The list of sources and metadata from the sources of the deployment. { # SourceMetadata represents the metadata for a resource at the source. "originalResourceCreateTime": "A String", # Output only. The time at which the resource was created at the source. @@ -581,6 +726,35 @@

Method Details

"sourceType": "A String", # Output only. The type of the source. }, ], + "sourceProject": "A String", # Optional. The project to which the deployment belongs. For GCP gateways, this will refer to the project identifier. For others like Edge/OPDK, this will refer to the org identifier. + "sourceUri": { # The attribute values associated with resource. # Optional. The uri where additional source specific information for this deployment can be found. This maps to the following system defined attribute: `projects/{project}/locations/{location}/attributes/system-source-uri` The number of values for this attribute will be based on the cardinality of the attribute. The same can be retrieved via GetAttribute API. + "attribute": "A String", # Output only. The name of the attribute. Format: projects/{project}/locations/{location}/attributes/{attribute} + "enumValues": { # The attribute values of data type enum. # The attribute values associated with a resource in case attribute data type is enum. + "values": [ # Required. The attribute values in case attribute data type is enum. + { # The value that can be assigned to the attribute when the data type is enum. + "description": "A String", # Optional. The detailed description of the allowed value. + "displayName": "A String", # Required. The display name of the allowed value. + "id": "A String", # Required. The ID of the allowed value. * If provided, the same will be used. The service will throw an error if the specified id is already used by another allowed value in the same attribute resource. * If not provided, a system generated id derived from the display name will be used. In this case, the service will handle conflict resolution by adding a system generated suffix in case of duplicates. This value should be 4-63 characters, and valid characters are /a-z-/. + "immutable": True or False, # Optional. When set to true, the allowed value cannot be updated or deleted by the user. It can only be true for System defined attributes. + }, + ], + }, + "jsonValues": { # The attribute values of data type string or JSON. # The attribute values associated with a resource in case attribute data type is JSON. + "values": [ # Required. The attribute values in case attribute data type is string or JSON. + "A String", + ], + }, + "stringValues": { # The attribute values of data type string or JSON. # The attribute values associated with a resource in case attribute data type is string. + "values": [ # Required. The attribute values in case attribute data type is string or JSON. + "A String", + ], + }, + "uriValues": { # The attribute values of data type string or JSON. # The attribute values associated with a resource in case attribute data type is URL, URI or IP, like gs://bucket-name/object-name. + "values": [ # Required. The attribute values in case attribute data type is string or JSON. + "A String", + ], + }, + }, "updateTime": "A String", # Output only. The time at which the deployment was last updated. } @@ -703,6 +877,34 @@

Method Details

], }, }, + "managementUrl": { # The attribute values associated with resource. # Optional. The uri where users can navigate to for the management of the deployment. This maps to the following system defined attribute: `projects/{project}/locations/{location}/attributes/system-management-url` The number of values for this attribute will be based on the cardinality of the attribute. The same can be retrieved via GetAttribute API. + "attribute": "A String", # Output only. The name of the attribute. Format: projects/{project}/locations/{location}/attributes/{attribute} + "enumValues": { # The attribute values of data type enum. # The attribute values associated with a resource in case attribute data type is enum. + "values": [ # Required. The attribute values in case attribute data type is enum. + { # The value that can be assigned to the attribute when the data type is enum. + "description": "A String", # Optional. The detailed description of the allowed value. + "displayName": "A String", # Required. The display name of the allowed value. + "id": "A String", # Required. The ID of the allowed value. * If provided, the same will be used. The service will throw an error if the specified id is already used by another allowed value in the same attribute resource. * If not provided, a system generated id derived from the display name will be used. In this case, the service will handle conflict resolution by adding a system generated suffix in case of duplicates. This value should be 4-63 characters, and valid characters are /a-z-/. + "immutable": True or False, # Optional. When set to true, the allowed value cannot be updated or deleted by the user. It can only be true for System defined attributes. + }, + ], + }, + "jsonValues": { # The attribute values of data type string or JSON. # The attribute values associated with a resource in case attribute data type is JSON. + "values": [ # Required. The attribute values in case attribute data type is string or JSON. + "A String", + ], + }, + "stringValues": { # The attribute values of data type string or JSON. # The attribute values associated with a resource in case attribute data type is string. + "values": [ # Required. The attribute values in case attribute data type is string or JSON. + "A String", + ], + }, + "uriValues": { # The attribute values of data type string or JSON. # The attribute values associated with a resource in case attribute data type is URL, URI or IP, like gs://bucket-name/object-name. + "values": [ # Required. The attribute values in case attribute data type is string or JSON. + "A String", + ], + }, + }, "name": "A String", # Identifier. The name of the deployment. Format: `projects/{project}/locations/{location}/deployments/{deployment}` "resourceUri": "A String", # Required. A uri that uniquely identfies the deployment within a particular gateway. For example, if the runtime resource is of type APIGEE_PROXY, then this field will be a combination of org, proxy name and environment. "slo": { # The attribute values associated with resource. # Optional. The SLO for this deployment. This maps to the following system defined attribute: `projects/{project}/locations/{location}/attributes/system-slo` attribute. The number of values for this attribute will be based on the cardinality of the attribute. The same can be retrieved via GetAttribute API. All values should be from the list of allowed values defined for the attribute. @@ -733,6 +935,7 @@

Method Details

], }, }, + "sourceEnvironment": "A String", # Optional. The environment at source for the deployment. For example: prod, dev, staging, etc. "sourceMetadata": [ # Output only. The list of sources and metadata from the sources of the deployment. { # SourceMetadata represents the metadata for a resource at the source. "originalResourceCreateTime": "A String", # Output only. The time at which the resource was created at the source. @@ -745,6 +948,35 @@

Method Details

"sourceType": "A String", # Output only. The type of the source. }, ], + "sourceProject": "A String", # Optional. The project to which the deployment belongs. For GCP gateways, this will refer to the project identifier. For others like Edge/OPDK, this will refer to the org identifier. + "sourceUri": { # The attribute values associated with resource. # Optional. The uri where additional source specific information for this deployment can be found. This maps to the following system defined attribute: `projects/{project}/locations/{location}/attributes/system-source-uri` The number of values for this attribute will be based on the cardinality of the attribute. The same can be retrieved via GetAttribute API. + "attribute": "A String", # Output only. The name of the attribute. Format: projects/{project}/locations/{location}/attributes/{attribute} + "enumValues": { # The attribute values of data type enum. # The attribute values associated with a resource in case attribute data type is enum. + "values": [ # Required. The attribute values in case attribute data type is enum. + { # The value that can be assigned to the attribute when the data type is enum. + "description": "A String", # Optional. The detailed description of the allowed value. + "displayName": "A String", # Required. The display name of the allowed value. + "id": "A String", # Required. The ID of the allowed value. * If provided, the same will be used. The service will throw an error if the specified id is already used by another allowed value in the same attribute resource. * If not provided, a system generated id derived from the display name will be used. In this case, the service will handle conflict resolution by adding a system generated suffix in case of duplicates. This value should be 4-63 characters, and valid characters are /a-z-/. + "immutable": True or False, # Optional. When set to true, the allowed value cannot be updated or deleted by the user. It can only be true for System defined attributes. + }, + ], + }, + "jsonValues": { # The attribute values of data type string or JSON. # The attribute values associated with a resource in case attribute data type is JSON. + "values": [ # Required. The attribute values in case attribute data type is string or JSON. + "A String", + ], + }, + "stringValues": { # The attribute values of data type string or JSON. # The attribute values associated with a resource in case attribute data type is string. + "values": [ # Required. The attribute values in case attribute data type is string or JSON. + "A String", + ], + }, + "uriValues": { # The attribute values of data type string or JSON. # The attribute values associated with a resource in case attribute data type is URL, URI or IP, like gs://bucket-name/object-name. + "values": [ # Required. The attribute values in case attribute data type is string or JSON. + "A String", + ], + }, + }, "updateTime": "A String", # Output only. The time at which the deployment was last updated. }, ], @@ -874,6 +1106,34 @@

Method Details

], }, }, + "managementUrl": { # The attribute values associated with resource. # Optional. The uri where users can navigate to for the management of the deployment. This maps to the following system defined attribute: `projects/{project}/locations/{location}/attributes/system-management-url` The number of values for this attribute will be based on the cardinality of the attribute. The same can be retrieved via GetAttribute API. + "attribute": "A String", # Output only. The name of the attribute. Format: projects/{project}/locations/{location}/attributes/{attribute} + "enumValues": { # The attribute values of data type enum. # The attribute values associated with a resource in case attribute data type is enum. + "values": [ # Required. The attribute values in case attribute data type is enum. + { # The value that can be assigned to the attribute when the data type is enum. + "description": "A String", # Optional. The detailed description of the allowed value. + "displayName": "A String", # Required. The display name of the allowed value. + "id": "A String", # Required. The ID of the allowed value. * If provided, the same will be used. The service will throw an error if the specified id is already used by another allowed value in the same attribute resource. * If not provided, a system generated id derived from the display name will be used. In this case, the service will handle conflict resolution by adding a system generated suffix in case of duplicates. This value should be 4-63 characters, and valid characters are /a-z-/. + "immutable": True or False, # Optional. When set to true, the allowed value cannot be updated or deleted by the user. It can only be true for System defined attributes. + }, + ], + }, + "jsonValues": { # The attribute values of data type string or JSON. # The attribute values associated with a resource in case attribute data type is JSON. + "values": [ # Required. The attribute values in case attribute data type is string or JSON. + "A String", + ], + }, + "stringValues": { # The attribute values of data type string or JSON. # The attribute values associated with a resource in case attribute data type is string. + "values": [ # Required. The attribute values in case attribute data type is string or JSON. + "A String", + ], + }, + "uriValues": { # The attribute values of data type string or JSON. # The attribute values associated with a resource in case attribute data type is URL, URI or IP, like gs://bucket-name/object-name. + "values": [ # Required. The attribute values in case attribute data type is string or JSON. + "A String", + ], + }, + }, "name": "A String", # Identifier. The name of the deployment. Format: `projects/{project}/locations/{location}/deployments/{deployment}` "resourceUri": "A String", # Required. A uri that uniquely identfies the deployment within a particular gateway. For example, if the runtime resource is of type APIGEE_PROXY, then this field will be a combination of org, proxy name and environment. "slo": { # The attribute values associated with resource. # Optional. The SLO for this deployment. This maps to the following system defined attribute: `projects/{project}/locations/{location}/attributes/system-slo` attribute. The number of values for this attribute will be based on the cardinality of the attribute. The same can be retrieved via GetAttribute API. All values should be from the list of allowed values defined for the attribute. @@ -904,6 +1164,7 @@

Method Details

], }, }, + "sourceEnvironment": "A String", # Optional. The environment at source for the deployment. For example: prod, dev, staging, etc. "sourceMetadata": [ # Output only. The list of sources and metadata from the sources of the deployment. { # SourceMetadata represents the metadata for a resource at the source. "originalResourceCreateTime": "A String", # Output only. The time at which the resource was created at the source. @@ -916,6 +1177,35 @@

Method Details

"sourceType": "A String", # Output only. The type of the source. }, ], + "sourceProject": "A String", # Optional. The project to which the deployment belongs. For GCP gateways, this will refer to the project identifier. For others like Edge/OPDK, this will refer to the org identifier. + "sourceUri": { # The attribute values associated with resource. # Optional. The uri where additional source specific information for this deployment can be found. This maps to the following system defined attribute: `projects/{project}/locations/{location}/attributes/system-source-uri` The number of values for this attribute will be based on the cardinality of the attribute. The same can be retrieved via GetAttribute API. + "attribute": "A String", # Output only. The name of the attribute. Format: projects/{project}/locations/{location}/attributes/{attribute} + "enumValues": { # The attribute values of data type enum. # The attribute values associated with a resource in case attribute data type is enum. + "values": [ # Required. The attribute values in case attribute data type is enum. + { # The value that can be assigned to the attribute when the data type is enum. + "description": "A String", # Optional. The detailed description of the allowed value. + "displayName": "A String", # Required. The display name of the allowed value. + "id": "A String", # Required. The ID of the allowed value. * If provided, the same will be used. The service will throw an error if the specified id is already used by another allowed value in the same attribute resource. * If not provided, a system generated id derived from the display name will be used. In this case, the service will handle conflict resolution by adding a system generated suffix in case of duplicates. This value should be 4-63 characters, and valid characters are /a-z-/. + "immutable": True or False, # Optional. When set to true, the allowed value cannot be updated or deleted by the user. It can only be true for System defined attributes. + }, + ], + }, + "jsonValues": { # The attribute values of data type string or JSON. # The attribute values associated with a resource in case attribute data type is JSON. + "values": [ # Required. The attribute values in case attribute data type is string or JSON. + "A String", + ], + }, + "stringValues": { # The attribute values of data type string or JSON. # The attribute values associated with a resource in case attribute data type is string. + "values": [ # Required. The attribute values in case attribute data type is string or JSON. + "A String", + ], + }, + "uriValues": { # The attribute values of data type string or JSON. # The attribute values associated with a resource in case attribute data type is URL, URI or IP, like gs://bucket-name/object-name. + "values": [ # Required. The attribute values in case attribute data type is string or JSON. + "A String", + ], + }, + }, "updateTime": "A String", # Output only. The time at which the deployment was last updated. } @@ -1027,6 +1317,34 @@

Method Details

], }, }, + "managementUrl": { # The attribute values associated with resource. # Optional. The uri where users can navigate to for the management of the deployment. This maps to the following system defined attribute: `projects/{project}/locations/{location}/attributes/system-management-url` The number of values for this attribute will be based on the cardinality of the attribute. The same can be retrieved via GetAttribute API. + "attribute": "A String", # Output only. The name of the attribute. Format: projects/{project}/locations/{location}/attributes/{attribute} + "enumValues": { # The attribute values of data type enum. # The attribute values associated with a resource in case attribute data type is enum. + "values": [ # Required. The attribute values in case attribute data type is enum. + { # The value that can be assigned to the attribute when the data type is enum. + "description": "A String", # Optional. The detailed description of the allowed value. + "displayName": "A String", # Required. The display name of the allowed value. + "id": "A String", # Required. The ID of the allowed value. * If provided, the same will be used. The service will throw an error if the specified id is already used by another allowed value in the same attribute resource. * If not provided, a system generated id derived from the display name will be used. In this case, the service will handle conflict resolution by adding a system generated suffix in case of duplicates. This value should be 4-63 characters, and valid characters are /a-z-/. + "immutable": True or False, # Optional. When set to true, the allowed value cannot be updated or deleted by the user. It can only be true for System defined attributes. + }, + ], + }, + "jsonValues": { # The attribute values of data type string or JSON. # The attribute values associated with a resource in case attribute data type is JSON. + "values": [ # Required. The attribute values in case attribute data type is string or JSON. + "A String", + ], + }, + "stringValues": { # The attribute values of data type string or JSON. # The attribute values associated with a resource in case attribute data type is string. + "values": [ # Required. The attribute values in case attribute data type is string or JSON. + "A String", + ], + }, + "uriValues": { # The attribute values of data type string or JSON. # The attribute values associated with a resource in case attribute data type is URL, URI or IP, like gs://bucket-name/object-name. + "values": [ # Required. The attribute values in case attribute data type is string or JSON. + "A String", + ], + }, + }, "name": "A String", # Identifier. The name of the deployment. Format: `projects/{project}/locations/{location}/deployments/{deployment}` "resourceUri": "A String", # Required. A uri that uniquely identfies the deployment within a particular gateway. For example, if the runtime resource is of type APIGEE_PROXY, then this field will be a combination of org, proxy name and environment. "slo": { # The attribute values associated with resource. # Optional. The SLO for this deployment. This maps to the following system defined attribute: `projects/{project}/locations/{location}/attributes/system-slo` attribute. The number of values for this attribute will be based on the cardinality of the attribute. The same can be retrieved via GetAttribute API. All values should be from the list of allowed values defined for the attribute. @@ -1057,6 +1375,7 @@

Method Details

], }, }, + "sourceEnvironment": "A String", # Optional. The environment at source for the deployment. For example: prod, dev, staging, etc. "sourceMetadata": [ # Output only. The list of sources and metadata from the sources of the deployment. { # SourceMetadata represents the metadata for a resource at the source. "originalResourceCreateTime": "A String", # Output only. The time at which the resource was created at the source. @@ -1069,6 +1388,35 @@

Method Details

"sourceType": "A String", # Output only. The type of the source. }, ], + "sourceProject": "A String", # Optional. The project to which the deployment belongs. For GCP gateways, this will refer to the project identifier. For others like Edge/OPDK, this will refer to the org identifier. + "sourceUri": { # The attribute values associated with resource. # Optional. The uri where additional source specific information for this deployment can be found. This maps to the following system defined attribute: `projects/{project}/locations/{location}/attributes/system-source-uri` The number of values for this attribute will be based on the cardinality of the attribute. The same can be retrieved via GetAttribute API. + "attribute": "A String", # Output only. The name of the attribute. Format: projects/{project}/locations/{location}/attributes/{attribute} + "enumValues": { # The attribute values of data type enum. # The attribute values associated with a resource in case attribute data type is enum. + "values": [ # Required. The attribute values in case attribute data type is enum. + { # The value that can be assigned to the attribute when the data type is enum. + "description": "A String", # Optional. The detailed description of the allowed value. + "displayName": "A String", # Required. The display name of the allowed value. + "id": "A String", # Required. The ID of the allowed value. * If provided, the same will be used. The service will throw an error if the specified id is already used by another allowed value in the same attribute resource. * If not provided, a system generated id derived from the display name will be used. In this case, the service will handle conflict resolution by adding a system generated suffix in case of duplicates. This value should be 4-63 characters, and valid characters are /a-z-/. + "immutable": True or False, # Optional. When set to true, the allowed value cannot be updated or deleted by the user. It can only be true for System defined attributes. + }, + ], + }, + "jsonValues": { # The attribute values of data type string or JSON. # The attribute values associated with a resource in case attribute data type is JSON. + "values": [ # Required. The attribute values in case attribute data type is string or JSON. + "A String", + ], + }, + "stringValues": { # The attribute values of data type string or JSON. # The attribute values associated with a resource in case attribute data type is string. + "values": [ # Required. The attribute values in case attribute data type is string or JSON. + "A String", + ], + }, + "uriValues": { # The attribute values of data type string or JSON. # The attribute values associated with a resource in case attribute data type is URL, URI or IP, like gs://bucket-name/object-name. + "values": [ # Required. The attribute values in case attribute data type is string or JSON. + "A String", + ], + }, + }, "updateTime": "A String", # Output only. The time at which the deployment was last updated. } diff --git a/docs/dyn/apihub_v1.projects.locations.html b/docs/dyn/apihub_v1.projects.locations.html index 0ee2e39f21..d38589902c 100644 --- a/docs/dyn/apihub_v1.projects.locations.html +++ b/docs/dyn/apihub_v1.projects.locations.html @@ -565,6 +565,34 @@

Method Details

], }, }, + "managementUrl": { # The attribute values associated with resource. # Optional. The uri where users can navigate to for the management of the deployment. This maps to the following system defined attribute: `projects/{project}/locations/{location}/attributes/system-management-url` The number of values for this attribute will be based on the cardinality of the attribute. The same can be retrieved via GetAttribute API. + "attribute": "A String", # Output only. The name of the attribute. Format: projects/{project}/locations/{location}/attributes/{attribute} + "enumValues": { # The attribute values of data type enum. # The attribute values associated with a resource in case attribute data type is enum. + "values": [ # Required. The attribute values in case attribute data type is enum. + { # The value that can be assigned to the attribute when the data type is enum. + "description": "A String", # Optional. The detailed description of the allowed value. + "displayName": "A String", # Required. The display name of the allowed value. + "id": "A String", # Required. The ID of the allowed value. * If provided, the same will be used. The service will throw an error if the specified id is already used by another allowed value in the same attribute resource. * If not provided, a system generated id derived from the display name will be used. In this case, the service will handle conflict resolution by adding a system generated suffix in case of duplicates. This value should be 4-63 characters, and valid characters are /a-z-/. + "immutable": True or False, # Optional. When set to true, the allowed value cannot be updated or deleted by the user. It can only be true for System defined attributes. + }, + ], + }, + "jsonValues": { # The attribute values of data type string or JSON. # The attribute values associated with a resource in case attribute data type is JSON. + "values": [ # Required. The attribute values in case attribute data type is string or JSON. + "A String", + ], + }, + "stringValues": { # The attribute values of data type string or JSON. # The attribute values associated with a resource in case attribute data type is string. + "values": [ # Required. The attribute values in case attribute data type is string or JSON. + "A String", + ], + }, + "uriValues": { # The attribute values of data type string or JSON. # The attribute values associated with a resource in case attribute data type is URL, URI or IP, like gs://bucket-name/object-name. + "values": [ # Required. The attribute values in case attribute data type is string or JSON. + "A String", + ], + }, + }, "name": "A String", # Identifier. The name of the deployment. Format: `projects/{project}/locations/{location}/deployments/{deployment}` "resourceUri": "A String", # Required. A uri that uniquely identfies the deployment within a particular gateway. For example, if the runtime resource is of type APIGEE_PROXY, then this field will be a combination of org, proxy name and environment. "slo": { # The attribute values associated with resource. # Optional. The SLO for this deployment. This maps to the following system defined attribute: `projects/{project}/locations/{location}/attributes/system-slo` attribute. The number of values for this attribute will be based on the cardinality of the attribute. The same can be retrieved via GetAttribute API. All values should be from the list of allowed values defined for the attribute. @@ -595,6 +623,7 @@

Method Details

], }, }, + "sourceEnvironment": "A String", # Optional. The environment at source for the deployment. For example: prod, dev, staging, etc. "sourceMetadata": [ # Output only. The list of sources and metadata from the sources of the deployment. { # SourceMetadata represents the metadata for a resource at the source. "originalResourceCreateTime": "A String", # Output only. The time at which the resource was created at the source. @@ -607,6 +636,35 @@

Method Details

"sourceType": "A String", # Output only. The type of the source. }, ], + "sourceProject": "A String", # Optional. The project to which the deployment belongs. For GCP gateways, this will refer to the project identifier. For others like Edge/OPDK, this will refer to the org identifier. + "sourceUri": { # The attribute values associated with resource. # Optional. The uri where additional source specific information for this deployment can be found. This maps to the following system defined attribute: `projects/{project}/locations/{location}/attributes/system-source-uri` The number of values for this attribute will be based on the cardinality of the attribute. The same can be retrieved via GetAttribute API. + "attribute": "A String", # Output only. The name of the attribute. Format: projects/{project}/locations/{location}/attributes/{attribute} + "enumValues": { # The attribute values of data type enum. # The attribute values associated with a resource in case attribute data type is enum. + "values": [ # Required. The attribute values in case attribute data type is enum. + { # The value that can be assigned to the attribute when the data type is enum. + "description": "A String", # Optional. The detailed description of the allowed value. + "displayName": "A String", # Required. The display name of the allowed value. + "id": "A String", # Required. The ID of the allowed value. * If provided, the same will be used. The service will throw an error if the specified id is already used by another allowed value in the same attribute resource. * If not provided, a system generated id derived from the display name will be used. In this case, the service will handle conflict resolution by adding a system generated suffix in case of duplicates. This value should be 4-63 characters, and valid characters are /a-z-/. + "immutable": True or False, # Optional. When set to true, the allowed value cannot be updated or deleted by the user. It can only be true for System defined attributes. + }, + ], + }, + "jsonValues": { # The attribute values of data type string or JSON. # The attribute values associated with a resource in case attribute data type is JSON. + "values": [ # Required. The attribute values in case attribute data type is string or JSON. + "A String", + ], + }, + "stringValues": { # The attribute values of data type string or JSON. # The attribute values associated with a resource in case attribute data type is string. + "values": [ # Required. The attribute values in case attribute data type is string or JSON. + "A String", + ], + }, + "uriValues": { # The attribute values of data type string or JSON. # The attribute values associated with a resource in case attribute data type is URL, URI or IP, like gs://bucket-name/object-name. + "values": [ # Required. The attribute values in case attribute data type is string or JSON. + "A String", + ], + }, + }, "updateTime": "A String", # Output only. The time at which the deployment was last updated. }, "originalCreateTime": "A String", # Optional. Timestamp indicating when the deployment was created at the source. @@ -1494,6 +1552,34 @@

Method Details

], }, }, + "managementUrl": { # The attribute values associated with resource. # Optional. The uri where users can navigate to for the management of the deployment. This maps to the following system defined attribute: `projects/{project}/locations/{location}/attributes/system-management-url` The number of values for this attribute will be based on the cardinality of the attribute. The same can be retrieved via GetAttribute API. + "attribute": "A String", # Output only. The name of the attribute. Format: projects/{project}/locations/{location}/attributes/{attribute} + "enumValues": { # The attribute values of data type enum. # The attribute values associated with a resource in case attribute data type is enum. + "values": [ # Required. The attribute values in case attribute data type is enum. + { # The value that can be assigned to the attribute when the data type is enum. + "description": "A String", # Optional. The detailed description of the allowed value. + "displayName": "A String", # Required. The display name of the allowed value. + "id": "A String", # Required. The ID of the allowed value. * If provided, the same will be used. The service will throw an error if the specified id is already used by another allowed value in the same attribute resource. * If not provided, a system generated id derived from the display name will be used. In this case, the service will handle conflict resolution by adding a system generated suffix in case of duplicates. This value should be 4-63 characters, and valid characters are /a-z-/. + "immutable": True or False, # Optional. When set to true, the allowed value cannot be updated or deleted by the user. It can only be true for System defined attributes. + }, + ], + }, + "jsonValues": { # The attribute values of data type string or JSON. # The attribute values associated with a resource in case attribute data type is JSON. + "values": [ # Required. The attribute values in case attribute data type is string or JSON. + "A String", + ], + }, + "stringValues": { # The attribute values of data type string or JSON. # The attribute values associated with a resource in case attribute data type is string. + "values": [ # Required. The attribute values in case attribute data type is string or JSON. + "A String", + ], + }, + "uriValues": { # The attribute values of data type string or JSON. # The attribute values associated with a resource in case attribute data type is URL, URI or IP, like gs://bucket-name/object-name. + "values": [ # Required. The attribute values in case attribute data type is string or JSON. + "A String", + ], + }, + }, "name": "A String", # Identifier. The name of the deployment. Format: `projects/{project}/locations/{location}/deployments/{deployment}` "resourceUri": "A String", # Required. A uri that uniquely identfies the deployment within a particular gateway. For example, if the runtime resource is of type APIGEE_PROXY, then this field will be a combination of org, proxy name and environment. "slo": { # The attribute values associated with resource. # Optional. The SLO for this deployment. This maps to the following system defined attribute: `projects/{project}/locations/{location}/attributes/system-slo` attribute. The number of values for this attribute will be based on the cardinality of the attribute. The same can be retrieved via GetAttribute API. All values should be from the list of allowed values defined for the attribute. @@ -1524,6 +1610,7 @@

Method Details

], }, }, + "sourceEnvironment": "A String", # Optional. The environment at source for the deployment. For example: prod, dev, staging, etc. "sourceMetadata": [ # Output only. The list of sources and metadata from the sources of the deployment. { # SourceMetadata represents the metadata for a resource at the source. "originalResourceCreateTime": "A String", # Output only. The time at which the resource was created at the source. @@ -1536,6 +1623,35 @@

Method Details

"sourceType": "A String", # Output only. The type of the source. }, ], + "sourceProject": "A String", # Optional. The project to which the deployment belongs. For GCP gateways, this will refer to the project identifier. For others like Edge/OPDK, this will refer to the org identifier. + "sourceUri": { # The attribute values associated with resource. # Optional. The uri where additional source specific information for this deployment can be found. This maps to the following system defined attribute: `projects/{project}/locations/{location}/attributes/system-source-uri` The number of values for this attribute will be based on the cardinality of the attribute. The same can be retrieved via GetAttribute API. + "attribute": "A String", # Output only. The name of the attribute. Format: projects/{project}/locations/{location}/attributes/{attribute} + "enumValues": { # The attribute values of data type enum. # The attribute values associated with a resource in case attribute data type is enum. + "values": [ # Required. The attribute values in case attribute data type is enum. + { # The value that can be assigned to the attribute when the data type is enum. + "description": "A String", # Optional. The detailed description of the allowed value. + "displayName": "A String", # Required. The display name of the allowed value. + "id": "A String", # Required. The ID of the allowed value. * If provided, the same will be used. The service will throw an error if the specified id is already used by another allowed value in the same attribute resource. * If not provided, a system generated id derived from the display name will be used. In this case, the service will handle conflict resolution by adding a system generated suffix in case of duplicates. This value should be 4-63 characters, and valid characters are /a-z-/. + "immutable": True or False, # Optional. When set to true, the allowed value cannot be updated or deleted by the user. It can only be true for System defined attributes. + }, + ], + }, + "jsonValues": { # The attribute values of data type string or JSON. # The attribute values associated with a resource in case attribute data type is JSON. + "values": [ # Required. The attribute values in case attribute data type is string or JSON. + "A String", + ], + }, + "stringValues": { # The attribute values of data type string or JSON. # The attribute values associated with a resource in case attribute data type is string. + "values": [ # Required. The attribute values in case attribute data type is string or JSON. + "A String", + ], + }, + "uriValues": { # The attribute values of data type string or JSON. # The attribute values associated with a resource in case attribute data type is URL, URI or IP, like gs://bucket-name/object-name. + "values": [ # Required. The attribute values in case attribute data type is string or JSON. + "A String", + ], + }, + }, "updateTime": "A String", # Output only. The time at which the deployment was last updated. }, "operation": { # Represents an operation contained in an API version in the API Hub. An operation is added/updated/deleted in an API version when a new spec is added or an existing spec is updated/deleted in a version. Currently, an operation will be created only corresponding to OpenAPI spec as parsing is supported for OpenAPI spec. Alternatively operations can be managed via create,update and delete APIs, creation of apiOperation can be possible only for version with no parsed operations and update/delete can be possible only for operations created via create API. # This represents ApiOperation resource in search results. Only name, description, spec and details fields are populated in search results. diff --git a/docs/dyn/apihub_v1.projects.locations.plugins.html b/docs/dyn/apihub_v1.projects.locations.plugins.html index d5d69cfc29..abfb1b35b6 100644 --- a/docs/dyn/apihub_v1.projects.locations.plugins.html +++ b/docs/dyn/apihub_v1.projects.locations.plugins.html @@ -177,6 +177,7 @@

Method Details

"documentation": { # Documentation details. # Optional. The documentation of the plugin, that explains how to set up and use the plugin. "externalUri": "A String", # Optional. The uri of the externally hosted documentation. }, + "gatewayType": "A String", # Optional. The type of the gateway. "hostingService": { # The information related to the service implemented by the plugin developer, used to invoke the plugin's functionality. # Optional. This field is optional. It is used to notify the plugin hosting service for any lifecycle changes of the plugin instance and trigger execution of plugin instance actions in case of API hub managed actions. This field should be provided if the plugin instance lifecycle of the developed plugin needs to be managed from API hub. Also, in this case the plugin hosting service interface needs to be implemented. This field should not be provided if the plugin wants to manage plugin instance lifecycle events outside of hub interface and use plugin framework for only registering of plugin and plugin instances to capture the source of data into hub. Note, in this case the plugin hosting service interface is not required to be implemented. Also, the plugin instance lifecycle actions will be disabled from API hub's UI. "serviceUri": "A String", # Optional. The URI of the service implemented by the plugin developer, used to invoke the plugin's functionality. This information is only required for user defined plugins. }, @@ -272,6 +273,7 @@

Method Details

"documentation": { # Documentation details. # Optional. The documentation of the plugin, that explains how to set up and use the plugin. "externalUri": "A String", # Optional. The uri of the externally hosted documentation. }, + "gatewayType": "A String", # Optional. The type of the gateway. "hostingService": { # The information related to the service implemented by the plugin developer, used to invoke the plugin's functionality. # Optional. This field is optional. It is used to notify the plugin hosting service for any lifecycle changes of the plugin instance and trigger execution of plugin instance actions in case of API hub managed actions. This field should be provided if the plugin instance lifecycle of the developed plugin needs to be managed from API hub. Also, in this case the plugin hosting service interface needs to be implemented. This field should not be provided if the plugin wants to manage plugin instance lifecycle events outside of hub interface and use plugin framework for only registering of plugin and plugin instances to capture the source of data into hub. Note, in this case the plugin hosting service interface is not required to be implemented. Also, the plugin instance lifecycle actions will be disabled from API hub's UI. "serviceUri": "A String", # Optional. The URI of the service implemented by the plugin developer, used to invoke the plugin's functionality. This information is only required for user defined plugins. }, @@ -414,6 +416,7 @@

Method Details

"documentation": { # Documentation details. # Optional. The documentation of the plugin, that explains how to set up and use the plugin. "externalUri": "A String", # Optional. The uri of the externally hosted documentation. }, + "gatewayType": "A String", # Optional. The type of the gateway. "hostingService": { # The information related to the service implemented by the plugin developer, used to invoke the plugin's functionality. # Optional. This field is optional. It is used to notify the plugin hosting service for any lifecycle changes of the plugin instance and trigger execution of plugin instance actions in case of API hub managed actions. This field should be provided if the plugin instance lifecycle of the developed plugin needs to be managed from API hub. Also, in this case the plugin hosting service interface needs to be implemented. This field should not be provided if the plugin wants to manage plugin instance lifecycle events outside of hub interface and use plugin framework for only registering of plugin and plugin instances to capture the source of data into hub. Note, in this case the plugin hosting service interface is not required to be implemented. Also, the plugin instance lifecycle actions will be disabled from API hub's UI. "serviceUri": "A String", # Optional. The URI of the service implemented by the plugin developer, used to invoke the plugin's functionality. This information is only required for user defined plugins. }, @@ -521,6 +524,7 @@

Method Details

"documentation": { # Documentation details. # Optional. The documentation of the plugin, that explains how to set up and use the plugin. "externalUri": "A String", # Optional. The uri of the externally hosted documentation. }, + "gatewayType": "A String", # Optional. The type of the gateway. "hostingService": { # The information related to the service implemented by the plugin developer, used to invoke the plugin's functionality. # Optional. This field is optional. It is used to notify the plugin hosting service for any lifecycle changes of the plugin instance and trigger execution of plugin instance actions in case of API hub managed actions. This field should be provided if the plugin instance lifecycle of the developed plugin needs to be managed from API hub. Also, in this case the plugin hosting service interface needs to be implemented. This field should not be provided if the plugin wants to manage plugin instance lifecycle events outside of hub interface and use plugin framework for only registering of plugin and plugin instances to capture the source of data into hub. Note, in this case the plugin hosting service interface is not required to be implemented. Also, the plugin instance lifecycle actions will be disabled from API hub's UI. "serviceUri": "A String", # Optional. The URI of the service implemented by the plugin developer, used to invoke the plugin's functionality. This information is only required for user defined plugins. }, @@ -622,6 +626,7 @@

Method Details

"documentation": { # Documentation details. # Optional. The documentation of the plugin, that explains how to set up and use the plugin. "externalUri": "A String", # Optional. The uri of the externally hosted documentation. }, + "gatewayType": "A String", # Optional. The type of the gateway. "hostingService": { # The information related to the service implemented by the plugin developer, used to invoke the plugin's functionality. # Optional. This field is optional. It is used to notify the plugin hosting service for any lifecycle changes of the plugin instance and trigger execution of plugin instance actions in case of API hub managed actions. This field should be provided if the plugin instance lifecycle of the developed plugin needs to be managed from API hub. Also, in this case the plugin hosting service interface needs to be implemented. This field should not be provided if the plugin wants to manage plugin instance lifecycle events outside of hub interface and use plugin framework for only registering of plugin and plugin instances to capture the source of data into hub. Note, in this case the plugin hosting service interface is not required to be implemented. Also, the plugin instance lifecycle actions will be disabled from API hub's UI. "serviceUri": "A String", # Optional. The URI of the service implemented by the plugin developer, used to invoke the plugin's functionality. This information is only required for user defined plugins. }, @@ -753,6 +758,7 @@

Method Details

"documentation": { # Documentation details. # Optional. The documentation of the plugin, that explains how to set up and use the plugin. "externalUri": "A String", # Optional. The uri of the externally hosted documentation. }, + "gatewayType": "A String", # Optional. The type of the gateway. "hostingService": { # The information related to the service implemented by the plugin developer, used to invoke the plugin's functionality. # Optional. This field is optional. It is used to notify the plugin hosting service for any lifecycle changes of the plugin instance and trigger execution of plugin instance actions in case of API hub managed actions. This field should be provided if the plugin instance lifecycle of the developed plugin needs to be managed from API hub. Also, in this case the plugin hosting service interface needs to be implemented. This field should not be provided if the plugin wants to manage plugin instance lifecycle events outside of hub interface and use plugin framework for only registering of plugin and plugin instances to capture the source of data into hub. Note, in this case the plugin hosting service interface is not required to be implemented. Also, the plugin instance lifecycle actions will be disabled from API hub's UI. "serviceUri": "A String", # Optional. The URI of the service implemented by the plugin developer, used to invoke the plugin's functionality. This information is only required for user defined plugins. }, diff --git a/docs/dyn/apihub_v1.projects.locations.plugins.instances.html b/docs/dyn/apihub_v1.projects.locations.plugins.instances.html index 335491a550..b64762ac33 100644 --- a/docs/dyn/apihub_v1.projects.locations.plugins.instances.html +++ b/docs/dyn/apihub_v1.projects.locations.plugins.instances.html @@ -101,6 +101,9 @@

Instance Methods

list_next()

Retrieves the next page of results.

+

+ patch(name, body=None, updateMask=None, x__xgafv=None)

+

Updates a plugin instance in the API hub. The following fields in the plugin_instance can be updated currently: * display_name * schedule_cron_expression The update_mask should be used to specify the fields being updated. To update the auth_config and additional_config of the plugin instance, use the ApplyPluginInstanceConfig method.

Method Details

close() @@ -135,8 +138,13 @@

Method Details

"startTime": "A String", # Output only. The last execution start time of the plugin instance. }, }, + "resourceConfig": { # The configuration of resources created for a given plugin instance action. # Output only. The configuration of resources created for a given plugin instance action. Note these will be returned only in case of Non-GCP plugins like OPDK. + "actionType": "A String", # Output only. The type of the action. + "pubsubTopic": "A String", # Output only. The pubsub topic to publish the data to. Format is projects/{project}/topics/{topic} + }, "scheduleCronExpression": "A String", # Optional. The schedule for this plugin instance action. This can only be set if the plugin supports API_HUB_SCHEDULE_TRIGGER mode for this action. "scheduleTimeZone": "A String", # Optional. The time zone for the schedule cron expression. If not provided, UTC will be used. + "serviceAccount": "A String", # Optional. The service account used to publish data. Note, the service account will only be accepted for non GCP plugins like OPDK. "state": "A String", # Output only. The current state of the plugin action in the plugin instance. }, ], @@ -204,6 +212,7 @@

Method Details

"displayName": "A String", # Required. The display name for this plugin instance. Max length is 255 characters. "errorMessage": "A String", # Output only. Error message describing the failure, if any, during Create, Delete or ApplyConfig operation corresponding to the plugin instance.This field will only be populated if the plugin instance is in the ERROR or FAILED state. "name": "A String", # Identifier. The unique name of the plugin instance resource. Format: `projects/{project}/locations/{location}/plugins/{plugin}/instances/{instance}` + "sourceProjectId": "A String", # Optional. The source project id of the plugin instance. This will be the id of runtime project in case of gcp based plugins and org id in case of non gcp based plugins. This is a required field. "state": "A String", # Output only. The current state of the plugin instance (e.g., enabled, disabled, provisioning). "updateTime": "A String", # Output only. Timestamp indicating when the plugin instance was last updated. } @@ -434,8 +443,13 @@

Method Details

"startTime": "A String", # Output only. The last execution start time of the plugin instance. }, }, + "resourceConfig": { # The configuration of resources created for a given plugin instance action. # Output only. The configuration of resources created for a given plugin instance action. Note these will be returned only in case of Non-GCP plugins like OPDK. + "actionType": "A String", # Output only. The type of the action. + "pubsubTopic": "A String", # Output only. The pubsub topic to publish the data to. Format is projects/{project}/topics/{topic} + }, "scheduleCronExpression": "A String", # Optional. The schedule for this plugin instance action. This can only be set if the plugin supports API_HUB_SCHEDULE_TRIGGER mode for this action. "scheduleTimeZone": "A String", # Optional. The time zone for the schedule cron expression. If not provided, UTC will be used. + "serviceAccount": "A String", # Optional. The service account used to publish data. Note, the service account will only be accepted for non GCP plugins like OPDK. "state": "A String", # Output only. The current state of the plugin action in the plugin instance. }, ], @@ -503,6 +517,7 @@

Method Details

"displayName": "A String", # Required. The display name for this plugin instance. Max length is 255 characters. "errorMessage": "A String", # Output only. Error message describing the failure, if any, during Create, Delete or ApplyConfig operation corresponding to the plugin instance.This field will only be populated if the plugin instance is in the ERROR or FAILED state. "name": "A String", # Identifier. The unique name of the plugin instance resource. Format: `projects/{project}/locations/{location}/plugins/{plugin}/instances/{instance}` + "sourceProjectId": "A String", # Optional. The source project id of the plugin instance. This will be the id of runtime project in case of gcp based plugins and org id in case of non gcp based plugins. This is a required field. "state": "A String", # Output only. The current state of the plugin instance (e.g., enabled, disabled, provisioning). "updateTime": "A String", # Output only. Timestamp indicating when the plugin instance was last updated. } @@ -547,8 +562,13 @@

Method Details

"startTime": "A String", # Output only. The last execution start time of the plugin instance. }, }, + "resourceConfig": { # The configuration of resources created for a given plugin instance action. # Output only. The configuration of resources created for a given plugin instance action. Note these will be returned only in case of Non-GCP plugins like OPDK. + "actionType": "A String", # Output only. The type of the action. + "pubsubTopic": "A String", # Output only. The pubsub topic to publish the data to. Format is projects/{project}/topics/{topic} + }, "scheduleCronExpression": "A String", # Optional. The schedule for this plugin instance action. This can only be set if the plugin supports API_HUB_SCHEDULE_TRIGGER mode for this action. "scheduleTimeZone": "A String", # Optional. The time zone for the schedule cron expression. If not provided, UTC will be used. + "serviceAccount": "A String", # Optional. The service account used to publish data. Note, the service account will only be accepted for non GCP plugins like OPDK. "state": "A String", # Output only. The current state of the plugin action in the plugin instance. }, ], @@ -616,6 +636,7 @@

Method Details

"displayName": "A String", # Required. The display name for this plugin instance. Max length is 255 characters. "errorMessage": "A String", # Output only. Error message describing the failure, if any, during Create, Delete or ApplyConfig operation corresponding to the plugin instance.This field will only be populated if the plugin instance is in the ERROR or FAILED state. "name": "A String", # Identifier. The unique name of the plugin instance resource. Format: `projects/{project}/locations/{location}/plugins/{plugin}/instances/{instance}` + "sourceProjectId": "A String", # Optional. The source project id of the plugin instance. This will be the id of runtime project in case of gcp based plugins and org id in case of non gcp based plugins. This is a required field. "state": "A String", # Output only. The current state of the plugin instance (e.g., enabled, disabled, provisioning). "updateTime": "A String", # Output only. Timestamp indicating when the plugin instance was last updated. }, @@ -637,4 +658,219 @@

Method Details

+
+ patch(name, body=None, updateMask=None, x__xgafv=None) +
Updates a plugin instance in the API hub. The following fields in the plugin_instance can be updated currently: * display_name * schedule_cron_expression The update_mask should be used to specify the fields being updated. To update the auth_config and additional_config of the plugin instance, use the ApplyPluginInstanceConfig method.
+
+Args:
+  name: string, Identifier. The unique name of the plugin instance resource. Format: `projects/{project}/locations/{location}/plugins/{plugin}/instances/{instance}` (required)
+  body: object, The request body.
+    The object takes the form of:
+
+{ # Represents a plugin instance resource in the API Hub. A PluginInstance is a specific instance of a hub plugin with its own configuration, state, and execution details.
+  "actions": [ # Required. The action status for the plugin instance.
+    { # PluginInstanceAction represents an action which can be executed in the plugin instance.
+      "actionId": "A String", # Required. This should map to one of the action id specified in actions_config in the plugin.
+      "curationConfig": { # The curation information for this plugin instance. # Optional. This configuration should be provided if the plugin action is publishing data to API hub curate layer.
+        "curationType": "A String", # Required. The curation type for this plugin instance.
+        "customCuration": { # Custom curation information for this plugin instance. # Optional. Custom curation information for this plugin instance.
+          "curation": "A String", # Required. The unique name of the curation resource. This will be the name of the curation resource in the format: `projects/{project}/locations/{location}/curations/{curation}`
+        },
+      },
+      "hubInstanceAction": { # The execution status for the plugin instance. # Optional. The execution information for the plugin instance action done corresponding to an API hub instance.
+        "currentExecutionState": "A String", # Output only. The current state of the execution.
+        "lastExecution": { # The result of the last execution of the plugin instance. # Output only. The last execution of the plugin instance.
+          "endTime": "A String", # Output only. The last execution end time of the plugin instance.
+          "errorMessage": "A String", # Output only. Error message describing the failure, if any, during the last execution.
+          "result": "A String", # Output only. The result of the last execution of the plugin instance.
+          "startTime": "A String", # Output only. The last execution start time of the plugin instance.
+        },
+      },
+      "resourceConfig": { # The configuration of resources created for a given plugin instance action. # Output only. The configuration of resources created for a given plugin instance action. Note these will be returned only in case of Non-GCP plugins like OPDK.
+        "actionType": "A String", # Output only. The type of the action.
+        "pubsubTopic": "A String", # Output only. The pubsub topic to publish the data to. Format is projects/{project}/topics/{topic}
+      },
+      "scheduleCronExpression": "A String", # Optional. The schedule for this plugin instance action. This can only be set if the plugin supports API_HUB_SCHEDULE_TRIGGER mode for this action.
+      "scheduleTimeZone": "A String", # Optional. The time zone for the schedule cron expression. If not provided, UTC will be used.
+      "serviceAccount": "A String", # Optional. The service account used to publish data. Note, the service account will only be accepted for non GCP plugins like OPDK.
+      "state": "A String", # Output only. The current state of the plugin action in the plugin instance.
+    },
+  ],
+  "additionalConfig": { # Optional. The additional information for this plugin instance corresponding to the additional config template of the plugin. This information will be sent to plugin hosting service on each call to plugin hosted service. The key will be the config_variable_template.display_name to uniquely identify the config variable.
+    "a_key": { # ConfigVariable represents a additional configuration variable present in a PluginInstance Config or AuthConfig, based on a ConfigVariableTemplate.
+      "boolValue": True or False, # Optional. The config variable value in case of config variable of type boolean.
+      "enumValue": { # ConfigValueOption represents an option for a config variable of type enum or multi select. # Optional. The config variable value in case of config variable of type enum.
+        "description": "A String", # Optional. Description of the option.
+        "displayName": "A String", # Required. Display name of the option.
+        "id": "A String", # Required. Id of the option.
+      },
+      "intValue": "A String", # Optional. The config variable value in case of config variable of type integer.
+      "key": "A String", # Output only. Key will be the id to uniquely identify the config variable.
+      "multiIntValues": { # The config variable value of data type multi int. # Optional. The config variable value in case of config variable of type multi integer.
+        "values": [ # Optional. The config variable value of data type multi int.
+          42,
+        ],
+      },
+      "multiSelectValues": { # The config variable value of data type multi select. # Optional. The config variable value in case of config variable of type multi select.
+        "values": [ # Optional. The config variable value of data type multi select.
+          { # ConfigValueOption represents an option for a config variable of type enum or multi select.
+            "description": "A String", # Optional. Description of the option.
+            "displayName": "A String", # Required. Display name of the option.
+            "id": "A String", # Required. Id of the option.
+          },
+        ],
+      },
+      "multiStringValues": { # The config variable value of data type multi string. # Optional. The config variable value in case of config variable of type multi string.
+        "values": [ # Optional. The config variable value of data type multi string.
+          "A String",
+        ],
+      },
+      "secretValue": { # Secret provides a reference to entries in Secret Manager. # Optional. The config variable value in case of config variable of type secret.
+        "secretVersion": "A String", # Required. The resource name of the secret version in the format, format as: `projects/*/secrets/*/versions/*`.
+      },
+      "stringValue": "A String", # Optional. The config variable value in case of config variable of type string.
+    },
+  },
+  "authConfig": { # AuthConfig represents the authentication information. # Optional. The authentication information for this plugin instance.
+    "apiKeyConfig": { # Config for authentication with API key. # Api Key Config.
+      "apiKey": { # Secret provides a reference to entries in Secret Manager. # Required. The name of the SecretManager secret version resource storing the API key. Format: `projects/{project}/secrets/{secrete}/versions/{version}`. The `secretmanager.versions.access` permission should be granted to the service account accessing the secret.
+        "secretVersion": "A String", # Required. The resource name of the secret version in the format, format as: `projects/*/secrets/*/versions/*`.
+      },
+      "httpElementLocation": "A String", # Required. The location of the API key. The default value is QUERY.
+      "name": "A String", # Required. The parameter name of the API key. E.g. If the API request is "https://example.com/act?api_key=", "api_key" would be the parameter name.
+    },
+    "authType": "A String", # Required. The authentication type.
+    "googleServiceAccountConfig": { # Config for Google service account authentication. # Google Service Account.
+      "serviceAccount": "A String", # Required. The service account to be used for authenticating request. The `iam.serviceAccounts.getAccessToken` permission should be granted on this service account to the impersonator service account.
+    },
+    "oauth2ClientCredentialsConfig": { # Parameters to support Oauth 2.0 client credentials grant authentication. See https://tools.ietf.org/html/rfc6749#section-1.3.4 for more details. # Oauth2.0 Client Credentials.
+      "clientId": "A String", # Required. The client identifier.
+      "clientSecret": { # Secret provides a reference to entries in Secret Manager. # Required. Secret version reference containing the client secret. The `secretmanager.versions.access` permission should be granted to the service account accessing the secret.
+        "secretVersion": "A String", # Required. The resource name of the secret version in the format, format as: `projects/*/secrets/*/versions/*`.
+      },
+    },
+    "userPasswordConfig": { # Parameters to support Username and Password Authentication. # User Password.
+      "password": { # Secret provides a reference to entries in Secret Manager. # Required. Secret version reference containing the password. The `secretmanager.versions.access` permission should be granted to the service account accessing the secret.
+        "secretVersion": "A String", # Required. The resource name of the secret version in the format, format as: `projects/*/secrets/*/versions/*`.
+      },
+      "username": "A String", # Required. Username.
+    },
+  },
+  "createTime": "A String", # Output only. Timestamp indicating when the plugin instance was created.
+  "displayName": "A String", # Required. The display name for this plugin instance. Max length is 255 characters.
+  "errorMessage": "A String", # Output only. Error message describing the failure, if any, during Create, Delete or ApplyConfig operation corresponding to the plugin instance.This field will only be populated if the plugin instance is in the ERROR or FAILED state.
+  "name": "A String", # Identifier. The unique name of the plugin instance resource. Format: `projects/{project}/locations/{location}/plugins/{plugin}/instances/{instance}`
+  "sourceProjectId": "A String", # Optional. The source project id of the plugin instance. This will be the id of runtime project in case of gcp based plugins and org id in case of non gcp based plugins. This is a required field.
+  "state": "A String", # Output only. The current state of the plugin instance (e.g., enabled, disabled, provisioning).
+  "updateTime": "A String", # Output only. Timestamp indicating when the plugin instance was last updated.
+}
+
+  updateMask: string, Optional. The list of fields to update.
+  x__xgafv: string, V1 error format.
+    Allowed values
+      1 - v1 error format
+      2 - v2 error format
+
+Returns:
+  An object of the form:
+
+    { # Represents a plugin instance resource in the API Hub. A PluginInstance is a specific instance of a hub plugin with its own configuration, state, and execution details.
+  "actions": [ # Required. The action status for the plugin instance.
+    { # PluginInstanceAction represents an action which can be executed in the plugin instance.
+      "actionId": "A String", # Required. This should map to one of the action id specified in actions_config in the plugin.
+      "curationConfig": { # The curation information for this plugin instance. # Optional. This configuration should be provided if the plugin action is publishing data to API hub curate layer.
+        "curationType": "A String", # Required. The curation type for this plugin instance.
+        "customCuration": { # Custom curation information for this plugin instance. # Optional. Custom curation information for this plugin instance.
+          "curation": "A String", # Required. The unique name of the curation resource. This will be the name of the curation resource in the format: `projects/{project}/locations/{location}/curations/{curation}`
+        },
+      },
+      "hubInstanceAction": { # The execution status for the plugin instance. # Optional. The execution information for the plugin instance action done corresponding to an API hub instance.
+        "currentExecutionState": "A String", # Output only. The current state of the execution.
+        "lastExecution": { # The result of the last execution of the plugin instance. # Output only. The last execution of the plugin instance.
+          "endTime": "A String", # Output only. The last execution end time of the plugin instance.
+          "errorMessage": "A String", # Output only. Error message describing the failure, if any, during the last execution.
+          "result": "A String", # Output only. The result of the last execution of the plugin instance.
+          "startTime": "A String", # Output only. The last execution start time of the plugin instance.
+        },
+      },
+      "resourceConfig": { # The configuration of resources created for a given plugin instance action. # Output only. The configuration of resources created for a given plugin instance action. Note these will be returned only in case of Non-GCP plugins like OPDK.
+        "actionType": "A String", # Output only. The type of the action.
+        "pubsubTopic": "A String", # Output only. The pubsub topic to publish the data to. Format is projects/{project}/topics/{topic}
+      },
+      "scheduleCronExpression": "A String", # Optional. The schedule for this plugin instance action. This can only be set if the plugin supports API_HUB_SCHEDULE_TRIGGER mode for this action.
+      "scheduleTimeZone": "A String", # Optional. The time zone for the schedule cron expression. If not provided, UTC will be used.
+      "serviceAccount": "A String", # Optional. The service account used to publish data. Note, the service account will only be accepted for non GCP plugins like OPDK.
+      "state": "A String", # Output only. The current state of the plugin action in the plugin instance.
+    },
+  ],
+  "additionalConfig": { # Optional. The additional information for this plugin instance corresponding to the additional config template of the plugin. This information will be sent to plugin hosting service on each call to plugin hosted service. The key will be the config_variable_template.display_name to uniquely identify the config variable.
+    "a_key": { # ConfigVariable represents a additional configuration variable present in a PluginInstance Config or AuthConfig, based on a ConfigVariableTemplate.
+      "boolValue": True or False, # Optional. The config variable value in case of config variable of type boolean.
+      "enumValue": { # ConfigValueOption represents an option for a config variable of type enum or multi select. # Optional. The config variable value in case of config variable of type enum.
+        "description": "A String", # Optional. Description of the option.
+        "displayName": "A String", # Required. Display name of the option.
+        "id": "A String", # Required. Id of the option.
+      },
+      "intValue": "A String", # Optional. The config variable value in case of config variable of type integer.
+      "key": "A String", # Output only. Key will be the id to uniquely identify the config variable.
+      "multiIntValues": { # The config variable value of data type multi int. # Optional. The config variable value in case of config variable of type multi integer.
+        "values": [ # Optional. The config variable value of data type multi int.
+          42,
+        ],
+      },
+      "multiSelectValues": { # The config variable value of data type multi select. # Optional. The config variable value in case of config variable of type multi select.
+        "values": [ # Optional. The config variable value of data type multi select.
+          { # ConfigValueOption represents an option for a config variable of type enum or multi select.
+            "description": "A String", # Optional. Description of the option.
+            "displayName": "A String", # Required. Display name of the option.
+            "id": "A String", # Required. Id of the option.
+          },
+        ],
+      },
+      "multiStringValues": { # The config variable value of data type multi string. # Optional. The config variable value in case of config variable of type multi string.
+        "values": [ # Optional. The config variable value of data type multi string.
+          "A String",
+        ],
+      },
+      "secretValue": { # Secret provides a reference to entries in Secret Manager. # Optional. The config variable value in case of config variable of type secret.
+        "secretVersion": "A String", # Required. The resource name of the secret version in the format, format as: `projects/*/secrets/*/versions/*`.
+      },
+      "stringValue": "A String", # Optional. The config variable value in case of config variable of type string.
+    },
+  },
+  "authConfig": { # AuthConfig represents the authentication information. # Optional. The authentication information for this plugin instance.
+    "apiKeyConfig": { # Config for authentication with API key. # Api Key Config.
+      "apiKey": { # Secret provides a reference to entries in Secret Manager. # Required. The name of the SecretManager secret version resource storing the API key. Format: `projects/{project}/secrets/{secrete}/versions/{version}`. The `secretmanager.versions.access` permission should be granted to the service account accessing the secret.
+        "secretVersion": "A String", # Required. The resource name of the secret version in the format, format as: `projects/*/secrets/*/versions/*`.
+      },
+      "httpElementLocation": "A String", # Required. The location of the API key. The default value is QUERY.
+      "name": "A String", # Required. The parameter name of the API key. E.g. If the API request is "https://example.com/act?api_key=", "api_key" would be the parameter name.
+    },
+    "authType": "A String", # Required. The authentication type.
+    "googleServiceAccountConfig": { # Config for Google service account authentication. # Google Service Account.
+      "serviceAccount": "A String", # Required. The service account to be used for authenticating request. The `iam.serviceAccounts.getAccessToken` permission should be granted on this service account to the impersonator service account.
+    },
+    "oauth2ClientCredentialsConfig": { # Parameters to support Oauth 2.0 client credentials grant authentication. See https://tools.ietf.org/html/rfc6749#section-1.3.4 for more details. # Oauth2.0 Client Credentials.
+      "clientId": "A String", # Required. The client identifier.
+      "clientSecret": { # Secret provides a reference to entries in Secret Manager. # Required. Secret version reference containing the client secret. The `secretmanager.versions.access` permission should be granted to the service account accessing the secret.
+        "secretVersion": "A String", # Required. The resource name of the secret version in the format, format as: `projects/*/secrets/*/versions/*`.
+      },
+    },
+    "userPasswordConfig": { # Parameters to support Username and Password Authentication. # User Password.
+      "password": { # Secret provides a reference to entries in Secret Manager. # Required. Secret version reference containing the password. The `secretmanager.versions.access` permission should be granted to the service account accessing the secret.
+        "secretVersion": "A String", # Required. The resource name of the secret version in the format, format as: `projects/*/secrets/*/versions/*`.
+      },
+      "username": "A String", # Required. Username.
+    },
+  },
+  "createTime": "A String", # Output only. Timestamp indicating when the plugin instance was created.
+  "displayName": "A String", # Required. The display name for this plugin instance. Max length is 255 characters.
+  "errorMessage": "A String", # Output only. Error message describing the failure, if any, during Create, Delete or ApplyConfig operation corresponding to the plugin instance.This field will only be populated if the plugin instance is in the ERROR or FAILED state.
+  "name": "A String", # Identifier. The unique name of the plugin instance resource. Format: `projects/{project}/locations/{location}/plugins/{plugin}/instances/{instance}`
+  "sourceProjectId": "A String", # Optional. The source project id of the plugin instance. This will be the id of runtime project in case of gcp based plugins and org id in case of non gcp based plugins. This is a required field.
+  "state": "A String", # Output only. The current state of the plugin instance (e.g., enabled, disabled, provisioning).
+  "updateTime": "A String", # Output only. Timestamp indicating when the plugin instance was last updated.
+}
+
+ \ No newline at end of file diff --git a/docs/dyn/backupdr_v1.projects.locations.backupPlanAssociations.html b/docs/dyn/backupdr_v1.projects.locations.backupPlanAssociations.html index 594ce5d90c..61c27feb70 100644 --- a/docs/dyn/backupdr_v1.projects.locations.backupPlanAssociations.html +++ b/docs/dyn/backupdr_v1.projects.locations.backupPlanAssociations.html @@ -83,6 +83,12 @@

Instance Methods

delete(name, requestId=None, x__xgafv=None)

Deletes a single BackupPlanAssociation.

+

+ fetchForResourceType(parent, filter=None, orderBy=None, pageSize=None, pageToken=None, resourceType=None, x__xgafv=None)

+

List BackupPlanAssociations for a given resource type.

+

+ fetchForResourceType_next()

+

Retrieves the next page of results.

get(name, x__xgafv=None)

Gets details of a single BackupPlanAssociation.

@@ -117,6 +123,9 @@

Method Details

"backupPlan": "A String", # Required. Resource name of backup plan which needs to be applied on workload. Format: projects/{project}/locations/{location}/backupPlans/{backupPlanId} "backupPlanRevisionId": "A String", # Output only. The user friendly revision ID of the `BackupPlanRevision`. Example: v0, v1, v2, etc. "backupPlanRevisionName": "A String", # Output only. The resource id of the `BackupPlanRevision`. Format: `projects/{project}/locations/{location}/backupPlans/{backup_plan}/revisions/{revision_id}` + "cloudSqlInstanceBackupPlanAssociationProperties": { # Cloud SQL instance's BPA properties. # Output only. Cloud SQL instance's backup plan association properties. + "instanceCreateTime": "A String", # Output only. The time when the instance was created. + }, "createTime": "A String", # Output only. The time when the instance was created. "dataSource": "A String", # Output only. Resource name of data source which will be used as storage location for backups taken. Format : projects/{project}/locations/{location}/backupVaults/{backupvault}/dataSources/{datasource} "name": "A String", # Output only. Identifier. The resource name of BackupPlanAssociation in below format Format : projects/{project}/locations/{location}/backupPlanAssociations/{backupPlanAssociationId} @@ -209,6 +218,77 @@

Method Details

} +
+ fetchForResourceType(parent, filter=None, orderBy=None, pageSize=None, pageToken=None, resourceType=None, x__xgafv=None) +
List BackupPlanAssociations for a given resource type.
+
+Args:
+  parent: string, Required. The parent resource name. Format: projects/{project}/locations/{location} (required)
+  filter: string, Optional. A filter expression that filters the results fetched in the response. The expression must specify the field name, a comparison operator, and the value that you want to use for filtering. Supported fields: * resource * backup_plan * state * data_source * cloud_sql_instance_backup_plan_association_properties.instance_create_time
+  orderBy: string, Optional. A comma-separated list of fields to order by, sorted in ascending order. Use "desc" after a field name for descending. Supported fields: * name
+  pageSize: integer, Optional. The maximum number of BackupPlanAssociations to return. The service may return fewer than this value. If unspecified, at most 50 BackupPlanAssociations will be returned. The maximum value is 100; values above 100 will be coerced to 100.
+  pageToken: string, Optional. A page token, received from a previous call of `FetchBackupPlanAssociationsForResourceType`. Provide this to retrieve the subsequent page. When paginating, all other parameters provided to `FetchBackupPlanAssociationsForResourceType` must match the call that provided the page token.
+  resourceType: string, Required. The type of the GCP resource. Ex: sql.googleapis.com/Instance
+  x__xgafv: string, V1 error format.
+    Allowed values
+      1 - v1 error format
+      2 - v2 error format
+
+Returns:
+  An object of the form:
+
+    { # Response for the FetchBackupPlanAssociationsForResourceType method.
+  "backupPlanAssociations": [ # Output only. The BackupPlanAssociations from the specified parent.
+    { # A BackupPlanAssociation represents a single BackupPlanAssociation which contains details like workload, backup plan etc
+      "backupPlan": "A String", # Required. Resource name of backup plan which needs to be applied on workload. Format: projects/{project}/locations/{location}/backupPlans/{backupPlanId}
+      "backupPlanRevisionId": "A String", # Output only. The user friendly revision ID of the `BackupPlanRevision`. Example: v0, v1, v2, etc.
+      "backupPlanRevisionName": "A String", # Output only. The resource id of the `BackupPlanRevision`. Format: `projects/{project}/locations/{location}/backupPlans/{backup_plan}/revisions/{revision_id}`
+      "cloudSqlInstanceBackupPlanAssociationProperties": { # Cloud SQL instance's BPA properties. # Output only. Cloud SQL instance's backup plan association properties.
+        "instanceCreateTime": "A String", # Output only. The time when the instance was created.
+      },
+      "createTime": "A String", # Output only. The time when the instance was created.
+      "dataSource": "A String", # Output only. Resource name of data source which will be used as storage location for backups taken. Format : projects/{project}/locations/{location}/backupVaults/{backupvault}/dataSources/{datasource}
+      "name": "A String", # Output only. Identifier. The resource name of BackupPlanAssociation in below format Format : projects/{project}/locations/{location}/backupPlanAssociations/{backupPlanAssociationId}
+      "resource": "A String", # Required. Immutable. Resource name of workload on which the backup plan is applied. The format can either be the resource name (e.g., "projects/my-project/zones/us-central1-a/instances/my-instance") or the full resource URI (e.g., "https://www.googleapis.com/compute/v1/projects/my-project/zones/us-central1-a/instances/my-instance").
+      "resourceType": "A String", # Required. Immutable.
+      "rulesConfigInfo": [ # Output only. The config info related to backup rules.
+        { # Message for rules config info.
+          "lastBackupError": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # Output only. google.rpc.Status object to store the last backup error.
+            "code": 42, # The status code, which should be an enum value of google.rpc.Code.
+            "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use.
+              {
+                "a_key": "", # Properties of the object. Contains field @type with type URL.
+              },
+            ],
+            "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
+          },
+          "lastBackupState": "A String", # Output only. The last backup state for rule.
+          "lastSuccessfulBackupConsistencyTime": "A String", # Output only. The point in time when the last successful backup was captured from the source.
+          "ruleId": "A String", # Output only. Backup Rule id fetched from backup plan.
+        },
+      ],
+      "state": "A String", # Output only. The BackupPlanAssociation resource state.
+      "updateTime": "A String", # Output only. The time when the instance was updated.
+    },
+  ],
+  "nextPageToken": "A String", # Output only. A token, which can be sent as `page_token` to retrieve the next page. If this field is omitted, there are no subsequent pages.
+}
+
+ +
+ fetchForResourceType_next() +
Retrieves the next page of results.
+
+        Args:
+          previous_request: The request for the previous page. (required)
+          previous_response: The response from the request for the previous page. (required)
+
+        Returns:
+          A request object that you can call 'execute()' on to request the next
+          page. Returns None if there are no more items in the collection.
+        
+
+
get(name, x__xgafv=None)
Gets details of a single BackupPlanAssociation.
@@ -227,6 +307,9 @@ 

Method Details

"backupPlan": "A String", # Required. Resource name of backup plan which needs to be applied on workload. Format: projects/{project}/locations/{location}/backupPlans/{backupPlanId} "backupPlanRevisionId": "A String", # Output only. The user friendly revision ID of the `BackupPlanRevision`. Example: v0, v1, v2, etc. "backupPlanRevisionName": "A String", # Output only. The resource id of the `BackupPlanRevision`. Format: `projects/{project}/locations/{location}/backupPlans/{backup_plan}/revisions/{revision_id}` + "cloudSqlInstanceBackupPlanAssociationProperties": { # Cloud SQL instance's BPA properties. # Output only. Cloud SQL instance's backup plan association properties. + "instanceCreateTime": "A String", # Output only. The time when the instance was created. + }, "createTime": "A String", # Output only. The time when the instance was created. "dataSource": "A String", # Output only. Resource name of data source which will be used as storage location for backups taken. Format : projects/{project}/locations/{location}/backupVaults/{backupvault}/dataSources/{datasource} "name": "A String", # Output only. Identifier. The resource name of BackupPlanAssociation in below format Format : projects/{project}/locations/{location}/backupPlanAssociations/{backupPlanAssociationId} @@ -276,6 +359,9 @@

Method Details

"backupPlan": "A String", # Required. Resource name of backup plan which needs to be applied on workload. Format: projects/{project}/locations/{location}/backupPlans/{backupPlanId} "backupPlanRevisionId": "A String", # Output only. The user friendly revision ID of the `BackupPlanRevision`. Example: v0, v1, v2, etc. "backupPlanRevisionName": "A String", # Output only. The resource id of the `BackupPlanRevision`. Format: `projects/{project}/locations/{location}/backupPlans/{backup_plan}/revisions/{revision_id}` + "cloudSqlInstanceBackupPlanAssociationProperties": { # Cloud SQL instance's BPA properties. # Output only. Cloud SQL instance's backup plan association properties. + "instanceCreateTime": "A String", # Output only. The time when the instance was created. + }, "createTime": "A String", # Output only. The time when the instance was created. "dataSource": "A String", # Output only. Resource name of data source which will be used as storage location for backups taken. Format : projects/{project}/locations/{location}/backupVaults/{backupvault}/dataSources/{datasource} "name": "A String", # Output only. Identifier. The resource name of BackupPlanAssociation in below format Format : projects/{project}/locations/{location}/backupPlanAssociations/{backupPlanAssociationId} @@ -335,6 +421,9 @@

Method Details

"backupPlan": "A String", # Required. Resource name of backup plan which needs to be applied on workload. Format: projects/{project}/locations/{location}/backupPlans/{backupPlanId} "backupPlanRevisionId": "A String", # Output only. The user friendly revision ID of the `BackupPlanRevision`. Example: v0, v1, v2, etc. "backupPlanRevisionName": "A String", # Output only. The resource id of the `BackupPlanRevision`. Format: `projects/{project}/locations/{location}/backupPlans/{backup_plan}/revisions/{revision_id}` + "cloudSqlInstanceBackupPlanAssociationProperties": { # Cloud SQL instance's BPA properties. # Output only. Cloud SQL instance's backup plan association properties. + "instanceCreateTime": "A String", # Output only. The time when the instance was created. + }, "createTime": "A String", # Output only. The time when the instance was created. "dataSource": "A String", # Output only. Resource name of data source which will be used as storage location for backups taken. Format : projects/{project}/locations/{location}/backupVaults/{backupvault}/dataSources/{datasource} "name": "A String", # Output only. Identifier. The resource name of BackupPlanAssociation in below format Format : projects/{project}/locations/{location}/backupPlanAssociations/{backupPlanAssociationId} diff --git a/docs/dyn/backupdr_v1.projects.locations.backupPlans.html b/docs/dyn/backupdr_v1.projects.locations.backupPlans.html index f87669065a..e207331c2b 100644 --- a/docs/dyn/backupdr_v1.projects.locations.backupPlans.html +++ b/docs/dyn/backupdr_v1.projects.locations.backupPlans.html @@ -118,7 +118,7 @@

Method Details

{ # A `BackupPlan` specifies some common fields, such as `description` as well as one or more `BackupRule` messages. Each `BackupRule` has a retention policy and defines a schedule by which the system is to perform backup workloads. "backupRules": [ # Required. The backup rules for this `BackupPlan`. There must be at least one `BackupRule` message. { # `BackupRule` binds the backup schedule to a retention policy. - "backupRetentionDays": 42, # Required. Configures the duration for which backup data will be kept. It is defined in “days”. The value should be greater than or equal to minimum enforced retention of the backup vault. Minimum value is 1 and maximum value is 36159 for custom retention on-demand backup. Minimum value is 1 and maximum value is 90 for hourly backups. Minimum value is 1 and maximum value is 186 for daily backups. Minimum value is 7 and maximum value is 366 for weekly backups. Minimum value is 30 and maximum value is 732 for monthly backups. Minimum value is 365 and maximum value is 36159 for yearly backups. + "backupRetentionDays": 42, # Required. Configures the duration for which backup data will be kept. It is defined in “days”. The value should be greater than or equal to minimum enforced retention of the backup vault. Minimum value is 1 and maximum value is 36159 for custom retention on-demand backup. Minimum and maximum values are workload specific for all other rules. "ruleId": "A String", # Required. Immutable. The unique id of this `BackupRule`. The `rule_id` is unique per `BackupPlan`.The `rule_id` must start with a lowercase letter followed by up to 62 lowercase letters, numbers, or hyphens. Pattern, /a-z{,62}/. "standardSchedule": { # `StandardSchedule` defines a schedule that run within the confines of a defined window of days. We can define recurrence type for schedule as HOURLY, DAILY, WEEKLY, MONTHLY or YEARLY. # Optional. Defines a schedule that runs within the confines of a defined window of time. "backupWindow": { # `BackupWindow` defines a window of the day during which backup jobs will run. # Required. A BackupWindow defines the window of day during which backup jobs will run. Jobs are queued at the beginning of the window and will be marked as `NOT_RUN` if they do not start by the end of the window. Note: running jobs will not be cancelled at the end of the window. @@ -152,6 +152,7 @@

Method Details

"labels": { # Optional. This collection of key/value pairs allows for custom labels to be supplied by the user. Example, {"tag": "Weekly"}. "a_key": "A String", }, + "logRetentionDays": "A String", # Optional. Required for CloudSQL resource_type Configures how long logs will be stored. It is defined in “days”. This value should be greater than or equal to minimum enforced log retention duration of the backup vault. "name": "A String", # Output only. Identifier. The resource name of the `BackupPlan`. Format: `projects/{project}/locations/{location}/backupPlans/{backup_plan}` "resourceType": "A String", # Required. "revisionId": "A String", # Output only. The user friendly revision ID of the `BackupPlanRevision`. Example: v0, v1, v2, etc. @@ -247,7 +248,7 @@

Method Details

{ # A `BackupPlan` specifies some common fields, such as `description` as well as one or more `BackupRule` messages. Each `BackupRule` has a retention policy and defines a schedule by which the system is to perform backup workloads. "backupRules": [ # Required. The backup rules for this `BackupPlan`. There must be at least one `BackupRule` message. { # `BackupRule` binds the backup schedule to a retention policy. - "backupRetentionDays": 42, # Required. Configures the duration for which backup data will be kept. It is defined in “days”. The value should be greater than or equal to minimum enforced retention of the backup vault. Minimum value is 1 and maximum value is 36159 for custom retention on-demand backup. Minimum value is 1 and maximum value is 90 for hourly backups. Minimum value is 1 and maximum value is 186 for daily backups. Minimum value is 7 and maximum value is 366 for weekly backups. Minimum value is 30 and maximum value is 732 for monthly backups. Minimum value is 365 and maximum value is 36159 for yearly backups. + "backupRetentionDays": 42, # Required. Configures the duration for which backup data will be kept. It is defined in “days”. The value should be greater than or equal to minimum enforced retention of the backup vault. Minimum value is 1 and maximum value is 36159 for custom retention on-demand backup. Minimum and maximum values are workload specific for all other rules. "ruleId": "A String", # Required. Immutable. The unique id of this `BackupRule`. The `rule_id` is unique per `BackupPlan`.The `rule_id` must start with a lowercase letter followed by up to 62 lowercase letters, numbers, or hyphens. Pattern, /a-z{,62}/. "standardSchedule": { # `StandardSchedule` defines a schedule that run within the confines of a defined window of days. We can define recurrence type for schedule as HOURLY, DAILY, WEEKLY, MONTHLY or YEARLY. # Optional. Defines a schedule that runs within the confines of a defined window of time. "backupWindow": { # `BackupWindow` defines a window of the day during which backup jobs will run. # Required. A BackupWindow defines the window of day during which backup jobs will run. Jobs are queued at the beginning of the window and will be marked as `NOT_RUN` if they do not start by the end of the window. Note: running jobs will not be cancelled at the end of the window. @@ -281,6 +282,7 @@

Method Details

"labels": { # Optional. This collection of key/value pairs allows for custom labels to be supplied by the user. Example, {"tag": "Weekly"}. "a_key": "A String", }, + "logRetentionDays": "A String", # Optional. Required for CloudSQL resource_type Configures how long logs will be stored. It is defined in “days”. This value should be greater than or equal to minimum enforced log retention duration of the backup vault. "name": "A String", # Output only. Identifier. The resource name of the `BackupPlan`. Format: `projects/{project}/locations/{location}/backupPlans/{backup_plan}` "resourceType": "A String", # Required. "revisionId": "A String", # Output only. The user friendly revision ID of the `BackupPlanRevision`. Example: v0, v1, v2, etc. @@ -316,7 +318,7 @@

Method Details

{ # A `BackupPlan` specifies some common fields, such as `description` as well as one or more `BackupRule` messages. Each `BackupRule` has a retention policy and defines a schedule by which the system is to perform backup workloads. "backupRules": [ # Required. The backup rules for this `BackupPlan`. There must be at least one `BackupRule` message. { # `BackupRule` binds the backup schedule to a retention policy. - "backupRetentionDays": 42, # Required. Configures the duration for which backup data will be kept. It is defined in “days”. The value should be greater than or equal to minimum enforced retention of the backup vault. Minimum value is 1 and maximum value is 36159 for custom retention on-demand backup. Minimum value is 1 and maximum value is 90 for hourly backups. Minimum value is 1 and maximum value is 186 for daily backups. Minimum value is 7 and maximum value is 366 for weekly backups. Minimum value is 30 and maximum value is 732 for monthly backups. Minimum value is 365 and maximum value is 36159 for yearly backups. + "backupRetentionDays": 42, # Required. Configures the duration for which backup data will be kept. It is defined in “days”. The value should be greater than or equal to minimum enforced retention of the backup vault. Minimum value is 1 and maximum value is 36159 for custom retention on-demand backup. Minimum and maximum values are workload specific for all other rules. "ruleId": "A String", # Required. Immutable. The unique id of this `BackupRule`. The `rule_id` is unique per `BackupPlan`.The `rule_id` must start with a lowercase letter followed by up to 62 lowercase letters, numbers, or hyphens. Pattern, /a-z{,62}/. "standardSchedule": { # `StandardSchedule` defines a schedule that run within the confines of a defined window of days. We can define recurrence type for schedule as HOURLY, DAILY, WEEKLY, MONTHLY or YEARLY. # Optional. Defines a schedule that runs within the confines of a defined window of time. "backupWindow": { # `BackupWindow` defines a window of the day during which backup jobs will run. # Required. A BackupWindow defines the window of day during which backup jobs will run. Jobs are queued at the beginning of the window and will be marked as `NOT_RUN` if they do not start by the end of the window. Note: running jobs will not be cancelled at the end of the window. @@ -350,6 +352,7 @@

Method Details

"labels": { # Optional. This collection of key/value pairs allows for custom labels to be supplied by the user. Example, {"tag": "Weekly"}. "a_key": "A String", }, + "logRetentionDays": "A String", # Optional. Required for CloudSQL resource_type Configures how long logs will be stored. It is defined in “days”. This value should be greater than or equal to minimum enforced log retention duration of the backup vault. "name": "A String", # Output only. Identifier. The resource name of the `BackupPlan`. Format: `projects/{project}/locations/{location}/backupPlans/{backup_plan}` "resourceType": "A String", # Required. "revisionId": "A String", # Output only. The user friendly revision ID of the `BackupPlanRevision`. Example: v0, v1, v2, etc. @@ -394,7 +397,7 @@

Method Details

{ # A `BackupPlan` specifies some common fields, such as `description` as well as one or more `BackupRule` messages. Each `BackupRule` has a retention policy and defines a schedule by which the system is to perform backup workloads. "backupRules": [ # Required. The backup rules for this `BackupPlan`. There must be at least one `BackupRule` message. { # `BackupRule` binds the backup schedule to a retention policy. - "backupRetentionDays": 42, # Required. Configures the duration for which backup data will be kept. It is defined in “days”. The value should be greater than or equal to minimum enforced retention of the backup vault. Minimum value is 1 and maximum value is 36159 for custom retention on-demand backup. Minimum value is 1 and maximum value is 90 for hourly backups. Minimum value is 1 and maximum value is 186 for daily backups. Minimum value is 7 and maximum value is 366 for weekly backups. Minimum value is 30 and maximum value is 732 for monthly backups. Minimum value is 365 and maximum value is 36159 for yearly backups. + "backupRetentionDays": 42, # Required. Configures the duration for which backup data will be kept. It is defined in “days”. The value should be greater than or equal to minimum enforced retention of the backup vault. Minimum value is 1 and maximum value is 36159 for custom retention on-demand backup. Minimum and maximum values are workload specific for all other rules. "ruleId": "A String", # Required. Immutable. The unique id of this `BackupRule`. The `rule_id` is unique per `BackupPlan`.The `rule_id` must start with a lowercase letter followed by up to 62 lowercase letters, numbers, or hyphens. Pattern, /a-z{,62}/. "standardSchedule": { # `StandardSchedule` defines a schedule that run within the confines of a defined window of days. We can define recurrence type for schedule as HOURLY, DAILY, WEEKLY, MONTHLY or YEARLY. # Optional. Defines a schedule that runs within the confines of a defined window of time. "backupWindow": { # `BackupWindow` defines a window of the day during which backup jobs will run. # Required. A BackupWindow defines the window of day during which backup jobs will run. Jobs are queued at the beginning of the window and will be marked as `NOT_RUN` if they do not start by the end of the window. Note: running jobs will not be cancelled at the end of the window. @@ -428,6 +431,7 @@

Method Details

"labels": { # Optional. This collection of key/value pairs allows for custom labels to be supplied by the user. Example, {"tag": "Weekly"}. "a_key": "A String", }, + "logRetentionDays": "A String", # Optional. Required for CloudSQL resource_type Configures how long logs will be stored. It is defined in “days”. This value should be greater than or equal to minimum enforced log retention duration of the backup vault. "name": "A String", # Output only. Identifier. The resource name of the `BackupPlan`. Format: `projects/{project}/locations/{location}/backupPlans/{backup_plan}` "resourceType": "A String", # Required. "revisionId": "A String", # Output only. The user friendly revision ID of the `BackupPlanRevision`. Example: v0, v1, v2, etc. diff --git a/docs/dyn/backupdr_v1.projects.locations.backupPlans.revisions.html b/docs/dyn/backupdr_v1.projects.locations.backupPlans.revisions.html index d2743901ab..2f54f2990b 100644 --- a/docs/dyn/backupdr_v1.projects.locations.backupPlans.revisions.html +++ b/docs/dyn/backupdr_v1.projects.locations.backupPlans.revisions.html @@ -110,7 +110,7 @@

Method Details

"backupPlanSnapshot": { # A `BackupPlan` specifies some common fields, such as `description` as well as one or more `BackupRule` messages. Each `BackupRule` has a retention policy and defines a schedule by which the system is to perform backup workloads. # The Backup Plan being encompassed by this revision. "backupRules": [ # Required. The backup rules for this `BackupPlan`. There must be at least one `BackupRule` message. { # `BackupRule` binds the backup schedule to a retention policy. - "backupRetentionDays": 42, # Required. Configures the duration for which backup data will be kept. It is defined in “days”. The value should be greater than or equal to minimum enforced retention of the backup vault. Minimum value is 1 and maximum value is 36159 for custom retention on-demand backup. Minimum value is 1 and maximum value is 90 for hourly backups. Minimum value is 1 and maximum value is 186 for daily backups. Minimum value is 7 and maximum value is 366 for weekly backups. Minimum value is 30 and maximum value is 732 for monthly backups. Minimum value is 365 and maximum value is 36159 for yearly backups. + "backupRetentionDays": 42, # Required. Configures the duration for which backup data will be kept. It is defined in “days”. The value should be greater than or equal to minimum enforced retention of the backup vault. Minimum value is 1 and maximum value is 36159 for custom retention on-demand backup. Minimum and maximum values are workload specific for all other rules. "ruleId": "A String", # Required. Immutable. The unique id of this `BackupRule`. The `rule_id` is unique per `BackupPlan`.The `rule_id` must start with a lowercase letter followed by up to 62 lowercase letters, numbers, or hyphens. Pattern, /a-z{,62}/. "standardSchedule": { # `StandardSchedule` defines a schedule that run within the confines of a defined window of days. We can define recurrence type for schedule as HOURLY, DAILY, WEEKLY, MONTHLY or YEARLY. # Optional. Defines a schedule that runs within the confines of a defined window of time. "backupWindow": { # `BackupWindow` defines a window of the day during which backup jobs will run. # Required. A BackupWindow defines the window of day during which backup jobs will run. Jobs are queued at the beginning of the window and will be marked as `NOT_RUN` if they do not start by the end of the window. Note: running jobs will not be cancelled at the end of the window. @@ -144,6 +144,7 @@

Method Details

"labels": { # Optional. This collection of key/value pairs allows for custom labels to be supplied by the user. Example, {"tag": "Weekly"}. "a_key": "A String", }, + "logRetentionDays": "A String", # Optional. Required for CloudSQL resource_type Configures how long logs will be stored. It is defined in “days”. This value should be greater than or equal to minimum enforced log retention duration of the backup vault. "name": "A String", # Output only. Identifier. The resource name of the `BackupPlan`. Format: `projects/{project}/locations/{location}/backupPlans/{backup_plan}` "resourceType": "A String", # Required. "revisionId": "A String", # Output only. The user friendly revision ID of the `BackupPlanRevision`. Example: v0, v1, v2, etc. @@ -183,7 +184,7 @@

Method Details

"backupPlanSnapshot": { # A `BackupPlan` specifies some common fields, such as `description` as well as one or more `BackupRule` messages. Each `BackupRule` has a retention policy and defines a schedule by which the system is to perform backup workloads. # The Backup Plan being encompassed by this revision. "backupRules": [ # Required. The backup rules for this `BackupPlan`. There must be at least one `BackupRule` message. { # `BackupRule` binds the backup schedule to a retention policy. - "backupRetentionDays": 42, # Required. Configures the duration for which backup data will be kept. It is defined in “days”. The value should be greater than or equal to minimum enforced retention of the backup vault. Minimum value is 1 and maximum value is 36159 for custom retention on-demand backup. Minimum value is 1 and maximum value is 90 for hourly backups. Minimum value is 1 and maximum value is 186 for daily backups. Minimum value is 7 and maximum value is 366 for weekly backups. Minimum value is 30 and maximum value is 732 for monthly backups. Minimum value is 365 and maximum value is 36159 for yearly backups. + "backupRetentionDays": 42, # Required. Configures the duration for which backup data will be kept. It is defined in “days”. The value should be greater than or equal to minimum enforced retention of the backup vault. Minimum value is 1 and maximum value is 36159 for custom retention on-demand backup. Minimum and maximum values are workload specific for all other rules. "ruleId": "A String", # Required. Immutable. The unique id of this `BackupRule`. The `rule_id` is unique per `BackupPlan`.The `rule_id` must start with a lowercase letter followed by up to 62 lowercase letters, numbers, or hyphens. Pattern, /a-z{,62}/. "standardSchedule": { # `StandardSchedule` defines a schedule that run within the confines of a defined window of days. We can define recurrence type for schedule as HOURLY, DAILY, WEEKLY, MONTHLY or YEARLY. # Optional. Defines a schedule that runs within the confines of a defined window of time. "backupWindow": { # `BackupWindow` defines a window of the day during which backup jobs will run. # Required. A BackupWindow defines the window of day during which backup jobs will run. Jobs are queued at the beginning of the window and will be marked as `NOT_RUN` if they do not start by the end of the window. Note: running jobs will not be cancelled at the end of the window. @@ -217,6 +218,7 @@

Method Details

"labels": { # Optional. This collection of key/value pairs allows for custom labels to be supplied by the user. Example, {"tag": "Weekly"}. "a_key": "A String", }, + "logRetentionDays": "A String", # Optional. Required for CloudSQL resource_type Configures how long logs will be stored. It is defined in “days”. This value should be greater than or equal to minimum enforced log retention duration of the backup vault. "name": "A String", # Output only. Identifier. The resource name of the `BackupPlan`. Format: `projects/{project}/locations/{location}/backupPlans/{backup_plan}` "resourceType": "A String", # Required. "revisionId": "A String", # Output only. The user friendly revision ID of the `BackupPlanRevision`. Example: v0, v1, v2, etc. diff --git a/docs/dyn/backupdr_v1.projects.locations.backupVaults.dataSources.backups.html b/docs/dyn/backupdr_v1.projects.locations.backupVaults.dataSources.backups.html index 1b2001b72e..f880426531 100644 --- a/docs/dyn/backupdr_v1.projects.locations.backupVaults.dataSources.backups.html +++ b/docs/dyn/backupdr_v1.projects.locations.backupVaults.dataSources.backups.html @@ -180,6 +180,12 @@

Method Details

}, ], "backupType": "A String", # Output only. Type of the backup, unspecified, scheduled or ondemand. + "cloudSqlInstanceBackupProperties": { # CloudSqlInstanceBackupProperties represents Cloud SQL Instance Backup properties. . # Output only. Cloud SQL specific backup properties. + "databaseInstalledVersion": "A String", # Output only. The installed database version of the Cloud SQL instance when the backup was taken. + "finalBackup": True or False, # Output only. Whether the backup is a final backup. + "instanceTier": "A String", # Output only. The tier (or machine type) for this instance. Example: `db-custom-1-3840` + "sourceInstance": "A String", # Output only. The source instance of the backup. Format: projects/{project}/instances/{instance} + }, "computeInstanceBackupProperties": { # ComputeInstanceBackupProperties represents Compute Engine instance backup properties. # Output only. Compute Engine specific backup properties. "canIpForward": True or False, # Enables instances created based on these properties to send packets with source IP addresses other than their own and receive packets with destination IP addresses other than their own. If these instances will be used as an IP gateway or it will be set as the next-hop in a Route resource, specify `true`. If unsure, leave this set to `false`. See the https://cloud.google.com/vpc/docs/using-routes#canipforward documentation for more information. "description": "A String", # An optional text description for the instances that are created from these properties. @@ -430,6 +436,12 @@

Method Details

}, ], "backupType": "A String", # Output only. Type of the backup, unspecified, scheduled or ondemand. + "cloudSqlInstanceBackupProperties": { # CloudSqlInstanceBackupProperties represents Cloud SQL Instance Backup properties. . # Output only. Cloud SQL specific backup properties. + "databaseInstalledVersion": "A String", # Output only. The installed database version of the Cloud SQL instance when the backup was taken. + "finalBackup": True or False, # Output only. Whether the backup is a final backup. + "instanceTier": "A String", # Output only. The tier (or machine type) for this instance. Example: `db-custom-1-3840` + "sourceInstance": "A String", # Output only. The source instance of the backup. Format: projects/{project}/instances/{instance} + }, "computeInstanceBackupProperties": { # ComputeInstanceBackupProperties represents Compute Engine instance backup properties. # Output only. Compute Engine specific backup properties. "canIpForward": True or False, # Enables instances created based on these properties to send packets with source IP addresses other than their own and receive packets with destination IP addresses other than their own. If these instances will be used as an IP gateway or it will be set as the next-hop in a Route resource, specify `true`. If unsure, leave this set to `false`. See the https://cloud.google.com/vpc/docs/using-routes#canipforward documentation for more information. "description": "A String", # An optional text description for the instances that are created from these properties. @@ -684,6 +696,12 @@

Method Details

}, ], "backupType": "A String", # Output only. Type of the backup, unspecified, scheduled or ondemand. + "cloudSqlInstanceBackupProperties": { # CloudSqlInstanceBackupProperties represents Cloud SQL Instance Backup properties. . # Output only. Cloud SQL specific backup properties. + "databaseInstalledVersion": "A String", # Output only. The installed database version of the Cloud SQL instance when the backup was taken. + "finalBackup": True or False, # Output only. Whether the backup is a final backup. + "instanceTier": "A String", # Output only. The tier (or machine type) for this instance. Example: `db-custom-1-3840` + "sourceInstance": "A String", # Output only. The source instance of the backup. Format: projects/{project}/instances/{instance} + }, "computeInstanceBackupProperties": { # ComputeInstanceBackupProperties represents Compute Engine instance backup properties. # Output only. Compute Engine specific backup properties. "canIpForward": True or False, # Enables instances created based on these properties to send packets with source IP addresses other than their own and receive packets with destination IP addresses other than their own. If these instances will be used as an IP gateway or it will be set as the next-hop in a Route resource, specify `true`. If unsure, leave this set to `false`. See the https://cloud.google.com/vpc/docs/using-routes#canipforward documentation for more information. "description": "A String", # An optional text description for the instances that are created from these properties. diff --git a/docs/dyn/backupdr_v1.projects.locations.backupVaults.dataSources.html b/docs/dyn/backupdr_v1.projects.locations.backupVaults.dataSources.html index cbb7b833d0..8ed187233f 100644 --- a/docs/dyn/backupdr_v1.projects.locations.backupVaults.dataSources.html +++ b/docs/dyn/backupdr_v1.projects.locations.backupVaults.dataSources.html @@ -299,6 +299,12 @@

Method Details

"type": "A String", # The type of the application. e.g. VMBackup }, "dataSourceGcpResource": { # DataSourceGcpResource is used for protected resources that are Google Cloud Resources. This name is easeier to understand than GcpResourceDataSource or GcpDataSourceResource # The backed up resource is a Google Cloud resource. The word 'DataSource' was included in the names to indicate that this is the representation of the Google Cloud resource used within the DataSource object. + "cloudSqlInstanceDatasourceProperties": { # CloudSqlInstanceDataSourceProperties represents the properties of a Cloud SQL resource that are stored in the DataSource. . # Output only. CloudSqlInstanceDataSourceProperties has a subset of Cloud SQL Instance properties that are useful at the Datasource level. + "databaseInstalledVersion": "A String", # Output only. The installed database version of the Cloud SQL instance. + "instanceCreateTime": "A String", # Output only. The instance creation timestamp. + "instanceTier": "A String", # Output only. The tier (or machine type) for this instance. Example: `db-custom-1-3840` + "name": "A String", # Output only. Name of the Cloud SQL instance backed up by the datasource. Format: projects/{project}/instances/{instance} + }, "computeInstanceDatasourceProperties": { # ComputeInstanceDataSourceProperties represents the properties of a ComputeEngine resource that are stored in the DataSource. # ComputeInstanceDataSourceProperties has a subset of Compute Instance properties that are useful at the Datasource level. "description": "A String", # The description of the Compute Engine instance. "machineType": "A String", # The machine type of the instance. @@ -424,6 +430,12 @@

Method Details

"type": "A String", # The type of the application. e.g. VMBackup }, "dataSourceGcpResource": { # DataSourceGcpResource is used for protected resources that are Google Cloud Resources. This name is easeier to understand than GcpResourceDataSource or GcpDataSourceResource # The backed up resource is a Google Cloud resource. The word 'DataSource' was included in the names to indicate that this is the representation of the Google Cloud resource used within the DataSource object. + "cloudSqlInstanceDatasourceProperties": { # CloudSqlInstanceDataSourceProperties represents the properties of a Cloud SQL resource that are stored in the DataSource. . # Output only. CloudSqlInstanceDataSourceProperties has a subset of Cloud SQL Instance properties that are useful at the Datasource level. + "databaseInstalledVersion": "A String", # Output only. The installed database version of the Cloud SQL instance. + "instanceCreateTime": "A String", # Output only. The instance creation timestamp. + "instanceTier": "A String", # Output only. The tier (or machine type) for this instance. Example: `db-custom-1-3840` + "name": "A String", # Output only. Name of the Cloud SQL instance backed up by the datasource. Format: projects/{project}/instances/{instance} + }, "computeInstanceDatasourceProperties": { # ComputeInstanceDataSourceProperties represents the properties of a ComputeEngine resource that are stored in the DataSource. # ComputeInstanceDataSourceProperties has a subset of Compute Instance properties that are useful at the Datasource level. "description": "A String", # The description of the Compute Engine instance. "machineType": "A String", # The machine type of the instance. @@ -529,6 +541,12 @@

Method Details

"type": "A String", # The type of the application. e.g. VMBackup }, "dataSourceGcpResource": { # DataSourceGcpResource is used for protected resources that are Google Cloud Resources. This name is easeier to understand than GcpResourceDataSource or GcpDataSourceResource # The backed up resource is a Google Cloud resource. The word 'DataSource' was included in the names to indicate that this is the representation of the Google Cloud resource used within the DataSource object. + "cloudSqlInstanceDatasourceProperties": { # CloudSqlInstanceDataSourceProperties represents the properties of a Cloud SQL resource that are stored in the DataSource. . # Output only. CloudSqlInstanceDataSourceProperties has a subset of Cloud SQL Instance properties that are useful at the Datasource level. + "databaseInstalledVersion": "A String", # Output only. The installed database version of the Cloud SQL instance. + "instanceCreateTime": "A String", # Output only. The instance creation timestamp. + "instanceTier": "A String", # Output only. The tier (or machine type) for this instance. Example: `db-custom-1-3840` + "name": "A String", # Output only. Name of the Cloud SQL instance backed up by the datasource. Format: projects/{project}/instances/{instance} + }, "computeInstanceDatasourceProperties": { # ComputeInstanceDataSourceProperties represents the properties of a ComputeEngine resource that are stored in the DataSource. # ComputeInstanceDataSourceProperties has a subset of Compute Instance properties that are useful at the Datasource level. "description": "A String", # The description of the Compute Engine instance. "machineType": "A String", # The machine type of the instance. diff --git a/docs/dyn/backupdr_v1.projects.locations.dataSourceReferences.html b/docs/dyn/backupdr_v1.projects.locations.dataSourceReferences.html new file mode 100644 index 0000000000..2bd6d9a702 --- /dev/null +++ b/docs/dyn/backupdr_v1.projects.locations.dataSourceReferences.html @@ -0,0 +1,195 @@ + + + +

Backup and DR Service API . projects . locations . dataSourceReferences

+

Instance Methods

+

+ close()

+

Close httplib2 connections.

+

+ fetchForResourceType(parent, filter=None, orderBy=None, pageSize=None, pageToken=None, resourceType=None, x__xgafv=None)

+

Fetch DataSourceReferences for a given project, location and resource type.

+

+ fetchForResourceType_next()

+

Retrieves the next page of results.

+

+ get(name, x__xgafv=None)

+

Gets details of a single DataSourceReference.

+

Method Details

+
+ close() +
Close httplib2 connections.
+
+ +
+ fetchForResourceType(parent, filter=None, orderBy=None, pageSize=None, pageToken=None, resourceType=None, x__xgafv=None) +
Fetch DataSourceReferences for a given project, location and resource type.
+
+Args:
+  parent: string, Required. The parent resource name. Format: projects/{project}/locations/{location} (required)
+  filter: string, Optional. A filter expression that filters the results fetched in the response. The expression must specify the field name, a comparison operator, and the value that you want to use for filtering. Supported fields: * data_source * data_source_gcp_resource_info.gcp_resourcename * data_source_backup_config_state * data_source_backup_count * data_source_backup_config_info.last_backup_state * data_source_gcp_resource_info.gcp_resourcename * data_source_gcp_resource_info.type * data_source_gcp_resource_info.location * data_source_gcp_resource_info.cloud_sql_instance_properties.instance_create_time
+  orderBy: string, Optional. A comma-separated list of fields to order by, sorted in ascending order. Use "desc" after a field name for descending. Supported fields: * name
+  pageSize: integer, Optional. The maximum number of DataSourceReferences to return. The service may return fewer than this value. If unspecified, at most 50 DataSourceReferences will be returned. The maximum value is 100; values above 100 will be coerced to 100.
+  pageToken: string, Optional. A page token, received from a previous call of `FetchDataSourceReferencesForResourceType`. Provide this to retrieve the subsequent page. When paginating, all other parameters provided to `FetchDataSourceReferencesForResourceType` must match the call that provided the page token.
+  resourceType: string, Required. The type of the GCP resource. Ex: sql.googleapis.com/Instance
+  x__xgafv: string, V1 error format.
+    Allowed values
+      1 - v1 error format
+      2 - v2 error format
+
+Returns:
+  An object of the form:
+
+    { # Response for the FetchDataSourceReferencesForResourceType method.
+  "dataSourceReferences": [ # The DataSourceReferences from the specified parent.
+    { # DataSourceReference is a reference to a DataSource resource.
+      "createTime": "A String", # Output only. The time when the DataSourceReference was created.
+      "dataSource": "A String", # Output only. The resource name of the DataSource. Format: projects/{project}/locations/{location}/backupVaults/{backupVault}/dataSources/{dataSource}
+      "dataSourceBackupConfigInfo": { # Information of backup configuration on the DataSource. # Output only. Information of backup configuration on the DataSource.
+        "lastBackupState": "A String", # Output only. The status of the last backup in this DataSource
+        "lastSuccessfulBackupConsistencyTime": "A String", # Output only. Timestamp of the last successful backup to this DataSource.
+      },
+      "dataSourceBackupConfigState": "A String", # Output only. The backup configuration state of the DataSource.
+      "dataSourceBackupCount": "A String", # Output only. Number of backups in the DataSource.
+      "dataSourceGcpResourceInfo": { # The GCP resource that the DataSource is associated with. # Output only. The GCP resource that the DataSource is associated with.
+        "cloudSqlInstanceProperties": { # CloudSqlInstanceDataSourceReferenceProperties represents the properties of a Cloud SQL resource that are stored in the DataSourceReference. . # Output only. The properties of the Cloud SQL instance.
+          "databaseInstalledVersion": "A String", # Output only. The installed database version of the Cloud SQL instance.
+          "instanceCreateTime": "A String", # Output only. The instance creation timestamp.
+          "instanceTier": "A String", # Output only. The tier (or machine type) for this instance. Example: `db-custom-1-3840`
+          "name": "A String", # Output only. Name of the Cloud SQL instance backed up by the datasource. Format: projects/{project}/instances/{instance}
+        },
+        "gcpResourcename": "A String", # Output only. The resource name of the GCP resource. Ex: projects/{project}/zones/{zone}/instances/{instance}
+        "location": "A String", # Output only. The location of the GCP resource. Ex: //"global"/"unspecified"
+        "type": "A String", # Output only. The type of the GCP resource. Ex: compute.googleapis.com/Instance
+      },
+      "name": "A String", # Identifier. The resource name of the DataSourceReference. Format: projects/{project}/locations/{location}/dataSourceReferences/{data_source_reference}
+    },
+  ],
+  "nextPageToken": "A String", # A token, which can be sent as `page_token` to retrieve the next page. If this field is omitted, there are no subsequent pages.
+}
+
+ +
+ fetchForResourceType_next() +
Retrieves the next page of results.
+
+        Args:
+          previous_request: The request for the previous page. (required)
+          previous_response: The response from the request for the previous page. (required)
+
+        Returns:
+          A request object that you can call 'execute()' on to request the next
+          page. Returns None if there are no more items in the collection.
+        
+
+ +
+ get(name, x__xgafv=None) +
Gets details of a single DataSourceReference.
+
+Args:
+  name: string, Required. The name of the DataSourceReference to retrieve. Format: projects/{project}/locations/{location}/dataSourceReferences/{data_source_reference} (required)
+  x__xgafv: string, V1 error format.
+    Allowed values
+      1 - v1 error format
+      2 - v2 error format
+
+Returns:
+  An object of the form:
+
+    { # DataSourceReference is a reference to a DataSource resource.
+  "createTime": "A String", # Output only. The time when the DataSourceReference was created.
+  "dataSource": "A String", # Output only. The resource name of the DataSource. Format: projects/{project}/locations/{location}/backupVaults/{backupVault}/dataSources/{dataSource}
+  "dataSourceBackupConfigInfo": { # Information of backup configuration on the DataSource. # Output only. Information of backup configuration on the DataSource.
+    "lastBackupState": "A String", # Output only. The status of the last backup in this DataSource
+    "lastSuccessfulBackupConsistencyTime": "A String", # Output only. Timestamp of the last successful backup to this DataSource.
+  },
+  "dataSourceBackupConfigState": "A String", # Output only. The backup configuration state of the DataSource.
+  "dataSourceBackupCount": "A String", # Output only. Number of backups in the DataSource.
+  "dataSourceGcpResourceInfo": { # The GCP resource that the DataSource is associated with. # Output only. The GCP resource that the DataSource is associated with.
+    "cloudSqlInstanceProperties": { # CloudSqlInstanceDataSourceReferenceProperties represents the properties of a Cloud SQL resource that are stored in the DataSourceReference. . # Output only. The properties of the Cloud SQL instance.
+      "databaseInstalledVersion": "A String", # Output only. The installed database version of the Cloud SQL instance.
+      "instanceCreateTime": "A String", # Output only. The instance creation timestamp.
+      "instanceTier": "A String", # Output only. The tier (or machine type) for this instance. Example: `db-custom-1-3840`
+      "name": "A String", # Output only. Name of the Cloud SQL instance backed up by the datasource. Format: projects/{project}/instances/{instance}
+    },
+    "gcpResourcename": "A String", # Output only. The resource name of the GCP resource. Ex: projects/{project}/zones/{zone}/instances/{instance}
+    "location": "A String", # Output only. The location of the GCP resource. Ex: //"global"/"unspecified"
+    "type": "A String", # Output only. The type of the GCP resource. Ex: compute.googleapis.com/Instance
+  },
+  "name": "A String", # Identifier. The resource name of the DataSourceReference. Format: projects/{project}/locations/{location}/dataSourceReferences/{data_source_reference}
+}
+
+ + \ No newline at end of file diff --git a/docs/dyn/backupdr_v1.projects.locations.html b/docs/dyn/backupdr_v1.projects.locations.html index 27cc209ac5..699b83eaf5 100644 --- a/docs/dyn/backupdr_v1.projects.locations.html +++ b/docs/dyn/backupdr_v1.projects.locations.html @@ -89,6 +89,11 @@

Instance Methods

Returns the backupVaults Resource.

+

+ dataSourceReferences() +

+

Returns the dataSourceReferences Resource.

+

managementServers()

diff --git a/docs/dyn/backupdr_v1.projects.locations.serviceConfig.html b/docs/dyn/backupdr_v1.projects.locations.serviceConfig.html index ca339fc241..f2763c64ef 100644 --- a/docs/dyn/backupdr_v1.projects.locations.serviceConfig.html +++ b/docs/dyn/backupdr_v1.projects.locations.serviceConfig.html @@ -96,6 +96,9 @@

Method Details

The object takes the form of: { # Request message for initializing the service. + "cloudSqlInstanceInitializationConfig": { # CloudSqlInstanceInitializationConfig contains the configuration for initializing a Cloud SQL instance. # Optional. The configuration for initializing a Cloud SQL instance. + "edition": "A String", # Required. The edition of the Cloud SQL instance. + }, "requestId": "A String", # Optional. An optional request ID to identify requests. Specify a unique request ID so that if you must retry your request, the server will know to ignore the request if it has already been completed. The server will guarantee that for at least 60 minutes since the first request. For example, consider a situation where you make an initial request and t he request times out. If you make the request again with the same request ID, the server can check if original operation with the same request ID was received, and if so, will ignore the second request. This prevents clients from accidentally creating duplicate commitments. The request ID must be a valid UUID with the exception that zero UUID is not supported (00000000-0000-0000-0000-000000000000). "resourceType": "A String", # Required. The resource type to which the default service config will be applied. Examples include, "compute.googleapis.com/Instance" and "storage.googleapis.com/Bucket". } diff --git a/docs/dyn/cloudasset_v1.v1.html b/docs/dyn/cloudasset_v1.v1.html index 9e00e7e53c..66ba0a7c89 100644 --- a/docs/dyn/cloudasset_v1.v1.html +++ b/docs/dyn/cloudasset_v1.v1.html @@ -2094,7 +2094,7 @@

Method Details

"outputConfig": { # Output configuration for export assets destination. # Required. Output configuration indicating where the results will be output to. "bigqueryDestination": { # A BigQuery destination for exporting assets to. # Destination on BigQuery. The output table stores the fields in asset Protobuf as columns in BigQuery. "dataset": "A String", # Required. The BigQuery dataset in format "projects/projectId/datasets/datasetId", to which the snapshot result should be exported. If this dataset does not exist, the export call returns an INVALID_ARGUMENT error. Setting the `contentType` for `exportAssets` determines the [schema](/asset-inventory/docs/exporting-to-bigquery#bigquery-schema) of the BigQuery table. Setting `separateTablesPerAssetType` to `TRUE` also influences the schema. - "force": True or False, # If the destination table already exists and this flag is `TRUE`, the table will be overwritten by the contents of assets snapshot. If the flag is `FALSE` or unset and the destination table already exists, the export call returns an INVALID_ARGUMEMT error. + "force": True or False, # If the destination table already exists and this flag is `TRUE`, the table will be overwritten by the contents of assets snapshot. If the flag is `FALSE` or unset and the destination table already exists, the export call returns an INVALID_ARGUMENT error. "partitionSpec": { # Specifications of BigQuery partitioned table as export destination. # [partition_spec] determines whether to export to partitioned table(s) and how to partition the data. If [partition_spec] is unset or [partition_spec.partition_key] is unset or `PARTITION_KEY_UNSPECIFIED`, the snapshot results will be exported to non-partitioned table(s). [force] will decide whether to overwrite existing table(s). If [partition_spec] is specified. First, the snapshot results will be written to partitioned table(s) with two additional timestamp columns, readTime and requestTime, one of which will be the partition key. Secondly, in the case when any destination table already exists, it will first try to update existing table's schema as necessary by appending additional columns. Then, if [force] is `TRUE`, the corresponding partition will be overwritten by the snapshot results (data in different partitions will remain intact); if [force] is unset or `FALSE`, it will append the data. An error will be returned if the schema update or data appension fails. "partitionKey": "A String", # The partition key for BigQuery partitioned table. }, diff --git a/docs/dyn/cloudasset_v1p7beta1.v1p7beta1.html b/docs/dyn/cloudasset_v1p7beta1.v1p7beta1.html index 56d97074d4..3ab5f5c6ed 100644 --- a/docs/dyn/cloudasset_v1p7beta1.v1p7beta1.html +++ b/docs/dyn/cloudasset_v1p7beta1.v1p7beta1.html @@ -103,7 +103,7 @@

Method Details

"outputConfig": { # Output configuration for export assets destination. # Required. Output configuration indicating where the results will be output to. "bigqueryDestination": { # A BigQuery destination for exporting assets to. # Destination on BigQuery. The output table stores the fields in asset proto as columns in BigQuery. "dataset": "A String", # Required. The BigQuery dataset in format "projects/projectId/datasets/datasetId", to which the snapshot result should be exported. If this dataset does not exist, the export call returns an INVALID_ARGUMENT error. - "force": True or False, # If the destination table already exists and this flag is `TRUE`, the table will be overwritten by the contents of assets snapshot. If the flag is `FALSE` or unset and the destination table already exists, the export call returns an INVALID_ARGUMEMT error. + "force": True or False, # If the destination table already exists and this flag is `TRUE`, the table will be overwritten by the contents of assets snapshot. If the flag is `FALSE` or unset and the destination table already exists, the export call returns an INVALID_ARGUMENT error. "partitionSpec": { # Specifications of BigQuery partitioned table as export destination. # [partition_spec] determines whether to export to partitioned table(s) and how to partition the data. If [partition_spec] is unset or [partition_spec.partition_key] is unset or `PARTITION_KEY_UNSPECIFIED`, the snapshot results will be exported to non-partitioned table(s). [force] will decide whether to overwrite existing table(s). If [partition_spec] is specified. First, the snapshot results will be written to partitioned table(s) with two additional timestamp columns, readTime and requestTime, one of which will be the partition key. Secondly, in the case when any destination table already exists, it will first try to update existing table's schema as necessary by appending additional columns. Then, if [force] is `TRUE`, the corresponding partition will be overwritten by the snapshot results (data in different partitions will remain intact); if [force] is unset or `FALSE`, it will append the data. An error will be returned if the schema update or data appension fails. "partitionKey": "A String", # The partition key for BigQuery partitioned table. }, diff --git a/docs/dyn/cloudidentity_v1.groups.html b/docs/dyn/cloudidentity_v1.groups.html index 561619a321..7ff642c0a8 100644 --- a/docs/dyn/cloudidentity_v1.groups.html +++ b/docs/dyn/cloudidentity_v1.groups.html @@ -155,7 +155,7 @@

Method Details

"id": "A String", # The ID of the entity. For Google-managed entities, the `id` should be the email address of an existing group or user. Email addresses need to adhere to [name guidelines for users and groups](https://support.google.com/a/answer/9193374). For external-identity-mapped entities, the `id` must be a string conforming to the Identity Source's requirements. Must be unique within a `namespace`. "namespace": "A String", # The namespace in which the entity exists. If not specified, the `EntityKey` represents a Google-managed entity such as a Google user or a Google Group. If specified, the `EntityKey` represents an external-identity-mapped group. The namespace must correspond to an identity source created in Admin Console and must be in the form of `identitysources/{identity_source}`. }, - "labels": { # Required. One or more label entries that apply to the Group. Currently supported labels contain a key with an empty value. Google Groups are the default type of group and have a label with a key of `cloudidentity.googleapis.com/groups.discussion_forum` and an empty value. Existing Google Groups can have an additional label with a key of `cloudidentity.googleapis.com/groups.security` and an empty value added to them. **This is an immutable change and the security label cannot be removed once added.** Dynamic groups have a label with a key of `cloudidentity.googleapis.com/groups.dynamic`. Identity-mapped groups for Cloud Search have a label with a key of `system/groups/external` and an empty value. + "labels": { # Required. One or more label entries that apply to the Group. Labels contain a key with an empty value. Google Groups are the default type of group and have a label with a key of `cloudidentity.googleapis.com/groups.discussion_forum` and an empty value. Existing Google Groups can have an additional label with a key of `cloudidentity.googleapis.com/groups.security` and an empty value added to them. **This is an immutable change and the security label cannot be removed once added.** Dynamic groups have a label with a key of `cloudidentity.googleapis.com/groups.dynamic`. Identity-mapped groups for Cloud Search have a label with a key of `system/groups/external` and an empty value. Google Groups can be [locked](https://support.google.com/a?p=locked-groups). To lock a group, add a label with a key of `cloudidentity.googleapis.com/groups.locked` and an empty value. Doing so locks the group. To unlock the group, remove this label. "a_key": "A String", }, "name": "A String", # Output only. The [resource name](https://cloud.google.com/apis/design/resource_names) of the `Group`. Shall be of the form `groups/{group}`. @@ -272,7 +272,7 @@

Method Details

"id": "A String", # The ID of the entity. For Google-managed entities, the `id` should be the email address of an existing group or user. Email addresses need to adhere to [name guidelines for users and groups](https://support.google.com/a/answer/9193374). For external-identity-mapped entities, the `id` must be a string conforming to the Identity Source's requirements. Must be unique within a `namespace`. "namespace": "A String", # The namespace in which the entity exists. If not specified, the `EntityKey` represents a Google-managed entity such as a Google user or a Google Group. If specified, the `EntityKey` represents an external-identity-mapped group. The namespace must correspond to an identity source created in Admin Console and must be in the form of `identitysources/{identity_source}`. }, - "labels": { # Required. One or more label entries that apply to the Group. Currently supported labels contain a key with an empty value. Google Groups are the default type of group and have a label with a key of `cloudidentity.googleapis.com/groups.discussion_forum` and an empty value. Existing Google Groups can have an additional label with a key of `cloudidentity.googleapis.com/groups.security` and an empty value added to them. **This is an immutable change and the security label cannot be removed once added.** Dynamic groups have a label with a key of `cloudidentity.googleapis.com/groups.dynamic`. Identity-mapped groups for Cloud Search have a label with a key of `system/groups/external` and an empty value. + "labels": { # Required. One or more label entries that apply to the Group. Labels contain a key with an empty value. Google Groups are the default type of group and have a label with a key of `cloudidentity.googleapis.com/groups.discussion_forum` and an empty value. Existing Google Groups can have an additional label with a key of `cloudidentity.googleapis.com/groups.security` and an empty value added to them. **This is an immutable change and the security label cannot be removed once added.** Dynamic groups have a label with a key of `cloudidentity.googleapis.com/groups.dynamic`. Identity-mapped groups for Cloud Search have a label with a key of `system/groups/external` and an empty value. Google Groups can be [locked](https://support.google.com/a?p=locked-groups). To lock a group, add a label with a key of `cloudidentity.googleapis.com/groups.locked` and an empty value. Doing so locks the group. To unlock the group, remove this label. "a_key": "A String", }, "name": "A String", # Output only. The [resource name](https://cloud.google.com/apis/design/resource_names) of the `Group`. Shall be of the form `groups/{group}`. @@ -356,7 +356,7 @@

Method Details

"id": "A String", # The ID of the entity. For Google-managed entities, the `id` should be the email address of an existing group or user. Email addresses need to adhere to [name guidelines for users and groups](https://support.google.com/a/answer/9193374). For external-identity-mapped entities, the `id` must be a string conforming to the Identity Source's requirements. Must be unique within a `namespace`. "namespace": "A String", # The namespace in which the entity exists. If not specified, the `EntityKey` represents a Google-managed entity such as a Google user or a Google Group. If specified, the `EntityKey` represents an external-identity-mapped group. The namespace must correspond to an identity source created in Admin Console and must be in the form of `identitysources/{identity_source}`. }, - "labels": { # Required. One or more label entries that apply to the Group. Currently supported labels contain a key with an empty value. Google Groups are the default type of group and have a label with a key of `cloudidentity.googleapis.com/groups.discussion_forum` and an empty value. Existing Google Groups can have an additional label with a key of `cloudidentity.googleapis.com/groups.security` and an empty value added to them. **This is an immutable change and the security label cannot be removed once added.** Dynamic groups have a label with a key of `cloudidentity.googleapis.com/groups.dynamic`. Identity-mapped groups for Cloud Search have a label with a key of `system/groups/external` and an empty value. + "labels": { # Required. One or more label entries that apply to the Group. Labels contain a key with an empty value. Google Groups are the default type of group and have a label with a key of `cloudidentity.googleapis.com/groups.discussion_forum` and an empty value. Existing Google Groups can have an additional label with a key of `cloudidentity.googleapis.com/groups.security` and an empty value added to them. **This is an immutable change and the security label cannot be removed once added.** Dynamic groups have a label with a key of `cloudidentity.googleapis.com/groups.dynamic`. Identity-mapped groups for Cloud Search have a label with a key of `system/groups/external` and an empty value. Google Groups can be [locked](https://support.google.com/a?p=locked-groups). To lock a group, add a label with a key of `cloudidentity.googleapis.com/groups.locked` and an empty value. Doing so locks the group. To unlock the group, remove this label. "a_key": "A String", }, "name": "A String", # Output only. The [resource name](https://cloud.google.com/apis/design/resource_names) of the `Group`. Shall be of the form `groups/{group}`. @@ -437,7 +437,7 @@

Method Details

"id": "A String", # The ID of the entity. For Google-managed entities, the `id` should be the email address of an existing group or user. Email addresses need to adhere to [name guidelines for users and groups](https://support.google.com/a/answer/9193374). For external-identity-mapped entities, the `id` must be a string conforming to the Identity Source's requirements. Must be unique within a `namespace`. "namespace": "A String", # The namespace in which the entity exists. If not specified, the `EntityKey` represents a Google-managed entity such as a Google user or a Google Group. If specified, the `EntityKey` represents an external-identity-mapped group. The namespace must correspond to an identity source created in Admin Console and must be in the form of `identitysources/{identity_source}`. }, - "labels": { # Required. One or more label entries that apply to the Group. Currently supported labels contain a key with an empty value. Google Groups are the default type of group and have a label with a key of `cloudidentity.googleapis.com/groups.discussion_forum` and an empty value. Existing Google Groups can have an additional label with a key of `cloudidentity.googleapis.com/groups.security` and an empty value added to them. **This is an immutable change and the security label cannot be removed once added.** Dynamic groups have a label with a key of `cloudidentity.googleapis.com/groups.dynamic`. Identity-mapped groups for Cloud Search have a label with a key of `system/groups/external` and an empty value. + "labels": { # Required. One or more label entries that apply to the Group. Labels contain a key with an empty value. Google Groups are the default type of group and have a label with a key of `cloudidentity.googleapis.com/groups.discussion_forum` and an empty value. Existing Google Groups can have an additional label with a key of `cloudidentity.googleapis.com/groups.security` and an empty value added to them. **This is an immutable change and the security label cannot be removed once added.** Dynamic groups have a label with a key of `cloudidentity.googleapis.com/groups.dynamic`. Identity-mapped groups for Cloud Search have a label with a key of `system/groups/external` and an empty value. Google Groups can be [locked](https://support.google.com/a?p=locked-groups). To lock a group, add a label with a key of `cloudidentity.googleapis.com/groups.locked` and an empty value. Doing so locks the group. To unlock the group, remove this label. "a_key": "A String", }, "name": "A String", # Output only. The [resource name](https://cloud.google.com/apis/design/resource_names) of the `Group`. Shall be of the form `groups/{group}`. @@ -524,7 +524,7 @@

Method Details

"id": "A String", # The ID of the entity. For Google-managed entities, the `id` should be the email address of an existing group or user. Email addresses need to adhere to [name guidelines for users and groups](https://support.google.com/a/answer/9193374). For external-identity-mapped entities, the `id` must be a string conforming to the Identity Source's requirements. Must be unique within a `namespace`. "namespace": "A String", # The namespace in which the entity exists. If not specified, the `EntityKey` represents a Google-managed entity such as a Google user or a Google Group. If specified, the `EntityKey` represents an external-identity-mapped group. The namespace must correspond to an identity source created in Admin Console and must be in the form of `identitysources/{identity_source}`. }, - "labels": { # Required. One or more label entries that apply to the Group. Currently supported labels contain a key with an empty value. Google Groups are the default type of group and have a label with a key of `cloudidentity.googleapis.com/groups.discussion_forum` and an empty value. Existing Google Groups can have an additional label with a key of `cloudidentity.googleapis.com/groups.security` and an empty value added to them. **This is an immutable change and the security label cannot be removed once added.** Dynamic groups have a label with a key of `cloudidentity.googleapis.com/groups.dynamic`. Identity-mapped groups for Cloud Search have a label with a key of `system/groups/external` and an empty value. + "labels": { # Required. One or more label entries that apply to the Group. Labels contain a key with an empty value. Google Groups are the default type of group and have a label with a key of `cloudidentity.googleapis.com/groups.discussion_forum` and an empty value. Existing Google Groups can have an additional label with a key of `cloudidentity.googleapis.com/groups.security` and an empty value added to them. **This is an immutable change and the security label cannot be removed once added.** Dynamic groups have a label with a key of `cloudidentity.googleapis.com/groups.dynamic`. Identity-mapped groups for Cloud Search have a label with a key of `system/groups/external` and an empty value. Google Groups can be [locked](https://support.google.com/a?p=locked-groups). To lock a group, add a label with a key of `cloudidentity.googleapis.com/groups.locked` and an empty value. Doing so locks the group. To unlock the group, remove this label. "a_key": "A String", }, "name": "A String", # Output only. The [resource name](https://cloud.google.com/apis/design/resource_names) of the `Group`. Shall be of the form `groups/{group}`. diff --git a/docs/dyn/cloudidentity_v1beta1.groups.html b/docs/dyn/cloudidentity_v1beta1.groups.html index d8bfd0bfad..9b136c76fc 100644 --- a/docs/dyn/cloudidentity_v1beta1.groups.html +++ b/docs/dyn/cloudidentity_v1beta1.groups.html @@ -155,7 +155,7 @@

Method Details

"id": "A String", # The ID of the entity. For Google-managed entities, the `id` must be the email address of an existing group or user. For external-identity-mapped entities, the `id` must be a string conforming to the Identity Source's requirements. Must be unique within a `namespace`. "namespace": "A String", # The namespace in which the entity exists. If not specified, the `EntityKey` represents a Google-managed entity such as a Google user or a Google Group. If specified, the `EntityKey` represents an external-identity-mapped group. The namespace must correspond to an identity source created in Admin Console and must be in the form of `identitysources/{identity_source_id}`. }, - "labels": { # Required. One or more label entries that apply to the Group. Labels contain a key with an empty value. Google Groups are the default type of group and have a label with a key of `cloudidentity.googleapis.com/groups.discussion_forum` and an empty value. Existing Google Groups can have an additional label with a key of `cloudidentity.googleapis.com/groups.security` and an empty value added to them. **This is an immutable change and the security label cannot be removed once added.** Dynamic groups have a label with a key of `cloudidentity.googleapis.com/groups.dynamic`. Identity-mapped groups for Cloud Search have a label with a key of `system/groups/external` and an empty value. (Beta) Google Groups can be [locked](https://support.google.com/a?p=locked-groups). To lock a group, add a label with a key of `cloudidentity.googleapis.com/groups.locked` and an empty value. Doing so locks the group. To unlock the group, remove this label. + "labels": { # Required. One or more label entries that apply to the Group. Labels contain a key with an empty value. Google Groups are the default type of group and have a label with a key of `cloudidentity.googleapis.com/groups.discussion_forum` and an empty value. Existing Google Groups can have an additional label with a key of `cloudidentity.googleapis.com/groups.security` and an empty value added to them. **This is an immutable change and the security label cannot be removed once added.** Dynamic groups have a label with a key of `cloudidentity.googleapis.com/groups.dynamic`. Identity-mapped groups for Cloud Search have a label with a key of `system/groups/external` and an empty value. Google Groups can be [locked](https://support.google.com/a?p=locked-groups). To lock a group, add a label with a key of `cloudidentity.googleapis.com/groups.locked` and an empty value. Doing so locks the group. To unlock the group, remove this label. "a_key": "A String", }, "name": "A String", # Output only. The [resource name](https://cloud.google.com/apis/design/resource_names) of the `Group`. Shall be of the form `groups/{group_id}`. @@ -279,7 +279,7 @@

Method Details

"id": "A String", # The ID of the entity. For Google-managed entities, the `id` must be the email address of an existing group or user. For external-identity-mapped entities, the `id` must be a string conforming to the Identity Source's requirements. Must be unique within a `namespace`. "namespace": "A String", # The namespace in which the entity exists. If not specified, the `EntityKey` represents a Google-managed entity such as a Google user or a Google Group. If specified, the `EntityKey` represents an external-identity-mapped group. The namespace must correspond to an identity source created in Admin Console and must be in the form of `identitysources/{identity_source_id}`. }, - "labels": { # Required. One or more label entries that apply to the Group. Labels contain a key with an empty value. Google Groups are the default type of group and have a label with a key of `cloudidentity.googleapis.com/groups.discussion_forum` and an empty value. Existing Google Groups can have an additional label with a key of `cloudidentity.googleapis.com/groups.security` and an empty value added to them. **This is an immutable change and the security label cannot be removed once added.** Dynamic groups have a label with a key of `cloudidentity.googleapis.com/groups.dynamic`. Identity-mapped groups for Cloud Search have a label with a key of `system/groups/external` and an empty value. (Beta) Google Groups can be [locked](https://support.google.com/a?p=locked-groups). To lock a group, add a label with a key of `cloudidentity.googleapis.com/groups.locked` and an empty value. Doing so locks the group. To unlock the group, remove this label. + "labels": { # Required. One or more label entries that apply to the Group. Labels contain a key with an empty value. Google Groups are the default type of group and have a label with a key of `cloudidentity.googleapis.com/groups.discussion_forum` and an empty value. Existing Google Groups can have an additional label with a key of `cloudidentity.googleapis.com/groups.security` and an empty value added to them. **This is an immutable change and the security label cannot be removed once added.** Dynamic groups have a label with a key of `cloudidentity.googleapis.com/groups.dynamic`. Identity-mapped groups for Cloud Search have a label with a key of `system/groups/external` and an empty value. Google Groups can be [locked](https://support.google.com/a?p=locked-groups). To lock a group, add a label with a key of `cloudidentity.googleapis.com/groups.locked` and an empty value. Doing so locks the group. To unlock the group, remove this label. "a_key": "A String", }, "name": "A String", # Output only. The [resource name](https://cloud.google.com/apis/design/resource_names) of the `Group`. Shall be of the form `groups/{group_id}`. @@ -370,7 +370,7 @@

Method Details

"id": "A String", # The ID of the entity. For Google-managed entities, the `id` must be the email address of an existing group or user. For external-identity-mapped entities, the `id` must be a string conforming to the Identity Source's requirements. Must be unique within a `namespace`. "namespace": "A String", # The namespace in which the entity exists. If not specified, the `EntityKey` represents a Google-managed entity such as a Google user or a Google Group. If specified, the `EntityKey` represents an external-identity-mapped group. The namespace must correspond to an identity source created in Admin Console and must be in the form of `identitysources/{identity_source_id}`. }, - "labels": { # Required. One or more label entries that apply to the Group. Labels contain a key with an empty value. Google Groups are the default type of group and have a label with a key of `cloudidentity.googleapis.com/groups.discussion_forum` and an empty value. Existing Google Groups can have an additional label with a key of `cloudidentity.googleapis.com/groups.security` and an empty value added to them. **This is an immutable change and the security label cannot be removed once added.** Dynamic groups have a label with a key of `cloudidentity.googleapis.com/groups.dynamic`. Identity-mapped groups for Cloud Search have a label with a key of `system/groups/external` and an empty value. (Beta) Google Groups can be [locked](https://support.google.com/a?p=locked-groups). To lock a group, add a label with a key of `cloudidentity.googleapis.com/groups.locked` and an empty value. Doing so locks the group. To unlock the group, remove this label. + "labels": { # Required. One or more label entries that apply to the Group. Labels contain a key with an empty value. Google Groups are the default type of group and have a label with a key of `cloudidentity.googleapis.com/groups.discussion_forum` and an empty value. Existing Google Groups can have an additional label with a key of `cloudidentity.googleapis.com/groups.security` and an empty value added to them. **This is an immutable change and the security label cannot be removed once added.** Dynamic groups have a label with a key of `cloudidentity.googleapis.com/groups.dynamic`. Identity-mapped groups for Cloud Search have a label with a key of `system/groups/external` and an empty value. Google Groups can be [locked](https://support.google.com/a?p=locked-groups). To lock a group, add a label with a key of `cloudidentity.googleapis.com/groups.locked` and an empty value. Doing so locks the group. To unlock the group, remove this label. "a_key": "A String", }, "name": "A String", # Output only. The [resource name](https://cloud.google.com/apis/design/resource_names) of the `Group`. Shall be of the form `groups/{group_id}`. @@ -458,7 +458,7 @@

Method Details

"id": "A String", # The ID of the entity. For Google-managed entities, the `id` must be the email address of an existing group or user. For external-identity-mapped entities, the `id` must be a string conforming to the Identity Source's requirements. Must be unique within a `namespace`. "namespace": "A String", # The namespace in which the entity exists. If not specified, the `EntityKey` represents a Google-managed entity such as a Google user or a Google Group. If specified, the `EntityKey` represents an external-identity-mapped group. The namespace must correspond to an identity source created in Admin Console and must be in the form of `identitysources/{identity_source_id}`. }, - "labels": { # Required. One or more label entries that apply to the Group. Labels contain a key with an empty value. Google Groups are the default type of group and have a label with a key of `cloudidentity.googleapis.com/groups.discussion_forum` and an empty value. Existing Google Groups can have an additional label with a key of `cloudidentity.googleapis.com/groups.security` and an empty value added to them. **This is an immutable change and the security label cannot be removed once added.** Dynamic groups have a label with a key of `cloudidentity.googleapis.com/groups.dynamic`. Identity-mapped groups for Cloud Search have a label with a key of `system/groups/external` and an empty value. (Beta) Google Groups can be [locked](https://support.google.com/a?p=locked-groups). To lock a group, add a label with a key of `cloudidentity.googleapis.com/groups.locked` and an empty value. Doing so locks the group. To unlock the group, remove this label. + "labels": { # Required. One or more label entries that apply to the Group. Labels contain a key with an empty value. Google Groups are the default type of group and have a label with a key of `cloudidentity.googleapis.com/groups.discussion_forum` and an empty value. Existing Google Groups can have an additional label with a key of `cloudidentity.googleapis.com/groups.security` and an empty value added to them. **This is an immutable change and the security label cannot be removed once added.** Dynamic groups have a label with a key of `cloudidentity.googleapis.com/groups.dynamic`. Identity-mapped groups for Cloud Search have a label with a key of `system/groups/external` and an empty value. Google Groups can be [locked](https://support.google.com/a?p=locked-groups). To lock a group, add a label with a key of `cloudidentity.googleapis.com/groups.locked` and an empty value. Doing so locks the group. To unlock the group, remove this label. "a_key": "A String", }, "name": "A String", # Output only. The [resource name](https://cloud.google.com/apis/design/resource_names) of the `Group`. Shall be of the form `groups/{group_id}`. @@ -552,7 +552,7 @@

Method Details

"id": "A String", # The ID of the entity. For Google-managed entities, the `id` must be the email address of an existing group or user. For external-identity-mapped entities, the `id` must be a string conforming to the Identity Source's requirements. Must be unique within a `namespace`. "namespace": "A String", # The namespace in which the entity exists. If not specified, the `EntityKey` represents a Google-managed entity such as a Google user or a Google Group. If specified, the `EntityKey` represents an external-identity-mapped group. The namespace must correspond to an identity source created in Admin Console and must be in the form of `identitysources/{identity_source_id}`. }, - "labels": { # Required. One or more label entries that apply to the Group. Labels contain a key with an empty value. Google Groups are the default type of group and have a label with a key of `cloudidentity.googleapis.com/groups.discussion_forum` and an empty value. Existing Google Groups can have an additional label with a key of `cloudidentity.googleapis.com/groups.security` and an empty value added to them. **This is an immutable change and the security label cannot be removed once added.** Dynamic groups have a label with a key of `cloudidentity.googleapis.com/groups.dynamic`. Identity-mapped groups for Cloud Search have a label with a key of `system/groups/external` and an empty value. (Beta) Google Groups can be [locked](https://support.google.com/a?p=locked-groups). To lock a group, add a label with a key of `cloudidentity.googleapis.com/groups.locked` and an empty value. Doing so locks the group. To unlock the group, remove this label. + "labels": { # Required. One or more label entries that apply to the Group. Labels contain a key with an empty value. Google Groups are the default type of group and have a label with a key of `cloudidentity.googleapis.com/groups.discussion_forum` and an empty value. Existing Google Groups can have an additional label with a key of `cloudidentity.googleapis.com/groups.security` and an empty value added to them. **This is an immutable change and the security label cannot be removed once added.** Dynamic groups have a label with a key of `cloudidentity.googleapis.com/groups.dynamic`. Identity-mapped groups for Cloud Search have a label with a key of `system/groups/external` and an empty value. Google Groups can be [locked](https://support.google.com/a?p=locked-groups). To lock a group, add a label with a key of `cloudidentity.googleapis.com/groups.locked` and an empty value. Doing so locks the group. To unlock the group, remove this label. "a_key": "A String", }, "name": "A String", # Output only. The [resource name](https://cloud.google.com/apis/design/resource_names) of the `Group`. Shall be of the form `groups/{group_id}`. diff --git a/docs/dyn/contactcenterinsights_v1.projects.locations.authorizedViewSets.authorizedViews.html b/docs/dyn/contactcenterinsights_v1.projects.locations.authorizedViewSets.authorizedViews.html index 24fe52a550..b5f598cbb5 100644 --- a/docs/dyn/contactcenterinsights_v1.projects.locations.authorizedViewSets.authorizedViews.html +++ b/docs/dyn/contactcenterinsights_v1.projects.locations.authorizedViewSets.authorizedViews.html @@ -133,7 +133,7 @@

Method Details

The object takes the form of: { # An AuthorizedView represents a view of accessible Insights resources (for example, Conversation and Scorecard). Who have read access to the AuthorizedView resource will have access to these Insight resources as well. - "conversationFilter": "A String", # A filter to reduce conversation results to a specific subset. The AuthorizedView's assigned permission (read/write) could be applied to the subset of conversations. If conversation_filter is empty, there is no restriction on the conversations that the AuthorizedView can access. Having *authorizedViews.get* access to the AuthorizedView means having the same read/write access to the Conversations (as well as metadata/annotations liked to the conversation) that this AuthorizedView has. + "conversationFilter": "A String", # A filter to reduce conversation results to a specific subset. The AuthorizedView's assigned permission (read/write) could be applied to the subset of conversations. If conversation_filter is empty, there is no restriction on the conversations that the AuthorizedView can access. Having *authorizedViews.get* access to the AuthorizedView means having the same read/write access to the Conversations (as well as metadata/annotations linked to the conversation) that this AuthorizedView has. "createTime": "A String", # Output only. The time at which the authorized view was created. "displayName": "A String", # Display Name. Limit 64 characters. "name": "A String", # Identifier. The resource name of the AuthorizedView. Format: projects/{project}/locations/{location}/authorizedViewSets/{authorized_view_set}/authorizedViews/{authorized_view} @@ -150,7 +150,7 @@

Method Details

An object of the form: { # An AuthorizedView represents a view of accessible Insights resources (for example, Conversation and Scorecard). Who have read access to the AuthorizedView resource will have access to these Insight resources as well. - "conversationFilter": "A String", # A filter to reduce conversation results to a specific subset. The AuthorizedView's assigned permission (read/write) could be applied to the subset of conversations. If conversation_filter is empty, there is no restriction on the conversations that the AuthorizedView can access. Having *authorizedViews.get* access to the AuthorizedView means having the same read/write access to the Conversations (as well as metadata/annotations liked to the conversation) that this AuthorizedView has. + "conversationFilter": "A String", # A filter to reduce conversation results to a specific subset. The AuthorizedView's assigned permission (read/write) could be applied to the subset of conversations. If conversation_filter is empty, there is no restriction on the conversations that the AuthorizedView can access. Having *authorizedViews.get* access to the AuthorizedView means having the same read/write access to the Conversations (as well as metadata/annotations linked to the conversation) that this AuthorizedView has. "createTime": "A String", # Output only. The time at which the authorized view was created. "displayName": "A String", # Display Name. Limit 64 characters. "name": "A String", # Identifier. The resource name of the AuthorizedView. Format: projects/{project}/locations/{location}/authorizedViewSets/{authorized_view_set}/authorizedViews/{authorized_view} @@ -191,7 +191,7 @@

Method Details

An object of the form: { # An AuthorizedView represents a view of accessible Insights resources (for example, Conversation and Scorecard). Who have read access to the AuthorizedView resource will have access to these Insight resources as well. - "conversationFilter": "A String", # A filter to reduce conversation results to a specific subset. The AuthorizedView's assigned permission (read/write) could be applied to the subset of conversations. If conversation_filter is empty, there is no restriction on the conversations that the AuthorizedView can access. Having *authorizedViews.get* access to the AuthorizedView means having the same read/write access to the Conversations (as well as metadata/annotations liked to the conversation) that this AuthorizedView has. + "conversationFilter": "A String", # A filter to reduce conversation results to a specific subset. The AuthorizedView's assigned permission (read/write) could be applied to the subset of conversations. If conversation_filter is empty, there is no restriction on the conversations that the AuthorizedView can access. Having *authorizedViews.get* access to the AuthorizedView means having the same read/write access to the Conversations (as well as metadata/annotations linked to the conversation) that this AuthorizedView has. "createTime": "A String", # Output only. The time at which the authorized view was created. "displayName": "A String", # Display Name. Limit 64 characters. "name": "A String", # Identifier. The resource name of the AuthorizedView. Format: projects/{project}/locations/{location}/authorizedViewSets/{authorized_view_set}/authorizedViews/{authorized_view} @@ -220,7 +220,7 @@

Method Details

{ # The response from a ListAuthorizedViews request. "authorizedViews": [ # The AuthorizedViews under the parent. { # An AuthorizedView represents a view of accessible Insights resources (for example, Conversation and Scorecard). Who have read access to the AuthorizedView resource will have access to these Insight resources as well. - "conversationFilter": "A String", # A filter to reduce conversation results to a specific subset. The AuthorizedView's assigned permission (read/write) could be applied to the subset of conversations. If conversation_filter is empty, there is no restriction on the conversations that the AuthorizedView can access. Having *authorizedViews.get* access to the AuthorizedView means having the same read/write access to the Conversations (as well as metadata/annotations liked to the conversation) that this AuthorizedView has. + "conversationFilter": "A String", # A filter to reduce conversation results to a specific subset. The AuthorizedView's assigned permission (read/write) could be applied to the subset of conversations. If conversation_filter is empty, there is no restriction on the conversations that the AuthorizedView can access. Having *authorizedViews.get* access to the AuthorizedView means having the same read/write access to the Conversations (as well as metadata/annotations linked to the conversation) that this AuthorizedView has. "createTime": "A String", # Output only. The time at which the authorized view was created. "displayName": "A String", # Display Name. Limit 64 characters. "name": "A String", # Identifier. The resource name of the AuthorizedView. Format: projects/{project}/locations/{location}/authorizedViewSets/{authorized_view_set}/authorizedViews/{authorized_view} @@ -255,7 +255,7 @@

Method Details

The object takes the form of: { # An AuthorizedView represents a view of accessible Insights resources (for example, Conversation and Scorecard). Who have read access to the AuthorizedView resource will have access to these Insight resources as well. - "conversationFilter": "A String", # A filter to reduce conversation results to a specific subset. The AuthorizedView's assigned permission (read/write) could be applied to the subset of conversations. If conversation_filter is empty, there is no restriction on the conversations that the AuthorizedView can access. Having *authorizedViews.get* access to the AuthorizedView means having the same read/write access to the Conversations (as well as metadata/annotations liked to the conversation) that this AuthorizedView has. + "conversationFilter": "A String", # A filter to reduce conversation results to a specific subset. The AuthorizedView's assigned permission (read/write) could be applied to the subset of conversations. If conversation_filter is empty, there is no restriction on the conversations that the AuthorizedView can access. Having *authorizedViews.get* access to the AuthorizedView means having the same read/write access to the Conversations (as well as metadata/annotations linked to the conversation) that this AuthorizedView has. "createTime": "A String", # Output only. The time at which the authorized view was created. "displayName": "A String", # Display Name. Limit 64 characters. "name": "A String", # Identifier. The resource name of the AuthorizedView. Format: projects/{project}/locations/{location}/authorizedViewSets/{authorized_view_set}/authorizedViews/{authorized_view} @@ -272,7 +272,7 @@

Method Details

An object of the form: { # An AuthorizedView represents a view of accessible Insights resources (for example, Conversation and Scorecard). Who have read access to the AuthorizedView resource will have access to these Insight resources as well. - "conversationFilter": "A String", # A filter to reduce conversation results to a specific subset. The AuthorizedView's assigned permission (read/write) could be applied to the subset of conversations. If conversation_filter is empty, there is no restriction on the conversations that the AuthorizedView can access. Having *authorizedViews.get* access to the AuthorizedView means having the same read/write access to the Conversations (as well as metadata/annotations liked to the conversation) that this AuthorizedView has. + "conversationFilter": "A String", # A filter to reduce conversation results to a specific subset. The AuthorizedView's assigned permission (read/write) could be applied to the subset of conversations. If conversation_filter is empty, there is no restriction on the conversations that the AuthorizedView can access. Having *authorizedViews.get* access to the AuthorizedView means having the same read/write access to the Conversations (as well as metadata/annotations linked to the conversation) that this AuthorizedView has. "createTime": "A String", # Output only. The time at which the authorized view was created. "displayName": "A String", # Display Name. Limit 64 characters. "name": "A String", # Identifier. The resource name of the AuthorizedView. Format: projects/{project}/locations/{location}/authorizedViewSets/{authorized_view_set}/authorizedViews/{authorized_view} @@ -424,7 +424,7 @@

Method Details

{ # The response from a ListAuthorizedViews request. "authorizedViews": [ # The AuthorizedViews under the parent. { # An AuthorizedView represents a view of accessible Insights resources (for example, Conversation and Scorecard). Who have read access to the AuthorizedView resource will have access to these Insight resources as well. - "conversationFilter": "A String", # A filter to reduce conversation results to a specific subset. The AuthorizedView's assigned permission (read/write) could be applied to the subset of conversations. If conversation_filter is empty, there is no restriction on the conversations that the AuthorizedView can access. Having *authorizedViews.get* access to the AuthorizedView means having the same read/write access to the Conversations (as well as metadata/annotations liked to the conversation) that this AuthorizedView has. + "conversationFilter": "A String", # A filter to reduce conversation results to a specific subset. The AuthorizedView's assigned permission (read/write) could be applied to the subset of conversations. If conversation_filter is empty, there is no restriction on the conversations that the AuthorizedView can access. Having *authorizedViews.get* access to the AuthorizedView means having the same read/write access to the Conversations (as well as metadata/annotations linked to the conversation) that this AuthorizedView has. "createTime": "A String", # Output only. The time at which the authorized view was created. "displayName": "A String", # Display Name. Limit 64 characters. "name": "A String", # Identifier. The resource name of the AuthorizedView. Format: projects/{project}/locations/{location}/authorizedViewSets/{authorized_view_set}/authorizedViews/{authorized_view} diff --git a/docs/dyn/datamigration_v1.projects.locations.privateConnections.html b/docs/dyn/datamigration_v1.projects.locations.privateConnections.html index e09a2bfdb9..28a27d429f 100644 --- a/docs/dyn/datamigration_v1.projects.locations.privateConnections.html +++ b/docs/dyn/datamigration_v1.projects.locations.privateConnections.html @@ -78,7 +78,7 @@

Instance Methods

close()

Close httplib2 connections.

- create(parent, body=None, privateConnectionId=None, requestId=None, skipValidation=None, x__xgafv=None)

+ create(parent, body=None, privateConnectionId=None, requestId=None, skipValidation=None, validateOnly=None, x__xgafv=None)

Creates a new private connection in a given project and location.

delete(name, requestId=None, x__xgafv=None)

@@ -108,7 +108,7 @@

Method Details

- create(parent, body=None, privateConnectionId=None, requestId=None, skipValidation=None, x__xgafv=None) + create(parent, body=None, privateConnectionId=None, requestId=None, skipValidation=None, validateOnly=None, x__xgafv=None)
Creates a new private connection in a given project and location.
 
 Args:
@@ -132,6 +132,9 @@ 

Method Details

"a_key": "A String", }, "name": "A String", # The name of the resource. + "pscInterfaceConfig": { # The PSC Interface configuration is used to create PSC Interface between DMS's internal VPC and the consumer's PSC. # PSC Interface configuration. + "networkAttachment": "A String", # Required. Fully qualified name of the Network Attachment that DMS will connect to. Format: `projects/{{project}}/regions/{{region}}/networkAttachments/{{name}}` + }, "satisfiesPzi": True or False, # Output only. Reserved for future use. "satisfiesPzs": True or False, # Output only. Reserved for future use. "state": "A String", # Output only. The state of the private connection. @@ -145,6 +148,7 @@

Method Details

privateConnectionId: string, Required. The private connection identifier. requestId: string, Optional. A unique ID used to identify the request. If the server receives two requests with the same ID, then the second request is ignored. It is recommended to always set this value to a UUID. The ID must contain only letters (a-z, A-Z), numbers (0-9), underscores (_), and hyphens (-). The maximum length is 40 characters. skipValidation: boolean, Optional. If set to true, will skip validations. + validateOnly: boolean, Optional. For PSC Interface only - get the tenant project before creating the resource. x__xgafv: string, V1 error format. Allowed values 1 - v1 error format @@ -240,6 +244,9 @@

Method Details

"a_key": "A String", }, "name": "A String", # The name of the resource. + "pscInterfaceConfig": { # The PSC Interface configuration is used to create PSC Interface between DMS's internal VPC and the consumer's PSC. # PSC Interface configuration. + "networkAttachment": "A String", # Required. Fully qualified name of the Network Attachment that DMS will connect to. Format: `projects/{{project}}/regions/{{region}}/networkAttachments/{{name}}` + }, "satisfiesPzi": True or False, # Output only. Reserved for future use. "satisfiesPzs": True or False, # Output only. Reserved for future use. "state": "A String", # Output only. The state of the private connection. @@ -336,6 +343,9 @@

Method Details

"a_key": "A String", }, "name": "A String", # The name of the resource. + "pscInterfaceConfig": { # The PSC Interface configuration is used to create PSC Interface between DMS's internal VPC and the consumer's PSC. # PSC Interface configuration. + "networkAttachment": "A String", # Required. Fully qualified name of the Network Attachment that DMS will connect to. Format: `projects/{{project}}/regions/{{region}}/networkAttachments/{{name}}` + }, "satisfiesPzi": True or False, # Output only. Reserved for future use. "satisfiesPzs": True or False, # Output only. Reserved for future use. "state": "A String", # Output only. The state of the private connection. diff --git a/docs/dyn/dataproc_v1.projects.locations.sessions.sparkApplications.html b/docs/dyn/dataproc_v1.projects.locations.sessions.sparkApplications.html index 75fbd068f9..54f3aa6198 100644 --- a/docs/dyn/dataproc_v1.projects.locations.sessions.sparkApplications.html +++ b/docs/dyn/dataproc_v1.projects.locations.sessions.sparkApplications.html @@ -150,13 +150,13 @@

Instance Methods

summarizeExecutors(name, parent=None, x__xgafv=None)

Obtain summary of Executor Summary for a Spark Application

- summarizeJobs(name, parent=None, x__xgafv=None)

+ summarizeJobs(name, jobIds=None, parent=None, x__xgafv=None)

Obtain summary of Jobs for a Spark Application

summarizeStageAttemptTasks(name, parent=None, stageAttemptId=None, stageId=None, x__xgafv=None)

Obtain summary of Tasks for a Spark Application Stage Attempt

- summarizeStages(name, parent=None, x__xgafv=None)

+ summarizeStages(name, parent=None, stageIds=None, x__xgafv=None)

Obtain summary of Stages for a Spark Application

write(name, body=None, x__xgafv=None)

@@ -3071,11 +3071,12 @@

Method Details

- summarizeJobs(name, parent=None, x__xgafv=None) + summarizeJobs(name, jobIds=None, parent=None, x__xgafv=None)
Obtain summary of Jobs for a Spark Application
 
 Args:
   name: string, Required. The fully qualified name of the session to retrieve in the format "projects/PROJECT_ID/locations/DATAPROC_REGION/sessions/SESSION_ID/sparkApplications/APPLICATION_ID" (required)
+  jobIds: string, Optional. List of Job IDs to filter by if provided. (repeated)
   parent: string, Required. Parent (Session) resource reference.
   x__xgafv: string, V1 error format.
     Allowed values
@@ -3141,12 +3142,13 @@ 

Method Details

- summarizeStages(name, parent=None, x__xgafv=None) + summarizeStages(name, parent=None, stageIds=None, x__xgafv=None)
Obtain summary of Stages for a Spark Application
 
 Args:
   name: string, Required. The fully qualified name of the session to retrieve in the format "projects/PROJECT_ID/locations/DATAPROC_REGION/sessions/SESSION_ID/sparkApplications/APPLICATION_ID" (required)
   parent: string, Required. Parent (Session) resource reference.
+  stageIds: string, Optional. List of Stage IDs to filter by if provided. (repeated)
   x__xgafv: string, V1 error format.
     Allowed values
       1 - v1 error format
diff --git a/docs/dyn/dataproc_v1.projects.locations.workflowTemplates.html b/docs/dyn/dataproc_v1.projects.locations.workflowTemplates.html
index 3f6eb87145..2eb555943c 100644
--- a/docs/dyn/dataproc_v1.projects.locations.workflowTemplates.html
+++ b/docs/dyn/dataproc_v1.projects.locations.workflowTemplates.html
@@ -588,8 +588,11 @@ 

Method Details

"lifecycleConfig": { # Specifies the cluster auto-delete schedule configuration. # Optional. Lifecycle setting for the cluster. "autoDeleteTime": "A String", # Optional. The time when cluster will be auto-deleted (see JSON representation of Timestamp (https://developers.google.com/protocol-buffers/docs/proto3#json)). "autoDeleteTtl": "A String", # Optional. The lifetime duration of cluster. The cluster will be auto-deleted at the end of this period. Minimum value is 10 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). + "autoStopTime": "A String", # Optional. The time when cluster will be auto-stopped (see JSON representation of Timestamp (https://developers.google.com/protocol-buffers/docs/proto3#json)). + "autoStopTtl": "A String", # Optional. The lifetime duration of the cluster. The cluster will be auto-stopped at the end of this period, calculated from the time of submission of the create or update cluster request. Minimum value is 10 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). "idleDeleteTtl": "A String", # Optional. The duration to keep the cluster alive while idling (when no jobs are running). Passing this threshold will cause the cluster to be deleted. Minimum value is 5 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). "idleStartTime": "A String", # Output only. The time when cluster became idle (most recent job finished) and became eligible for deletion due to idleness (see JSON representation of Timestamp (https://developers.google.com/protocol-buffers/docs/proto3#json)). + "idleStopTtl": "A String", # Optional. The duration to keep the cluster started while idling (when no jobs are running). Passing this threshold will cause the cluster to be stopped. Minimum value is 5 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). }, "masterConfig": { # The config settings for Compute Engine resources in an instance group, such as a master or worker group. # Optional. The Compute Engine config settings for the cluster's master instance. "accelerators": [ # Optional. The Compute Engine accelerator configuration for these instances. @@ -1295,8 +1298,11 @@

Method Details

"lifecycleConfig": { # Specifies the cluster auto-delete schedule configuration. # Optional. Lifecycle setting for the cluster. "autoDeleteTime": "A String", # Optional. The time when cluster will be auto-deleted (see JSON representation of Timestamp (https://developers.google.com/protocol-buffers/docs/proto3#json)). "autoDeleteTtl": "A String", # Optional. The lifetime duration of cluster. The cluster will be auto-deleted at the end of this period. Minimum value is 10 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). + "autoStopTime": "A String", # Optional. The time when cluster will be auto-stopped (see JSON representation of Timestamp (https://developers.google.com/protocol-buffers/docs/proto3#json)). + "autoStopTtl": "A String", # Optional. The lifetime duration of the cluster. The cluster will be auto-stopped at the end of this period, calculated from the time of submission of the create or update cluster request. Minimum value is 10 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). "idleDeleteTtl": "A String", # Optional. The duration to keep the cluster alive while idling (when no jobs are running). Passing this threshold will cause the cluster to be deleted. Minimum value is 5 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). "idleStartTime": "A String", # Output only. The time when cluster became idle (most recent job finished) and became eligible for deletion due to idleness (see JSON representation of Timestamp (https://developers.google.com/protocol-buffers/docs/proto3#json)). + "idleStopTtl": "A String", # Optional. The duration to keep the cluster started while idling (when no jobs are running). Passing this threshold will cause the cluster to be stopped. Minimum value is 5 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). }, "masterConfig": { # The config settings for Compute Engine resources in an instance group, such as a master or worker group. # Optional. The Compute Engine config settings for the cluster's master instance. "accelerators": [ # Optional. The Compute Engine accelerator configuration for these instances. @@ -2029,8 +2035,11 @@

Method Details

"lifecycleConfig": { # Specifies the cluster auto-delete schedule configuration. # Optional. Lifecycle setting for the cluster. "autoDeleteTime": "A String", # Optional. The time when cluster will be auto-deleted (see JSON representation of Timestamp (https://developers.google.com/protocol-buffers/docs/proto3#json)). "autoDeleteTtl": "A String", # Optional. The lifetime duration of cluster. The cluster will be auto-deleted at the end of this period. Minimum value is 10 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). + "autoStopTime": "A String", # Optional. The time when cluster will be auto-stopped (see JSON representation of Timestamp (https://developers.google.com/protocol-buffers/docs/proto3#json)). + "autoStopTtl": "A String", # Optional. The lifetime duration of the cluster. The cluster will be auto-stopped at the end of this period, calculated from the time of submission of the create or update cluster request. Minimum value is 10 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). "idleDeleteTtl": "A String", # Optional. The duration to keep the cluster alive while idling (when no jobs are running). Passing this threshold will cause the cluster to be deleted. Minimum value is 5 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). "idleStartTime": "A String", # Output only. The time when cluster became idle (most recent job finished) and became eligible for deletion due to idleness (see JSON representation of Timestamp (https://developers.google.com/protocol-buffers/docs/proto3#json)). + "idleStopTtl": "A String", # Optional. The duration to keep the cluster started while idling (when no jobs are running). Passing this threshold will cause the cluster to be stopped. Minimum value is 5 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). }, "masterConfig": { # The config settings for Compute Engine resources in an instance group, such as a master or worker group. # Optional. The Compute Engine config settings for the cluster's master instance. "accelerators": [ # Optional. The Compute Engine accelerator configuration for these instances. @@ -2827,8 +2836,11 @@

Method Details

"lifecycleConfig": { # Specifies the cluster auto-delete schedule configuration. # Optional. Lifecycle setting for the cluster. "autoDeleteTime": "A String", # Optional. The time when cluster will be auto-deleted (see JSON representation of Timestamp (https://developers.google.com/protocol-buffers/docs/proto3#json)). "autoDeleteTtl": "A String", # Optional. The lifetime duration of cluster. The cluster will be auto-deleted at the end of this period. Minimum value is 10 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). + "autoStopTime": "A String", # Optional. The time when cluster will be auto-stopped (see JSON representation of Timestamp (https://developers.google.com/protocol-buffers/docs/proto3#json)). + "autoStopTtl": "A String", # Optional. The lifetime duration of the cluster. The cluster will be auto-stopped at the end of this period, calculated from the time of submission of the create or update cluster request. Minimum value is 10 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). "idleDeleteTtl": "A String", # Optional. The duration to keep the cluster alive while idling (when no jobs are running). Passing this threshold will cause the cluster to be deleted. Minimum value is 5 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). "idleStartTime": "A String", # Output only. The time when cluster became idle (most recent job finished) and became eligible for deletion due to idleness (see JSON representation of Timestamp (https://developers.google.com/protocol-buffers/docs/proto3#json)). + "idleStopTtl": "A String", # Optional. The duration to keep the cluster started while idling (when no jobs are running). Passing this threshold will cause the cluster to be stopped. Minimum value is 5 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). }, "masterConfig": { # The config settings for Compute Engine resources in an instance group, such as a master or worker group. # Optional. The Compute Engine config settings for the cluster's master instance. "accelerators": [ # Optional. The Compute Engine accelerator configuration for these instances. @@ -3575,8 +3587,11 @@

Method Details

"lifecycleConfig": { # Specifies the cluster auto-delete schedule configuration. # Optional. Lifecycle setting for the cluster. "autoDeleteTime": "A String", # Optional. The time when cluster will be auto-deleted (see JSON representation of Timestamp (https://developers.google.com/protocol-buffers/docs/proto3#json)). "autoDeleteTtl": "A String", # Optional. The lifetime duration of cluster. The cluster will be auto-deleted at the end of this period. Minimum value is 10 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). + "autoStopTime": "A String", # Optional. The time when cluster will be auto-stopped (see JSON representation of Timestamp (https://developers.google.com/protocol-buffers/docs/proto3#json)). + "autoStopTtl": "A String", # Optional. The lifetime duration of the cluster. The cluster will be auto-stopped at the end of this period, calculated from the time of submission of the create or update cluster request. Minimum value is 10 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). "idleDeleteTtl": "A String", # Optional. The duration to keep the cluster alive while idling (when no jobs are running). Passing this threshold will cause the cluster to be deleted. Minimum value is 5 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). "idleStartTime": "A String", # Output only. The time when cluster became idle (most recent job finished) and became eligible for deletion due to idleness (see JSON representation of Timestamp (https://developers.google.com/protocol-buffers/docs/proto3#json)). + "idleStopTtl": "A String", # Optional. The duration to keep the cluster started while idling (when no jobs are running). Passing this threshold will cause the cluster to be stopped. Minimum value is 5 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). }, "masterConfig": { # The config settings for Compute Engine resources in an instance group, such as a master or worker group. # Optional. The Compute Engine config settings for the cluster's master instance. "accelerators": [ # Optional. The Compute Engine accelerator configuration for these instances. @@ -4391,8 +4406,11 @@

Method Details

"lifecycleConfig": { # Specifies the cluster auto-delete schedule configuration. # Optional. Lifecycle setting for the cluster. "autoDeleteTime": "A String", # Optional. The time when cluster will be auto-deleted (see JSON representation of Timestamp (https://developers.google.com/protocol-buffers/docs/proto3#json)). "autoDeleteTtl": "A String", # Optional. The lifetime duration of cluster. The cluster will be auto-deleted at the end of this period. Minimum value is 10 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). + "autoStopTime": "A String", # Optional. The time when cluster will be auto-stopped (see JSON representation of Timestamp (https://developers.google.com/protocol-buffers/docs/proto3#json)). + "autoStopTtl": "A String", # Optional. The lifetime duration of the cluster. The cluster will be auto-stopped at the end of this period, calculated from the time of submission of the create or update cluster request. Minimum value is 10 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). "idleDeleteTtl": "A String", # Optional. The duration to keep the cluster alive while idling (when no jobs are running). Passing this threshold will cause the cluster to be deleted. Minimum value is 5 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). "idleStartTime": "A String", # Output only. The time when cluster became idle (most recent job finished) and became eligible for deletion due to idleness (see JSON representation of Timestamp (https://developers.google.com/protocol-buffers/docs/proto3#json)). + "idleStopTtl": "A String", # Optional. The duration to keep the cluster started while idling (when no jobs are running). Passing this threshold will cause the cluster to be stopped. Minimum value is 5 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). }, "masterConfig": { # The config settings for Compute Engine resources in an instance group, such as a master or worker group. # Optional. The Compute Engine config settings for the cluster's master instance. "accelerators": [ # Optional. The Compute Engine accelerator configuration for these instances. @@ -5098,8 +5116,11 @@

Method Details

"lifecycleConfig": { # Specifies the cluster auto-delete schedule configuration. # Optional. Lifecycle setting for the cluster. "autoDeleteTime": "A String", # Optional. The time when cluster will be auto-deleted (see JSON representation of Timestamp (https://developers.google.com/protocol-buffers/docs/proto3#json)). "autoDeleteTtl": "A String", # Optional. The lifetime duration of cluster. The cluster will be auto-deleted at the end of this period. Minimum value is 10 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). + "autoStopTime": "A String", # Optional. The time when cluster will be auto-stopped (see JSON representation of Timestamp (https://developers.google.com/protocol-buffers/docs/proto3#json)). + "autoStopTtl": "A String", # Optional. The lifetime duration of the cluster. The cluster will be auto-stopped at the end of this period, calculated from the time of submission of the create or update cluster request. Minimum value is 10 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). "idleDeleteTtl": "A String", # Optional. The duration to keep the cluster alive while idling (when no jobs are running). Passing this threshold will cause the cluster to be deleted. Minimum value is 5 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). "idleStartTime": "A String", # Output only. The time when cluster became idle (most recent job finished) and became eligible for deletion due to idleness (see JSON representation of Timestamp (https://developers.google.com/protocol-buffers/docs/proto3#json)). + "idleStopTtl": "A String", # Optional. The duration to keep the cluster started while idling (when no jobs are running). Passing this threshold will cause the cluster to be stopped. Minimum value is 5 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). }, "masterConfig": { # The config settings for Compute Engine resources in an instance group, such as a master or worker group. # Optional. The Compute Engine config settings for the cluster's master instance. "accelerators": [ # Optional. The Compute Engine accelerator configuration for these instances. diff --git a/docs/dyn/dataproc_v1.projects.regions.clusters.html b/docs/dyn/dataproc_v1.projects.regions.clusters.html index 4b7fc9f3a3..a657c423cd 100644 --- a/docs/dyn/dataproc_v1.projects.regions.clusters.html +++ b/docs/dyn/dataproc_v1.projects.regions.clusters.html @@ -332,8 +332,11 @@

Method Details

"lifecycleConfig": { # Specifies the cluster auto-delete schedule configuration. # Optional. Lifecycle setting for the cluster. "autoDeleteTime": "A String", # Optional. The time when cluster will be auto-deleted (see JSON representation of Timestamp (https://developers.google.com/protocol-buffers/docs/proto3#json)). "autoDeleteTtl": "A String", # Optional. The lifetime duration of cluster. The cluster will be auto-deleted at the end of this period. Minimum value is 10 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). + "autoStopTime": "A String", # Optional. The time when cluster will be auto-stopped (see JSON representation of Timestamp (https://developers.google.com/protocol-buffers/docs/proto3#json)). + "autoStopTtl": "A String", # Optional. The lifetime duration of the cluster. The cluster will be auto-stopped at the end of this period, calculated from the time of submission of the create or update cluster request. Minimum value is 10 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). "idleDeleteTtl": "A String", # Optional. The duration to keep the cluster alive while idling (when no jobs are running). Passing this threshold will cause the cluster to be deleted. Minimum value is 5 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). "idleStartTime": "A String", # Output only. The time when cluster became idle (most recent job finished) and became eligible for deletion due to idleness (see JSON representation of Timestamp (https://developers.google.com/protocol-buffers/docs/proto3#json)). + "idleStopTtl": "A String", # Optional. The duration to keep the cluster started while idling (when no jobs are running). Passing this threshold will cause the cluster to be stopped. Minimum value is 5 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). }, "masterConfig": { # The config settings for Compute Engine resources in an instance group, such as a master or worker group. # Optional. The Compute Engine config settings for the cluster's master instance. "accelerators": [ # Optional. The Compute Engine accelerator configuration for these instances. @@ -988,8 +991,11 @@

Method Details

"lifecycleConfig": { # Specifies the cluster auto-delete schedule configuration. # Optional. Lifecycle setting for the cluster. "autoDeleteTime": "A String", # Optional. The time when cluster will be auto-deleted (see JSON representation of Timestamp (https://developers.google.com/protocol-buffers/docs/proto3#json)). "autoDeleteTtl": "A String", # Optional. The lifetime duration of cluster. The cluster will be auto-deleted at the end of this period. Minimum value is 10 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). + "autoStopTime": "A String", # Optional. The time when cluster will be auto-stopped (see JSON representation of Timestamp (https://developers.google.com/protocol-buffers/docs/proto3#json)). + "autoStopTtl": "A String", # Optional. The lifetime duration of the cluster. The cluster will be auto-stopped at the end of this period, calculated from the time of submission of the create or update cluster request. Minimum value is 10 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). "idleDeleteTtl": "A String", # Optional. The duration to keep the cluster alive while idling (when no jobs are running). Passing this threshold will cause the cluster to be deleted. Minimum value is 5 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). "idleStartTime": "A String", # Output only. The time when cluster became idle (most recent job finished) and became eligible for deletion due to idleness (see JSON representation of Timestamp (https://developers.google.com/protocol-buffers/docs/proto3#json)). + "idleStopTtl": "A String", # Optional. The duration to keep the cluster started while idling (when no jobs are running). Passing this threshold will cause the cluster to be stopped. Minimum value is 5 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). }, "masterConfig": { # The config settings for Compute Engine resources in an instance group, such as a master or worker group. # Optional. The Compute Engine config settings for the cluster's master instance. "accelerators": [ # Optional. The Compute Engine accelerator configuration for these instances. @@ -1605,8 +1611,11 @@

Method Details

"lifecycleConfig": { # Specifies the cluster auto-delete schedule configuration. # Optional. Lifecycle setting for the cluster. "autoDeleteTime": "A String", # Optional. The time when cluster will be auto-deleted (see JSON representation of Timestamp (https://developers.google.com/protocol-buffers/docs/proto3#json)). "autoDeleteTtl": "A String", # Optional. The lifetime duration of cluster. The cluster will be auto-deleted at the end of this period. Minimum value is 10 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). + "autoStopTime": "A String", # Optional. The time when cluster will be auto-stopped (see JSON representation of Timestamp (https://developers.google.com/protocol-buffers/docs/proto3#json)). + "autoStopTtl": "A String", # Optional. The lifetime duration of the cluster. The cluster will be auto-stopped at the end of this period, calculated from the time of submission of the create or update cluster request. Minimum value is 10 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). "idleDeleteTtl": "A String", # Optional. The duration to keep the cluster alive while idling (when no jobs are running). Passing this threshold will cause the cluster to be deleted. Minimum value is 5 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). "idleStartTime": "A String", # Output only. The time when cluster became idle (most recent job finished) and became eligible for deletion due to idleness (see JSON representation of Timestamp (https://developers.google.com/protocol-buffers/docs/proto3#json)). + "idleStopTtl": "A String", # Optional. The duration to keep the cluster started while idling (when no jobs are running). Passing this threshold will cause the cluster to be stopped. Minimum value is 5 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). }, "masterConfig": { # The config settings for Compute Engine resources in an instance group, such as a master or worker group. # Optional. The Compute Engine config settings for the cluster's master instance. "accelerators": [ # Optional. The Compute Engine accelerator configuration for these instances. @@ -2142,8 +2151,11 @@

Method Details

"lifecycleConfig": { # Specifies the cluster auto-delete schedule configuration. # Optional. Lifecycle setting for the cluster. "autoDeleteTime": "A String", # Optional. The time when cluster will be auto-deleted (see JSON representation of Timestamp (https://developers.google.com/protocol-buffers/docs/proto3#json)). "autoDeleteTtl": "A String", # Optional. The lifetime duration of cluster. The cluster will be auto-deleted at the end of this period. Minimum value is 10 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). + "autoStopTime": "A String", # Optional. The time when cluster will be auto-stopped (see JSON representation of Timestamp (https://developers.google.com/protocol-buffers/docs/proto3#json)). + "autoStopTtl": "A String", # Optional. The lifetime duration of the cluster. The cluster will be auto-stopped at the end of this period, calculated from the time of submission of the create or update cluster request. Minimum value is 10 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). "idleDeleteTtl": "A String", # Optional. The duration to keep the cluster alive while idling (when no jobs are running). Passing this threshold will cause the cluster to be deleted. Minimum value is 5 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). "idleStartTime": "A String", # Output only. The time when cluster became idle (most recent job finished) and became eligible for deletion due to idleness (see JSON representation of Timestamp (https://developers.google.com/protocol-buffers/docs/proto3#json)). + "idleStopTtl": "A String", # Optional. The duration to keep the cluster started while idling (when no jobs are running). Passing this threshold will cause the cluster to be stopped. Minimum value is 5 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). }, "masterConfig": { # The config settings for Compute Engine resources in an instance group, such as a master or worker group. # Optional. The Compute Engine config settings for the cluster's master instance. "accelerators": [ # Optional. The Compute Engine accelerator configuration for these instances. diff --git a/docs/dyn/dataproc_v1.projects.regions.workflowTemplates.html b/docs/dyn/dataproc_v1.projects.regions.workflowTemplates.html index 1688cdf82c..1031ba29f3 100644 --- a/docs/dyn/dataproc_v1.projects.regions.workflowTemplates.html +++ b/docs/dyn/dataproc_v1.projects.regions.workflowTemplates.html @@ -588,8 +588,11 @@

Method Details

"lifecycleConfig": { # Specifies the cluster auto-delete schedule configuration. # Optional. Lifecycle setting for the cluster. "autoDeleteTime": "A String", # Optional. The time when cluster will be auto-deleted (see JSON representation of Timestamp (https://developers.google.com/protocol-buffers/docs/proto3#json)). "autoDeleteTtl": "A String", # Optional. The lifetime duration of cluster. The cluster will be auto-deleted at the end of this period. Minimum value is 10 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). + "autoStopTime": "A String", # Optional. The time when cluster will be auto-stopped (see JSON representation of Timestamp (https://developers.google.com/protocol-buffers/docs/proto3#json)). + "autoStopTtl": "A String", # Optional. The lifetime duration of the cluster. The cluster will be auto-stopped at the end of this period, calculated from the time of submission of the create or update cluster request. Minimum value is 10 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). "idleDeleteTtl": "A String", # Optional. The duration to keep the cluster alive while idling (when no jobs are running). Passing this threshold will cause the cluster to be deleted. Minimum value is 5 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). "idleStartTime": "A String", # Output only. The time when cluster became idle (most recent job finished) and became eligible for deletion due to idleness (see JSON representation of Timestamp (https://developers.google.com/protocol-buffers/docs/proto3#json)). + "idleStopTtl": "A String", # Optional. The duration to keep the cluster started while idling (when no jobs are running). Passing this threshold will cause the cluster to be stopped. Minimum value is 5 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). }, "masterConfig": { # The config settings for Compute Engine resources in an instance group, such as a master or worker group. # Optional. The Compute Engine config settings for the cluster's master instance. "accelerators": [ # Optional. The Compute Engine accelerator configuration for these instances. @@ -1295,8 +1298,11 @@

Method Details

"lifecycleConfig": { # Specifies the cluster auto-delete schedule configuration. # Optional. Lifecycle setting for the cluster. "autoDeleteTime": "A String", # Optional. The time when cluster will be auto-deleted (see JSON representation of Timestamp (https://developers.google.com/protocol-buffers/docs/proto3#json)). "autoDeleteTtl": "A String", # Optional. The lifetime duration of cluster. The cluster will be auto-deleted at the end of this period. Minimum value is 10 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). + "autoStopTime": "A String", # Optional. The time when cluster will be auto-stopped (see JSON representation of Timestamp (https://developers.google.com/protocol-buffers/docs/proto3#json)). + "autoStopTtl": "A String", # Optional. The lifetime duration of the cluster. The cluster will be auto-stopped at the end of this period, calculated from the time of submission of the create or update cluster request. Minimum value is 10 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). "idleDeleteTtl": "A String", # Optional. The duration to keep the cluster alive while idling (when no jobs are running). Passing this threshold will cause the cluster to be deleted. Minimum value is 5 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). "idleStartTime": "A String", # Output only. The time when cluster became idle (most recent job finished) and became eligible for deletion due to idleness (see JSON representation of Timestamp (https://developers.google.com/protocol-buffers/docs/proto3#json)). + "idleStopTtl": "A String", # Optional. The duration to keep the cluster started while idling (when no jobs are running). Passing this threshold will cause the cluster to be stopped. Minimum value is 5 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). }, "masterConfig": { # The config settings for Compute Engine resources in an instance group, such as a master or worker group. # Optional. The Compute Engine config settings for the cluster's master instance. "accelerators": [ # Optional. The Compute Engine accelerator configuration for these instances. @@ -2029,8 +2035,11 @@

Method Details

"lifecycleConfig": { # Specifies the cluster auto-delete schedule configuration. # Optional. Lifecycle setting for the cluster. "autoDeleteTime": "A String", # Optional. The time when cluster will be auto-deleted (see JSON representation of Timestamp (https://developers.google.com/protocol-buffers/docs/proto3#json)). "autoDeleteTtl": "A String", # Optional. The lifetime duration of cluster. The cluster will be auto-deleted at the end of this period. Minimum value is 10 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). + "autoStopTime": "A String", # Optional. The time when cluster will be auto-stopped (see JSON representation of Timestamp (https://developers.google.com/protocol-buffers/docs/proto3#json)). + "autoStopTtl": "A String", # Optional. The lifetime duration of the cluster. The cluster will be auto-stopped at the end of this period, calculated from the time of submission of the create or update cluster request. Minimum value is 10 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). "idleDeleteTtl": "A String", # Optional. The duration to keep the cluster alive while idling (when no jobs are running). Passing this threshold will cause the cluster to be deleted. Minimum value is 5 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). "idleStartTime": "A String", # Output only. The time when cluster became idle (most recent job finished) and became eligible for deletion due to idleness (see JSON representation of Timestamp (https://developers.google.com/protocol-buffers/docs/proto3#json)). + "idleStopTtl": "A String", # Optional. The duration to keep the cluster started while idling (when no jobs are running). Passing this threshold will cause the cluster to be stopped. Minimum value is 5 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). }, "masterConfig": { # The config settings for Compute Engine resources in an instance group, such as a master or worker group. # Optional. The Compute Engine config settings for the cluster's master instance. "accelerators": [ # Optional. The Compute Engine accelerator configuration for these instances. @@ -2827,8 +2836,11 @@

Method Details

"lifecycleConfig": { # Specifies the cluster auto-delete schedule configuration. # Optional. Lifecycle setting for the cluster. "autoDeleteTime": "A String", # Optional. The time when cluster will be auto-deleted (see JSON representation of Timestamp (https://developers.google.com/protocol-buffers/docs/proto3#json)). "autoDeleteTtl": "A String", # Optional. The lifetime duration of cluster. The cluster will be auto-deleted at the end of this period. Minimum value is 10 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). + "autoStopTime": "A String", # Optional. The time when cluster will be auto-stopped (see JSON representation of Timestamp (https://developers.google.com/protocol-buffers/docs/proto3#json)). + "autoStopTtl": "A String", # Optional. The lifetime duration of the cluster. The cluster will be auto-stopped at the end of this period, calculated from the time of submission of the create or update cluster request. Minimum value is 10 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). "idleDeleteTtl": "A String", # Optional. The duration to keep the cluster alive while idling (when no jobs are running). Passing this threshold will cause the cluster to be deleted. Minimum value is 5 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). "idleStartTime": "A String", # Output only. The time when cluster became idle (most recent job finished) and became eligible for deletion due to idleness (see JSON representation of Timestamp (https://developers.google.com/protocol-buffers/docs/proto3#json)). + "idleStopTtl": "A String", # Optional. The duration to keep the cluster started while idling (when no jobs are running). Passing this threshold will cause the cluster to be stopped. Minimum value is 5 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). }, "masterConfig": { # The config settings for Compute Engine resources in an instance group, such as a master or worker group. # Optional. The Compute Engine config settings for the cluster's master instance. "accelerators": [ # Optional. The Compute Engine accelerator configuration for these instances. @@ -3575,8 +3587,11 @@

Method Details

"lifecycleConfig": { # Specifies the cluster auto-delete schedule configuration. # Optional. Lifecycle setting for the cluster. "autoDeleteTime": "A String", # Optional. The time when cluster will be auto-deleted (see JSON representation of Timestamp (https://developers.google.com/protocol-buffers/docs/proto3#json)). "autoDeleteTtl": "A String", # Optional. The lifetime duration of cluster. The cluster will be auto-deleted at the end of this period. Minimum value is 10 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). + "autoStopTime": "A String", # Optional. The time when cluster will be auto-stopped (see JSON representation of Timestamp (https://developers.google.com/protocol-buffers/docs/proto3#json)). + "autoStopTtl": "A String", # Optional. The lifetime duration of the cluster. The cluster will be auto-stopped at the end of this period, calculated from the time of submission of the create or update cluster request. Minimum value is 10 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). "idleDeleteTtl": "A String", # Optional. The duration to keep the cluster alive while idling (when no jobs are running). Passing this threshold will cause the cluster to be deleted. Minimum value is 5 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). "idleStartTime": "A String", # Output only. The time when cluster became idle (most recent job finished) and became eligible for deletion due to idleness (see JSON representation of Timestamp (https://developers.google.com/protocol-buffers/docs/proto3#json)). + "idleStopTtl": "A String", # Optional. The duration to keep the cluster started while idling (when no jobs are running). Passing this threshold will cause the cluster to be stopped. Minimum value is 5 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). }, "masterConfig": { # The config settings for Compute Engine resources in an instance group, such as a master or worker group. # Optional. The Compute Engine config settings for the cluster's master instance. "accelerators": [ # Optional. The Compute Engine accelerator configuration for these instances. @@ -4391,8 +4406,11 @@

Method Details

"lifecycleConfig": { # Specifies the cluster auto-delete schedule configuration. # Optional. Lifecycle setting for the cluster. "autoDeleteTime": "A String", # Optional. The time when cluster will be auto-deleted (see JSON representation of Timestamp (https://developers.google.com/protocol-buffers/docs/proto3#json)). "autoDeleteTtl": "A String", # Optional. The lifetime duration of cluster. The cluster will be auto-deleted at the end of this period. Minimum value is 10 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). + "autoStopTime": "A String", # Optional. The time when cluster will be auto-stopped (see JSON representation of Timestamp (https://developers.google.com/protocol-buffers/docs/proto3#json)). + "autoStopTtl": "A String", # Optional. The lifetime duration of the cluster. The cluster will be auto-stopped at the end of this period, calculated from the time of submission of the create or update cluster request. Minimum value is 10 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). "idleDeleteTtl": "A String", # Optional. The duration to keep the cluster alive while idling (when no jobs are running). Passing this threshold will cause the cluster to be deleted. Minimum value is 5 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). "idleStartTime": "A String", # Output only. The time when cluster became idle (most recent job finished) and became eligible for deletion due to idleness (see JSON representation of Timestamp (https://developers.google.com/protocol-buffers/docs/proto3#json)). + "idleStopTtl": "A String", # Optional. The duration to keep the cluster started while idling (when no jobs are running). Passing this threshold will cause the cluster to be stopped. Minimum value is 5 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). }, "masterConfig": { # The config settings for Compute Engine resources in an instance group, such as a master or worker group. # Optional. The Compute Engine config settings for the cluster's master instance. "accelerators": [ # Optional. The Compute Engine accelerator configuration for these instances. @@ -5098,8 +5116,11 @@

Method Details

"lifecycleConfig": { # Specifies the cluster auto-delete schedule configuration. # Optional. Lifecycle setting for the cluster. "autoDeleteTime": "A String", # Optional. The time when cluster will be auto-deleted (see JSON representation of Timestamp (https://developers.google.com/protocol-buffers/docs/proto3#json)). "autoDeleteTtl": "A String", # Optional. The lifetime duration of cluster. The cluster will be auto-deleted at the end of this period. Minimum value is 10 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). + "autoStopTime": "A String", # Optional. The time when cluster will be auto-stopped (see JSON representation of Timestamp (https://developers.google.com/protocol-buffers/docs/proto3#json)). + "autoStopTtl": "A String", # Optional. The lifetime duration of the cluster. The cluster will be auto-stopped at the end of this period, calculated from the time of submission of the create or update cluster request. Minimum value is 10 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). "idleDeleteTtl": "A String", # Optional. The duration to keep the cluster alive while idling (when no jobs are running). Passing this threshold will cause the cluster to be deleted. Minimum value is 5 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). "idleStartTime": "A String", # Output only. The time when cluster became idle (most recent job finished) and became eligible for deletion due to idleness (see JSON representation of Timestamp (https://developers.google.com/protocol-buffers/docs/proto3#json)). + "idleStopTtl": "A String", # Optional. The duration to keep the cluster started while idling (when no jobs are running). Passing this threshold will cause the cluster to be stopped. Minimum value is 5 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)). }, "masterConfig": { # The config settings for Compute Engine resources in an instance group, such as a master or worker group. # Optional. The Compute Engine config settings for the cluster's master instance. "accelerators": [ # Optional. The Compute Engine accelerator configuration for these instances. diff --git a/docs/dyn/datastream_v1.projects.locations.streams.html b/docs/dyn/datastream_v1.projects.locations.streams.html index f2214d19d9..ff79ce3dda 100644 --- a/docs/dyn/datastream_v1.projects.locations.streams.html +++ b/docs/dyn/datastream_v1.projects.locations.streams.html @@ -277,6 +277,7 @@

Method Details

"kmsKeyName": "A String", # Describes the Cloud KMS encryption key that will be used to protect destination BigQuery table. The BigQuery Service Account associated with your project requires access to this encryption key. i.e. projects/{project}/locations/{location}/keyRings/{key_ring}/cryptoKeys/{cryptoKey}. See https://cloud.google.com/bigquery/docs/customer-managed-encryption for more information. "location": "A String", # Required. The geographic location where the dataset should reside. See https://cloud.google.com/bigquery/docs/locations for supported locations. }, + "projectId": "A String", # Optional. The project id of the BigQuery dataset. If not specified, the project will be inferred from the stream resource. }, }, "destinationConnectionProfile": "A String", # Required. Destination connection profile resource. Format: `projects/{project}/locations/{location}/connectionProfiles/{name}` @@ -863,6 +864,7 @@

Method Details

"kmsKeyName": "A String", # Describes the Cloud KMS encryption key that will be used to protect destination BigQuery table. The BigQuery Service Account associated with your project requires access to this encryption key. i.e. projects/{project}/locations/{location}/keyRings/{key_ring}/cryptoKeys/{cryptoKey}. See https://cloud.google.com/bigquery/docs/customer-managed-encryption for more information. "location": "A String", # Required. The geographic location where the dataset should reside. See https://cloud.google.com/bigquery/docs/locations for supported locations. }, + "projectId": "A String", # Optional. The project id of the BigQuery dataset. If not specified, the project will be inferred from the stream resource. }, }, "destinationConnectionProfile": "A String", # Required. Destination connection profile resource. Format: `projects/{project}/locations/{location}/connectionProfiles/{name}` @@ -1388,6 +1390,7 @@

Method Details

"kmsKeyName": "A String", # Describes the Cloud KMS encryption key that will be used to protect destination BigQuery table. The BigQuery Service Account associated with your project requires access to this encryption key. i.e. projects/{project}/locations/{location}/keyRings/{key_ring}/cryptoKeys/{cryptoKey}. See https://cloud.google.com/bigquery/docs/customer-managed-encryption for more information. "location": "A String", # Required. The geographic location where the dataset should reside. See https://cloud.google.com/bigquery/docs/locations for supported locations. }, + "projectId": "A String", # Optional. The project id of the BigQuery dataset. If not specified, the project will be inferred from the stream resource. }, }, "destinationConnectionProfile": "A String", # Required. Destination connection profile resource. Format: `projects/{project}/locations/{location}/connectionProfiles/{name}` @@ -1920,6 +1923,7 @@

Method Details

"kmsKeyName": "A String", # Describes the Cloud KMS encryption key that will be used to protect destination BigQuery table. The BigQuery Service Account associated with your project requires access to this encryption key. i.e. projects/{project}/locations/{location}/keyRings/{key_ring}/cryptoKeys/{cryptoKey}. See https://cloud.google.com/bigquery/docs/customer-managed-encryption for more information. "location": "A String", # Required. The geographic location where the dataset should reside. See https://cloud.google.com/bigquery/docs/locations for supported locations. }, + "projectId": "A String", # Optional. The project id of the BigQuery dataset. If not specified, the project will be inferred from the stream resource. }, }, "destinationConnectionProfile": "A String", # Required. Destination connection profile resource. Format: `projects/{project}/locations/{location}/connectionProfiles/{name}` diff --git a/docs/dyn/dialogflow_v2.projects.conversationProfiles.html b/docs/dyn/dialogflow_v2.projects.conversationProfiles.html index 02a2b937a9..17862150d6 100644 --- a/docs/dyn/dialogflow_v2.projects.conversationProfiles.html +++ b/docs/dyn/dialogflow_v2.projects.conversationProfiles.html @@ -172,7 +172,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. - "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -233,7 +233,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. - "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -383,7 +383,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. - "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -444,7 +444,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. - "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -619,7 +619,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. - "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -680,7 +680,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. - "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -841,7 +841,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. - "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -902,7 +902,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. - "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -1071,7 +1071,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. - "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -1132,7 +1132,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. - "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -1283,7 +1283,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. - "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -1344,7 +1344,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. - "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -1487,7 +1487,7 @@

Method Details

"participantRole": "A String", # Required. The participant role to add or update the suggestion feature config. Only HUMAN_AGENT or END_USER can be used. "suggestionFeatureConfig": { # Config for suggestion features. # Required. The suggestion feature config to add or update. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. - "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. diff --git a/docs/dyn/dialogflow_v2.projects.conversations.html b/docs/dyn/dialogflow_v2.projects.conversations.html index 17bda7da25..a7f6dd8500 100644 --- a/docs/dyn/dialogflow_v2.projects.conversations.html +++ b/docs/dyn/dialogflow_v2.projects.conversations.html @@ -166,7 +166,7 @@

Method Details

"mimeType": "A String", # Optional. The mime type of the content. }, ], - "sdp": "A String", # Optional. SDP of the call. It's initially the SDP answer to the endpoint, but maybe later updated for the purpose of making the link active, etc. + "sdp": "A String", # Optional. SDP of the call. It's initially the SDP answer to the incoming call, but maybe later updated for the purpose of making the link active, etc. "sipHeaders": [ # Output only. The SIP headers from the initial SIP INVITE. { # The SIP headers from the initial SIP INVITE. "name": "A String", # Optional. The name of the header. @@ -219,7 +219,7 @@

Method Details

"mimeType": "A String", # Optional. The mime type of the content. }, ], - "sdp": "A String", # Optional. SDP of the call. It's initially the SDP answer to the endpoint, but maybe later updated for the purpose of making the link active, etc. + "sdp": "A String", # Optional. SDP of the call. It's initially the SDP answer to the incoming call, but maybe later updated for the purpose of making the link active, etc. "sipHeaders": [ # Output only. The SIP headers from the initial SIP INVITE. { # The SIP headers from the initial SIP INVITE. "name": "A String", # Optional. The name of the header. @@ -271,7 +271,7 @@

Method Details

"mimeType": "A String", # Optional. The mime type of the content. }, ], - "sdp": "A String", # Optional. SDP of the call. It's initially the SDP answer to the endpoint, but maybe later updated for the purpose of making the link active, etc. + "sdp": "A String", # Optional. SDP of the call. It's initially the SDP answer to the incoming call, but maybe later updated for the purpose of making the link active, etc. "sipHeaders": [ # Output only. The SIP headers from the initial SIP INVITE. { # The SIP headers from the initial SIP INVITE. "name": "A String", # Optional. The name of the header. @@ -329,7 +329,7 @@

Method Details

"mimeType": "A String", # Optional. The mime type of the content. }, ], - "sdp": "A String", # Optional. SDP of the call. It's initially the SDP answer to the endpoint, but maybe later updated for the purpose of making the link active, etc. + "sdp": "A String", # Optional. SDP of the call. It's initially the SDP answer to the incoming call, but maybe later updated for the purpose of making the link active, etc. "sipHeaders": [ # Output only. The SIP headers from the initial SIP INVITE. { # The SIP headers from the initial SIP INVITE. "name": "A String", # Optional. The name of the header. @@ -392,7 +392,7 @@

Method Details

"mimeType": "A String", # Optional. The mime type of the content. }, ], - "sdp": "A String", # Optional. SDP of the call. It's initially the SDP answer to the endpoint, but maybe later updated for the purpose of making the link active, etc. + "sdp": "A String", # Optional. SDP of the call. It's initially the SDP answer to the incoming call, but maybe later updated for the purpose of making the link active, etc. "sipHeaders": [ # Output only. The SIP headers from the initial SIP INVITE. { # The SIP headers from the initial SIP INVITE. "name": "A String", # Optional. The name of the header. diff --git a/docs/dyn/dialogflow_v2.projects.locations.conversationProfiles.html b/docs/dyn/dialogflow_v2.projects.locations.conversationProfiles.html index 78f8cbf24b..36d3937bac 100644 --- a/docs/dyn/dialogflow_v2.projects.locations.conversationProfiles.html +++ b/docs/dyn/dialogflow_v2.projects.locations.conversationProfiles.html @@ -172,7 +172,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. - "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -233,7 +233,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. - "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -383,7 +383,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. - "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -444,7 +444,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. - "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -619,7 +619,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. - "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -680,7 +680,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. - "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -841,7 +841,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. - "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -902,7 +902,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. - "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -1071,7 +1071,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. - "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -1132,7 +1132,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. - "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -1283,7 +1283,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. - "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -1344,7 +1344,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. - "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -1487,7 +1487,7 @@

Method Details

"participantRole": "A String", # Required. The participant role to add or update the suggestion feature config. Only HUMAN_AGENT or END_USER can be used. "suggestionFeatureConfig": { # Config for suggestion features. # Required. The suggestion feature config to add or update. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. - "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. diff --git a/docs/dyn/dialogflow_v2.projects.locations.conversations.html b/docs/dyn/dialogflow_v2.projects.locations.conversations.html index 1d8460148f..2333956650 100644 --- a/docs/dyn/dialogflow_v2.projects.locations.conversations.html +++ b/docs/dyn/dialogflow_v2.projects.locations.conversations.html @@ -169,7 +169,7 @@

Method Details

"mimeType": "A String", # Optional. The mime type of the content. }, ], - "sdp": "A String", # Optional. SDP of the call. It's initially the SDP answer to the endpoint, but maybe later updated for the purpose of making the link active, etc. + "sdp": "A String", # Optional. SDP of the call. It's initially the SDP answer to the incoming call, but maybe later updated for the purpose of making the link active, etc. "sipHeaders": [ # Output only. The SIP headers from the initial SIP INVITE. { # The SIP headers from the initial SIP INVITE. "name": "A String", # Optional. The name of the header. @@ -222,7 +222,7 @@

Method Details

"mimeType": "A String", # Optional. The mime type of the content. }, ], - "sdp": "A String", # Optional. SDP of the call. It's initially the SDP answer to the endpoint, but maybe later updated for the purpose of making the link active, etc. + "sdp": "A String", # Optional. SDP of the call. It's initially the SDP answer to the incoming call, but maybe later updated for the purpose of making the link active, etc. "sipHeaders": [ # Output only. The SIP headers from the initial SIP INVITE. { # The SIP headers from the initial SIP INVITE. "name": "A String", # Optional. The name of the header. @@ -274,7 +274,7 @@

Method Details

"mimeType": "A String", # Optional. The mime type of the content. }, ], - "sdp": "A String", # Optional. SDP of the call. It's initially the SDP answer to the endpoint, but maybe later updated for the purpose of making the link active, etc. + "sdp": "A String", # Optional. SDP of the call. It's initially the SDP answer to the incoming call, but maybe later updated for the purpose of making the link active, etc. "sipHeaders": [ # Output only. The SIP headers from the initial SIP INVITE. { # The SIP headers from the initial SIP INVITE. "name": "A String", # Optional. The name of the header. @@ -332,7 +332,7 @@

Method Details

"mimeType": "A String", # Optional. The mime type of the content. }, ], - "sdp": "A String", # Optional. SDP of the call. It's initially the SDP answer to the endpoint, but maybe later updated for the purpose of making the link active, etc. + "sdp": "A String", # Optional. SDP of the call. It's initially the SDP answer to the incoming call, but maybe later updated for the purpose of making the link active, etc. "sipHeaders": [ # Output only. The SIP headers from the initial SIP INVITE. { # The SIP headers from the initial SIP INVITE. "name": "A String", # Optional. The name of the header. @@ -447,7 +447,7 @@

Method Details

"mimeType": "A String", # Optional. The mime type of the content. }, ], - "sdp": "A String", # Optional. SDP of the call. It's initially the SDP answer to the endpoint, but maybe later updated for the purpose of making the link active, etc. + "sdp": "A String", # Optional. SDP of the call. It's initially the SDP answer to the incoming call, but maybe later updated for the purpose of making the link active, etc. "sipHeaders": [ # Output only. The SIP headers from the initial SIP INVITE. { # The SIP headers from the initial SIP INVITE. "name": "A String", # Optional. The name of the header. diff --git a/docs/dyn/dialogflow_v2.projects.locations.suggestions.html b/docs/dyn/dialogflow_v2.projects.locations.suggestions.html index 4bfab84726..d3293e05e3 100644 --- a/docs/dyn/dialogflow_v2.projects.locations.suggestions.html +++ b/docs/dyn/dialogflow_v2.projects.locations.suggestions.html @@ -112,7 +112,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. - "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -173,7 +173,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. - "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. diff --git a/docs/dyn/dialogflow_v2.projects.suggestions.html b/docs/dyn/dialogflow_v2.projects.suggestions.html index 3a40cf405f..f221770f93 100644 --- a/docs/dyn/dialogflow_v2.projects.suggestions.html +++ b/docs/dyn/dialogflow_v2.projects.suggestions.html @@ -112,7 +112,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. - "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -173,7 +173,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. - "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. diff --git a/docs/dyn/dialogflow_v2beta1.projects.agent.environments.html b/docs/dyn/dialogflow_v2beta1.projects.agent.environments.html index 317448cba2..8ac47638bf 100644 --- a/docs/dyn/dialogflow_v2beta1.projects.agent.environments.html +++ b/docs/dyn/dialogflow_v2beta1.projects.agent.environments.html @@ -170,7 +170,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -233,7 +232,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -320,7 +318,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -436,7 +433,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -517,7 +513,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -581,7 +576,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. diff --git a/docs/dyn/dialogflow_v2beta1.projects.agent.environments.users.sessions.html b/docs/dyn/dialogflow_v2beta1.projects.agent.environments.users.sessions.html index 1b1b511c5e..44ab5ca319 100644 --- a/docs/dyn/dialogflow_v2beta1.projects.agent.environments.users.sessions.html +++ b/docs/dyn/dialogflow_v2beta1.projects.agent.environments.users.sessions.html @@ -146,7 +146,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -952,7 +951,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. diff --git a/docs/dyn/dialogflow_v2beta1.projects.agent.sessions.html b/docs/dyn/dialogflow_v2beta1.projects.agent.sessions.html index e943d15265..ba38f695a5 100644 --- a/docs/dyn/dialogflow_v2beta1.projects.agent.sessions.html +++ b/docs/dyn/dialogflow_v2beta1.projects.agent.sessions.html @@ -146,7 +146,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -952,7 +951,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. diff --git a/docs/dyn/dialogflow_v2beta1.projects.conversationProfiles.html b/docs/dyn/dialogflow_v2beta1.projects.conversationProfiles.html index ffa5f7d0fb..c131c2a774 100644 --- a/docs/dyn/dialogflow_v2beta1.projects.conversationProfiles.html +++ b/docs/dyn/dialogflow_v2beta1.projects.conversationProfiles.html @@ -172,7 +172,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. - "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -233,7 +233,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. - "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -355,7 +355,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -384,7 +383,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. - "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -445,7 +444,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. - "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -567,7 +566,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -621,7 +619,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. - "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -682,7 +680,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. - "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -804,7 +802,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -844,7 +841,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. - "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -905,7 +902,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. - "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -1027,7 +1024,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -1075,7 +1071,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. - "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -1136,7 +1132,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. - "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -1258,7 +1254,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -1288,7 +1283,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. - "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -1349,7 +1344,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. - "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -1471,7 +1466,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -1493,7 +1487,7 @@

Method Details

"participantRole": "A String", # Required. The participant role to add or update the suggestion feature config. Only HUMAN_AGENT or END_USER can be used. "suggestionFeatureConfig": { # Config for suggestion features. # Required. The suggestion feature config to add or update. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. - "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. diff --git a/docs/dyn/dialogflow_v2beta1.projects.conversations.html b/docs/dyn/dialogflow_v2beta1.projects.conversations.html index 5cd7cbd506..c81cee1e61 100644 --- a/docs/dyn/dialogflow_v2beta1.projects.conversations.html +++ b/docs/dyn/dialogflow_v2beta1.projects.conversations.html @@ -166,7 +166,7 @@

Method Details

"mimeType": "A String", # Optional. The mime type of the content. }, ], - "sdp": "A String", # Optional. SDP of the call. It's initially the SDP answer to the endpoint, but maybe later updated for the purpose of making the link active, etc. + "sdp": "A String", # Optional. SDP of the call. It's initially the SDP answer to the incoming call, but maybe later updated for the purpose of making the link active, etc. "sipHeaders": [ # Output only. The SIP headers from the initial SIP INVITE. { # The SIP headers from the initial SIP INVITE. "name": "A String", # Optional. The name of the header. @@ -219,7 +219,7 @@

Method Details

"mimeType": "A String", # Optional. The mime type of the content. }, ], - "sdp": "A String", # Optional. SDP of the call. It's initially the SDP answer to the endpoint, but maybe later updated for the purpose of making the link active, etc. + "sdp": "A String", # Optional. SDP of the call. It's initially the SDP answer to the incoming call, but maybe later updated for the purpose of making the link active, etc. "sipHeaders": [ # Output only. The SIP headers from the initial SIP INVITE. { # The SIP headers from the initial SIP INVITE. "name": "A String", # Optional. The name of the header. @@ -271,7 +271,7 @@

Method Details

"mimeType": "A String", # Optional. The mime type of the content. }, ], - "sdp": "A String", # Optional. SDP of the call. It's initially the SDP answer to the endpoint, but maybe later updated for the purpose of making the link active, etc. + "sdp": "A String", # Optional. SDP of the call. It's initially the SDP answer to the incoming call, but maybe later updated for the purpose of making the link active, etc. "sipHeaders": [ # Output only. The SIP headers from the initial SIP INVITE. { # The SIP headers from the initial SIP INVITE. "name": "A String", # Optional. The name of the header. @@ -329,7 +329,7 @@

Method Details

"mimeType": "A String", # Optional. The mime type of the content. }, ], - "sdp": "A String", # Optional. SDP of the call. It's initially the SDP answer to the endpoint, but maybe later updated for the purpose of making the link active, etc. + "sdp": "A String", # Optional. SDP of the call. It's initially the SDP answer to the incoming call, but maybe later updated for the purpose of making the link active, etc. "sipHeaders": [ # Output only. The SIP headers from the initial SIP INVITE. { # The SIP headers from the initial SIP INVITE. "name": "A String", # Optional. The name of the header. @@ -392,7 +392,7 @@

Method Details

"mimeType": "A String", # Optional. The mime type of the content. }, ], - "sdp": "A String", # Optional. SDP of the call. It's initially the SDP answer to the endpoint, but maybe later updated for the purpose of making the link active, etc. + "sdp": "A String", # Optional. SDP of the call. It's initially the SDP answer to the incoming call, but maybe later updated for the purpose of making the link active, etc. "sipHeaders": [ # Output only. The SIP headers from the initial SIP INVITE. { # The SIP headers from the initial SIP INVITE. "name": "A String", # Optional. The name of the header. diff --git a/docs/dyn/dialogflow_v2beta1.projects.conversations.participants.html b/docs/dyn/dialogflow_v2beta1.projects.conversations.participants.html index 2679a73aab..61e6cf36a1 100644 --- a/docs/dyn/dialogflow_v2beta1.projects.conversations.participants.html +++ b/docs/dyn/dialogflow_v2beta1.projects.conversations.participants.html @@ -234,7 +234,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -968,7 +967,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -4758,7 +4756,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. diff --git a/docs/dyn/dialogflow_v2beta1.projects.locations.agent.environments.html b/docs/dyn/dialogflow_v2beta1.projects.locations.agent.environments.html index 602bb67deb..ffd34a6147 100644 --- a/docs/dyn/dialogflow_v2beta1.projects.locations.agent.environments.html +++ b/docs/dyn/dialogflow_v2beta1.projects.locations.agent.environments.html @@ -170,7 +170,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -233,7 +232,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -320,7 +318,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -436,7 +433,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -517,7 +513,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -581,7 +576,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. diff --git a/docs/dyn/dialogflow_v2beta1.projects.locations.agent.environments.users.sessions.html b/docs/dyn/dialogflow_v2beta1.projects.locations.agent.environments.users.sessions.html index 11e5f0018e..b6e78918dc 100644 --- a/docs/dyn/dialogflow_v2beta1.projects.locations.agent.environments.users.sessions.html +++ b/docs/dyn/dialogflow_v2beta1.projects.locations.agent.environments.users.sessions.html @@ -146,7 +146,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -952,7 +951,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. diff --git a/docs/dyn/dialogflow_v2beta1.projects.locations.agent.sessions.html b/docs/dyn/dialogflow_v2beta1.projects.locations.agent.sessions.html index f03ebd80a4..eba9089a0f 100644 --- a/docs/dyn/dialogflow_v2beta1.projects.locations.agent.sessions.html +++ b/docs/dyn/dialogflow_v2beta1.projects.locations.agent.sessions.html @@ -146,7 +146,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -952,7 +951,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. diff --git a/docs/dyn/dialogflow_v2beta1.projects.locations.conversationProfiles.html b/docs/dyn/dialogflow_v2beta1.projects.locations.conversationProfiles.html index b1f00ff085..d15e2b6283 100644 --- a/docs/dyn/dialogflow_v2beta1.projects.locations.conversationProfiles.html +++ b/docs/dyn/dialogflow_v2beta1.projects.locations.conversationProfiles.html @@ -172,7 +172,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. - "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -233,7 +233,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. - "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -355,7 +355,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -384,7 +383,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. - "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -445,7 +444,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. - "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -567,7 +566,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -621,7 +619,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. - "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -682,7 +680,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. - "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -804,7 +802,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -844,7 +841,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. - "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -905,7 +902,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. - "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -1027,7 +1024,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -1075,7 +1071,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. - "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -1136,7 +1132,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. - "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -1258,7 +1254,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -1288,7 +1283,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. - "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -1349,7 +1344,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. - "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -1471,7 +1466,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -1493,7 +1487,7 @@

Method Details

"participantRole": "A String", # Required. The participant role to add or update the suggestion feature config. Only HUMAN_AGENT or END_USER can be used. "suggestionFeatureConfig": { # Config for suggestion features. # Required. The suggestion feature config to add or update. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. - "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. diff --git a/docs/dyn/dialogflow_v2beta1.projects.locations.conversations.html b/docs/dyn/dialogflow_v2beta1.projects.locations.conversations.html index be725dd2e3..000e77a831 100644 --- a/docs/dyn/dialogflow_v2beta1.projects.locations.conversations.html +++ b/docs/dyn/dialogflow_v2beta1.projects.locations.conversations.html @@ -169,7 +169,7 @@

Method Details

"mimeType": "A String", # Optional. The mime type of the content. }, ], - "sdp": "A String", # Optional. SDP of the call. It's initially the SDP answer to the endpoint, but maybe later updated for the purpose of making the link active, etc. + "sdp": "A String", # Optional. SDP of the call. It's initially the SDP answer to the incoming call, but maybe later updated for the purpose of making the link active, etc. "sipHeaders": [ # Output only. The SIP headers from the initial SIP INVITE. { # The SIP headers from the initial SIP INVITE. "name": "A String", # Optional. The name of the header. @@ -222,7 +222,7 @@

Method Details

"mimeType": "A String", # Optional. The mime type of the content. }, ], - "sdp": "A String", # Optional. SDP of the call. It's initially the SDP answer to the endpoint, but maybe later updated for the purpose of making the link active, etc. + "sdp": "A String", # Optional. SDP of the call. It's initially the SDP answer to the incoming call, but maybe later updated for the purpose of making the link active, etc. "sipHeaders": [ # Output only. The SIP headers from the initial SIP INVITE. { # The SIP headers from the initial SIP INVITE. "name": "A String", # Optional. The name of the header. @@ -274,7 +274,7 @@

Method Details

"mimeType": "A String", # Optional. The mime type of the content. }, ], - "sdp": "A String", # Optional. SDP of the call. It's initially the SDP answer to the endpoint, but maybe later updated for the purpose of making the link active, etc. + "sdp": "A String", # Optional. SDP of the call. It's initially the SDP answer to the incoming call, but maybe later updated for the purpose of making the link active, etc. "sipHeaders": [ # Output only. The SIP headers from the initial SIP INVITE. { # The SIP headers from the initial SIP INVITE. "name": "A String", # Optional. The name of the header. @@ -332,7 +332,7 @@

Method Details

"mimeType": "A String", # Optional. The mime type of the content. }, ], - "sdp": "A String", # Optional. SDP of the call. It's initially the SDP answer to the endpoint, but maybe later updated for the purpose of making the link active, etc. + "sdp": "A String", # Optional. SDP of the call. It's initially the SDP answer to the incoming call, but maybe later updated for the purpose of making the link active, etc. "sipHeaders": [ # Output only. The SIP headers from the initial SIP INVITE. { # The SIP headers from the initial SIP INVITE. "name": "A String", # Optional. The name of the header. @@ -447,7 +447,7 @@

Method Details

"mimeType": "A String", # Optional. The mime type of the content. }, ], - "sdp": "A String", # Optional. SDP of the call. It's initially the SDP answer to the endpoint, but maybe later updated for the purpose of making the link active, etc. + "sdp": "A String", # Optional. SDP of the call. It's initially the SDP answer to the incoming call, but maybe later updated for the purpose of making the link active, etc. "sipHeaders": [ # Output only. The SIP headers from the initial SIP INVITE. { # The SIP headers from the initial SIP INVITE. "name": "A String", # Optional. The name of the header. diff --git a/docs/dyn/dialogflow_v2beta1.projects.locations.conversations.participants.html b/docs/dyn/dialogflow_v2beta1.projects.locations.conversations.participants.html index 26bab13f7b..8640d98533 100644 --- a/docs/dyn/dialogflow_v2beta1.projects.locations.conversations.participants.html +++ b/docs/dyn/dialogflow_v2beta1.projects.locations.conversations.participants.html @@ -234,7 +234,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -968,7 +967,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -4758,7 +4756,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. diff --git a/docs/dyn/dialogflow_v2beta1.projects.locations.suggestions.html b/docs/dyn/dialogflow_v2beta1.projects.locations.suggestions.html index f6b9673736..9ea0fce4b7 100644 --- a/docs/dyn/dialogflow_v2beta1.projects.locations.suggestions.html +++ b/docs/dyn/dialogflow_v2beta1.projects.locations.suggestions.html @@ -112,7 +112,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. - "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -173,7 +173,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. - "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -295,7 +295,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. diff --git a/docs/dyn/dialogflow_v2beta1.projects.suggestions.html b/docs/dyn/dialogflow_v2beta1.projects.suggestions.html index 92f329c213..3599f88dd2 100644 --- a/docs/dyn/dialogflow_v2beta1.projects.suggestions.html +++ b/docs/dyn/dialogflow_v2beta1.projects.suggestions.html @@ -112,7 +112,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. - "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -173,7 +173,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. - "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -295,7 +295,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. diff --git a/docs/dyn/dialogflow_v3.projects.locations.agents.environments.sessions.html b/docs/dyn/dialogflow_v3.projects.locations.agents.environments.sessions.html index 2ccd6a1d88..2b481aabf2 100644 --- a/docs/dyn/dialogflow_v3.projects.locations.agents.environments.sessions.html +++ b/docs/dyn/dialogflow_v3.projects.locations.agents.environments.sessions.html @@ -121,7 +121,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer substitutes a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -282,7 +281,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer substitutes a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -1479,6 +1477,9 @@

Method Details

"searchSnippet": { # Search snippet details. # Details of the snippet. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, "snippetIndex": 42, # Index of the snippet in `search_snippets` field. @@ -1503,6 +1504,9 @@

Method Details

{ # Search snippet details. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, ], @@ -1869,7 +1873,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer substitutes a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -1898,7 +1901,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer substitutes a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -3095,6 +3097,9 @@

Method Details

"searchSnippet": { # Search snippet details. # Details of the snippet. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, "snippetIndex": 42, # Index of the snippet in `search_snippets` field. @@ -3119,6 +3124,9 @@

Method Details

{ # Search snippet details. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, ], @@ -4656,7 +4664,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer substitutes a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -4817,7 +4824,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer substitutes a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -6014,6 +6020,9 @@

Method Details

"searchSnippet": { # Search snippet details. # Details of the snippet. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, "snippetIndex": 42, # Index of the snippet in `search_snippets` field. @@ -6038,6 +6047,9 @@

Method Details

{ # Search snippet details. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, ], diff --git a/docs/dyn/dialogflow_v3.projects.locations.agents.html b/docs/dyn/dialogflow_v3.projects.locations.agents.html index bdb69d66fb..df30512bbb 100644 --- a/docs/dyn/dialogflow_v3.projects.locations.agents.html +++ b/docs/dyn/dialogflow_v3.projects.locations.agents.html @@ -271,7 +271,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer substitutes a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -372,7 +371,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer substitutes a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -547,7 +545,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer substitutes a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -756,7 +753,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer substitutes a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -876,7 +872,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer substitutes a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -978,7 +973,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer substitutes a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. diff --git a/docs/dyn/dialogflow_v3.projects.locations.agents.sessions.html b/docs/dyn/dialogflow_v3.projects.locations.agents.sessions.html index e395b38e9e..835b7c4b7f 100644 --- a/docs/dyn/dialogflow_v3.projects.locations.agents.sessions.html +++ b/docs/dyn/dialogflow_v3.projects.locations.agents.sessions.html @@ -124,7 +124,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer substitutes a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -285,7 +284,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer substitutes a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -1482,6 +1480,9 @@

Method Details

"searchSnippet": { # Search snippet details. # Details of the snippet. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, "snippetIndex": 42, # Index of the snippet in `search_snippets` field. @@ -1506,6 +1507,9 @@

Method Details

{ # Search snippet details. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, ], @@ -1872,7 +1876,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer substitutes a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -1901,7 +1904,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer substitutes a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -3098,6 +3100,9 @@

Method Details

"searchSnippet": { # Search snippet details. # Details of the snippet. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, "snippetIndex": 42, # Index of the snippet in `search_snippets` field. @@ -3122,6 +3127,9 @@

Method Details

{ # Search snippet details. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, ], @@ -4659,7 +4667,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer substitutes a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -4820,7 +4827,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer substitutes a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -6017,6 +6023,9 @@

Method Details

"searchSnippet": { # Search snippet details. # Details of the snippet. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, "snippetIndex": 42, # Index of the snippet in `search_snippets` field. @@ -6041,6 +6050,9 @@

Method Details

{ # Search snippet details. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, ], diff --git a/docs/dyn/dialogflow_v3beta1.projects.locations.agents.conversations.html b/docs/dyn/dialogflow_v3beta1.projects.locations.agents.conversations.html index 7114df9757..1a7b26ff07 100644 --- a/docs/dyn/dialogflow_v3beta1.projects.locations.agents.conversations.html +++ b/docs/dyn/dialogflow_v3beta1.projects.locations.agents.conversations.html @@ -881,7 +881,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -2652,6 +2651,9 @@

Method Details

"searchSnippet": { # Search snippet details. # Details of the snippet. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, "snippetIndex": 42, # Index of the snippet in `search_snippets` field. @@ -2676,6 +2678,9 @@

Method Details

{ # Search snippet details. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, ], @@ -2837,6 +2842,9 @@

Method Details

"searchSnippet": { # Search snippet details. # Details of the snippet. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, "snippetIndex": 42, # Index of the snippet in `search_snippets` field. @@ -2861,6 +2869,9 @@

Method Details

{ # Search snippet details. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, ], @@ -3099,7 +3110,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -3253,7 +3263,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -5024,6 +5033,9 @@

Method Details

"searchSnippet": { # Search snippet details. # Details of the snippet. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, "snippetIndex": 42, # Index of the snippet in `search_snippets` field. @@ -5048,6 +5060,9 @@

Method Details

{ # Search snippet details. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, ], @@ -5209,6 +5224,9 @@

Method Details

"searchSnippet": { # Search snippet details. # Details of the snippet. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, "snippetIndex": 42, # Index of the snippet in `search_snippets` field. @@ -5233,6 +5251,9 @@

Method Details

{ # Search snippet details. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, ], @@ -7406,7 +7427,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -9177,6 +9197,9 @@

Method Details

"searchSnippet": { # Search snippet details. # Details of the snippet. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, "snippetIndex": 42, # Index of the snippet in `search_snippets` field. @@ -9201,6 +9224,9 @@

Method Details

{ # Search snippet details. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, ], @@ -9362,6 +9388,9 @@

Method Details

"searchSnippet": { # Search snippet details. # Details of the snippet. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, "snippetIndex": 42, # Index of the snippet in `search_snippets` field. @@ -9386,6 +9415,9 @@

Method Details

{ # Search snippet details. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, ], @@ -9624,7 +9656,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -9778,7 +9809,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -11549,6 +11579,9 @@

Method Details

"searchSnippet": { # Search snippet details. # Details of the snippet. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, "snippetIndex": 42, # Index of the snippet in `search_snippets` field. @@ -11573,6 +11606,9 @@

Method Details

{ # Search snippet details. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, ], @@ -11734,6 +11770,9 @@

Method Details

"searchSnippet": { # Search snippet details. # Details of the snippet. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, "snippetIndex": 42, # Index of the snippet in `search_snippets` field. @@ -11758,6 +11797,9 @@

Method Details

{ # Search snippet details. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, ], diff --git a/docs/dyn/dialogflow_v3beta1.projects.locations.agents.environments.sessions.html b/docs/dyn/dialogflow_v3beta1.projects.locations.agents.environments.sessions.html index 8b67ed4c45..3fd4c5da75 100644 --- a/docs/dyn/dialogflow_v3beta1.projects.locations.agents.environments.sessions.html +++ b/docs/dyn/dialogflow_v3beta1.projects.locations.agents.environments.sessions.html @@ -121,7 +121,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -283,7 +282,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -2054,6 +2052,9 @@

Method Details

"searchSnippet": { # Search snippet details. # Details of the snippet. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, "snippetIndex": 42, # Index of the snippet in `search_snippets` field. @@ -2078,6 +2079,9 @@

Method Details

{ # Search snippet details. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, ], @@ -2239,6 +2243,9 @@

Method Details

"searchSnippet": { # Search snippet details. # Details of the snippet. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, "snippetIndex": 42, # Index of the snippet in `search_snippets` field. @@ -2263,6 +2270,9 @@

Method Details

{ # Search snippet details. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, ], @@ -2684,7 +2694,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -2713,7 +2722,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -4484,6 +4492,9 @@

Method Details

"searchSnippet": { # Search snippet details. # Details of the snippet. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, "snippetIndex": 42, # Index of the snippet in `search_snippets` field. @@ -4508,6 +4519,9 @@

Method Details

{ # Search snippet details. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, ], @@ -4669,6 +4683,9 @@

Method Details

"searchSnippet": { # Search snippet details. # Details of the snippet. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, "snippetIndex": 42, # Index of the snippet in `search_snippets` field. @@ -4693,6 +4710,9 @@

Method Details

{ # Search snippet details. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, ], @@ -6273,7 +6293,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -6435,7 +6454,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -8206,6 +8224,9 @@

Method Details

"searchSnippet": { # Search snippet details. # Details of the snippet. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, "snippetIndex": 42, # Index of the snippet in `search_snippets` field. @@ -8230,6 +8251,9 @@

Method Details

{ # Search snippet details. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, ], @@ -8391,6 +8415,9 @@

Method Details

"searchSnippet": { # Search snippet details. # Details of the snippet. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, "snippetIndex": 42, # Index of the snippet in `search_snippets` field. @@ -8415,6 +8442,9 @@

Method Details

{ # Search snippet details. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, ], diff --git a/docs/dyn/dialogflow_v3beta1.projects.locations.agents.html b/docs/dyn/dialogflow_v3beta1.projects.locations.agents.html index 16acd291a2..81643fe6be 100644 --- a/docs/dyn/dialogflow_v3beta1.projects.locations.agents.html +++ b/docs/dyn/dialogflow_v3beta1.projects.locations.agents.html @@ -289,7 +289,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -403,7 +402,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -591,7 +589,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -813,7 +810,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -946,7 +942,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -1061,7 +1056,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. diff --git a/docs/dyn/dialogflow_v3beta1.projects.locations.agents.playbooks.examples.html b/docs/dyn/dialogflow_v3beta1.projects.locations.agents.playbooks.examples.html index 51329aeaac..5a5968f9e7 100644 --- a/docs/dyn/dialogflow_v3beta1.projects.locations.agents.playbooks.examples.html +++ b/docs/dyn/dialogflow_v3beta1.projects.locations.agents.playbooks.examples.html @@ -259,6 +259,9 @@

Method Details

"searchSnippet": { # Search snippet details. # Details of the snippet. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, "snippetIndex": 42, # Index of the snippet in `search_snippets` field. @@ -283,6 +286,9 @@

Method Details

{ # Search snippet details. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, ], @@ -492,6 +498,9 @@

Method Details

"searchSnippet": { # Search snippet details. # Details of the snippet. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, "snippetIndex": 42, # Index of the snippet in `search_snippets` field. @@ -516,6 +525,9 @@

Method Details

{ # Search snippet details. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, ], @@ -750,6 +762,9 @@

Method Details

"searchSnippet": { # Search snippet details. # Details of the snippet. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, "snippetIndex": 42, # Index of the snippet in `search_snippets` field. @@ -774,6 +789,9 @@

Method Details

{ # Search snippet details. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, ], @@ -995,6 +1013,9 @@

Method Details

"searchSnippet": { # Search snippet details. # Details of the snippet. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, "snippetIndex": 42, # Index of the snippet in `search_snippets` field. @@ -1019,6 +1040,9 @@

Method Details

{ # Search snippet details. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, ], @@ -1247,6 +1271,9 @@

Method Details

"searchSnippet": { # Search snippet details. # Details of the snippet. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, "snippetIndex": 42, # Index of the snippet in `search_snippets` field. @@ -1271,6 +1298,9 @@

Method Details

{ # Search snippet details. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, ], @@ -1481,6 +1511,9 @@

Method Details

"searchSnippet": { # Search snippet details. # Details of the snippet. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, "snippetIndex": 42, # Index of the snippet in `search_snippets` field. @@ -1505,6 +1538,9 @@

Method Details

{ # Search snippet details. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, ], diff --git a/docs/dyn/dialogflow_v3beta1.projects.locations.agents.playbooks.versions.html b/docs/dyn/dialogflow_v3beta1.projects.locations.agents.playbooks.versions.html index f7ee394cb0..1284fd20ed 100644 --- a/docs/dyn/dialogflow_v3beta1.projects.locations.agents.playbooks.versions.html +++ b/docs/dyn/dialogflow_v3beta1.projects.locations.agents.playbooks.versions.html @@ -262,6 +262,9 @@

Method Details

"searchSnippet": { # Search snippet details. # Details of the snippet. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, "snippetIndex": 42, # Index of the snippet in `search_snippets` field. @@ -286,6 +289,9 @@

Method Details

{ # Search snippet details. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, ], @@ -929,6 +935,9 @@

Method Details

"searchSnippet": { # Search snippet details. # Details of the snippet. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, "snippetIndex": 42, # Index of the snippet in `search_snippets` field. @@ -953,6 +962,9 @@

Method Details

{ # Search snippet details. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, ], @@ -1621,6 +1633,9 @@

Method Details

"searchSnippet": { # Search snippet details. # Details of the snippet. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, "snippetIndex": 42, # Index of the snippet in `search_snippets` field. @@ -1645,6 +1660,9 @@

Method Details

{ # Search snippet details. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, ], @@ -2300,6 +2318,9 @@

Method Details

"searchSnippet": { # Search snippet details. # Details of the snippet. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, "snippetIndex": 42, # Index of the snippet in `search_snippets` field. @@ -2324,6 +2345,9 @@

Method Details

{ # Search snippet details. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, ], diff --git a/docs/dyn/dialogflow_v3beta1.projects.locations.agents.sessions.html b/docs/dyn/dialogflow_v3beta1.projects.locations.agents.sessions.html index 369de72b3d..8d1c4169ef 100644 --- a/docs/dyn/dialogflow_v3beta1.projects.locations.agents.sessions.html +++ b/docs/dyn/dialogflow_v3beta1.projects.locations.agents.sessions.html @@ -124,7 +124,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -286,7 +285,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -2057,6 +2055,9 @@

Method Details

"searchSnippet": { # Search snippet details. # Details of the snippet. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, "snippetIndex": 42, # Index of the snippet in `search_snippets` field. @@ -2081,6 +2082,9 @@

Method Details

{ # Search snippet details. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, ], @@ -2242,6 +2246,9 @@

Method Details

"searchSnippet": { # Search snippet details. # Details of the snippet. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, "snippetIndex": 42, # Index of the snippet in `search_snippets` field. @@ -2266,6 +2273,9 @@

Method Details

{ # Search snippet details. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, ], @@ -2687,7 +2697,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -2716,7 +2725,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -4487,6 +4495,9 @@

Method Details

"searchSnippet": { # Search snippet details. # Details of the snippet. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, "snippetIndex": 42, # Index of the snippet in `search_snippets` field. @@ -4511,6 +4522,9 @@

Method Details

{ # Search snippet details. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, ], @@ -4672,6 +4686,9 @@

Method Details

"searchSnippet": { # Search snippet details. # Details of the snippet. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, "snippetIndex": 42, # Index of the snippet in `search_snippets` field. @@ -4696,6 +4713,9 @@

Method Details

{ # Search snippet details. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, ], @@ -6276,7 +6296,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -6438,7 +6457,6 @@

Method Details

"speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error. "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio. "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices). - "secondLanguageModel": "A String", # Optional. The name of the second voice (non-English) if language detection is supported. "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request. }, "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. @@ -8209,6 +8227,9 @@

Method Details

"searchSnippet": { # Search snippet details. # Details of the snippet. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, "snippetIndex": 42, # Index of the snippet in `search_snippets` field. @@ -8233,6 +8254,9 @@

Method Details

{ # Search snippet details. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, ], @@ -8394,6 +8418,9 @@

Method Details

"searchSnippet": { # Search snippet details. # Details of the snippet. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, "snippetIndex": 42, # Index of the snippet in `search_snippets` field. @@ -8418,6 +8445,9 @@

Method Details

{ # Search snippet details. "documentTitle": "A String", # Title of the enclosing document. "documentUri": "A String", # Uri for the document. Present if specified for the document. + "metadata": { # Metadata associated with the document. + "a_key": "", # Properties of the object. + }, "text": "A String", # Text included in the prompt. }, ], diff --git a/docs/dyn/discoveryengine_v1alpha.projects.locations.collections.dataConnector.connectorRuns.html b/docs/dyn/discoveryengine_v1alpha.projects.locations.collections.dataConnector.connectorRuns.html index 51978eb6d7..28a06179b7 100644 --- a/docs/dyn/discoveryengine_v1alpha.projects.locations.collections.dataConnector.connectorRuns.html +++ b/docs/dyn/discoveryengine_v1alpha.projects.locations.collections.dataConnector.connectorRuns.html @@ -111,9 +111,9 @@

Method Details

"endTime": "A String", # Output only. The time when the connector run ended. "entityRuns": [ # Output only. The details of the entities synced at the ConnectorRun. Each ConnectorRun consists of syncing one or more entities. { # Represents an entity that was synced in this ConnectorRun. - "deletedRecordCount": "A String", # The number of documents deleted. + "deletedRecordCount": "A String", # Optional. The number of documents deleted. "entityName": "A String", # The name of the source entity. - "errorRecordCount": "A String", # The total number of documents failed at sync at any stage (extraction, indexing, etc). + "errorRecordCount": "A String", # Optional. The total number of documents failed at sync at indexing stage. "errors": [ # The errors from the entity's sync run. Only exist if running into an error state. Contains error code and error message. { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). "code": 42, # The status code, which should be an enum value of google.rpc.Code. @@ -125,15 +125,15 @@

Method Details

"message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client. }, ], - "extractedRecordCount": "A String", # The number of documents extracted from connector source, ready to be ingested to VAIS. - "indexedRecordCount": "A String", # The number of documents indexed. + "extractedRecordCount": "A String", # Optional. The number of documents extracted from connector source, ready to be ingested to VAIS. + "indexedRecordCount": "A String", # Optional. The number of documents indexed. "progress": { # Represents the progress of a sync run. # Metadata to generate the progress bar. "currentCount": "A String", # The current progress. "percentile": 3.14, # Derived. The percentile of the progress.current_count / total_count. The value is between [0, 1.0] inclusive. "totalCount": "A String", # The total. }, - "scheduledRecordCount": "A String", # The number of documents scheduled to be crawled/extracted from connector source. This only applies to third party connectors. - "sourceApiRequestCount": "A String", # The number of requests sent to 3p API. + "scheduledRecordCount": "A String", # Optional. The number of documents scheduled to be crawled/extracted from connector source. This only applies to third party connectors. + "sourceApiRequestCount": "A String", # Optional. The number of requests sent to 3p API. "state": "A String", # The state of the entity's sync run. "stateUpdateTime": "A String", # Timestamp at which the entity sync state was last updated. "statsUpdateTime": "A String", # The timestamp for either extracted_documents_count, indexed_documents_count and error_documents_count was last updated. diff --git a/docs/dyn/discoveryengine_v1alpha.projects.locations.collections.dataConnector.html b/docs/dyn/discoveryengine_v1alpha.projects.locations.collections.dataConnector.html index 75300dc43f..3592a75245 100644 --- a/docs/dyn/discoveryengine_v1alpha.projects.locations.collections.dataConnector.html +++ b/docs/dyn/discoveryengine_v1alpha.projects.locations.collections.dataConnector.html @@ -156,9 +156,9 @@

Method Details

"endTime": "A String", # Output only. The time when the connector run ended. "entityRuns": [ # Output only. The details of the entities synced at the ConnectorRun. Each ConnectorRun consists of syncing one or more entities. { # Represents an entity that was synced in this ConnectorRun. - "deletedRecordCount": "A String", # The number of documents deleted. + "deletedRecordCount": "A String", # Optional. The number of documents deleted. "entityName": "A String", # The name of the source entity. - "errorRecordCount": "A String", # The total number of documents failed at sync at any stage (extraction, indexing, etc). + "errorRecordCount": "A String", # Optional. The total number of documents failed at sync at indexing stage. "errors": [ # The errors from the entity's sync run. Only exist if running into an error state. Contains error code and error message. { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). "code": 42, # The status code, which should be an enum value of google.rpc.Code. @@ -170,15 +170,15 @@

Method Details

"message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client. }, ], - "extractedRecordCount": "A String", # The number of documents extracted from connector source, ready to be ingested to VAIS. - "indexedRecordCount": "A String", # The number of documents indexed. + "extractedRecordCount": "A String", # Optional. The number of documents extracted from connector source, ready to be ingested to VAIS. + "indexedRecordCount": "A String", # Optional. The number of documents indexed. "progress": { # Represents the progress of a sync run. # Metadata to generate the progress bar. "currentCount": "A String", # The current progress. "percentile": 3.14, # Derived. The percentile of the progress.current_count / total_count. The value is between [0, 1.0] inclusive. "totalCount": "A String", # The total. }, - "scheduledRecordCount": "A String", # The number of documents scheduled to be crawled/extracted from connector source. This only applies to third party connectors. - "sourceApiRequestCount": "A String", # The number of requests sent to 3p API. + "scheduledRecordCount": "A String", # Optional. The number of documents scheduled to be crawled/extracted from connector source. This only applies to third party connectors. + "sourceApiRequestCount": "A String", # Optional. The number of requests sent to 3p API. "state": "A String", # The state of the entity's sync run. "stateUpdateTime": "A String", # Timestamp at which the entity sync state was last updated. "statsUpdateTime": "A String", # The timestamp for either extracted_documents_count, indexed_documents_count and error_documents_count was last updated. diff --git a/docs/dyn/discoveryengine_v1alpha.projects.locations.collections.dataStores.widgetConfigs.html b/docs/dyn/discoveryengine_v1alpha.projects.locations.collections.dataStores.widgetConfigs.html index 45721b2e05..1cbe6130fa 100644 --- a/docs/dyn/discoveryengine_v1alpha.projects.locations.collections.dataStores.widgetConfigs.html +++ b/docs/dyn/discoveryengine_v1alpha.projects.locations.collections.dataStores.widgetConfigs.html @@ -123,6 +123,7 @@

Method Details

{ # Read-only collection component that contains data store collections fields that may be used for filtering "connectorIconLink": "A String", # Output only. The icon link of the connector source. "dataSource": "A String", # The name of the data source, retrieved from `Collection.data_connector.data_source`. + "dataSourceDisplayName": "A String", # Output only. The display name of the data source. "dataStoreComponents": [ # For the data store collection, list of the children data stores. { # Read-only data store component that contains data stores fields that may be used for filtering, it's the child of `CollectionComponent`. "contentConfig": "A String", # Output only. The content config of the data store. diff --git a/docs/dyn/discoveryengine_v1alpha.projects.locations.collections.engines.widgetConfigs.html b/docs/dyn/discoveryengine_v1alpha.projects.locations.collections.engines.widgetConfigs.html index e49da1de62..04860b1660 100644 --- a/docs/dyn/discoveryengine_v1alpha.projects.locations.collections.engines.widgetConfigs.html +++ b/docs/dyn/discoveryengine_v1alpha.projects.locations.collections.engines.widgetConfigs.html @@ -123,6 +123,7 @@

Method Details

{ # Read-only collection component that contains data store collections fields that may be used for filtering "connectorIconLink": "A String", # Output only. The icon link of the connector source. "dataSource": "A String", # The name of the data source, retrieved from `Collection.data_connector.data_source`. + "dataSourceDisplayName": "A String", # Output only. The display name of the data source. "dataStoreComponents": [ # For the data store collection, list of the children data stores. { # Read-only data store component that contains data stores fields that may be used for filtering, it's the child of `CollectionComponent`. "contentConfig": "A String", # Output only. The content config of the data store. diff --git a/docs/dyn/discoveryengine_v1alpha.projects.locations.collections.html b/docs/dyn/discoveryengine_v1alpha.projects.locations.collections.html index 09696cb597..8ff3f943a9 100644 --- a/docs/dyn/discoveryengine_v1alpha.projects.locations.collections.html +++ b/docs/dyn/discoveryengine_v1alpha.projects.locations.collections.html @@ -233,6 +233,11 @@

Method Details

"authParams": { # Optional. Any authentication parameters specific to EUA connectors. "a_key": "", # Properties of the object. }, + "tenant": { # Tenant information for a connector source. This includes some of the same information stored in the Credential message, but is limited to only what is needed to provide a list of accessible tenants to the user. # Optional. The tenant project the connector is connected to. + "displayName": "A String", # Optional display name for the tenant, e.g. "My Slack Team". + "id": "A String", # The tenant's instance ID. Examples: Jira ("8594f221-9797-5f78-1fa4-485e198d7cd0"), Slack ("T123456"). + "uri": "A String", # The URI of the tenant, if applicable. For example, the URI of a Jira instance is https://my-jira-instance.atlassian.net, and a Slack tenant does not have a URI. + }, }, "entities": [ # List of entities from the connected data source to ingest. { # Represents an entity in the data source. For example, the `Account` object in Salesforce. @@ -421,6 +426,11 @@

Method Details

"authParams": { # Optional. Any authentication parameters specific to EUA connectors. "a_key": "", # Properties of the object. }, + "tenant": { # Tenant information for a connector source. This includes some of the same information stored in the Credential message, but is limited to only what is needed to provide a list of accessible tenants to the user. # Optional. The tenant project the connector is connected to. + "displayName": "A String", # Optional display name for the tenant, e.g. "My Slack Team". + "id": "A String", # The tenant's instance ID. Examples: Jira ("8594f221-9797-5f78-1fa4-485e198d7cd0"), Slack ("T123456"). + "uri": "A String", # The URI of the tenant, if applicable. For example, the URI of a Jira instance is https://my-jira-instance.atlassian.net, and a Slack tenant does not have a URI. + }, }, "entities": [ # List of entities from the connected data source to ingest. { # Represents an entity in the data source. For example, the `Account` object in Salesforce. @@ -613,6 +623,11 @@

Method Details

"authParams": { # Optional. Any authentication parameters specific to EUA connectors. "a_key": "", # Properties of the object. }, + "tenant": { # Tenant information for a connector source. This includes some of the same information stored in the Credential message, but is limited to only what is needed to provide a list of accessible tenants to the user. # Optional. The tenant project the connector is connected to. + "displayName": "A String", # Optional display name for the tenant, e.g. "My Slack Team". + "id": "A String", # The tenant's instance ID. Examples: Jira ("8594f221-9797-5f78-1fa4-485e198d7cd0"), Slack ("T123456"). + "uri": "A String", # The URI of the tenant, if applicable. For example, the URI of a Jira instance is https://my-jira-instance.atlassian.net, and a Slack tenant does not have a URI. + }, }, "entities": [ # List of entities from the connected data source to ingest. { # Represents an entity in the data source. For example, the `Account` object in Salesforce. @@ -815,6 +830,11 @@

Method Details

"authParams": { # Optional. Any authentication parameters specific to EUA connectors. "a_key": "", # Properties of the object. }, + "tenant": { # Tenant information for a connector source. This includes some of the same information stored in the Credential message, but is limited to only what is needed to provide a list of accessible tenants to the user. # Optional. The tenant project the connector is connected to. + "displayName": "A String", # Optional display name for the tenant, e.g. "My Slack Team". + "id": "A String", # The tenant's instance ID. Examples: Jira ("8594f221-9797-5f78-1fa4-485e198d7cd0"), Slack ("T123456"). + "uri": "A String", # The URI of the tenant, if applicable. For example, the URI of a Jira instance is https://my-jira-instance.atlassian.net, and a Slack tenant does not have a URI. + }, }, "entities": [ # List of entities from the connected data source to ingest. { # Represents an entity in the data source. For example, the `Account` object in Salesforce. @@ -1027,6 +1047,11 @@

Method Details

"authParams": { # Optional. Any authentication parameters specific to EUA connectors. "a_key": "", # Properties of the object. }, + "tenant": { # Tenant information for a connector source. This includes some of the same information stored in the Credential message, but is limited to only what is needed to provide a list of accessible tenants to the user. # Optional. The tenant project the connector is connected to. + "displayName": "A String", # Optional display name for the tenant, e.g. "My Slack Team". + "id": "A String", # The tenant's instance ID. Examples: Jira ("8594f221-9797-5f78-1fa4-485e198d7cd0"), Slack ("T123456"). + "uri": "A String", # The URI of the tenant, if applicable. For example, the URI of a Jira instance is https://my-jira-instance.atlassian.net, and a Slack tenant does not have a URI. + }, }, "entities": [ # List of entities from the connected data source to ingest. { # Represents an entity in the data source. For example, the `Account` object in Salesforce. @@ -1206,6 +1231,11 @@

Method Details

"authParams": { # Optional. Any authentication parameters specific to EUA connectors. "a_key": "", # Properties of the object. }, + "tenant": { # Tenant information for a connector source. This includes some of the same information stored in the Credential message, but is limited to only what is needed to provide a list of accessible tenants to the user. # Optional. The tenant project the connector is connected to. + "displayName": "A String", # Optional display name for the tenant, e.g. "My Slack Team". + "id": "A String", # The tenant's instance ID. Examples: Jira ("8594f221-9797-5f78-1fa4-485e198d7cd0"), Slack ("T123456"). + "uri": "A String", # The URI of the tenant, if applicable. For example, the URI of a Jira instance is https://my-jira-instance.atlassian.net, and a Slack tenant does not have a URI. + }, }, "entities": [ # List of entities from the connected data source to ingest. { # Represents an entity in the data source. For example, the `Account` object in Salesforce. diff --git a/docs/dyn/discoveryengine_v1alpha.projects.locations.dataStores.widgetConfigs.html b/docs/dyn/discoveryengine_v1alpha.projects.locations.dataStores.widgetConfigs.html index 1205d33d80..0f9508a629 100644 --- a/docs/dyn/discoveryengine_v1alpha.projects.locations.dataStores.widgetConfigs.html +++ b/docs/dyn/discoveryengine_v1alpha.projects.locations.dataStores.widgetConfigs.html @@ -123,6 +123,7 @@

Method Details

{ # Read-only collection component that contains data store collections fields that may be used for filtering "connectorIconLink": "A String", # Output only. The icon link of the connector source. "dataSource": "A String", # The name of the data source, retrieved from `Collection.data_connector.data_source`. + "dataSourceDisplayName": "A String", # Output only. The display name of the data source. "dataStoreComponents": [ # For the data store collection, list of the children data stores. { # Read-only data store component that contains data stores fields that may be used for filtering, it's the child of `CollectionComponent`. "contentConfig": "A String", # Output only. The content config of the data store. diff --git a/docs/dyn/discoveryengine_v1alpha.projects.locations.html b/docs/dyn/discoveryengine_v1alpha.projects.locations.html index c5ed493eea..b08280f8cf 100644 --- a/docs/dyn/discoveryengine_v1alpha.projects.locations.html +++ b/docs/dyn/discoveryengine_v1alpha.projects.locations.html @@ -993,6 +993,11 @@

Method Details

"authParams": { # Optional. Any authentication parameters specific to EUA connectors. "a_key": "", # Properties of the object. }, + "tenant": { # Tenant information for a connector source. This includes some of the same information stored in the Credential message, but is limited to only what is needed to provide a list of accessible tenants to the user. # Optional. The tenant project the connector is connected to. + "displayName": "A String", # Optional display name for the tenant, e.g. "My Slack Team". + "id": "A String", # The tenant's instance ID. Examples: Jira ("8594f221-9797-5f78-1fa4-485e198d7cd0"), Slack ("T123456"). + "uri": "A String", # The URI of the tenant, if applicable. For example, the URI of a Jira instance is https://my-jira-instance.atlassian.net, and a Slack tenant does not have a URI. + }, }, "entities": [ # List of entities from the connected data source to ingest. { # Represents an entity in the data source. For example, the `Account` object in Salesforce. diff --git a/docs/dyn/dlp_v2.organizations.locations.fileStoreDataProfiles.html b/docs/dyn/dlp_v2.organizations.locations.fileStoreDataProfiles.html index da04fd617c..707bc022f1 100644 --- a/docs/dyn/dlp_v2.organizations.locations.fileStoreDataProfiles.html +++ b/docs/dyn/dlp_v2.organizations.locations.fileStoreDataProfiles.html @@ -724,6 +724,14 @@

Method Details

"dataStorageLocations": [ # For resources that have multiple storage locations, these are those regions. For Cloud Storage this is the list of regions chosen for dual-region storage. `file_store_location` will normally be the corresponding multi-region for the list of individual locations. The first region is always picked as the processing and storage location for the data profile. "A String", ], + "domains": [ # Domains associated with the profile. + { # A domain represents a thematic category that a data profile can fall under. + "category": "A String", # A domain category that this profile is related to. + "signals": [ # The collection of signals that influenced selection of the category. + "A String", + ], + }, + ], "fileClusterSummaries": [ # FileClusterSummary per each cluster. { # The file cluster summary. "dataRiskLevel": { # Score is a summary of all elements in the data profile. A higher number means more risk. # The data risk level of this cluster. RISK_LOW if nothing has been scanned. @@ -1475,6 +1483,14 @@

Method Details

"dataStorageLocations": [ # For resources that have multiple storage locations, these are those regions. For Cloud Storage this is the list of regions chosen for dual-region storage. `file_store_location` will normally be the corresponding multi-region for the list of individual locations. The first region is always picked as the processing and storage location for the data profile. "A String", ], + "domains": [ # Domains associated with the profile. + { # A domain represents a thematic category that a data profile can fall under. + "category": "A String", # A domain category that this profile is related to. + "signals": [ # The collection of signals that influenced selection of the category. + "A String", + ], + }, + ], "fileClusterSummaries": [ # FileClusterSummary per each cluster. { # The file cluster summary. "dataRiskLevel": { # Score is a summary of all elements in the data profile. A higher number means more risk. # The data risk level of this cluster. RISK_LOW if nothing has been scanned. diff --git a/docs/dyn/dlp_v2.organizations.locations.tableDataProfiles.html b/docs/dyn/dlp_v2.organizations.locations.tableDataProfiles.html index c4edb171d6..5705c48d4c 100644 --- a/docs/dyn/dlp_v2.organizations.locations.tableDataProfiles.html +++ b/docs/dyn/dlp_v2.organizations.locations.tableDataProfiles.html @@ -724,6 +724,14 @@

Method Details

"datasetId": "A String", # If the resource is BigQuery, the dataset ID. "datasetLocation": "A String", # If supported, the location where the dataset's data is stored. See https://cloud.google.com/bigquery/docs/locations for supported locations. "datasetProjectId": "A String", # The Google Cloud project ID that owns the resource. + "domains": [ # Domains associated with the profile. + { # A domain represents a thematic category that a data profile can fall under. + "category": "A String", # A domain category that this profile is related to. + "signals": [ # The collection of signals that influenced selection of the category. + "A String", + ], + }, + ], "encryptionStatus": "A String", # How the table is encrypted. "expirationTime": "A String", # Optional. The time when this table expires. "failedColumnCount": "A String", # The number of columns skipped in the table because of an error. @@ -788,7 +796,7 @@

Method Details

"sensitivityScore": { # Score is calculated from of all elements in the data profile. A higher level means the data is more sensitive. # The sensitivity score of this table. "score": "A String", # The sensitivity score applied to the resource. }, - "state": "A String", # State of a profile. + "state": "A String", # State of a profile. This will always be set to DONE when the table data profile is written to another service like BigQuery or Pub/Sub. "tableId": "A String", # The table ID. "tableSizeBytes": "A String", # The size of the table when the profile was generated. "tags": [ # The tags attached to the table, including any tags attached during profiling. Because tags are attached to Cloud SQL instances rather than Cloud SQL tables, this field is empty for Cloud SQL table profiles. @@ -1419,6 +1427,14 @@

Method Details

"datasetId": "A String", # If the resource is BigQuery, the dataset ID. "datasetLocation": "A String", # If supported, the location where the dataset's data is stored. See https://cloud.google.com/bigquery/docs/locations for supported locations. "datasetProjectId": "A String", # The Google Cloud project ID that owns the resource. + "domains": [ # Domains associated with the profile. + { # A domain represents a thematic category that a data profile can fall under. + "category": "A String", # A domain category that this profile is related to. + "signals": [ # The collection of signals that influenced selection of the category. + "A String", + ], + }, + ], "encryptionStatus": "A String", # How the table is encrypted. "expirationTime": "A String", # Optional. The time when this table expires. "failedColumnCount": "A String", # The number of columns skipped in the table because of an error. @@ -1483,7 +1499,7 @@

Method Details

"sensitivityScore": { # Score is calculated from of all elements in the data profile. A higher level means the data is more sensitive. # The sensitivity score of this table. "score": "A String", # The sensitivity score applied to the resource. }, - "state": "A String", # State of a profile. + "state": "A String", # State of a profile. This will always be set to DONE when the table data profile is written to another service like BigQuery or Pub/Sub. "tableId": "A String", # The table ID. "tableSizeBytes": "A String", # The size of the table when the profile was generated. "tags": [ # The tags attached to the table, including any tags attached during profiling. Because tags are attached to Cloud SQL instances rather than Cloud SQL tables, this field is empty for Cloud SQL table profiles. diff --git a/docs/dyn/dlp_v2.projects.locations.fileStoreDataProfiles.html b/docs/dyn/dlp_v2.projects.locations.fileStoreDataProfiles.html index d1ba581d4c..a1fadf510c 100644 --- a/docs/dyn/dlp_v2.projects.locations.fileStoreDataProfiles.html +++ b/docs/dyn/dlp_v2.projects.locations.fileStoreDataProfiles.html @@ -724,6 +724,14 @@

Method Details

"dataStorageLocations": [ # For resources that have multiple storage locations, these are those regions. For Cloud Storage this is the list of regions chosen for dual-region storage. `file_store_location` will normally be the corresponding multi-region for the list of individual locations. The first region is always picked as the processing and storage location for the data profile. "A String", ], + "domains": [ # Domains associated with the profile. + { # A domain represents a thematic category that a data profile can fall under. + "category": "A String", # A domain category that this profile is related to. + "signals": [ # The collection of signals that influenced selection of the category. + "A String", + ], + }, + ], "fileClusterSummaries": [ # FileClusterSummary per each cluster. { # The file cluster summary. "dataRiskLevel": { # Score is a summary of all elements in the data profile. A higher number means more risk. # The data risk level of this cluster. RISK_LOW if nothing has been scanned. @@ -1475,6 +1483,14 @@

Method Details

"dataStorageLocations": [ # For resources that have multiple storage locations, these are those regions. For Cloud Storage this is the list of regions chosen for dual-region storage. `file_store_location` will normally be the corresponding multi-region for the list of individual locations. The first region is always picked as the processing and storage location for the data profile. "A String", ], + "domains": [ # Domains associated with the profile. + { # A domain represents a thematic category that a data profile can fall under. + "category": "A String", # A domain category that this profile is related to. + "signals": [ # The collection of signals that influenced selection of the category. + "A String", + ], + }, + ], "fileClusterSummaries": [ # FileClusterSummary per each cluster. { # The file cluster summary. "dataRiskLevel": { # Score is a summary of all elements in the data profile. A higher number means more risk. # The data risk level of this cluster. RISK_LOW if nothing has been scanned. diff --git a/docs/dyn/dlp_v2.projects.locations.tableDataProfiles.html b/docs/dyn/dlp_v2.projects.locations.tableDataProfiles.html index 21016dc106..fd77919619 100644 --- a/docs/dyn/dlp_v2.projects.locations.tableDataProfiles.html +++ b/docs/dyn/dlp_v2.projects.locations.tableDataProfiles.html @@ -724,6 +724,14 @@

Method Details

"datasetId": "A String", # If the resource is BigQuery, the dataset ID. "datasetLocation": "A String", # If supported, the location where the dataset's data is stored. See https://cloud.google.com/bigquery/docs/locations for supported locations. "datasetProjectId": "A String", # The Google Cloud project ID that owns the resource. + "domains": [ # Domains associated with the profile. + { # A domain represents a thematic category that a data profile can fall under. + "category": "A String", # A domain category that this profile is related to. + "signals": [ # The collection of signals that influenced selection of the category. + "A String", + ], + }, + ], "encryptionStatus": "A String", # How the table is encrypted. "expirationTime": "A String", # Optional. The time when this table expires. "failedColumnCount": "A String", # The number of columns skipped in the table because of an error. @@ -788,7 +796,7 @@

Method Details

"sensitivityScore": { # Score is calculated from of all elements in the data profile. A higher level means the data is more sensitive. # The sensitivity score of this table. "score": "A String", # The sensitivity score applied to the resource. }, - "state": "A String", # State of a profile. + "state": "A String", # State of a profile. This will always be set to DONE when the table data profile is written to another service like BigQuery or Pub/Sub. "tableId": "A String", # The table ID. "tableSizeBytes": "A String", # The size of the table when the profile was generated. "tags": [ # The tags attached to the table, including any tags attached during profiling. Because tags are attached to Cloud SQL instances rather than Cloud SQL tables, this field is empty for Cloud SQL table profiles. @@ -1419,6 +1427,14 @@

Method Details

"datasetId": "A String", # If the resource is BigQuery, the dataset ID. "datasetLocation": "A String", # If supported, the location where the dataset's data is stored. See https://cloud.google.com/bigquery/docs/locations for supported locations. "datasetProjectId": "A String", # The Google Cloud project ID that owns the resource. + "domains": [ # Domains associated with the profile. + { # A domain represents a thematic category that a data profile can fall under. + "category": "A String", # A domain category that this profile is related to. + "signals": [ # The collection of signals that influenced selection of the category. + "A String", + ], + }, + ], "encryptionStatus": "A String", # How the table is encrypted. "expirationTime": "A String", # Optional. The time when this table expires. "failedColumnCount": "A String", # The number of columns skipped in the table because of an error. @@ -1483,7 +1499,7 @@

Method Details

"sensitivityScore": { # Score is calculated from of all elements in the data profile. A higher level means the data is more sensitive. # The sensitivity score of this table. "score": "A String", # The sensitivity score applied to the resource. }, - "state": "A String", # State of a profile. + "state": "A String", # State of a profile. This will always be set to DONE when the table data profile is written to another service like BigQuery or Pub/Sub. "tableId": "A String", # The table ID. "tableSizeBytes": "A String", # The size of the table when the profile was generated. "tags": [ # The tags attached to the table, including any tags attached during profiling. Because tags are attached to Cloud SQL instances rather than Cloud SQL tables, this field is empty for Cloud SQL table profiles. diff --git a/docs/dyn/drive_v3.changes.html b/docs/dyn/drive_v3.changes.html index d9fc0d633d..1d158b677d 100644 --- a/docs/dyn/drive_v3.changes.html +++ b/docs/dyn/drive_v3.changes.html @@ -161,6 +161,7 @@

Method Details

"canAddChildren": True or False, # Output only. Whether the current user can add children to folders in this shared drive. "canChangeCopyRequiresWriterPermissionRestriction": True or False, # Output only. Whether the current user can change the `copyRequiresWriterPermission` restriction of this shared drive. "canChangeDomainUsersOnlyRestriction": True or False, # Output only. Whether the current user can change the `domainUsersOnly` restriction of this shared drive. + "canChangeDownloadRestriction": True or False, # Output only. Whether the current user can change organizer-applied download restrictions of this shared drive. "canChangeDriveBackground": True or False, # Output only. Whether the current user can change the background of this shared drive. "canChangeDriveMembersOnlyRestriction": True or False, # Output only. Whether the current user can change the `driveMembersOnly` restriction of this shared drive. "canChangeSharingFoldersRequiresOrganizerPermissionRestriction": True or False, # Output only. Whether the current user can change the `sharingFoldersRequiresOrganizerPermission` restriction of this shared drive. @@ -211,6 +212,7 @@

Method Details

"canAddFolderFromAnotherDrive": True or False, # Output only. Whether the current user can add a folder from another drive (different shared drive or My Drive) to this folder. This is false when the item is not a folder. Only populated for items in shared drives. "canAddMyDriveParent": True or False, # Output only. Whether the current user can add a parent for the item without removing an existing parent in the same request. Not populated for shared drive files. "canChangeCopyRequiresWriterPermission": True or False, # Output only. Whether the current user can change the `copyRequiresWriterPermission` restriction of this file. + "canChangeItemDownloadRestriction": True or False, # Output only. Whether the current user can change the owner-applied download restrictions of the file. "canChangeSecurityUpdateEnabled": True or False, # Output only. Whether the current user can change the securityUpdateEnabled field on link share metadata. "canChangeViewersCanCopyContent": True or False, # Deprecated: Output only. "canComment": True or False, # Output only. Whether the current user can comment on this file. @@ -278,6 +280,16 @@

Method Details

"copyRequiresWriterPermission": True or False, # Whether the options to copy, print, or download this file, should be disabled for readers and commenters. "createdTime": "A String", # The time at which the file was created (RFC 3339 date-time). "description": "A String", # A short description of the file. + "downloadRestrictions": { # Download restrictions applied to the file. # Download restrictions applied on the file. + "effectiveDownloadRestrictionWithContext": { # A restriction for copy and download of the file. # The effective download restriction applied to this file. This considers all restriction settings and DLP rules. + "restrictedForReaders": True or False, # Whether download and copy is restricted for readers. + "restrictedForWriters": True or False, # Whether download and copy is restricted for writers. If true, download is also restricted for readers. + }, + "itemDownloadRestriction": { # A restriction for copy and download of the file. # The download restriction of the file applied directly by the owner or organizer. This does not take into account shared drive settings or DLP rules. + "restrictedForReaders": True or False, # Whether download and copy is restricted for readers. + "restrictedForWriters": True or False, # Whether download and copy is restricted for writers. If true, download is also restricted for readers. + }, + }, "driveId": "A String", # Output only. ID of the shared drive the file resides in. Only populated for items in shared drives. "explicitlyTrashed": True or False, # Output only. Whether the file has been explicitly trashed, as opposed to recursively trashed from a parent folder. "exportLinks": { # Output only. Links for exporting Docs Editors files to specific formats. @@ -498,6 +510,7 @@

Method Details

"canAddChildren": True or False, # Whether the current user can add children to folders in this Team Drive. "canChangeCopyRequiresWriterPermissionRestriction": True or False, # Whether the current user can change the `copyRequiresWriterPermission` restriction of this Team Drive. "canChangeDomainUsersOnlyRestriction": True or False, # Whether the current user can change the `domainUsersOnly` restriction of this Team Drive. + "canChangeDownloadRestriction": True or False, # Whether the current user can change organizer-applied download restrictions of this shared drive. "canChangeSharingFoldersRequiresOrganizerPermissionRestriction": True or False, # Whether the current user can change the `sharingFoldersRequiresOrganizerPermission` restriction of this Team Drive. "canChangeTeamDriveBackground": True or False, # Whether the current user can change the background of this Team Drive. "canChangeTeamMembersOnlyRestriction": True or False, # Whether the current user can change the `teamMembersOnly` restriction of this Team Drive. diff --git a/docs/dyn/drive_v3.drives.html b/docs/dyn/drive_v3.drives.html index 05fecb5d5b..5cc1038eb6 100644 --- a/docs/dyn/drive_v3.drives.html +++ b/docs/dyn/drive_v3.drives.html @@ -128,6 +128,7 @@

Method Details

"canAddChildren": True or False, # Output only. Whether the current user can add children to folders in this shared drive. "canChangeCopyRequiresWriterPermissionRestriction": True or False, # Output only. Whether the current user can change the `copyRequiresWriterPermission` restriction of this shared drive. "canChangeDomainUsersOnlyRestriction": True or False, # Output only. Whether the current user can change the `domainUsersOnly` restriction of this shared drive. + "canChangeDownloadRestriction": True or False, # Output only. Whether the current user can change organizer-applied download restrictions of this shared drive. "canChangeDriveBackground": True or False, # Output only. Whether the current user can change the background of this shared drive. "canChangeDriveMembersOnlyRestriction": True or False, # Output only. Whether the current user can change the `driveMembersOnly` restriction of this shared drive. "canChangeSharingFoldersRequiresOrganizerPermissionRestriction": True or False, # Output only. Whether the current user can change the `sharingFoldersRequiresOrganizerPermission` restriction of this shared drive. @@ -187,6 +188,7 @@

Method Details

"canAddChildren": True or False, # Output only. Whether the current user can add children to folders in this shared drive. "canChangeCopyRequiresWriterPermissionRestriction": True or False, # Output only. Whether the current user can change the `copyRequiresWriterPermission` restriction of this shared drive. "canChangeDomainUsersOnlyRestriction": True or False, # Output only. Whether the current user can change the `domainUsersOnly` restriction of this shared drive. + "canChangeDownloadRestriction": True or False, # Output only. Whether the current user can change organizer-applied download restrictions of this shared drive. "canChangeDriveBackground": True or False, # Output only. Whether the current user can change the background of this shared drive. "canChangeDriveMembersOnlyRestriction": True or False, # Output only. Whether the current user can change the `driveMembersOnly` restriction of this shared drive. "canChangeSharingFoldersRequiresOrganizerPermissionRestriction": True or False, # Output only. Whether the current user can change the `sharingFoldersRequiresOrganizerPermission` restriction of this shared drive. @@ -269,6 +271,7 @@

Method Details

"canAddChildren": True or False, # Output only. Whether the current user can add children to folders in this shared drive. "canChangeCopyRequiresWriterPermissionRestriction": True or False, # Output only. Whether the current user can change the `copyRequiresWriterPermission` restriction of this shared drive. "canChangeDomainUsersOnlyRestriction": True or False, # Output only. Whether the current user can change the `domainUsersOnly` restriction of this shared drive. + "canChangeDownloadRestriction": True or False, # Output only. Whether the current user can change organizer-applied download restrictions of this shared drive. "canChangeDriveBackground": True or False, # Output only. Whether the current user can change the background of this shared drive. "canChangeDriveMembersOnlyRestriction": True or False, # Output only. Whether the current user can change the `driveMembersOnly` restriction of this shared drive. "canChangeSharingFoldersRequiresOrganizerPermissionRestriction": True or False, # Output only. Whether the current user can change the `sharingFoldersRequiresOrganizerPermission` restriction of this shared drive. @@ -335,6 +338,7 @@

Method Details

"canAddChildren": True or False, # Output only. Whether the current user can add children to folders in this shared drive. "canChangeCopyRequiresWriterPermissionRestriction": True or False, # Output only. Whether the current user can change the `copyRequiresWriterPermission` restriction of this shared drive. "canChangeDomainUsersOnlyRestriction": True or False, # Output only. Whether the current user can change the `domainUsersOnly` restriction of this shared drive. + "canChangeDownloadRestriction": True or False, # Output only. Whether the current user can change organizer-applied download restrictions of this shared drive. "canChangeDriveBackground": True or False, # Output only. Whether the current user can change the background of this shared drive. "canChangeDriveMembersOnlyRestriction": True or False, # Output only. Whether the current user can change the `driveMembersOnly` restriction of this shared drive. "canChangeSharingFoldersRequiresOrganizerPermissionRestriction": True or False, # Output only. Whether the current user can change the `sharingFoldersRequiresOrganizerPermission` restriction of this shared drive. @@ -406,6 +410,7 @@

Method Details

"canAddChildren": True or False, # Output only. Whether the current user can add children to folders in this shared drive. "canChangeCopyRequiresWriterPermissionRestriction": True or False, # Output only. Whether the current user can change the `copyRequiresWriterPermission` restriction of this shared drive. "canChangeDomainUsersOnlyRestriction": True or False, # Output only. Whether the current user can change the `domainUsersOnly` restriction of this shared drive. + "canChangeDownloadRestriction": True or False, # Output only. Whether the current user can change organizer-applied download restrictions of this shared drive. "canChangeDriveBackground": True or False, # Output only. Whether the current user can change the background of this shared drive. "canChangeDriveMembersOnlyRestriction": True or False, # Output only. Whether the current user can change the `driveMembersOnly` restriction of this shared drive. "canChangeSharingFoldersRequiresOrganizerPermissionRestriction": True or False, # Output only. Whether the current user can change the `sharingFoldersRequiresOrganizerPermission` restriction of this shared drive. @@ -490,6 +495,7 @@

Method Details

"canAddChildren": True or False, # Output only. Whether the current user can add children to folders in this shared drive. "canChangeCopyRequiresWriterPermissionRestriction": True or False, # Output only. Whether the current user can change the `copyRequiresWriterPermission` restriction of this shared drive. "canChangeDomainUsersOnlyRestriction": True or False, # Output only. Whether the current user can change the `domainUsersOnly` restriction of this shared drive. + "canChangeDownloadRestriction": True or False, # Output only. Whether the current user can change organizer-applied download restrictions of this shared drive. "canChangeDriveBackground": True or False, # Output only. Whether the current user can change the background of this shared drive. "canChangeDriveMembersOnlyRestriction": True or False, # Output only. Whether the current user can change the `driveMembersOnly` restriction of this shared drive. "canChangeSharingFoldersRequiresOrganizerPermissionRestriction": True or False, # Output only. Whether the current user can change the `sharingFoldersRequiresOrganizerPermission` restriction of this shared drive. @@ -551,6 +557,7 @@

Method Details

"canAddChildren": True or False, # Output only. Whether the current user can add children to folders in this shared drive. "canChangeCopyRequiresWriterPermissionRestriction": True or False, # Output only. Whether the current user can change the `copyRequiresWriterPermission` restriction of this shared drive. "canChangeDomainUsersOnlyRestriction": True or False, # Output only. Whether the current user can change the `domainUsersOnly` restriction of this shared drive. + "canChangeDownloadRestriction": True or False, # Output only. Whether the current user can change organizer-applied download restrictions of this shared drive. "canChangeDriveBackground": True or False, # Output only. Whether the current user can change the background of this shared drive. "canChangeDriveMembersOnlyRestriction": True or False, # Output only. Whether the current user can change the `driveMembersOnly` restriction of this shared drive. "canChangeSharingFoldersRequiresOrganizerPermissionRestriction": True or False, # Output only. Whether the current user can change the `sharingFoldersRequiresOrganizerPermission` restriction of this shared drive. @@ -611,6 +618,7 @@

Method Details

"canAddChildren": True or False, # Output only. Whether the current user can add children to folders in this shared drive. "canChangeCopyRequiresWriterPermissionRestriction": True or False, # Output only. Whether the current user can change the `copyRequiresWriterPermission` restriction of this shared drive. "canChangeDomainUsersOnlyRestriction": True or False, # Output only. Whether the current user can change the `domainUsersOnly` restriction of this shared drive. + "canChangeDownloadRestriction": True or False, # Output only. Whether the current user can change organizer-applied download restrictions of this shared drive. "canChangeDriveBackground": True or False, # Output only. Whether the current user can change the background of this shared drive. "canChangeDriveMembersOnlyRestriction": True or False, # Output only. Whether the current user can change the `driveMembersOnly` restriction of this shared drive. "canChangeSharingFoldersRequiresOrganizerPermissionRestriction": True or False, # Output only. Whether the current user can change the `sharingFoldersRequiresOrganizerPermission` restriction of this shared drive. diff --git a/docs/dyn/drive_v3.files.html b/docs/dyn/drive_v3.files.html index 825c08e620..a3ca58b888 100644 --- a/docs/dyn/drive_v3.files.html +++ b/docs/dyn/drive_v3.files.html @@ -154,6 +154,7 @@

Method Details

"canAddFolderFromAnotherDrive": True or False, # Output only. Whether the current user can add a folder from another drive (different shared drive or My Drive) to this folder. This is false when the item is not a folder. Only populated for items in shared drives. "canAddMyDriveParent": True or False, # Output only. Whether the current user can add a parent for the item without removing an existing parent in the same request. Not populated for shared drive files. "canChangeCopyRequiresWriterPermission": True or False, # Output only. Whether the current user can change the `copyRequiresWriterPermission` restriction of this file. + "canChangeItemDownloadRestriction": True or False, # Output only. Whether the current user can change the owner-applied download restrictions of the file. "canChangeSecurityUpdateEnabled": True or False, # Output only. Whether the current user can change the securityUpdateEnabled field on link share metadata. "canChangeViewersCanCopyContent": True or False, # Deprecated: Output only. "canComment": True or False, # Output only. Whether the current user can comment on this file. @@ -221,6 +222,16 @@

Method Details

"copyRequiresWriterPermission": True or False, # Whether the options to copy, print, or download this file, should be disabled for readers and commenters. "createdTime": "A String", # The time at which the file was created (RFC 3339 date-time). "description": "A String", # A short description of the file. + "downloadRestrictions": { # Download restrictions applied to the file. # Download restrictions applied on the file. + "effectiveDownloadRestrictionWithContext": { # A restriction for copy and download of the file. # The effective download restriction applied to this file. This considers all restriction settings and DLP rules. + "restrictedForReaders": True or False, # Whether download and copy is restricted for readers. + "restrictedForWriters": True or False, # Whether download and copy is restricted for writers. If true, download is also restricted for readers. + }, + "itemDownloadRestriction": { # A restriction for copy and download of the file. # The download restriction of the file applied directly by the owner or organizer. This does not take into account shared drive settings or DLP rules. + "restrictedForReaders": True or False, # Whether download and copy is restricted for readers. + "restrictedForWriters": True or False, # Whether download and copy is restricted for writers. If true, download is also restricted for readers. + }, + }, "driveId": "A String", # Output only. ID of the shared drive the file resides in. Only populated for items in shared drives. "explicitlyTrashed": True or False, # Output only. Whether the file has been explicitly trashed, as opposed to recursively trashed from a parent folder. "exportLinks": { # Output only. Links for exporting Docs Editors files to specific formats. @@ -454,6 +465,7 @@

Method Details

"canAddFolderFromAnotherDrive": True or False, # Output only. Whether the current user can add a folder from another drive (different shared drive or My Drive) to this folder. This is false when the item is not a folder. Only populated for items in shared drives. "canAddMyDriveParent": True or False, # Output only. Whether the current user can add a parent for the item without removing an existing parent in the same request. Not populated for shared drive files. "canChangeCopyRequiresWriterPermission": True or False, # Output only. Whether the current user can change the `copyRequiresWriterPermission` restriction of this file. + "canChangeItemDownloadRestriction": True or False, # Output only. Whether the current user can change the owner-applied download restrictions of the file. "canChangeSecurityUpdateEnabled": True or False, # Output only. Whether the current user can change the securityUpdateEnabled field on link share metadata. "canChangeViewersCanCopyContent": True or False, # Deprecated: Output only. "canComment": True or False, # Output only. Whether the current user can comment on this file. @@ -521,6 +533,16 @@

Method Details

"copyRequiresWriterPermission": True or False, # Whether the options to copy, print, or download this file, should be disabled for readers and commenters. "createdTime": "A String", # The time at which the file was created (RFC 3339 date-time). "description": "A String", # A short description of the file. + "downloadRestrictions": { # Download restrictions applied to the file. # Download restrictions applied on the file. + "effectiveDownloadRestrictionWithContext": { # A restriction for copy and download of the file. # The effective download restriction applied to this file. This considers all restriction settings and DLP rules. + "restrictedForReaders": True or False, # Whether download and copy is restricted for readers. + "restrictedForWriters": True or False, # Whether download and copy is restricted for writers. If true, download is also restricted for readers. + }, + "itemDownloadRestriction": { # A restriction for copy and download of the file. # The download restriction of the file applied directly by the owner or organizer. This does not take into account shared drive settings or DLP rules. + "restrictedForReaders": True or False, # Whether download and copy is restricted for readers. + "restrictedForWriters": True or False, # Whether download and copy is restricted for writers. If true, download is also restricted for readers. + }, + }, "driveId": "A String", # Output only. ID of the shared drive the file resides in. Only populated for items in shared drives. "explicitlyTrashed": True or False, # Output only. Whether the file has been explicitly trashed, as opposed to recursively trashed from a parent folder. "exportLinks": { # Output only. Links for exporting Docs Editors files to specific formats. @@ -747,6 +769,7 @@

Method Details

"canAddFolderFromAnotherDrive": True or False, # Output only. Whether the current user can add a folder from another drive (different shared drive or My Drive) to this folder. This is false when the item is not a folder. Only populated for items in shared drives. "canAddMyDriveParent": True or False, # Output only. Whether the current user can add a parent for the item without removing an existing parent in the same request. Not populated for shared drive files. "canChangeCopyRequiresWriterPermission": True or False, # Output only. Whether the current user can change the `copyRequiresWriterPermission` restriction of this file. + "canChangeItemDownloadRestriction": True or False, # Output only. Whether the current user can change the owner-applied download restrictions of the file. "canChangeSecurityUpdateEnabled": True or False, # Output only. Whether the current user can change the securityUpdateEnabled field on link share metadata. "canChangeViewersCanCopyContent": True or False, # Deprecated: Output only. "canComment": True or False, # Output only. Whether the current user can comment on this file. @@ -814,6 +837,16 @@

Method Details

"copyRequiresWriterPermission": True or False, # Whether the options to copy, print, or download this file, should be disabled for readers and commenters. "createdTime": "A String", # The time at which the file was created (RFC 3339 date-time). "description": "A String", # A short description of the file. + "downloadRestrictions": { # Download restrictions applied to the file. # Download restrictions applied on the file. + "effectiveDownloadRestrictionWithContext": { # A restriction for copy and download of the file. # The effective download restriction applied to this file. This considers all restriction settings and DLP rules. + "restrictedForReaders": True or False, # Whether download and copy is restricted for readers. + "restrictedForWriters": True or False, # Whether download and copy is restricted for writers. If true, download is also restricted for readers. + }, + "itemDownloadRestriction": { # A restriction for copy and download of the file. # The download restriction of the file applied directly by the owner or organizer. This does not take into account shared drive settings or DLP rules. + "restrictedForReaders": True or False, # Whether download and copy is restricted for readers. + "restrictedForWriters": True or False, # Whether download and copy is restricted for writers. If true, download is also restricted for readers. + }, + }, "driveId": "A String", # Output only. ID of the shared drive the file resides in. Only populated for items in shared drives. "explicitlyTrashed": True or False, # Output only. Whether the file has been explicitly trashed, as opposed to recursively trashed from a parent folder. "exportLinks": { # Output only. Links for exporting Docs Editors files to specific formats. @@ -1050,6 +1083,7 @@

Method Details

"canAddFolderFromAnotherDrive": True or False, # Output only. Whether the current user can add a folder from another drive (different shared drive or My Drive) to this folder. This is false when the item is not a folder. Only populated for items in shared drives. "canAddMyDriveParent": True or False, # Output only. Whether the current user can add a parent for the item without removing an existing parent in the same request. Not populated for shared drive files. "canChangeCopyRequiresWriterPermission": True or False, # Output only. Whether the current user can change the `copyRequiresWriterPermission` restriction of this file. + "canChangeItemDownloadRestriction": True or False, # Output only. Whether the current user can change the owner-applied download restrictions of the file. "canChangeSecurityUpdateEnabled": True or False, # Output only. Whether the current user can change the securityUpdateEnabled field on link share metadata. "canChangeViewersCanCopyContent": True or False, # Deprecated: Output only. "canComment": True or False, # Output only. Whether the current user can comment on this file. @@ -1117,6 +1151,16 @@

Method Details

"copyRequiresWriterPermission": True or False, # Whether the options to copy, print, or download this file, should be disabled for readers and commenters. "createdTime": "A String", # The time at which the file was created (RFC 3339 date-time). "description": "A String", # A short description of the file. + "downloadRestrictions": { # Download restrictions applied to the file. # Download restrictions applied on the file. + "effectiveDownloadRestrictionWithContext": { # A restriction for copy and download of the file. # The effective download restriction applied to this file. This considers all restriction settings and DLP rules. + "restrictedForReaders": True or False, # Whether download and copy is restricted for readers. + "restrictedForWriters": True or False, # Whether download and copy is restricted for writers. If true, download is also restricted for readers. + }, + "itemDownloadRestriction": { # A restriction for copy and download of the file. # The download restriction of the file applied directly by the owner or organizer. This does not take into account shared drive settings or DLP rules. + "restrictedForReaders": True or False, # Whether download and copy is restricted for readers. + "restrictedForWriters": True or False, # Whether download and copy is restricted for writers. If true, download is also restricted for readers. + }, + }, "driveId": "A String", # Output only. ID of the shared drive the file resides in. Only populated for items in shared drives. "explicitlyTrashed": True or False, # Output only. Whether the file has been explicitly trashed, as opposed to recursively trashed from a parent folder. "exportLinks": { # Output only. Links for exporting Docs Editors files to specific formats. @@ -1474,6 +1518,7 @@

Method Details

"canAddFolderFromAnotherDrive": True or False, # Output only. Whether the current user can add a folder from another drive (different shared drive or My Drive) to this folder. This is false when the item is not a folder. Only populated for items in shared drives. "canAddMyDriveParent": True or False, # Output only. Whether the current user can add a parent for the item without removing an existing parent in the same request. Not populated for shared drive files. "canChangeCopyRequiresWriterPermission": True or False, # Output only. Whether the current user can change the `copyRequiresWriterPermission` restriction of this file. + "canChangeItemDownloadRestriction": True or False, # Output only. Whether the current user can change the owner-applied download restrictions of the file. "canChangeSecurityUpdateEnabled": True or False, # Output only. Whether the current user can change the securityUpdateEnabled field on link share metadata. "canChangeViewersCanCopyContent": True or False, # Deprecated: Output only. "canComment": True or False, # Output only. Whether the current user can comment on this file. @@ -1541,6 +1586,16 @@

Method Details

"copyRequiresWriterPermission": True or False, # Whether the options to copy, print, or download this file, should be disabled for readers and commenters. "createdTime": "A String", # The time at which the file was created (RFC 3339 date-time). "description": "A String", # A short description of the file. + "downloadRestrictions": { # Download restrictions applied to the file. # Download restrictions applied on the file. + "effectiveDownloadRestrictionWithContext": { # A restriction for copy and download of the file. # The effective download restriction applied to this file. This considers all restriction settings and DLP rules. + "restrictedForReaders": True or False, # Whether download and copy is restricted for readers. + "restrictedForWriters": True or False, # Whether download and copy is restricted for writers. If true, download is also restricted for readers. + }, + "itemDownloadRestriction": { # A restriction for copy and download of the file. # The download restriction of the file applied directly by the owner or organizer. This does not take into account shared drive settings or DLP rules. + "restrictedForReaders": True or False, # Whether download and copy is restricted for readers. + "restrictedForWriters": True or False, # Whether download and copy is restricted for writers. If true, download is also restricted for readers. + }, + }, "driveId": "A String", # Output only. ID of the shared drive the file resides in. Only populated for items in shared drives. "explicitlyTrashed": True or False, # Output only. Whether the file has been explicitly trashed, as opposed to recursively trashed from a parent folder. "exportLinks": { # Output only. Links for exporting Docs Editors files to specific formats. @@ -1814,6 +1869,7 @@

Method Details

"canAddFolderFromAnotherDrive": True or False, # Output only. Whether the current user can add a folder from another drive (different shared drive or My Drive) to this folder. This is false when the item is not a folder. Only populated for items in shared drives. "canAddMyDriveParent": True or False, # Output only. Whether the current user can add a parent for the item without removing an existing parent in the same request. Not populated for shared drive files. "canChangeCopyRequiresWriterPermission": True or False, # Output only. Whether the current user can change the `copyRequiresWriterPermission` restriction of this file. + "canChangeItemDownloadRestriction": True or False, # Output only. Whether the current user can change the owner-applied download restrictions of the file. "canChangeSecurityUpdateEnabled": True or False, # Output only. Whether the current user can change the securityUpdateEnabled field on link share metadata. "canChangeViewersCanCopyContent": True or False, # Deprecated: Output only. "canComment": True or False, # Output only. Whether the current user can comment on this file. @@ -1881,6 +1937,16 @@

Method Details

"copyRequiresWriterPermission": True or False, # Whether the options to copy, print, or download this file, should be disabled for readers and commenters. "createdTime": "A String", # The time at which the file was created (RFC 3339 date-time). "description": "A String", # A short description of the file. + "downloadRestrictions": { # Download restrictions applied to the file. # Download restrictions applied on the file. + "effectiveDownloadRestrictionWithContext": { # A restriction for copy and download of the file. # The effective download restriction applied to this file. This considers all restriction settings and DLP rules. + "restrictedForReaders": True or False, # Whether download and copy is restricted for readers. + "restrictedForWriters": True or False, # Whether download and copy is restricted for writers. If true, download is also restricted for readers. + }, + "itemDownloadRestriction": { # A restriction for copy and download of the file. # The download restriction of the file applied directly by the owner or organizer. This does not take into account shared drive settings or DLP rules. + "restrictedForReaders": True or False, # Whether download and copy is restricted for readers. + "restrictedForWriters": True or False, # Whether download and copy is restricted for writers. If true, download is also restricted for readers. + }, + }, "driveId": "A String", # Output only. ID of the shared drive the file resides in. Only populated for items in shared drives. "explicitlyTrashed": True or False, # Output only. Whether the file has been explicitly trashed, as opposed to recursively trashed from a parent folder. "exportLinks": { # Output only. Links for exporting Docs Editors files to specific formats. @@ -2290,6 +2356,7 @@

Method Details

"canAddFolderFromAnotherDrive": True or False, # Output only. Whether the current user can add a folder from another drive (different shared drive or My Drive) to this folder. This is false when the item is not a folder. Only populated for items in shared drives. "canAddMyDriveParent": True or False, # Output only. Whether the current user can add a parent for the item without removing an existing parent in the same request. Not populated for shared drive files. "canChangeCopyRequiresWriterPermission": True or False, # Output only. Whether the current user can change the `copyRequiresWriterPermission` restriction of this file. + "canChangeItemDownloadRestriction": True or False, # Output only. Whether the current user can change the owner-applied download restrictions of the file. "canChangeSecurityUpdateEnabled": True or False, # Output only. Whether the current user can change the securityUpdateEnabled field on link share metadata. "canChangeViewersCanCopyContent": True or False, # Deprecated: Output only. "canComment": True or False, # Output only. Whether the current user can comment on this file. @@ -2357,6 +2424,16 @@

Method Details

"copyRequiresWriterPermission": True or False, # Whether the options to copy, print, or download this file, should be disabled for readers and commenters. "createdTime": "A String", # The time at which the file was created (RFC 3339 date-time). "description": "A String", # A short description of the file. + "downloadRestrictions": { # Download restrictions applied to the file. # Download restrictions applied on the file. + "effectiveDownloadRestrictionWithContext": { # A restriction for copy and download of the file. # The effective download restriction applied to this file. This considers all restriction settings and DLP rules. + "restrictedForReaders": True or False, # Whether download and copy is restricted for readers. + "restrictedForWriters": True or False, # Whether download and copy is restricted for writers. If true, download is also restricted for readers. + }, + "itemDownloadRestriction": { # A restriction for copy and download of the file. # The download restriction of the file applied directly by the owner or organizer. This does not take into account shared drive settings or DLP rules. + "restrictedForReaders": True or False, # Whether download and copy is restricted for readers. + "restrictedForWriters": True or False, # Whether download and copy is restricted for writers. If true, download is also restricted for readers. + }, + }, "driveId": "A String", # Output only. ID of the shared drive the file resides in. Only populated for items in shared drives. "explicitlyTrashed": True or False, # Output only. Whether the file has been explicitly trashed, as opposed to recursively trashed from a parent folder. "exportLinks": { # Output only. Links for exporting Docs Editors files to specific formats. @@ -2594,6 +2671,7 @@

Method Details

"canAddFolderFromAnotherDrive": True or False, # Output only. Whether the current user can add a folder from another drive (different shared drive or My Drive) to this folder. This is false when the item is not a folder. Only populated for items in shared drives. "canAddMyDriveParent": True or False, # Output only. Whether the current user can add a parent for the item without removing an existing parent in the same request. Not populated for shared drive files. "canChangeCopyRequiresWriterPermission": True or False, # Output only. Whether the current user can change the `copyRequiresWriterPermission` restriction of this file. + "canChangeItemDownloadRestriction": True or False, # Output only. Whether the current user can change the owner-applied download restrictions of the file. "canChangeSecurityUpdateEnabled": True or False, # Output only. Whether the current user can change the securityUpdateEnabled field on link share metadata. "canChangeViewersCanCopyContent": True or False, # Deprecated: Output only. "canComment": True or False, # Output only. Whether the current user can comment on this file. @@ -2661,6 +2739,16 @@

Method Details

"copyRequiresWriterPermission": True or False, # Whether the options to copy, print, or download this file, should be disabled for readers and commenters. "createdTime": "A String", # The time at which the file was created (RFC 3339 date-time). "description": "A String", # A short description of the file. + "downloadRestrictions": { # Download restrictions applied to the file. # Download restrictions applied on the file. + "effectiveDownloadRestrictionWithContext": { # A restriction for copy and download of the file. # The effective download restriction applied to this file. This considers all restriction settings and DLP rules. + "restrictedForReaders": True or False, # Whether download and copy is restricted for readers. + "restrictedForWriters": True or False, # Whether download and copy is restricted for writers. If true, download is also restricted for readers. + }, + "itemDownloadRestriction": { # A restriction for copy and download of the file. # The download restriction of the file applied directly by the owner or organizer. This does not take into account shared drive settings or DLP rules. + "restrictedForReaders": True or False, # Whether download and copy is restricted for readers. + "restrictedForWriters": True or False, # Whether download and copy is restricted for writers. If true, download is also restricted for readers. + }, + }, "driveId": "A String", # Output only. ID of the shared drive the file resides in. Only populated for items in shared drives. "explicitlyTrashed": True or False, # Output only. Whether the file has been explicitly trashed, as opposed to recursively trashed from a parent folder. "exportLinks": { # Output only. Links for exporting Docs Editors files to specific formats. diff --git a/docs/dyn/drive_v3.teamdrives.html b/docs/dyn/drive_v3.teamdrives.html index e6007994c7..ec89f0b2b9 100644 --- a/docs/dyn/drive_v3.teamdrives.html +++ b/docs/dyn/drive_v3.teamdrives.html @@ -122,6 +122,7 @@

Method Details

"canAddChildren": True or False, # Whether the current user can add children to folders in this Team Drive. "canChangeCopyRequiresWriterPermissionRestriction": True or False, # Whether the current user can change the `copyRequiresWriterPermission` restriction of this Team Drive. "canChangeDomainUsersOnlyRestriction": True or False, # Whether the current user can change the `domainUsersOnly` restriction of this Team Drive. + "canChangeDownloadRestriction": True or False, # Whether the current user can change organizer-applied download restrictions of this shared drive. "canChangeSharingFoldersRequiresOrganizerPermissionRestriction": True or False, # Whether the current user can change the `sharingFoldersRequiresOrganizerPermission` restriction of this Team Drive. "canChangeTeamDriveBackground": True or False, # Whether the current user can change the background of this Team Drive. "canChangeTeamMembersOnlyRestriction": True or False, # Whether the current user can change the `teamMembersOnly` restriction of this Team Drive. @@ -181,6 +182,7 @@

Method Details

"canAddChildren": True or False, # Whether the current user can add children to folders in this Team Drive. "canChangeCopyRequiresWriterPermissionRestriction": True or False, # Whether the current user can change the `copyRequiresWriterPermission` restriction of this Team Drive. "canChangeDomainUsersOnlyRestriction": True or False, # Whether the current user can change the `domainUsersOnly` restriction of this Team Drive. + "canChangeDownloadRestriction": True or False, # Whether the current user can change organizer-applied download restrictions of this shared drive. "canChangeSharingFoldersRequiresOrganizerPermissionRestriction": True or False, # Whether the current user can change the `sharingFoldersRequiresOrganizerPermission` restriction of this Team Drive. "canChangeTeamDriveBackground": True or False, # Whether the current user can change the background of this Team Drive. "canChangeTeamMembersOnlyRestriction": True or False, # Whether the current user can change the `teamMembersOnly` restriction of this Team Drive. @@ -261,6 +263,7 @@

Method Details

"canAddChildren": True or False, # Whether the current user can add children to folders in this Team Drive. "canChangeCopyRequiresWriterPermissionRestriction": True or False, # Whether the current user can change the `copyRequiresWriterPermission` restriction of this Team Drive. "canChangeDomainUsersOnlyRestriction": True or False, # Whether the current user can change the `domainUsersOnly` restriction of this Team Drive. + "canChangeDownloadRestriction": True or False, # Whether the current user can change organizer-applied download restrictions of this shared drive. "canChangeSharingFoldersRequiresOrganizerPermissionRestriction": True or False, # Whether the current user can change the `sharingFoldersRequiresOrganizerPermission` restriction of this Team Drive. "canChangeTeamDriveBackground": True or False, # Whether the current user can change the background of this Team Drive. "canChangeTeamMembersOnlyRestriction": True or False, # Whether the current user can change the `teamMembersOnly` restriction of this Team Drive. @@ -334,6 +337,7 @@

Method Details

"canAddChildren": True or False, # Whether the current user can add children to folders in this Team Drive. "canChangeCopyRequiresWriterPermissionRestriction": True or False, # Whether the current user can change the `copyRequiresWriterPermission` restriction of this Team Drive. "canChangeDomainUsersOnlyRestriction": True or False, # Whether the current user can change the `domainUsersOnly` restriction of this Team Drive. + "canChangeDownloadRestriction": True or False, # Whether the current user can change organizer-applied download restrictions of this shared drive. "canChangeSharingFoldersRequiresOrganizerPermissionRestriction": True or False, # Whether the current user can change the `sharingFoldersRequiresOrganizerPermission` restriction of this Team Drive. "canChangeTeamDriveBackground": True or False, # Whether the current user can change the background of this Team Drive. "canChangeTeamMembersOnlyRestriction": True or False, # Whether the current user can change the `teamMembersOnly` restriction of this Team Drive. @@ -411,6 +415,7 @@

Method Details

"canAddChildren": True or False, # Whether the current user can add children to folders in this Team Drive. "canChangeCopyRequiresWriterPermissionRestriction": True or False, # Whether the current user can change the `copyRequiresWriterPermission` restriction of this Team Drive. "canChangeDomainUsersOnlyRestriction": True or False, # Whether the current user can change the `domainUsersOnly` restriction of this Team Drive. + "canChangeDownloadRestriction": True or False, # Whether the current user can change organizer-applied download restrictions of this shared drive. "canChangeSharingFoldersRequiresOrganizerPermissionRestriction": True or False, # Whether the current user can change the `sharingFoldersRequiresOrganizerPermission` restriction of this Team Drive. "canChangeTeamDriveBackground": True or False, # Whether the current user can change the background of this Team Drive. "canChangeTeamMembersOnlyRestriction": True or False, # Whether the current user can change the `teamMembersOnly` restriction of this Team Drive. @@ -471,6 +476,7 @@

Method Details

"canAddChildren": True or False, # Whether the current user can add children to folders in this Team Drive. "canChangeCopyRequiresWriterPermissionRestriction": True or False, # Whether the current user can change the `copyRequiresWriterPermission` restriction of this Team Drive. "canChangeDomainUsersOnlyRestriction": True or False, # Whether the current user can change the `domainUsersOnly` restriction of this Team Drive. + "canChangeDownloadRestriction": True or False, # Whether the current user can change organizer-applied download restrictions of this shared drive. "canChangeSharingFoldersRequiresOrganizerPermissionRestriction": True or False, # Whether the current user can change the `sharingFoldersRequiresOrganizerPermission` restriction of this Team Drive. "canChangeTeamDriveBackground": True or False, # Whether the current user can change the background of this Team Drive. "canChangeTeamMembersOnlyRestriction": True or False, # Whether the current user can change the `teamMembersOnly` restriction of this Team Drive. diff --git a/docs/dyn/gmail_v1.users.messages.html b/docs/dyn/gmail_v1.users.messages.html index 412bbdab56..cbfb74e716 100644 --- a/docs/dyn/gmail_v1.users.messages.html +++ b/docs/dyn/gmail_v1.users.messages.html @@ -102,7 +102,7 @@

Instance Methods

Directly inserts a message into only this user's mailbox similar to `IMAP APPEND`, bypassing most scanning and classification. Does not send a message.

list(userId, includeSpamTrash=None, labelIds=None, maxResults=None, pageToken=None, q=None, x__xgafv=None)

-

Lists the messages in the user's mailbox.

+

Lists the messages in the user's mailbox. For example usage, see [List Gmail messages](https://developers.google.com/workspace/gmail/api/guides/list-messages).

list_next()

Retrieves the next page of results.

@@ -425,7 +425,7 @@

Method Details

list(userId, includeSpamTrash=None, labelIds=None, maxResults=None, pageToken=None, q=None, x__xgafv=None) -
Lists the messages in the user's mailbox.
+  
Lists the messages in the user's mailbox. For example usage, see [List Gmail messages](https://developers.google.com/workspace/gmail/api/guides/list-messages).
 
 Args:
   userId: string, The user's email address. The special value `me` can be used to indicate the authenticated user. (required)
diff --git a/docs/dyn/iam_v1.projects.locations.workloadIdentityPools.html b/docs/dyn/iam_v1.projects.locations.workloadIdentityPools.html
index 7066bf5fdf..057ca00d3b 100644
--- a/docs/dyn/iam_v1.projects.locations.workloadIdentityPools.html
+++ b/docs/dyn/iam_v1.projects.locations.workloadIdentityPools.html
@@ -89,9 +89,6 @@ 

Instance Methods

Returns the providers Resource.

-

- addAttestationRule(resource, body=None, x__xgafv=None)

-

Add an AttestationRule on a WorkloadIdentityPoolManagedIdentity. The total attestation rules after addition must not exceed 50.

close()

Close httplib2 connections.

@@ -110,24 +107,12 @@

Instance Methods

list(parent, pageSize=None, pageToken=None, showDeleted=None, x__xgafv=None)

Lists all non-deleted WorkloadIdentityPools in a project. If `show_deleted` is set to `true`, then deleted pools are also listed.

-

- listAttestationRules(resource, filter=None, pageSize=None, pageToken=None, x__xgafv=None)

-

List all AttestationRule on a WorkloadIdentityPoolManagedIdentity.

-

- listAttestationRules_next()

-

Retrieves the next page of results.

list_next()

Retrieves the next page of results.

patch(name, body=None, updateMask=None, x__xgafv=None)

Updates an existing WorkloadIdentityPool.

-

- removeAttestationRule(resource, body=None, x__xgafv=None)

-

Remove an AttestationRule on a WorkloadIdentityPoolManagedIdentity.

-

- setAttestationRules(resource, body=None, x__xgafv=None)

-

Set all AttestationRule on a WorkloadIdentityPoolManagedIdentity. A maximum of 50 AttestationRules can be set.

setIamPolicy(resource, body=None, x__xgafv=None)

Sets the IAM policies on a WorkloadIdentityPool

@@ -138,50 +123,6 @@

Instance Methods

undelete(name, body=None, x__xgafv=None)

Undeletes a WorkloadIdentityPool, as long as it was deleted fewer than 30 days ago.

Method Details

-
- addAttestationRule(resource, body=None, x__xgafv=None) -
Add an AttestationRule on a WorkloadIdentityPoolManagedIdentity. The total attestation rules after addition must not exceed 50.
-
-Args:
-  resource: string, Required. The resource name of the managed identity or namespace resource to add an attestation rule to. (required)
-  body: object, The request body.
-    The object takes the form of:
-
-{ # Request message for AddAttestationRule.
-  "attestationRule": { # Defines which workloads can receive an identity within a pool. When an AttestationRule is defined under a managed identity, matching workloads may receive that identity. # Required. The attestation rule to be added.
-    "googleCloudResource": "A String", # Optional. A single workload operating on Google Cloud. For example: `//compute.googleapis.com/projects/123/uid/zones/us-central1-a/instances/12345`.
-  },
-}
-
-  x__xgafv: string, V1 error format.
-    Allowed values
-      1 - v1 error format
-      2 - v2 error format
-
-Returns:
-  An object of the form:
-
-    { # This resource represents a long-running operation that is the result of a network API call.
-  "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available.
-  "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation.
-    "code": 42, # The status code, which should be an enum value of google.rpc.Code.
-    "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use.
-      {
-        "a_key": "", # Properties of the object. Contains field @type with type URL.
-      },
-    ],
-    "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
-  },
-  "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any.
-    "a_key": "", # Properties of the object. Contains field @type with type URL.
-  },
-  "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`.
-  "response": { # The normal, successful response of the operation. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`.
-    "a_key": "", # Properties of the object. Contains field @type with type URL.
-  },
-}
-
-
close()
Close httplib2 connections.
@@ -457,47 +398,6 @@

Method Details

}
-
- listAttestationRules(resource, filter=None, pageSize=None, pageToken=None, x__xgafv=None) -
List all AttestationRule on a WorkloadIdentityPoolManagedIdentity.
-
-Args:
-  resource: string, Required. The resource name of the managed identity or namespace resource to list attestation rules of. (required)
-  filter: string, Optional. A query filter. Supports the following function: * `container_ids()`: Returns only the AttestationRules under the specific container ids. The function expects a comma-delimited list with only project numbers and must use the format `projects/`. For example: `container_ids(projects/, projects/,...)`.
-  pageSize: integer, Optional. The maximum number of AttestationRules to return. If unspecified, at most 50 AttestationRules are returned. The maximum value is 100; values above 100 are truncated to 100.
-  pageToken: string, Optional. A page token, received from a previous `ListWorkloadIdentityPoolProviderKeys` call. Provide this to retrieve the subsequent page.
-  x__xgafv: string, V1 error format.
-    Allowed values
-      1 - v1 error format
-      2 - v2 error format
-
-Returns:
-  An object of the form:
-
-    { # Response message for ListAttestationRules.
-  "attestationRules": [ # A list of AttestationRules.
-    { # Defines which workloads can receive an identity within a pool. When an AttestationRule is defined under a managed identity, matching workloads may receive that identity.
-      "googleCloudResource": "A String", # Optional. A single workload operating on Google Cloud. For example: `//compute.googleapis.com/projects/123/uid/zones/us-central1-a/instances/12345`.
-    },
-  ],
-  "nextPageToken": "A String", # Optional. A token, which can be sent as `page_token` to retrieve the next page. If this field is omitted, there are no subsequent pages.
-}
-
- -
- listAttestationRules_next() -
Retrieves the next page of results.
-
-        Args:
-          previous_request: The request for the previous page. (required)
-          previous_response: The response from the request for the previous page. (required)
-
-        Returns:
-          A request object that you can call 'execute()' on to request the next
-          page. Returns None if there are no more items in the collection.
-        
-
-
list_next()
Retrieves the next page of results.
@@ -585,96 +485,6 @@ 

Method Details

}
-
- removeAttestationRule(resource, body=None, x__xgafv=None) -
Remove an AttestationRule on a WorkloadIdentityPoolManagedIdentity.
-
-Args:
-  resource: string, Required. The resource name of the managed identity or namespace resource to remove an attestation rule from. (required)
-  body: object, The request body.
-    The object takes the form of:
-
-{ # Request message for RemoveAttestationRule.
-  "attestationRule": { # Defines which workloads can receive an identity within a pool. When an AttestationRule is defined under a managed identity, matching workloads may receive that identity. # Required. The attestation rule to be removed.
-    "googleCloudResource": "A String", # Optional. A single workload operating on Google Cloud. For example: `//compute.googleapis.com/projects/123/uid/zones/us-central1-a/instances/12345`.
-  },
-}
-
-  x__xgafv: string, V1 error format.
-    Allowed values
-      1 - v1 error format
-      2 - v2 error format
-
-Returns:
-  An object of the form:
-
-    { # This resource represents a long-running operation that is the result of a network API call.
-  "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available.
-  "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation.
-    "code": 42, # The status code, which should be an enum value of google.rpc.Code.
-    "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use.
-      {
-        "a_key": "", # Properties of the object. Contains field @type with type URL.
-      },
-    ],
-    "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
-  },
-  "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any.
-    "a_key": "", # Properties of the object. Contains field @type with type URL.
-  },
-  "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`.
-  "response": { # The normal, successful response of the operation. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`.
-    "a_key": "", # Properties of the object. Contains field @type with type URL.
-  },
-}
-
- -
- setAttestationRules(resource, body=None, x__xgafv=None) -
Set all AttestationRule on a WorkloadIdentityPoolManagedIdentity. A maximum of 50 AttestationRules can be set.
-
-Args:
-  resource: string, Required. The resource name of the managed identity or namespace resource to add an attestation rule to. (required)
-  body: object, The request body.
-    The object takes the form of:
-
-{ # Request message for SetAttestationRules.
-  "attestationRules": [ # Required. The attestation rules to be set. At most 50 attestation rules can be set.
-    { # Defines which workloads can receive an identity within a pool. When an AttestationRule is defined under a managed identity, matching workloads may receive that identity.
-      "googleCloudResource": "A String", # Optional. A single workload operating on Google Cloud. For example: `//compute.googleapis.com/projects/123/uid/zones/us-central1-a/instances/12345`.
-    },
-  ],
-}
-
-  x__xgafv: string, V1 error format.
-    Allowed values
-      1 - v1 error format
-      2 - v2 error format
-
-Returns:
-  An object of the form:
-
-    { # This resource represents a long-running operation that is the result of a network API call.
-  "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available.
-  "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation.
-    "code": 42, # The status code, which should be an enum value of google.rpc.Code.
-    "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use.
-      {
-        "a_key": "", # Properties of the object. Contains field @type with type URL.
-      },
-    ],
-    "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
-  },
-  "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any.
-    "a_key": "", # Properties of the object. Contains field @type with type URL.
-  },
-  "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`.
-  "response": { # The normal, successful response of the operation. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`.
-    "a_key": "", # Properties of the object. Contains field @type with type URL.
-  },
-}
-
-
setIamPolicy(resource, body=None, x__xgafv=None)
Sets the IAM policies on a WorkloadIdentityPool
diff --git a/docs/dyn/merchantapi_accounts_v1beta.accounts.programs.checkoutSettings.html b/docs/dyn/merchantapi_accounts_v1beta.accounts.programs.checkoutSettings.html
new file mode 100644
index 0000000000..3f8c55d179
--- /dev/null
+++ b/docs/dyn/merchantapi_accounts_v1beta.accounts.programs.checkoutSettings.html
@@ -0,0 +1,263 @@
+
+
+
+

Merchant API . accounts . programs . checkoutSettings

+

Instance Methods

+

+ close()

+

Close httplib2 connections.

+

+ create(parent, body=None, x__xgafv=None)

+

Creates `CheckoutSettings` for the given merchant.

+

+ deleteCheckoutSettings(name, x__xgafv=None)

+

Deletes `CheckoutSettings` and unenrolls merchant from `Checkout` program.

+

+ getCheckoutSettings(name, x__xgafv=None)

+

Gets `CheckoutSettings` for the given merchant. This includes information about review state, enrollment state and URL settings.

+

+ updateCheckoutSettings(name, body=None, updateMask=None, x__xgafv=None)

+

Updates `CheckoutSettings` for the given merchant.

+

Method Details

+
+ close() +
Close httplib2 connections.
+
+ +
+ create(parent, body=None, x__xgafv=None) +
Creates `CheckoutSettings` for the given merchant.
+
+Args:
+  parent: string, Required. The merchant account for which the `CheckoutSettings` will be created. (required)
+  body: object, The request body.
+    The object takes the form of:
+
+{ # [CheckoutSettings](https://support.google.com/merchants/answer/13945960) for a specific merchant.
+  "effectiveEnrollmentState": "A String", # Output only. The effective value of enrollment_state for a given merchant ID. If account level settings are present then this value will be a copy of the account level settings. Otherwise, it will have the value of the parent account (for only marketplace sellers).
+  "effectiveReviewState": "A String", # Output only. The effective value of `review_state` for a given merchant ID. If account level settings are present then this value will be a copy of the account level settings. Otherwise, it will have the value of the parent account (for only marketplace sellers).
+  "effectiveUriSettings": { # URL settings for cart or checkout URL. # Output only. The effective value of `uri_settings` for a given merchant. If account level settings are present then this value will be a copy of url settings. Otherwise, it will have the value of the parent account (for only marketplace sellers).
+    "cartUriTemplate": "A String", # Cart URL template. When the placeholders are expanded will redirect the buyer to the cart page on the merchant website with the selected item in cart. For more details, check the [help center doc](https://support.google.com/merchants/answer/13945960#method1&zippy=%2Cproduct-level-url-formatting%2Caccount-level-url-formatting)
+    "checkoutUriTemplate": "A String", # Checkout URL template. When the placeholders are expanded will redirect the buyer to the merchant checkout page with the item in the cart. For more details, check the [help center doc](https://support.google.com/merchants/answer/13945960#method1&zippy=%2Cproduct-level-url-formatting%2Caccount-level-url-formatting)
+  },
+  "eligibleDestinations": [ # Optional. The destinations to which the checkout program applies, valid destination values are `SHOPPING_ADS`, `FREE_LISTINGS`
+    "A String",
+  ],
+  "enrollmentState": "A String", # Output only. Reflects the merchant enrollment state in `Checkout` program.
+  "name": "A String", # Identifier. The resource name of the program configuration settings. Format: `accounts/{account}/programs/{program}/checkoutSettings`
+  "reviewState": "A String", # Output only. Reflects the merchant review state in `Checkout` program. This is set based on the data quality reviews of the URL provided by the merchant. A merchant with enrollment state as `ENROLLED` can be in the following review states: `IN_REVIEW`, `APPROVED` or `DISAPPROVED`. A merchant must be in an `enrollment_state` of `ENROLLED` before a review can begin for the merchant.For more details, check the help center doc.
+  "uriSettings": { # URL settings for cart or checkout URL. # URI settings for cart or checkout URL.
+    "cartUriTemplate": "A String", # Cart URL template. When the placeholders are expanded will redirect the buyer to the cart page on the merchant website with the selected item in cart. For more details, check the [help center doc](https://support.google.com/merchants/answer/13945960#method1&zippy=%2Cproduct-level-url-formatting%2Caccount-level-url-formatting)
+    "checkoutUriTemplate": "A String", # Checkout URL template. When the placeholders are expanded will redirect the buyer to the merchant checkout page with the item in the cart. For more details, check the [help center doc](https://support.google.com/merchants/answer/13945960#method1&zippy=%2Cproduct-level-url-formatting%2Caccount-level-url-formatting)
+  },
+}
+
+  x__xgafv: string, V1 error format.
+    Allowed values
+      1 - v1 error format
+      2 - v2 error format
+
+Returns:
+  An object of the form:
+
+    { # [CheckoutSettings](https://support.google.com/merchants/answer/13945960) for a specific merchant.
+  "effectiveEnrollmentState": "A String", # Output only. The effective value of enrollment_state for a given merchant ID. If account level settings are present then this value will be a copy of the account level settings. Otherwise, it will have the value of the parent account (for only marketplace sellers).
+  "effectiveReviewState": "A String", # Output only. The effective value of `review_state` for a given merchant ID. If account level settings are present then this value will be a copy of the account level settings. Otherwise, it will have the value of the parent account (for only marketplace sellers).
+  "effectiveUriSettings": { # URL settings for cart or checkout URL. # Output only. The effective value of `uri_settings` for a given merchant. If account level settings are present then this value will be a copy of url settings. Otherwise, it will have the value of the parent account (for only marketplace sellers).
+    "cartUriTemplate": "A String", # Cart URL template. When the placeholders are expanded will redirect the buyer to the cart page on the merchant website with the selected item in cart. For more details, check the [help center doc](https://support.google.com/merchants/answer/13945960#method1&zippy=%2Cproduct-level-url-formatting%2Caccount-level-url-formatting)
+    "checkoutUriTemplate": "A String", # Checkout URL template. When the placeholders are expanded will redirect the buyer to the merchant checkout page with the item in the cart. For more details, check the [help center doc](https://support.google.com/merchants/answer/13945960#method1&zippy=%2Cproduct-level-url-formatting%2Caccount-level-url-formatting)
+  },
+  "eligibleDestinations": [ # Optional. The destinations to which the checkout program applies, valid destination values are `SHOPPING_ADS`, `FREE_LISTINGS`
+    "A String",
+  ],
+  "enrollmentState": "A String", # Output only. Reflects the merchant enrollment state in `Checkout` program.
+  "name": "A String", # Identifier. The resource name of the program configuration settings. Format: `accounts/{account}/programs/{program}/checkoutSettings`
+  "reviewState": "A String", # Output only. Reflects the merchant review state in `Checkout` program. This is set based on the data quality reviews of the URL provided by the merchant. A merchant with enrollment state as `ENROLLED` can be in the following review states: `IN_REVIEW`, `APPROVED` or `DISAPPROVED`. A merchant must be in an `enrollment_state` of `ENROLLED` before a review can begin for the merchant.For more details, check the help center doc.
+  "uriSettings": { # URL settings for cart or checkout URL. # URI settings for cart or checkout URL.
+    "cartUriTemplate": "A String", # Cart URL template. When the placeholders are expanded will redirect the buyer to the cart page on the merchant website with the selected item in cart. For more details, check the [help center doc](https://support.google.com/merchants/answer/13945960#method1&zippy=%2Cproduct-level-url-formatting%2Caccount-level-url-formatting)
+    "checkoutUriTemplate": "A String", # Checkout URL template. When the placeholders are expanded will redirect the buyer to the merchant checkout page with the item in the cart. For more details, check the [help center doc](https://support.google.com/merchants/answer/13945960#method1&zippy=%2Cproduct-level-url-formatting%2Caccount-level-url-formatting)
+  },
+}
+
+ +
+ deleteCheckoutSettings(name, x__xgafv=None) +
Deletes `CheckoutSettings` and unenrolls merchant from `Checkout` program.
+
+Args:
+  name: string, Required. The name/identifier of the merchant account. Format: `accounts/{account}/programs/{program}/checkoutSettings` (required)
+  x__xgafv: string, V1 error format.
+    Allowed values
+      1 - v1 error format
+      2 - v2 error format
+
+Returns:
+  An object of the form:
+
+    { # A generic empty message that you can re-use to avoid defining duplicated empty messages in your APIs. A typical example is to use it as the request or the response type of an API method. For instance: service Foo { rpc Bar(google.protobuf.Empty) returns (google.protobuf.Empty); }
+}
+
+ +
+ getCheckoutSettings(name, x__xgafv=None) +
Gets `CheckoutSettings` for the given merchant. This includes information about review state, enrollment state and URL settings.
+
+Args:
+  name: string, Required. The name/identifier of the merchant account. Format: `accounts/{account}/programs/{program}/checkoutSettings` (required)
+  x__xgafv: string, V1 error format.
+    Allowed values
+      1 - v1 error format
+      2 - v2 error format
+
+Returns:
+  An object of the form:
+
+    { # [CheckoutSettings](https://support.google.com/merchants/answer/13945960) for a specific merchant.
+  "effectiveEnrollmentState": "A String", # Output only. The effective value of enrollment_state for a given merchant ID. If account level settings are present then this value will be a copy of the account level settings. Otherwise, it will have the value of the parent account (for only marketplace sellers).
+  "effectiveReviewState": "A String", # Output only. The effective value of `review_state` for a given merchant ID. If account level settings are present then this value will be a copy of the account level settings. Otherwise, it will have the value of the parent account (for only marketplace sellers).
+  "effectiveUriSettings": { # URL settings for cart or checkout URL. # Output only. The effective value of `uri_settings` for a given merchant. If account level settings are present then this value will be a copy of url settings. Otherwise, it will have the value of the parent account (for only marketplace sellers).
+    "cartUriTemplate": "A String", # Cart URL template. When the placeholders are expanded will redirect the buyer to the cart page on the merchant website with the selected item in cart. For more details, check the [help center doc](https://support.google.com/merchants/answer/13945960#method1&zippy=%2Cproduct-level-url-formatting%2Caccount-level-url-formatting)
+    "checkoutUriTemplate": "A String", # Checkout URL template. When the placeholders are expanded will redirect the buyer to the merchant checkout page with the item in the cart. For more details, check the [help center doc](https://support.google.com/merchants/answer/13945960#method1&zippy=%2Cproduct-level-url-formatting%2Caccount-level-url-formatting)
+  },
+  "eligibleDestinations": [ # Optional. The destinations to which the checkout program applies, valid destination values are `SHOPPING_ADS`, `FREE_LISTINGS`
+    "A String",
+  ],
+  "enrollmentState": "A String", # Output only. Reflects the merchant enrollment state in `Checkout` program.
+  "name": "A String", # Identifier. The resource name of the program configuration settings. Format: `accounts/{account}/programs/{program}/checkoutSettings`
+  "reviewState": "A String", # Output only. Reflects the merchant review state in `Checkout` program. This is set based on the data quality reviews of the URL provided by the merchant. A merchant with enrollment state as `ENROLLED` can be in the following review states: `IN_REVIEW`, `APPROVED` or `DISAPPROVED`. A merchant must be in an `enrollment_state` of `ENROLLED` before a review can begin for the merchant.For more details, check the help center doc.
+  "uriSettings": { # URL settings for cart or checkout URL. # URI settings for cart or checkout URL.
+    "cartUriTemplate": "A String", # Cart URL template. When the placeholders are expanded will redirect the buyer to the cart page on the merchant website with the selected item in cart. For more details, check the [help center doc](https://support.google.com/merchants/answer/13945960#method1&zippy=%2Cproduct-level-url-formatting%2Caccount-level-url-formatting)
+    "checkoutUriTemplate": "A String", # Checkout URL template. When the placeholders are expanded will redirect the buyer to the merchant checkout page with the item in the cart. For more details, check the [help center doc](https://support.google.com/merchants/answer/13945960#method1&zippy=%2Cproduct-level-url-formatting%2Caccount-level-url-formatting)
+  },
+}
+
+ +
+ updateCheckoutSettings(name, body=None, updateMask=None, x__xgafv=None) +
Updates `CheckoutSettings` for the given merchant.
+
+Args:
+  name: string, Identifier. The resource name of the program configuration settings. Format: `accounts/{account}/programs/{program}/checkoutSettings` (required)
+  body: object, The request body.
+    The object takes the form of:
+
+{ # [CheckoutSettings](https://support.google.com/merchants/answer/13945960) for a specific merchant.
+  "effectiveEnrollmentState": "A String", # Output only. The effective value of enrollment_state for a given merchant ID. If account level settings are present then this value will be a copy of the account level settings. Otherwise, it will have the value of the parent account (for only marketplace sellers).
+  "effectiveReviewState": "A String", # Output only. The effective value of `review_state` for a given merchant ID. If account level settings are present then this value will be a copy of the account level settings. Otherwise, it will have the value of the parent account (for only marketplace sellers).
+  "effectiveUriSettings": { # URL settings for cart or checkout URL. # Output only. The effective value of `uri_settings` for a given merchant. If account level settings are present then this value will be a copy of url settings. Otherwise, it will have the value of the parent account (for only marketplace sellers).
+    "cartUriTemplate": "A String", # Cart URL template. When the placeholders are expanded will redirect the buyer to the cart page on the merchant website with the selected item in cart. For more details, check the [help center doc](https://support.google.com/merchants/answer/13945960#method1&zippy=%2Cproduct-level-url-formatting%2Caccount-level-url-formatting)
+    "checkoutUriTemplate": "A String", # Checkout URL template. When the placeholders are expanded will redirect the buyer to the merchant checkout page with the item in the cart. For more details, check the [help center doc](https://support.google.com/merchants/answer/13945960#method1&zippy=%2Cproduct-level-url-formatting%2Caccount-level-url-formatting)
+  },
+  "eligibleDestinations": [ # Optional. The destinations to which the checkout program applies, valid destination values are `SHOPPING_ADS`, `FREE_LISTINGS`
+    "A String",
+  ],
+  "enrollmentState": "A String", # Output only. Reflects the merchant enrollment state in `Checkout` program.
+  "name": "A String", # Identifier. The resource name of the program configuration settings. Format: `accounts/{account}/programs/{program}/checkoutSettings`
+  "reviewState": "A String", # Output only. Reflects the merchant review state in `Checkout` program. This is set based on the data quality reviews of the URL provided by the merchant. A merchant with enrollment state as `ENROLLED` can be in the following review states: `IN_REVIEW`, `APPROVED` or `DISAPPROVED`. A merchant must be in an `enrollment_state` of `ENROLLED` before a review can begin for the merchant.For more details, check the help center doc.
+  "uriSettings": { # URL settings for cart or checkout URL. # URI settings for cart or checkout URL.
+    "cartUriTemplate": "A String", # Cart URL template. When the placeholders are expanded will redirect the buyer to the cart page on the merchant website with the selected item in cart. For more details, check the [help center doc](https://support.google.com/merchants/answer/13945960#method1&zippy=%2Cproduct-level-url-formatting%2Caccount-level-url-formatting)
+    "checkoutUriTemplate": "A String", # Checkout URL template. When the placeholders are expanded will redirect the buyer to the merchant checkout page with the item in the cart. For more details, check the [help center doc](https://support.google.com/merchants/answer/13945960#method1&zippy=%2Cproduct-level-url-formatting%2Caccount-level-url-formatting)
+  },
+}
+
+  updateMask: string, Required. List of fields being updated. The following fields are supported (in both `snake_case` and `lowerCamelCase`): - `eligible_destinations` - `uri_settings`
+  x__xgafv: string, V1 error format.
+    Allowed values
+      1 - v1 error format
+      2 - v2 error format
+
+Returns:
+  An object of the form:
+
+    { # [CheckoutSettings](https://support.google.com/merchants/answer/13945960) for a specific merchant.
+  "effectiveEnrollmentState": "A String", # Output only. The effective value of enrollment_state for a given merchant ID. If account level settings are present then this value will be a copy of the account level settings. Otherwise, it will have the value of the parent account (for only marketplace sellers).
+  "effectiveReviewState": "A String", # Output only. The effective value of `review_state` for a given merchant ID. If account level settings are present then this value will be a copy of the account level settings. Otherwise, it will have the value of the parent account (for only marketplace sellers).
+  "effectiveUriSettings": { # URL settings for cart or checkout URL. # Output only. The effective value of `uri_settings` for a given merchant. If account level settings are present then this value will be a copy of url settings. Otherwise, it will have the value of the parent account (for only marketplace sellers).
+    "cartUriTemplate": "A String", # Cart URL template. When the placeholders are expanded will redirect the buyer to the cart page on the merchant website with the selected item in cart. For more details, check the [help center doc](https://support.google.com/merchants/answer/13945960#method1&zippy=%2Cproduct-level-url-formatting%2Caccount-level-url-formatting)
+    "checkoutUriTemplate": "A String", # Checkout URL template. When the placeholders are expanded will redirect the buyer to the merchant checkout page with the item in the cart. For more details, check the [help center doc](https://support.google.com/merchants/answer/13945960#method1&zippy=%2Cproduct-level-url-formatting%2Caccount-level-url-formatting)
+  },
+  "eligibleDestinations": [ # Optional. The destinations to which the checkout program applies, valid destination values are `SHOPPING_ADS`, `FREE_LISTINGS`
+    "A String",
+  ],
+  "enrollmentState": "A String", # Output only. Reflects the merchant enrollment state in `Checkout` program.
+  "name": "A String", # Identifier. The resource name of the program configuration settings. Format: `accounts/{account}/programs/{program}/checkoutSettings`
+  "reviewState": "A String", # Output only. Reflects the merchant review state in `Checkout` program. This is set based on the data quality reviews of the URL provided by the merchant. A merchant with enrollment state as `ENROLLED` can be in the following review states: `IN_REVIEW`, `APPROVED` or `DISAPPROVED`. A merchant must be in an `enrollment_state` of `ENROLLED` before a review can begin for the merchant.For more details, check the help center doc.
+  "uriSettings": { # URL settings for cart or checkout URL. # URI settings for cart or checkout URL.
+    "cartUriTemplate": "A String", # Cart URL template. When the placeholders are expanded will redirect the buyer to the cart page on the merchant website with the selected item in cart. For more details, check the [help center doc](https://support.google.com/merchants/answer/13945960#method1&zippy=%2Cproduct-level-url-formatting%2Caccount-level-url-formatting)
+    "checkoutUriTemplate": "A String", # Checkout URL template. When the placeholders are expanded will redirect the buyer to the merchant checkout page with the item in the cart. For more details, check the [help center doc](https://support.google.com/merchants/answer/13945960#method1&zippy=%2Cproduct-level-url-formatting%2Caccount-level-url-formatting)
+  },
+}
+
+ + \ No newline at end of file diff --git a/docs/dyn/merchantapi_accounts_v1beta.accounts.programs.html b/docs/dyn/merchantapi_accounts_v1beta.accounts.programs.html index 909daf71ab..82be62c491 100644 --- a/docs/dyn/merchantapi_accounts_v1beta.accounts.programs.html +++ b/docs/dyn/merchantapi_accounts_v1beta.accounts.programs.html @@ -74,6 +74,11 @@

Merchant API . accounts . programs

Instance Methods

+

+ checkoutSettings() +

+

Returns the checkoutSettings Resource.

+

close()

Close httplib2 connections.

diff --git a/docs/dyn/merchantapi_promotions_v1beta.accounts.promotions.html b/docs/dyn/merchantapi_promotions_v1beta.accounts.promotions.html index b6bcedd9e1..46809844b5 100644 --- a/docs/dyn/merchantapi_promotions_v1beta.accounts.promotions.html +++ b/docs/dyn/merchantapi_promotions_v1beta.accounts.promotions.html @@ -148,7 +148,7 @@

Method Details

"amountMicros": "A String", # The price represented as a number in micros (1 million micros is an equivalent to one's currency standard unit, for example, 1 USD = 1000000 micros). "currencyCode": "A String", # The currency of the price using three-letter acronyms according to [ISO 4217](http://en.wikipedia.org/wiki/ISO_4217). }, - "minimumPurchaseQuantity": "A String", # Optional. [Minimum purchase quantity](https://support.google.com/merchants/answer/13838182?ref_topic=13773355&sjid=17642868584668136159-NC) for the promotion. + "minimumPurchaseQuantity": "A String", "moneyOffAmount": { # The price represented as a number and currency. # Optional. The [money off amount](https://support.google.com/merchants/answer/13838101?ref_topic=13773355&sjid=17642868584668136159-NC) offered in the promotion. "amountMicros": "A String", # The price represented as a number in micros (1 million micros is an equivalent to one's currency standard unit, for example, 1 USD = 1000000 micros). "currencyCode": "A String", # The currency of the price using three-letter acronyms according to [ISO 4217](http://en.wikipedia.org/wiki/ISO_4217). @@ -278,7 +278,7 @@

Method Details

"amountMicros": "A String", # The price represented as a number in micros (1 million micros is an equivalent to one's currency standard unit, for example, 1 USD = 1000000 micros). "currencyCode": "A String", # The currency of the price using three-letter acronyms according to [ISO 4217](http://en.wikipedia.org/wiki/ISO_4217). }, - "minimumPurchaseQuantity": "A String", # Optional. [Minimum purchase quantity](https://support.google.com/merchants/answer/13838182?ref_topic=13773355&sjid=17642868584668136159-NC) for the promotion. + "minimumPurchaseQuantity": "A String", "moneyOffAmount": { # The price represented as a number and currency. # Optional. The [money off amount](https://support.google.com/merchants/answer/13838101?ref_topic=13773355&sjid=17642868584668136159-NC) offered in the promotion. "amountMicros": "A String", # The price represented as a number in micros (1 million micros is an equivalent to one's currency standard unit, for example, 1 USD = 1000000 micros). "currencyCode": "A String", # The currency of the price using three-letter acronyms according to [ISO 4217](http://en.wikipedia.org/wiki/ISO_4217). @@ -405,7 +405,7 @@

Method Details

"amountMicros": "A String", # The price represented as a number in micros (1 million micros is an equivalent to one's currency standard unit, for example, 1 USD = 1000000 micros). "currencyCode": "A String", # The currency of the price using three-letter acronyms according to [ISO 4217](http://en.wikipedia.org/wiki/ISO_4217). }, - "minimumPurchaseQuantity": "A String", # Optional. [Minimum purchase quantity](https://support.google.com/merchants/answer/13838182?ref_topic=13773355&sjid=17642868584668136159-NC) for the promotion. + "minimumPurchaseQuantity": "A String", "moneyOffAmount": { # The price represented as a number and currency. # Optional. The [money off amount](https://support.google.com/merchants/answer/13838101?ref_topic=13773355&sjid=17642868584668136159-NC) offered in the promotion. "amountMicros": "A String", # The price represented as a number in micros (1 million micros is an equivalent to one's currency standard unit, for example, 1 USD = 1000000 micros). "currencyCode": "A String", # The currency of the price using three-letter acronyms according to [ISO 4217](http://en.wikipedia.org/wiki/ISO_4217). @@ -543,7 +543,7 @@

Method Details

"amountMicros": "A String", # The price represented as a number in micros (1 million micros is an equivalent to one's currency standard unit, for example, 1 USD = 1000000 micros). "currencyCode": "A String", # The currency of the price using three-letter acronyms according to [ISO 4217](http://en.wikipedia.org/wiki/ISO_4217). }, - "minimumPurchaseQuantity": "A String", # Optional. [Minimum purchase quantity](https://support.google.com/merchants/answer/13838182?ref_topic=13773355&sjid=17642868584668136159-NC) for the promotion. + "minimumPurchaseQuantity": "A String", "moneyOffAmount": { # The price represented as a number and currency. # Optional. The [money off amount](https://support.google.com/merchants/answer/13838101?ref_topic=13773355&sjid=17642868584668136159-NC) offered in the promotion. "amountMicros": "A String", # The price represented as a number in micros (1 million micros is an equivalent to one's currency standard unit, for example, 1 USD = 1000000 micros). "currencyCode": "A String", # The currency of the price using three-letter acronyms according to [ISO 4217](http://en.wikipedia.org/wiki/ISO_4217). diff --git a/docs/dyn/networkmanagement_v1.projects.locations.vpcFlowLogsConfigs.html b/docs/dyn/networkmanagement_v1.projects.locations.vpcFlowLogsConfigs.html index f37ca06f07..dd04efec17 100644 --- a/docs/dyn/networkmanagement_v1.projects.locations.vpcFlowLogsConfigs.html +++ b/docs/dyn/networkmanagement_v1.projects.locations.vpcFlowLogsConfigs.html @@ -79,7 +79,7 @@

Instance Methods

Close httplib2 connections.

create(parent, body=None, vpcFlowLogsConfigId=None, x__xgafv=None)

-

Creates a new `VpcFlowLogsConfig`. If a configuration with the exact same settings already exists (even if the ID is different), the creation fails. Notes: 1. Creating a configuration with state=DISABLED will fail 2. The following fields are not considered as `settings` for the purpose of the check mentioned above, therefore - creating another configuration with the same fields but different values for the following fields will fail as well: * name * create_time * update_time * labels * description

+

Creates a new `VpcFlowLogsConfig`. If a configuration with the exact same settings already exists (even if the ID is different), the creation fails. Notes: 1. Creating a configuration with `state=DISABLED` will fail 2. The following fields are not considered as settings for the purpose of the check mentioned above, therefore - creating another configuration with the same fields but different values for the following fields will fail as well: * name * create_time * update_time * labels * description

delete(name, x__xgafv=None)

Deletes a specific `VpcFlowLogsConfig`.

@@ -94,7 +94,7 @@

Instance Methods

Retrieves the next page of results.

patch(name, body=None, updateMask=None, x__xgafv=None)

-

Updates an existing `VpcFlowLogsConfig`. If a configuration with the exact same settings already exists (even if the ID is different), the creation fails. Notes: 1. Updating a configuration with state=DISABLED will fail. 2. The following fields are not considered as `settings` for the purpose of the check mentioned above, therefore - updating another configuration with the same fields but different values for the following fields will fail as well: * name * create_time * update_time * labels * description

+

Updates an existing `VpcFlowLogsConfig`. If a configuration with the exact same settings already exists (even if the ID is different), the creation fails. Notes: 1. Updating a configuration with `state=DISABLED` will fail. 2. The following fields are not considered as settings for the purpose of the check mentioned above, therefore - updating another configuration with the same fields but different values for the following fields will fail as well: * name * create_time * update_time * labels * description

Method Details

close() @@ -103,10 +103,10 @@

Method Details

create(parent, body=None, vpcFlowLogsConfigId=None, x__xgafv=None) -
Creates a new `VpcFlowLogsConfig`. If a configuration with the exact same settings already exists (even if the ID is different), the creation fails. Notes: 1. Creating a configuration with state=DISABLED will fail 2. The following fields are not considered as `settings` for the purpose of the check mentioned above, therefore - creating another configuration with the same fields but different values for the following fields will fail as well: * name * create_time * update_time * labels * description
+  
Creates a new `VpcFlowLogsConfig`. If a configuration with the exact same settings already exists (even if the ID is different), the creation fails. Notes: 1. Creating a configuration with `state=DISABLED` will fail 2. The following fields are not considered as settings for the purpose of the check mentioned above, therefore - creating another configuration with the same fields but different values for the following fields will fail as well: * name * create_time * update_time * labels * description
 
 Args:
-  parent: string, Required. The parent resource of the VPC Flow Logs configuration to create: `projects/{project_id}/locations/global` `organizations/{organization_id}/locations/global` (required)
+  parent: string, Required. The parent resource of the VpcFlowLogsConfig to create, in one of the following formats: - For project-level resources: `projects/{project_id}/locations/global` - For organization-level resources: `organizations/{organization_id}/locations/global` (required)
   body: object, The request body.
     The object takes the form of:
 
@@ -124,9 +124,9 @@ 

Method Details

"metadataFields": [ # Optional. Custom metadata fields to include in the reported VPC flow logs. Can only be specified if "metadata" was set to CUSTOM_METADATA. "A String", ], - "name": "A String", # Identifier. Unique name of the configuration using one of the forms: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` `organizations/{organization_number}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` + "name": "A String", # Identifier. Unique name of the configuration. The name can have one of the following forms: - For project-level configurations: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` - For organization-level configurations: `organizations/{organization_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` "state": "A String", # Optional. The state of the VPC Flow Log configuration. Default value is ENABLED. When creating a new configuration, it must be enabled. Setting state=DISABLED will pause the log generation for this config. - "targetResourceState": "A String", # Output only. A diagnostic bit - describes the state of the configured target resource for diagnostic purposes. + "targetResourceState": "A String", # Output only. Describes the state of the configured target resource for diagnostic purposes. "updateTime": "A String", # Output only. The time the config was updated. "vpnTunnel": "A String", # Traffic will be logged from the VPN Tunnel. Format: projects/{project_id}/regions/{region}/vpnTunnels/{name} } @@ -166,7 +166,7 @@

Method Details

Deletes a specific `VpcFlowLogsConfig`.
 
 Args:
-  name: string, Required. `VpcFlowLogsConfig` resource name using one of the form: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config} `organizations/{organization_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config}` (required)
+  name: string, Required. The resource name of the VpcFlowLogsConfig, in one of the following formats: - For a project-level resource: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` - For an organization-level resource: `organizations/{organization_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` (required)
   x__xgafv: string, V1 error format.
     Allowed values
       1 - v1 error format
@@ -201,7 +201,7 @@ 

Method Details

Gets the details of a specific `VpcFlowLogsConfig`.
 
 Args:
-  name: string, Required. `VpcFlowLogsConfig` resource name using the form: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config}` `organizations/{organization_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config}` (required)
+  name: string, Required. The resource name of the VpcFlowLogsConfig, in one of the following formats: - For project-level resources: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` - For organization-level resources: `organizations/{organization_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` (required)
   x__xgafv: string, V1 error format.
     Allowed values
       1 - v1 error format
@@ -224,9 +224,9 @@ 

Method Details

"metadataFields": [ # Optional. Custom metadata fields to include in the reported VPC flow logs. Can only be specified if "metadata" was set to CUSTOM_METADATA. "A String", ], - "name": "A String", # Identifier. Unique name of the configuration using one of the forms: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` `organizations/{organization_number}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` + "name": "A String", # Identifier. Unique name of the configuration. The name can have one of the following forms: - For project-level configurations: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` - For organization-level configurations: `organizations/{organization_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` "state": "A String", # Optional. The state of the VPC Flow Log configuration. Default value is ENABLED. When creating a new configuration, it must be enabled. Setting state=DISABLED will pause the log generation for this config. - "targetResourceState": "A String", # Output only. A diagnostic bit - describes the state of the configured target resource for diagnostic purposes. + "targetResourceState": "A String", # Output only. Describes the state of the configured target resource for diagnostic purposes. "updateTime": "A String", # Output only. The time the config was updated. "vpnTunnel": "A String", # Traffic will be logged from the VPN Tunnel. Format: projects/{project_id}/regions/{region}/vpnTunnels/{name} }
@@ -237,7 +237,7 @@

Method Details

Lists all `VpcFlowLogsConfigs` in a given project.
 
 Args:
-  parent: string, Required. The parent resource of the VpcFlowLogsConfig: `projects/{project_id}/locations/global` `organizations/{organization_id}/locations/global` (required)
+  parent: string, Required. The parent resource of the VpcFlowLogsConfig, in one of the following formats: - For project-level resourcs: `projects/{project_id}/locations/global` - For organization-level resources: `organizations/{organization_id}/locations/global` (required)
   filter: string, Optional. Lists the `VpcFlowLogsConfigs` that match the filter expression. A filter expression must use the supported [CEL logic operators] (https://cloud.google.com/vpc/docs/about-flow-logs-records#supported_cel_logic_operators).
   orderBy: string, Optional. Field to use to sort the list.
   pageSize: integer, Optional. Number of `VpcFlowLogsConfigs` to return.
@@ -270,9 +270,9 @@ 

Method Details

"metadataFields": [ # Optional. Custom metadata fields to include in the reported VPC flow logs. Can only be specified if "metadata" was set to CUSTOM_METADATA. "A String", ], - "name": "A String", # Identifier. Unique name of the configuration using one of the forms: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` `organizations/{organization_number}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` + "name": "A String", # Identifier. Unique name of the configuration. The name can have one of the following forms: - For project-level configurations: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` - For organization-level configurations: `organizations/{organization_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` "state": "A String", # Optional. The state of the VPC Flow Log configuration. Default value is ENABLED. When creating a new configuration, it must be enabled. Setting state=DISABLED will pause the log generation for this config. - "targetResourceState": "A String", # Output only. A diagnostic bit - describes the state of the configured target resource for diagnostic purposes. + "targetResourceState": "A String", # Output only. Describes the state of the configured target resource for diagnostic purposes. "updateTime": "A String", # Output only. The time the config was updated. "vpnTunnel": "A String", # Traffic will be logged from the VPN Tunnel. Format: projects/{project_id}/regions/{region}/vpnTunnels/{name} }, @@ -296,10 +296,10 @@

Method Details

patch(name, body=None, updateMask=None, x__xgafv=None) -
Updates an existing `VpcFlowLogsConfig`. If a configuration with the exact same settings already exists (even if the ID is different), the creation fails. Notes: 1. Updating a configuration with state=DISABLED will fail. 2. The following fields are not considered as `settings` for the purpose of the check mentioned above, therefore - updating another configuration with the same fields but different values for the following fields will fail as well: * name * create_time * update_time * labels * description
+  
Updates an existing `VpcFlowLogsConfig`. If a configuration with the exact same settings already exists (even if the ID is different), the creation fails. Notes: 1. Updating a configuration with `state=DISABLED` will fail. 2. The following fields are not considered as settings for the purpose of the check mentioned above, therefore - updating another configuration with the same fields but different values for the following fields will fail as well: * name * create_time * update_time * labels * description
 
 Args:
-  name: string, Identifier. Unique name of the configuration using one of the forms: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` `organizations/{organization_number}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` (required)
+  name: string, Identifier. Unique name of the configuration. The name can have one of the following forms: - For project-level configurations: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` - For organization-level configurations: `organizations/{organization_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` (required)
   body: object, The request body.
     The object takes the form of:
 
@@ -317,14 +317,14 @@ 

Method Details

"metadataFields": [ # Optional. Custom metadata fields to include in the reported VPC flow logs. Can only be specified if "metadata" was set to CUSTOM_METADATA. "A String", ], - "name": "A String", # Identifier. Unique name of the configuration using one of the forms: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` `organizations/{organization_number}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` + "name": "A String", # Identifier. Unique name of the configuration. The name can have one of the following forms: - For project-level configurations: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` - For organization-level configurations: `organizations/{organization_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` "state": "A String", # Optional. The state of the VPC Flow Log configuration. Default value is ENABLED. When creating a new configuration, it must be enabled. Setting state=DISABLED will pause the log generation for this config. - "targetResourceState": "A String", # Output only. A diagnostic bit - describes the state of the configured target resource for diagnostic purposes. + "targetResourceState": "A String", # Output only. Describes the state of the configured target resource for diagnostic purposes. "updateTime": "A String", # Output only. The time the config was updated. "vpnTunnel": "A String", # Traffic will be logged from the VPN Tunnel. Format: projects/{project_id}/regions/{region}/vpnTunnels/{name} } - updateMask: string, Required. Mask of fields to update. At least one path must be supplied in this field. + updateMask: string, Required. Mask of fields to update. At least one path must be supplied in this field. For example, to change the state of the configuration to ENABLED, specify `update_mask` = `"state"`, and the `vpc_flow_logs_config` would be: `vpc_flow_logs_config = { name = "projects/my-project/locations/global/vpcFlowLogsConfigs/my-config" state = "ENABLED" }` x__xgafv: string, V1 error format. Allowed values 1 - v1 error format diff --git a/docs/dyn/networkmanagement_v1beta1.organizations.locations.vpcFlowLogsConfigs.html b/docs/dyn/networkmanagement_v1beta1.organizations.locations.vpcFlowLogsConfigs.html index 6b74521dde..ddbf79616d 100644 --- a/docs/dyn/networkmanagement_v1beta1.organizations.locations.vpcFlowLogsConfigs.html +++ b/docs/dyn/networkmanagement_v1beta1.organizations.locations.vpcFlowLogsConfigs.html @@ -79,7 +79,7 @@

Instance Methods

Close httplib2 connections.

create(parent, body=None, vpcFlowLogsConfigId=None, x__xgafv=None)

-

Creates a new `VpcFlowLogsConfig`. If a configuration with the exact same settings already exists (even if the ID is different), the creation fails. Notes: 1. Creating a configuration with state=DISABLED will fail 2. The following fields are not considered as `settings` for the purpose of the check mentioned above, therefore - creating another configuration with the same fields but different values for the following fields will fail as well: * name * create_time * update_time * labels * description

+

Creates a new `VpcFlowLogsConfig`. If a configuration with the exact same settings already exists (even if the ID is different), the creation fails. Notes: 1. Creating a configuration with `state=DISABLED` will fail 2. The following fields are not considered as settings for the purpose of the check mentioned above, therefore - creating another configuration with the same fields but different values for the following fields will fail as well: * name * create_time * update_time * labels * description

delete(name, x__xgafv=None)

Deletes a specific `VpcFlowLogsConfig`.

@@ -94,7 +94,7 @@

Instance Methods

Retrieves the next page of results.

patch(name, body=None, updateMask=None, x__xgafv=None)

-

Updates an existing `VpcFlowLogsConfig`. If a configuration with the exact same settings already exists (even if the ID is different), the creation fails. Notes: 1. Updating a configuration with state=DISABLED will fail 2. The following fields are not considered as `settings` for the purpose of the check mentioned above, therefore - updating another configuration with the same fields but different values for the following fields will fail as well: * name * create_time * update_time * labels * description

+

Updates an existing `VpcFlowLogsConfig`. If a configuration with the exact same settings already exists (even if the ID is different), the creation fails. Notes: 1. Updating a configuration with `state=DISABLED` will fail 2. The following fields are not considered as settings for the purpose of the check mentioned above, therefore - updating another configuration with the same fields but different values for the following fields will fail as well: * name * create_time * update_time * labels * description

Method Details

close() @@ -103,10 +103,10 @@

Method Details

create(parent, body=None, vpcFlowLogsConfigId=None, x__xgafv=None) -
Creates a new `VpcFlowLogsConfig`. If a configuration with the exact same settings already exists (even if the ID is different), the creation fails. Notes: 1. Creating a configuration with state=DISABLED will fail 2. The following fields are not considered as `settings` for the purpose of the check mentioned above, therefore - creating another configuration with the same fields but different values for the following fields will fail as well: * name * create_time * update_time * labels * description
+  
Creates a new `VpcFlowLogsConfig`. If a configuration with the exact same settings already exists (even if the ID is different), the creation fails. Notes: 1. Creating a configuration with `state=DISABLED` will fail 2. The following fields are not considered as settings for the purpose of the check mentioned above, therefore - creating another configuration with the same fields but different values for the following fields will fail as well: * name * create_time * update_time * labels * description
 
 Args:
-  parent: string, Required. The parent resource of the VPC Flow Logs configuration to create: `projects/{project_id}/locations/global` `organizations/{organization_id}/locations/global` (required)
+  parent: string, Required. The parent resource of the VpcFlowLogsConfig to create, in one of the following formats: - For project-level resources: `projects/{project_id}/locations/global` - For organization-level resources: `organizations/{organization_id}/locations/global` (required)
   body: object, The request body.
     The object takes the form of:
 
@@ -125,11 +125,11 @@ 

Method Details

"metadataFields": [ # Optional. Custom metadata fields to include in the reported VPC flow logs. Can only be specified if "metadata" was set to CUSTOM_METADATA. "A String", ], - "name": "A String", # Identifier. Unique name of the configuration using one of the forms: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` `organizations/{organization_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` + "name": "A String", # Identifier. Unique name of the configuration. The name can have one of the following forms: - For project-level configurations: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` - For organization-level configurations: `organizations/{organization_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` "network": "A String", # Traffic will be logged from VMs, VPN tunnels and Interconnect Attachments within the network. Format: projects/{project_id}/global/networks/{name} "state": "A String", # Optional. The state of the VPC Flow Log configuration. Default value is ENABLED. When creating a new configuration, it must be enabled. Setting state=DISABLED will pause the log generation for this config. "subnet": "A String", # Traffic will be logged from VMs within the subnetwork. Format: projects/{project_id}/regions/{region}/subnetworks/{name} - "targetResourceState": "A String", # Output only. A diagnostic bit - describes the state of the configured target resource for diagnostic purposes. + "targetResourceState": "A String", # Output only. Describes the state of the configured target resource for diagnostic purposes. "updateTime": "A String", # Output only. The time the config was updated. "vpnTunnel": "A String", # Traffic will be logged from the VPN Tunnel. Format: projects/{project_id}/regions/{region}/vpnTunnels/{name} } @@ -169,7 +169,7 @@

Method Details

Deletes a specific `VpcFlowLogsConfig`.
 
 Args:
-  name: string, Required. `VpcFlowLogsConfig` resource name using one of the form: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config}` `organizations/{organization_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config}` (required)
+  name: string, Required. The resource name of the VpcFlowLogsConfig, in one of the following formats: - For a project-level resource: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` - For an organization-level resource: `organizations/{organization_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` (required)
   x__xgafv: string, V1 error format.
     Allowed values
       1 - v1 error format
@@ -204,7 +204,7 @@ 

Method Details

Gets the details of a specific `VpcFlowLogsConfig`.
 
 Args:
-  name: string, Required. `VpcFlowLogsConfig` resource name using the form: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config}` `organizations/{organization_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config}` (required)
+  name: string, Required. The resource name of the VpcFlowLogsConfig, in one of the following formats: - For project-level resources: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` - For organization-level resources: `organizations/{organization_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` (required)
   x__xgafv: string, V1 error format.
     Allowed values
       1 - v1 error format
@@ -228,11 +228,11 @@ 

Method Details

"metadataFields": [ # Optional. Custom metadata fields to include in the reported VPC flow logs. Can only be specified if "metadata" was set to CUSTOM_METADATA. "A String", ], - "name": "A String", # Identifier. Unique name of the configuration using one of the forms: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` `organizations/{organization_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` + "name": "A String", # Identifier. Unique name of the configuration. The name can have one of the following forms: - For project-level configurations: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` - For organization-level configurations: `organizations/{organization_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` "network": "A String", # Traffic will be logged from VMs, VPN tunnels and Interconnect Attachments within the network. Format: projects/{project_id}/global/networks/{name} "state": "A String", # Optional. The state of the VPC Flow Log configuration. Default value is ENABLED. When creating a new configuration, it must be enabled. Setting state=DISABLED will pause the log generation for this config. "subnet": "A String", # Traffic will be logged from VMs within the subnetwork. Format: projects/{project_id}/regions/{region}/subnetworks/{name} - "targetResourceState": "A String", # Output only. A diagnostic bit - describes the state of the configured target resource for diagnostic purposes. + "targetResourceState": "A String", # Output only. Describes the state of the configured target resource for diagnostic purposes. "updateTime": "A String", # Output only. The time the config was updated. "vpnTunnel": "A String", # Traffic will be logged from the VPN Tunnel. Format: projects/{project_id}/regions/{region}/vpnTunnels/{name} }
@@ -243,7 +243,7 @@

Method Details

Lists all `VpcFlowLogsConfigs` in a given organization.
 
 Args:
-  parent: string, Required. The parent resource of the VpcFlowLogsConfig: `projects/{project_id}/locations/global` `organizations/{organization_id}/locations/global` (required)
+  parent: string, Required. The parent resource of the VpcFlowLogsConfig, in one of the following formats: - For project-level resourcs: `projects/{project_id}/locations/global` - For organization-level resources: `organizations/{organization_id}/locations/global` (required)
   filter: string, Optional. Lists the `VpcFlowLogsConfigs` that match the filter expression. A filter expression must use the supported [CEL logic operators] (https://cloud.google.com/vpc/docs/about-flow-logs-records#supported_cel_logic_operators).
   orderBy: string, Optional. Field to use to sort the list.
   pageSize: integer, Optional. Number of `VpcFlowLogsConfigs` to return.
@@ -277,11 +277,11 @@ 

Method Details

"metadataFields": [ # Optional. Custom metadata fields to include in the reported VPC flow logs. Can only be specified if "metadata" was set to CUSTOM_METADATA. "A String", ], - "name": "A String", # Identifier. Unique name of the configuration using one of the forms: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` `organizations/{organization_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` + "name": "A String", # Identifier. Unique name of the configuration. The name can have one of the following forms: - For project-level configurations: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` - For organization-level configurations: `organizations/{organization_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` "network": "A String", # Traffic will be logged from VMs, VPN tunnels and Interconnect Attachments within the network. Format: projects/{project_id}/global/networks/{name} "state": "A String", # Optional. The state of the VPC Flow Log configuration. Default value is ENABLED. When creating a new configuration, it must be enabled. Setting state=DISABLED will pause the log generation for this config. "subnet": "A String", # Traffic will be logged from VMs within the subnetwork. Format: projects/{project_id}/regions/{region}/subnetworks/{name} - "targetResourceState": "A String", # Output only. A diagnostic bit - describes the state of the configured target resource for diagnostic purposes. + "targetResourceState": "A String", # Output only. Describes the state of the configured target resource for diagnostic purposes. "updateTime": "A String", # Output only. The time the config was updated. "vpnTunnel": "A String", # Traffic will be logged from the VPN Tunnel. Format: projects/{project_id}/regions/{region}/vpnTunnels/{name} }, @@ -305,10 +305,10 @@

Method Details

patch(name, body=None, updateMask=None, x__xgafv=None) -
Updates an existing `VpcFlowLogsConfig`. If a configuration with the exact same settings already exists (even if the ID is different), the creation fails. Notes: 1. Updating a configuration with state=DISABLED will fail 2. The following fields are not considered as `settings` for the purpose of the check mentioned above, therefore - updating another configuration with the same fields but different values for the following fields will fail as well: * name * create_time * update_time * labels * description
+  
Updates an existing `VpcFlowLogsConfig`. If a configuration with the exact same settings already exists (even if the ID is different), the creation fails. Notes: 1. Updating a configuration with `state=DISABLED` will fail 2. The following fields are not considered as settings for the purpose of the check mentioned above, therefore - updating another configuration with the same fields but different values for the following fields will fail as well: * name * create_time * update_time * labels * description
 
 Args:
-  name: string, Identifier. Unique name of the configuration using one of the forms: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` `organizations/{organization_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` (required)
+  name: string, Identifier. Unique name of the configuration. The name can have one of the following forms: - For project-level configurations: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` - For organization-level configurations: `organizations/{organization_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` (required)
   body: object, The request body.
     The object takes the form of:
 
@@ -327,16 +327,16 @@ 

Method Details

"metadataFields": [ # Optional. Custom metadata fields to include in the reported VPC flow logs. Can only be specified if "metadata" was set to CUSTOM_METADATA. "A String", ], - "name": "A String", # Identifier. Unique name of the configuration using one of the forms: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` `organizations/{organization_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` + "name": "A String", # Identifier. Unique name of the configuration. The name can have one of the following forms: - For project-level configurations: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` - For organization-level configurations: `organizations/{organization_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` "network": "A String", # Traffic will be logged from VMs, VPN tunnels and Interconnect Attachments within the network. Format: projects/{project_id}/global/networks/{name} "state": "A String", # Optional. The state of the VPC Flow Log configuration. Default value is ENABLED. When creating a new configuration, it must be enabled. Setting state=DISABLED will pause the log generation for this config. "subnet": "A String", # Traffic will be logged from VMs within the subnetwork. Format: projects/{project_id}/regions/{region}/subnetworks/{name} - "targetResourceState": "A String", # Output only. A diagnostic bit - describes the state of the configured target resource for diagnostic purposes. + "targetResourceState": "A String", # Output only. Describes the state of the configured target resource for diagnostic purposes. "updateTime": "A String", # Output only. The time the config was updated. "vpnTunnel": "A String", # Traffic will be logged from the VPN Tunnel. Format: projects/{project_id}/regions/{region}/vpnTunnels/{name} } - updateMask: string, Required. Mask of fields to update. At least one path must be supplied in this field. + updateMask: string, Required. Mask of fields to update. At least one path must be supplied in this field. For example, to change the state of the configuration to ENABLED, specify `update_mask` = `"state"`, and the `vpc_flow_logs_config` would be: `vpc_flow_logs_config = { name = "projects/my-project/locations/global/vpcFlowLogsConfigs/my-config" state = "ENABLED" }` x__xgafv: string, V1 error format. Allowed values 1 - v1 error format diff --git a/docs/dyn/networkmanagement_v1beta1.projects.locations.vpcFlowLogsConfigs.html b/docs/dyn/networkmanagement_v1beta1.projects.locations.vpcFlowLogsConfigs.html index eabd5f81a0..d8f2af214f 100644 --- a/docs/dyn/networkmanagement_v1beta1.projects.locations.vpcFlowLogsConfigs.html +++ b/docs/dyn/networkmanagement_v1beta1.projects.locations.vpcFlowLogsConfigs.html @@ -79,7 +79,7 @@

Instance Methods

Close httplib2 connections.

create(parent, body=None, vpcFlowLogsConfigId=None, x__xgafv=None)

-

Creates a new `VpcFlowLogsConfig`. If a configuration with the exact same settings already exists (even if the ID is different), the creation fails. Notes: 1. Creating a configuration with state=DISABLED will fail 2. The following fields are not considered as `settings` for the purpose of the check mentioned above, therefore - creating another configuration with the same fields but different values for the following fields will fail as well: * name * create_time * update_time * labels * description

+

Creates a new `VpcFlowLogsConfig`. If a configuration with the exact same settings already exists (even if the ID is different), the creation fails. Notes: 1. Creating a configuration with `state=DISABLED` will fail 2. The following fields are not considered as settings for the purpose of the check mentioned above, therefore - creating another configuration with the same fields but different values for the following fields will fail as well: * name * create_time * update_time * labels * description

delete(name, x__xgafv=None)

Deletes a specific `VpcFlowLogsConfig`.

@@ -94,10 +94,10 @@

Instance Methods

Retrieves the next page of results.

patch(name, body=None, updateMask=None, x__xgafv=None)

-

Updates an existing `VpcFlowLogsConfig`. If a configuration with the exact same settings already exists (even if the ID is different), the creation fails. Notes: 1. Updating a configuration with state=DISABLED will fail 2. The following fields are not considered as `settings` for the purpose of the check mentioned above, therefore - updating another configuration with the same fields but different values for the following fields will fail as well: * name * create_time * update_time * labels * description

+

Updates an existing `VpcFlowLogsConfig`. If a configuration with the exact same settings already exists (even if the ID is different), the creation fails. Notes: 1. Updating a configuration with `state=DISABLED` will fail 2. The following fields are not considered as settings for the purpose of the check mentioned above, therefore - updating another configuration with the same fields but different values for the following fields will fail as well: * name * create_time * update_time * labels * description

queryOrgVpcFlowLogsConfigs(parent, filter=None, pageSize=None, pageToken=None, x__xgafv=None)

-

QueryOrgVpcFlowLogsConfigs lists Organization resources for a given project.

+

QueryOrgVpcFlowLogsConfigs returns a list of all organization-level VPC Flow Logs configurations applicable to the specified project.

queryOrgVpcFlowLogsConfigs_next()

Retrieves the next page of results.

@@ -109,10 +109,10 @@

Method Details

create(parent, body=None, vpcFlowLogsConfigId=None, x__xgafv=None) -
Creates a new `VpcFlowLogsConfig`. If a configuration with the exact same settings already exists (even if the ID is different), the creation fails. Notes: 1. Creating a configuration with state=DISABLED will fail 2. The following fields are not considered as `settings` for the purpose of the check mentioned above, therefore - creating another configuration with the same fields but different values for the following fields will fail as well: * name * create_time * update_time * labels * description
+  
Creates a new `VpcFlowLogsConfig`. If a configuration with the exact same settings already exists (even if the ID is different), the creation fails. Notes: 1. Creating a configuration with `state=DISABLED` will fail 2. The following fields are not considered as settings for the purpose of the check mentioned above, therefore - creating another configuration with the same fields but different values for the following fields will fail as well: * name * create_time * update_time * labels * description
 
 Args:
-  parent: string, Required. The parent resource of the VPC Flow Logs configuration to create: `projects/{project_id}/locations/global` `organizations/{organization_id}/locations/global` (required)
+  parent: string, Required. The parent resource of the VpcFlowLogsConfig to create, in one of the following formats: - For project-level resources: `projects/{project_id}/locations/global` - For organization-level resources: `organizations/{organization_id}/locations/global` (required)
   body: object, The request body.
     The object takes the form of:
 
@@ -131,11 +131,11 @@ 

Method Details

"metadataFields": [ # Optional. Custom metadata fields to include in the reported VPC flow logs. Can only be specified if "metadata" was set to CUSTOM_METADATA. "A String", ], - "name": "A String", # Identifier. Unique name of the configuration using one of the forms: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` `organizations/{organization_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` + "name": "A String", # Identifier. Unique name of the configuration. The name can have one of the following forms: - For project-level configurations: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` - For organization-level configurations: `organizations/{organization_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` "network": "A String", # Traffic will be logged from VMs, VPN tunnels and Interconnect Attachments within the network. Format: projects/{project_id}/global/networks/{name} "state": "A String", # Optional. The state of the VPC Flow Log configuration. Default value is ENABLED. When creating a new configuration, it must be enabled. Setting state=DISABLED will pause the log generation for this config. "subnet": "A String", # Traffic will be logged from VMs within the subnetwork. Format: projects/{project_id}/regions/{region}/subnetworks/{name} - "targetResourceState": "A String", # Output only. A diagnostic bit - describes the state of the configured target resource for diagnostic purposes. + "targetResourceState": "A String", # Output only. Describes the state of the configured target resource for diagnostic purposes. "updateTime": "A String", # Output only. The time the config was updated. "vpnTunnel": "A String", # Traffic will be logged from the VPN Tunnel. Format: projects/{project_id}/regions/{region}/vpnTunnels/{name} } @@ -175,7 +175,7 @@

Method Details

Deletes a specific `VpcFlowLogsConfig`.
 
 Args:
-  name: string, Required. `VpcFlowLogsConfig` resource name using one of the form: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config}` `organizations/{organization_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config}` (required)
+  name: string, Required. The resource name of the VpcFlowLogsConfig, in one of the following formats: - For a project-level resource: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` - For an organization-level resource: `organizations/{organization_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` (required)
   x__xgafv: string, V1 error format.
     Allowed values
       1 - v1 error format
@@ -210,7 +210,7 @@ 

Method Details

Gets the details of a specific `VpcFlowLogsConfig`.
 
 Args:
-  name: string, Required. `VpcFlowLogsConfig` resource name using the form: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config}` `organizations/{organization_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config}` (required)
+  name: string, Required. The resource name of the VpcFlowLogsConfig, in one of the following formats: - For project-level resources: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` - For organization-level resources: `organizations/{organization_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` (required)
   x__xgafv: string, V1 error format.
     Allowed values
       1 - v1 error format
@@ -234,11 +234,11 @@ 

Method Details

"metadataFields": [ # Optional. Custom metadata fields to include in the reported VPC flow logs. Can only be specified if "metadata" was set to CUSTOM_METADATA. "A String", ], - "name": "A String", # Identifier. Unique name of the configuration using one of the forms: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` `organizations/{organization_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` + "name": "A String", # Identifier. Unique name of the configuration. The name can have one of the following forms: - For project-level configurations: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` - For organization-level configurations: `organizations/{organization_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` "network": "A String", # Traffic will be logged from VMs, VPN tunnels and Interconnect Attachments within the network. Format: projects/{project_id}/global/networks/{name} "state": "A String", # Optional. The state of the VPC Flow Log configuration. Default value is ENABLED. When creating a new configuration, it must be enabled. Setting state=DISABLED will pause the log generation for this config. "subnet": "A String", # Traffic will be logged from VMs within the subnetwork. Format: projects/{project_id}/regions/{region}/subnetworks/{name} - "targetResourceState": "A String", # Output only. A diagnostic bit - describes the state of the configured target resource for diagnostic purposes. + "targetResourceState": "A String", # Output only. Describes the state of the configured target resource for diagnostic purposes. "updateTime": "A String", # Output only. The time the config was updated. "vpnTunnel": "A String", # Traffic will be logged from the VPN Tunnel. Format: projects/{project_id}/regions/{region}/vpnTunnels/{name} }
@@ -249,7 +249,7 @@

Method Details

Lists all `VpcFlowLogsConfigs` in a given project.
 
 Args:
-  parent: string, Required. The parent resource of the VpcFlowLogsConfig: `projects/{project_id}/locations/global` `organizations/{organization_id}/locations/global` (required)
+  parent: string, Required. The parent resource of the VpcFlowLogsConfig, in one of the following formats: - For project-level resourcs: `projects/{project_id}/locations/global` - For organization-level resources: `organizations/{organization_id}/locations/global` (required)
   filter: string, Optional. Lists the `VpcFlowLogsConfigs` that match the filter expression. A filter expression must use the supported [CEL logic operators] (https://cloud.google.com/vpc/docs/about-flow-logs-records#supported_cel_logic_operators).
   orderBy: string, Optional. Field to use to sort the list.
   pageSize: integer, Optional. Number of `VpcFlowLogsConfigs` to return.
@@ -283,11 +283,11 @@ 

Method Details

"metadataFields": [ # Optional. Custom metadata fields to include in the reported VPC flow logs. Can only be specified if "metadata" was set to CUSTOM_METADATA. "A String", ], - "name": "A String", # Identifier. Unique name of the configuration using one of the forms: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` `organizations/{organization_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` + "name": "A String", # Identifier. Unique name of the configuration. The name can have one of the following forms: - For project-level configurations: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` - For organization-level configurations: `organizations/{organization_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` "network": "A String", # Traffic will be logged from VMs, VPN tunnels and Interconnect Attachments within the network. Format: projects/{project_id}/global/networks/{name} "state": "A String", # Optional. The state of the VPC Flow Log configuration. Default value is ENABLED. When creating a new configuration, it must be enabled. Setting state=DISABLED will pause the log generation for this config. "subnet": "A String", # Traffic will be logged from VMs within the subnetwork. Format: projects/{project_id}/regions/{region}/subnetworks/{name} - "targetResourceState": "A String", # Output only. A diagnostic bit - describes the state of the configured target resource for diagnostic purposes. + "targetResourceState": "A String", # Output only. Describes the state of the configured target resource for diagnostic purposes. "updateTime": "A String", # Output only. The time the config was updated. "vpnTunnel": "A String", # Traffic will be logged from the VPN Tunnel. Format: projects/{project_id}/regions/{region}/vpnTunnels/{name} }, @@ -311,10 +311,10 @@

Method Details

patch(name, body=None, updateMask=None, x__xgafv=None) -
Updates an existing `VpcFlowLogsConfig`. If a configuration with the exact same settings already exists (even if the ID is different), the creation fails. Notes: 1. Updating a configuration with state=DISABLED will fail 2. The following fields are not considered as `settings` for the purpose of the check mentioned above, therefore - updating another configuration with the same fields but different values for the following fields will fail as well: * name * create_time * update_time * labels * description
+  
Updates an existing `VpcFlowLogsConfig`. If a configuration with the exact same settings already exists (even if the ID is different), the creation fails. Notes: 1. Updating a configuration with `state=DISABLED` will fail 2. The following fields are not considered as settings for the purpose of the check mentioned above, therefore - updating another configuration with the same fields but different values for the following fields will fail as well: * name * create_time * update_time * labels * description
 
 Args:
-  name: string, Identifier. Unique name of the configuration using one of the forms: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` `organizations/{organization_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` (required)
+  name: string, Identifier. Unique name of the configuration. The name can have one of the following forms: - For project-level configurations: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` - For organization-level configurations: `organizations/{organization_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` (required)
   body: object, The request body.
     The object takes the form of:
 
@@ -333,16 +333,16 @@ 

Method Details

"metadataFields": [ # Optional. Custom metadata fields to include in the reported VPC flow logs. Can only be specified if "metadata" was set to CUSTOM_METADATA. "A String", ], - "name": "A String", # Identifier. Unique name of the configuration using one of the forms: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` `organizations/{organization_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` + "name": "A String", # Identifier. Unique name of the configuration. The name can have one of the following forms: - For project-level configurations: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` - For organization-level configurations: `organizations/{organization_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` "network": "A String", # Traffic will be logged from VMs, VPN tunnels and Interconnect Attachments within the network. Format: projects/{project_id}/global/networks/{name} "state": "A String", # Optional. The state of the VPC Flow Log configuration. Default value is ENABLED. When creating a new configuration, it must be enabled. Setting state=DISABLED will pause the log generation for this config. "subnet": "A String", # Traffic will be logged from VMs within the subnetwork. Format: projects/{project_id}/regions/{region}/subnetworks/{name} - "targetResourceState": "A String", # Output only. A diagnostic bit - describes the state of the configured target resource for diagnostic purposes. + "targetResourceState": "A String", # Output only. Describes the state of the configured target resource for diagnostic purposes. "updateTime": "A String", # Output only. The time the config was updated. "vpnTunnel": "A String", # Traffic will be logged from the VPN Tunnel. Format: projects/{project_id}/regions/{region}/vpnTunnels/{name} } - updateMask: string, Required. Mask of fields to update. At least one path must be supplied in this field. + updateMask: string, Required. Mask of fields to update. At least one path must be supplied in this field. For example, to change the state of the configuration to ENABLED, specify `update_mask` = `"state"`, and the `vpc_flow_logs_config` would be: `vpc_flow_logs_config = { name = "projects/my-project/locations/global/vpcFlowLogsConfigs/my-config" state = "ENABLED" }` x__xgafv: string, V1 error format. Allowed values 1 - v1 error format @@ -374,10 +374,10 @@

Method Details

queryOrgVpcFlowLogsConfigs(parent, filter=None, pageSize=None, pageToken=None, x__xgafv=None) -
QueryOrgVpcFlowLogsConfigs lists Organization resources for a given project.
+  
QueryOrgVpcFlowLogsConfigs returns a list of all organization-level VPC Flow Logs configurations applicable to the specified project.
 
 Args:
-  parent: string, Required. The parent resource of the VpcFlowLogsConfig: `projects/{project_id}/locations/global` (required)
+  parent: string, Required. The parent resource of the VpcFlowLogsConfig, specified in the following format: `projects/{project_id}/locations/global` (required)
   filter: string, Optional. Lists the `VpcFlowLogsConfigs` that match the filter expression. A filter expression must use the supported [CEL logic operators] (https://cloud.google.com/vpc/docs/about-flow-logs-records#supported_cel_logic_operators).
   pageSize: integer, Optional. Number of `VpcFlowLogsConfigs` to return.
   pageToken: string, Optional. Page token from an earlier query, as returned in `next_page_token`.
@@ -410,11 +410,11 @@ 

Method Details

"metadataFields": [ # Optional. Custom metadata fields to include in the reported VPC flow logs. Can only be specified if "metadata" was set to CUSTOM_METADATA. "A String", ], - "name": "A String", # Identifier. Unique name of the configuration using one of the forms: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` `organizations/{organization_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` + "name": "A String", # Identifier. Unique name of the configuration. The name can have one of the following forms: - For project-level configurations: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` - For organization-level configurations: `organizations/{organization_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` "network": "A String", # Traffic will be logged from VMs, VPN tunnels and Interconnect Attachments within the network. Format: projects/{project_id}/global/networks/{name} "state": "A String", # Optional. The state of the VPC Flow Log configuration. Default value is ENABLED. When creating a new configuration, it must be enabled. Setting state=DISABLED will pause the log generation for this config. "subnet": "A String", # Traffic will be logged from VMs within the subnetwork. Format: projects/{project_id}/regions/{region}/subnetworks/{name} - "targetResourceState": "A String", # Output only. A diagnostic bit - describes the state of the configured target resource for diagnostic purposes. + "targetResourceState": "A String", # Output only. Describes the state of the configured target resource for diagnostic purposes. "updateTime": "A String", # Output only. The time the config was updated. "vpnTunnel": "A String", # Traffic will be logged from the VPN Tunnel. Format: projects/{project_id}/regions/{region}/vpnTunnels/{name} }, diff --git a/docs/dyn/networkservices_v1beta1.projects.locations.lbEdgeExtensions.html b/docs/dyn/networkservices_v1beta1.projects.locations.lbEdgeExtensions.html index d5e5ff84ee..0f1f1e5f03 100644 --- a/docs/dyn/networkservices_v1beta1.projects.locations.lbEdgeExtensions.html +++ b/docs/dyn/networkservices_v1beta1.projects.locations.lbEdgeExtensions.html @@ -110,7 +110,7 @@

Method Details

body: object, The request body. The object takes the form of: -{ # `LbEdgeExtension` is a resource that lets the extension service influence the Backend Service selection or Cloud CDN cache keys by modifying the request headers. +{ # `LbEdgeExtension` is a resource that lets the extension service influence the selection of backend services and Cloud CDN cache keys by modifying request headers. "createTime": "A String", # Output only. The timestamp when the resource was created. "description": "A String", # Optional. A human-readable description of the resource. "extensionChains": [ # Required. A set of ordered extension chains that contain the match conditions and extensions to execute. Match conditions for each extension chain are evaluated in sequence for a given request. The first extension chain that has a condition that matches the request is executed. Any subsequent extension chains do not execute. Limited to 5 extension chains per resource. @@ -234,7 +234,7 @@

Method Details

Returns: An object of the form: - { # `LbEdgeExtension` is a resource that lets the extension service influence the Backend Service selection or Cloud CDN cache keys by modifying the request headers. + { # `LbEdgeExtension` is a resource that lets the extension service influence the selection of backend services and Cloud CDN cache keys by modifying request headers. "createTime": "A String", # Output only. The timestamp when the resource was created. "description": "A String", # Optional. A human-readable description of the resource. "extensionChains": [ # Required. A set of ordered extension chains that contain the match conditions and extensions to execute. Match conditions for each extension chain are evaluated in sequence for a given request. The first extension chain that has a condition that matches the request is executed. Any subsequent extension chains do not execute. Limited to 5 extension chains per resource. @@ -298,7 +298,7 @@

Method Details

{ # Message for response to listing `LbEdgeExtension` resources. "lbEdgeExtensions": [ # The list of `LbEdgeExtension` resources. - { # `LbEdgeExtension` is a resource that lets the extension service influence the Backend Service selection or Cloud CDN cache keys by modifying the request headers. + { # `LbEdgeExtension` is a resource that lets the extension service influence the selection of backend services and Cloud CDN cache keys by modifying request headers. "createTime": "A String", # Output only. The timestamp when the resource was created. "description": "A String", # Optional. A human-readable description of the resource. "extensionChains": [ # Required. A set of ordered extension chains that contain the match conditions and extensions to execute. Match conditions for each extension chain are evaluated in sequence for a given request. The first extension chain that has a condition that matches the request is executed. Any subsequent extension chains do not execute. Limited to 5 extension chains per resource. @@ -371,7 +371,7 @@

Method Details

body: object, The request body. The object takes the form of: -{ # `LbEdgeExtension` is a resource that lets the extension service influence the Backend Service selection or Cloud CDN cache keys by modifying the request headers. +{ # `LbEdgeExtension` is a resource that lets the extension service influence the selection of backend services and Cloud CDN cache keys by modifying request headers. "createTime": "A String", # Output only. The timestamp when the resource was created. "description": "A String", # Optional. A human-readable description of the resource. "extensionChains": [ # Required. A set of ordered extension chains that contain the match conditions and extensions to execute. Match conditions for each extension chain are evaluated in sequence for a given request. The first extension chain that has a condition that matches the request is executed. Any subsequent extension chains do not execute. Limited to 5 extension chains per resource. diff --git a/docs/dyn/places_v1.places.html b/docs/dyn/places_v1.places.html index 61a549194b..135763c874 100644 --- a/docs/dyn/places_v1.places.html +++ b/docs/dyn/places_v1.places.html @@ -336,7 +336,7 @@

Method Details

"nextCloseTime": "A String", # The next time the current opening hours period ends up to 7 days in the future. This field is only populated if the opening hours period is active at the time of serving the request. "nextOpenTime": "A String", # The next time the current opening hours period starts up to 7 days in the future. This field is only populated if the opening hours period is not active at the time of serving the request. "openNow": True or False, # Whether the opening hours period is currently active. For regular opening hours and current opening hours, this field means whether the place is open. For secondary opening hours and current secondary opening hours, this field means whether the secondary hours of this place is active. - "periods": [ # NOTE: The ordering of the `periods` array is independent of the ordering of the `weekday_descriptions` array. Do not assume they will begin on the same day. + "periods": [ # The periods that this place is open during the week. The periods are in chronological order, in the place-local timezone. An empty (but not absent) value indicates a place that is never open, e.g. because it is closed temporarily for renovations. The starting day of `periods` is NOT fixed and should not be assumed to be Sunday. The API determines the start day based on a variety of factors. For example, for a 24/7 business, the first period may begin on the day of the request. For other businesses, it might be the first day of the week that they are open. NOTE: The ordering of the `periods` array is independent of the ordering of the `weekday_descriptions` array. Do not assume they will begin on the same day. { # A period the place remains in open_now status. "close": { # Status changing points. # The time that the place starts to be closed. "date": { # Represents a whole or partial calendar date, such as a birthday. The time of day and time zone are either specified elsewhere or are insignificant. The date is relative to the Gregorian Calendar. This can represent one of the following: * A full date, with non-zero year, month, and day values. * A month and day, with a zero year (for example, an anniversary). * A year on its own, with a zero month and a zero day. * A year and month, with a zero day (for example, a credit card expiration date). Related types: * google.type.TimeOfDay * google.type.DateTime * google.protobuf.Timestamp # Date in the local timezone for the place. @@ -381,7 +381,7 @@

Method Details

"nextCloseTime": "A String", # The next time the current opening hours period ends up to 7 days in the future. This field is only populated if the opening hours period is active at the time of serving the request. "nextOpenTime": "A String", # The next time the current opening hours period starts up to 7 days in the future. This field is only populated if the opening hours period is not active at the time of serving the request. "openNow": True or False, # Whether the opening hours period is currently active. For regular opening hours and current opening hours, this field means whether the place is open. For secondary opening hours and current secondary opening hours, this field means whether the secondary hours of this place is active. - "periods": [ # NOTE: The ordering of the `periods` array is independent of the ordering of the `weekday_descriptions` array. Do not assume they will begin on the same day. + "periods": [ # The periods that this place is open during the week. The periods are in chronological order, in the place-local timezone. An empty (but not absent) value indicates a place that is never open, e.g. because it is closed temporarily for renovations. The starting day of `periods` is NOT fixed and should not be assumed to be Sunday. The API determines the start day based on a variety of factors. For example, for a 24/7 business, the first period may begin on the day of the request. For other businesses, it might be the first day of the week that they are open. NOTE: The ordering of the `periods` array is independent of the ordering of the `weekday_descriptions` array. Do not assume they will begin on the same day. { # A period the place remains in open_now status. "close": { # Status changing points. # The time that the place starts to be closed. "date": { # Represents a whole or partial calendar date, such as a birthday. The time of day and time zone are either specified elsewhere or are insignificant. The date is relative to the Gregorian Calendar. This can represent one of the following: * A full date, with non-zero year, month, and day values. * A month and day, with a zero year (for example, an anniversary). * A year on its own, with a zero month and a zero day. * A year and month, with a zero day (for example, a credit card expiration date). Related types: * google.type.TimeOfDay * google.type.DateTime * google.protobuf.Timestamp # Date in the local timezone for the place. @@ -627,8 +627,8 @@

Method Details

"units": "A String", # The whole units of the amount. For example if `currencyCode` is `"USD"`, then 1 unit is one US dollar. }, }, - "primaryType": "A String", # The primary type of the given result. This type must one of the Places API supported types. For example, "restaurant", "cafe", "airport", etc. A place can only have a single primary type. For the complete list of possible values, see Table A and Table B at https://developers.google.com/maps/documentation/places/web-service/place-types - "primaryTypeDisplayName": { # Localized variant of a text in a particular language. # The display name of the primary type, localized to the request language if applicable. For the complete list of possible values, see Table A and Table B at https://developers.google.com/maps/documentation/places/web-service/place-types + "primaryType": "A String", # The primary type of the given result. This type must be one of the Places API supported types. For example, "restaurant", "cafe", "airport", etc. A place can only have a single primary type. For the complete list of possible values, see Table A and Table B at https://developers.google.com/maps/documentation/places/web-service/place-types. The primary type may be missing if the place's primary type is not a supported type. When a primary type is present, it is always one of the types in the `types` field. + "primaryTypeDisplayName": { # Localized variant of a text in a particular language. # The display name of the primary type, localized to the request language if applicable. For the complete list of possible values, see Table A and Table B at https://developers.google.com/maps/documentation/places/web-service/place-types. The primary type may be missing if the place's primary type is not a supported type. "languageCode": "A String", # The text's BCP-47 language code, such as "en-US" or "sr-Latn". For more information, see http://www.unicode.org/reports/tr35/#Unicode_locale_identifier. "text": "A String", # Localized string in the language corresponding to language_code below. }, @@ -638,7 +638,7 @@

Method Details

"nextCloseTime": "A String", # The next time the current opening hours period ends up to 7 days in the future. This field is only populated if the opening hours period is active at the time of serving the request. "nextOpenTime": "A String", # The next time the current opening hours period starts up to 7 days in the future. This field is only populated if the opening hours period is not active at the time of serving the request. "openNow": True or False, # Whether the opening hours period is currently active. For regular opening hours and current opening hours, this field means whether the place is open. For secondary opening hours and current secondary opening hours, this field means whether the secondary hours of this place is active. - "periods": [ # NOTE: The ordering of the `periods` array is independent of the ordering of the `weekday_descriptions` array. Do not assume they will begin on the same day. + "periods": [ # The periods that this place is open during the week. The periods are in chronological order, in the place-local timezone. An empty (but not absent) value indicates a place that is never open, e.g. because it is closed temporarily for renovations. The starting day of `periods` is NOT fixed and should not be assumed to be Sunday. The API determines the start day based on a variety of factors. For example, for a 24/7 business, the first period may begin on the day of the request. For other businesses, it might be the first day of the week that they are open. NOTE: The ordering of the `periods` array is independent of the ordering of the `weekday_descriptions` array. Do not assume they will begin on the same day. { # A period the place remains in open_now status. "close": { # Status changing points. # The time that the place starts to be closed. "date": { # Represents a whole or partial calendar date, such as a birthday. The time of day and time zone are either specified elsewhere or are insignificant. The date is relative to the Gregorian Calendar. This can represent one of the following: * A full date, with non-zero year, month, and day values. * A month and day, with a zero year (for example, an anniversary). * A year on its own, with a zero month and a zero day. * A year and month, with a zero day (for example, a credit card expiration date). Related types: * google.type.TimeOfDay * google.type.DateTime * google.protobuf.Timestamp # Date in the local timezone for the place. @@ -683,7 +683,7 @@

Method Details

"nextCloseTime": "A String", # The next time the current opening hours period ends up to 7 days in the future. This field is only populated if the opening hours period is active at the time of serving the request. "nextOpenTime": "A String", # The next time the current opening hours period starts up to 7 days in the future. This field is only populated if the opening hours period is not active at the time of serving the request. "openNow": True or False, # Whether the opening hours period is currently active. For regular opening hours and current opening hours, this field means whether the place is open. For secondary opening hours and current secondary opening hours, this field means whether the secondary hours of this place is active. - "periods": [ # NOTE: The ordering of the `periods` array is independent of the ordering of the `weekday_descriptions` array. Do not assume they will begin on the same day. + "periods": [ # The periods that this place is open during the week. The periods are in chronological order, in the place-local timezone. An empty (but not absent) value indicates a place that is never open, e.g. because it is closed temporarily for renovations. The starting day of `periods` is NOT fixed and should not be assumed to be Sunday. The API determines the start day based on a variety of factors. For example, for a 24/7 business, the first period may begin on the day of the request. For other businesses, it might be the first day of the week that they are open. NOTE: The ordering of the `periods` array is independent of the ordering of the `weekday_descriptions` array. Do not assume they will begin on the same day. { # A period the place remains in open_now status. "close": { # Status changing points. # The time that the place starts to be closed. "date": { # Represents a whole or partial calendar date, such as a birthday. The time of day and time zone are either specified elsewhere or are insignificant. The date is relative to the Gregorian Calendar. This can represent one of the following: * A full date, with non-zero year, month, and day values. * A month and day, with a zero year (for example, an anniversary). * A year on its own, with a zero month and a zero day. * A year and month, with a zero day (for example, a credit card expiration date). Related types: * google.type.TimeOfDay * google.type.DateTime * google.protobuf.Timestamp # Date in the local timezone for the place. @@ -928,7 +928,7 @@

Method Details

"nextCloseTime": "A String", # The next time the current opening hours period ends up to 7 days in the future. This field is only populated if the opening hours period is active at the time of serving the request. "nextOpenTime": "A String", # The next time the current opening hours period starts up to 7 days in the future. This field is only populated if the opening hours period is not active at the time of serving the request. "openNow": True or False, # Whether the opening hours period is currently active. For regular opening hours and current opening hours, this field means whether the place is open. For secondary opening hours and current secondary opening hours, this field means whether the secondary hours of this place is active. - "periods": [ # NOTE: The ordering of the `periods` array is independent of the ordering of the `weekday_descriptions` array. Do not assume they will begin on the same day. + "periods": [ # The periods that this place is open during the week. The periods are in chronological order, in the place-local timezone. An empty (but not absent) value indicates a place that is never open, e.g. because it is closed temporarily for renovations. The starting day of `periods` is NOT fixed and should not be assumed to be Sunday. The API determines the start day based on a variety of factors. For example, for a 24/7 business, the first period may begin on the day of the request. For other businesses, it might be the first day of the week that they are open. NOTE: The ordering of the `periods` array is independent of the ordering of the `weekday_descriptions` array. Do not assume they will begin on the same day. { # A period the place remains in open_now status. "close": { # Status changing points. # The time that the place starts to be closed. "date": { # Represents a whole or partial calendar date, such as a birthday. The time of day and time zone are either specified elsewhere or are insignificant. The date is relative to the Gregorian Calendar. This can represent one of the following: * A full date, with non-zero year, month, and day values. * A month and day, with a zero year (for example, an anniversary). * A year on its own, with a zero month and a zero day. * A year and month, with a zero day (for example, a credit card expiration date). Related types: * google.type.TimeOfDay * google.type.DateTime * google.protobuf.Timestamp # Date in the local timezone for the place. @@ -973,7 +973,7 @@

Method Details

"nextCloseTime": "A String", # The next time the current opening hours period ends up to 7 days in the future. This field is only populated if the opening hours period is active at the time of serving the request. "nextOpenTime": "A String", # The next time the current opening hours period starts up to 7 days in the future. This field is only populated if the opening hours period is not active at the time of serving the request. "openNow": True or False, # Whether the opening hours period is currently active. For regular opening hours and current opening hours, this field means whether the place is open. For secondary opening hours and current secondary opening hours, this field means whether the secondary hours of this place is active. - "periods": [ # NOTE: The ordering of the `periods` array is independent of the ordering of the `weekday_descriptions` array. Do not assume they will begin on the same day. + "periods": [ # The periods that this place is open during the week. The periods are in chronological order, in the place-local timezone. An empty (but not absent) value indicates a place that is never open, e.g. because it is closed temporarily for renovations. The starting day of `periods` is NOT fixed and should not be assumed to be Sunday. The API determines the start day based on a variety of factors. For example, for a 24/7 business, the first period may begin on the day of the request. For other businesses, it might be the first day of the week that they are open. NOTE: The ordering of the `periods` array is independent of the ordering of the `weekday_descriptions` array. Do not assume they will begin on the same day. { # A period the place remains in open_now status. "close": { # Status changing points. # The time that the place starts to be closed. "date": { # Represents a whole or partial calendar date, such as a birthday. The time of day and time zone are either specified elsewhere or are insignificant. The date is relative to the Gregorian Calendar. This can represent one of the following: * A full date, with non-zero year, month, and day values. * A month and day, with a zero year (for example, an anniversary). * A year on its own, with a zero month and a zero day. * A year and month, with a zero day (for example, a credit card expiration date). Related types: * google.type.TimeOfDay * google.type.DateTime * google.protobuf.Timestamp # Date in the local timezone for the place. @@ -1219,8 +1219,8 @@

Method Details

"units": "A String", # The whole units of the amount. For example if `currencyCode` is `"USD"`, then 1 unit is one US dollar. }, }, - "primaryType": "A String", # The primary type of the given result. This type must one of the Places API supported types. For example, "restaurant", "cafe", "airport", etc. A place can only have a single primary type. For the complete list of possible values, see Table A and Table B at https://developers.google.com/maps/documentation/places/web-service/place-types - "primaryTypeDisplayName": { # Localized variant of a text in a particular language. # The display name of the primary type, localized to the request language if applicable. For the complete list of possible values, see Table A and Table B at https://developers.google.com/maps/documentation/places/web-service/place-types + "primaryType": "A String", # The primary type of the given result. This type must be one of the Places API supported types. For example, "restaurant", "cafe", "airport", etc. A place can only have a single primary type. For the complete list of possible values, see Table A and Table B at https://developers.google.com/maps/documentation/places/web-service/place-types. The primary type may be missing if the place's primary type is not a supported type. When a primary type is present, it is always one of the types in the `types` field. + "primaryTypeDisplayName": { # Localized variant of a text in a particular language. # The display name of the primary type, localized to the request language if applicable. For the complete list of possible values, see Table A and Table B at https://developers.google.com/maps/documentation/places/web-service/place-types. The primary type may be missing if the place's primary type is not a supported type. "languageCode": "A String", # The text's BCP-47 language code, such as "en-US" or "sr-Latn". For more information, see http://www.unicode.org/reports/tr35/#Unicode_locale_identifier. "text": "A String", # Localized string in the language corresponding to language_code below. }, @@ -1230,7 +1230,7 @@

Method Details

"nextCloseTime": "A String", # The next time the current opening hours period ends up to 7 days in the future. This field is only populated if the opening hours period is active at the time of serving the request. "nextOpenTime": "A String", # The next time the current opening hours period starts up to 7 days in the future. This field is only populated if the opening hours period is not active at the time of serving the request. "openNow": True or False, # Whether the opening hours period is currently active. For regular opening hours and current opening hours, this field means whether the place is open. For secondary opening hours and current secondary opening hours, this field means whether the secondary hours of this place is active. - "periods": [ # NOTE: The ordering of the `periods` array is independent of the ordering of the `weekday_descriptions` array. Do not assume they will begin on the same day. + "periods": [ # The periods that this place is open during the week. The periods are in chronological order, in the place-local timezone. An empty (but not absent) value indicates a place that is never open, e.g. because it is closed temporarily for renovations. The starting day of `periods` is NOT fixed and should not be assumed to be Sunday. The API determines the start day based on a variety of factors. For example, for a 24/7 business, the first period may begin on the day of the request. For other businesses, it might be the first day of the week that they are open. NOTE: The ordering of the `periods` array is independent of the ordering of the `weekday_descriptions` array. Do not assume they will begin on the same day. { # A period the place remains in open_now status. "close": { # Status changing points. # The time that the place starts to be closed. "date": { # Represents a whole or partial calendar date, such as a birthday. The time of day and time zone are either specified elsewhere or are insignificant. The date is relative to the Gregorian Calendar. This can represent one of the following: * A full date, with non-zero year, month, and day values. * A month and day, with a zero year (for example, an anniversary). * A year on its own, with a zero month and a zero day. * A year and month, with a zero day (for example, a credit card expiration date). Related types: * google.type.TimeOfDay * google.type.DateTime * google.protobuf.Timestamp # Date in the local timezone for the place. @@ -1275,7 +1275,7 @@

Method Details

"nextCloseTime": "A String", # The next time the current opening hours period ends up to 7 days in the future. This field is only populated if the opening hours period is active at the time of serving the request. "nextOpenTime": "A String", # The next time the current opening hours period starts up to 7 days in the future. This field is only populated if the opening hours period is not active at the time of serving the request. "openNow": True or False, # Whether the opening hours period is currently active. For regular opening hours and current opening hours, this field means whether the place is open. For secondary opening hours and current secondary opening hours, this field means whether the secondary hours of this place is active. - "periods": [ # NOTE: The ordering of the `periods` array is independent of the ordering of the `weekday_descriptions` array. Do not assume they will begin on the same day. + "periods": [ # The periods that this place is open during the week. The periods are in chronological order, in the place-local timezone. An empty (but not absent) value indicates a place that is never open, e.g. because it is closed temporarily for renovations. The starting day of `periods` is NOT fixed and should not be assumed to be Sunday. The API determines the start day based on a variety of factors. For example, for a 24/7 business, the first period may begin on the day of the request. For other businesses, it might be the first day of the week that they are open. NOTE: The ordering of the `periods` array is independent of the ordering of the `weekday_descriptions` array. Do not assume they will begin on the same day. { # A period the place remains in open_now status. "close": { # Status changing points. # The time that the place starts to be closed. "date": { # Represents a whole or partial calendar date, such as a birthday. The time of day and time zone are either specified elsewhere or are insignificant. The date is relative to the Gregorian Calendar. This can represent one of the following: * A full date, with non-zero year, month, and day values. * A month and day, with a zero year (for example, an anniversary). * A year on its own, with a zero month and a zero day. * A year and month, with a zero day (for example, a credit card expiration date). Related types: * google.type.TimeOfDay * google.type.DateTime * google.protobuf.Timestamp # Date in the local timezone for the place. @@ -1650,7 +1650,7 @@

Method Details

"nextCloseTime": "A String", # The next time the current opening hours period ends up to 7 days in the future. This field is only populated if the opening hours period is active at the time of serving the request. "nextOpenTime": "A String", # The next time the current opening hours period starts up to 7 days in the future. This field is only populated if the opening hours period is not active at the time of serving the request. "openNow": True or False, # Whether the opening hours period is currently active. For regular opening hours and current opening hours, this field means whether the place is open. For secondary opening hours and current secondary opening hours, this field means whether the secondary hours of this place is active. - "periods": [ # NOTE: The ordering of the `periods` array is independent of the ordering of the `weekday_descriptions` array. Do not assume they will begin on the same day. + "periods": [ # The periods that this place is open during the week. The periods are in chronological order, in the place-local timezone. An empty (but not absent) value indicates a place that is never open, e.g. because it is closed temporarily for renovations. The starting day of `periods` is NOT fixed and should not be assumed to be Sunday. The API determines the start day based on a variety of factors. For example, for a 24/7 business, the first period may begin on the day of the request. For other businesses, it might be the first day of the week that they are open. NOTE: The ordering of the `periods` array is independent of the ordering of the `weekday_descriptions` array. Do not assume they will begin on the same day. { # A period the place remains in open_now status. "close": { # Status changing points. # The time that the place starts to be closed. "date": { # Represents a whole or partial calendar date, such as a birthday. The time of day and time zone are either specified elsewhere or are insignificant. The date is relative to the Gregorian Calendar. This can represent one of the following: * A full date, with non-zero year, month, and day values. * A month and day, with a zero year (for example, an anniversary). * A year on its own, with a zero month and a zero day. * A year and month, with a zero day (for example, a credit card expiration date). Related types: * google.type.TimeOfDay * google.type.DateTime * google.protobuf.Timestamp # Date in the local timezone for the place. @@ -1695,7 +1695,7 @@

Method Details

"nextCloseTime": "A String", # The next time the current opening hours period ends up to 7 days in the future. This field is only populated if the opening hours period is active at the time of serving the request. "nextOpenTime": "A String", # The next time the current opening hours period starts up to 7 days in the future. This field is only populated if the opening hours period is not active at the time of serving the request. "openNow": True or False, # Whether the opening hours period is currently active. For regular opening hours and current opening hours, this field means whether the place is open. For secondary opening hours and current secondary opening hours, this field means whether the secondary hours of this place is active. - "periods": [ # NOTE: The ordering of the `periods` array is independent of the ordering of the `weekday_descriptions` array. Do not assume they will begin on the same day. + "periods": [ # The periods that this place is open during the week. The periods are in chronological order, in the place-local timezone. An empty (but not absent) value indicates a place that is never open, e.g. because it is closed temporarily for renovations. The starting day of `periods` is NOT fixed and should not be assumed to be Sunday. The API determines the start day based on a variety of factors. For example, for a 24/7 business, the first period may begin on the day of the request. For other businesses, it might be the first day of the week that they are open. NOTE: The ordering of the `periods` array is independent of the ordering of the `weekday_descriptions` array. Do not assume they will begin on the same day. { # A period the place remains in open_now status. "close": { # Status changing points. # The time that the place starts to be closed. "date": { # Represents a whole or partial calendar date, such as a birthday. The time of day and time zone are either specified elsewhere or are insignificant. The date is relative to the Gregorian Calendar. This can represent one of the following: * A full date, with non-zero year, month, and day values. * A month and day, with a zero year (for example, an anniversary). * A year on its own, with a zero month and a zero day. * A year and month, with a zero day (for example, a credit card expiration date). Related types: * google.type.TimeOfDay * google.type.DateTime * google.protobuf.Timestamp # Date in the local timezone for the place. @@ -1941,8 +1941,8 @@

Method Details

"units": "A String", # The whole units of the amount. For example if `currencyCode` is `"USD"`, then 1 unit is one US dollar. }, }, - "primaryType": "A String", # The primary type of the given result. This type must one of the Places API supported types. For example, "restaurant", "cafe", "airport", etc. A place can only have a single primary type. For the complete list of possible values, see Table A and Table B at https://developers.google.com/maps/documentation/places/web-service/place-types - "primaryTypeDisplayName": { # Localized variant of a text in a particular language. # The display name of the primary type, localized to the request language if applicable. For the complete list of possible values, see Table A and Table B at https://developers.google.com/maps/documentation/places/web-service/place-types + "primaryType": "A String", # The primary type of the given result. This type must be one of the Places API supported types. For example, "restaurant", "cafe", "airport", etc. A place can only have a single primary type. For the complete list of possible values, see Table A and Table B at https://developers.google.com/maps/documentation/places/web-service/place-types. The primary type may be missing if the place's primary type is not a supported type. When a primary type is present, it is always one of the types in the `types` field. + "primaryTypeDisplayName": { # Localized variant of a text in a particular language. # The display name of the primary type, localized to the request language if applicable. For the complete list of possible values, see Table A and Table B at https://developers.google.com/maps/documentation/places/web-service/place-types. The primary type may be missing if the place's primary type is not a supported type. "languageCode": "A String", # The text's BCP-47 language code, such as "en-US" or "sr-Latn". For more information, see http://www.unicode.org/reports/tr35/#Unicode_locale_identifier. "text": "A String", # Localized string in the language corresponding to language_code below. }, @@ -1952,7 +1952,7 @@

Method Details

"nextCloseTime": "A String", # The next time the current opening hours period ends up to 7 days in the future. This field is only populated if the opening hours period is active at the time of serving the request. "nextOpenTime": "A String", # The next time the current opening hours period starts up to 7 days in the future. This field is only populated if the opening hours period is not active at the time of serving the request. "openNow": True or False, # Whether the opening hours period is currently active. For regular opening hours and current opening hours, this field means whether the place is open. For secondary opening hours and current secondary opening hours, this field means whether the secondary hours of this place is active. - "periods": [ # NOTE: The ordering of the `periods` array is independent of the ordering of the `weekday_descriptions` array. Do not assume they will begin on the same day. + "periods": [ # The periods that this place is open during the week. The periods are in chronological order, in the place-local timezone. An empty (but not absent) value indicates a place that is never open, e.g. because it is closed temporarily for renovations. The starting day of `periods` is NOT fixed and should not be assumed to be Sunday. The API determines the start day based on a variety of factors. For example, for a 24/7 business, the first period may begin on the day of the request. For other businesses, it might be the first day of the week that they are open. NOTE: The ordering of the `periods` array is independent of the ordering of the `weekday_descriptions` array. Do not assume they will begin on the same day. { # A period the place remains in open_now status. "close": { # Status changing points. # The time that the place starts to be closed. "date": { # Represents a whole or partial calendar date, such as a birthday. The time of day and time zone are either specified elsewhere or are insignificant. The date is relative to the Gregorian Calendar. This can represent one of the following: * A full date, with non-zero year, month, and day values. * A month and day, with a zero year (for example, an anniversary). * A year on its own, with a zero month and a zero day. * A year and month, with a zero day (for example, a credit card expiration date). Related types: * google.type.TimeOfDay * google.type.DateTime * google.protobuf.Timestamp # Date in the local timezone for the place. @@ -1997,7 +1997,7 @@

Method Details

"nextCloseTime": "A String", # The next time the current opening hours period ends up to 7 days in the future. This field is only populated if the opening hours period is active at the time of serving the request. "nextOpenTime": "A String", # The next time the current opening hours period starts up to 7 days in the future. This field is only populated if the opening hours period is not active at the time of serving the request. "openNow": True or False, # Whether the opening hours period is currently active. For regular opening hours and current opening hours, this field means whether the place is open. For secondary opening hours and current secondary opening hours, this field means whether the secondary hours of this place is active. - "periods": [ # NOTE: The ordering of the `periods` array is independent of the ordering of the `weekday_descriptions` array. Do not assume they will begin on the same day. + "periods": [ # The periods that this place is open during the week. The periods are in chronological order, in the place-local timezone. An empty (but not absent) value indicates a place that is never open, e.g. because it is closed temporarily for renovations. The starting day of `periods` is NOT fixed and should not be assumed to be Sunday. The API determines the start day based on a variety of factors. For example, for a 24/7 business, the first period may begin on the day of the request. For other businesses, it might be the first day of the week that they are open. NOTE: The ordering of the `periods` array is independent of the ordering of the `weekday_descriptions` array. Do not assume they will begin on the same day. { # A period the place remains in open_now status. "close": { # Status changing points. # The time that the place starts to be closed. "date": { # Represents a whole or partial calendar date, such as a birthday. The time of day and time zone are either specified elsewhere or are insignificant. The date is relative to the Gregorian Calendar. This can represent one of the following: * A full date, with non-zero year, month, and day values. * A month and day, with a zero year (for example, an anniversary). * A year on its own, with a zero month and a zero day. * A year and month, with a zero day (for example, a credit card expiration date). Related types: * google.type.TimeOfDay * google.type.DateTime * google.protobuf.Timestamp # Date in the local timezone for the place. diff --git a/docs/dyn/policysimulator_v1.folders.locations.accessPolicySimulations.html b/docs/dyn/policysimulator_v1.folders.locations.accessPolicySimulations.html new file mode 100644 index 0000000000..039da3ec43 --- /dev/null +++ b/docs/dyn/policysimulator_v1.folders.locations.accessPolicySimulations.html @@ -0,0 +1,91 @@ + + + +

Policy Simulator API . folders . locations . accessPolicySimulations

+

Instance Methods

+

+ operations() +

+

Returns the operations Resource.

+ +

+ close()

+

Close httplib2 connections.

+

Method Details

+
+ close() +
Close httplib2 connections.
+
+ + \ No newline at end of file diff --git a/docs/dyn/policysimulator_v1.folders.locations.accessPolicySimulations.operations.html b/docs/dyn/policysimulator_v1.folders.locations.accessPolicySimulations.operations.html new file mode 100644 index 0000000000..22807101bf --- /dev/null +++ b/docs/dyn/policysimulator_v1.folders.locations.accessPolicySimulations.operations.html @@ -0,0 +1,124 @@ + + + +

Policy Simulator API . folders . locations . accessPolicySimulations . operations

+

Instance Methods

+

+ close()

+

Close httplib2 connections.

+

+ get(name, x__xgafv=None)

+

Gets the latest state of a long-running operation. Clients can use this method to poll the operation result at intervals as recommended by the API service.

+

Method Details

+
+ close() +
Close httplib2 connections.
+
+ +
+ get(name, x__xgafv=None) +
Gets the latest state of a long-running operation. Clients can use this method to poll the operation result at intervals as recommended by the API service.
+
+Args:
+  name: string, The name of the operation resource. (required)
+  x__xgafv: string, V1 error format.
+    Allowed values
+      1 - v1 error format
+      2 - v2 error format
+
+Returns:
+  An object of the form:
+
+    { # This resource represents a long-running operation that is the result of a network API call.
+  "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available.
+  "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation.
+    "code": 42, # The status code, which should be an enum value of google.rpc.Code.
+    "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use.
+      {
+        "a_key": "", # Properties of the object. Contains field @type with type URL.
+      },
+    ],
+    "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
+  },
+  "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+  "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`.
+  "response": { # The normal, successful response of the operation. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+}
+
+ + \ No newline at end of file diff --git a/docs/dyn/policysimulator_v1.folders.locations.html b/docs/dyn/policysimulator_v1.folders.locations.html index 349d2834ca..0e4220713a 100644 --- a/docs/dyn/policysimulator_v1.folders.locations.html +++ b/docs/dyn/policysimulator_v1.folders.locations.html @@ -74,6 +74,11 @@

Policy Simulator API . folders . locations

Instance Methods

+

+ accessPolicySimulations() +

+

Returns the accessPolicySimulations Resource.

+

orgPolicyViolationsPreviews()

diff --git a/docs/dyn/policysimulator_v1.organizations.locations.accessPolicySimulations.html b/docs/dyn/policysimulator_v1.organizations.locations.accessPolicySimulations.html new file mode 100644 index 0000000000..24830d0126 --- /dev/null +++ b/docs/dyn/policysimulator_v1.organizations.locations.accessPolicySimulations.html @@ -0,0 +1,91 @@ + + + +

Policy Simulator API . organizations . locations . accessPolicySimulations

+

Instance Methods

+

+ operations() +

+

Returns the operations Resource.

+ +

+ close()

+

Close httplib2 connections.

+

Method Details

+
+ close() +
Close httplib2 connections.
+
+ + \ No newline at end of file diff --git a/docs/dyn/policysimulator_v1.organizations.locations.accessPolicySimulations.operations.html b/docs/dyn/policysimulator_v1.organizations.locations.accessPolicySimulations.operations.html new file mode 100644 index 0000000000..69f41204ac --- /dev/null +++ b/docs/dyn/policysimulator_v1.organizations.locations.accessPolicySimulations.operations.html @@ -0,0 +1,124 @@ + + + +

Policy Simulator API . organizations . locations . accessPolicySimulations . operations

+

Instance Methods

+

+ close()

+

Close httplib2 connections.

+

+ get(name, x__xgafv=None)

+

Gets the latest state of a long-running operation. Clients can use this method to poll the operation result at intervals as recommended by the API service.

+

Method Details

+
+ close() +
Close httplib2 connections.
+
+ +
+ get(name, x__xgafv=None) +
Gets the latest state of a long-running operation. Clients can use this method to poll the operation result at intervals as recommended by the API service.
+
+Args:
+  name: string, The name of the operation resource. (required)
+  x__xgafv: string, V1 error format.
+    Allowed values
+      1 - v1 error format
+      2 - v2 error format
+
+Returns:
+  An object of the form:
+
+    { # This resource represents a long-running operation that is the result of a network API call.
+  "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available.
+  "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation.
+    "code": 42, # The status code, which should be an enum value of google.rpc.Code.
+    "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use.
+      {
+        "a_key": "", # Properties of the object. Contains field @type with type URL.
+      },
+    ],
+    "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
+  },
+  "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+  "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`.
+  "response": { # The normal, successful response of the operation. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+}
+
+ + \ No newline at end of file diff --git a/docs/dyn/policysimulator_v1.organizations.locations.html b/docs/dyn/policysimulator_v1.organizations.locations.html index 189b33fc92..12e84896c1 100644 --- a/docs/dyn/policysimulator_v1.organizations.locations.html +++ b/docs/dyn/policysimulator_v1.organizations.locations.html @@ -74,6 +74,11 @@

Policy Simulator API . organizations . locations

Instance Methods

+

+ accessPolicySimulations() +

+

Returns the accessPolicySimulations Resource.

+

orgPolicyViolationsPreviews()

diff --git a/docs/dyn/policysimulator_v1.projects.locations.accessPolicySimulations.html b/docs/dyn/policysimulator_v1.projects.locations.accessPolicySimulations.html new file mode 100644 index 0000000000..7c41a8721d --- /dev/null +++ b/docs/dyn/policysimulator_v1.projects.locations.accessPolicySimulations.html @@ -0,0 +1,91 @@ + + + +

Policy Simulator API . projects . locations . accessPolicySimulations

+

Instance Methods

+

+ operations() +

+

Returns the operations Resource.

+ +

+ close()

+

Close httplib2 connections.

+

Method Details

+
+ close() +
Close httplib2 connections.
+
+ + \ No newline at end of file diff --git a/docs/dyn/policysimulator_v1.projects.locations.accessPolicySimulations.operations.html b/docs/dyn/policysimulator_v1.projects.locations.accessPolicySimulations.operations.html new file mode 100644 index 0000000000..70f4ea8814 --- /dev/null +++ b/docs/dyn/policysimulator_v1.projects.locations.accessPolicySimulations.operations.html @@ -0,0 +1,124 @@ + + + +

Policy Simulator API . projects . locations . accessPolicySimulations . operations

+

Instance Methods

+

+ close()

+

Close httplib2 connections.

+

+ get(name, x__xgafv=None)

+

Gets the latest state of a long-running operation. Clients can use this method to poll the operation result at intervals as recommended by the API service.

+

Method Details

+
+ close() +
Close httplib2 connections.
+
+ +
+ get(name, x__xgafv=None) +
Gets the latest state of a long-running operation. Clients can use this method to poll the operation result at intervals as recommended by the API service.
+
+Args:
+  name: string, The name of the operation resource. (required)
+  x__xgafv: string, V1 error format.
+    Allowed values
+      1 - v1 error format
+      2 - v2 error format
+
+Returns:
+  An object of the form:
+
+    { # This resource represents a long-running operation that is the result of a network API call.
+  "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available.
+  "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation.
+    "code": 42, # The status code, which should be an enum value of google.rpc.Code.
+    "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use.
+      {
+        "a_key": "", # Properties of the object. Contains field @type with type URL.
+      },
+    ],
+    "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
+  },
+  "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+  "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`.
+  "response": { # The normal, successful response of the operation. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+}
+
+ + \ No newline at end of file diff --git a/docs/dyn/policysimulator_v1.projects.locations.html b/docs/dyn/policysimulator_v1.projects.locations.html index 59d0d8a220..e7ec0f34f2 100644 --- a/docs/dyn/policysimulator_v1.projects.locations.html +++ b/docs/dyn/policysimulator_v1.projects.locations.html @@ -74,6 +74,11 @@

Policy Simulator API . projects . locations

Instance Methods

+

+ accessPolicySimulations() +

+

Returns the accessPolicySimulations Resource.

+

orgPolicyViolationsPreviews()

diff --git a/docs/dyn/policysimulator_v1alpha.folders.locations.accessPolicySimulations.html b/docs/dyn/policysimulator_v1alpha.folders.locations.accessPolicySimulations.html new file mode 100644 index 0000000000..365db66eeb --- /dev/null +++ b/docs/dyn/policysimulator_v1alpha.folders.locations.accessPolicySimulations.html @@ -0,0 +1,91 @@ + + + +

Policy Simulator API . folders . locations . accessPolicySimulations

+

Instance Methods

+

+ operations() +

+

Returns the operations Resource.

+ +

+ close()

+

Close httplib2 connections.

+

Method Details

+
+ close() +
Close httplib2 connections.
+
+ + \ No newline at end of file diff --git a/docs/dyn/policysimulator_v1alpha.folders.locations.accessPolicySimulations.operations.html b/docs/dyn/policysimulator_v1alpha.folders.locations.accessPolicySimulations.operations.html new file mode 100644 index 0000000000..71b633b403 --- /dev/null +++ b/docs/dyn/policysimulator_v1alpha.folders.locations.accessPolicySimulations.operations.html @@ -0,0 +1,124 @@ + + + +

Policy Simulator API . folders . locations . accessPolicySimulations . operations

+

Instance Methods

+

+ close()

+

Close httplib2 connections.

+

+ get(name, x__xgafv=None)

+

Gets the latest state of a long-running operation. Clients can use this method to poll the operation result at intervals as recommended by the API service.

+

Method Details

+
+ close() +
Close httplib2 connections.
+
+ +
+ get(name, x__xgafv=None) +
Gets the latest state of a long-running operation. Clients can use this method to poll the operation result at intervals as recommended by the API service.
+
+Args:
+  name: string, The name of the operation resource. (required)
+  x__xgafv: string, V1 error format.
+    Allowed values
+      1 - v1 error format
+      2 - v2 error format
+
+Returns:
+  An object of the form:
+
+    { # This resource represents a long-running operation that is the result of a network API call.
+  "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available.
+  "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation.
+    "code": 42, # The status code, which should be an enum value of google.rpc.Code.
+    "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use.
+      {
+        "a_key": "", # Properties of the object. Contains field @type with type URL.
+      },
+    ],
+    "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
+  },
+  "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+  "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`.
+  "response": { # The normal, successful response of the operation. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+}
+
+ + \ No newline at end of file diff --git a/docs/dyn/policysimulator_v1alpha.folders.locations.html b/docs/dyn/policysimulator_v1alpha.folders.locations.html index 5a5dfffbaa..6294696ba9 100644 --- a/docs/dyn/policysimulator_v1alpha.folders.locations.html +++ b/docs/dyn/policysimulator_v1alpha.folders.locations.html @@ -74,6 +74,11 @@

Policy Simulator API . folders . locations

Instance Methods

+

+ accessPolicySimulations() +

+

Returns the accessPolicySimulations Resource.

+

orgPolicyViolationsPreviews()

diff --git a/docs/dyn/policysimulator_v1alpha.organizations.locations.accessPolicySimulations.html b/docs/dyn/policysimulator_v1alpha.organizations.locations.accessPolicySimulations.html new file mode 100644 index 0000000000..f849f37fa8 --- /dev/null +++ b/docs/dyn/policysimulator_v1alpha.organizations.locations.accessPolicySimulations.html @@ -0,0 +1,91 @@ + + + +

Policy Simulator API . organizations . locations . accessPolicySimulations

+

Instance Methods

+

+ operations() +

+

Returns the operations Resource.

+ +

+ close()

+

Close httplib2 connections.

+

Method Details

+
+ close() +
Close httplib2 connections.
+
+ + \ No newline at end of file diff --git a/docs/dyn/policysimulator_v1alpha.organizations.locations.accessPolicySimulations.operations.html b/docs/dyn/policysimulator_v1alpha.organizations.locations.accessPolicySimulations.operations.html new file mode 100644 index 0000000000..3286dc3494 --- /dev/null +++ b/docs/dyn/policysimulator_v1alpha.organizations.locations.accessPolicySimulations.operations.html @@ -0,0 +1,124 @@ + + + +

Policy Simulator API . organizations . locations . accessPolicySimulations . operations

+

Instance Methods

+

+ close()

+

Close httplib2 connections.

+

+ get(name, x__xgafv=None)

+

Gets the latest state of a long-running operation. Clients can use this method to poll the operation result at intervals as recommended by the API service.

+

Method Details

+
+ close() +
Close httplib2 connections.
+
+ +
+ get(name, x__xgafv=None) +
Gets the latest state of a long-running operation. Clients can use this method to poll the operation result at intervals as recommended by the API service.
+
+Args:
+  name: string, The name of the operation resource. (required)
+  x__xgafv: string, V1 error format.
+    Allowed values
+      1 - v1 error format
+      2 - v2 error format
+
+Returns:
+  An object of the form:
+
+    { # This resource represents a long-running operation that is the result of a network API call.
+  "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available.
+  "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation.
+    "code": 42, # The status code, which should be an enum value of google.rpc.Code.
+    "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use.
+      {
+        "a_key": "", # Properties of the object. Contains field @type with type URL.
+      },
+    ],
+    "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
+  },
+  "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+  "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`.
+  "response": { # The normal, successful response of the operation. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+}
+
+ + \ No newline at end of file diff --git a/docs/dyn/policysimulator_v1alpha.organizations.locations.html b/docs/dyn/policysimulator_v1alpha.organizations.locations.html index 4ae5951dd5..4050183d76 100644 --- a/docs/dyn/policysimulator_v1alpha.organizations.locations.html +++ b/docs/dyn/policysimulator_v1alpha.organizations.locations.html @@ -74,6 +74,11 @@

Policy Simulator API . organizations . locations

Instance Methods

+

+ accessPolicySimulations() +

+

Returns the accessPolicySimulations Resource.

+

orgPolicyViolationsPreviews()

diff --git a/docs/dyn/policysimulator_v1alpha.projects.locations.accessPolicySimulations.html b/docs/dyn/policysimulator_v1alpha.projects.locations.accessPolicySimulations.html new file mode 100644 index 0000000000..8c99b8171b --- /dev/null +++ b/docs/dyn/policysimulator_v1alpha.projects.locations.accessPolicySimulations.html @@ -0,0 +1,91 @@ + + + +

Policy Simulator API . projects . locations . accessPolicySimulations

+

Instance Methods

+

+ operations() +

+

Returns the operations Resource.

+ +

+ close()

+

Close httplib2 connections.

+

Method Details

+
+ close() +
Close httplib2 connections.
+
+ + \ No newline at end of file diff --git a/docs/dyn/policysimulator_v1alpha.projects.locations.accessPolicySimulations.operations.html b/docs/dyn/policysimulator_v1alpha.projects.locations.accessPolicySimulations.operations.html new file mode 100644 index 0000000000..bd1d3350f2 --- /dev/null +++ b/docs/dyn/policysimulator_v1alpha.projects.locations.accessPolicySimulations.operations.html @@ -0,0 +1,124 @@ + + + +

Policy Simulator API . projects . locations . accessPolicySimulations . operations

+

Instance Methods

+

+ close()

+

Close httplib2 connections.

+

+ get(name, x__xgafv=None)

+

Gets the latest state of a long-running operation. Clients can use this method to poll the operation result at intervals as recommended by the API service.

+

Method Details

+
+ close() +
Close httplib2 connections.
+
+ +
+ get(name, x__xgafv=None) +
Gets the latest state of a long-running operation. Clients can use this method to poll the operation result at intervals as recommended by the API service.
+
+Args:
+  name: string, The name of the operation resource. (required)
+  x__xgafv: string, V1 error format.
+    Allowed values
+      1 - v1 error format
+      2 - v2 error format
+
+Returns:
+  An object of the form:
+
+    { # This resource represents a long-running operation that is the result of a network API call.
+  "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available.
+  "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation.
+    "code": 42, # The status code, which should be an enum value of google.rpc.Code.
+    "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use.
+      {
+        "a_key": "", # Properties of the object. Contains field @type with type URL.
+      },
+    ],
+    "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
+  },
+  "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+  "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`.
+  "response": { # The normal, successful response of the operation. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+}
+
+ + \ No newline at end of file diff --git a/docs/dyn/policysimulator_v1alpha.projects.locations.html b/docs/dyn/policysimulator_v1alpha.projects.locations.html index 172742833c..5881735dfd 100644 --- a/docs/dyn/policysimulator_v1alpha.projects.locations.html +++ b/docs/dyn/policysimulator_v1alpha.projects.locations.html @@ -74,6 +74,11 @@

Policy Simulator API . projects . locations

Instance Methods

+

+ accessPolicySimulations() +

+

Returns the accessPolicySimulations Resource.

+

orgPolicyViolationsPreviews()

diff --git a/docs/dyn/policysimulator_v1beta.folders.locations.accessPolicySimulations.html b/docs/dyn/policysimulator_v1beta.folders.locations.accessPolicySimulations.html new file mode 100644 index 0000000000..97878f21f1 --- /dev/null +++ b/docs/dyn/policysimulator_v1beta.folders.locations.accessPolicySimulations.html @@ -0,0 +1,91 @@ + + + +

Policy Simulator API . folders . locations . accessPolicySimulations

+

Instance Methods

+

+ operations() +

+

Returns the operations Resource.

+ +

+ close()

+

Close httplib2 connections.

+

Method Details

+
+ close() +
Close httplib2 connections.
+
+ + \ No newline at end of file diff --git a/docs/dyn/policysimulator_v1beta.folders.locations.accessPolicySimulations.operations.html b/docs/dyn/policysimulator_v1beta.folders.locations.accessPolicySimulations.operations.html new file mode 100644 index 0000000000..3952e106d6 --- /dev/null +++ b/docs/dyn/policysimulator_v1beta.folders.locations.accessPolicySimulations.operations.html @@ -0,0 +1,124 @@ + + + +

Policy Simulator API . folders . locations . accessPolicySimulations . operations

+

Instance Methods

+

+ close()

+

Close httplib2 connections.

+

+ get(name, x__xgafv=None)

+

Gets the latest state of a long-running operation. Clients can use this method to poll the operation result at intervals as recommended by the API service.

+

Method Details

+
+ close() +
Close httplib2 connections.
+
+ +
+ get(name, x__xgafv=None) +
Gets the latest state of a long-running operation. Clients can use this method to poll the operation result at intervals as recommended by the API service.
+
+Args:
+  name: string, The name of the operation resource. (required)
+  x__xgafv: string, V1 error format.
+    Allowed values
+      1 - v1 error format
+      2 - v2 error format
+
+Returns:
+  An object of the form:
+
+    { # This resource represents a long-running operation that is the result of a network API call.
+  "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available.
+  "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation.
+    "code": 42, # The status code, which should be an enum value of google.rpc.Code.
+    "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use.
+      {
+        "a_key": "", # Properties of the object. Contains field @type with type URL.
+      },
+    ],
+    "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
+  },
+  "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+  "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`.
+  "response": { # The normal, successful response of the operation. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+}
+
+ + \ No newline at end of file diff --git a/docs/dyn/policysimulator_v1beta.folders.locations.html b/docs/dyn/policysimulator_v1beta.folders.locations.html index d6c907bae1..1c097efdbf 100644 --- a/docs/dyn/policysimulator_v1beta.folders.locations.html +++ b/docs/dyn/policysimulator_v1beta.folders.locations.html @@ -74,6 +74,11 @@

Policy Simulator API . folders . locations

Instance Methods

+

+ accessPolicySimulations() +

+

Returns the accessPolicySimulations Resource.

+

orgPolicyViolationsPreviews()

diff --git a/docs/dyn/policysimulator_v1beta.organizations.locations.accessPolicySimulations.html b/docs/dyn/policysimulator_v1beta.organizations.locations.accessPolicySimulations.html new file mode 100644 index 0000000000..b43c5a972f --- /dev/null +++ b/docs/dyn/policysimulator_v1beta.organizations.locations.accessPolicySimulations.html @@ -0,0 +1,91 @@ + + + +

Policy Simulator API . organizations . locations . accessPolicySimulations

+

Instance Methods

+

+ operations() +

+

Returns the operations Resource.

+ +

+ close()

+

Close httplib2 connections.

+

Method Details

+
+ close() +
Close httplib2 connections.
+
+ + \ No newline at end of file diff --git a/docs/dyn/policysimulator_v1beta.organizations.locations.accessPolicySimulations.operations.html b/docs/dyn/policysimulator_v1beta.organizations.locations.accessPolicySimulations.operations.html new file mode 100644 index 0000000000..dbd793ca7d --- /dev/null +++ b/docs/dyn/policysimulator_v1beta.organizations.locations.accessPolicySimulations.operations.html @@ -0,0 +1,124 @@ + + + +

Policy Simulator API . organizations . locations . accessPolicySimulations . operations

+

Instance Methods

+

+ close()

+

Close httplib2 connections.

+

+ get(name, x__xgafv=None)

+

Gets the latest state of a long-running operation. Clients can use this method to poll the operation result at intervals as recommended by the API service.

+

Method Details

+
+ close() +
Close httplib2 connections.
+
+ +
+ get(name, x__xgafv=None) +
Gets the latest state of a long-running operation. Clients can use this method to poll the operation result at intervals as recommended by the API service.
+
+Args:
+  name: string, The name of the operation resource. (required)
+  x__xgafv: string, V1 error format.
+    Allowed values
+      1 - v1 error format
+      2 - v2 error format
+
+Returns:
+  An object of the form:
+
+    { # This resource represents a long-running operation that is the result of a network API call.
+  "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available.
+  "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation.
+    "code": 42, # The status code, which should be an enum value of google.rpc.Code.
+    "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use.
+      {
+        "a_key": "", # Properties of the object. Contains field @type with type URL.
+      },
+    ],
+    "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
+  },
+  "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+  "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`.
+  "response": { # The normal, successful response of the operation. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+}
+
+ + \ No newline at end of file diff --git a/docs/dyn/policysimulator_v1beta.organizations.locations.html b/docs/dyn/policysimulator_v1beta.organizations.locations.html index 5d211a0a50..f9b9d6e46c 100644 --- a/docs/dyn/policysimulator_v1beta.organizations.locations.html +++ b/docs/dyn/policysimulator_v1beta.organizations.locations.html @@ -74,6 +74,11 @@

Policy Simulator API . organizations . locations

Instance Methods

+

+ accessPolicySimulations() +

+

Returns the accessPolicySimulations Resource.

+

orgPolicyViolationsPreviews()

diff --git a/docs/dyn/policysimulator_v1beta.projects.locations.accessPolicySimulations.html b/docs/dyn/policysimulator_v1beta.projects.locations.accessPolicySimulations.html new file mode 100644 index 0000000000..140c8f5b1b --- /dev/null +++ b/docs/dyn/policysimulator_v1beta.projects.locations.accessPolicySimulations.html @@ -0,0 +1,91 @@ + + + +

Policy Simulator API . projects . locations . accessPolicySimulations

+

Instance Methods

+

+ operations() +

+

Returns the operations Resource.

+ +

+ close()

+

Close httplib2 connections.

+

Method Details

+
+ close() +
Close httplib2 connections.
+
+ + \ No newline at end of file diff --git a/docs/dyn/policysimulator_v1beta.projects.locations.accessPolicySimulations.operations.html b/docs/dyn/policysimulator_v1beta.projects.locations.accessPolicySimulations.operations.html new file mode 100644 index 0000000000..7ece0b9ffe --- /dev/null +++ b/docs/dyn/policysimulator_v1beta.projects.locations.accessPolicySimulations.operations.html @@ -0,0 +1,124 @@ + + + +

Policy Simulator API . projects . locations . accessPolicySimulations . operations

+

Instance Methods

+

+ close()

+

Close httplib2 connections.

+

+ get(name, x__xgafv=None)

+

Gets the latest state of a long-running operation. Clients can use this method to poll the operation result at intervals as recommended by the API service.

+

Method Details

+
+ close() +
Close httplib2 connections.
+
+ +
+ get(name, x__xgafv=None) +
Gets the latest state of a long-running operation. Clients can use this method to poll the operation result at intervals as recommended by the API service.
+
+Args:
+  name: string, The name of the operation resource. (required)
+  x__xgafv: string, V1 error format.
+    Allowed values
+      1 - v1 error format
+      2 - v2 error format
+
+Returns:
+  An object of the form:
+
+    { # This resource represents a long-running operation that is the result of a network API call.
+  "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available.
+  "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation.
+    "code": 42, # The status code, which should be an enum value of google.rpc.Code.
+    "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use.
+      {
+        "a_key": "", # Properties of the object. Contains field @type with type URL.
+      },
+    ],
+    "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
+  },
+  "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+  "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`.
+  "response": { # The normal, successful response of the operation. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+}
+
+ + \ No newline at end of file diff --git a/docs/dyn/policysimulator_v1beta.projects.locations.html b/docs/dyn/policysimulator_v1beta.projects.locations.html index 73593ef947..ce67e0ff4c 100644 --- a/docs/dyn/policysimulator_v1beta.projects.locations.html +++ b/docs/dyn/policysimulator_v1beta.projects.locations.html @@ -74,6 +74,11 @@

Policy Simulator API . projects . locations

Instance Methods

+

+ accessPolicySimulations() +

+

Returns the accessPolicySimulations Resource.

+

orgPolicyViolationsPreviews()

diff --git a/docs/dyn/redis_v1.projects.locations.clusters.html b/docs/dyn/redis_v1.projects.locations.clusters.html index 94ebe63f6b..ff73fb4995 100644 --- a/docs/dyn/redis_v1.projects.locations.clusters.html +++ b/docs/dyn/redis_v1.projects.locations.clusters.html @@ -163,6 +163,7 @@

Method Details

The object takes the form of: { # A cluster instance. + "allowFewerZonesDeployment": True or False, # Optional. Immutable. Allows customers to specify if they are okay with deploying a multi-zone cluster in less than 3 zones. Once set, if there is a zonal outage during the cluster creation, the cluster will only be deployed in 2 zones, and stay within the 2 zones for its lifecycle. "asyncClusterEndpointsDeletionEnabled": True or False, # Optional. If true, cluster endpoints that are created and registered by customers can be deleted asynchronously. That is, such a cluster endpoint can be de-registered before the forwarding rules in the cluster endpoint are deleted. "authorizationMode": "A String", # Optional. The authorization mode of the Redis cluster. If not provided, auth feature is disabled for the cluster. "automatedBackupConfig": { # The automated backup config for a cluster. # Optional. The automated backup config for the cluster. @@ -423,6 +424,7 @@

Method Details

An object of the form: { # A cluster instance. + "allowFewerZonesDeployment": True or False, # Optional. Immutable. Allows customers to specify if they are okay with deploying a multi-zone cluster in less than 3 zones. Once set, if there is a zonal outage during the cluster creation, the cluster will only be deployed in 2 zones, and stay within the 2 zones for its lifecycle. "asyncClusterEndpointsDeletionEnabled": True or False, # Optional. If true, cluster endpoints that are created and registered by customers can be deleted asynchronously. That is, such a cluster endpoint can be de-registered before the forwarding rules in the cluster endpoint are deleted. "authorizationMode": "A String", # Optional. The authorization mode of the Redis cluster. If not provided, auth feature is disabled for the cluster. "automatedBackupConfig": { # The automated backup config for a cluster. # Optional. The automated backup config for the cluster. @@ -649,6 +651,7 @@

Method Details

{ # Response for ListClusters. "clusters": [ # A list of Redis clusters in the project in the specified location, or across all locations. If the `location_id` in the parent field of the request is "-", all regions available to the project are queried, and the results aggregated. If in such an aggregated query a location is unavailable, a placeholder Redis entry is included in the response with the `name` field set to a value of the form `projects/{project_id}/locations/{location_id}/clusters/`- and the `status` field set to ERROR and `status_message` field set to "location not available for ListClusters". { # A cluster instance. + "allowFewerZonesDeployment": True or False, # Optional. Immutable. Allows customers to specify if they are okay with deploying a multi-zone cluster in less than 3 zones. Once set, if there is a zonal outage during the cluster creation, the cluster will only be deployed in 2 zones, and stay within the 2 zones for its lifecycle. "asyncClusterEndpointsDeletionEnabled": True or False, # Optional. If true, cluster endpoints that are created and registered by customers can be deleted asynchronously. That is, such a cluster endpoint can be de-registered before the forwarding rules in the cluster endpoint are deleted. "authorizationMode": "A String", # Optional. The authorization mode of the Redis cluster. If not provided, auth feature is disabled for the cluster. "automatedBackupConfig": { # The automated backup config for a cluster. # Optional. The automated backup config for the cluster. @@ -858,6 +861,7 @@

Method Details

The object takes the form of: { # A cluster instance. + "allowFewerZonesDeployment": True or False, # Optional. Immutable. Allows customers to specify if they are okay with deploying a multi-zone cluster in less than 3 zones. Once set, if there is a zonal outage during the cluster creation, the cluster will only be deployed in 2 zones, and stay within the 2 zones for its lifecycle. "asyncClusterEndpointsDeletionEnabled": True or False, # Optional. If true, cluster endpoints that are created and registered by customers can be deleted asynchronously. That is, such a cluster endpoint can be de-registered before the forwarding rules in the cluster endpoint are deleted. "authorizationMode": "A String", # Optional. The authorization mode of the Redis cluster. If not provided, auth feature is disabled for the cluster. "automatedBackupConfig": { # The automated backup config for a cluster. # Optional. The automated backup config for the cluster. diff --git a/docs/dyn/redis_v1beta1.projects.locations.clusters.html b/docs/dyn/redis_v1beta1.projects.locations.clusters.html index 87e547fcbc..1d0ee0f55e 100644 --- a/docs/dyn/redis_v1beta1.projects.locations.clusters.html +++ b/docs/dyn/redis_v1beta1.projects.locations.clusters.html @@ -163,6 +163,7 @@

Method Details

The object takes the form of: { # A cluster instance. + "allowFewerZonesDeployment": True or False, # Optional. Immutable. Allows customers to specify if they are okay with deploying a multi-zone cluster in less than 3 zones. Once set, if there is a zonal outage during the cluster creation, the cluster will only be deployed in 2 zones, and stay within the 2 zones for its lifecycle. "asyncClusterEndpointsDeletionEnabled": True or False, # Optional. If true, cluster endpoints that are created and registered by customers can be deleted asynchronously. That is, such a cluster endpoint can be de-registered before the forwarding rules in the cluster endpoint are deleted. "authorizationMode": "A String", # Optional. The authorization mode of the Redis cluster. If not provided, auth feature is disabled for the cluster. "automatedBackupConfig": { # The automated backup config for a cluster. # Optional. The automated backup config for the cluster. @@ -423,6 +424,7 @@

Method Details

An object of the form: { # A cluster instance. + "allowFewerZonesDeployment": True or False, # Optional. Immutable. Allows customers to specify if they are okay with deploying a multi-zone cluster in less than 3 zones. Once set, if there is a zonal outage during the cluster creation, the cluster will only be deployed in 2 zones, and stay within the 2 zones for its lifecycle. "asyncClusterEndpointsDeletionEnabled": True or False, # Optional. If true, cluster endpoints that are created and registered by customers can be deleted asynchronously. That is, such a cluster endpoint can be de-registered before the forwarding rules in the cluster endpoint are deleted. "authorizationMode": "A String", # Optional. The authorization mode of the Redis cluster. If not provided, auth feature is disabled for the cluster. "automatedBackupConfig": { # The automated backup config for a cluster. # Optional. The automated backup config for the cluster. @@ -649,6 +651,7 @@

Method Details

{ # Response for ListClusters. "clusters": [ # A list of Redis clusters in the project in the specified location, or across all locations. If the `location_id` in the parent field of the request is "-", all regions available to the project are queried, and the results aggregated. If in such an aggregated query a location is unavailable, a placeholder Redis entry is included in the response with the `name` field set to a value of the form `projects/{project_id}/locations/{location_id}/clusters/`- and the `status` field set to ERROR and `status_message` field set to "location not available for ListClusters". { # A cluster instance. + "allowFewerZonesDeployment": True or False, # Optional. Immutable. Allows customers to specify if they are okay with deploying a multi-zone cluster in less than 3 zones. Once set, if there is a zonal outage during the cluster creation, the cluster will only be deployed in 2 zones, and stay within the 2 zones for its lifecycle. "asyncClusterEndpointsDeletionEnabled": True or False, # Optional. If true, cluster endpoints that are created and registered by customers can be deleted asynchronously. That is, such a cluster endpoint can be de-registered before the forwarding rules in the cluster endpoint are deleted. "authorizationMode": "A String", # Optional. The authorization mode of the Redis cluster. If not provided, auth feature is disabled for the cluster. "automatedBackupConfig": { # The automated backup config for a cluster. # Optional. The automated backup config for the cluster. @@ -858,6 +861,7 @@

Method Details

The object takes the form of: { # A cluster instance. + "allowFewerZonesDeployment": True or False, # Optional. Immutable. Allows customers to specify if they are okay with deploying a multi-zone cluster in less than 3 zones. Once set, if there is a zonal outage during the cluster creation, the cluster will only be deployed in 2 zones, and stay within the 2 zones for its lifecycle. "asyncClusterEndpointsDeletionEnabled": True or False, # Optional. If true, cluster endpoints that are created and registered by customers can be deleted asynchronously. That is, such a cluster endpoint can be de-registered before the forwarding rules in the cluster endpoint are deleted. "authorizationMode": "A String", # Optional. The authorization mode of the Redis cluster. If not provided, auth feature is disabled for the cluster. "automatedBackupConfig": { # The automated backup config for a cluster. # Optional. The automated backup config for the cluster. diff --git a/docs/dyn/retail_v2.projects.locations.catalogs.html b/docs/dyn/retail_v2.projects.locations.catalogs.html index 4224d256ba..28c78fe451 100644 --- a/docs/dyn/retail_v2.projects.locations.catalogs.html +++ b/docs/dyn/retail_v2.projects.locations.catalogs.html @@ -455,11 +455,21 @@

Method Details

], "example": [ # Optional. A list of examples for intent classification. { # An example for intent classification. + "classifiedPositive": True or False, # Required. Whether the example is classified positively. "intentType": "A String", # Optional. The intent_type must match one of the predefined intent types defined at https://cloud.google.com/retail/docs/reference/rpc/google.cloud.retail.v2alpha#querytype "query": "A String", # Required. Example query. "reason": "A String", # Optional. The reason for the intent classification. This is used to explain the intent classification decision. }, ], + "inlineSource": { # Inline source for intent classifications. # Optional. Inline source for intent classifications. + "inlineForceIntents": [ # Optional. A list of inline force intent classifications. + { # An inline force intent classification configuration. + "intentType": "A String", # Optional. The intent_type must match one of the predefined intent types defined at https://cloud.google.com/retail/docs/reference/rpc/google.cloud.retail.v2alpha#querytype + "operation": "A String", # Optional. The operation to perform for the query. + "query": "A String", # Optional. A example query. + }, + ], + }, "modelPreamble": "A String", # Optional. Customers can use the preamble to specify any requirements for blocklisting intent classification. This preamble will be added to the blocklisting intent classification model prompt. }, "retailerDisplayName": "A String", # Optional. The retailer's display name that could be used in our LLM answers. Example - "Google" @@ -887,11 +897,21 @@

Method Details

], "example": [ # Optional. A list of examples for intent classification. { # An example for intent classification. + "classifiedPositive": True or False, # Required. Whether the example is classified positively. "intentType": "A String", # Optional. The intent_type must match one of the predefined intent types defined at https://cloud.google.com/retail/docs/reference/rpc/google.cloud.retail.v2alpha#querytype "query": "A String", # Required. Example query. "reason": "A String", # Optional. The reason for the intent classification. This is used to explain the intent classification decision. }, ], + "inlineSource": { # Inline source for intent classifications. # Optional. Inline source for intent classifications. + "inlineForceIntents": [ # Optional. A list of inline force intent classifications. + { # An inline force intent classification configuration. + "intentType": "A String", # Optional. The intent_type must match one of the predefined intent types defined at https://cloud.google.com/retail/docs/reference/rpc/google.cloud.retail.v2alpha#querytype + "operation": "A String", # Optional. The operation to perform for the query. + "query": "A String", # Optional. A example query. + }, + ], + }, "modelPreamble": "A String", # Optional. Customers can use the preamble to specify any requirements for blocklisting intent classification. This preamble will be added to the blocklisting intent classification model prompt. }, "retailerDisplayName": "A String", # Optional. The retailer's display name that could be used in our LLM answers. Example - "Google" @@ -917,11 +937,21 @@

Method Details

], "example": [ # Optional. A list of examples for intent classification. { # An example for intent classification. + "classifiedPositive": True or False, # Required. Whether the example is classified positively. "intentType": "A String", # Optional. The intent_type must match one of the predefined intent types defined at https://cloud.google.com/retail/docs/reference/rpc/google.cloud.retail.v2alpha#querytype "query": "A String", # Required. Example query. "reason": "A String", # Optional. The reason for the intent classification. This is used to explain the intent classification decision. }, ], + "inlineSource": { # Inline source for intent classifications. # Optional. Inline source for intent classifications. + "inlineForceIntents": [ # Optional. A list of inline force intent classifications. + { # An inline force intent classification configuration. + "intentType": "A String", # Optional. The intent_type must match one of the predefined intent types defined at https://cloud.google.com/retail/docs/reference/rpc/google.cloud.retail.v2alpha#querytype + "operation": "A String", # Optional. The operation to perform for the query. + "query": "A String", # Optional. A example query. + }, + ], + }, "modelPreamble": "A String", # Optional. Customers can use the preamble to specify any requirements for blocklisting intent classification. This preamble will be added to the blocklisting intent classification model prompt. }, "retailerDisplayName": "A String", # Optional. The retailer's display name that could be used in our LLM answers. Example - "Google" diff --git a/docs/dyn/retail_v2.projects.locations.catalogs.placements.html b/docs/dyn/retail_v2.projects.locations.catalogs.placements.html index 4e5e4fb2a2..107ede9567 100644 --- a/docs/dyn/retail_v2.projects.locations.catalogs.placements.html +++ b/docs/dyn/retail_v2.projects.locations.catalogs.placements.html @@ -484,6 +484,13 @@

Method Details

], "tileNavigationRequested": True or False, # This field specifies whether the customer would like to request tile navigation. }, + "userAttributes": { # Optional. The user attributes that could be used for personalization of search results. * Populate at most 100 key-value pairs per query. * Only supports string keys and repeated string values. * Duplcate keys are not allowed within a single query. Example: user_attributes: [ { key: "pets" value { values: "dog" values: "cat" } }, { key: "state" value { values: "CA" } } ] + "a_key": { # A list of string values. + "values": [ # String values. + "A String", + ], + }, + }, "userInfo": { # Information of an end user. # User information. "directUserRequest": True or False, # True if the request is made directly from the end user, in which case the ip_address and user_agent can be populated from the HTTP request. This flag should be set only if the API request is made directly from the end user such as a mobile app (and not if a gateway or a server is processing and pushing the user events). This should not be set when using the JavaScript tag in UserEventService.CollectUserEvent. "ipAddress": "A String", # The end user's IP address. This field is used to extract location information for personalization. This field must be either an IPv4 address (e.g. "104.133.9.80") or an IPv6 address (e.g. "2001:0db8:85a3:0000:0000:8a2e:0370:7334"). Otherwise, an INVALID_ARGUMENT error is returned. This should not be set when: * setting SearchRequest.user_info. * using the JavaScript tag in UserEventService.CollectUserEvent or if direct_user_request is set. diff --git a/docs/dyn/retail_v2.projects.locations.catalogs.servingConfigs.html b/docs/dyn/retail_v2.projects.locations.catalogs.servingConfigs.html index b67794ef69..141d9e00f8 100644 --- a/docs/dyn/retail_v2.projects.locations.catalogs.servingConfigs.html +++ b/docs/dyn/retail_v2.projects.locations.catalogs.servingConfigs.html @@ -1035,6 +1035,13 @@

Method Details

], "tileNavigationRequested": True or False, # This field specifies whether the customer would like to request tile navigation. }, + "userAttributes": { # Optional. The user attributes that could be used for personalization of search results. * Populate at most 100 key-value pairs per query. * Only supports string keys and repeated string values. * Duplcate keys are not allowed within a single query. Example: user_attributes: [ { key: "pets" value { values: "dog" values: "cat" } }, { key: "state" value { values: "CA" } } ] + "a_key": { # A list of string values. + "values": [ # String values. + "A String", + ], + }, + }, "userInfo": { # Information of an end user. # User information. "directUserRequest": True or False, # True if the request is made directly from the end user, in which case the ip_address and user_agent can be populated from the HTTP request. This flag should be set only if the API request is made directly from the end user such as a mobile app (and not if a gateway or a server is processing and pushing the user events). This should not be set when using the JavaScript tag in UserEventService.CollectUserEvent. "ipAddress": "A String", # The end user's IP address. This field is used to extract location information for personalization. This field must be either an IPv4 address (e.g. "104.133.9.80") or an IPv6 address (e.g. "2001:0db8:85a3:0000:0000:8a2e:0370:7334"). Otherwise, an INVALID_ARGUMENT error is returned. This should not be set when: * setting SearchRequest.user_info. * using the JavaScript tag in UserEventService.CollectUserEvent or if direct_user_request is set. diff --git a/docs/dyn/retail_v2alpha.projects.locations.catalogs.html b/docs/dyn/retail_v2alpha.projects.locations.catalogs.html index 82055a4e15..3a1d57f92b 100644 --- a/docs/dyn/retail_v2alpha.projects.locations.catalogs.html +++ b/docs/dyn/retail_v2alpha.projects.locations.catalogs.html @@ -482,11 +482,21 @@

Method Details

], "example": [ # Optional. A list of examples for intent classification. { # An example for intent classification. + "classifiedPositive": True or False, # Required. Whether the example is classified positively. "intentType": "A String", # Optional. The intent_type must match one of the predefined intent types defined at https://cloud.google.com/retail/docs/reference/rpc/google.cloud.retail.v2alpha#querytype "query": "A String", # Required. Example query. "reason": "A String", # Optional. The reason for the intent classification. This is used to explain the intent classification decision. }, ], + "inlineSource": { # Inline source for intent classifications. # Optional. Inline source for intent classifications. + "inlineForceIntents": [ # Optional. A list of inline force intent classifications. + { # An inline force intent classification configuration. + "intentType": "A String", # Optional. The intent_type must match one of the predefined intent types defined at https://cloud.google.com/retail/docs/reference/rpc/google.cloud.retail.v2alpha#querytype + "operation": "A String", # Optional. The operation to perform for the query. + "query": "A String", # Optional. A example query. + }, + ], + }, "modelPreamble": "A String", # Optional. Customers can use the preamble to specify any requirements for blocklisting intent classification. This preamble will be added to the blocklisting intent classification model prompt. }, "retailerDisplayName": "A String", # Optional. The retailer's display name that could be used in our LLM answers. Example - "Google" @@ -976,11 +986,21 @@

Method Details

], "example": [ # Optional. A list of examples for intent classification. { # An example for intent classification. + "classifiedPositive": True or False, # Required. Whether the example is classified positively. "intentType": "A String", # Optional. The intent_type must match one of the predefined intent types defined at https://cloud.google.com/retail/docs/reference/rpc/google.cloud.retail.v2alpha#querytype "query": "A String", # Required. Example query. "reason": "A String", # Optional. The reason for the intent classification. This is used to explain the intent classification decision. }, ], + "inlineSource": { # Inline source for intent classifications. # Optional. Inline source for intent classifications. + "inlineForceIntents": [ # Optional. A list of inline force intent classifications. + { # An inline force intent classification configuration. + "intentType": "A String", # Optional. The intent_type must match one of the predefined intent types defined at https://cloud.google.com/retail/docs/reference/rpc/google.cloud.retail.v2alpha#querytype + "operation": "A String", # Optional. The operation to perform for the query. + "query": "A String", # Optional. A example query. + }, + ], + }, "modelPreamble": "A String", # Optional. Customers can use the preamble to specify any requirements for blocklisting intent classification. This preamble will be added to the blocklisting intent classification model prompt. }, "retailerDisplayName": "A String", # Optional. The retailer's display name that could be used in our LLM answers. Example - "Google" @@ -1006,11 +1026,21 @@

Method Details

], "example": [ # Optional. A list of examples for intent classification. { # An example for intent classification. + "classifiedPositive": True or False, # Required. Whether the example is classified positively. "intentType": "A String", # Optional. The intent_type must match one of the predefined intent types defined at https://cloud.google.com/retail/docs/reference/rpc/google.cloud.retail.v2alpha#querytype "query": "A String", # Required. Example query. "reason": "A String", # Optional. The reason for the intent classification. This is used to explain the intent classification decision. }, ], + "inlineSource": { # Inline source for intent classifications. # Optional. Inline source for intent classifications. + "inlineForceIntents": [ # Optional. A list of inline force intent classifications. + { # An inline force intent classification configuration. + "intentType": "A String", # Optional. The intent_type must match one of the predefined intent types defined at https://cloud.google.com/retail/docs/reference/rpc/google.cloud.retail.v2alpha#querytype + "operation": "A String", # Optional. The operation to perform for the query. + "query": "A String", # Optional. A example query. + }, + ], + }, "modelPreamble": "A String", # Optional. Customers can use the preamble to specify any requirements for blocklisting intent classification. This preamble will be added to the blocklisting intent classification model prompt. }, "retailerDisplayName": "A String", # Optional. The retailer's display name that could be used in our LLM answers. Example - "Google" diff --git a/docs/dyn/retail_v2beta.projects.locations.catalogs.html b/docs/dyn/retail_v2beta.projects.locations.catalogs.html index f87e65555a..a808246b87 100644 --- a/docs/dyn/retail_v2beta.projects.locations.catalogs.html +++ b/docs/dyn/retail_v2beta.projects.locations.catalogs.html @@ -456,11 +456,21 @@

Method Details

], "example": [ # Optional. A list of examples for intent classification. { # An example for intent classification. + "classifiedPositive": True or False, # Required. Whether the example is classified positively. "intentType": "A String", # Optional. The intent_type must match one of the predefined intent types defined at https://cloud.google.com/retail/docs/reference/rpc/google.cloud.retail.v2alpha#querytype "query": "A String", # Required. Example query. "reason": "A String", # Optional. The reason for the intent classification. This is used to explain the intent classification decision. }, ], + "inlineSource": { # Inline source for intent classifications. # Optional. Inline source for intent classifications. + "inlineForceIntents": [ # Optional. A list of inline force intent classifications. + { # An inline force intent classification configuration. + "intentType": "A String", # Optional. The intent_type must match one of the predefined intent types defined at https://cloud.google.com/retail/docs/reference/rpc/google.cloud.retail.v2alpha#querytype + "operation": "A String", # Optional. The operation to perform for the query. + "query": "A String", # Optional. A example query. + }, + ], + }, "modelPreamble": "A String", # Optional. Customers can use the preamble to specify any requirements for blocklisting intent classification. This preamble will be added to the blocklisting intent classification model prompt. }, "retailerDisplayName": "A String", # Optional. The retailer's display name that could be used in our LLM answers. Example - "Google" @@ -950,11 +960,21 @@

Method Details

], "example": [ # Optional. A list of examples for intent classification. { # An example for intent classification. + "classifiedPositive": True or False, # Required. Whether the example is classified positively. "intentType": "A String", # Optional. The intent_type must match one of the predefined intent types defined at https://cloud.google.com/retail/docs/reference/rpc/google.cloud.retail.v2alpha#querytype "query": "A String", # Required. Example query. "reason": "A String", # Optional. The reason for the intent classification. This is used to explain the intent classification decision. }, ], + "inlineSource": { # Inline source for intent classifications. # Optional. Inline source for intent classifications. + "inlineForceIntents": [ # Optional. A list of inline force intent classifications. + { # An inline force intent classification configuration. + "intentType": "A String", # Optional. The intent_type must match one of the predefined intent types defined at https://cloud.google.com/retail/docs/reference/rpc/google.cloud.retail.v2alpha#querytype + "operation": "A String", # Optional. The operation to perform for the query. + "query": "A String", # Optional. A example query. + }, + ], + }, "modelPreamble": "A String", # Optional. Customers can use the preamble to specify any requirements for blocklisting intent classification. This preamble will be added to the blocklisting intent classification model prompt. }, "retailerDisplayName": "A String", # Optional. The retailer's display name that could be used in our LLM answers. Example - "Google" @@ -980,11 +1000,21 @@

Method Details

], "example": [ # Optional. A list of examples for intent classification. { # An example for intent classification. + "classifiedPositive": True or False, # Required. Whether the example is classified positively. "intentType": "A String", # Optional. The intent_type must match one of the predefined intent types defined at https://cloud.google.com/retail/docs/reference/rpc/google.cloud.retail.v2alpha#querytype "query": "A String", # Required. Example query. "reason": "A String", # Optional. The reason for the intent classification. This is used to explain the intent classification decision. }, ], + "inlineSource": { # Inline source for intent classifications. # Optional. Inline source for intent classifications. + "inlineForceIntents": [ # Optional. A list of inline force intent classifications. + { # An inline force intent classification configuration. + "intentType": "A String", # Optional. The intent_type must match one of the predefined intent types defined at https://cloud.google.com/retail/docs/reference/rpc/google.cloud.retail.v2alpha#querytype + "operation": "A String", # Optional. The operation to perform for the query. + "query": "A String", # Optional. A example query. + }, + ], + }, "modelPreamble": "A String", # Optional. Customers can use the preamble to specify any requirements for blocklisting intent classification. This preamble will be added to the blocklisting intent classification model prompt. }, "retailerDisplayName": "A String", # Optional. The retailer's display name that could be used in our LLM answers. Example - "Google" diff --git a/docs/dyn/retail_v2beta.projects.locations.catalogs.placements.html b/docs/dyn/retail_v2beta.projects.locations.catalogs.placements.html index ab40542dd2..808a9d9d6d 100644 --- a/docs/dyn/retail_v2beta.projects.locations.catalogs.placements.html +++ b/docs/dyn/retail_v2beta.projects.locations.catalogs.placements.html @@ -484,6 +484,13 @@

Method Details

], "tileNavigationRequested": True or False, # This field specifies whether the customer would like to request tile navigation. }, + "userAttributes": { # Optional. The user attributes that could be used for personalization of search results. * Populate at most 100 key-value pairs per query. * Only supports string keys and repeated string values. * Duplcate keys are not allowed within a single query. Example: user_attributes: [ { key: "pets" value { values: "dog" values: "cat" } }, { key: "state" value { values: "CA" } } ] + "a_key": { # A list of string values. + "values": [ # String values. + "A String", + ], + }, + }, "userInfo": { # Information of an end user. # User information. "directUserRequest": True or False, # True if the request is made directly from the end user, in which case the ip_address and user_agent can be populated from the HTTP request. This flag should be set only if the API request is made directly from the end user such as a mobile app (and not if a gateway or a server is processing and pushing the user events). This should not be set when using the JavaScript tag in UserEventService.CollectUserEvent. "ipAddress": "A String", # The end user's IP address. This field is used to extract location information for personalization. This field must be either an IPv4 address (e.g. "104.133.9.80") or an IPv6 address (e.g. "2001:0db8:85a3:0000:0000:8a2e:0370:7334"). Otherwise, an INVALID_ARGUMENT error is returned. This should not be set when: * setting SearchRequest.user_info. * using the JavaScript tag in UserEventService.CollectUserEvent or if direct_user_request is set. diff --git a/docs/dyn/retail_v2beta.projects.locations.catalogs.servingConfigs.html b/docs/dyn/retail_v2beta.projects.locations.catalogs.servingConfigs.html index 78586bd9a3..a1966ce1ad 100644 --- a/docs/dyn/retail_v2beta.projects.locations.catalogs.servingConfigs.html +++ b/docs/dyn/retail_v2beta.projects.locations.catalogs.servingConfigs.html @@ -1035,6 +1035,13 @@

Method Details

], "tileNavigationRequested": True or False, # This field specifies whether the customer would like to request tile navigation. }, + "userAttributes": { # Optional. The user attributes that could be used for personalization of search results. * Populate at most 100 key-value pairs per query. * Only supports string keys and repeated string values. * Duplcate keys are not allowed within a single query. Example: user_attributes: [ { key: "pets" value { values: "dog" values: "cat" } }, { key: "state" value { values: "CA" } } ] + "a_key": { # A list of string values. + "values": [ # String values. + "A String", + ], + }, + }, "userInfo": { # Information of an end user. # User information. "directUserRequest": True or False, # True if the request is made directly from the end user, in which case the ip_address and user_agent can be populated from the HTTP request. This flag should be set only if the API request is made directly from the end user such as a mobile app (and not if a gateway or a server is processing and pushing the user events). This should not be set when using the JavaScript tag in UserEventService.CollectUserEvent. "ipAddress": "A String", # The end user's IP address. This field is used to extract location information for personalization. This field must be either an IPv4 address (e.g. "104.133.9.80") or an IPv6 address (e.g. "2001:0db8:85a3:0000:0000:8a2e:0370:7334"). Otherwise, an INVALID_ARGUMENT error is returned. This should not be set when: * setting SearchRequest.user_info. * using the JavaScript tag in UserEventService.CollectUserEvent or if direct_user_request is set. diff --git a/docs/dyn/script_v1.projects.html b/docs/dyn/script_v1.projects.html index d120f03d24..306fc02805 100644 --- a/docs/dyn/script_v1.projects.html +++ b/docs/dyn/script_v1.projects.html @@ -116,7 +116,7 @@

Method Details

body: object, The request body. The object takes the form of: -{ # Request to create a script project. Request to create a script project. +{ # Request to create a script project. "parentId": "A String", # The Drive ID of a parent file that the created script project is bound to. This is usually the ID of a Google Doc, Google Sheet, Google Form, or Google Slides file. If not set, a standalone script project is created. "title": "A String", # The title for the project. } diff --git a/docs/dyn/securitycenter_v1.folders.securityHealthAnalyticsSettings.customModules.html b/docs/dyn/securitycenter_v1.folders.securityHealthAnalyticsSettings.customModules.html index 8ef862c784..badaec42e1 100644 --- a/docs/dyn/securitycenter_v1.folders.securityHealthAnalyticsSettings.customModules.html +++ b/docs/dyn/securitycenter_v1.folders.securityHealthAnalyticsSettings.customModules.html @@ -1204,6 +1204,7 @@

Method Details

"sha256": "A String", # SHA256 hash of the first hashed_size bytes of the file encoded as a hex string. If hashed_size == size, sha256 represents the SHA256 hash of the entire file. "size": "A String", # Size of the file in bytes. }, + "userId": "A String", # The ID of the user that executed the process. E.g. If this is the root user this will always be 0. }, ], "resourceName": "A String", # For findings on Google Cloud resources, the full resource name of the Google Cloud resource this finding is for. See: https://cloud.google.com/apis/design/resource_names#full_resource_name When the finding is for a non-Google Cloud resource, the resourceName can be a customer or partner defined string. This field is immutable after creation time. diff --git a/docs/dyn/securitycenter_v1.folders.sources.findings.html b/docs/dyn/securitycenter_v1.folders.sources.findings.html index de1c0a57c1..de54d07c16 100644 --- a/docs/dyn/securitycenter_v1.folders.sources.findings.html +++ b/docs/dyn/securitycenter_v1.folders.sources.findings.html @@ -773,6 +773,7 @@

Method Details

"sha256": "A String", # SHA256 hash of the first hashed_size bytes of the file encoded as a hex string. If hashed_size == size, sha256 represents the SHA256 hash of the entire file. "size": "A String", # Size of the file in bytes. }, + "userId": "A String", # The ID of the user that executed the process. E.g. If this is the root user this will always be 0. }, ], "resourceName": "A String", # For findings on Google Cloud resources, the full resource name of the Google Cloud resource this finding is for. See: https://cloud.google.com/apis/design/resource_names#full_resource_name When the finding is for a non-Google Cloud resource, the resourceName can be a customer or partner defined string. This field is immutable after creation time. @@ -1563,6 +1564,7 @@

Method Details

"sha256": "A String", # SHA256 hash of the first hashed_size bytes of the file encoded as a hex string. If hashed_size == size, sha256 represents the SHA256 hash of the entire file. "size": "A String", # Size of the file in bytes. }, + "userId": "A String", # The ID of the user that executed the process. E.g. If this is the root user this will always be 0. }, ], "resourceName": "A String", # For findings on Google Cloud resources, the full resource name of the Google Cloud resource this finding is for. See: https://cloud.google.com/apis/design/resource_names#full_resource_name When the finding is for a non-Google Cloud resource, the resourceName can be a customer or partner defined string. This field is immutable after creation time. @@ -2267,6 +2269,7 @@

Method Details

"sha256": "A String", # SHA256 hash of the first hashed_size bytes of the file encoded as a hex string. If hashed_size == size, sha256 represents the SHA256 hash of the entire file. "size": "A String", # Size of the file in bytes. }, + "userId": "A String", # The ID of the user that executed the process. E.g. If this is the root user this will always be 0. }, ], "resourceName": "A String", # For findings on Google Cloud resources, the full resource name of the Google Cloud resource this finding is for. See: https://cloud.google.com/apis/design/resource_names#full_resource_name When the finding is for a non-Google Cloud resource, the resourceName can be a customer or partner defined string. This field is immutable after creation time. @@ -2984,6 +2987,7 @@

Method Details

"sha256": "A String", # SHA256 hash of the first hashed_size bytes of the file encoded as a hex string. If hashed_size == size, sha256 represents the SHA256 hash of the entire file. "size": "A String", # Size of the file in bytes. }, + "userId": "A String", # The ID of the user that executed the process. E.g. If this is the root user this will always be 0. }, ], "resourceName": "A String", # For findings on Google Cloud resources, the full resource name of the Google Cloud resource this finding is for. See: https://cloud.google.com/apis/design/resource_names#full_resource_name When the finding is for a non-Google Cloud resource, the resourceName can be a customer or partner defined string. This field is immutable after creation time. @@ -3702,6 +3706,7 @@

Method Details

"sha256": "A String", # SHA256 hash of the first hashed_size bytes of the file encoded as a hex string. If hashed_size == size, sha256 represents the SHA256 hash of the entire file. "size": "A String", # Size of the file in bytes. }, + "userId": "A String", # The ID of the user that executed the process. E.g. If this is the root user this will always be 0. }, ], "resourceName": "A String", # For findings on Google Cloud resources, the full resource name of the Google Cloud resource this finding is for. See: https://cloud.google.com/apis/design/resource_names#full_resource_name When the finding is for a non-Google Cloud resource, the resourceName can be a customer or partner defined string. This field is immutable after creation time. diff --git a/docs/dyn/securitycenter_v1.organizations.securityHealthAnalyticsSettings.customModules.html b/docs/dyn/securitycenter_v1.organizations.securityHealthAnalyticsSettings.customModules.html index 8af9ced1be..2e246ffdac 100644 --- a/docs/dyn/securitycenter_v1.organizations.securityHealthAnalyticsSettings.customModules.html +++ b/docs/dyn/securitycenter_v1.organizations.securityHealthAnalyticsSettings.customModules.html @@ -1204,6 +1204,7 @@

Method Details

"sha256": "A String", # SHA256 hash of the first hashed_size bytes of the file encoded as a hex string. If hashed_size == size, sha256 represents the SHA256 hash of the entire file. "size": "A String", # Size of the file in bytes. }, + "userId": "A String", # The ID of the user that executed the process. E.g. If this is the root user this will always be 0. }, ], "resourceName": "A String", # For findings on Google Cloud resources, the full resource name of the Google Cloud resource this finding is for. See: https://cloud.google.com/apis/design/resource_names#full_resource_name When the finding is for a non-Google Cloud resource, the resourceName can be a customer or partner defined string. This field is immutable after creation time. diff --git a/docs/dyn/securitycenter_v1.organizations.sources.findings.html b/docs/dyn/securitycenter_v1.organizations.sources.findings.html index a433eeef66..c9f1c8b352 100644 --- a/docs/dyn/securitycenter_v1.organizations.sources.findings.html +++ b/docs/dyn/securitycenter_v1.organizations.sources.findings.html @@ -706,6 +706,7 @@

Method Details

"sha256": "A String", # SHA256 hash of the first hashed_size bytes of the file encoded as a hex string. If hashed_size == size, sha256 represents the SHA256 hash of the entire file. "size": "A String", # Size of the file in bytes. }, + "userId": "A String", # The ID of the user that executed the process. E.g. If this is the root user this will always be 0. }, ], "resourceName": "A String", # For findings on Google Cloud resources, the full resource name of the Google Cloud resource this finding is for. See: https://cloud.google.com/apis/design/resource_names#full_resource_name When the finding is for a non-Google Cloud resource, the resourceName can be a customer or partner defined string. This field is immutable after creation time. @@ -1410,6 +1411,7 @@

Method Details

"sha256": "A String", # SHA256 hash of the first hashed_size bytes of the file encoded as a hex string. If hashed_size == size, sha256 represents the SHA256 hash of the entire file. "size": "A String", # Size of the file in bytes. }, + "userId": "A String", # The ID of the user that executed the process. E.g. If this is the root user this will always be 0. }, ], "resourceName": "A String", # For findings on Google Cloud resources, the full resource name of the Google Cloud resource this finding is for. See: https://cloud.google.com/apis/design/resource_names#full_resource_name When the finding is for a non-Google Cloud resource, the resourceName can be a customer or partner defined string. This field is immutable after creation time. @@ -2185,6 +2187,7 @@

Method Details

"sha256": "A String", # SHA256 hash of the first hashed_size bytes of the file encoded as a hex string. If hashed_size == size, sha256 represents the SHA256 hash of the entire file. "size": "A String", # Size of the file in bytes. }, + "userId": "A String", # The ID of the user that executed the process. E.g. If this is the root user this will always be 0. }, ], "resourceName": "A String", # For findings on Google Cloud resources, the full resource name of the Google Cloud resource this finding is for. See: https://cloud.google.com/apis/design/resource_names#full_resource_name When the finding is for a non-Google Cloud resource, the resourceName can be a customer or partner defined string. This field is immutable after creation time. @@ -2975,6 +2978,7 @@

Method Details

"sha256": "A String", # SHA256 hash of the first hashed_size bytes of the file encoded as a hex string. If hashed_size == size, sha256 represents the SHA256 hash of the entire file. "size": "A String", # Size of the file in bytes. }, + "userId": "A String", # The ID of the user that executed the process. E.g. If this is the root user this will always be 0. }, ], "resourceName": "A String", # For findings on Google Cloud resources, the full resource name of the Google Cloud resource this finding is for. See: https://cloud.google.com/apis/design/resource_names#full_resource_name When the finding is for a non-Google Cloud resource, the resourceName can be a customer or partner defined string. This field is immutable after creation time. @@ -3679,6 +3683,7 @@

Method Details

"sha256": "A String", # SHA256 hash of the first hashed_size bytes of the file encoded as a hex string. If hashed_size == size, sha256 represents the SHA256 hash of the entire file. "size": "A String", # Size of the file in bytes. }, + "userId": "A String", # The ID of the user that executed the process. E.g. If this is the root user this will always be 0. }, ], "resourceName": "A String", # For findings on Google Cloud resources, the full resource name of the Google Cloud resource this finding is for. See: https://cloud.google.com/apis/design/resource_names#full_resource_name When the finding is for a non-Google Cloud resource, the resourceName can be a customer or partner defined string. This field is immutable after creation time. @@ -4396,6 +4401,7 @@

Method Details

"sha256": "A String", # SHA256 hash of the first hashed_size bytes of the file encoded as a hex string. If hashed_size == size, sha256 represents the SHA256 hash of the entire file. "size": "A String", # Size of the file in bytes. }, + "userId": "A String", # The ID of the user that executed the process. E.g. If this is the root user this will always be 0. }, ], "resourceName": "A String", # For findings on Google Cloud resources, the full resource name of the Google Cloud resource this finding is for. See: https://cloud.google.com/apis/design/resource_names#full_resource_name When the finding is for a non-Google Cloud resource, the resourceName can be a customer or partner defined string. This field is immutable after creation time. @@ -5114,6 +5120,7 @@

Method Details

"sha256": "A String", # SHA256 hash of the first hashed_size bytes of the file encoded as a hex string. If hashed_size == size, sha256 represents the SHA256 hash of the entire file. "size": "A String", # Size of the file in bytes. }, + "userId": "A String", # The ID of the user that executed the process. E.g. If this is the root user this will always be 0. }, ], "resourceName": "A String", # For findings on Google Cloud resources, the full resource name of the Google Cloud resource this finding is for. See: https://cloud.google.com/apis/design/resource_names#full_resource_name When the finding is for a non-Google Cloud resource, the resourceName can be a customer or partner defined string. This field is immutable after creation time. diff --git a/docs/dyn/securitycenter_v1.projects.securityHealthAnalyticsSettings.customModules.html b/docs/dyn/securitycenter_v1.projects.securityHealthAnalyticsSettings.customModules.html index ccc342e2ce..e757d05d47 100644 --- a/docs/dyn/securitycenter_v1.projects.securityHealthAnalyticsSettings.customModules.html +++ b/docs/dyn/securitycenter_v1.projects.securityHealthAnalyticsSettings.customModules.html @@ -1204,6 +1204,7 @@

Method Details

"sha256": "A String", # SHA256 hash of the first hashed_size bytes of the file encoded as a hex string. If hashed_size == size, sha256 represents the SHA256 hash of the entire file. "size": "A String", # Size of the file in bytes. }, + "userId": "A String", # The ID of the user that executed the process. E.g. If this is the root user this will always be 0. }, ], "resourceName": "A String", # For findings on Google Cloud resources, the full resource name of the Google Cloud resource this finding is for. See: https://cloud.google.com/apis/design/resource_names#full_resource_name When the finding is for a non-Google Cloud resource, the resourceName can be a customer or partner defined string. This field is immutable after creation time. diff --git a/docs/dyn/securitycenter_v1.projects.sources.findings.html b/docs/dyn/securitycenter_v1.projects.sources.findings.html index ed4901785e..cb7f43371e 100644 --- a/docs/dyn/securitycenter_v1.projects.sources.findings.html +++ b/docs/dyn/securitycenter_v1.projects.sources.findings.html @@ -773,6 +773,7 @@

Method Details

"sha256": "A String", # SHA256 hash of the first hashed_size bytes of the file encoded as a hex string. If hashed_size == size, sha256 represents the SHA256 hash of the entire file. "size": "A String", # Size of the file in bytes. }, + "userId": "A String", # The ID of the user that executed the process. E.g. If this is the root user this will always be 0. }, ], "resourceName": "A String", # For findings on Google Cloud resources, the full resource name of the Google Cloud resource this finding is for. See: https://cloud.google.com/apis/design/resource_names#full_resource_name When the finding is for a non-Google Cloud resource, the resourceName can be a customer or partner defined string. This field is immutable after creation time. @@ -1563,6 +1564,7 @@

Method Details

"sha256": "A String", # SHA256 hash of the first hashed_size bytes of the file encoded as a hex string. If hashed_size == size, sha256 represents the SHA256 hash of the entire file. "size": "A String", # Size of the file in bytes. }, + "userId": "A String", # The ID of the user that executed the process. E.g. If this is the root user this will always be 0. }, ], "resourceName": "A String", # For findings on Google Cloud resources, the full resource name of the Google Cloud resource this finding is for. See: https://cloud.google.com/apis/design/resource_names#full_resource_name When the finding is for a non-Google Cloud resource, the resourceName can be a customer or partner defined string. This field is immutable after creation time. @@ -2267,6 +2269,7 @@

Method Details

"sha256": "A String", # SHA256 hash of the first hashed_size bytes of the file encoded as a hex string. If hashed_size == size, sha256 represents the SHA256 hash of the entire file. "size": "A String", # Size of the file in bytes. }, + "userId": "A String", # The ID of the user that executed the process. E.g. If this is the root user this will always be 0. }, ], "resourceName": "A String", # For findings on Google Cloud resources, the full resource name of the Google Cloud resource this finding is for. See: https://cloud.google.com/apis/design/resource_names#full_resource_name When the finding is for a non-Google Cloud resource, the resourceName can be a customer or partner defined string. This field is immutable after creation time. @@ -2984,6 +2987,7 @@

Method Details

"sha256": "A String", # SHA256 hash of the first hashed_size bytes of the file encoded as a hex string. If hashed_size == size, sha256 represents the SHA256 hash of the entire file. "size": "A String", # Size of the file in bytes. }, + "userId": "A String", # The ID of the user that executed the process. E.g. If this is the root user this will always be 0. }, ], "resourceName": "A String", # For findings on Google Cloud resources, the full resource name of the Google Cloud resource this finding is for. See: https://cloud.google.com/apis/design/resource_names#full_resource_name When the finding is for a non-Google Cloud resource, the resourceName can be a customer or partner defined string. This field is immutable after creation time. @@ -3702,6 +3706,7 @@

Method Details

"sha256": "A String", # SHA256 hash of the first hashed_size bytes of the file encoded as a hex string. If hashed_size == size, sha256 represents the SHA256 hash of the entire file. "size": "A String", # Size of the file in bytes. }, + "userId": "A String", # The ID of the user that executed the process. E.g. If this is the root user this will always be 0. }, ], "resourceName": "A String", # For findings on Google Cloud resources, the full resource name of the Google Cloud resource this finding is for. See: https://cloud.google.com/apis/design/resource_names#full_resource_name When the finding is for a non-Google Cloud resource, the resourceName can be a customer or partner defined string. This field is immutable after creation time. diff --git a/docs/dyn/spanner_v1.projects.instances.databases.sessions.html b/docs/dyn/spanner_v1.projects.instances.databases.sessions.html index ffc8c23891..81ae623479 100644 --- a/docs/dyn/spanner_v1.projects.instances.databases.sessions.html +++ b/docs/dyn/spanner_v1.projects.instances.databases.sessions.html @@ -446,7 +446,7 @@

Method Details

], }, }, - "options": { # Transactions: Each session can have at most one active transaction at a time (note that standalone reads and queries use a transaction internally and do count towards the one transaction limit). After the active transaction is completed, the session can immediately be re-used for the next transaction. It is not necessary to create a new session for each transaction. Transaction modes: Cloud Spanner supports three transaction modes: 1. Locking read-write. This type of transaction is the only way to write data into Cloud Spanner. These transactions rely on pessimistic locking and, if necessary, two-phase commit. Locking read-write transactions may abort, requiring the application to retry. 2. Snapshot read-only. Snapshot read-only transactions provide guaranteed consistency across several reads, but do not allow writes. Snapshot read-only transactions can be configured to read at timestamps in the past, or configured to perform a strong read (where Spanner selects a timestamp such that the read is guaranteed to see the effects of all transactions that have committed before the start of the read). Snapshot read-only transactions do not need to be committed. Queries on change streams must be performed with the snapshot read-only transaction mode, specifying a strong read. See TransactionOptions.ReadOnly.strong for more details. 3. Partitioned DML. This type of transaction is used to execute a single Partitioned DML statement. Partitioned DML partitions the key space and runs the DML statement over each partition in parallel using separate, internal transactions that commit independently. Partitioned DML transactions do not need to be committed. For transactions that only read, snapshot read-only transactions provide simpler semantics and are almost always faster. In particular, read-only transactions do not take locks, so they do not conflict with read-write transactions. As a consequence of not taking locks, they also do not abort, so retry loops are not needed. Transactions may only read-write data in a single database. They may, however, read-write data in different tables within that database. Locking read-write transactions: Locking transactions may be used to atomically read-modify-write data anywhere in a database. This type of transaction is externally consistent. Clients should attempt to minimize the amount of time a transaction is active. Faster transactions commit with higher probability and cause less contention. Cloud Spanner attempts to keep read locks active as long as the transaction continues to do reads, and the transaction has not been terminated by Commit or Rollback. Long periods of inactivity at the client may cause Cloud Spanner to release a transaction's locks and abort it. Conceptually, a read-write transaction consists of zero or more reads or SQL statements followed by Commit. At any time before Commit, the client can send a Rollback request to abort the transaction. Semantics: Cloud Spanner can commit the transaction if all read locks it acquired are still valid at commit time, and it is able to acquire write locks for all writes. Cloud Spanner can abort the transaction for any reason. If a commit attempt returns `ABORTED`, Cloud Spanner guarantees that the transaction has not modified any user data in Cloud Spanner. Unless the transaction commits, Cloud Spanner makes no guarantees about how long the transaction's locks were held for. It is an error to use Cloud Spanner locks for any sort of mutual exclusion other than between Cloud Spanner transactions themselves. Retrying aborted transactions: When a transaction aborts, the application can choose to retry the whole transaction again. To maximize the chances of successfully committing the retry, the client should execute the retry in the same session as the original attempt. The original session's lock priority increases with each consecutive abort, meaning that each attempt has a slightly better chance of success than the previous. Note that the lock priority is preserved per session (not per transaction). Lock priority is set by the first read or write in the first attempt of a read-write transaction. If the application starts a new session to retry the whole transaction, the transaction loses its original lock priority. Moreover, the lock priority is only preserved if the transaction fails with an `ABORTED` error. Under some circumstances (for example, many transactions attempting to modify the same row(s)), a transaction can abort many times in a short period before successfully committing. Thus, it is not a good idea to cap the number of retries a transaction can attempt; instead, it is better to limit the total amount of time spent retrying. Idle transactions: A transaction is considered idle if it has no outstanding reads or SQL queries and has not started a read or SQL query within the last 10 seconds. Idle transactions can be aborted by Cloud Spanner so that they don't hold on to locks indefinitely. If an idle transaction is aborted, the commit fails with error `ABORTED`. If this behavior is undesirable, periodically executing a simple SQL query in the transaction (for example, `SELECT 1`) prevents the transaction from becoming idle. Snapshot read-only transactions: Snapshot read-only transactions provides a simpler method than locking read-write transactions for doing several consistent reads. However, this type of transaction does not support writes. Snapshot transactions do not take locks. Instead, they work by choosing a Cloud Spanner timestamp, then executing all reads at that timestamp. Since they do not acquire locks, they do not block concurrent read-write transactions. Unlike locking read-write transactions, snapshot read-only transactions never abort. They can fail if the chosen read timestamp is garbage collected; however, the default garbage collection policy is generous enough that most applications do not need to worry about this in practice. Snapshot read-only transactions do not need to call Commit or Rollback (and in fact are not permitted to do so). To execute a snapshot transaction, the client specifies a timestamp bound, which tells Cloud Spanner how to choose a read timestamp. The types of timestamp bound are: - Strong (the default). - Bounded staleness. - Exact staleness. If the Cloud Spanner database to be read is geographically distributed, stale read-only transactions can execute more quickly than strong or read-write transactions, because they are able to execute far from the leader replica. Each type of timestamp bound is discussed in detail below. Strong: Strong reads are guaranteed to see the effects of all transactions that have committed before the start of the read. Furthermore, all rows yielded by a single read are consistent with each other -- if any part of the read observes a transaction, all parts of the read see the transaction. Strong reads are not repeatable: two consecutive strong read-only transactions might return inconsistent results if there are concurrent writes. If consistency across reads is required, the reads should be executed within a transaction or at an exact read timestamp. Queries on change streams (see below for more details) must also specify the strong read timestamp bound. See TransactionOptions.ReadOnly.strong. Exact staleness: These timestamp bounds execute reads at a user-specified timestamp. Reads at a timestamp are guaranteed to see a consistent prefix of the global transaction history: they observe modifications done by all transactions with a commit timestamp less than or equal to the read timestamp, and observe none of the modifications done by transactions with a larger commit timestamp. They block until all conflicting transactions that can be assigned commit timestamps <= the read timestamp have finished. The timestamp can either be expressed as an absolute Cloud Spanner commit timestamp or a staleness relative to the current time. These modes do not require a "negotiation phase" to pick a timestamp. As a result, they execute slightly faster than the equivalent boundedly stale concurrency modes. On the other hand, boundedly stale reads usually return fresher results. See TransactionOptions.ReadOnly.read_timestamp and TransactionOptions.ReadOnly.exact_staleness. Bounded staleness: Bounded staleness modes allow Cloud Spanner to pick the read timestamp, subject to a user-provided staleness bound. Cloud Spanner chooses the newest timestamp within the staleness bound that allows execution of the reads at the closest available replica without blocking. All rows yielded are consistent with each other -- if any part of the read observes a transaction, all parts of the read see the transaction. Boundedly stale reads are not repeatable: two stale reads, even if they use the same staleness bound, can execute at different timestamps and thus return inconsistent results. Boundedly stale reads execute in two phases: the first phase negotiates a timestamp among all replicas needed to serve the read. In the second phase, reads are executed at the negotiated timestamp. As a result of the two phase execution, bounded staleness reads are usually a little slower than comparable exact staleness reads. However, they are typically able to return fresher results, and are more likely to execute at the closest replica. Because the timestamp negotiation requires up-front knowledge of which rows are read, it can only be used with single-use read-only transactions. See TransactionOptions.ReadOnly.max_staleness and TransactionOptions.ReadOnly.min_read_timestamp. Old read timestamps and garbage collection: Cloud Spanner continuously garbage collects deleted and overwritten data in the background to reclaim storage space. This process is known as "version GC". By default, version GC reclaims versions after they are one hour old. Because of this, Cloud Spanner can't perform reads at read timestamps more than one hour in the past. This restriction also applies to in-progress reads and/or SQL queries whose timestamp become too old while executing. Reads and SQL queries with too-old read timestamps fail with the error `FAILED_PRECONDITION`. You can configure and extend the `VERSION_RETENTION_PERIOD` of a database up to a period as long as one week, which allows Cloud Spanner to perform reads up to one week in the past. Querying change Streams: A Change Stream is a schema object that can be configured to watch data changes on the entire database, a set of tables, or a set of columns in a database. When a change stream is created, Spanner automatically defines a corresponding SQL Table-Valued Function (TVF) that can be used to query the change records in the associated change stream using the ExecuteStreamingSql API. The name of the TVF for a change stream is generated from the name of the change stream: READ_. All queries on change stream TVFs must be executed using the ExecuteStreamingSql API with a single-use read-only transaction with a strong read-only timestamp_bound. The change stream TVF allows users to specify the start_timestamp and end_timestamp for the time range of interest. All change records within the retention period is accessible using the strong read-only timestamp_bound. All other TransactionOptions are invalid for change stream queries. In addition, if TransactionOptions.read_only.return_read_timestamp is set to true, a special value of 2^63 - 2 is returned in the Transaction message that describes the transaction, instead of a valid read timestamp. This special value should be discarded and not used for any subsequent queries. Please see https://cloud.google.com/spanner/docs/change-streams for more details on how to query the change stream TVFs. Partitioned DML transactions: Partitioned DML transactions are used to execute DML statements with a different execution strategy that provides different, and often better, scalability properties for large, table-wide operations than DML in a ReadWrite transaction. Smaller scoped statements, such as an OLTP workload, should prefer using ReadWrite transactions. Partitioned DML partitions the keyspace and runs the DML statement on each partition in separate, internal transactions. These transactions commit automatically when complete, and run independently from one another. To reduce lock contention, this execution strategy only acquires read locks on rows that match the WHERE clause of the statement. Additionally, the smaller per-partition transactions hold locks for less time. That said, Partitioned DML is not a drop-in replacement for standard DML used in ReadWrite transactions. - The DML statement must be fully-partitionable. Specifically, the statement must be expressible as the union of many statements which each access only a single row of the table. - The statement is not applied atomically to all rows of the table. Rather, the statement is applied atomically to partitions of the table, in independent transactions. Secondary index rows are updated atomically with the base table rows. - Partitioned DML does not guarantee exactly-once execution semantics against a partition. The statement is applied at least once to each partition. It is strongly recommended that the DML statement should be idempotent to avoid unexpected results. For instance, it is potentially dangerous to run a statement such as `UPDATE table SET column = column + 1` as it could be run multiple times against some rows. - The partitions are committed automatically - there is no support for Commit or Rollback. If the call returns an error, or if the client issuing the ExecuteSql call dies, it is possible that some rows had the statement executed on them successfully. It is also possible that statement was never executed against other rows. - Partitioned DML transactions may only contain the execution of a single DML statement via ExecuteSql or ExecuteStreamingSql. - If any error is encountered during the execution of the partitioned DML operation (for instance, a UNIQUE INDEX violation, division by zero, or a value that can't be stored due to schema constraints), then the operation is stopped at that point and an error is returned. It is possible that at this point, some partitions have been committed (or even committed multiple times), and other partitions have not been run at all. Given the above, Partitioned DML is good fit for large, database-wide, operations that are idempotent, such as deleting old rows from a very large table. # Required. Options for the new transaction. + "options": { # Options to use for transactions. # Required. Options for the new transaction. "excludeTxnFromChangeStreams": True or False, # When `exclude_txn_from_change_streams` is set to `true`, it prevents read or write transactions from being tracked in change streams. * If the DDL option `allow_txn_exclusion` is set to `true`, then the updates made within this transaction aren't recorded in the change stream. * If you don't set the DDL option `allow_txn_exclusion` or if it's set to `false`, then the updates made within this transaction are recorded in the change stream. When `exclude_txn_from_change_streams` is set to `false` or not set, modifications from this transaction are recorded in all change streams that are tracking columns modified by these transactions. The `exclude_txn_from_change_streams` option can only be specified for read-write or partitioned DML transactions, otherwise the API returns an `INVALID_ARGUMENT` error. "isolationLevel": "A String", # Isolation level for the transaction. "partitionedDml": { # Message type to initiate a Partitioned DML transaction. # Partitioned DML transaction. Authorization to begin a Partitioned DML transaction requires `spanner.databases.beginPartitionedDmlTransaction` permission on the `session` resource. @@ -590,7 +590,7 @@

Method Details

"transactionTag": "A String", # A tag used for statistics collection about this transaction. Both `request_tag` and `transaction_tag` can be specified for a read or query that belongs to a transaction. The value of transaction_tag should be the same for all requests belonging to the same transaction. If this request doesn't belong to any transaction, `transaction_tag` is ignored. Legal characters for `transaction_tag` values are all printable characters (ASCII 32 - 126) and the length of a `transaction_tag` is limited to 50 characters. Values that exceed this limit are truncated. Any leading underscore (_) characters are removed from the string. }, "returnCommitStats": True or False, # If `true`, then statistics related to the transaction is included in the CommitResponse. Default value is `false`. - "singleUseTransaction": { # Transactions: Each session can have at most one active transaction at a time (note that standalone reads and queries use a transaction internally and do count towards the one transaction limit). After the active transaction is completed, the session can immediately be re-used for the next transaction. It is not necessary to create a new session for each transaction. Transaction modes: Cloud Spanner supports three transaction modes: 1. Locking read-write. This type of transaction is the only way to write data into Cloud Spanner. These transactions rely on pessimistic locking and, if necessary, two-phase commit. Locking read-write transactions may abort, requiring the application to retry. 2. Snapshot read-only. Snapshot read-only transactions provide guaranteed consistency across several reads, but do not allow writes. Snapshot read-only transactions can be configured to read at timestamps in the past, or configured to perform a strong read (where Spanner selects a timestamp such that the read is guaranteed to see the effects of all transactions that have committed before the start of the read). Snapshot read-only transactions do not need to be committed. Queries on change streams must be performed with the snapshot read-only transaction mode, specifying a strong read. See TransactionOptions.ReadOnly.strong for more details. 3. Partitioned DML. This type of transaction is used to execute a single Partitioned DML statement. Partitioned DML partitions the key space and runs the DML statement over each partition in parallel using separate, internal transactions that commit independently. Partitioned DML transactions do not need to be committed. For transactions that only read, snapshot read-only transactions provide simpler semantics and are almost always faster. In particular, read-only transactions do not take locks, so they do not conflict with read-write transactions. As a consequence of not taking locks, they also do not abort, so retry loops are not needed. Transactions may only read-write data in a single database. They may, however, read-write data in different tables within that database. Locking read-write transactions: Locking transactions may be used to atomically read-modify-write data anywhere in a database. This type of transaction is externally consistent. Clients should attempt to minimize the amount of time a transaction is active. Faster transactions commit with higher probability and cause less contention. Cloud Spanner attempts to keep read locks active as long as the transaction continues to do reads, and the transaction has not been terminated by Commit or Rollback. Long periods of inactivity at the client may cause Cloud Spanner to release a transaction's locks and abort it. Conceptually, a read-write transaction consists of zero or more reads or SQL statements followed by Commit. At any time before Commit, the client can send a Rollback request to abort the transaction. Semantics: Cloud Spanner can commit the transaction if all read locks it acquired are still valid at commit time, and it is able to acquire write locks for all writes. Cloud Spanner can abort the transaction for any reason. If a commit attempt returns `ABORTED`, Cloud Spanner guarantees that the transaction has not modified any user data in Cloud Spanner. Unless the transaction commits, Cloud Spanner makes no guarantees about how long the transaction's locks were held for. It is an error to use Cloud Spanner locks for any sort of mutual exclusion other than between Cloud Spanner transactions themselves. Retrying aborted transactions: When a transaction aborts, the application can choose to retry the whole transaction again. To maximize the chances of successfully committing the retry, the client should execute the retry in the same session as the original attempt. The original session's lock priority increases with each consecutive abort, meaning that each attempt has a slightly better chance of success than the previous. Note that the lock priority is preserved per session (not per transaction). Lock priority is set by the first read or write in the first attempt of a read-write transaction. If the application starts a new session to retry the whole transaction, the transaction loses its original lock priority. Moreover, the lock priority is only preserved if the transaction fails with an `ABORTED` error. Under some circumstances (for example, many transactions attempting to modify the same row(s)), a transaction can abort many times in a short period before successfully committing. Thus, it is not a good idea to cap the number of retries a transaction can attempt; instead, it is better to limit the total amount of time spent retrying. Idle transactions: A transaction is considered idle if it has no outstanding reads or SQL queries and has not started a read or SQL query within the last 10 seconds. Idle transactions can be aborted by Cloud Spanner so that they don't hold on to locks indefinitely. If an idle transaction is aborted, the commit fails with error `ABORTED`. If this behavior is undesirable, periodically executing a simple SQL query in the transaction (for example, `SELECT 1`) prevents the transaction from becoming idle. Snapshot read-only transactions: Snapshot read-only transactions provides a simpler method than locking read-write transactions for doing several consistent reads. However, this type of transaction does not support writes. Snapshot transactions do not take locks. Instead, they work by choosing a Cloud Spanner timestamp, then executing all reads at that timestamp. Since they do not acquire locks, they do not block concurrent read-write transactions. Unlike locking read-write transactions, snapshot read-only transactions never abort. They can fail if the chosen read timestamp is garbage collected; however, the default garbage collection policy is generous enough that most applications do not need to worry about this in practice. Snapshot read-only transactions do not need to call Commit or Rollback (and in fact are not permitted to do so). To execute a snapshot transaction, the client specifies a timestamp bound, which tells Cloud Spanner how to choose a read timestamp. The types of timestamp bound are: - Strong (the default). - Bounded staleness. - Exact staleness. If the Cloud Spanner database to be read is geographically distributed, stale read-only transactions can execute more quickly than strong or read-write transactions, because they are able to execute far from the leader replica. Each type of timestamp bound is discussed in detail below. Strong: Strong reads are guaranteed to see the effects of all transactions that have committed before the start of the read. Furthermore, all rows yielded by a single read are consistent with each other -- if any part of the read observes a transaction, all parts of the read see the transaction. Strong reads are not repeatable: two consecutive strong read-only transactions might return inconsistent results if there are concurrent writes. If consistency across reads is required, the reads should be executed within a transaction or at an exact read timestamp. Queries on change streams (see below for more details) must also specify the strong read timestamp bound. See TransactionOptions.ReadOnly.strong. Exact staleness: These timestamp bounds execute reads at a user-specified timestamp. Reads at a timestamp are guaranteed to see a consistent prefix of the global transaction history: they observe modifications done by all transactions with a commit timestamp less than or equal to the read timestamp, and observe none of the modifications done by transactions with a larger commit timestamp. They block until all conflicting transactions that can be assigned commit timestamps <= the read timestamp have finished. The timestamp can either be expressed as an absolute Cloud Spanner commit timestamp or a staleness relative to the current time. These modes do not require a "negotiation phase" to pick a timestamp. As a result, they execute slightly faster than the equivalent boundedly stale concurrency modes. On the other hand, boundedly stale reads usually return fresher results. See TransactionOptions.ReadOnly.read_timestamp and TransactionOptions.ReadOnly.exact_staleness. Bounded staleness: Bounded staleness modes allow Cloud Spanner to pick the read timestamp, subject to a user-provided staleness bound. Cloud Spanner chooses the newest timestamp within the staleness bound that allows execution of the reads at the closest available replica without blocking. All rows yielded are consistent with each other -- if any part of the read observes a transaction, all parts of the read see the transaction. Boundedly stale reads are not repeatable: two stale reads, even if they use the same staleness bound, can execute at different timestamps and thus return inconsistent results. Boundedly stale reads execute in two phases: the first phase negotiates a timestamp among all replicas needed to serve the read. In the second phase, reads are executed at the negotiated timestamp. As a result of the two phase execution, bounded staleness reads are usually a little slower than comparable exact staleness reads. However, they are typically able to return fresher results, and are more likely to execute at the closest replica. Because the timestamp negotiation requires up-front knowledge of which rows are read, it can only be used with single-use read-only transactions. See TransactionOptions.ReadOnly.max_staleness and TransactionOptions.ReadOnly.min_read_timestamp. Old read timestamps and garbage collection: Cloud Spanner continuously garbage collects deleted and overwritten data in the background to reclaim storage space. This process is known as "version GC". By default, version GC reclaims versions after they are one hour old. Because of this, Cloud Spanner can't perform reads at read timestamps more than one hour in the past. This restriction also applies to in-progress reads and/or SQL queries whose timestamp become too old while executing. Reads and SQL queries with too-old read timestamps fail with the error `FAILED_PRECONDITION`. You can configure and extend the `VERSION_RETENTION_PERIOD` of a database up to a period as long as one week, which allows Cloud Spanner to perform reads up to one week in the past. Querying change Streams: A Change Stream is a schema object that can be configured to watch data changes on the entire database, a set of tables, or a set of columns in a database. When a change stream is created, Spanner automatically defines a corresponding SQL Table-Valued Function (TVF) that can be used to query the change records in the associated change stream using the ExecuteStreamingSql API. The name of the TVF for a change stream is generated from the name of the change stream: READ_. All queries on change stream TVFs must be executed using the ExecuteStreamingSql API with a single-use read-only transaction with a strong read-only timestamp_bound. The change stream TVF allows users to specify the start_timestamp and end_timestamp for the time range of interest. All change records within the retention period is accessible using the strong read-only timestamp_bound. All other TransactionOptions are invalid for change stream queries. In addition, if TransactionOptions.read_only.return_read_timestamp is set to true, a special value of 2^63 - 2 is returned in the Transaction message that describes the transaction, instead of a valid read timestamp. This special value should be discarded and not used for any subsequent queries. Please see https://cloud.google.com/spanner/docs/change-streams for more details on how to query the change stream TVFs. Partitioned DML transactions: Partitioned DML transactions are used to execute DML statements with a different execution strategy that provides different, and often better, scalability properties for large, table-wide operations than DML in a ReadWrite transaction. Smaller scoped statements, such as an OLTP workload, should prefer using ReadWrite transactions. Partitioned DML partitions the keyspace and runs the DML statement on each partition in separate, internal transactions. These transactions commit automatically when complete, and run independently from one another. To reduce lock contention, this execution strategy only acquires read locks on rows that match the WHERE clause of the statement. Additionally, the smaller per-partition transactions hold locks for less time. That said, Partitioned DML is not a drop-in replacement for standard DML used in ReadWrite transactions. - The DML statement must be fully-partitionable. Specifically, the statement must be expressible as the union of many statements which each access only a single row of the table. - The statement is not applied atomically to all rows of the table. Rather, the statement is applied atomically to partitions of the table, in independent transactions. Secondary index rows are updated atomically with the base table rows. - Partitioned DML does not guarantee exactly-once execution semantics against a partition. The statement is applied at least once to each partition. It is strongly recommended that the DML statement should be idempotent to avoid unexpected results. For instance, it is potentially dangerous to run a statement such as `UPDATE table SET column = column + 1` as it could be run multiple times against some rows. - The partitions are committed automatically - there is no support for Commit or Rollback. If the call returns an error, or if the client issuing the ExecuteSql call dies, it is possible that some rows had the statement executed on them successfully. It is also possible that statement was never executed against other rows. - Partitioned DML transactions may only contain the execution of a single DML statement via ExecuteSql or ExecuteStreamingSql. - If any error is encountered during the execution of the partitioned DML operation (for instance, a UNIQUE INDEX violation, division by zero, or a value that can't be stored due to schema constraints), then the operation is stopped at that point and an error is returned. It is possible that at this point, some partitions have been committed (or even committed multiple times), and other partitions have not been run at all. Given the above, Partitioned DML is good fit for large, database-wide, operations that are idempotent, such as deleting old rows from a very large table. # Execute mutations in a temporary transaction. Note that unlike commit of a previously-started transaction, commit with a temporary transaction is non-idempotent. That is, if the `CommitRequest` is sent to Cloud Spanner more than once (for instance, due to retries in the application, or in the transport library), it's possible that the mutations are executed more than once. If this is undesirable, use BeginTransaction and Commit instead. + "singleUseTransaction": { # Options to use for transactions. # Execute mutations in a temporary transaction. Note that unlike commit of a previously-started transaction, commit with a temporary transaction is non-idempotent. That is, if the `CommitRequest` is sent to Cloud Spanner more than once (for instance, due to retries in the application, or in the transport library), it's possible that the mutations are executed more than once. If this is undesirable, use BeginTransaction and Commit instead. "excludeTxnFromChangeStreams": True or False, # When `exclude_txn_from_change_streams` is set to `true`, it prevents read or write transactions from being tracked in change streams. * If the DDL option `allow_txn_exclusion` is set to `true`, then the updates made within this transaction aren't recorded in the change stream. * If you don't set the DDL option `allow_txn_exclusion` or if it's set to `false`, then the updates made within this transaction are recorded in the change stream. When `exclude_txn_from_change_streams` is set to `false` or not set, modifications from this transaction are recorded in all change streams that are tracking columns modified by these transactions. The `exclude_txn_from_change_streams` option can only be specified for read-write or partitioned DML transactions, otherwise the API returns an `INVALID_ARGUMENT` error. "isolationLevel": "A String", # Isolation level for the transaction. "partitionedDml": { # Message type to initiate a Partitioned DML transaction. # Partitioned DML transaction. Authorization to begin a Partitioned DML transaction requires `spanner.databases.beginPartitionedDmlTransaction` permission on the `session` resource. @@ -733,7 +733,7 @@

Method Details

}, ], "transaction": { # This message is used to select the transaction in which a Read or ExecuteSql call runs. See TransactionOptions for more information about transactions. # Required. The transaction to use. Must be a read-write transaction. To protect against replays, single-use transactions are not supported. The caller must either supply an existing transaction ID or begin a new transaction. - "begin": { # Transactions: Each session can have at most one active transaction at a time (note that standalone reads and queries use a transaction internally and do count towards the one transaction limit). After the active transaction is completed, the session can immediately be re-used for the next transaction. It is not necessary to create a new session for each transaction. Transaction modes: Cloud Spanner supports three transaction modes: 1. Locking read-write. This type of transaction is the only way to write data into Cloud Spanner. These transactions rely on pessimistic locking and, if necessary, two-phase commit. Locking read-write transactions may abort, requiring the application to retry. 2. Snapshot read-only. Snapshot read-only transactions provide guaranteed consistency across several reads, but do not allow writes. Snapshot read-only transactions can be configured to read at timestamps in the past, or configured to perform a strong read (where Spanner selects a timestamp such that the read is guaranteed to see the effects of all transactions that have committed before the start of the read). Snapshot read-only transactions do not need to be committed. Queries on change streams must be performed with the snapshot read-only transaction mode, specifying a strong read. See TransactionOptions.ReadOnly.strong for more details. 3. Partitioned DML. This type of transaction is used to execute a single Partitioned DML statement. Partitioned DML partitions the key space and runs the DML statement over each partition in parallel using separate, internal transactions that commit independently. Partitioned DML transactions do not need to be committed. For transactions that only read, snapshot read-only transactions provide simpler semantics and are almost always faster. In particular, read-only transactions do not take locks, so they do not conflict with read-write transactions. As a consequence of not taking locks, they also do not abort, so retry loops are not needed. Transactions may only read-write data in a single database. They may, however, read-write data in different tables within that database. Locking read-write transactions: Locking transactions may be used to atomically read-modify-write data anywhere in a database. This type of transaction is externally consistent. Clients should attempt to minimize the amount of time a transaction is active. Faster transactions commit with higher probability and cause less contention. Cloud Spanner attempts to keep read locks active as long as the transaction continues to do reads, and the transaction has not been terminated by Commit or Rollback. Long periods of inactivity at the client may cause Cloud Spanner to release a transaction's locks and abort it. Conceptually, a read-write transaction consists of zero or more reads or SQL statements followed by Commit. At any time before Commit, the client can send a Rollback request to abort the transaction. Semantics: Cloud Spanner can commit the transaction if all read locks it acquired are still valid at commit time, and it is able to acquire write locks for all writes. Cloud Spanner can abort the transaction for any reason. If a commit attempt returns `ABORTED`, Cloud Spanner guarantees that the transaction has not modified any user data in Cloud Spanner. Unless the transaction commits, Cloud Spanner makes no guarantees about how long the transaction's locks were held for. It is an error to use Cloud Spanner locks for any sort of mutual exclusion other than between Cloud Spanner transactions themselves. Retrying aborted transactions: When a transaction aborts, the application can choose to retry the whole transaction again. To maximize the chances of successfully committing the retry, the client should execute the retry in the same session as the original attempt. The original session's lock priority increases with each consecutive abort, meaning that each attempt has a slightly better chance of success than the previous. Note that the lock priority is preserved per session (not per transaction). Lock priority is set by the first read or write in the first attempt of a read-write transaction. If the application starts a new session to retry the whole transaction, the transaction loses its original lock priority. Moreover, the lock priority is only preserved if the transaction fails with an `ABORTED` error. Under some circumstances (for example, many transactions attempting to modify the same row(s)), a transaction can abort many times in a short period before successfully committing. Thus, it is not a good idea to cap the number of retries a transaction can attempt; instead, it is better to limit the total amount of time spent retrying. Idle transactions: A transaction is considered idle if it has no outstanding reads or SQL queries and has not started a read or SQL query within the last 10 seconds. Idle transactions can be aborted by Cloud Spanner so that they don't hold on to locks indefinitely. If an idle transaction is aborted, the commit fails with error `ABORTED`. If this behavior is undesirable, periodically executing a simple SQL query in the transaction (for example, `SELECT 1`) prevents the transaction from becoming idle. Snapshot read-only transactions: Snapshot read-only transactions provides a simpler method than locking read-write transactions for doing several consistent reads. However, this type of transaction does not support writes. Snapshot transactions do not take locks. Instead, they work by choosing a Cloud Spanner timestamp, then executing all reads at that timestamp. Since they do not acquire locks, they do not block concurrent read-write transactions. Unlike locking read-write transactions, snapshot read-only transactions never abort. They can fail if the chosen read timestamp is garbage collected; however, the default garbage collection policy is generous enough that most applications do not need to worry about this in practice. Snapshot read-only transactions do not need to call Commit or Rollback (and in fact are not permitted to do so). To execute a snapshot transaction, the client specifies a timestamp bound, which tells Cloud Spanner how to choose a read timestamp. The types of timestamp bound are: - Strong (the default). - Bounded staleness. - Exact staleness. If the Cloud Spanner database to be read is geographically distributed, stale read-only transactions can execute more quickly than strong or read-write transactions, because they are able to execute far from the leader replica. Each type of timestamp bound is discussed in detail below. Strong: Strong reads are guaranteed to see the effects of all transactions that have committed before the start of the read. Furthermore, all rows yielded by a single read are consistent with each other -- if any part of the read observes a transaction, all parts of the read see the transaction. Strong reads are not repeatable: two consecutive strong read-only transactions might return inconsistent results if there are concurrent writes. If consistency across reads is required, the reads should be executed within a transaction or at an exact read timestamp. Queries on change streams (see below for more details) must also specify the strong read timestamp bound. See TransactionOptions.ReadOnly.strong. Exact staleness: These timestamp bounds execute reads at a user-specified timestamp. Reads at a timestamp are guaranteed to see a consistent prefix of the global transaction history: they observe modifications done by all transactions with a commit timestamp less than or equal to the read timestamp, and observe none of the modifications done by transactions with a larger commit timestamp. They block until all conflicting transactions that can be assigned commit timestamps <= the read timestamp have finished. The timestamp can either be expressed as an absolute Cloud Spanner commit timestamp or a staleness relative to the current time. These modes do not require a "negotiation phase" to pick a timestamp. As a result, they execute slightly faster than the equivalent boundedly stale concurrency modes. On the other hand, boundedly stale reads usually return fresher results. See TransactionOptions.ReadOnly.read_timestamp and TransactionOptions.ReadOnly.exact_staleness. Bounded staleness: Bounded staleness modes allow Cloud Spanner to pick the read timestamp, subject to a user-provided staleness bound. Cloud Spanner chooses the newest timestamp within the staleness bound that allows execution of the reads at the closest available replica without blocking. All rows yielded are consistent with each other -- if any part of the read observes a transaction, all parts of the read see the transaction. Boundedly stale reads are not repeatable: two stale reads, even if they use the same staleness bound, can execute at different timestamps and thus return inconsistent results. Boundedly stale reads execute in two phases: the first phase negotiates a timestamp among all replicas needed to serve the read. In the second phase, reads are executed at the negotiated timestamp. As a result of the two phase execution, bounded staleness reads are usually a little slower than comparable exact staleness reads. However, they are typically able to return fresher results, and are more likely to execute at the closest replica. Because the timestamp negotiation requires up-front knowledge of which rows are read, it can only be used with single-use read-only transactions. See TransactionOptions.ReadOnly.max_staleness and TransactionOptions.ReadOnly.min_read_timestamp. Old read timestamps and garbage collection: Cloud Spanner continuously garbage collects deleted and overwritten data in the background to reclaim storage space. This process is known as "version GC". By default, version GC reclaims versions after they are one hour old. Because of this, Cloud Spanner can't perform reads at read timestamps more than one hour in the past. This restriction also applies to in-progress reads and/or SQL queries whose timestamp become too old while executing. Reads and SQL queries with too-old read timestamps fail with the error `FAILED_PRECONDITION`. You can configure and extend the `VERSION_RETENTION_PERIOD` of a database up to a period as long as one week, which allows Cloud Spanner to perform reads up to one week in the past. Querying change Streams: A Change Stream is a schema object that can be configured to watch data changes on the entire database, a set of tables, or a set of columns in a database. When a change stream is created, Spanner automatically defines a corresponding SQL Table-Valued Function (TVF) that can be used to query the change records in the associated change stream using the ExecuteStreamingSql API. The name of the TVF for a change stream is generated from the name of the change stream: READ_. All queries on change stream TVFs must be executed using the ExecuteStreamingSql API with a single-use read-only transaction with a strong read-only timestamp_bound. The change stream TVF allows users to specify the start_timestamp and end_timestamp for the time range of interest. All change records within the retention period is accessible using the strong read-only timestamp_bound. All other TransactionOptions are invalid for change stream queries. In addition, if TransactionOptions.read_only.return_read_timestamp is set to true, a special value of 2^63 - 2 is returned in the Transaction message that describes the transaction, instead of a valid read timestamp. This special value should be discarded and not used for any subsequent queries. Please see https://cloud.google.com/spanner/docs/change-streams for more details on how to query the change stream TVFs. Partitioned DML transactions: Partitioned DML transactions are used to execute DML statements with a different execution strategy that provides different, and often better, scalability properties for large, table-wide operations than DML in a ReadWrite transaction. Smaller scoped statements, such as an OLTP workload, should prefer using ReadWrite transactions. Partitioned DML partitions the keyspace and runs the DML statement on each partition in separate, internal transactions. These transactions commit automatically when complete, and run independently from one another. To reduce lock contention, this execution strategy only acquires read locks on rows that match the WHERE clause of the statement. Additionally, the smaller per-partition transactions hold locks for less time. That said, Partitioned DML is not a drop-in replacement for standard DML used in ReadWrite transactions. - The DML statement must be fully-partitionable. Specifically, the statement must be expressible as the union of many statements which each access only a single row of the table. - The statement is not applied atomically to all rows of the table. Rather, the statement is applied atomically to partitions of the table, in independent transactions. Secondary index rows are updated atomically with the base table rows. - Partitioned DML does not guarantee exactly-once execution semantics against a partition. The statement is applied at least once to each partition. It is strongly recommended that the DML statement should be idempotent to avoid unexpected results. For instance, it is potentially dangerous to run a statement such as `UPDATE table SET column = column + 1` as it could be run multiple times against some rows. - The partitions are committed automatically - there is no support for Commit or Rollback. If the call returns an error, or if the client issuing the ExecuteSql call dies, it is possible that some rows had the statement executed on them successfully. It is also possible that statement was never executed against other rows. - Partitioned DML transactions may only contain the execution of a single DML statement via ExecuteSql or ExecuteStreamingSql. - If any error is encountered during the execution of the partitioned DML operation (for instance, a UNIQUE INDEX violation, division by zero, or a value that can't be stored due to schema constraints), then the operation is stopped at that point and an error is returned. It is possible that at this point, some partitions have been committed (or even committed multiple times), and other partitions have not been run at all. Given the above, Partitioned DML is good fit for large, database-wide, operations that are idempotent, such as deleting old rows from a very large table. # Begin a new transaction and execute this read or SQL query in it. The transaction ID of the new transaction is returned in ResultSetMetadata.transaction, which is a Transaction. + "begin": { # Options to use for transactions. # Begin a new transaction and execute this read or SQL query in it. The transaction ID of the new transaction is returned in ResultSetMetadata.transaction, which is a Transaction. "excludeTxnFromChangeStreams": True or False, # When `exclude_txn_from_change_streams` is set to `true`, it prevents read or write transactions from being tracked in change streams. * If the DDL option `allow_txn_exclusion` is set to `true`, then the updates made within this transaction aren't recorded in the change stream. * If you don't set the DDL option `allow_txn_exclusion` or if it's set to `false`, then the updates made within this transaction are recorded in the change stream. When `exclude_txn_from_change_streams` is set to `false` or not set, modifications from this transaction are recorded in all change streams that are tracking columns modified by these transactions. The `exclude_txn_from_change_streams` option can only be specified for read-write or partitioned DML transactions, otherwise the API returns an `INVALID_ARGUMENT` error. "isolationLevel": "A String", # Isolation level for the transaction. "partitionedDml": { # Message type to initiate a Partitioned DML transaction. # Partitioned DML transaction. Authorization to begin a Partitioned DML transaction requires `spanner.databases.beginPartitionedDmlTransaction` permission on the `session` resource. @@ -752,7 +752,7 @@

Method Details

}, }, "id": "A String", # Execute the read or SQL query in a previously-started transaction. - "singleUse": { # Transactions: Each session can have at most one active transaction at a time (note that standalone reads and queries use a transaction internally and do count towards the one transaction limit). After the active transaction is completed, the session can immediately be re-used for the next transaction. It is not necessary to create a new session for each transaction. Transaction modes: Cloud Spanner supports three transaction modes: 1. Locking read-write. This type of transaction is the only way to write data into Cloud Spanner. These transactions rely on pessimistic locking and, if necessary, two-phase commit. Locking read-write transactions may abort, requiring the application to retry. 2. Snapshot read-only. Snapshot read-only transactions provide guaranteed consistency across several reads, but do not allow writes. Snapshot read-only transactions can be configured to read at timestamps in the past, or configured to perform a strong read (where Spanner selects a timestamp such that the read is guaranteed to see the effects of all transactions that have committed before the start of the read). Snapshot read-only transactions do not need to be committed. Queries on change streams must be performed with the snapshot read-only transaction mode, specifying a strong read. See TransactionOptions.ReadOnly.strong for more details. 3. Partitioned DML. This type of transaction is used to execute a single Partitioned DML statement. Partitioned DML partitions the key space and runs the DML statement over each partition in parallel using separate, internal transactions that commit independently. Partitioned DML transactions do not need to be committed. For transactions that only read, snapshot read-only transactions provide simpler semantics and are almost always faster. In particular, read-only transactions do not take locks, so they do not conflict with read-write transactions. As a consequence of not taking locks, they also do not abort, so retry loops are not needed. Transactions may only read-write data in a single database. They may, however, read-write data in different tables within that database. Locking read-write transactions: Locking transactions may be used to atomically read-modify-write data anywhere in a database. This type of transaction is externally consistent. Clients should attempt to minimize the amount of time a transaction is active. Faster transactions commit with higher probability and cause less contention. Cloud Spanner attempts to keep read locks active as long as the transaction continues to do reads, and the transaction has not been terminated by Commit or Rollback. Long periods of inactivity at the client may cause Cloud Spanner to release a transaction's locks and abort it. Conceptually, a read-write transaction consists of zero or more reads or SQL statements followed by Commit. At any time before Commit, the client can send a Rollback request to abort the transaction. Semantics: Cloud Spanner can commit the transaction if all read locks it acquired are still valid at commit time, and it is able to acquire write locks for all writes. Cloud Spanner can abort the transaction for any reason. If a commit attempt returns `ABORTED`, Cloud Spanner guarantees that the transaction has not modified any user data in Cloud Spanner. Unless the transaction commits, Cloud Spanner makes no guarantees about how long the transaction's locks were held for. It is an error to use Cloud Spanner locks for any sort of mutual exclusion other than between Cloud Spanner transactions themselves. Retrying aborted transactions: When a transaction aborts, the application can choose to retry the whole transaction again. To maximize the chances of successfully committing the retry, the client should execute the retry in the same session as the original attempt. The original session's lock priority increases with each consecutive abort, meaning that each attempt has a slightly better chance of success than the previous. Note that the lock priority is preserved per session (not per transaction). Lock priority is set by the first read or write in the first attempt of a read-write transaction. If the application starts a new session to retry the whole transaction, the transaction loses its original lock priority. Moreover, the lock priority is only preserved if the transaction fails with an `ABORTED` error. Under some circumstances (for example, many transactions attempting to modify the same row(s)), a transaction can abort many times in a short period before successfully committing. Thus, it is not a good idea to cap the number of retries a transaction can attempt; instead, it is better to limit the total amount of time spent retrying. Idle transactions: A transaction is considered idle if it has no outstanding reads or SQL queries and has not started a read or SQL query within the last 10 seconds. Idle transactions can be aborted by Cloud Spanner so that they don't hold on to locks indefinitely. If an idle transaction is aborted, the commit fails with error `ABORTED`. If this behavior is undesirable, periodically executing a simple SQL query in the transaction (for example, `SELECT 1`) prevents the transaction from becoming idle. Snapshot read-only transactions: Snapshot read-only transactions provides a simpler method than locking read-write transactions for doing several consistent reads. However, this type of transaction does not support writes. Snapshot transactions do not take locks. Instead, they work by choosing a Cloud Spanner timestamp, then executing all reads at that timestamp. Since they do not acquire locks, they do not block concurrent read-write transactions. Unlike locking read-write transactions, snapshot read-only transactions never abort. They can fail if the chosen read timestamp is garbage collected; however, the default garbage collection policy is generous enough that most applications do not need to worry about this in practice. Snapshot read-only transactions do not need to call Commit or Rollback (and in fact are not permitted to do so). To execute a snapshot transaction, the client specifies a timestamp bound, which tells Cloud Spanner how to choose a read timestamp. The types of timestamp bound are: - Strong (the default). - Bounded staleness. - Exact staleness. If the Cloud Spanner database to be read is geographically distributed, stale read-only transactions can execute more quickly than strong or read-write transactions, because they are able to execute far from the leader replica. Each type of timestamp bound is discussed in detail below. Strong: Strong reads are guaranteed to see the effects of all transactions that have committed before the start of the read. Furthermore, all rows yielded by a single read are consistent with each other -- if any part of the read observes a transaction, all parts of the read see the transaction. Strong reads are not repeatable: two consecutive strong read-only transactions might return inconsistent results if there are concurrent writes. If consistency across reads is required, the reads should be executed within a transaction or at an exact read timestamp. Queries on change streams (see below for more details) must also specify the strong read timestamp bound. See TransactionOptions.ReadOnly.strong. Exact staleness: These timestamp bounds execute reads at a user-specified timestamp. Reads at a timestamp are guaranteed to see a consistent prefix of the global transaction history: they observe modifications done by all transactions with a commit timestamp less than or equal to the read timestamp, and observe none of the modifications done by transactions with a larger commit timestamp. They block until all conflicting transactions that can be assigned commit timestamps <= the read timestamp have finished. The timestamp can either be expressed as an absolute Cloud Spanner commit timestamp or a staleness relative to the current time. These modes do not require a "negotiation phase" to pick a timestamp. As a result, they execute slightly faster than the equivalent boundedly stale concurrency modes. On the other hand, boundedly stale reads usually return fresher results. See TransactionOptions.ReadOnly.read_timestamp and TransactionOptions.ReadOnly.exact_staleness. Bounded staleness: Bounded staleness modes allow Cloud Spanner to pick the read timestamp, subject to a user-provided staleness bound. Cloud Spanner chooses the newest timestamp within the staleness bound that allows execution of the reads at the closest available replica without blocking. All rows yielded are consistent with each other -- if any part of the read observes a transaction, all parts of the read see the transaction. Boundedly stale reads are not repeatable: two stale reads, even if they use the same staleness bound, can execute at different timestamps and thus return inconsistent results. Boundedly stale reads execute in two phases: the first phase negotiates a timestamp among all replicas needed to serve the read. In the second phase, reads are executed at the negotiated timestamp. As a result of the two phase execution, bounded staleness reads are usually a little slower than comparable exact staleness reads. However, they are typically able to return fresher results, and are more likely to execute at the closest replica. Because the timestamp negotiation requires up-front knowledge of which rows are read, it can only be used with single-use read-only transactions. See TransactionOptions.ReadOnly.max_staleness and TransactionOptions.ReadOnly.min_read_timestamp. Old read timestamps and garbage collection: Cloud Spanner continuously garbage collects deleted and overwritten data in the background to reclaim storage space. This process is known as "version GC". By default, version GC reclaims versions after they are one hour old. Because of this, Cloud Spanner can't perform reads at read timestamps more than one hour in the past. This restriction also applies to in-progress reads and/or SQL queries whose timestamp become too old while executing. Reads and SQL queries with too-old read timestamps fail with the error `FAILED_PRECONDITION`. You can configure and extend the `VERSION_RETENTION_PERIOD` of a database up to a period as long as one week, which allows Cloud Spanner to perform reads up to one week in the past. Querying change Streams: A Change Stream is a schema object that can be configured to watch data changes on the entire database, a set of tables, or a set of columns in a database. When a change stream is created, Spanner automatically defines a corresponding SQL Table-Valued Function (TVF) that can be used to query the change records in the associated change stream using the ExecuteStreamingSql API. The name of the TVF for a change stream is generated from the name of the change stream: READ_. All queries on change stream TVFs must be executed using the ExecuteStreamingSql API with a single-use read-only transaction with a strong read-only timestamp_bound. The change stream TVF allows users to specify the start_timestamp and end_timestamp for the time range of interest. All change records within the retention period is accessible using the strong read-only timestamp_bound. All other TransactionOptions are invalid for change stream queries. In addition, if TransactionOptions.read_only.return_read_timestamp is set to true, a special value of 2^63 - 2 is returned in the Transaction message that describes the transaction, instead of a valid read timestamp. This special value should be discarded and not used for any subsequent queries. Please see https://cloud.google.com/spanner/docs/change-streams for more details on how to query the change stream TVFs. Partitioned DML transactions: Partitioned DML transactions are used to execute DML statements with a different execution strategy that provides different, and often better, scalability properties for large, table-wide operations than DML in a ReadWrite transaction. Smaller scoped statements, such as an OLTP workload, should prefer using ReadWrite transactions. Partitioned DML partitions the keyspace and runs the DML statement on each partition in separate, internal transactions. These transactions commit automatically when complete, and run independently from one another. To reduce lock contention, this execution strategy only acquires read locks on rows that match the WHERE clause of the statement. Additionally, the smaller per-partition transactions hold locks for less time. That said, Partitioned DML is not a drop-in replacement for standard DML used in ReadWrite transactions. - The DML statement must be fully-partitionable. Specifically, the statement must be expressible as the union of many statements which each access only a single row of the table. - The statement is not applied atomically to all rows of the table. Rather, the statement is applied atomically to partitions of the table, in independent transactions. Secondary index rows are updated atomically with the base table rows. - Partitioned DML does not guarantee exactly-once execution semantics against a partition. The statement is applied at least once to each partition. It is strongly recommended that the DML statement should be idempotent to avoid unexpected results. For instance, it is potentially dangerous to run a statement such as `UPDATE table SET column = column + 1` as it could be run multiple times against some rows. - The partitions are committed automatically - there is no support for Commit or Rollback. If the call returns an error, or if the client issuing the ExecuteSql call dies, it is possible that some rows had the statement executed on them successfully. It is also possible that statement was never executed against other rows. - Partitioned DML transactions may only contain the execution of a single DML statement via ExecuteSql or ExecuteStreamingSql. - If any error is encountered during the execution of the partitioned DML operation (for instance, a UNIQUE INDEX violation, division by zero, or a value that can't be stored due to schema constraints), then the operation is stopped at that point and an error is returned. It is possible that at this point, some partitions have been committed (or even committed multiple times), and other partitions have not been run at all. Given the above, Partitioned DML is good fit for large, database-wide, operations that are idempotent, such as deleting old rows from a very large table. # Execute the read or SQL query in a temporary transaction. This is the most efficient way to execute a transaction that consists of a single SQL query. + "singleUse": { # Options to use for transactions. # Execute the read or SQL query in a temporary transaction. This is the most efficient way to execute a transaction that consists of a single SQL query. "excludeTxnFromChangeStreams": True or False, # When `exclude_txn_from_change_streams` is set to `true`, it prevents read or write transactions from being tracked in change streams. * If the DDL option `allow_txn_exclusion` is set to `true`, then the updates made within this transaction aren't recorded in the change stream. * If you don't set the DDL option `allow_txn_exclusion` or if it's set to `false`, then the updates made within this transaction are recorded in the change stream. When `exclude_txn_from_change_streams` is set to `false` or not set, modifications from this transaction are recorded in all change streams that are tracking columns modified by these transactions. The `exclude_txn_from_change_streams` option can only be specified for read-write or partitioned DML transactions, otherwise the API returns an `INVALID_ARGUMENT` error. "isolationLevel": "A String", # Isolation level for the transaction. "partitionedDml": { # Message type to initiate a Partitioned DML transaction. # Partitioned DML transaction. Authorization to begin a Partitioned DML transaction requires `spanner.databases.beginPartitionedDmlTransaction` permission on the `session` resource. @@ -947,7 +947,7 @@

Method Details

"seqno": "A String", # A per-transaction sequence number used to identify this request. This field makes each request idempotent such that if the request is received multiple times, at most one succeeds. The sequence number must be monotonically increasing within the transaction. If a request arrives for the first time with an out-of-order sequence number, the transaction can be aborted. Replays of previously handled requests yield the same response as the first execution. Required for DML statements. Ignored for queries. "sql": "A String", # Required. The SQL string. "transaction": { # This message is used to select the transaction in which a Read or ExecuteSql call runs. See TransactionOptions for more information about transactions. # The transaction to use. For queries, if none is provided, the default is a temporary read-only transaction with strong concurrency. Standard DML statements require a read-write transaction. To protect against replays, single-use transactions are not supported. The caller must either supply an existing transaction ID or begin a new transaction. Partitioned DML requires an existing Partitioned DML transaction ID. - "begin": { # Transactions: Each session can have at most one active transaction at a time (note that standalone reads and queries use a transaction internally and do count towards the one transaction limit). After the active transaction is completed, the session can immediately be re-used for the next transaction. It is not necessary to create a new session for each transaction. Transaction modes: Cloud Spanner supports three transaction modes: 1. Locking read-write. This type of transaction is the only way to write data into Cloud Spanner. These transactions rely on pessimistic locking and, if necessary, two-phase commit. Locking read-write transactions may abort, requiring the application to retry. 2. Snapshot read-only. Snapshot read-only transactions provide guaranteed consistency across several reads, but do not allow writes. Snapshot read-only transactions can be configured to read at timestamps in the past, or configured to perform a strong read (where Spanner selects a timestamp such that the read is guaranteed to see the effects of all transactions that have committed before the start of the read). Snapshot read-only transactions do not need to be committed. Queries on change streams must be performed with the snapshot read-only transaction mode, specifying a strong read. See TransactionOptions.ReadOnly.strong for more details. 3. Partitioned DML. This type of transaction is used to execute a single Partitioned DML statement. Partitioned DML partitions the key space and runs the DML statement over each partition in parallel using separate, internal transactions that commit independently. Partitioned DML transactions do not need to be committed. For transactions that only read, snapshot read-only transactions provide simpler semantics and are almost always faster. In particular, read-only transactions do not take locks, so they do not conflict with read-write transactions. As a consequence of not taking locks, they also do not abort, so retry loops are not needed. Transactions may only read-write data in a single database. They may, however, read-write data in different tables within that database. Locking read-write transactions: Locking transactions may be used to atomically read-modify-write data anywhere in a database. This type of transaction is externally consistent. Clients should attempt to minimize the amount of time a transaction is active. Faster transactions commit with higher probability and cause less contention. Cloud Spanner attempts to keep read locks active as long as the transaction continues to do reads, and the transaction has not been terminated by Commit or Rollback. Long periods of inactivity at the client may cause Cloud Spanner to release a transaction's locks and abort it. Conceptually, a read-write transaction consists of zero or more reads or SQL statements followed by Commit. At any time before Commit, the client can send a Rollback request to abort the transaction. Semantics: Cloud Spanner can commit the transaction if all read locks it acquired are still valid at commit time, and it is able to acquire write locks for all writes. Cloud Spanner can abort the transaction for any reason. If a commit attempt returns `ABORTED`, Cloud Spanner guarantees that the transaction has not modified any user data in Cloud Spanner. Unless the transaction commits, Cloud Spanner makes no guarantees about how long the transaction's locks were held for. It is an error to use Cloud Spanner locks for any sort of mutual exclusion other than between Cloud Spanner transactions themselves. Retrying aborted transactions: When a transaction aborts, the application can choose to retry the whole transaction again. To maximize the chances of successfully committing the retry, the client should execute the retry in the same session as the original attempt. The original session's lock priority increases with each consecutive abort, meaning that each attempt has a slightly better chance of success than the previous. Note that the lock priority is preserved per session (not per transaction). Lock priority is set by the first read or write in the first attempt of a read-write transaction. If the application starts a new session to retry the whole transaction, the transaction loses its original lock priority. Moreover, the lock priority is only preserved if the transaction fails with an `ABORTED` error. Under some circumstances (for example, many transactions attempting to modify the same row(s)), a transaction can abort many times in a short period before successfully committing. Thus, it is not a good idea to cap the number of retries a transaction can attempt; instead, it is better to limit the total amount of time spent retrying. Idle transactions: A transaction is considered idle if it has no outstanding reads or SQL queries and has not started a read or SQL query within the last 10 seconds. Idle transactions can be aborted by Cloud Spanner so that they don't hold on to locks indefinitely. If an idle transaction is aborted, the commit fails with error `ABORTED`. If this behavior is undesirable, periodically executing a simple SQL query in the transaction (for example, `SELECT 1`) prevents the transaction from becoming idle. Snapshot read-only transactions: Snapshot read-only transactions provides a simpler method than locking read-write transactions for doing several consistent reads. However, this type of transaction does not support writes. Snapshot transactions do not take locks. Instead, they work by choosing a Cloud Spanner timestamp, then executing all reads at that timestamp. Since they do not acquire locks, they do not block concurrent read-write transactions. Unlike locking read-write transactions, snapshot read-only transactions never abort. They can fail if the chosen read timestamp is garbage collected; however, the default garbage collection policy is generous enough that most applications do not need to worry about this in practice. Snapshot read-only transactions do not need to call Commit or Rollback (and in fact are not permitted to do so). To execute a snapshot transaction, the client specifies a timestamp bound, which tells Cloud Spanner how to choose a read timestamp. The types of timestamp bound are: - Strong (the default). - Bounded staleness. - Exact staleness. If the Cloud Spanner database to be read is geographically distributed, stale read-only transactions can execute more quickly than strong or read-write transactions, because they are able to execute far from the leader replica. Each type of timestamp bound is discussed in detail below. Strong: Strong reads are guaranteed to see the effects of all transactions that have committed before the start of the read. Furthermore, all rows yielded by a single read are consistent with each other -- if any part of the read observes a transaction, all parts of the read see the transaction. Strong reads are not repeatable: two consecutive strong read-only transactions might return inconsistent results if there are concurrent writes. If consistency across reads is required, the reads should be executed within a transaction or at an exact read timestamp. Queries on change streams (see below for more details) must also specify the strong read timestamp bound. See TransactionOptions.ReadOnly.strong. Exact staleness: These timestamp bounds execute reads at a user-specified timestamp. Reads at a timestamp are guaranteed to see a consistent prefix of the global transaction history: they observe modifications done by all transactions with a commit timestamp less than or equal to the read timestamp, and observe none of the modifications done by transactions with a larger commit timestamp. They block until all conflicting transactions that can be assigned commit timestamps <= the read timestamp have finished. The timestamp can either be expressed as an absolute Cloud Spanner commit timestamp or a staleness relative to the current time. These modes do not require a "negotiation phase" to pick a timestamp. As a result, they execute slightly faster than the equivalent boundedly stale concurrency modes. On the other hand, boundedly stale reads usually return fresher results. See TransactionOptions.ReadOnly.read_timestamp and TransactionOptions.ReadOnly.exact_staleness. Bounded staleness: Bounded staleness modes allow Cloud Spanner to pick the read timestamp, subject to a user-provided staleness bound. Cloud Spanner chooses the newest timestamp within the staleness bound that allows execution of the reads at the closest available replica without blocking. All rows yielded are consistent with each other -- if any part of the read observes a transaction, all parts of the read see the transaction. Boundedly stale reads are not repeatable: two stale reads, even if they use the same staleness bound, can execute at different timestamps and thus return inconsistent results. Boundedly stale reads execute in two phases: the first phase negotiates a timestamp among all replicas needed to serve the read. In the second phase, reads are executed at the negotiated timestamp. As a result of the two phase execution, bounded staleness reads are usually a little slower than comparable exact staleness reads. However, they are typically able to return fresher results, and are more likely to execute at the closest replica. Because the timestamp negotiation requires up-front knowledge of which rows are read, it can only be used with single-use read-only transactions. See TransactionOptions.ReadOnly.max_staleness and TransactionOptions.ReadOnly.min_read_timestamp. Old read timestamps and garbage collection: Cloud Spanner continuously garbage collects deleted and overwritten data in the background to reclaim storage space. This process is known as "version GC". By default, version GC reclaims versions after they are one hour old. Because of this, Cloud Spanner can't perform reads at read timestamps more than one hour in the past. This restriction also applies to in-progress reads and/or SQL queries whose timestamp become too old while executing. Reads and SQL queries with too-old read timestamps fail with the error `FAILED_PRECONDITION`. You can configure and extend the `VERSION_RETENTION_PERIOD` of a database up to a period as long as one week, which allows Cloud Spanner to perform reads up to one week in the past. Querying change Streams: A Change Stream is a schema object that can be configured to watch data changes on the entire database, a set of tables, or a set of columns in a database. When a change stream is created, Spanner automatically defines a corresponding SQL Table-Valued Function (TVF) that can be used to query the change records in the associated change stream using the ExecuteStreamingSql API. The name of the TVF for a change stream is generated from the name of the change stream: READ_. All queries on change stream TVFs must be executed using the ExecuteStreamingSql API with a single-use read-only transaction with a strong read-only timestamp_bound. The change stream TVF allows users to specify the start_timestamp and end_timestamp for the time range of interest. All change records within the retention period is accessible using the strong read-only timestamp_bound. All other TransactionOptions are invalid for change stream queries. In addition, if TransactionOptions.read_only.return_read_timestamp is set to true, a special value of 2^63 - 2 is returned in the Transaction message that describes the transaction, instead of a valid read timestamp. This special value should be discarded and not used for any subsequent queries. Please see https://cloud.google.com/spanner/docs/change-streams for more details on how to query the change stream TVFs. Partitioned DML transactions: Partitioned DML transactions are used to execute DML statements with a different execution strategy that provides different, and often better, scalability properties for large, table-wide operations than DML in a ReadWrite transaction. Smaller scoped statements, such as an OLTP workload, should prefer using ReadWrite transactions. Partitioned DML partitions the keyspace and runs the DML statement on each partition in separate, internal transactions. These transactions commit automatically when complete, and run independently from one another. To reduce lock contention, this execution strategy only acquires read locks on rows that match the WHERE clause of the statement. Additionally, the smaller per-partition transactions hold locks for less time. That said, Partitioned DML is not a drop-in replacement for standard DML used in ReadWrite transactions. - The DML statement must be fully-partitionable. Specifically, the statement must be expressible as the union of many statements which each access only a single row of the table. - The statement is not applied atomically to all rows of the table. Rather, the statement is applied atomically to partitions of the table, in independent transactions. Secondary index rows are updated atomically with the base table rows. - Partitioned DML does not guarantee exactly-once execution semantics against a partition. The statement is applied at least once to each partition. It is strongly recommended that the DML statement should be idempotent to avoid unexpected results. For instance, it is potentially dangerous to run a statement such as `UPDATE table SET column = column + 1` as it could be run multiple times against some rows. - The partitions are committed automatically - there is no support for Commit or Rollback. If the call returns an error, or if the client issuing the ExecuteSql call dies, it is possible that some rows had the statement executed on them successfully. It is also possible that statement was never executed against other rows. - Partitioned DML transactions may only contain the execution of a single DML statement via ExecuteSql or ExecuteStreamingSql. - If any error is encountered during the execution of the partitioned DML operation (for instance, a UNIQUE INDEX violation, division by zero, or a value that can't be stored due to schema constraints), then the operation is stopped at that point and an error is returned. It is possible that at this point, some partitions have been committed (or even committed multiple times), and other partitions have not been run at all. Given the above, Partitioned DML is good fit for large, database-wide, operations that are idempotent, such as deleting old rows from a very large table. # Begin a new transaction and execute this read or SQL query in it. The transaction ID of the new transaction is returned in ResultSetMetadata.transaction, which is a Transaction. + "begin": { # Options to use for transactions. # Begin a new transaction and execute this read or SQL query in it. The transaction ID of the new transaction is returned in ResultSetMetadata.transaction, which is a Transaction. "excludeTxnFromChangeStreams": True or False, # When `exclude_txn_from_change_streams` is set to `true`, it prevents read or write transactions from being tracked in change streams. * If the DDL option `allow_txn_exclusion` is set to `true`, then the updates made within this transaction aren't recorded in the change stream. * If you don't set the DDL option `allow_txn_exclusion` or if it's set to `false`, then the updates made within this transaction are recorded in the change stream. When `exclude_txn_from_change_streams` is set to `false` or not set, modifications from this transaction are recorded in all change streams that are tracking columns modified by these transactions. The `exclude_txn_from_change_streams` option can only be specified for read-write or partitioned DML transactions, otherwise the API returns an `INVALID_ARGUMENT` error. "isolationLevel": "A String", # Isolation level for the transaction. "partitionedDml": { # Message type to initiate a Partitioned DML transaction. # Partitioned DML transaction. Authorization to begin a Partitioned DML transaction requires `spanner.databases.beginPartitionedDmlTransaction` permission on the `session` resource. @@ -966,7 +966,7 @@

Method Details

}, }, "id": "A String", # Execute the read or SQL query in a previously-started transaction. - "singleUse": { # Transactions: Each session can have at most one active transaction at a time (note that standalone reads and queries use a transaction internally and do count towards the one transaction limit). After the active transaction is completed, the session can immediately be re-used for the next transaction. It is not necessary to create a new session for each transaction. Transaction modes: Cloud Spanner supports three transaction modes: 1. Locking read-write. This type of transaction is the only way to write data into Cloud Spanner. These transactions rely on pessimistic locking and, if necessary, two-phase commit. Locking read-write transactions may abort, requiring the application to retry. 2. Snapshot read-only. Snapshot read-only transactions provide guaranteed consistency across several reads, but do not allow writes. Snapshot read-only transactions can be configured to read at timestamps in the past, or configured to perform a strong read (where Spanner selects a timestamp such that the read is guaranteed to see the effects of all transactions that have committed before the start of the read). Snapshot read-only transactions do not need to be committed. Queries on change streams must be performed with the snapshot read-only transaction mode, specifying a strong read. See TransactionOptions.ReadOnly.strong for more details. 3. Partitioned DML. This type of transaction is used to execute a single Partitioned DML statement. Partitioned DML partitions the key space and runs the DML statement over each partition in parallel using separate, internal transactions that commit independently. Partitioned DML transactions do not need to be committed. For transactions that only read, snapshot read-only transactions provide simpler semantics and are almost always faster. In particular, read-only transactions do not take locks, so they do not conflict with read-write transactions. As a consequence of not taking locks, they also do not abort, so retry loops are not needed. Transactions may only read-write data in a single database. They may, however, read-write data in different tables within that database. Locking read-write transactions: Locking transactions may be used to atomically read-modify-write data anywhere in a database. This type of transaction is externally consistent. Clients should attempt to minimize the amount of time a transaction is active. Faster transactions commit with higher probability and cause less contention. Cloud Spanner attempts to keep read locks active as long as the transaction continues to do reads, and the transaction has not been terminated by Commit or Rollback. Long periods of inactivity at the client may cause Cloud Spanner to release a transaction's locks and abort it. Conceptually, a read-write transaction consists of zero or more reads or SQL statements followed by Commit. At any time before Commit, the client can send a Rollback request to abort the transaction. Semantics: Cloud Spanner can commit the transaction if all read locks it acquired are still valid at commit time, and it is able to acquire write locks for all writes. Cloud Spanner can abort the transaction for any reason. If a commit attempt returns `ABORTED`, Cloud Spanner guarantees that the transaction has not modified any user data in Cloud Spanner. Unless the transaction commits, Cloud Spanner makes no guarantees about how long the transaction's locks were held for. It is an error to use Cloud Spanner locks for any sort of mutual exclusion other than between Cloud Spanner transactions themselves. Retrying aborted transactions: When a transaction aborts, the application can choose to retry the whole transaction again. To maximize the chances of successfully committing the retry, the client should execute the retry in the same session as the original attempt. The original session's lock priority increases with each consecutive abort, meaning that each attempt has a slightly better chance of success than the previous. Note that the lock priority is preserved per session (not per transaction). Lock priority is set by the first read or write in the first attempt of a read-write transaction. If the application starts a new session to retry the whole transaction, the transaction loses its original lock priority. Moreover, the lock priority is only preserved if the transaction fails with an `ABORTED` error. Under some circumstances (for example, many transactions attempting to modify the same row(s)), a transaction can abort many times in a short period before successfully committing. Thus, it is not a good idea to cap the number of retries a transaction can attempt; instead, it is better to limit the total amount of time spent retrying. Idle transactions: A transaction is considered idle if it has no outstanding reads or SQL queries and has not started a read or SQL query within the last 10 seconds. Idle transactions can be aborted by Cloud Spanner so that they don't hold on to locks indefinitely. If an idle transaction is aborted, the commit fails with error `ABORTED`. If this behavior is undesirable, periodically executing a simple SQL query in the transaction (for example, `SELECT 1`) prevents the transaction from becoming idle. Snapshot read-only transactions: Snapshot read-only transactions provides a simpler method than locking read-write transactions for doing several consistent reads. However, this type of transaction does not support writes. Snapshot transactions do not take locks. Instead, they work by choosing a Cloud Spanner timestamp, then executing all reads at that timestamp. Since they do not acquire locks, they do not block concurrent read-write transactions. Unlike locking read-write transactions, snapshot read-only transactions never abort. They can fail if the chosen read timestamp is garbage collected; however, the default garbage collection policy is generous enough that most applications do not need to worry about this in practice. Snapshot read-only transactions do not need to call Commit or Rollback (and in fact are not permitted to do so). To execute a snapshot transaction, the client specifies a timestamp bound, which tells Cloud Spanner how to choose a read timestamp. The types of timestamp bound are: - Strong (the default). - Bounded staleness. - Exact staleness. If the Cloud Spanner database to be read is geographically distributed, stale read-only transactions can execute more quickly than strong or read-write transactions, because they are able to execute far from the leader replica. Each type of timestamp bound is discussed in detail below. Strong: Strong reads are guaranteed to see the effects of all transactions that have committed before the start of the read. Furthermore, all rows yielded by a single read are consistent with each other -- if any part of the read observes a transaction, all parts of the read see the transaction. Strong reads are not repeatable: two consecutive strong read-only transactions might return inconsistent results if there are concurrent writes. If consistency across reads is required, the reads should be executed within a transaction or at an exact read timestamp. Queries on change streams (see below for more details) must also specify the strong read timestamp bound. See TransactionOptions.ReadOnly.strong. Exact staleness: These timestamp bounds execute reads at a user-specified timestamp. Reads at a timestamp are guaranteed to see a consistent prefix of the global transaction history: they observe modifications done by all transactions with a commit timestamp less than or equal to the read timestamp, and observe none of the modifications done by transactions with a larger commit timestamp. They block until all conflicting transactions that can be assigned commit timestamps <= the read timestamp have finished. The timestamp can either be expressed as an absolute Cloud Spanner commit timestamp or a staleness relative to the current time. These modes do not require a "negotiation phase" to pick a timestamp. As a result, they execute slightly faster than the equivalent boundedly stale concurrency modes. On the other hand, boundedly stale reads usually return fresher results. See TransactionOptions.ReadOnly.read_timestamp and TransactionOptions.ReadOnly.exact_staleness. Bounded staleness: Bounded staleness modes allow Cloud Spanner to pick the read timestamp, subject to a user-provided staleness bound. Cloud Spanner chooses the newest timestamp within the staleness bound that allows execution of the reads at the closest available replica without blocking. All rows yielded are consistent with each other -- if any part of the read observes a transaction, all parts of the read see the transaction. Boundedly stale reads are not repeatable: two stale reads, even if they use the same staleness bound, can execute at different timestamps and thus return inconsistent results. Boundedly stale reads execute in two phases: the first phase negotiates a timestamp among all replicas needed to serve the read. In the second phase, reads are executed at the negotiated timestamp. As a result of the two phase execution, bounded staleness reads are usually a little slower than comparable exact staleness reads. However, they are typically able to return fresher results, and are more likely to execute at the closest replica. Because the timestamp negotiation requires up-front knowledge of which rows are read, it can only be used with single-use read-only transactions. See TransactionOptions.ReadOnly.max_staleness and TransactionOptions.ReadOnly.min_read_timestamp. Old read timestamps and garbage collection: Cloud Spanner continuously garbage collects deleted and overwritten data in the background to reclaim storage space. This process is known as "version GC". By default, version GC reclaims versions after they are one hour old. Because of this, Cloud Spanner can't perform reads at read timestamps more than one hour in the past. This restriction also applies to in-progress reads and/or SQL queries whose timestamp become too old while executing. Reads and SQL queries with too-old read timestamps fail with the error `FAILED_PRECONDITION`. You can configure and extend the `VERSION_RETENTION_PERIOD` of a database up to a period as long as one week, which allows Cloud Spanner to perform reads up to one week in the past. Querying change Streams: A Change Stream is a schema object that can be configured to watch data changes on the entire database, a set of tables, or a set of columns in a database. When a change stream is created, Spanner automatically defines a corresponding SQL Table-Valued Function (TVF) that can be used to query the change records in the associated change stream using the ExecuteStreamingSql API. The name of the TVF for a change stream is generated from the name of the change stream: READ_. All queries on change stream TVFs must be executed using the ExecuteStreamingSql API with a single-use read-only transaction with a strong read-only timestamp_bound. The change stream TVF allows users to specify the start_timestamp and end_timestamp for the time range of interest. All change records within the retention period is accessible using the strong read-only timestamp_bound. All other TransactionOptions are invalid for change stream queries. In addition, if TransactionOptions.read_only.return_read_timestamp is set to true, a special value of 2^63 - 2 is returned in the Transaction message that describes the transaction, instead of a valid read timestamp. This special value should be discarded and not used for any subsequent queries. Please see https://cloud.google.com/spanner/docs/change-streams for more details on how to query the change stream TVFs. Partitioned DML transactions: Partitioned DML transactions are used to execute DML statements with a different execution strategy that provides different, and often better, scalability properties for large, table-wide operations than DML in a ReadWrite transaction. Smaller scoped statements, such as an OLTP workload, should prefer using ReadWrite transactions. Partitioned DML partitions the keyspace and runs the DML statement on each partition in separate, internal transactions. These transactions commit automatically when complete, and run independently from one another. To reduce lock contention, this execution strategy only acquires read locks on rows that match the WHERE clause of the statement. Additionally, the smaller per-partition transactions hold locks for less time. That said, Partitioned DML is not a drop-in replacement for standard DML used in ReadWrite transactions. - The DML statement must be fully-partitionable. Specifically, the statement must be expressible as the union of many statements which each access only a single row of the table. - The statement is not applied atomically to all rows of the table. Rather, the statement is applied atomically to partitions of the table, in independent transactions. Secondary index rows are updated atomically with the base table rows. - Partitioned DML does not guarantee exactly-once execution semantics against a partition. The statement is applied at least once to each partition. It is strongly recommended that the DML statement should be idempotent to avoid unexpected results. For instance, it is potentially dangerous to run a statement such as `UPDATE table SET column = column + 1` as it could be run multiple times against some rows. - The partitions are committed automatically - there is no support for Commit or Rollback. If the call returns an error, or if the client issuing the ExecuteSql call dies, it is possible that some rows had the statement executed on them successfully. It is also possible that statement was never executed against other rows. - Partitioned DML transactions may only contain the execution of a single DML statement via ExecuteSql or ExecuteStreamingSql. - If any error is encountered during the execution of the partitioned DML operation (for instance, a UNIQUE INDEX violation, division by zero, or a value that can't be stored due to schema constraints), then the operation is stopped at that point and an error is returned. It is possible that at this point, some partitions have been committed (or even committed multiple times), and other partitions have not been run at all. Given the above, Partitioned DML is good fit for large, database-wide, operations that are idempotent, such as deleting old rows from a very large table. # Execute the read or SQL query in a temporary transaction. This is the most efficient way to execute a transaction that consists of a single SQL query. + "singleUse": { # Options to use for transactions. # Execute the read or SQL query in a temporary transaction. This is the most efficient way to execute a transaction that consists of a single SQL query. "excludeTxnFromChangeStreams": True or False, # When `exclude_txn_from_change_streams` is set to `true`, it prevents read or write transactions from being tracked in change streams. * If the DDL option `allow_txn_exclusion` is set to `true`, then the updates made within this transaction aren't recorded in the change stream. * If you don't set the DDL option `allow_txn_exclusion` or if it's set to `false`, then the updates made within this transaction are recorded in the change stream. When `exclude_txn_from_change_streams` is set to `false` or not set, modifications from this transaction are recorded in all change streams that are tracking columns modified by these transactions. The `exclude_txn_from_change_streams` option can only be specified for read-write or partitioned DML transactions, otherwise the API returns an `INVALID_ARGUMENT` error. "isolationLevel": "A String", # Isolation level for the transaction. "partitionedDml": { # Message type to initiate a Partitioned DML transaction. # Partitioned DML transaction. Authorization to begin a Partitioned DML transaction requires `spanner.databases.beginPartitionedDmlTransaction` permission on the `session` resource. @@ -1144,7 +1144,7 @@

Method Details

"seqno": "A String", # A per-transaction sequence number used to identify this request. This field makes each request idempotent such that if the request is received multiple times, at most one succeeds. The sequence number must be monotonically increasing within the transaction. If a request arrives for the first time with an out-of-order sequence number, the transaction can be aborted. Replays of previously handled requests yield the same response as the first execution. Required for DML statements. Ignored for queries. "sql": "A String", # Required. The SQL string. "transaction": { # This message is used to select the transaction in which a Read or ExecuteSql call runs. See TransactionOptions for more information about transactions. # The transaction to use. For queries, if none is provided, the default is a temporary read-only transaction with strong concurrency. Standard DML statements require a read-write transaction. To protect against replays, single-use transactions are not supported. The caller must either supply an existing transaction ID or begin a new transaction. Partitioned DML requires an existing Partitioned DML transaction ID. - "begin": { # Transactions: Each session can have at most one active transaction at a time (note that standalone reads and queries use a transaction internally and do count towards the one transaction limit). After the active transaction is completed, the session can immediately be re-used for the next transaction. It is not necessary to create a new session for each transaction. Transaction modes: Cloud Spanner supports three transaction modes: 1. Locking read-write. This type of transaction is the only way to write data into Cloud Spanner. These transactions rely on pessimistic locking and, if necessary, two-phase commit. Locking read-write transactions may abort, requiring the application to retry. 2. Snapshot read-only. Snapshot read-only transactions provide guaranteed consistency across several reads, but do not allow writes. Snapshot read-only transactions can be configured to read at timestamps in the past, or configured to perform a strong read (where Spanner selects a timestamp such that the read is guaranteed to see the effects of all transactions that have committed before the start of the read). Snapshot read-only transactions do not need to be committed. Queries on change streams must be performed with the snapshot read-only transaction mode, specifying a strong read. See TransactionOptions.ReadOnly.strong for more details. 3. Partitioned DML. This type of transaction is used to execute a single Partitioned DML statement. Partitioned DML partitions the key space and runs the DML statement over each partition in parallel using separate, internal transactions that commit independently. Partitioned DML transactions do not need to be committed. For transactions that only read, snapshot read-only transactions provide simpler semantics and are almost always faster. In particular, read-only transactions do not take locks, so they do not conflict with read-write transactions. As a consequence of not taking locks, they also do not abort, so retry loops are not needed. Transactions may only read-write data in a single database. They may, however, read-write data in different tables within that database. Locking read-write transactions: Locking transactions may be used to atomically read-modify-write data anywhere in a database. This type of transaction is externally consistent. Clients should attempt to minimize the amount of time a transaction is active. Faster transactions commit with higher probability and cause less contention. Cloud Spanner attempts to keep read locks active as long as the transaction continues to do reads, and the transaction has not been terminated by Commit or Rollback. Long periods of inactivity at the client may cause Cloud Spanner to release a transaction's locks and abort it. Conceptually, a read-write transaction consists of zero or more reads or SQL statements followed by Commit. At any time before Commit, the client can send a Rollback request to abort the transaction. Semantics: Cloud Spanner can commit the transaction if all read locks it acquired are still valid at commit time, and it is able to acquire write locks for all writes. Cloud Spanner can abort the transaction for any reason. If a commit attempt returns `ABORTED`, Cloud Spanner guarantees that the transaction has not modified any user data in Cloud Spanner. Unless the transaction commits, Cloud Spanner makes no guarantees about how long the transaction's locks were held for. It is an error to use Cloud Spanner locks for any sort of mutual exclusion other than between Cloud Spanner transactions themselves. Retrying aborted transactions: When a transaction aborts, the application can choose to retry the whole transaction again. To maximize the chances of successfully committing the retry, the client should execute the retry in the same session as the original attempt. The original session's lock priority increases with each consecutive abort, meaning that each attempt has a slightly better chance of success than the previous. Note that the lock priority is preserved per session (not per transaction). Lock priority is set by the first read or write in the first attempt of a read-write transaction. If the application starts a new session to retry the whole transaction, the transaction loses its original lock priority. Moreover, the lock priority is only preserved if the transaction fails with an `ABORTED` error. Under some circumstances (for example, many transactions attempting to modify the same row(s)), a transaction can abort many times in a short period before successfully committing. Thus, it is not a good idea to cap the number of retries a transaction can attempt; instead, it is better to limit the total amount of time spent retrying. Idle transactions: A transaction is considered idle if it has no outstanding reads or SQL queries and has not started a read or SQL query within the last 10 seconds. Idle transactions can be aborted by Cloud Spanner so that they don't hold on to locks indefinitely. If an idle transaction is aborted, the commit fails with error `ABORTED`. If this behavior is undesirable, periodically executing a simple SQL query in the transaction (for example, `SELECT 1`) prevents the transaction from becoming idle. Snapshot read-only transactions: Snapshot read-only transactions provides a simpler method than locking read-write transactions for doing several consistent reads. However, this type of transaction does not support writes. Snapshot transactions do not take locks. Instead, they work by choosing a Cloud Spanner timestamp, then executing all reads at that timestamp. Since they do not acquire locks, they do not block concurrent read-write transactions. Unlike locking read-write transactions, snapshot read-only transactions never abort. They can fail if the chosen read timestamp is garbage collected; however, the default garbage collection policy is generous enough that most applications do not need to worry about this in practice. Snapshot read-only transactions do not need to call Commit or Rollback (and in fact are not permitted to do so). To execute a snapshot transaction, the client specifies a timestamp bound, which tells Cloud Spanner how to choose a read timestamp. The types of timestamp bound are: - Strong (the default). - Bounded staleness. - Exact staleness. If the Cloud Spanner database to be read is geographically distributed, stale read-only transactions can execute more quickly than strong or read-write transactions, because they are able to execute far from the leader replica. Each type of timestamp bound is discussed in detail below. Strong: Strong reads are guaranteed to see the effects of all transactions that have committed before the start of the read. Furthermore, all rows yielded by a single read are consistent with each other -- if any part of the read observes a transaction, all parts of the read see the transaction. Strong reads are not repeatable: two consecutive strong read-only transactions might return inconsistent results if there are concurrent writes. If consistency across reads is required, the reads should be executed within a transaction or at an exact read timestamp. Queries on change streams (see below for more details) must also specify the strong read timestamp bound. See TransactionOptions.ReadOnly.strong. Exact staleness: These timestamp bounds execute reads at a user-specified timestamp. Reads at a timestamp are guaranteed to see a consistent prefix of the global transaction history: they observe modifications done by all transactions with a commit timestamp less than or equal to the read timestamp, and observe none of the modifications done by transactions with a larger commit timestamp. They block until all conflicting transactions that can be assigned commit timestamps <= the read timestamp have finished. The timestamp can either be expressed as an absolute Cloud Spanner commit timestamp or a staleness relative to the current time. These modes do not require a "negotiation phase" to pick a timestamp. As a result, they execute slightly faster than the equivalent boundedly stale concurrency modes. On the other hand, boundedly stale reads usually return fresher results. See TransactionOptions.ReadOnly.read_timestamp and TransactionOptions.ReadOnly.exact_staleness. Bounded staleness: Bounded staleness modes allow Cloud Spanner to pick the read timestamp, subject to a user-provided staleness bound. Cloud Spanner chooses the newest timestamp within the staleness bound that allows execution of the reads at the closest available replica without blocking. All rows yielded are consistent with each other -- if any part of the read observes a transaction, all parts of the read see the transaction. Boundedly stale reads are not repeatable: two stale reads, even if they use the same staleness bound, can execute at different timestamps and thus return inconsistent results. Boundedly stale reads execute in two phases: the first phase negotiates a timestamp among all replicas needed to serve the read. In the second phase, reads are executed at the negotiated timestamp. As a result of the two phase execution, bounded staleness reads are usually a little slower than comparable exact staleness reads. However, they are typically able to return fresher results, and are more likely to execute at the closest replica. Because the timestamp negotiation requires up-front knowledge of which rows are read, it can only be used with single-use read-only transactions. See TransactionOptions.ReadOnly.max_staleness and TransactionOptions.ReadOnly.min_read_timestamp. Old read timestamps and garbage collection: Cloud Spanner continuously garbage collects deleted and overwritten data in the background to reclaim storage space. This process is known as "version GC". By default, version GC reclaims versions after they are one hour old. Because of this, Cloud Spanner can't perform reads at read timestamps more than one hour in the past. This restriction also applies to in-progress reads and/or SQL queries whose timestamp become too old while executing. Reads and SQL queries with too-old read timestamps fail with the error `FAILED_PRECONDITION`. You can configure and extend the `VERSION_RETENTION_PERIOD` of a database up to a period as long as one week, which allows Cloud Spanner to perform reads up to one week in the past. Querying change Streams: A Change Stream is a schema object that can be configured to watch data changes on the entire database, a set of tables, or a set of columns in a database. When a change stream is created, Spanner automatically defines a corresponding SQL Table-Valued Function (TVF) that can be used to query the change records in the associated change stream using the ExecuteStreamingSql API. The name of the TVF for a change stream is generated from the name of the change stream: READ_. All queries on change stream TVFs must be executed using the ExecuteStreamingSql API with a single-use read-only transaction with a strong read-only timestamp_bound. The change stream TVF allows users to specify the start_timestamp and end_timestamp for the time range of interest. All change records within the retention period is accessible using the strong read-only timestamp_bound. All other TransactionOptions are invalid for change stream queries. In addition, if TransactionOptions.read_only.return_read_timestamp is set to true, a special value of 2^63 - 2 is returned in the Transaction message that describes the transaction, instead of a valid read timestamp. This special value should be discarded and not used for any subsequent queries. Please see https://cloud.google.com/spanner/docs/change-streams for more details on how to query the change stream TVFs. Partitioned DML transactions: Partitioned DML transactions are used to execute DML statements with a different execution strategy that provides different, and often better, scalability properties for large, table-wide operations than DML in a ReadWrite transaction. Smaller scoped statements, such as an OLTP workload, should prefer using ReadWrite transactions. Partitioned DML partitions the keyspace and runs the DML statement on each partition in separate, internal transactions. These transactions commit automatically when complete, and run independently from one another. To reduce lock contention, this execution strategy only acquires read locks on rows that match the WHERE clause of the statement. Additionally, the smaller per-partition transactions hold locks for less time. That said, Partitioned DML is not a drop-in replacement for standard DML used in ReadWrite transactions. - The DML statement must be fully-partitionable. Specifically, the statement must be expressible as the union of many statements which each access only a single row of the table. - The statement is not applied atomically to all rows of the table. Rather, the statement is applied atomically to partitions of the table, in independent transactions. Secondary index rows are updated atomically with the base table rows. - Partitioned DML does not guarantee exactly-once execution semantics against a partition. The statement is applied at least once to each partition. It is strongly recommended that the DML statement should be idempotent to avoid unexpected results. For instance, it is potentially dangerous to run a statement such as `UPDATE table SET column = column + 1` as it could be run multiple times against some rows. - The partitions are committed automatically - there is no support for Commit or Rollback. If the call returns an error, or if the client issuing the ExecuteSql call dies, it is possible that some rows had the statement executed on them successfully. It is also possible that statement was never executed against other rows. - Partitioned DML transactions may only contain the execution of a single DML statement via ExecuteSql or ExecuteStreamingSql. - If any error is encountered during the execution of the partitioned DML operation (for instance, a UNIQUE INDEX violation, division by zero, or a value that can't be stored due to schema constraints), then the operation is stopped at that point and an error is returned. It is possible that at this point, some partitions have been committed (or even committed multiple times), and other partitions have not been run at all. Given the above, Partitioned DML is good fit for large, database-wide, operations that are idempotent, such as deleting old rows from a very large table. # Begin a new transaction and execute this read or SQL query in it. The transaction ID of the new transaction is returned in ResultSetMetadata.transaction, which is a Transaction. + "begin": { # Options to use for transactions. # Begin a new transaction and execute this read or SQL query in it. The transaction ID of the new transaction is returned in ResultSetMetadata.transaction, which is a Transaction. "excludeTxnFromChangeStreams": True or False, # When `exclude_txn_from_change_streams` is set to `true`, it prevents read or write transactions from being tracked in change streams. * If the DDL option `allow_txn_exclusion` is set to `true`, then the updates made within this transaction aren't recorded in the change stream. * If you don't set the DDL option `allow_txn_exclusion` or if it's set to `false`, then the updates made within this transaction are recorded in the change stream. When `exclude_txn_from_change_streams` is set to `false` or not set, modifications from this transaction are recorded in all change streams that are tracking columns modified by these transactions. The `exclude_txn_from_change_streams` option can only be specified for read-write or partitioned DML transactions, otherwise the API returns an `INVALID_ARGUMENT` error. "isolationLevel": "A String", # Isolation level for the transaction. "partitionedDml": { # Message type to initiate a Partitioned DML transaction. # Partitioned DML transaction. Authorization to begin a Partitioned DML transaction requires `spanner.databases.beginPartitionedDmlTransaction` permission on the `session` resource. @@ -1163,7 +1163,7 @@

Method Details

}, }, "id": "A String", # Execute the read or SQL query in a previously-started transaction. - "singleUse": { # Transactions: Each session can have at most one active transaction at a time (note that standalone reads and queries use a transaction internally and do count towards the one transaction limit). After the active transaction is completed, the session can immediately be re-used for the next transaction. It is not necessary to create a new session for each transaction. Transaction modes: Cloud Spanner supports three transaction modes: 1. Locking read-write. This type of transaction is the only way to write data into Cloud Spanner. These transactions rely on pessimistic locking and, if necessary, two-phase commit. Locking read-write transactions may abort, requiring the application to retry. 2. Snapshot read-only. Snapshot read-only transactions provide guaranteed consistency across several reads, but do not allow writes. Snapshot read-only transactions can be configured to read at timestamps in the past, or configured to perform a strong read (where Spanner selects a timestamp such that the read is guaranteed to see the effects of all transactions that have committed before the start of the read). Snapshot read-only transactions do not need to be committed. Queries on change streams must be performed with the snapshot read-only transaction mode, specifying a strong read. See TransactionOptions.ReadOnly.strong for more details. 3. Partitioned DML. This type of transaction is used to execute a single Partitioned DML statement. Partitioned DML partitions the key space and runs the DML statement over each partition in parallel using separate, internal transactions that commit independently. Partitioned DML transactions do not need to be committed. For transactions that only read, snapshot read-only transactions provide simpler semantics and are almost always faster. In particular, read-only transactions do not take locks, so they do not conflict with read-write transactions. As a consequence of not taking locks, they also do not abort, so retry loops are not needed. Transactions may only read-write data in a single database. They may, however, read-write data in different tables within that database. Locking read-write transactions: Locking transactions may be used to atomically read-modify-write data anywhere in a database. This type of transaction is externally consistent. Clients should attempt to minimize the amount of time a transaction is active. Faster transactions commit with higher probability and cause less contention. Cloud Spanner attempts to keep read locks active as long as the transaction continues to do reads, and the transaction has not been terminated by Commit or Rollback. Long periods of inactivity at the client may cause Cloud Spanner to release a transaction's locks and abort it. Conceptually, a read-write transaction consists of zero or more reads or SQL statements followed by Commit. At any time before Commit, the client can send a Rollback request to abort the transaction. Semantics: Cloud Spanner can commit the transaction if all read locks it acquired are still valid at commit time, and it is able to acquire write locks for all writes. Cloud Spanner can abort the transaction for any reason. If a commit attempt returns `ABORTED`, Cloud Spanner guarantees that the transaction has not modified any user data in Cloud Spanner. Unless the transaction commits, Cloud Spanner makes no guarantees about how long the transaction's locks were held for. It is an error to use Cloud Spanner locks for any sort of mutual exclusion other than between Cloud Spanner transactions themselves. Retrying aborted transactions: When a transaction aborts, the application can choose to retry the whole transaction again. To maximize the chances of successfully committing the retry, the client should execute the retry in the same session as the original attempt. The original session's lock priority increases with each consecutive abort, meaning that each attempt has a slightly better chance of success than the previous. Note that the lock priority is preserved per session (not per transaction). Lock priority is set by the first read or write in the first attempt of a read-write transaction. If the application starts a new session to retry the whole transaction, the transaction loses its original lock priority. Moreover, the lock priority is only preserved if the transaction fails with an `ABORTED` error. Under some circumstances (for example, many transactions attempting to modify the same row(s)), a transaction can abort many times in a short period before successfully committing. Thus, it is not a good idea to cap the number of retries a transaction can attempt; instead, it is better to limit the total amount of time spent retrying. Idle transactions: A transaction is considered idle if it has no outstanding reads or SQL queries and has not started a read or SQL query within the last 10 seconds. Idle transactions can be aborted by Cloud Spanner so that they don't hold on to locks indefinitely. If an idle transaction is aborted, the commit fails with error `ABORTED`. If this behavior is undesirable, periodically executing a simple SQL query in the transaction (for example, `SELECT 1`) prevents the transaction from becoming idle. Snapshot read-only transactions: Snapshot read-only transactions provides a simpler method than locking read-write transactions for doing several consistent reads. However, this type of transaction does not support writes. Snapshot transactions do not take locks. Instead, they work by choosing a Cloud Spanner timestamp, then executing all reads at that timestamp. Since they do not acquire locks, they do not block concurrent read-write transactions. Unlike locking read-write transactions, snapshot read-only transactions never abort. They can fail if the chosen read timestamp is garbage collected; however, the default garbage collection policy is generous enough that most applications do not need to worry about this in practice. Snapshot read-only transactions do not need to call Commit or Rollback (and in fact are not permitted to do so). To execute a snapshot transaction, the client specifies a timestamp bound, which tells Cloud Spanner how to choose a read timestamp. The types of timestamp bound are: - Strong (the default). - Bounded staleness. - Exact staleness. If the Cloud Spanner database to be read is geographically distributed, stale read-only transactions can execute more quickly than strong or read-write transactions, because they are able to execute far from the leader replica. Each type of timestamp bound is discussed in detail below. Strong: Strong reads are guaranteed to see the effects of all transactions that have committed before the start of the read. Furthermore, all rows yielded by a single read are consistent with each other -- if any part of the read observes a transaction, all parts of the read see the transaction. Strong reads are not repeatable: two consecutive strong read-only transactions might return inconsistent results if there are concurrent writes. If consistency across reads is required, the reads should be executed within a transaction or at an exact read timestamp. Queries on change streams (see below for more details) must also specify the strong read timestamp bound. See TransactionOptions.ReadOnly.strong. Exact staleness: These timestamp bounds execute reads at a user-specified timestamp. Reads at a timestamp are guaranteed to see a consistent prefix of the global transaction history: they observe modifications done by all transactions with a commit timestamp less than or equal to the read timestamp, and observe none of the modifications done by transactions with a larger commit timestamp. They block until all conflicting transactions that can be assigned commit timestamps <= the read timestamp have finished. The timestamp can either be expressed as an absolute Cloud Spanner commit timestamp or a staleness relative to the current time. These modes do not require a "negotiation phase" to pick a timestamp. As a result, they execute slightly faster than the equivalent boundedly stale concurrency modes. On the other hand, boundedly stale reads usually return fresher results. See TransactionOptions.ReadOnly.read_timestamp and TransactionOptions.ReadOnly.exact_staleness. Bounded staleness: Bounded staleness modes allow Cloud Spanner to pick the read timestamp, subject to a user-provided staleness bound. Cloud Spanner chooses the newest timestamp within the staleness bound that allows execution of the reads at the closest available replica without blocking. All rows yielded are consistent with each other -- if any part of the read observes a transaction, all parts of the read see the transaction. Boundedly stale reads are not repeatable: two stale reads, even if they use the same staleness bound, can execute at different timestamps and thus return inconsistent results. Boundedly stale reads execute in two phases: the first phase negotiates a timestamp among all replicas needed to serve the read. In the second phase, reads are executed at the negotiated timestamp. As a result of the two phase execution, bounded staleness reads are usually a little slower than comparable exact staleness reads. However, they are typically able to return fresher results, and are more likely to execute at the closest replica. Because the timestamp negotiation requires up-front knowledge of which rows are read, it can only be used with single-use read-only transactions. See TransactionOptions.ReadOnly.max_staleness and TransactionOptions.ReadOnly.min_read_timestamp. Old read timestamps and garbage collection: Cloud Spanner continuously garbage collects deleted and overwritten data in the background to reclaim storage space. This process is known as "version GC". By default, version GC reclaims versions after they are one hour old. Because of this, Cloud Spanner can't perform reads at read timestamps more than one hour in the past. This restriction also applies to in-progress reads and/or SQL queries whose timestamp become too old while executing. Reads and SQL queries with too-old read timestamps fail with the error `FAILED_PRECONDITION`. You can configure and extend the `VERSION_RETENTION_PERIOD` of a database up to a period as long as one week, which allows Cloud Spanner to perform reads up to one week in the past. Querying change Streams: A Change Stream is a schema object that can be configured to watch data changes on the entire database, a set of tables, or a set of columns in a database. When a change stream is created, Spanner automatically defines a corresponding SQL Table-Valued Function (TVF) that can be used to query the change records in the associated change stream using the ExecuteStreamingSql API. The name of the TVF for a change stream is generated from the name of the change stream: READ_. All queries on change stream TVFs must be executed using the ExecuteStreamingSql API with a single-use read-only transaction with a strong read-only timestamp_bound. The change stream TVF allows users to specify the start_timestamp and end_timestamp for the time range of interest. All change records within the retention period is accessible using the strong read-only timestamp_bound. All other TransactionOptions are invalid for change stream queries. In addition, if TransactionOptions.read_only.return_read_timestamp is set to true, a special value of 2^63 - 2 is returned in the Transaction message that describes the transaction, instead of a valid read timestamp. This special value should be discarded and not used for any subsequent queries. Please see https://cloud.google.com/spanner/docs/change-streams for more details on how to query the change stream TVFs. Partitioned DML transactions: Partitioned DML transactions are used to execute DML statements with a different execution strategy that provides different, and often better, scalability properties for large, table-wide operations than DML in a ReadWrite transaction. Smaller scoped statements, such as an OLTP workload, should prefer using ReadWrite transactions. Partitioned DML partitions the keyspace and runs the DML statement on each partition in separate, internal transactions. These transactions commit automatically when complete, and run independently from one another. To reduce lock contention, this execution strategy only acquires read locks on rows that match the WHERE clause of the statement. Additionally, the smaller per-partition transactions hold locks for less time. That said, Partitioned DML is not a drop-in replacement for standard DML used in ReadWrite transactions. - The DML statement must be fully-partitionable. Specifically, the statement must be expressible as the union of many statements which each access only a single row of the table. - The statement is not applied atomically to all rows of the table. Rather, the statement is applied atomically to partitions of the table, in independent transactions. Secondary index rows are updated atomically with the base table rows. - Partitioned DML does not guarantee exactly-once execution semantics against a partition. The statement is applied at least once to each partition. It is strongly recommended that the DML statement should be idempotent to avoid unexpected results. For instance, it is potentially dangerous to run a statement such as `UPDATE table SET column = column + 1` as it could be run multiple times against some rows. - The partitions are committed automatically - there is no support for Commit or Rollback. If the call returns an error, or if the client issuing the ExecuteSql call dies, it is possible that some rows had the statement executed on them successfully. It is also possible that statement was never executed against other rows. - Partitioned DML transactions may only contain the execution of a single DML statement via ExecuteSql or ExecuteStreamingSql. - If any error is encountered during the execution of the partitioned DML operation (for instance, a UNIQUE INDEX violation, division by zero, or a value that can't be stored due to schema constraints), then the operation is stopped at that point and an error is returned. It is possible that at this point, some partitions have been committed (or even committed multiple times), and other partitions have not been run at all. Given the above, Partitioned DML is good fit for large, database-wide, operations that are idempotent, such as deleting old rows from a very large table. # Execute the read or SQL query in a temporary transaction. This is the most efficient way to execute a transaction that consists of a single SQL query. + "singleUse": { # Options to use for transactions. # Execute the read or SQL query in a temporary transaction. This is the most efficient way to execute a transaction that consists of a single SQL query. "excludeTxnFromChangeStreams": True or False, # When `exclude_txn_from_change_streams` is set to `true`, it prevents read or write transactions from being tracked in change streams. * If the DDL option `allow_txn_exclusion` is set to `true`, then the updates made within this transaction aren't recorded in the change stream. * If you don't set the DDL option `allow_txn_exclusion` or if it's set to `false`, then the updates made within this transaction are recorded in the change stream. When `exclude_txn_from_change_streams` is set to `false` or not set, modifications from this transaction are recorded in all change streams that are tracking columns modified by these transactions. The `exclude_txn_from_change_streams` option can only be specified for read-write or partitioned DML transactions, otherwise the API returns an `INVALID_ARGUMENT` error. "isolationLevel": "A String", # Isolation level for the transaction. "partitionedDml": { # Message type to initiate a Partitioned DML transaction. # Partitioned DML transaction. Authorization to begin a Partitioned DML transaction requires `spanner.databases.beginPartitionedDmlTransaction` permission on the `session` resource. @@ -1386,7 +1386,7 @@

Method Details

}, "sql": "A String", # Required. The query request to generate partitions for. The request fails if the query isn't root partitionable. For a query to be root partitionable, it needs to satisfy a few conditions. For example, if the query execution plan contains a distributed union operator, then it must be the first operator in the plan. For more information about other conditions, see [Read data in parallel](https://cloud.google.com/spanner/docs/reads#read_data_in_parallel). The query request must not contain DML commands, such as `INSERT`, `UPDATE`, or `DELETE`. Use `ExecuteStreamingSql` with a `PartitionedDml` transaction for large, partition-friendly DML operations. "transaction": { # This message is used to select the transaction in which a Read or ExecuteSql call runs. See TransactionOptions for more information about transactions. # Read-only snapshot transactions are supported, read and write and single-use transactions are not. - "begin": { # Transactions: Each session can have at most one active transaction at a time (note that standalone reads and queries use a transaction internally and do count towards the one transaction limit). After the active transaction is completed, the session can immediately be re-used for the next transaction. It is not necessary to create a new session for each transaction. Transaction modes: Cloud Spanner supports three transaction modes: 1. Locking read-write. This type of transaction is the only way to write data into Cloud Spanner. These transactions rely on pessimistic locking and, if necessary, two-phase commit. Locking read-write transactions may abort, requiring the application to retry. 2. Snapshot read-only. Snapshot read-only transactions provide guaranteed consistency across several reads, but do not allow writes. Snapshot read-only transactions can be configured to read at timestamps in the past, or configured to perform a strong read (where Spanner selects a timestamp such that the read is guaranteed to see the effects of all transactions that have committed before the start of the read). Snapshot read-only transactions do not need to be committed. Queries on change streams must be performed with the snapshot read-only transaction mode, specifying a strong read. See TransactionOptions.ReadOnly.strong for more details. 3. Partitioned DML. This type of transaction is used to execute a single Partitioned DML statement. Partitioned DML partitions the key space and runs the DML statement over each partition in parallel using separate, internal transactions that commit independently. Partitioned DML transactions do not need to be committed. For transactions that only read, snapshot read-only transactions provide simpler semantics and are almost always faster. In particular, read-only transactions do not take locks, so they do not conflict with read-write transactions. As a consequence of not taking locks, they also do not abort, so retry loops are not needed. Transactions may only read-write data in a single database. They may, however, read-write data in different tables within that database. Locking read-write transactions: Locking transactions may be used to atomically read-modify-write data anywhere in a database. This type of transaction is externally consistent. Clients should attempt to minimize the amount of time a transaction is active. Faster transactions commit with higher probability and cause less contention. Cloud Spanner attempts to keep read locks active as long as the transaction continues to do reads, and the transaction has not been terminated by Commit or Rollback. Long periods of inactivity at the client may cause Cloud Spanner to release a transaction's locks and abort it. Conceptually, a read-write transaction consists of zero or more reads or SQL statements followed by Commit. At any time before Commit, the client can send a Rollback request to abort the transaction. Semantics: Cloud Spanner can commit the transaction if all read locks it acquired are still valid at commit time, and it is able to acquire write locks for all writes. Cloud Spanner can abort the transaction for any reason. If a commit attempt returns `ABORTED`, Cloud Spanner guarantees that the transaction has not modified any user data in Cloud Spanner. Unless the transaction commits, Cloud Spanner makes no guarantees about how long the transaction's locks were held for. It is an error to use Cloud Spanner locks for any sort of mutual exclusion other than between Cloud Spanner transactions themselves. Retrying aborted transactions: When a transaction aborts, the application can choose to retry the whole transaction again. To maximize the chances of successfully committing the retry, the client should execute the retry in the same session as the original attempt. The original session's lock priority increases with each consecutive abort, meaning that each attempt has a slightly better chance of success than the previous. Note that the lock priority is preserved per session (not per transaction). Lock priority is set by the first read or write in the first attempt of a read-write transaction. If the application starts a new session to retry the whole transaction, the transaction loses its original lock priority. Moreover, the lock priority is only preserved if the transaction fails with an `ABORTED` error. Under some circumstances (for example, many transactions attempting to modify the same row(s)), a transaction can abort many times in a short period before successfully committing. Thus, it is not a good idea to cap the number of retries a transaction can attempt; instead, it is better to limit the total amount of time spent retrying. Idle transactions: A transaction is considered idle if it has no outstanding reads or SQL queries and has not started a read or SQL query within the last 10 seconds. Idle transactions can be aborted by Cloud Spanner so that they don't hold on to locks indefinitely. If an idle transaction is aborted, the commit fails with error `ABORTED`. If this behavior is undesirable, periodically executing a simple SQL query in the transaction (for example, `SELECT 1`) prevents the transaction from becoming idle. Snapshot read-only transactions: Snapshot read-only transactions provides a simpler method than locking read-write transactions for doing several consistent reads. However, this type of transaction does not support writes. Snapshot transactions do not take locks. Instead, they work by choosing a Cloud Spanner timestamp, then executing all reads at that timestamp. Since they do not acquire locks, they do not block concurrent read-write transactions. Unlike locking read-write transactions, snapshot read-only transactions never abort. They can fail if the chosen read timestamp is garbage collected; however, the default garbage collection policy is generous enough that most applications do not need to worry about this in practice. Snapshot read-only transactions do not need to call Commit or Rollback (and in fact are not permitted to do so). To execute a snapshot transaction, the client specifies a timestamp bound, which tells Cloud Spanner how to choose a read timestamp. The types of timestamp bound are: - Strong (the default). - Bounded staleness. - Exact staleness. If the Cloud Spanner database to be read is geographically distributed, stale read-only transactions can execute more quickly than strong or read-write transactions, because they are able to execute far from the leader replica. Each type of timestamp bound is discussed in detail below. Strong: Strong reads are guaranteed to see the effects of all transactions that have committed before the start of the read. Furthermore, all rows yielded by a single read are consistent with each other -- if any part of the read observes a transaction, all parts of the read see the transaction. Strong reads are not repeatable: two consecutive strong read-only transactions might return inconsistent results if there are concurrent writes. If consistency across reads is required, the reads should be executed within a transaction or at an exact read timestamp. Queries on change streams (see below for more details) must also specify the strong read timestamp bound. See TransactionOptions.ReadOnly.strong. Exact staleness: These timestamp bounds execute reads at a user-specified timestamp. Reads at a timestamp are guaranteed to see a consistent prefix of the global transaction history: they observe modifications done by all transactions with a commit timestamp less than or equal to the read timestamp, and observe none of the modifications done by transactions with a larger commit timestamp. They block until all conflicting transactions that can be assigned commit timestamps <= the read timestamp have finished. The timestamp can either be expressed as an absolute Cloud Spanner commit timestamp or a staleness relative to the current time. These modes do not require a "negotiation phase" to pick a timestamp. As a result, they execute slightly faster than the equivalent boundedly stale concurrency modes. On the other hand, boundedly stale reads usually return fresher results. See TransactionOptions.ReadOnly.read_timestamp and TransactionOptions.ReadOnly.exact_staleness. Bounded staleness: Bounded staleness modes allow Cloud Spanner to pick the read timestamp, subject to a user-provided staleness bound. Cloud Spanner chooses the newest timestamp within the staleness bound that allows execution of the reads at the closest available replica without blocking. All rows yielded are consistent with each other -- if any part of the read observes a transaction, all parts of the read see the transaction. Boundedly stale reads are not repeatable: two stale reads, even if they use the same staleness bound, can execute at different timestamps and thus return inconsistent results. Boundedly stale reads execute in two phases: the first phase negotiates a timestamp among all replicas needed to serve the read. In the second phase, reads are executed at the negotiated timestamp. As a result of the two phase execution, bounded staleness reads are usually a little slower than comparable exact staleness reads. However, they are typically able to return fresher results, and are more likely to execute at the closest replica. Because the timestamp negotiation requires up-front knowledge of which rows are read, it can only be used with single-use read-only transactions. See TransactionOptions.ReadOnly.max_staleness and TransactionOptions.ReadOnly.min_read_timestamp. Old read timestamps and garbage collection: Cloud Spanner continuously garbage collects deleted and overwritten data in the background to reclaim storage space. This process is known as "version GC". By default, version GC reclaims versions after they are one hour old. Because of this, Cloud Spanner can't perform reads at read timestamps more than one hour in the past. This restriction also applies to in-progress reads and/or SQL queries whose timestamp become too old while executing. Reads and SQL queries with too-old read timestamps fail with the error `FAILED_PRECONDITION`. You can configure and extend the `VERSION_RETENTION_PERIOD` of a database up to a period as long as one week, which allows Cloud Spanner to perform reads up to one week in the past. Querying change Streams: A Change Stream is a schema object that can be configured to watch data changes on the entire database, a set of tables, or a set of columns in a database. When a change stream is created, Spanner automatically defines a corresponding SQL Table-Valued Function (TVF) that can be used to query the change records in the associated change stream using the ExecuteStreamingSql API. The name of the TVF for a change stream is generated from the name of the change stream: READ_. All queries on change stream TVFs must be executed using the ExecuteStreamingSql API with a single-use read-only transaction with a strong read-only timestamp_bound. The change stream TVF allows users to specify the start_timestamp and end_timestamp for the time range of interest. All change records within the retention period is accessible using the strong read-only timestamp_bound. All other TransactionOptions are invalid for change stream queries. In addition, if TransactionOptions.read_only.return_read_timestamp is set to true, a special value of 2^63 - 2 is returned in the Transaction message that describes the transaction, instead of a valid read timestamp. This special value should be discarded and not used for any subsequent queries. Please see https://cloud.google.com/spanner/docs/change-streams for more details on how to query the change stream TVFs. Partitioned DML transactions: Partitioned DML transactions are used to execute DML statements with a different execution strategy that provides different, and often better, scalability properties for large, table-wide operations than DML in a ReadWrite transaction. Smaller scoped statements, such as an OLTP workload, should prefer using ReadWrite transactions. Partitioned DML partitions the keyspace and runs the DML statement on each partition in separate, internal transactions. These transactions commit automatically when complete, and run independently from one another. To reduce lock contention, this execution strategy only acquires read locks on rows that match the WHERE clause of the statement. Additionally, the smaller per-partition transactions hold locks for less time. That said, Partitioned DML is not a drop-in replacement for standard DML used in ReadWrite transactions. - The DML statement must be fully-partitionable. Specifically, the statement must be expressible as the union of many statements which each access only a single row of the table. - The statement is not applied atomically to all rows of the table. Rather, the statement is applied atomically to partitions of the table, in independent transactions. Secondary index rows are updated atomically with the base table rows. - Partitioned DML does not guarantee exactly-once execution semantics against a partition. The statement is applied at least once to each partition. It is strongly recommended that the DML statement should be idempotent to avoid unexpected results. For instance, it is potentially dangerous to run a statement such as `UPDATE table SET column = column + 1` as it could be run multiple times against some rows. - The partitions are committed automatically - there is no support for Commit or Rollback. If the call returns an error, or if the client issuing the ExecuteSql call dies, it is possible that some rows had the statement executed on them successfully. It is also possible that statement was never executed against other rows. - Partitioned DML transactions may only contain the execution of a single DML statement via ExecuteSql or ExecuteStreamingSql. - If any error is encountered during the execution of the partitioned DML operation (for instance, a UNIQUE INDEX violation, division by zero, or a value that can't be stored due to schema constraints), then the operation is stopped at that point and an error is returned. It is possible that at this point, some partitions have been committed (or even committed multiple times), and other partitions have not been run at all. Given the above, Partitioned DML is good fit for large, database-wide, operations that are idempotent, such as deleting old rows from a very large table. # Begin a new transaction and execute this read or SQL query in it. The transaction ID of the new transaction is returned in ResultSetMetadata.transaction, which is a Transaction. + "begin": { # Options to use for transactions. # Begin a new transaction and execute this read or SQL query in it. The transaction ID of the new transaction is returned in ResultSetMetadata.transaction, which is a Transaction. "excludeTxnFromChangeStreams": True or False, # When `exclude_txn_from_change_streams` is set to `true`, it prevents read or write transactions from being tracked in change streams. * If the DDL option `allow_txn_exclusion` is set to `true`, then the updates made within this transaction aren't recorded in the change stream. * If you don't set the DDL option `allow_txn_exclusion` or if it's set to `false`, then the updates made within this transaction are recorded in the change stream. When `exclude_txn_from_change_streams` is set to `false` or not set, modifications from this transaction are recorded in all change streams that are tracking columns modified by these transactions. The `exclude_txn_from_change_streams` option can only be specified for read-write or partitioned DML transactions, otherwise the API returns an `INVALID_ARGUMENT` error. "isolationLevel": "A String", # Isolation level for the transaction. "partitionedDml": { # Message type to initiate a Partitioned DML transaction. # Partitioned DML transaction. Authorization to begin a Partitioned DML transaction requires `spanner.databases.beginPartitionedDmlTransaction` permission on the `session` resource. @@ -1405,7 +1405,7 @@

Method Details

}, }, "id": "A String", # Execute the read or SQL query in a previously-started transaction. - "singleUse": { # Transactions: Each session can have at most one active transaction at a time (note that standalone reads and queries use a transaction internally and do count towards the one transaction limit). After the active transaction is completed, the session can immediately be re-used for the next transaction. It is not necessary to create a new session for each transaction. Transaction modes: Cloud Spanner supports three transaction modes: 1. Locking read-write. This type of transaction is the only way to write data into Cloud Spanner. These transactions rely on pessimistic locking and, if necessary, two-phase commit. Locking read-write transactions may abort, requiring the application to retry. 2. Snapshot read-only. Snapshot read-only transactions provide guaranteed consistency across several reads, but do not allow writes. Snapshot read-only transactions can be configured to read at timestamps in the past, or configured to perform a strong read (where Spanner selects a timestamp such that the read is guaranteed to see the effects of all transactions that have committed before the start of the read). Snapshot read-only transactions do not need to be committed. Queries on change streams must be performed with the snapshot read-only transaction mode, specifying a strong read. See TransactionOptions.ReadOnly.strong for more details. 3. Partitioned DML. This type of transaction is used to execute a single Partitioned DML statement. Partitioned DML partitions the key space and runs the DML statement over each partition in parallel using separate, internal transactions that commit independently. Partitioned DML transactions do not need to be committed. For transactions that only read, snapshot read-only transactions provide simpler semantics and are almost always faster. In particular, read-only transactions do not take locks, so they do not conflict with read-write transactions. As a consequence of not taking locks, they also do not abort, so retry loops are not needed. Transactions may only read-write data in a single database. They may, however, read-write data in different tables within that database. Locking read-write transactions: Locking transactions may be used to atomically read-modify-write data anywhere in a database. This type of transaction is externally consistent. Clients should attempt to minimize the amount of time a transaction is active. Faster transactions commit with higher probability and cause less contention. Cloud Spanner attempts to keep read locks active as long as the transaction continues to do reads, and the transaction has not been terminated by Commit or Rollback. Long periods of inactivity at the client may cause Cloud Spanner to release a transaction's locks and abort it. Conceptually, a read-write transaction consists of zero or more reads or SQL statements followed by Commit. At any time before Commit, the client can send a Rollback request to abort the transaction. Semantics: Cloud Spanner can commit the transaction if all read locks it acquired are still valid at commit time, and it is able to acquire write locks for all writes. Cloud Spanner can abort the transaction for any reason. If a commit attempt returns `ABORTED`, Cloud Spanner guarantees that the transaction has not modified any user data in Cloud Spanner. Unless the transaction commits, Cloud Spanner makes no guarantees about how long the transaction's locks were held for. It is an error to use Cloud Spanner locks for any sort of mutual exclusion other than between Cloud Spanner transactions themselves. Retrying aborted transactions: When a transaction aborts, the application can choose to retry the whole transaction again. To maximize the chances of successfully committing the retry, the client should execute the retry in the same session as the original attempt. The original session's lock priority increases with each consecutive abort, meaning that each attempt has a slightly better chance of success than the previous. Note that the lock priority is preserved per session (not per transaction). Lock priority is set by the first read or write in the first attempt of a read-write transaction. If the application starts a new session to retry the whole transaction, the transaction loses its original lock priority. Moreover, the lock priority is only preserved if the transaction fails with an `ABORTED` error. Under some circumstances (for example, many transactions attempting to modify the same row(s)), a transaction can abort many times in a short period before successfully committing. Thus, it is not a good idea to cap the number of retries a transaction can attempt; instead, it is better to limit the total amount of time spent retrying. Idle transactions: A transaction is considered idle if it has no outstanding reads or SQL queries and has not started a read or SQL query within the last 10 seconds. Idle transactions can be aborted by Cloud Spanner so that they don't hold on to locks indefinitely. If an idle transaction is aborted, the commit fails with error `ABORTED`. If this behavior is undesirable, periodically executing a simple SQL query in the transaction (for example, `SELECT 1`) prevents the transaction from becoming idle. Snapshot read-only transactions: Snapshot read-only transactions provides a simpler method than locking read-write transactions for doing several consistent reads. However, this type of transaction does not support writes. Snapshot transactions do not take locks. Instead, they work by choosing a Cloud Spanner timestamp, then executing all reads at that timestamp. Since they do not acquire locks, they do not block concurrent read-write transactions. Unlike locking read-write transactions, snapshot read-only transactions never abort. They can fail if the chosen read timestamp is garbage collected; however, the default garbage collection policy is generous enough that most applications do not need to worry about this in practice. Snapshot read-only transactions do not need to call Commit or Rollback (and in fact are not permitted to do so). To execute a snapshot transaction, the client specifies a timestamp bound, which tells Cloud Spanner how to choose a read timestamp. The types of timestamp bound are: - Strong (the default). - Bounded staleness. - Exact staleness. If the Cloud Spanner database to be read is geographically distributed, stale read-only transactions can execute more quickly than strong or read-write transactions, because they are able to execute far from the leader replica. Each type of timestamp bound is discussed in detail below. Strong: Strong reads are guaranteed to see the effects of all transactions that have committed before the start of the read. Furthermore, all rows yielded by a single read are consistent with each other -- if any part of the read observes a transaction, all parts of the read see the transaction. Strong reads are not repeatable: two consecutive strong read-only transactions might return inconsistent results if there are concurrent writes. If consistency across reads is required, the reads should be executed within a transaction or at an exact read timestamp. Queries on change streams (see below for more details) must also specify the strong read timestamp bound. See TransactionOptions.ReadOnly.strong. Exact staleness: These timestamp bounds execute reads at a user-specified timestamp. Reads at a timestamp are guaranteed to see a consistent prefix of the global transaction history: they observe modifications done by all transactions with a commit timestamp less than or equal to the read timestamp, and observe none of the modifications done by transactions with a larger commit timestamp. They block until all conflicting transactions that can be assigned commit timestamps <= the read timestamp have finished. The timestamp can either be expressed as an absolute Cloud Spanner commit timestamp or a staleness relative to the current time. These modes do not require a "negotiation phase" to pick a timestamp. As a result, they execute slightly faster than the equivalent boundedly stale concurrency modes. On the other hand, boundedly stale reads usually return fresher results. See TransactionOptions.ReadOnly.read_timestamp and TransactionOptions.ReadOnly.exact_staleness. Bounded staleness: Bounded staleness modes allow Cloud Spanner to pick the read timestamp, subject to a user-provided staleness bound. Cloud Spanner chooses the newest timestamp within the staleness bound that allows execution of the reads at the closest available replica without blocking. All rows yielded are consistent with each other -- if any part of the read observes a transaction, all parts of the read see the transaction. Boundedly stale reads are not repeatable: two stale reads, even if they use the same staleness bound, can execute at different timestamps and thus return inconsistent results. Boundedly stale reads execute in two phases: the first phase negotiates a timestamp among all replicas needed to serve the read. In the second phase, reads are executed at the negotiated timestamp. As a result of the two phase execution, bounded staleness reads are usually a little slower than comparable exact staleness reads. However, they are typically able to return fresher results, and are more likely to execute at the closest replica. Because the timestamp negotiation requires up-front knowledge of which rows are read, it can only be used with single-use read-only transactions. See TransactionOptions.ReadOnly.max_staleness and TransactionOptions.ReadOnly.min_read_timestamp. Old read timestamps and garbage collection: Cloud Spanner continuously garbage collects deleted and overwritten data in the background to reclaim storage space. This process is known as "version GC". By default, version GC reclaims versions after they are one hour old. Because of this, Cloud Spanner can't perform reads at read timestamps more than one hour in the past. This restriction also applies to in-progress reads and/or SQL queries whose timestamp become too old while executing. Reads and SQL queries with too-old read timestamps fail with the error `FAILED_PRECONDITION`. You can configure and extend the `VERSION_RETENTION_PERIOD` of a database up to a period as long as one week, which allows Cloud Spanner to perform reads up to one week in the past. Querying change Streams: A Change Stream is a schema object that can be configured to watch data changes on the entire database, a set of tables, or a set of columns in a database. When a change stream is created, Spanner automatically defines a corresponding SQL Table-Valued Function (TVF) that can be used to query the change records in the associated change stream using the ExecuteStreamingSql API. The name of the TVF for a change stream is generated from the name of the change stream: READ_. All queries on change stream TVFs must be executed using the ExecuteStreamingSql API with a single-use read-only transaction with a strong read-only timestamp_bound. The change stream TVF allows users to specify the start_timestamp and end_timestamp for the time range of interest. All change records within the retention period is accessible using the strong read-only timestamp_bound. All other TransactionOptions are invalid for change stream queries. In addition, if TransactionOptions.read_only.return_read_timestamp is set to true, a special value of 2^63 - 2 is returned in the Transaction message that describes the transaction, instead of a valid read timestamp. This special value should be discarded and not used for any subsequent queries. Please see https://cloud.google.com/spanner/docs/change-streams for more details on how to query the change stream TVFs. Partitioned DML transactions: Partitioned DML transactions are used to execute DML statements with a different execution strategy that provides different, and often better, scalability properties for large, table-wide operations than DML in a ReadWrite transaction. Smaller scoped statements, such as an OLTP workload, should prefer using ReadWrite transactions. Partitioned DML partitions the keyspace and runs the DML statement on each partition in separate, internal transactions. These transactions commit automatically when complete, and run independently from one another. To reduce lock contention, this execution strategy only acquires read locks on rows that match the WHERE clause of the statement. Additionally, the smaller per-partition transactions hold locks for less time. That said, Partitioned DML is not a drop-in replacement for standard DML used in ReadWrite transactions. - The DML statement must be fully-partitionable. Specifically, the statement must be expressible as the union of many statements which each access only a single row of the table. - The statement is not applied atomically to all rows of the table. Rather, the statement is applied atomically to partitions of the table, in independent transactions. Secondary index rows are updated atomically with the base table rows. - Partitioned DML does not guarantee exactly-once execution semantics against a partition. The statement is applied at least once to each partition. It is strongly recommended that the DML statement should be idempotent to avoid unexpected results. For instance, it is potentially dangerous to run a statement such as `UPDATE table SET column = column + 1` as it could be run multiple times against some rows. - The partitions are committed automatically - there is no support for Commit or Rollback. If the call returns an error, or if the client issuing the ExecuteSql call dies, it is possible that some rows had the statement executed on them successfully. It is also possible that statement was never executed against other rows. - Partitioned DML transactions may only contain the execution of a single DML statement via ExecuteSql or ExecuteStreamingSql. - If any error is encountered during the execution of the partitioned DML operation (for instance, a UNIQUE INDEX violation, division by zero, or a value that can't be stored due to schema constraints), then the operation is stopped at that point and an error is returned. It is possible that at this point, some partitions have been committed (or even committed multiple times), and other partitions have not been run at all. Given the above, Partitioned DML is good fit for large, database-wide, operations that are idempotent, such as deleting old rows from a very large table. # Execute the read or SQL query in a temporary transaction. This is the most efficient way to execute a transaction that consists of a single SQL query. + "singleUse": { # Options to use for transactions. # Execute the read or SQL query in a temporary transaction. This is the most efficient way to execute a transaction that consists of a single SQL query. "excludeTxnFromChangeStreams": True or False, # When `exclude_txn_from_change_streams` is set to `true`, it prevents read or write transactions from being tracked in change streams. * If the DDL option `allow_txn_exclusion` is set to `true`, then the updates made within this transaction aren't recorded in the change stream. * If you don't set the DDL option `allow_txn_exclusion` or if it's set to `false`, then the updates made within this transaction are recorded in the change stream. When `exclude_txn_from_change_streams` is set to `false` or not set, modifications from this transaction are recorded in all change streams that are tracking columns modified by these transactions. The `exclude_txn_from_change_streams` option can only be specified for read-write or partitioned DML transactions, otherwise the API returns an `INVALID_ARGUMENT` error. "isolationLevel": "A String", # Isolation level for the transaction. "partitionedDml": { # Message type to initiate a Partitioned DML transaction. # Partitioned DML transaction. Authorization to begin a Partitioned DML transaction requires `spanner.databases.beginPartitionedDmlTransaction` permission on the `session` resource. @@ -1495,7 +1495,7 @@

Method Details

}, "table": "A String", # Required. The name of the table in the database to be read. "transaction": { # This message is used to select the transaction in which a Read or ExecuteSql call runs. See TransactionOptions for more information about transactions. # Read only snapshot transactions are supported, read/write and single use transactions are not. - "begin": { # Transactions: Each session can have at most one active transaction at a time (note that standalone reads and queries use a transaction internally and do count towards the one transaction limit). After the active transaction is completed, the session can immediately be re-used for the next transaction. It is not necessary to create a new session for each transaction. Transaction modes: Cloud Spanner supports three transaction modes: 1. Locking read-write. This type of transaction is the only way to write data into Cloud Spanner. These transactions rely on pessimistic locking and, if necessary, two-phase commit. Locking read-write transactions may abort, requiring the application to retry. 2. Snapshot read-only. Snapshot read-only transactions provide guaranteed consistency across several reads, but do not allow writes. Snapshot read-only transactions can be configured to read at timestamps in the past, or configured to perform a strong read (where Spanner selects a timestamp such that the read is guaranteed to see the effects of all transactions that have committed before the start of the read). Snapshot read-only transactions do not need to be committed. Queries on change streams must be performed with the snapshot read-only transaction mode, specifying a strong read. See TransactionOptions.ReadOnly.strong for more details. 3. Partitioned DML. This type of transaction is used to execute a single Partitioned DML statement. Partitioned DML partitions the key space and runs the DML statement over each partition in parallel using separate, internal transactions that commit independently. Partitioned DML transactions do not need to be committed. For transactions that only read, snapshot read-only transactions provide simpler semantics and are almost always faster. In particular, read-only transactions do not take locks, so they do not conflict with read-write transactions. As a consequence of not taking locks, they also do not abort, so retry loops are not needed. Transactions may only read-write data in a single database. They may, however, read-write data in different tables within that database. Locking read-write transactions: Locking transactions may be used to atomically read-modify-write data anywhere in a database. This type of transaction is externally consistent. Clients should attempt to minimize the amount of time a transaction is active. Faster transactions commit with higher probability and cause less contention. Cloud Spanner attempts to keep read locks active as long as the transaction continues to do reads, and the transaction has not been terminated by Commit or Rollback. Long periods of inactivity at the client may cause Cloud Spanner to release a transaction's locks and abort it. Conceptually, a read-write transaction consists of zero or more reads or SQL statements followed by Commit. At any time before Commit, the client can send a Rollback request to abort the transaction. Semantics: Cloud Spanner can commit the transaction if all read locks it acquired are still valid at commit time, and it is able to acquire write locks for all writes. Cloud Spanner can abort the transaction for any reason. If a commit attempt returns `ABORTED`, Cloud Spanner guarantees that the transaction has not modified any user data in Cloud Spanner. Unless the transaction commits, Cloud Spanner makes no guarantees about how long the transaction's locks were held for. It is an error to use Cloud Spanner locks for any sort of mutual exclusion other than between Cloud Spanner transactions themselves. Retrying aborted transactions: When a transaction aborts, the application can choose to retry the whole transaction again. To maximize the chances of successfully committing the retry, the client should execute the retry in the same session as the original attempt. The original session's lock priority increases with each consecutive abort, meaning that each attempt has a slightly better chance of success than the previous. Note that the lock priority is preserved per session (not per transaction). Lock priority is set by the first read or write in the first attempt of a read-write transaction. If the application starts a new session to retry the whole transaction, the transaction loses its original lock priority. Moreover, the lock priority is only preserved if the transaction fails with an `ABORTED` error. Under some circumstances (for example, many transactions attempting to modify the same row(s)), a transaction can abort many times in a short period before successfully committing. Thus, it is not a good idea to cap the number of retries a transaction can attempt; instead, it is better to limit the total amount of time spent retrying. Idle transactions: A transaction is considered idle if it has no outstanding reads or SQL queries and has not started a read or SQL query within the last 10 seconds. Idle transactions can be aborted by Cloud Spanner so that they don't hold on to locks indefinitely. If an idle transaction is aborted, the commit fails with error `ABORTED`. If this behavior is undesirable, periodically executing a simple SQL query in the transaction (for example, `SELECT 1`) prevents the transaction from becoming idle. Snapshot read-only transactions: Snapshot read-only transactions provides a simpler method than locking read-write transactions for doing several consistent reads. However, this type of transaction does not support writes. Snapshot transactions do not take locks. Instead, they work by choosing a Cloud Spanner timestamp, then executing all reads at that timestamp. Since they do not acquire locks, they do not block concurrent read-write transactions. Unlike locking read-write transactions, snapshot read-only transactions never abort. They can fail if the chosen read timestamp is garbage collected; however, the default garbage collection policy is generous enough that most applications do not need to worry about this in practice. Snapshot read-only transactions do not need to call Commit or Rollback (and in fact are not permitted to do so). To execute a snapshot transaction, the client specifies a timestamp bound, which tells Cloud Spanner how to choose a read timestamp. The types of timestamp bound are: - Strong (the default). - Bounded staleness. - Exact staleness. If the Cloud Spanner database to be read is geographically distributed, stale read-only transactions can execute more quickly than strong or read-write transactions, because they are able to execute far from the leader replica. Each type of timestamp bound is discussed in detail below. Strong: Strong reads are guaranteed to see the effects of all transactions that have committed before the start of the read. Furthermore, all rows yielded by a single read are consistent with each other -- if any part of the read observes a transaction, all parts of the read see the transaction. Strong reads are not repeatable: two consecutive strong read-only transactions might return inconsistent results if there are concurrent writes. If consistency across reads is required, the reads should be executed within a transaction or at an exact read timestamp. Queries on change streams (see below for more details) must also specify the strong read timestamp bound. See TransactionOptions.ReadOnly.strong. Exact staleness: These timestamp bounds execute reads at a user-specified timestamp. Reads at a timestamp are guaranteed to see a consistent prefix of the global transaction history: they observe modifications done by all transactions with a commit timestamp less than or equal to the read timestamp, and observe none of the modifications done by transactions with a larger commit timestamp. They block until all conflicting transactions that can be assigned commit timestamps <= the read timestamp have finished. The timestamp can either be expressed as an absolute Cloud Spanner commit timestamp or a staleness relative to the current time. These modes do not require a "negotiation phase" to pick a timestamp. As a result, they execute slightly faster than the equivalent boundedly stale concurrency modes. On the other hand, boundedly stale reads usually return fresher results. See TransactionOptions.ReadOnly.read_timestamp and TransactionOptions.ReadOnly.exact_staleness. Bounded staleness: Bounded staleness modes allow Cloud Spanner to pick the read timestamp, subject to a user-provided staleness bound. Cloud Spanner chooses the newest timestamp within the staleness bound that allows execution of the reads at the closest available replica without blocking. All rows yielded are consistent with each other -- if any part of the read observes a transaction, all parts of the read see the transaction. Boundedly stale reads are not repeatable: two stale reads, even if they use the same staleness bound, can execute at different timestamps and thus return inconsistent results. Boundedly stale reads execute in two phases: the first phase negotiates a timestamp among all replicas needed to serve the read. In the second phase, reads are executed at the negotiated timestamp. As a result of the two phase execution, bounded staleness reads are usually a little slower than comparable exact staleness reads. However, they are typically able to return fresher results, and are more likely to execute at the closest replica. Because the timestamp negotiation requires up-front knowledge of which rows are read, it can only be used with single-use read-only transactions. See TransactionOptions.ReadOnly.max_staleness and TransactionOptions.ReadOnly.min_read_timestamp. Old read timestamps and garbage collection: Cloud Spanner continuously garbage collects deleted and overwritten data in the background to reclaim storage space. This process is known as "version GC". By default, version GC reclaims versions after they are one hour old. Because of this, Cloud Spanner can't perform reads at read timestamps more than one hour in the past. This restriction also applies to in-progress reads and/or SQL queries whose timestamp become too old while executing. Reads and SQL queries with too-old read timestamps fail with the error `FAILED_PRECONDITION`. You can configure and extend the `VERSION_RETENTION_PERIOD` of a database up to a period as long as one week, which allows Cloud Spanner to perform reads up to one week in the past. Querying change Streams: A Change Stream is a schema object that can be configured to watch data changes on the entire database, a set of tables, or a set of columns in a database. When a change stream is created, Spanner automatically defines a corresponding SQL Table-Valued Function (TVF) that can be used to query the change records in the associated change stream using the ExecuteStreamingSql API. The name of the TVF for a change stream is generated from the name of the change stream: READ_. All queries on change stream TVFs must be executed using the ExecuteStreamingSql API with a single-use read-only transaction with a strong read-only timestamp_bound. The change stream TVF allows users to specify the start_timestamp and end_timestamp for the time range of interest. All change records within the retention period is accessible using the strong read-only timestamp_bound. All other TransactionOptions are invalid for change stream queries. In addition, if TransactionOptions.read_only.return_read_timestamp is set to true, a special value of 2^63 - 2 is returned in the Transaction message that describes the transaction, instead of a valid read timestamp. This special value should be discarded and not used for any subsequent queries. Please see https://cloud.google.com/spanner/docs/change-streams for more details on how to query the change stream TVFs. Partitioned DML transactions: Partitioned DML transactions are used to execute DML statements with a different execution strategy that provides different, and often better, scalability properties for large, table-wide operations than DML in a ReadWrite transaction. Smaller scoped statements, such as an OLTP workload, should prefer using ReadWrite transactions. Partitioned DML partitions the keyspace and runs the DML statement on each partition in separate, internal transactions. These transactions commit automatically when complete, and run independently from one another. To reduce lock contention, this execution strategy only acquires read locks on rows that match the WHERE clause of the statement. Additionally, the smaller per-partition transactions hold locks for less time. That said, Partitioned DML is not a drop-in replacement for standard DML used in ReadWrite transactions. - The DML statement must be fully-partitionable. Specifically, the statement must be expressible as the union of many statements which each access only a single row of the table. - The statement is not applied atomically to all rows of the table. Rather, the statement is applied atomically to partitions of the table, in independent transactions. Secondary index rows are updated atomically with the base table rows. - Partitioned DML does not guarantee exactly-once execution semantics against a partition. The statement is applied at least once to each partition. It is strongly recommended that the DML statement should be idempotent to avoid unexpected results. For instance, it is potentially dangerous to run a statement such as `UPDATE table SET column = column + 1` as it could be run multiple times against some rows. - The partitions are committed automatically - there is no support for Commit or Rollback. If the call returns an error, or if the client issuing the ExecuteSql call dies, it is possible that some rows had the statement executed on them successfully. It is also possible that statement was never executed against other rows. - Partitioned DML transactions may only contain the execution of a single DML statement via ExecuteSql or ExecuteStreamingSql. - If any error is encountered during the execution of the partitioned DML operation (for instance, a UNIQUE INDEX violation, division by zero, or a value that can't be stored due to schema constraints), then the operation is stopped at that point and an error is returned. It is possible that at this point, some partitions have been committed (or even committed multiple times), and other partitions have not been run at all. Given the above, Partitioned DML is good fit for large, database-wide, operations that are idempotent, such as deleting old rows from a very large table. # Begin a new transaction and execute this read or SQL query in it. The transaction ID of the new transaction is returned in ResultSetMetadata.transaction, which is a Transaction. + "begin": { # Options to use for transactions. # Begin a new transaction and execute this read or SQL query in it. The transaction ID of the new transaction is returned in ResultSetMetadata.transaction, which is a Transaction. "excludeTxnFromChangeStreams": True or False, # When `exclude_txn_from_change_streams` is set to `true`, it prevents read or write transactions from being tracked in change streams. * If the DDL option `allow_txn_exclusion` is set to `true`, then the updates made within this transaction aren't recorded in the change stream. * If you don't set the DDL option `allow_txn_exclusion` or if it's set to `false`, then the updates made within this transaction are recorded in the change stream. When `exclude_txn_from_change_streams` is set to `false` or not set, modifications from this transaction are recorded in all change streams that are tracking columns modified by these transactions. The `exclude_txn_from_change_streams` option can only be specified for read-write or partitioned DML transactions, otherwise the API returns an `INVALID_ARGUMENT` error. "isolationLevel": "A String", # Isolation level for the transaction. "partitionedDml": { # Message type to initiate a Partitioned DML transaction. # Partitioned DML transaction. Authorization to begin a Partitioned DML transaction requires `spanner.databases.beginPartitionedDmlTransaction` permission on the `session` resource. @@ -1514,7 +1514,7 @@

Method Details

}, }, "id": "A String", # Execute the read or SQL query in a previously-started transaction. - "singleUse": { # Transactions: Each session can have at most one active transaction at a time (note that standalone reads and queries use a transaction internally and do count towards the one transaction limit). After the active transaction is completed, the session can immediately be re-used for the next transaction. It is not necessary to create a new session for each transaction. Transaction modes: Cloud Spanner supports three transaction modes: 1. Locking read-write. This type of transaction is the only way to write data into Cloud Spanner. These transactions rely on pessimistic locking and, if necessary, two-phase commit. Locking read-write transactions may abort, requiring the application to retry. 2. Snapshot read-only. Snapshot read-only transactions provide guaranteed consistency across several reads, but do not allow writes. Snapshot read-only transactions can be configured to read at timestamps in the past, or configured to perform a strong read (where Spanner selects a timestamp such that the read is guaranteed to see the effects of all transactions that have committed before the start of the read). Snapshot read-only transactions do not need to be committed. Queries on change streams must be performed with the snapshot read-only transaction mode, specifying a strong read. See TransactionOptions.ReadOnly.strong for more details. 3. Partitioned DML. This type of transaction is used to execute a single Partitioned DML statement. Partitioned DML partitions the key space and runs the DML statement over each partition in parallel using separate, internal transactions that commit independently. Partitioned DML transactions do not need to be committed. For transactions that only read, snapshot read-only transactions provide simpler semantics and are almost always faster. In particular, read-only transactions do not take locks, so they do not conflict with read-write transactions. As a consequence of not taking locks, they also do not abort, so retry loops are not needed. Transactions may only read-write data in a single database. They may, however, read-write data in different tables within that database. Locking read-write transactions: Locking transactions may be used to atomically read-modify-write data anywhere in a database. This type of transaction is externally consistent. Clients should attempt to minimize the amount of time a transaction is active. Faster transactions commit with higher probability and cause less contention. Cloud Spanner attempts to keep read locks active as long as the transaction continues to do reads, and the transaction has not been terminated by Commit or Rollback. Long periods of inactivity at the client may cause Cloud Spanner to release a transaction's locks and abort it. Conceptually, a read-write transaction consists of zero or more reads or SQL statements followed by Commit. At any time before Commit, the client can send a Rollback request to abort the transaction. Semantics: Cloud Spanner can commit the transaction if all read locks it acquired are still valid at commit time, and it is able to acquire write locks for all writes. Cloud Spanner can abort the transaction for any reason. If a commit attempt returns `ABORTED`, Cloud Spanner guarantees that the transaction has not modified any user data in Cloud Spanner. Unless the transaction commits, Cloud Spanner makes no guarantees about how long the transaction's locks were held for. It is an error to use Cloud Spanner locks for any sort of mutual exclusion other than between Cloud Spanner transactions themselves. Retrying aborted transactions: When a transaction aborts, the application can choose to retry the whole transaction again. To maximize the chances of successfully committing the retry, the client should execute the retry in the same session as the original attempt. The original session's lock priority increases with each consecutive abort, meaning that each attempt has a slightly better chance of success than the previous. Note that the lock priority is preserved per session (not per transaction). Lock priority is set by the first read or write in the first attempt of a read-write transaction. If the application starts a new session to retry the whole transaction, the transaction loses its original lock priority. Moreover, the lock priority is only preserved if the transaction fails with an `ABORTED` error. Under some circumstances (for example, many transactions attempting to modify the same row(s)), a transaction can abort many times in a short period before successfully committing. Thus, it is not a good idea to cap the number of retries a transaction can attempt; instead, it is better to limit the total amount of time spent retrying. Idle transactions: A transaction is considered idle if it has no outstanding reads or SQL queries and has not started a read or SQL query within the last 10 seconds. Idle transactions can be aborted by Cloud Spanner so that they don't hold on to locks indefinitely. If an idle transaction is aborted, the commit fails with error `ABORTED`. If this behavior is undesirable, periodically executing a simple SQL query in the transaction (for example, `SELECT 1`) prevents the transaction from becoming idle. Snapshot read-only transactions: Snapshot read-only transactions provides a simpler method than locking read-write transactions for doing several consistent reads. However, this type of transaction does not support writes. Snapshot transactions do not take locks. Instead, they work by choosing a Cloud Spanner timestamp, then executing all reads at that timestamp. Since they do not acquire locks, they do not block concurrent read-write transactions. Unlike locking read-write transactions, snapshot read-only transactions never abort. They can fail if the chosen read timestamp is garbage collected; however, the default garbage collection policy is generous enough that most applications do not need to worry about this in practice. Snapshot read-only transactions do not need to call Commit or Rollback (and in fact are not permitted to do so). To execute a snapshot transaction, the client specifies a timestamp bound, which tells Cloud Spanner how to choose a read timestamp. The types of timestamp bound are: - Strong (the default). - Bounded staleness. - Exact staleness. If the Cloud Spanner database to be read is geographically distributed, stale read-only transactions can execute more quickly than strong or read-write transactions, because they are able to execute far from the leader replica. Each type of timestamp bound is discussed in detail below. Strong: Strong reads are guaranteed to see the effects of all transactions that have committed before the start of the read. Furthermore, all rows yielded by a single read are consistent with each other -- if any part of the read observes a transaction, all parts of the read see the transaction. Strong reads are not repeatable: two consecutive strong read-only transactions might return inconsistent results if there are concurrent writes. If consistency across reads is required, the reads should be executed within a transaction or at an exact read timestamp. Queries on change streams (see below for more details) must also specify the strong read timestamp bound. See TransactionOptions.ReadOnly.strong. Exact staleness: These timestamp bounds execute reads at a user-specified timestamp. Reads at a timestamp are guaranteed to see a consistent prefix of the global transaction history: they observe modifications done by all transactions with a commit timestamp less than or equal to the read timestamp, and observe none of the modifications done by transactions with a larger commit timestamp. They block until all conflicting transactions that can be assigned commit timestamps <= the read timestamp have finished. The timestamp can either be expressed as an absolute Cloud Spanner commit timestamp or a staleness relative to the current time. These modes do not require a "negotiation phase" to pick a timestamp. As a result, they execute slightly faster than the equivalent boundedly stale concurrency modes. On the other hand, boundedly stale reads usually return fresher results. See TransactionOptions.ReadOnly.read_timestamp and TransactionOptions.ReadOnly.exact_staleness. Bounded staleness: Bounded staleness modes allow Cloud Spanner to pick the read timestamp, subject to a user-provided staleness bound. Cloud Spanner chooses the newest timestamp within the staleness bound that allows execution of the reads at the closest available replica without blocking. All rows yielded are consistent with each other -- if any part of the read observes a transaction, all parts of the read see the transaction. Boundedly stale reads are not repeatable: two stale reads, even if they use the same staleness bound, can execute at different timestamps and thus return inconsistent results. Boundedly stale reads execute in two phases: the first phase negotiates a timestamp among all replicas needed to serve the read. In the second phase, reads are executed at the negotiated timestamp. As a result of the two phase execution, bounded staleness reads are usually a little slower than comparable exact staleness reads. However, they are typically able to return fresher results, and are more likely to execute at the closest replica. Because the timestamp negotiation requires up-front knowledge of which rows are read, it can only be used with single-use read-only transactions. See TransactionOptions.ReadOnly.max_staleness and TransactionOptions.ReadOnly.min_read_timestamp. Old read timestamps and garbage collection: Cloud Spanner continuously garbage collects deleted and overwritten data in the background to reclaim storage space. This process is known as "version GC". By default, version GC reclaims versions after they are one hour old. Because of this, Cloud Spanner can't perform reads at read timestamps more than one hour in the past. This restriction also applies to in-progress reads and/or SQL queries whose timestamp become too old while executing. Reads and SQL queries with too-old read timestamps fail with the error `FAILED_PRECONDITION`. You can configure and extend the `VERSION_RETENTION_PERIOD` of a database up to a period as long as one week, which allows Cloud Spanner to perform reads up to one week in the past. Querying change Streams: A Change Stream is a schema object that can be configured to watch data changes on the entire database, a set of tables, or a set of columns in a database. When a change stream is created, Spanner automatically defines a corresponding SQL Table-Valued Function (TVF) that can be used to query the change records in the associated change stream using the ExecuteStreamingSql API. The name of the TVF for a change stream is generated from the name of the change stream: READ_. All queries on change stream TVFs must be executed using the ExecuteStreamingSql API with a single-use read-only transaction with a strong read-only timestamp_bound. The change stream TVF allows users to specify the start_timestamp and end_timestamp for the time range of interest. All change records within the retention period is accessible using the strong read-only timestamp_bound. All other TransactionOptions are invalid for change stream queries. In addition, if TransactionOptions.read_only.return_read_timestamp is set to true, a special value of 2^63 - 2 is returned in the Transaction message that describes the transaction, instead of a valid read timestamp. This special value should be discarded and not used for any subsequent queries. Please see https://cloud.google.com/spanner/docs/change-streams for more details on how to query the change stream TVFs. Partitioned DML transactions: Partitioned DML transactions are used to execute DML statements with a different execution strategy that provides different, and often better, scalability properties for large, table-wide operations than DML in a ReadWrite transaction. Smaller scoped statements, such as an OLTP workload, should prefer using ReadWrite transactions. Partitioned DML partitions the keyspace and runs the DML statement on each partition in separate, internal transactions. These transactions commit automatically when complete, and run independently from one another. To reduce lock contention, this execution strategy only acquires read locks on rows that match the WHERE clause of the statement. Additionally, the smaller per-partition transactions hold locks for less time. That said, Partitioned DML is not a drop-in replacement for standard DML used in ReadWrite transactions. - The DML statement must be fully-partitionable. Specifically, the statement must be expressible as the union of many statements which each access only a single row of the table. - The statement is not applied atomically to all rows of the table. Rather, the statement is applied atomically to partitions of the table, in independent transactions. Secondary index rows are updated atomically with the base table rows. - Partitioned DML does not guarantee exactly-once execution semantics against a partition. The statement is applied at least once to each partition. It is strongly recommended that the DML statement should be idempotent to avoid unexpected results. For instance, it is potentially dangerous to run a statement such as `UPDATE table SET column = column + 1` as it could be run multiple times against some rows. - The partitions are committed automatically - there is no support for Commit or Rollback. If the call returns an error, or if the client issuing the ExecuteSql call dies, it is possible that some rows had the statement executed on them successfully. It is also possible that statement was never executed against other rows. - Partitioned DML transactions may only contain the execution of a single DML statement via ExecuteSql or ExecuteStreamingSql. - If any error is encountered during the execution of the partitioned DML operation (for instance, a UNIQUE INDEX violation, division by zero, or a value that can't be stored due to schema constraints), then the operation is stopped at that point and an error is returned. It is possible that at this point, some partitions have been committed (or even committed multiple times), and other partitions have not been run at all. Given the above, Partitioned DML is good fit for large, database-wide, operations that are idempotent, such as deleting old rows from a very large table. # Execute the read or SQL query in a temporary transaction. This is the most efficient way to execute a transaction that consists of a single SQL query. + "singleUse": { # Options to use for transactions. # Execute the read or SQL query in a temporary transaction. This is the most efficient way to execute a transaction that consists of a single SQL query. "excludeTxnFromChangeStreams": True or False, # When `exclude_txn_from_change_streams` is set to `true`, it prevents read or write transactions from being tracked in change streams. * If the DDL option `allow_txn_exclusion` is set to `true`, then the updates made within this transaction aren't recorded in the change stream. * If you don't set the DDL option `allow_txn_exclusion` or if it's set to `false`, then the updates made within this transaction are recorded in the change stream. When `exclude_txn_from_change_streams` is set to `false` or not set, modifications from this transaction are recorded in all change streams that are tracking columns modified by these transactions. The `exclude_txn_from_change_streams` option can only be specified for read-write or partitioned DML transactions, otherwise the API returns an `INVALID_ARGUMENT` error. "isolationLevel": "A String", # Isolation level for the transaction. "partitionedDml": { # Message type to initiate a Partitioned DML transaction. # Partitioned DML transaction. Authorization to begin a Partitioned DML transaction requires `spanner.databases.beginPartitionedDmlTransaction` permission on the `session` resource. @@ -1630,7 +1630,7 @@

Method Details

"resumeToken": "A String", # If this request is resuming a previously interrupted read, `resume_token` should be copied from the last PartialResultSet yielded before the interruption. Doing this enables the new read to resume where the last read left off. The rest of the request parameters must exactly match the request that yielded this token. "table": "A String", # Required. The name of the table in the database to be read. "transaction": { # This message is used to select the transaction in which a Read or ExecuteSql call runs. See TransactionOptions for more information about transactions. # The transaction to use. If none is provided, the default is a temporary read-only transaction with strong concurrency. - "begin": { # Transactions: Each session can have at most one active transaction at a time (note that standalone reads and queries use a transaction internally and do count towards the one transaction limit). After the active transaction is completed, the session can immediately be re-used for the next transaction. It is not necessary to create a new session for each transaction. Transaction modes: Cloud Spanner supports three transaction modes: 1. Locking read-write. This type of transaction is the only way to write data into Cloud Spanner. These transactions rely on pessimistic locking and, if necessary, two-phase commit. Locking read-write transactions may abort, requiring the application to retry. 2. Snapshot read-only. Snapshot read-only transactions provide guaranteed consistency across several reads, but do not allow writes. Snapshot read-only transactions can be configured to read at timestamps in the past, or configured to perform a strong read (where Spanner selects a timestamp such that the read is guaranteed to see the effects of all transactions that have committed before the start of the read). Snapshot read-only transactions do not need to be committed. Queries on change streams must be performed with the snapshot read-only transaction mode, specifying a strong read. See TransactionOptions.ReadOnly.strong for more details. 3. Partitioned DML. This type of transaction is used to execute a single Partitioned DML statement. Partitioned DML partitions the key space and runs the DML statement over each partition in parallel using separate, internal transactions that commit independently. Partitioned DML transactions do not need to be committed. For transactions that only read, snapshot read-only transactions provide simpler semantics and are almost always faster. In particular, read-only transactions do not take locks, so they do not conflict with read-write transactions. As a consequence of not taking locks, they also do not abort, so retry loops are not needed. Transactions may only read-write data in a single database. They may, however, read-write data in different tables within that database. Locking read-write transactions: Locking transactions may be used to atomically read-modify-write data anywhere in a database. This type of transaction is externally consistent. Clients should attempt to minimize the amount of time a transaction is active. Faster transactions commit with higher probability and cause less contention. Cloud Spanner attempts to keep read locks active as long as the transaction continues to do reads, and the transaction has not been terminated by Commit or Rollback. Long periods of inactivity at the client may cause Cloud Spanner to release a transaction's locks and abort it. Conceptually, a read-write transaction consists of zero or more reads or SQL statements followed by Commit. At any time before Commit, the client can send a Rollback request to abort the transaction. Semantics: Cloud Spanner can commit the transaction if all read locks it acquired are still valid at commit time, and it is able to acquire write locks for all writes. Cloud Spanner can abort the transaction for any reason. If a commit attempt returns `ABORTED`, Cloud Spanner guarantees that the transaction has not modified any user data in Cloud Spanner. Unless the transaction commits, Cloud Spanner makes no guarantees about how long the transaction's locks were held for. It is an error to use Cloud Spanner locks for any sort of mutual exclusion other than between Cloud Spanner transactions themselves. Retrying aborted transactions: When a transaction aborts, the application can choose to retry the whole transaction again. To maximize the chances of successfully committing the retry, the client should execute the retry in the same session as the original attempt. The original session's lock priority increases with each consecutive abort, meaning that each attempt has a slightly better chance of success than the previous. Note that the lock priority is preserved per session (not per transaction). Lock priority is set by the first read or write in the first attempt of a read-write transaction. If the application starts a new session to retry the whole transaction, the transaction loses its original lock priority. Moreover, the lock priority is only preserved if the transaction fails with an `ABORTED` error. Under some circumstances (for example, many transactions attempting to modify the same row(s)), a transaction can abort many times in a short period before successfully committing. Thus, it is not a good idea to cap the number of retries a transaction can attempt; instead, it is better to limit the total amount of time spent retrying. Idle transactions: A transaction is considered idle if it has no outstanding reads or SQL queries and has not started a read or SQL query within the last 10 seconds. Idle transactions can be aborted by Cloud Spanner so that they don't hold on to locks indefinitely. If an idle transaction is aborted, the commit fails with error `ABORTED`. If this behavior is undesirable, periodically executing a simple SQL query in the transaction (for example, `SELECT 1`) prevents the transaction from becoming idle. Snapshot read-only transactions: Snapshot read-only transactions provides a simpler method than locking read-write transactions for doing several consistent reads. However, this type of transaction does not support writes. Snapshot transactions do not take locks. Instead, they work by choosing a Cloud Spanner timestamp, then executing all reads at that timestamp. Since they do not acquire locks, they do not block concurrent read-write transactions. Unlike locking read-write transactions, snapshot read-only transactions never abort. They can fail if the chosen read timestamp is garbage collected; however, the default garbage collection policy is generous enough that most applications do not need to worry about this in practice. Snapshot read-only transactions do not need to call Commit or Rollback (and in fact are not permitted to do so). To execute a snapshot transaction, the client specifies a timestamp bound, which tells Cloud Spanner how to choose a read timestamp. The types of timestamp bound are: - Strong (the default). - Bounded staleness. - Exact staleness. If the Cloud Spanner database to be read is geographically distributed, stale read-only transactions can execute more quickly than strong or read-write transactions, because they are able to execute far from the leader replica. Each type of timestamp bound is discussed in detail below. Strong: Strong reads are guaranteed to see the effects of all transactions that have committed before the start of the read. Furthermore, all rows yielded by a single read are consistent with each other -- if any part of the read observes a transaction, all parts of the read see the transaction. Strong reads are not repeatable: two consecutive strong read-only transactions might return inconsistent results if there are concurrent writes. If consistency across reads is required, the reads should be executed within a transaction or at an exact read timestamp. Queries on change streams (see below for more details) must also specify the strong read timestamp bound. See TransactionOptions.ReadOnly.strong. Exact staleness: These timestamp bounds execute reads at a user-specified timestamp. Reads at a timestamp are guaranteed to see a consistent prefix of the global transaction history: they observe modifications done by all transactions with a commit timestamp less than or equal to the read timestamp, and observe none of the modifications done by transactions with a larger commit timestamp. They block until all conflicting transactions that can be assigned commit timestamps <= the read timestamp have finished. The timestamp can either be expressed as an absolute Cloud Spanner commit timestamp or a staleness relative to the current time. These modes do not require a "negotiation phase" to pick a timestamp. As a result, they execute slightly faster than the equivalent boundedly stale concurrency modes. On the other hand, boundedly stale reads usually return fresher results. See TransactionOptions.ReadOnly.read_timestamp and TransactionOptions.ReadOnly.exact_staleness. Bounded staleness: Bounded staleness modes allow Cloud Spanner to pick the read timestamp, subject to a user-provided staleness bound. Cloud Spanner chooses the newest timestamp within the staleness bound that allows execution of the reads at the closest available replica without blocking. All rows yielded are consistent with each other -- if any part of the read observes a transaction, all parts of the read see the transaction. Boundedly stale reads are not repeatable: two stale reads, even if they use the same staleness bound, can execute at different timestamps and thus return inconsistent results. Boundedly stale reads execute in two phases: the first phase negotiates a timestamp among all replicas needed to serve the read. In the second phase, reads are executed at the negotiated timestamp. As a result of the two phase execution, bounded staleness reads are usually a little slower than comparable exact staleness reads. However, they are typically able to return fresher results, and are more likely to execute at the closest replica. Because the timestamp negotiation requires up-front knowledge of which rows are read, it can only be used with single-use read-only transactions. See TransactionOptions.ReadOnly.max_staleness and TransactionOptions.ReadOnly.min_read_timestamp. Old read timestamps and garbage collection: Cloud Spanner continuously garbage collects deleted and overwritten data in the background to reclaim storage space. This process is known as "version GC". By default, version GC reclaims versions after they are one hour old. Because of this, Cloud Spanner can't perform reads at read timestamps more than one hour in the past. This restriction also applies to in-progress reads and/or SQL queries whose timestamp become too old while executing. Reads and SQL queries with too-old read timestamps fail with the error `FAILED_PRECONDITION`. You can configure and extend the `VERSION_RETENTION_PERIOD` of a database up to a period as long as one week, which allows Cloud Spanner to perform reads up to one week in the past. Querying change Streams: A Change Stream is a schema object that can be configured to watch data changes on the entire database, a set of tables, or a set of columns in a database. When a change stream is created, Spanner automatically defines a corresponding SQL Table-Valued Function (TVF) that can be used to query the change records in the associated change stream using the ExecuteStreamingSql API. The name of the TVF for a change stream is generated from the name of the change stream: READ_. All queries on change stream TVFs must be executed using the ExecuteStreamingSql API with a single-use read-only transaction with a strong read-only timestamp_bound. The change stream TVF allows users to specify the start_timestamp and end_timestamp for the time range of interest. All change records within the retention period is accessible using the strong read-only timestamp_bound. All other TransactionOptions are invalid for change stream queries. In addition, if TransactionOptions.read_only.return_read_timestamp is set to true, a special value of 2^63 - 2 is returned in the Transaction message that describes the transaction, instead of a valid read timestamp. This special value should be discarded and not used for any subsequent queries. Please see https://cloud.google.com/spanner/docs/change-streams for more details on how to query the change stream TVFs. Partitioned DML transactions: Partitioned DML transactions are used to execute DML statements with a different execution strategy that provides different, and often better, scalability properties for large, table-wide operations than DML in a ReadWrite transaction. Smaller scoped statements, such as an OLTP workload, should prefer using ReadWrite transactions. Partitioned DML partitions the keyspace and runs the DML statement on each partition in separate, internal transactions. These transactions commit automatically when complete, and run independently from one another. To reduce lock contention, this execution strategy only acquires read locks on rows that match the WHERE clause of the statement. Additionally, the smaller per-partition transactions hold locks for less time. That said, Partitioned DML is not a drop-in replacement for standard DML used in ReadWrite transactions. - The DML statement must be fully-partitionable. Specifically, the statement must be expressible as the union of many statements which each access only a single row of the table. - The statement is not applied atomically to all rows of the table. Rather, the statement is applied atomically to partitions of the table, in independent transactions. Secondary index rows are updated atomically with the base table rows. - Partitioned DML does not guarantee exactly-once execution semantics against a partition. The statement is applied at least once to each partition. It is strongly recommended that the DML statement should be idempotent to avoid unexpected results. For instance, it is potentially dangerous to run a statement such as `UPDATE table SET column = column + 1` as it could be run multiple times against some rows. - The partitions are committed automatically - there is no support for Commit or Rollback. If the call returns an error, or if the client issuing the ExecuteSql call dies, it is possible that some rows had the statement executed on them successfully. It is also possible that statement was never executed against other rows. - Partitioned DML transactions may only contain the execution of a single DML statement via ExecuteSql or ExecuteStreamingSql. - If any error is encountered during the execution of the partitioned DML operation (for instance, a UNIQUE INDEX violation, division by zero, or a value that can't be stored due to schema constraints), then the operation is stopped at that point and an error is returned. It is possible that at this point, some partitions have been committed (or even committed multiple times), and other partitions have not been run at all. Given the above, Partitioned DML is good fit for large, database-wide, operations that are idempotent, such as deleting old rows from a very large table. # Begin a new transaction and execute this read or SQL query in it. The transaction ID of the new transaction is returned in ResultSetMetadata.transaction, which is a Transaction. + "begin": { # Options to use for transactions. # Begin a new transaction and execute this read or SQL query in it. The transaction ID of the new transaction is returned in ResultSetMetadata.transaction, which is a Transaction. "excludeTxnFromChangeStreams": True or False, # When `exclude_txn_from_change_streams` is set to `true`, it prevents read or write transactions from being tracked in change streams. * If the DDL option `allow_txn_exclusion` is set to `true`, then the updates made within this transaction aren't recorded in the change stream. * If you don't set the DDL option `allow_txn_exclusion` or if it's set to `false`, then the updates made within this transaction are recorded in the change stream. When `exclude_txn_from_change_streams` is set to `false` or not set, modifications from this transaction are recorded in all change streams that are tracking columns modified by these transactions. The `exclude_txn_from_change_streams` option can only be specified for read-write or partitioned DML transactions, otherwise the API returns an `INVALID_ARGUMENT` error. "isolationLevel": "A String", # Isolation level for the transaction. "partitionedDml": { # Message type to initiate a Partitioned DML transaction. # Partitioned DML transaction. Authorization to begin a Partitioned DML transaction requires `spanner.databases.beginPartitionedDmlTransaction` permission on the `session` resource. @@ -1649,7 +1649,7 @@

Method Details

}, }, "id": "A String", # Execute the read or SQL query in a previously-started transaction. - "singleUse": { # Transactions: Each session can have at most one active transaction at a time (note that standalone reads and queries use a transaction internally and do count towards the one transaction limit). After the active transaction is completed, the session can immediately be re-used for the next transaction. It is not necessary to create a new session for each transaction. Transaction modes: Cloud Spanner supports three transaction modes: 1. Locking read-write. This type of transaction is the only way to write data into Cloud Spanner. These transactions rely on pessimistic locking and, if necessary, two-phase commit. Locking read-write transactions may abort, requiring the application to retry. 2. Snapshot read-only. Snapshot read-only transactions provide guaranteed consistency across several reads, but do not allow writes. Snapshot read-only transactions can be configured to read at timestamps in the past, or configured to perform a strong read (where Spanner selects a timestamp such that the read is guaranteed to see the effects of all transactions that have committed before the start of the read). Snapshot read-only transactions do not need to be committed. Queries on change streams must be performed with the snapshot read-only transaction mode, specifying a strong read. See TransactionOptions.ReadOnly.strong for more details. 3. Partitioned DML. This type of transaction is used to execute a single Partitioned DML statement. Partitioned DML partitions the key space and runs the DML statement over each partition in parallel using separate, internal transactions that commit independently. Partitioned DML transactions do not need to be committed. For transactions that only read, snapshot read-only transactions provide simpler semantics and are almost always faster. In particular, read-only transactions do not take locks, so they do not conflict with read-write transactions. As a consequence of not taking locks, they also do not abort, so retry loops are not needed. Transactions may only read-write data in a single database. They may, however, read-write data in different tables within that database. Locking read-write transactions: Locking transactions may be used to atomically read-modify-write data anywhere in a database. This type of transaction is externally consistent. Clients should attempt to minimize the amount of time a transaction is active. Faster transactions commit with higher probability and cause less contention. Cloud Spanner attempts to keep read locks active as long as the transaction continues to do reads, and the transaction has not been terminated by Commit or Rollback. Long periods of inactivity at the client may cause Cloud Spanner to release a transaction's locks and abort it. Conceptually, a read-write transaction consists of zero or more reads or SQL statements followed by Commit. At any time before Commit, the client can send a Rollback request to abort the transaction. Semantics: Cloud Spanner can commit the transaction if all read locks it acquired are still valid at commit time, and it is able to acquire write locks for all writes. Cloud Spanner can abort the transaction for any reason. If a commit attempt returns `ABORTED`, Cloud Spanner guarantees that the transaction has not modified any user data in Cloud Spanner. Unless the transaction commits, Cloud Spanner makes no guarantees about how long the transaction's locks were held for. It is an error to use Cloud Spanner locks for any sort of mutual exclusion other than between Cloud Spanner transactions themselves. Retrying aborted transactions: When a transaction aborts, the application can choose to retry the whole transaction again. To maximize the chances of successfully committing the retry, the client should execute the retry in the same session as the original attempt. The original session's lock priority increases with each consecutive abort, meaning that each attempt has a slightly better chance of success than the previous. Note that the lock priority is preserved per session (not per transaction). Lock priority is set by the first read or write in the first attempt of a read-write transaction. If the application starts a new session to retry the whole transaction, the transaction loses its original lock priority. Moreover, the lock priority is only preserved if the transaction fails with an `ABORTED` error. Under some circumstances (for example, many transactions attempting to modify the same row(s)), a transaction can abort many times in a short period before successfully committing. Thus, it is not a good idea to cap the number of retries a transaction can attempt; instead, it is better to limit the total amount of time spent retrying. Idle transactions: A transaction is considered idle if it has no outstanding reads or SQL queries and has not started a read or SQL query within the last 10 seconds. Idle transactions can be aborted by Cloud Spanner so that they don't hold on to locks indefinitely. If an idle transaction is aborted, the commit fails with error `ABORTED`. If this behavior is undesirable, periodically executing a simple SQL query in the transaction (for example, `SELECT 1`) prevents the transaction from becoming idle. Snapshot read-only transactions: Snapshot read-only transactions provides a simpler method than locking read-write transactions for doing several consistent reads. However, this type of transaction does not support writes. Snapshot transactions do not take locks. Instead, they work by choosing a Cloud Spanner timestamp, then executing all reads at that timestamp. Since they do not acquire locks, they do not block concurrent read-write transactions. Unlike locking read-write transactions, snapshot read-only transactions never abort. They can fail if the chosen read timestamp is garbage collected; however, the default garbage collection policy is generous enough that most applications do not need to worry about this in practice. Snapshot read-only transactions do not need to call Commit or Rollback (and in fact are not permitted to do so). To execute a snapshot transaction, the client specifies a timestamp bound, which tells Cloud Spanner how to choose a read timestamp. The types of timestamp bound are: - Strong (the default). - Bounded staleness. - Exact staleness. If the Cloud Spanner database to be read is geographically distributed, stale read-only transactions can execute more quickly than strong or read-write transactions, because they are able to execute far from the leader replica. Each type of timestamp bound is discussed in detail below. Strong: Strong reads are guaranteed to see the effects of all transactions that have committed before the start of the read. Furthermore, all rows yielded by a single read are consistent with each other -- if any part of the read observes a transaction, all parts of the read see the transaction. Strong reads are not repeatable: two consecutive strong read-only transactions might return inconsistent results if there are concurrent writes. If consistency across reads is required, the reads should be executed within a transaction or at an exact read timestamp. Queries on change streams (see below for more details) must also specify the strong read timestamp bound. See TransactionOptions.ReadOnly.strong. Exact staleness: These timestamp bounds execute reads at a user-specified timestamp. Reads at a timestamp are guaranteed to see a consistent prefix of the global transaction history: they observe modifications done by all transactions with a commit timestamp less than or equal to the read timestamp, and observe none of the modifications done by transactions with a larger commit timestamp. They block until all conflicting transactions that can be assigned commit timestamps <= the read timestamp have finished. The timestamp can either be expressed as an absolute Cloud Spanner commit timestamp or a staleness relative to the current time. These modes do not require a "negotiation phase" to pick a timestamp. As a result, they execute slightly faster than the equivalent boundedly stale concurrency modes. On the other hand, boundedly stale reads usually return fresher results. See TransactionOptions.ReadOnly.read_timestamp and TransactionOptions.ReadOnly.exact_staleness. Bounded staleness: Bounded staleness modes allow Cloud Spanner to pick the read timestamp, subject to a user-provided staleness bound. Cloud Spanner chooses the newest timestamp within the staleness bound that allows execution of the reads at the closest available replica without blocking. All rows yielded are consistent with each other -- if any part of the read observes a transaction, all parts of the read see the transaction. Boundedly stale reads are not repeatable: two stale reads, even if they use the same staleness bound, can execute at different timestamps and thus return inconsistent results. Boundedly stale reads execute in two phases: the first phase negotiates a timestamp among all replicas needed to serve the read. In the second phase, reads are executed at the negotiated timestamp. As a result of the two phase execution, bounded staleness reads are usually a little slower than comparable exact staleness reads. However, they are typically able to return fresher results, and are more likely to execute at the closest replica. Because the timestamp negotiation requires up-front knowledge of which rows are read, it can only be used with single-use read-only transactions. See TransactionOptions.ReadOnly.max_staleness and TransactionOptions.ReadOnly.min_read_timestamp. Old read timestamps and garbage collection: Cloud Spanner continuously garbage collects deleted and overwritten data in the background to reclaim storage space. This process is known as "version GC". By default, version GC reclaims versions after they are one hour old. Because of this, Cloud Spanner can't perform reads at read timestamps more than one hour in the past. This restriction also applies to in-progress reads and/or SQL queries whose timestamp become too old while executing. Reads and SQL queries with too-old read timestamps fail with the error `FAILED_PRECONDITION`. You can configure and extend the `VERSION_RETENTION_PERIOD` of a database up to a period as long as one week, which allows Cloud Spanner to perform reads up to one week in the past. Querying change Streams: A Change Stream is a schema object that can be configured to watch data changes on the entire database, a set of tables, or a set of columns in a database. When a change stream is created, Spanner automatically defines a corresponding SQL Table-Valued Function (TVF) that can be used to query the change records in the associated change stream using the ExecuteStreamingSql API. The name of the TVF for a change stream is generated from the name of the change stream: READ_. All queries on change stream TVFs must be executed using the ExecuteStreamingSql API with a single-use read-only transaction with a strong read-only timestamp_bound. The change stream TVF allows users to specify the start_timestamp and end_timestamp for the time range of interest. All change records within the retention period is accessible using the strong read-only timestamp_bound. All other TransactionOptions are invalid for change stream queries. In addition, if TransactionOptions.read_only.return_read_timestamp is set to true, a special value of 2^63 - 2 is returned in the Transaction message that describes the transaction, instead of a valid read timestamp. This special value should be discarded and not used for any subsequent queries. Please see https://cloud.google.com/spanner/docs/change-streams for more details on how to query the change stream TVFs. Partitioned DML transactions: Partitioned DML transactions are used to execute DML statements with a different execution strategy that provides different, and often better, scalability properties for large, table-wide operations than DML in a ReadWrite transaction. Smaller scoped statements, such as an OLTP workload, should prefer using ReadWrite transactions. Partitioned DML partitions the keyspace and runs the DML statement on each partition in separate, internal transactions. These transactions commit automatically when complete, and run independently from one another. To reduce lock contention, this execution strategy only acquires read locks on rows that match the WHERE clause of the statement. Additionally, the smaller per-partition transactions hold locks for less time. That said, Partitioned DML is not a drop-in replacement for standard DML used in ReadWrite transactions. - The DML statement must be fully-partitionable. Specifically, the statement must be expressible as the union of many statements which each access only a single row of the table. - The statement is not applied atomically to all rows of the table. Rather, the statement is applied atomically to partitions of the table, in independent transactions. Secondary index rows are updated atomically with the base table rows. - Partitioned DML does not guarantee exactly-once execution semantics against a partition. The statement is applied at least once to each partition. It is strongly recommended that the DML statement should be idempotent to avoid unexpected results. For instance, it is potentially dangerous to run a statement such as `UPDATE table SET column = column + 1` as it could be run multiple times against some rows. - The partitions are committed automatically - there is no support for Commit or Rollback. If the call returns an error, or if the client issuing the ExecuteSql call dies, it is possible that some rows had the statement executed on them successfully. It is also possible that statement was never executed against other rows. - Partitioned DML transactions may only contain the execution of a single DML statement via ExecuteSql or ExecuteStreamingSql. - If any error is encountered during the execution of the partitioned DML operation (for instance, a UNIQUE INDEX violation, division by zero, or a value that can't be stored due to schema constraints), then the operation is stopped at that point and an error is returned. It is possible that at this point, some partitions have been committed (or even committed multiple times), and other partitions have not been run at all. Given the above, Partitioned DML is good fit for large, database-wide, operations that are idempotent, such as deleting old rows from a very large table. # Execute the read or SQL query in a temporary transaction. This is the most efficient way to execute a transaction that consists of a single SQL query. + "singleUse": { # Options to use for transactions. # Execute the read or SQL query in a temporary transaction. This is the most efficient way to execute a transaction that consists of a single SQL query. "excludeTxnFromChangeStreams": True or False, # When `exclude_txn_from_change_streams` is set to `true`, it prevents read or write transactions from being tracked in change streams. * If the DDL option `allow_txn_exclusion` is set to `true`, then the updates made within this transaction aren't recorded in the change stream. * If you don't set the DDL option `allow_txn_exclusion` or if it's set to `false`, then the updates made within this transaction are recorded in the change stream. When `exclude_txn_from_change_streams` is set to `false` or not set, modifications from this transaction are recorded in all change streams that are tracking columns modified by these transactions. The `exclude_txn_from_change_streams` option can only be specified for read-write or partitioned DML transactions, otherwise the API returns an `INVALID_ARGUMENT` error. "isolationLevel": "A String", # Isolation level for the transaction. "partitionedDml": { # Message type to initiate a Partitioned DML transaction. # Partitioned DML transaction. Authorization to begin a Partitioned DML transaction requires `spanner.databases.beginPartitionedDmlTransaction` permission on the `session` resource. @@ -1857,7 +1857,7 @@

Method Details

"resumeToken": "A String", # If this request is resuming a previously interrupted read, `resume_token` should be copied from the last PartialResultSet yielded before the interruption. Doing this enables the new read to resume where the last read left off. The rest of the request parameters must exactly match the request that yielded this token. "table": "A String", # Required. The name of the table in the database to be read. "transaction": { # This message is used to select the transaction in which a Read or ExecuteSql call runs. See TransactionOptions for more information about transactions. # The transaction to use. If none is provided, the default is a temporary read-only transaction with strong concurrency. - "begin": { # Transactions: Each session can have at most one active transaction at a time (note that standalone reads and queries use a transaction internally and do count towards the one transaction limit). After the active transaction is completed, the session can immediately be re-used for the next transaction. It is not necessary to create a new session for each transaction. Transaction modes: Cloud Spanner supports three transaction modes: 1. Locking read-write. This type of transaction is the only way to write data into Cloud Spanner. These transactions rely on pessimistic locking and, if necessary, two-phase commit. Locking read-write transactions may abort, requiring the application to retry. 2. Snapshot read-only. Snapshot read-only transactions provide guaranteed consistency across several reads, but do not allow writes. Snapshot read-only transactions can be configured to read at timestamps in the past, or configured to perform a strong read (where Spanner selects a timestamp such that the read is guaranteed to see the effects of all transactions that have committed before the start of the read). Snapshot read-only transactions do not need to be committed. Queries on change streams must be performed with the snapshot read-only transaction mode, specifying a strong read. See TransactionOptions.ReadOnly.strong for more details. 3. Partitioned DML. This type of transaction is used to execute a single Partitioned DML statement. Partitioned DML partitions the key space and runs the DML statement over each partition in parallel using separate, internal transactions that commit independently. Partitioned DML transactions do not need to be committed. For transactions that only read, snapshot read-only transactions provide simpler semantics and are almost always faster. In particular, read-only transactions do not take locks, so they do not conflict with read-write transactions. As a consequence of not taking locks, they also do not abort, so retry loops are not needed. Transactions may only read-write data in a single database. They may, however, read-write data in different tables within that database. Locking read-write transactions: Locking transactions may be used to atomically read-modify-write data anywhere in a database. This type of transaction is externally consistent. Clients should attempt to minimize the amount of time a transaction is active. Faster transactions commit with higher probability and cause less contention. Cloud Spanner attempts to keep read locks active as long as the transaction continues to do reads, and the transaction has not been terminated by Commit or Rollback. Long periods of inactivity at the client may cause Cloud Spanner to release a transaction's locks and abort it. Conceptually, a read-write transaction consists of zero or more reads or SQL statements followed by Commit. At any time before Commit, the client can send a Rollback request to abort the transaction. Semantics: Cloud Spanner can commit the transaction if all read locks it acquired are still valid at commit time, and it is able to acquire write locks for all writes. Cloud Spanner can abort the transaction for any reason. If a commit attempt returns `ABORTED`, Cloud Spanner guarantees that the transaction has not modified any user data in Cloud Spanner. Unless the transaction commits, Cloud Spanner makes no guarantees about how long the transaction's locks were held for. It is an error to use Cloud Spanner locks for any sort of mutual exclusion other than between Cloud Spanner transactions themselves. Retrying aborted transactions: When a transaction aborts, the application can choose to retry the whole transaction again. To maximize the chances of successfully committing the retry, the client should execute the retry in the same session as the original attempt. The original session's lock priority increases with each consecutive abort, meaning that each attempt has a slightly better chance of success than the previous. Note that the lock priority is preserved per session (not per transaction). Lock priority is set by the first read or write in the first attempt of a read-write transaction. If the application starts a new session to retry the whole transaction, the transaction loses its original lock priority. Moreover, the lock priority is only preserved if the transaction fails with an `ABORTED` error. Under some circumstances (for example, many transactions attempting to modify the same row(s)), a transaction can abort many times in a short period before successfully committing. Thus, it is not a good idea to cap the number of retries a transaction can attempt; instead, it is better to limit the total amount of time spent retrying. Idle transactions: A transaction is considered idle if it has no outstanding reads or SQL queries and has not started a read or SQL query within the last 10 seconds. Idle transactions can be aborted by Cloud Spanner so that they don't hold on to locks indefinitely. If an idle transaction is aborted, the commit fails with error `ABORTED`. If this behavior is undesirable, periodically executing a simple SQL query in the transaction (for example, `SELECT 1`) prevents the transaction from becoming idle. Snapshot read-only transactions: Snapshot read-only transactions provides a simpler method than locking read-write transactions for doing several consistent reads. However, this type of transaction does not support writes. Snapshot transactions do not take locks. Instead, they work by choosing a Cloud Spanner timestamp, then executing all reads at that timestamp. Since they do not acquire locks, they do not block concurrent read-write transactions. Unlike locking read-write transactions, snapshot read-only transactions never abort. They can fail if the chosen read timestamp is garbage collected; however, the default garbage collection policy is generous enough that most applications do not need to worry about this in practice. Snapshot read-only transactions do not need to call Commit or Rollback (and in fact are not permitted to do so). To execute a snapshot transaction, the client specifies a timestamp bound, which tells Cloud Spanner how to choose a read timestamp. The types of timestamp bound are: - Strong (the default). - Bounded staleness. - Exact staleness. If the Cloud Spanner database to be read is geographically distributed, stale read-only transactions can execute more quickly than strong or read-write transactions, because they are able to execute far from the leader replica. Each type of timestamp bound is discussed in detail below. Strong: Strong reads are guaranteed to see the effects of all transactions that have committed before the start of the read. Furthermore, all rows yielded by a single read are consistent with each other -- if any part of the read observes a transaction, all parts of the read see the transaction. Strong reads are not repeatable: two consecutive strong read-only transactions might return inconsistent results if there are concurrent writes. If consistency across reads is required, the reads should be executed within a transaction or at an exact read timestamp. Queries on change streams (see below for more details) must also specify the strong read timestamp bound. See TransactionOptions.ReadOnly.strong. Exact staleness: These timestamp bounds execute reads at a user-specified timestamp. Reads at a timestamp are guaranteed to see a consistent prefix of the global transaction history: they observe modifications done by all transactions with a commit timestamp less than or equal to the read timestamp, and observe none of the modifications done by transactions with a larger commit timestamp. They block until all conflicting transactions that can be assigned commit timestamps <= the read timestamp have finished. The timestamp can either be expressed as an absolute Cloud Spanner commit timestamp or a staleness relative to the current time. These modes do not require a "negotiation phase" to pick a timestamp. As a result, they execute slightly faster than the equivalent boundedly stale concurrency modes. On the other hand, boundedly stale reads usually return fresher results. See TransactionOptions.ReadOnly.read_timestamp and TransactionOptions.ReadOnly.exact_staleness. Bounded staleness: Bounded staleness modes allow Cloud Spanner to pick the read timestamp, subject to a user-provided staleness bound. Cloud Spanner chooses the newest timestamp within the staleness bound that allows execution of the reads at the closest available replica without blocking. All rows yielded are consistent with each other -- if any part of the read observes a transaction, all parts of the read see the transaction. Boundedly stale reads are not repeatable: two stale reads, even if they use the same staleness bound, can execute at different timestamps and thus return inconsistent results. Boundedly stale reads execute in two phases: the first phase negotiates a timestamp among all replicas needed to serve the read. In the second phase, reads are executed at the negotiated timestamp. As a result of the two phase execution, bounded staleness reads are usually a little slower than comparable exact staleness reads. However, they are typically able to return fresher results, and are more likely to execute at the closest replica. Because the timestamp negotiation requires up-front knowledge of which rows are read, it can only be used with single-use read-only transactions. See TransactionOptions.ReadOnly.max_staleness and TransactionOptions.ReadOnly.min_read_timestamp. Old read timestamps and garbage collection: Cloud Spanner continuously garbage collects deleted and overwritten data in the background to reclaim storage space. This process is known as "version GC". By default, version GC reclaims versions after they are one hour old. Because of this, Cloud Spanner can't perform reads at read timestamps more than one hour in the past. This restriction also applies to in-progress reads and/or SQL queries whose timestamp become too old while executing. Reads and SQL queries with too-old read timestamps fail with the error `FAILED_PRECONDITION`. You can configure and extend the `VERSION_RETENTION_PERIOD` of a database up to a period as long as one week, which allows Cloud Spanner to perform reads up to one week in the past. Querying change Streams: A Change Stream is a schema object that can be configured to watch data changes on the entire database, a set of tables, or a set of columns in a database. When a change stream is created, Spanner automatically defines a corresponding SQL Table-Valued Function (TVF) that can be used to query the change records in the associated change stream using the ExecuteStreamingSql API. The name of the TVF for a change stream is generated from the name of the change stream: READ_. All queries on change stream TVFs must be executed using the ExecuteStreamingSql API with a single-use read-only transaction with a strong read-only timestamp_bound. The change stream TVF allows users to specify the start_timestamp and end_timestamp for the time range of interest. All change records within the retention period is accessible using the strong read-only timestamp_bound. All other TransactionOptions are invalid for change stream queries. In addition, if TransactionOptions.read_only.return_read_timestamp is set to true, a special value of 2^63 - 2 is returned in the Transaction message that describes the transaction, instead of a valid read timestamp. This special value should be discarded and not used for any subsequent queries. Please see https://cloud.google.com/spanner/docs/change-streams for more details on how to query the change stream TVFs. Partitioned DML transactions: Partitioned DML transactions are used to execute DML statements with a different execution strategy that provides different, and often better, scalability properties for large, table-wide operations than DML in a ReadWrite transaction. Smaller scoped statements, such as an OLTP workload, should prefer using ReadWrite transactions. Partitioned DML partitions the keyspace and runs the DML statement on each partition in separate, internal transactions. These transactions commit automatically when complete, and run independently from one another. To reduce lock contention, this execution strategy only acquires read locks on rows that match the WHERE clause of the statement. Additionally, the smaller per-partition transactions hold locks for less time. That said, Partitioned DML is not a drop-in replacement for standard DML used in ReadWrite transactions. - The DML statement must be fully-partitionable. Specifically, the statement must be expressible as the union of many statements which each access only a single row of the table. - The statement is not applied atomically to all rows of the table. Rather, the statement is applied atomically to partitions of the table, in independent transactions. Secondary index rows are updated atomically with the base table rows. - Partitioned DML does not guarantee exactly-once execution semantics against a partition. The statement is applied at least once to each partition. It is strongly recommended that the DML statement should be idempotent to avoid unexpected results. For instance, it is potentially dangerous to run a statement such as `UPDATE table SET column = column + 1` as it could be run multiple times against some rows. - The partitions are committed automatically - there is no support for Commit or Rollback. If the call returns an error, or if the client issuing the ExecuteSql call dies, it is possible that some rows had the statement executed on them successfully. It is also possible that statement was never executed against other rows. - Partitioned DML transactions may only contain the execution of a single DML statement via ExecuteSql or ExecuteStreamingSql. - If any error is encountered during the execution of the partitioned DML operation (for instance, a UNIQUE INDEX violation, division by zero, or a value that can't be stored due to schema constraints), then the operation is stopped at that point and an error is returned. It is possible that at this point, some partitions have been committed (or even committed multiple times), and other partitions have not been run at all. Given the above, Partitioned DML is good fit for large, database-wide, operations that are idempotent, such as deleting old rows from a very large table. # Begin a new transaction and execute this read or SQL query in it. The transaction ID of the new transaction is returned in ResultSetMetadata.transaction, which is a Transaction. + "begin": { # Options to use for transactions. # Begin a new transaction and execute this read or SQL query in it. The transaction ID of the new transaction is returned in ResultSetMetadata.transaction, which is a Transaction. "excludeTxnFromChangeStreams": True or False, # When `exclude_txn_from_change_streams` is set to `true`, it prevents read or write transactions from being tracked in change streams. * If the DDL option `allow_txn_exclusion` is set to `true`, then the updates made within this transaction aren't recorded in the change stream. * If you don't set the DDL option `allow_txn_exclusion` or if it's set to `false`, then the updates made within this transaction are recorded in the change stream. When `exclude_txn_from_change_streams` is set to `false` or not set, modifications from this transaction are recorded in all change streams that are tracking columns modified by these transactions. The `exclude_txn_from_change_streams` option can only be specified for read-write or partitioned DML transactions, otherwise the API returns an `INVALID_ARGUMENT` error. "isolationLevel": "A String", # Isolation level for the transaction. "partitionedDml": { # Message type to initiate a Partitioned DML transaction. # Partitioned DML transaction. Authorization to begin a Partitioned DML transaction requires `spanner.databases.beginPartitionedDmlTransaction` permission on the `session` resource. @@ -1876,7 +1876,7 @@

Method Details

}, }, "id": "A String", # Execute the read or SQL query in a previously-started transaction. - "singleUse": { # Transactions: Each session can have at most one active transaction at a time (note that standalone reads and queries use a transaction internally and do count towards the one transaction limit). After the active transaction is completed, the session can immediately be re-used for the next transaction. It is not necessary to create a new session for each transaction. Transaction modes: Cloud Spanner supports three transaction modes: 1. Locking read-write. This type of transaction is the only way to write data into Cloud Spanner. These transactions rely on pessimistic locking and, if necessary, two-phase commit. Locking read-write transactions may abort, requiring the application to retry. 2. Snapshot read-only. Snapshot read-only transactions provide guaranteed consistency across several reads, but do not allow writes. Snapshot read-only transactions can be configured to read at timestamps in the past, or configured to perform a strong read (where Spanner selects a timestamp such that the read is guaranteed to see the effects of all transactions that have committed before the start of the read). Snapshot read-only transactions do not need to be committed. Queries on change streams must be performed with the snapshot read-only transaction mode, specifying a strong read. See TransactionOptions.ReadOnly.strong for more details. 3. Partitioned DML. This type of transaction is used to execute a single Partitioned DML statement. Partitioned DML partitions the key space and runs the DML statement over each partition in parallel using separate, internal transactions that commit independently. Partitioned DML transactions do not need to be committed. For transactions that only read, snapshot read-only transactions provide simpler semantics and are almost always faster. In particular, read-only transactions do not take locks, so they do not conflict with read-write transactions. As a consequence of not taking locks, they also do not abort, so retry loops are not needed. Transactions may only read-write data in a single database. They may, however, read-write data in different tables within that database. Locking read-write transactions: Locking transactions may be used to atomically read-modify-write data anywhere in a database. This type of transaction is externally consistent. Clients should attempt to minimize the amount of time a transaction is active. Faster transactions commit with higher probability and cause less contention. Cloud Spanner attempts to keep read locks active as long as the transaction continues to do reads, and the transaction has not been terminated by Commit or Rollback. Long periods of inactivity at the client may cause Cloud Spanner to release a transaction's locks and abort it. Conceptually, a read-write transaction consists of zero or more reads or SQL statements followed by Commit. At any time before Commit, the client can send a Rollback request to abort the transaction. Semantics: Cloud Spanner can commit the transaction if all read locks it acquired are still valid at commit time, and it is able to acquire write locks for all writes. Cloud Spanner can abort the transaction for any reason. If a commit attempt returns `ABORTED`, Cloud Spanner guarantees that the transaction has not modified any user data in Cloud Spanner. Unless the transaction commits, Cloud Spanner makes no guarantees about how long the transaction's locks were held for. It is an error to use Cloud Spanner locks for any sort of mutual exclusion other than between Cloud Spanner transactions themselves. Retrying aborted transactions: When a transaction aborts, the application can choose to retry the whole transaction again. To maximize the chances of successfully committing the retry, the client should execute the retry in the same session as the original attempt. The original session's lock priority increases with each consecutive abort, meaning that each attempt has a slightly better chance of success than the previous. Note that the lock priority is preserved per session (not per transaction). Lock priority is set by the first read or write in the first attempt of a read-write transaction. If the application starts a new session to retry the whole transaction, the transaction loses its original lock priority. Moreover, the lock priority is only preserved if the transaction fails with an `ABORTED` error. Under some circumstances (for example, many transactions attempting to modify the same row(s)), a transaction can abort many times in a short period before successfully committing. Thus, it is not a good idea to cap the number of retries a transaction can attempt; instead, it is better to limit the total amount of time spent retrying. Idle transactions: A transaction is considered idle if it has no outstanding reads or SQL queries and has not started a read or SQL query within the last 10 seconds. Idle transactions can be aborted by Cloud Spanner so that they don't hold on to locks indefinitely. If an idle transaction is aborted, the commit fails with error `ABORTED`. If this behavior is undesirable, periodically executing a simple SQL query in the transaction (for example, `SELECT 1`) prevents the transaction from becoming idle. Snapshot read-only transactions: Snapshot read-only transactions provides a simpler method than locking read-write transactions for doing several consistent reads. However, this type of transaction does not support writes. Snapshot transactions do not take locks. Instead, they work by choosing a Cloud Spanner timestamp, then executing all reads at that timestamp. Since they do not acquire locks, they do not block concurrent read-write transactions. Unlike locking read-write transactions, snapshot read-only transactions never abort. They can fail if the chosen read timestamp is garbage collected; however, the default garbage collection policy is generous enough that most applications do not need to worry about this in practice. Snapshot read-only transactions do not need to call Commit or Rollback (and in fact are not permitted to do so). To execute a snapshot transaction, the client specifies a timestamp bound, which tells Cloud Spanner how to choose a read timestamp. The types of timestamp bound are: - Strong (the default). - Bounded staleness. - Exact staleness. If the Cloud Spanner database to be read is geographically distributed, stale read-only transactions can execute more quickly than strong or read-write transactions, because they are able to execute far from the leader replica. Each type of timestamp bound is discussed in detail below. Strong: Strong reads are guaranteed to see the effects of all transactions that have committed before the start of the read. Furthermore, all rows yielded by a single read are consistent with each other -- if any part of the read observes a transaction, all parts of the read see the transaction. Strong reads are not repeatable: two consecutive strong read-only transactions might return inconsistent results if there are concurrent writes. If consistency across reads is required, the reads should be executed within a transaction or at an exact read timestamp. Queries on change streams (see below for more details) must also specify the strong read timestamp bound. See TransactionOptions.ReadOnly.strong. Exact staleness: These timestamp bounds execute reads at a user-specified timestamp. Reads at a timestamp are guaranteed to see a consistent prefix of the global transaction history: they observe modifications done by all transactions with a commit timestamp less than or equal to the read timestamp, and observe none of the modifications done by transactions with a larger commit timestamp. They block until all conflicting transactions that can be assigned commit timestamps <= the read timestamp have finished. The timestamp can either be expressed as an absolute Cloud Spanner commit timestamp or a staleness relative to the current time. These modes do not require a "negotiation phase" to pick a timestamp. As a result, they execute slightly faster than the equivalent boundedly stale concurrency modes. On the other hand, boundedly stale reads usually return fresher results. See TransactionOptions.ReadOnly.read_timestamp and TransactionOptions.ReadOnly.exact_staleness. Bounded staleness: Bounded staleness modes allow Cloud Spanner to pick the read timestamp, subject to a user-provided staleness bound. Cloud Spanner chooses the newest timestamp within the staleness bound that allows execution of the reads at the closest available replica without blocking. All rows yielded are consistent with each other -- if any part of the read observes a transaction, all parts of the read see the transaction. Boundedly stale reads are not repeatable: two stale reads, even if they use the same staleness bound, can execute at different timestamps and thus return inconsistent results. Boundedly stale reads execute in two phases: the first phase negotiates a timestamp among all replicas needed to serve the read. In the second phase, reads are executed at the negotiated timestamp. As a result of the two phase execution, bounded staleness reads are usually a little slower than comparable exact staleness reads. However, they are typically able to return fresher results, and are more likely to execute at the closest replica. Because the timestamp negotiation requires up-front knowledge of which rows are read, it can only be used with single-use read-only transactions. See TransactionOptions.ReadOnly.max_staleness and TransactionOptions.ReadOnly.min_read_timestamp. Old read timestamps and garbage collection: Cloud Spanner continuously garbage collects deleted and overwritten data in the background to reclaim storage space. This process is known as "version GC". By default, version GC reclaims versions after they are one hour old. Because of this, Cloud Spanner can't perform reads at read timestamps more than one hour in the past. This restriction also applies to in-progress reads and/or SQL queries whose timestamp become too old while executing. Reads and SQL queries with too-old read timestamps fail with the error `FAILED_PRECONDITION`. You can configure and extend the `VERSION_RETENTION_PERIOD` of a database up to a period as long as one week, which allows Cloud Spanner to perform reads up to one week in the past. Querying change Streams: A Change Stream is a schema object that can be configured to watch data changes on the entire database, a set of tables, or a set of columns in a database. When a change stream is created, Spanner automatically defines a corresponding SQL Table-Valued Function (TVF) that can be used to query the change records in the associated change stream using the ExecuteStreamingSql API. The name of the TVF for a change stream is generated from the name of the change stream: READ_. All queries on change stream TVFs must be executed using the ExecuteStreamingSql API with a single-use read-only transaction with a strong read-only timestamp_bound. The change stream TVF allows users to specify the start_timestamp and end_timestamp for the time range of interest. All change records within the retention period is accessible using the strong read-only timestamp_bound. All other TransactionOptions are invalid for change stream queries. In addition, if TransactionOptions.read_only.return_read_timestamp is set to true, a special value of 2^63 - 2 is returned in the Transaction message that describes the transaction, instead of a valid read timestamp. This special value should be discarded and not used for any subsequent queries. Please see https://cloud.google.com/spanner/docs/change-streams for more details on how to query the change stream TVFs. Partitioned DML transactions: Partitioned DML transactions are used to execute DML statements with a different execution strategy that provides different, and often better, scalability properties for large, table-wide operations than DML in a ReadWrite transaction. Smaller scoped statements, such as an OLTP workload, should prefer using ReadWrite transactions. Partitioned DML partitions the keyspace and runs the DML statement on each partition in separate, internal transactions. These transactions commit automatically when complete, and run independently from one another. To reduce lock contention, this execution strategy only acquires read locks on rows that match the WHERE clause of the statement. Additionally, the smaller per-partition transactions hold locks for less time. That said, Partitioned DML is not a drop-in replacement for standard DML used in ReadWrite transactions. - The DML statement must be fully-partitionable. Specifically, the statement must be expressible as the union of many statements which each access only a single row of the table. - The statement is not applied atomically to all rows of the table. Rather, the statement is applied atomically to partitions of the table, in independent transactions. Secondary index rows are updated atomically with the base table rows. - Partitioned DML does not guarantee exactly-once execution semantics against a partition. The statement is applied at least once to each partition. It is strongly recommended that the DML statement should be idempotent to avoid unexpected results. For instance, it is potentially dangerous to run a statement such as `UPDATE table SET column = column + 1` as it could be run multiple times against some rows. - The partitions are committed automatically - there is no support for Commit or Rollback. If the call returns an error, or if the client issuing the ExecuteSql call dies, it is possible that some rows had the statement executed on them successfully. It is also possible that statement was never executed against other rows. - Partitioned DML transactions may only contain the execution of a single DML statement via ExecuteSql or ExecuteStreamingSql. - If any error is encountered during the execution of the partitioned DML operation (for instance, a UNIQUE INDEX violation, division by zero, or a value that can't be stored due to schema constraints), then the operation is stopped at that point and an error is returned. It is possible that at this point, some partitions have been committed (or even committed multiple times), and other partitions have not been run at all. Given the above, Partitioned DML is good fit for large, database-wide, operations that are idempotent, such as deleting old rows from a very large table. # Execute the read or SQL query in a temporary transaction. This is the most efficient way to execute a transaction that consists of a single SQL query. + "singleUse": { # Options to use for transactions. # Execute the read or SQL query in a temporary transaction. This is the most efficient way to execute a transaction that consists of a single SQL query. "excludeTxnFromChangeStreams": True or False, # When `exclude_txn_from_change_streams` is set to `true`, it prevents read or write transactions from being tracked in change streams. * If the DDL option `allow_txn_exclusion` is set to `true`, then the updates made within this transaction aren't recorded in the change stream. * If you don't set the DDL option `allow_txn_exclusion` or if it's set to `false`, then the updates made within this transaction are recorded in the change stream. When `exclude_txn_from_change_streams` is set to `false` or not set, modifications from this transaction are recorded in all change streams that are tracking columns modified by these transactions. The `exclude_txn_from_change_streams` option can only be specified for read-write or partitioned DML transactions, otherwise the API returns an `INVALID_ARGUMENT` error. "isolationLevel": "A String", # Isolation level for the transaction. "partitionedDml": { # Message type to initiate a Partitioned DML transaction. # Partitioned DML transaction. Authorization to begin a Partitioned DML transaction requires `spanner.databases.beginPartitionedDmlTransaction` permission on the `session` resource. diff --git a/docs/dyn/storage_v1.objects.html b/docs/dyn/storage_v1.objects.html index 3910a314f6..9b8b4237ce 100644 --- a/docs/dyn/storage_v1.objects.html +++ b/docs/dyn/storage_v1.objects.html @@ -102,7 +102,7 @@

Instance Methods

insert(bucket, body=None, contentEncoding=None, ifGenerationMatch=None, ifGenerationNotMatch=None, ifMetagenerationMatch=None, ifMetagenerationNotMatch=None, kmsKeyName=None, media_body=None, media_mime_type=None, name=None, predefinedAcl=None, projection=None, userProject=None)

Stores a new object and metadata.

- list(bucket, delimiter=None, endOffset=None, includeFoldersAsPrefixes=None, includeTrailingDelimiter=None, matchGlob=None, maxResults=None, pageToken=None, prefix=None, projection=None, softDeleted=None, startOffset=None, userProject=None, versions=None)

+ list(bucket, delimiter=None, endOffset=None, filter=None, includeFoldersAsPrefixes=None, includeTrailingDelimiter=None, matchGlob=None, maxResults=None, pageToken=None, prefix=None, projection=None, softDeleted=None, startOffset=None, userProject=None, versions=None)

Retrieves a list of objects matching the criteria.

list_next()

@@ -233,6 +233,15 @@

Method Details

"contentEncoding": "A String", # Content-Encoding of the object data. "contentLanguage": "A String", # Content-Language of the object data. "contentType": "A String", # Content-Type of the object data. If an object is stored without a Content-Type, it is served as application/octet-stream. + "contexts": { # User-defined or system-defined object contexts. Each object context is a key-payload pair, where the key provides the identification and the payload holds the associated value and additional metadata. + "custom": { # User-defined object contexts. + "a_key": { # The payload of a single user-defined object context. # A single user-defined object context. + "createTime": "A String", # The time at which the object context was created in RFC 3339 format. + "updateTime": "A String", # The time at which the object context was last updated in RFC 3339 format. + "value": "A String", # The value of the object context. + }, + }, + }, "crc32c": "A String", # CRC32c checksum, as described in RFC 4960, Appendix B; encoded using base64 in big-endian byte order. For more information about using the CRC32c checksum, see [Data Validation and Change Detection](https://cloud.google.com/storage/docs/data-validation). "customTime": "A String", # A timestamp in RFC 3339 format specified by the user for an object. "customerEncryption": { # Metadata of customer-supplied encryption key, if the object is encrypted by such a key. @@ -341,6 +350,15 @@

Method Details

"contentEncoding": "A String", # Content-Encoding of the object data. "contentLanguage": "A String", # Content-Language of the object data. "contentType": "A String", # Content-Type of the object data. If an object is stored without a Content-Type, it is served as application/octet-stream. + "contexts": { # User-defined or system-defined object contexts. Each object context is a key-payload pair, where the key provides the identification and the payload holds the associated value and additional metadata. + "custom": { # User-defined object contexts. + "a_key": { # The payload of a single user-defined object context. # A single user-defined object context. + "createTime": "A String", # The time at which the object context was created in RFC 3339 format. + "updateTime": "A String", # The time at which the object context was last updated in RFC 3339 format. + "value": "A String", # The value of the object context. + }, + }, + }, "crc32c": "A String", # CRC32c checksum, as described in RFC 4960, Appendix B; encoded using base64 in big-endian byte order. For more information about using the CRC32c checksum, see [Data Validation and Change Detection](https://cloud.google.com/storage/docs/data-validation). "customTime": "A String", # A timestamp in RFC 3339 format specified by the user for an object. "customerEncryption": { # Metadata of customer-supplied encryption key, if the object is encrypted by such a key. @@ -435,6 +453,15 @@

Method Details

"contentEncoding": "A String", # Content-Encoding of the object data. "contentLanguage": "A String", # Content-Language of the object data. "contentType": "A String", # Content-Type of the object data. If an object is stored without a Content-Type, it is served as application/octet-stream. + "contexts": { # User-defined or system-defined object contexts. Each object context is a key-payload pair, where the key provides the identification and the payload holds the associated value and additional metadata. + "custom": { # User-defined object contexts. + "a_key": { # The payload of a single user-defined object context. # A single user-defined object context. + "createTime": "A String", # The time at which the object context was created in RFC 3339 format. + "updateTime": "A String", # The time at which the object context was last updated in RFC 3339 format. + "value": "A String", # The value of the object context. + }, + }, + }, "crc32c": "A String", # CRC32c checksum, as described in RFC 4960, Appendix B; encoded using base64 in big-endian byte order. For more information about using the CRC32c checksum, see [Data Validation and Change Detection](https://cloud.google.com/storage/docs/data-validation). "customTime": "A String", # A timestamp in RFC 3339 format specified by the user for an object. "customerEncryption": { # Metadata of customer-supplied encryption key, if the object is encrypted by such a key. @@ -543,6 +570,15 @@

Method Details

"contentEncoding": "A String", # Content-Encoding of the object data. "contentLanguage": "A String", # Content-Language of the object data. "contentType": "A String", # Content-Type of the object data. If an object is stored without a Content-Type, it is served as application/octet-stream. + "contexts": { # User-defined or system-defined object contexts. Each object context is a key-payload pair, where the key provides the identification and the payload holds the associated value and additional metadata. + "custom": { # User-defined object contexts. + "a_key": { # The payload of a single user-defined object context. # A single user-defined object context. + "createTime": "A String", # The time at which the object context was created in RFC 3339 format. + "updateTime": "A String", # The time at which the object context was last updated in RFC 3339 format. + "value": "A String", # The value of the object context. + }, + }, + }, "crc32c": "A String", # CRC32c checksum, as described in RFC 4960, Appendix B; encoded using base64 in big-endian byte order. For more information about using the CRC32c checksum, see [Data Validation and Change Detection](https://cloud.google.com/storage/docs/data-validation). "customTime": "A String", # A timestamp in RFC 3339 format specified by the user for an object. "customerEncryption": { # Metadata of customer-supplied encryption key, if the object is encrypted by such a key. @@ -664,6 +700,15 @@

Method Details

"contentEncoding": "A String", # Content-Encoding of the object data. "contentLanguage": "A String", # Content-Language of the object data. "contentType": "A String", # Content-Type of the object data. If an object is stored without a Content-Type, it is served as application/octet-stream. + "contexts": { # User-defined or system-defined object contexts. Each object context is a key-payload pair, where the key provides the identification and the payload holds the associated value and additional metadata. + "custom": { # User-defined object contexts. + "a_key": { # The payload of a single user-defined object context. # A single user-defined object context. + "createTime": "A String", # The time at which the object context was created in RFC 3339 format. + "updateTime": "A String", # The time at which the object context was last updated in RFC 3339 format. + "value": "A String", # The value of the object context. + }, + }, + }, "crc32c": "A String", # CRC32c checksum, as described in RFC 4960, Appendix B; encoded using base64 in big-endian byte order. For more information about using the CRC32c checksum, see [Data Validation and Change Detection](https://cloud.google.com/storage/docs/data-validation). "customTime": "A String", # A timestamp in RFC 3339 format specified by the user for an object. "customerEncryption": { # Metadata of customer-supplied encryption key, if the object is encrypted by such a key. @@ -835,6 +880,15 @@

Method Details

"contentEncoding": "A String", # Content-Encoding of the object data. "contentLanguage": "A String", # Content-Language of the object data. "contentType": "A String", # Content-Type of the object data. If an object is stored without a Content-Type, it is served as application/octet-stream. + "contexts": { # User-defined or system-defined object contexts. Each object context is a key-payload pair, where the key provides the identification and the payload holds the associated value and additional metadata. + "custom": { # User-defined object contexts. + "a_key": { # The payload of a single user-defined object context. # A single user-defined object context. + "createTime": "A String", # The time at which the object context was created in RFC 3339 format. + "updateTime": "A String", # The time at which the object context was last updated in RFC 3339 format. + "value": "A String", # The value of the object context. + }, + }, + }, "crc32c": "A String", # CRC32c checksum, as described in RFC 4960, Appendix B; encoded using base64 in big-endian byte order. For more information about using the CRC32c checksum, see [Data Validation and Change Detection](https://cloud.google.com/storage/docs/data-validation). "customTime": "A String", # A timestamp in RFC 3339 format specified by the user for an object. "customerEncryption": { # Metadata of customer-supplied encryption key, if the object is encrypted by such a key. @@ -942,6 +996,15 @@

Method Details

"contentEncoding": "A String", # Content-Encoding of the object data. "contentLanguage": "A String", # Content-Language of the object data. "contentType": "A String", # Content-Type of the object data. If an object is stored without a Content-Type, it is served as application/octet-stream. + "contexts": { # User-defined or system-defined object contexts. Each object context is a key-payload pair, where the key provides the identification and the payload holds the associated value and additional metadata. + "custom": { # User-defined object contexts. + "a_key": { # The payload of a single user-defined object context. # A single user-defined object context. + "createTime": "A String", # The time at which the object context was created in RFC 3339 format. + "updateTime": "A String", # The time at which the object context was last updated in RFC 3339 format. + "value": "A String", # The value of the object context. + }, + }, + }, "crc32c": "A String", # CRC32c checksum, as described in RFC 4960, Appendix B; encoded using base64 in big-endian byte order. For more information about using the CRC32c checksum, see [Data Validation and Change Detection](https://cloud.google.com/storage/docs/data-validation). "customTime": "A String", # A timestamp in RFC 3339 format specified by the user for an object. "customerEncryption": { # Metadata of customer-supplied encryption key, if the object is encrypted by such a key. @@ -986,13 +1049,14 @@

Method Details

- list(bucket, delimiter=None, endOffset=None, includeFoldersAsPrefixes=None, includeTrailingDelimiter=None, matchGlob=None, maxResults=None, pageToken=None, prefix=None, projection=None, softDeleted=None, startOffset=None, userProject=None, versions=None) + list(bucket, delimiter=None, endOffset=None, filter=None, includeFoldersAsPrefixes=None, includeTrailingDelimiter=None, matchGlob=None, maxResults=None, pageToken=None, prefix=None, projection=None, softDeleted=None, startOffset=None, userProject=None, versions=None)
Retrieves a list of objects matching the criteria.
 
 Args:
   bucket: string, Name of the bucket in which to look for objects. (required)
   delimiter: string, Returns results in a directory-like mode. items will contain only objects whose names, aside from the prefix, do not contain delimiter. Objects whose names, aside from the prefix, contain delimiter will have their name, truncated after the delimiter, returned in prefixes. Duplicate prefixes are omitted.
   endOffset: string, Filter results to objects whose names are lexicographically before endOffset. If startOffset is also set, the objects listed will have names between startOffset (inclusive) and endOffset (exclusive).
+  filter: string, Filter the returned objects. Currently only supported for the contexts field. If delimiter is set, the returned prefixes are exempt from this filter.
   includeFoldersAsPrefixes: boolean, Only applicable if delimiter is set to '/'. If true, will also include folders and managed folders (besides objects) in the returned prefixes.
   includeTrailingDelimiter: boolean, If true, objects that end in exactly one instance of delimiter will have their metadata included in items in addition to prefixes.
   matchGlob: string, Filter results to objects and prefixes that match this glob pattern.
@@ -1052,6 +1116,15 @@ 

Method Details

"contentEncoding": "A String", # Content-Encoding of the object data. "contentLanguage": "A String", # Content-Language of the object data. "contentType": "A String", # Content-Type of the object data. If an object is stored without a Content-Type, it is served as application/octet-stream. + "contexts": { # User-defined or system-defined object contexts. Each object context is a key-payload pair, where the key provides the identification and the payload holds the associated value and additional metadata. + "custom": { # User-defined object contexts. + "a_key": { # The payload of a single user-defined object context. # A single user-defined object context. + "createTime": "A String", # The time at which the object context was created in RFC 3339 format. + "updateTime": "A String", # The time at which the object context was last updated in RFC 3339 format. + "value": "A String", # The value of the object context. + }, + }, + }, "crc32c": "A String", # CRC32c checksum, as described in RFC 4960, Appendix B; encoded using base64 in big-endian byte order. For more information about using the CRC32c checksum, see [Data Validation and Change Detection](https://cloud.google.com/storage/docs/data-validation). "customTime": "A String", # A timestamp in RFC 3339 format specified by the user for an object. "customerEncryption": { # Metadata of customer-supplied encryption key, if the object is encrypted by such a key. @@ -1180,6 +1253,15 @@

Method Details

"contentEncoding": "A String", # Content-Encoding of the object data. "contentLanguage": "A String", # Content-Language of the object data. "contentType": "A String", # Content-Type of the object data. If an object is stored without a Content-Type, it is served as application/octet-stream. + "contexts": { # User-defined or system-defined object contexts. Each object context is a key-payload pair, where the key provides the identification and the payload holds the associated value and additional metadata. + "custom": { # User-defined object contexts. + "a_key": { # The payload of a single user-defined object context. # A single user-defined object context. + "createTime": "A String", # The time at which the object context was created in RFC 3339 format. + "updateTime": "A String", # The time at which the object context was last updated in RFC 3339 format. + "value": "A String", # The value of the object context. + }, + }, + }, "crc32c": "A String", # CRC32c checksum, as described in RFC 4960, Appendix B; encoded using base64 in big-endian byte order. For more information about using the CRC32c checksum, see [Data Validation and Change Detection](https://cloud.google.com/storage/docs/data-validation). "customTime": "A String", # A timestamp in RFC 3339 format specified by the user for an object. "customerEncryption": { # Metadata of customer-supplied encryption key, if the object is encrypted by such a key. @@ -1272,6 +1354,15 @@

Method Details

"contentEncoding": "A String", # Content-Encoding of the object data. "contentLanguage": "A String", # Content-Language of the object data. "contentType": "A String", # Content-Type of the object data. If an object is stored without a Content-Type, it is served as application/octet-stream. + "contexts": { # User-defined or system-defined object contexts. Each object context is a key-payload pair, where the key provides the identification and the payload holds the associated value and additional metadata. + "custom": { # User-defined object contexts. + "a_key": { # The payload of a single user-defined object context. # A single user-defined object context. + "createTime": "A String", # The time at which the object context was created in RFC 3339 format. + "updateTime": "A String", # The time at which the object context was last updated in RFC 3339 format. + "value": "A String", # The value of the object context. + }, + }, + }, "crc32c": "A String", # CRC32c checksum, as described in RFC 4960, Appendix B; encoded using base64 in big-endian byte order. For more information about using the CRC32c checksum, see [Data Validation and Change Detection](https://cloud.google.com/storage/docs/data-validation). "customTime": "A String", # A timestamp in RFC 3339 format specified by the user for an object. "customerEncryption": { # Metadata of customer-supplied encryption key, if the object is encrypted by such a key. @@ -1376,6 +1467,15 @@

Method Details

"contentEncoding": "A String", # Content-Encoding of the object data. "contentLanguage": "A String", # Content-Language of the object data. "contentType": "A String", # Content-Type of the object data. If an object is stored without a Content-Type, it is served as application/octet-stream. + "contexts": { # User-defined or system-defined object contexts. Each object context is a key-payload pair, where the key provides the identification and the payload holds the associated value and additional metadata. + "custom": { # User-defined object contexts. + "a_key": { # The payload of a single user-defined object context. # A single user-defined object context. + "createTime": "A String", # The time at which the object context was created in RFC 3339 format. + "updateTime": "A String", # The time at which the object context was last updated in RFC 3339 format. + "value": "A String", # The value of the object context. + }, + }, + }, "crc32c": "A String", # CRC32c checksum, as described in RFC 4960, Appendix B; encoded using base64 in big-endian byte order. For more information about using the CRC32c checksum, see [Data Validation and Change Detection](https://cloud.google.com/storage/docs/data-validation). "customTime": "A String", # A timestamp in RFC 3339 format specified by the user for an object. "customerEncryption": { # Metadata of customer-supplied encryption key, if the object is encrypted by such a key. @@ -1481,6 +1581,15 @@

Method Details

"contentEncoding": "A String", # Content-Encoding of the object data. "contentLanguage": "A String", # Content-Language of the object data. "contentType": "A String", # Content-Type of the object data. If an object is stored without a Content-Type, it is served as application/octet-stream. + "contexts": { # User-defined or system-defined object contexts. Each object context is a key-payload pair, where the key provides the identification and the payload holds the associated value and additional metadata. + "custom": { # User-defined object contexts. + "a_key": { # The payload of a single user-defined object context. # A single user-defined object context. + "createTime": "A String", # The time at which the object context was created in RFC 3339 format. + "updateTime": "A String", # The time at which the object context was last updated in RFC 3339 format. + "value": "A String", # The value of the object context. + }, + }, + }, "crc32c": "A String", # CRC32c checksum, as described in RFC 4960, Appendix B; encoded using base64 in big-endian byte order. For more information about using the CRC32c checksum, see [Data Validation and Change Detection](https://cloud.google.com/storage/docs/data-validation). "customTime": "A String", # A timestamp in RFC 3339 format specified by the user for an object. "customerEncryption": { # Metadata of customer-supplied encryption key, if the object is encrypted by such a key. @@ -1575,6 +1684,15 @@

Method Details

"contentEncoding": "A String", # Content-Encoding of the object data. "contentLanguage": "A String", # Content-Language of the object data. "contentType": "A String", # Content-Type of the object data. If an object is stored without a Content-Type, it is served as application/octet-stream. + "contexts": { # User-defined or system-defined object contexts. Each object context is a key-payload pair, where the key provides the identification and the payload holds the associated value and additional metadata. + "custom": { # User-defined object contexts. + "a_key": { # The payload of a single user-defined object context. # A single user-defined object context. + "createTime": "A String", # The time at which the object context was created in RFC 3339 format. + "updateTime": "A String", # The time at which the object context was last updated in RFC 3339 format. + "value": "A String", # The value of the object context. + }, + }, + }, "crc32c": "A String", # CRC32c checksum, as described in RFC 4960, Appendix B; encoded using base64 in big-endian byte order. For more information about using the CRC32c checksum, see [Data Validation and Change Detection](https://cloud.google.com/storage/docs/data-validation). "customTime": "A String", # A timestamp in RFC 3339 format specified by the user for an object. "customerEncryption": { # Metadata of customer-supplied encryption key, if the object is encrypted by such a key. @@ -1689,6 +1807,15 @@

Method Details

"contentEncoding": "A String", # Content-Encoding of the object data. "contentLanguage": "A String", # Content-Language of the object data. "contentType": "A String", # Content-Type of the object data. If an object is stored without a Content-Type, it is served as application/octet-stream. + "contexts": { # User-defined or system-defined object contexts. Each object context is a key-payload pair, where the key provides the identification and the payload holds the associated value and additional metadata. + "custom": { # User-defined object contexts. + "a_key": { # The payload of a single user-defined object context. # A single user-defined object context. + "createTime": "A String", # The time at which the object context was created in RFC 3339 format. + "updateTime": "A String", # The time at which the object context was last updated in RFC 3339 format. + "value": "A String", # The value of the object context. + }, + }, + }, "crc32c": "A String", # CRC32c checksum, as described in RFC 4960, Appendix B; encoded using base64 in big-endian byte order. For more information about using the CRC32c checksum, see [Data Validation and Change Detection](https://cloud.google.com/storage/docs/data-validation). "customTime": "A String", # A timestamp in RFC 3339 format specified by the user for an object. "customerEncryption": { # Metadata of customer-supplied encryption key, if the object is encrypted by such a key. @@ -1923,6 +2050,15 @@

Method Details

"contentEncoding": "A String", # Content-Encoding of the object data. "contentLanguage": "A String", # Content-Language of the object data. "contentType": "A String", # Content-Type of the object data. If an object is stored without a Content-Type, it is served as application/octet-stream. + "contexts": { # User-defined or system-defined object contexts. Each object context is a key-payload pair, where the key provides the identification and the payload holds the associated value and additional metadata. + "custom": { # User-defined object contexts. + "a_key": { # The payload of a single user-defined object context. # A single user-defined object context. + "createTime": "A String", # The time at which the object context was created in RFC 3339 format. + "updateTime": "A String", # The time at which the object context was last updated in RFC 3339 format. + "value": "A String", # The value of the object context. + }, + }, + }, "crc32c": "A String", # CRC32c checksum, as described in RFC 4960, Appendix B; encoded using base64 in big-endian byte order. For more information about using the CRC32c checksum, see [Data Validation and Change Detection](https://cloud.google.com/storage/docs/data-validation). "customTime": "A String", # A timestamp in RFC 3339 format specified by the user for an object. "customerEncryption": { # Metadata of customer-supplied encryption key, if the object is encrypted by such a key. @@ -2027,6 +2163,15 @@

Method Details

"contentEncoding": "A String", # Content-Encoding of the object data. "contentLanguage": "A String", # Content-Language of the object data. "contentType": "A String", # Content-Type of the object data. If an object is stored without a Content-Type, it is served as application/octet-stream. + "contexts": { # User-defined or system-defined object contexts. Each object context is a key-payload pair, where the key provides the identification and the payload holds the associated value and additional metadata. + "custom": { # User-defined object contexts. + "a_key": { # The payload of a single user-defined object context. # A single user-defined object context. + "createTime": "A String", # The time at which the object context was created in RFC 3339 format. + "updateTime": "A String", # The time at which the object context was last updated in RFC 3339 format. + "value": "A String", # The value of the object context. + }, + }, + }, "crc32c": "A String", # CRC32c checksum, as described in RFC 4960, Appendix B; encoded using base64 in big-endian byte order. For more information about using the CRC32c checksum, see [Data Validation and Change Detection](https://cloud.google.com/storage/docs/data-validation). "customTime": "A String", # A timestamp in RFC 3339 format specified by the user for an object. "customerEncryption": { # Metadata of customer-supplied encryption key, if the object is encrypted by such a key. diff --git a/docs/dyn/workspaceevents_v1.subscriptions.html b/docs/dyn/workspaceevents_v1.subscriptions.html index 640a2b4db0..cab66c5952 100644 --- a/docs/dyn/workspaceevents_v1.subscriptions.html +++ b/docs/dyn/workspaceevents_v1.subscriptions.html @@ -340,7 +340,7 @@

Method Details

"updateTime": "A String", # Output only. The last time that the subscription is updated. } - updateMask: string, Optional. The field to update. If omitted, updates any fields included in the request. You can update one of the following fields in a subscription: * `expire_time`: The timestamp when the subscription expires. * `ttl`: The time-to-live (TTL) or duration of the subscription. * `event_types`: The list of event types to receive about the target resource. To fully replace the subscription (the equivalent of `PUT`), use `*`. Any omitted fields are updated with empty values. + updateMask: string, Optional. The field to update. If omitted, updates any fields included in the request. You can update one of the following fields in a subscription: * `expire_time`: The timestamp when the subscription expires. * `ttl`: The time-to-live (TTL) or duration of the subscription. * `event_types`: The list of event types to receive about the target resource. When using the `*` wildcard (equivalent to `PUT`), omitted fields are set to empty values and rejected if they're invalid. validateOnly: boolean, Optional. If set to `true`, validates and previews the request, but doesn't update the subscription. x__xgafv: string, V1 error format. Allowed values diff --git a/docs/dyn/workstations_v1beta.projects.locations.workstationClusters.workstationConfigs.html b/docs/dyn/workstations_v1beta.projects.locations.workstationClusters.workstationConfigs.html index 2fec2e70a0..ec322bb47e 100644 --- a/docs/dyn/workstations_v1beta.projects.locations.workstationClusters.workstationConfigs.html +++ b/docs/dyn/workstations_v1beta.projects.locations.workstationClusters.workstationConfigs.html @@ -209,6 +209,13 @@

Method Details

"id": "A String", # Required. The ID to be used for the boost configuration. "machineType": "A String", # Optional. The type of machine that boosted VM instances will use—for example, `e2-standard-4`. For more information about machine types that Cloud Workstations supports, see the list of [available machine types](https://cloud.google.com/workstations/docs/available-machine-types). Defaults to `e2-standard-4`. "poolSize": 42, # Optional. The number of boost VMs that the system should keep idle so that workstations can be boosted quickly. Defaults to `0`. + "reservationAffinity": { # ReservationAffinity is the configuration of the desired reservation from which instances can consume resources. # Optional. [ReservationAffinity](https://cloud.google.com/compute/docs/instances/reserving-zonal-resources) specifies a reservation that can be consumed to create boost VM instances. If SPECIFIC_RESERVATION is specified, Cloud Workstations will only create VMs in the zone where the reservation is located. This would affect availability since the service will no longer be resilient to zonal outages. If ANY_RESERVATION is specified, creating reservations in both zones that the config creates VMs in will ensure higher availability. **Important Considerations for Reservation Affinity:** * This feature is intended for advanced users and requires familiarity with Google Compute Engine reservations. * Using reservations incurs charges, regardless of utilization. * The resources in the pool will consume the specified reservation. Take this into account when setting the pool size. + "consumeReservationType": "A String", # Optional. Corresponds to the type of reservation consumption. + "key": "A String", # Optional. Corresponds to the label key of reservation resource. + "values": [ # Optional. Corresponds to the label values of reservation resources. Valid values are either a name to a reservation in the same project or "projects/{project}/reservations/{reservation}" to target a shared reservation in the same zone but in a different project. + "A String", + ], + }, }, ], "bootDiskSizeGb": 42, # Optional. The size of the boot disk for the VM in gigabytes (GB). The minimum boot disk size is `30` GB. Defaults to `50` GB. @@ -221,6 +228,13 @@

Method Details

"machineType": "A String", # Optional. The type of machine to use for VM instances—for example, `"e2-standard-4"`. For more information about machine types that Cloud Workstations supports, see the list of [available machine types](https://cloud.google.com/workstations/docs/available-machine-types). "poolSize": 42, # Optional. The number of VMs that the system should keep idle so that new workstations can be started quickly for new users. Defaults to `0` in the API. "pooledInstances": 42, # Output only. Number of instances currently available in the pool for faster workstation startup. + "reservationAffinity": { # ReservationAffinity is the configuration of the desired reservation from which instances can consume resources. # Optional. [ReservationAffinity](https://cloud.google.com/compute/docs/instances/reserving-zonal-resources) specifies a reservation that can be consumed to create boost VM instances. If SPECIFIC_RESERVATION is specified, Cloud Workstations will only create VMs in the zone where the reservation is located. This would affect availability since the service will no longer be resilient to zonal outages. If ANY_RESERVATION is specified, creating reservations in both zones that the config creates VMs in will ensure higher availability. **Important Considerations for Reservation Affinity:** * This feature is intended for advanced users and requires familiarity with Google Compute Engine reservations. * Using reservations incurs charges, regardless of utilization. * The resources in the pool will consume the specified reservation. Take this into account when setting the pool size. + "consumeReservationType": "A String", # Optional. Corresponds to the type of reservation consumption. + "key": "A String", # Optional. Corresponds to the label key of reservation resource. + "values": [ # Optional. Corresponds to the label values of reservation resources. Valid values are either a name to a reservation in the same project or "projects/{project}/reservations/{reservation}" to target a shared reservation in the same zone but in a different project. + "A String", + ], + }, "serviceAccount": "A String", # Optional. The email address of the service account for Cloud Workstations VMs created with this configuration. When specified, be sure that the service account has `logging.logEntries.create` and `monitoring.timeSeries.create` permissions on the project so it can write logs out to Cloud Logging. If using a custom container image, the service account must have [Artifact Registry Reader](https://cloud.google.com/artifact-registry/docs/access-control#roles) permission to pull the specified image. If you as the administrator want to be able to `ssh` into the underlying VM, you need to set this value to a service account for which you have the `iam.serviceAccounts.actAs` permission. Conversely, if you don't want anyone to be able to `ssh` into the underlying VM, use a service account where no one has that permission. If not set, VMs run with a service account provided by the Cloud Workstations service, and the image must be publicly accessible. "serviceAccountScopes": [ # Optional. Scopes to grant to the service_account. When specified, users of workstations under this configuration must have `iam.serviceAccounts.actAs` on the service account. "A String", @@ -439,6 +453,13 @@

Method Details

"id": "A String", # Required. The ID to be used for the boost configuration. "machineType": "A String", # Optional. The type of machine that boosted VM instances will use—for example, `e2-standard-4`. For more information about machine types that Cloud Workstations supports, see the list of [available machine types](https://cloud.google.com/workstations/docs/available-machine-types). Defaults to `e2-standard-4`. "poolSize": 42, # Optional. The number of boost VMs that the system should keep idle so that workstations can be boosted quickly. Defaults to `0`. + "reservationAffinity": { # ReservationAffinity is the configuration of the desired reservation from which instances can consume resources. # Optional. [ReservationAffinity](https://cloud.google.com/compute/docs/instances/reserving-zonal-resources) specifies a reservation that can be consumed to create boost VM instances. If SPECIFIC_RESERVATION is specified, Cloud Workstations will only create VMs in the zone where the reservation is located. This would affect availability since the service will no longer be resilient to zonal outages. If ANY_RESERVATION is specified, creating reservations in both zones that the config creates VMs in will ensure higher availability. **Important Considerations for Reservation Affinity:** * This feature is intended for advanced users and requires familiarity with Google Compute Engine reservations. * Using reservations incurs charges, regardless of utilization. * The resources in the pool will consume the specified reservation. Take this into account when setting the pool size. + "consumeReservationType": "A String", # Optional. Corresponds to the type of reservation consumption. + "key": "A String", # Optional. Corresponds to the label key of reservation resource. + "values": [ # Optional. Corresponds to the label values of reservation resources. Valid values are either a name to a reservation in the same project or "projects/{project}/reservations/{reservation}" to target a shared reservation in the same zone but in a different project. + "A String", + ], + }, }, ], "bootDiskSizeGb": 42, # Optional. The size of the boot disk for the VM in gigabytes (GB). The minimum boot disk size is `30` GB. Defaults to `50` GB. @@ -451,6 +472,13 @@

Method Details

"machineType": "A String", # Optional. The type of machine to use for VM instances—for example, `"e2-standard-4"`. For more information about machine types that Cloud Workstations supports, see the list of [available machine types](https://cloud.google.com/workstations/docs/available-machine-types). "poolSize": 42, # Optional. The number of VMs that the system should keep idle so that new workstations can be started quickly for new users. Defaults to `0` in the API. "pooledInstances": 42, # Output only. Number of instances currently available in the pool for faster workstation startup. + "reservationAffinity": { # ReservationAffinity is the configuration of the desired reservation from which instances can consume resources. # Optional. [ReservationAffinity](https://cloud.google.com/compute/docs/instances/reserving-zonal-resources) specifies a reservation that can be consumed to create boost VM instances. If SPECIFIC_RESERVATION is specified, Cloud Workstations will only create VMs in the zone where the reservation is located. This would affect availability since the service will no longer be resilient to zonal outages. If ANY_RESERVATION is specified, creating reservations in both zones that the config creates VMs in will ensure higher availability. **Important Considerations for Reservation Affinity:** * This feature is intended for advanced users and requires familiarity with Google Compute Engine reservations. * Using reservations incurs charges, regardless of utilization. * The resources in the pool will consume the specified reservation. Take this into account when setting the pool size. + "consumeReservationType": "A String", # Optional. Corresponds to the type of reservation consumption. + "key": "A String", # Optional. Corresponds to the label key of reservation resource. + "values": [ # Optional. Corresponds to the label values of reservation resources. Valid values are either a name to a reservation in the same project or "projects/{project}/reservations/{reservation}" to target a shared reservation in the same zone but in a different project. + "A String", + ], + }, "serviceAccount": "A String", # Optional. The email address of the service account for Cloud Workstations VMs created with this configuration. When specified, be sure that the service account has `logging.logEntries.create` and `monitoring.timeSeries.create` permissions on the project so it can write logs out to Cloud Logging. If using a custom container image, the service account must have [Artifact Registry Reader](https://cloud.google.com/artifact-registry/docs/access-control#roles) permission to pull the specified image. If you as the administrator want to be able to `ssh` into the underlying VM, you need to set this value to a service account for which you have the `iam.serviceAccounts.actAs` permission. Conversely, if you don't want anyone to be able to `ssh` into the underlying VM, use a service account where no one has that permission. If not set, VMs run with a service account provided by the Cloud Workstations service, and the image must be publicly accessible. "serviceAccountScopes": [ # Optional. Scopes to grant to the service_account. When specified, users of workstations under this configuration must have `iam.serviceAccounts.actAs` on the service account. "A String", @@ -658,6 +686,13 @@

Method Details

"id": "A String", # Required. The ID to be used for the boost configuration. "machineType": "A String", # Optional. The type of machine that boosted VM instances will use—for example, `e2-standard-4`. For more information about machine types that Cloud Workstations supports, see the list of [available machine types](https://cloud.google.com/workstations/docs/available-machine-types). Defaults to `e2-standard-4`. "poolSize": 42, # Optional. The number of boost VMs that the system should keep idle so that workstations can be boosted quickly. Defaults to `0`. + "reservationAffinity": { # ReservationAffinity is the configuration of the desired reservation from which instances can consume resources. # Optional. [ReservationAffinity](https://cloud.google.com/compute/docs/instances/reserving-zonal-resources) specifies a reservation that can be consumed to create boost VM instances. If SPECIFIC_RESERVATION is specified, Cloud Workstations will only create VMs in the zone where the reservation is located. This would affect availability since the service will no longer be resilient to zonal outages. If ANY_RESERVATION is specified, creating reservations in both zones that the config creates VMs in will ensure higher availability. **Important Considerations for Reservation Affinity:** * This feature is intended for advanced users and requires familiarity with Google Compute Engine reservations. * Using reservations incurs charges, regardless of utilization. * The resources in the pool will consume the specified reservation. Take this into account when setting the pool size. + "consumeReservationType": "A String", # Optional. Corresponds to the type of reservation consumption. + "key": "A String", # Optional. Corresponds to the label key of reservation resource. + "values": [ # Optional. Corresponds to the label values of reservation resources. Valid values are either a name to a reservation in the same project or "projects/{project}/reservations/{reservation}" to target a shared reservation in the same zone but in a different project. + "A String", + ], + }, }, ], "bootDiskSizeGb": 42, # Optional. The size of the boot disk for the VM in gigabytes (GB). The minimum boot disk size is `30` GB. Defaults to `50` GB. @@ -670,6 +705,13 @@

Method Details

"machineType": "A String", # Optional. The type of machine to use for VM instances—for example, `"e2-standard-4"`. For more information about machine types that Cloud Workstations supports, see the list of [available machine types](https://cloud.google.com/workstations/docs/available-machine-types). "poolSize": 42, # Optional. The number of VMs that the system should keep idle so that new workstations can be started quickly for new users. Defaults to `0` in the API. "pooledInstances": 42, # Output only. Number of instances currently available in the pool for faster workstation startup. + "reservationAffinity": { # ReservationAffinity is the configuration of the desired reservation from which instances can consume resources. # Optional. [ReservationAffinity](https://cloud.google.com/compute/docs/instances/reserving-zonal-resources) specifies a reservation that can be consumed to create boost VM instances. If SPECIFIC_RESERVATION is specified, Cloud Workstations will only create VMs in the zone where the reservation is located. This would affect availability since the service will no longer be resilient to zonal outages. If ANY_RESERVATION is specified, creating reservations in both zones that the config creates VMs in will ensure higher availability. **Important Considerations for Reservation Affinity:** * This feature is intended for advanced users and requires familiarity with Google Compute Engine reservations. * Using reservations incurs charges, regardless of utilization. * The resources in the pool will consume the specified reservation. Take this into account when setting the pool size. + "consumeReservationType": "A String", # Optional. Corresponds to the type of reservation consumption. + "key": "A String", # Optional. Corresponds to the label key of reservation resource. + "values": [ # Optional. Corresponds to the label values of reservation resources. Valid values are either a name to a reservation in the same project or "projects/{project}/reservations/{reservation}" to target a shared reservation in the same zone but in a different project. + "A String", + ], + }, "serviceAccount": "A String", # Optional. The email address of the service account for Cloud Workstations VMs created with this configuration. When specified, be sure that the service account has `logging.logEntries.create` and `monitoring.timeSeries.create` permissions on the project so it can write logs out to Cloud Logging. If using a custom container image, the service account must have [Artifact Registry Reader](https://cloud.google.com/artifact-registry/docs/access-control#roles) permission to pull the specified image. If you as the administrator want to be able to `ssh` into the underlying VM, you need to set this value to a service account for which you have the `iam.serviceAccounts.actAs` permission. Conversely, if you don't want anyone to be able to `ssh` into the underlying VM, use a service account where no one has that permission. If not set, VMs run with a service account provided by the Cloud Workstations service, and the image must be publicly accessible. "serviceAccountScopes": [ # Optional. Scopes to grant to the service_account. When specified, users of workstations under this configuration must have `iam.serviceAccounts.actAs` on the service account. "A String", @@ -830,6 +872,13 @@

Method Details

"id": "A String", # Required. The ID to be used for the boost configuration. "machineType": "A String", # Optional. The type of machine that boosted VM instances will use—for example, `e2-standard-4`. For more information about machine types that Cloud Workstations supports, see the list of [available machine types](https://cloud.google.com/workstations/docs/available-machine-types). Defaults to `e2-standard-4`. "poolSize": 42, # Optional. The number of boost VMs that the system should keep idle so that workstations can be boosted quickly. Defaults to `0`. + "reservationAffinity": { # ReservationAffinity is the configuration of the desired reservation from which instances can consume resources. # Optional. [ReservationAffinity](https://cloud.google.com/compute/docs/instances/reserving-zonal-resources) specifies a reservation that can be consumed to create boost VM instances. If SPECIFIC_RESERVATION is specified, Cloud Workstations will only create VMs in the zone where the reservation is located. This would affect availability since the service will no longer be resilient to zonal outages. If ANY_RESERVATION is specified, creating reservations in both zones that the config creates VMs in will ensure higher availability. **Important Considerations for Reservation Affinity:** * This feature is intended for advanced users and requires familiarity with Google Compute Engine reservations. * Using reservations incurs charges, regardless of utilization. * The resources in the pool will consume the specified reservation. Take this into account when setting the pool size. + "consumeReservationType": "A String", # Optional. Corresponds to the type of reservation consumption. + "key": "A String", # Optional. Corresponds to the label key of reservation resource. + "values": [ # Optional. Corresponds to the label values of reservation resources. Valid values are either a name to a reservation in the same project or "projects/{project}/reservations/{reservation}" to target a shared reservation in the same zone but in a different project. + "A String", + ], + }, }, ], "bootDiskSizeGb": 42, # Optional. The size of the boot disk for the VM in gigabytes (GB). The minimum boot disk size is `30` GB. Defaults to `50` GB. @@ -842,6 +891,13 @@

Method Details

"machineType": "A String", # Optional. The type of machine to use for VM instances—for example, `"e2-standard-4"`. For more information about machine types that Cloud Workstations supports, see the list of [available machine types](https://cloud.google.com/workstations/docs/available-machine-types). "poolSize": 42, # Optional. The number of VMs that the system should keep idle so that new workstations can be started quickly for new users. Defaults to `0` in the API. "pooledInstances": 42, # Output only. Number of instances currently available in the pool for faster workstation startup. + "reservationAffinity": { # ReservationAffinity is the configuration of the desired reservation from which instances can consume resources. # Optional. [ReservationAffinity](https://cloud.google.com/compute/docs/instances/reserving-zonal-resources) specifies a reservation that can be consumed to create boost VM instances. If SPECIFIC_RESERVATION is specified, Cloud Workstations will only create VMs in the zone where the reservation is located. This would affect availability since the service will no longer be resilient to zonal outages. If ANY_RESERVATION is specified, creating reservations in both zones that the config creates VMs in will ensure higher availability. **Important Considerations for Reservation Affinity:** * This feature is intended for advanced users and requires familiarity with Google Compute Engine reservations. * Using reservations incurs charges, regardless of utilization. * The resources in the pool will consume the specified reservation. Take this into account when setting the pool size. + "consumeReservationType": "A String", # Optional. Corresponds to the type of reservation consumption. + "key": "A String", # Optional. Corresponds to the label key of reservation resource. + "values": [ # Optional. Corresponds to the label values of reservation resources. Valid values are either a name to a reservation in the same project or "projects/{project}/reservations/{reservation}" to target a shared reservation in the same zone but in a different project. + "A String", + ], + }, "serviceAccount": "A String", # Optional. The email address of the service account for Cloud Workstations VMs created with this configuration. When specified, be sure that the service account has `logging.logEntries.create` and `monitoring.timeSeries.create` permissions on the project so it can write logs out to Cloud Logging. If using a custom container image, the service account must have [Artifact Registry Reader](https://cloud.google.com/artifact-registry/docs/access-control#roles) permission to pull the specified image. If you as the administrator want to be able to `ssh` into the underlying VM, you need to set this value to a service account for which you have the `iam.serviceAccounts.actAs` permission. Conversely, if you don't want anyone to be able to `ssh` into the underlying VM, use a service account where no one has that permission. If not set, VMs run with a service account provided by the Cloud Workstations service, and the image must be publicly accessible. "serviceAccountScopes": [ # Optional. Scopes to grant to the service_account. When specified, users of workstations under this configuration must have `iam.serviceAccounts.actAs` on the service account. "A String", @@ -1017,6 +1073,13 @@

Method Details

"id": "A String", # Required. The ID to be used for the boost configuration. "machineType": "A String", # Optional. The type of machine that boosted VM instances will use—for example, `e2-standard-4`. For more information about machine types that Cloud Workstations supports, see the list of [available machine types](https://cloud.google.com/workstations/docs/available-machine-types). Defaults to `e2-standard-4`. "poolSize": 42, # Optional. The number of boost VMs that the system should keep idle so that workstations can be boosted quickly. Defaults to `0`. + "reservationAffinity": { # ReservationAffinity is the configuration of the desired reservation from which instances can consume resources. # Optional. [ReservationAffinity](https://cloud.google.com/compute/docs/instances/reserving-zonal-resources) specifies a reservation that can be consumed to create boost VM instances. If SPECIFIC_RESERVATION is specified, Cloud Workstations will only create VMs in the zone where the reservation is located. This would affect availability since the service will no longer be resilient to zonal outages. If ANY_RESERVATION is specified, creating reservations in both zones that the config creates VMs in will ensure higher availability. **Important Considerations for Reservation Affinity:** * This feature is intended for advanced users and requires familiarity with Google Compute Engine reservations. * Using reservations incurs charges, regardless of utilization. * The resources in the pool will consume the specified reservation. Take this into account when setting the pool size. + "consumeReservationType": "A String", # Optional. Corresponds to the type of reservation consumption. + "key": "A String", # Optional. Corresponds to the label key of reservation resource. + "values": [ # Optional. Corresponds to the label values of reservation resources. Valid values are either a name to a reservation in the same project or "projects/{project}/reservations/{reservation}" to target a shared reservation in the same zone but in a different project. + "A String", + ], + }, }, ], "bootDiskSizeGb": 42, # Optional. The size of the boot disk for the VM in gigabytes (GB). The minimum boot disk size is `30` GB. Defaults to `50` GB. @@ -1029,6 +1092,13 @@

Method Details

"machineType": "A String", # Optional. The type of machine to use for VM instances—for example, `"e2-standard-4"`. For more information about machine types that Cloud Workstations supports, see the list of [available machine types](https://cloud.google.com/workstations/docs/available-machine-types). "poolSize": 42, # Optional. The number of VMs that the system should keep idle so that new workstations can be started quickly for new users. Defaults to `0` in the API. "pooledInstances": 42, # Output only. Number of instances currently available in the pool for faster workstation startup. + "reservationAffinity": { # ReservationAffinity is the configuration of the desired reservation from which instances can consume resources. # Optional. [ReservationAffinity](https://cloud.google.com/compute/docs/instances/reserving-zonal-resources) specifies a reservation that can be consumed to create boost VM instances. If SPECIFIC_RESERVATION is specified, Cloud Workstations will only create VMs in the zone where the reservation is located. This would affect availability since the service will no longer be resilient to zonal outages. If ANY_RESERVATION is specified, creating reservations in both zones that the config creates VMs in will ensure higher availability. **Important Considerations for Reservation Affinity:** * This feature is intended for advanced users and requires familiarity with Google Compute Engine reservations. * Using reservations incurs charges, regardless of utilization. * The resources in the pool will consume the specified reservation. Take this into account when setting the pool size. + "consumeReservationType": "A String", # Optional. Corresponds to the type of reservation consumption. + "key": "A String", # Optional. Corresponds to the label key of reservation resource. + "values": [ # Optional. Corresponds to the label values of reservation resources. Valid values are either a name to a reservation in the same project or "projects/{project}/reservations/{reservation}" to target a shared reservation in the same zone but in a different project. + "A String", + ], + }, "serviceAccount": "A String", # Optional. The email address of the service account for Cloud Workstations VMs created with this configuration. When specified, be sure that the service account has `logging.logEntries.create` and `monitoring.timeSeries.create` permissions on the project so it can write logs out to Cloud Logging. If using a custom container image, the service account must have [Artifact Registry Reader](https://cloud.google.com/artifact-registry/docs/access-control#roles) permission to pull the specified image. If you as the administrator want to be able to `ssh` into the underlying VM, you need to set this value to a service account for which you have the `iam.serviceAccounts.actAs` permission. Conversely, if you don't want anyone to be able to `ssh` into the underlying VM, use a service account where no one has that permission. If not set, VMs run with a service account provided by the Cloud Workstations service, and the image must be publicly accessible. "serviceAccountScopes": [ # Optional. Scopes to grant to the service_account. When specified, users of workstations under this configuration must have `iam.serviceAccounts.actAs` on the service account. "A String", diff --git a/googleapiclient/discovery_cache/documents/adsenseplatform.v1alpha.json b/googleapiclient/discovery_cache/documents/adsenseplatform.v1alpha.json index 5fe039bc94..bcf7395d5c 100644 --- a/googleapiclient/discovery_cache/documents/adsenseplatform.v1alpha.json +++ b/googleapiclient/discovery_cache/documents/adsenseplatform.v1alpha.json @@ -203,7 +203,8 @@ "$ref": "PlatformChildSite" }, "scopes": [ -"https://www.googleapis.com/auth/adsense" +"https://www.googleapis.com/auth/adsense", +"https://www.googleapis.com/auth/adsense.readonly" ] }, "list": { @@ -305,7 +306,8 @@ "$ref": "PlatformGroup" }, "scopes": [ -"https://www.googleapis.com/auth/adsense" +"https://www.googleapis.com/auth/adsense", +"https://www.googleapis.com/auth/adsense.readonly" ] }, "list": { @@ -344,6 +346,40 @@ "https://www.googleapis.com/auth/adsense", "https://www.googleapis.com/auth/adsense.readonly" ] +}, +"patch": { +"description": "Update a Platform Group.", +"flatPath": "v1alpha/accounts/{accountsId}/platforms/{platformsId}/groups/{groupsId}", +"httpMethod": "PATCH", +"id": "adsenseplatform.accounts.platforms.groups.patch", +"parameterOrder": [ +"name" +], +"parameters": { +"name": { +"description": "Identifier. Format: accounts/{account}/platforms/{platform}/groups/{platform_group}", +"location": "path", +"pattern": "^accounts/[^/]+/platforms/[^/]+/groups/[^/]+$", +"required": true, +"type": "string" +}, +"updateMask": { +"description": "Optional. The list of fields to update - currently only supports updating the `description` field.", +"format": "google-fieldmask", +"location": "query", +"type": "string" +} +}, +"path": "v1alpha/{+name}", +"request": { +"$ref": "PlatformGroup" +}, +"response": { +"$ref": "PlatformGroup" +}, +"scopes": [ +"https://www.googleapis.com/auth/adsense" +] } } } @@ -689,7 +725,7 @@ } } }, -"revision": "20250623", +"revision": "20250703", "rootUrl": "https://adsenseplatform.googleapis.com/", "schemas": { "Account": { diff --git a/googleapiclient/discovery_cache/documents/aiplatform.v1.json b/googleapiclient/discovery_cache/documents/aiplatform.v1.json index 341b5bdedd..9b49dd909f 100644 --- a/googleapiclient/discovery_cache/documents/aiplatform.v1.json +++ b/googleapiclient/discovery_cache/documents/aiplatform.v1.json @@ -101,6 +101,11 @@ }, { "description": "Locational Endpoint", +"endpointUrl": "https://europe-west10-aiplatform.googleapis.com/", +"location": "europe-west10" +}, +{ +"description": "Locational Endpoint", "endpointUrl": "https://europe-west2-aiplatform.googleapis.com/", "location": "europe-west2" }, @@ -20248,7 +20253,7 @@ } } }, -"revision": "20250613", +"revision": "20250701", "rootUrl": "https://aiplatform.googleapis.com/", "schemas": { "CloudAiLargeModelsVisionGenerateVideoResponse": { @@ -27256,7 +27261,7 @@ "id": "GoogleCloudAiplatformV1FeatureValueMetadata", "properties": { "generateTime": { -"description": "Feature generation timestamp. Typically, it is provided by user at feature ingestion time. If not, feature store will use the system timestamp when the data is ingested into feature store. For streaming ingestion, the time, aligned by days, must be no older than five years (1825 days) and no later than one year (366 days) in the future.", +"description": "Feature generation timestamp. Typically, it is provided by user at feature ingestion time. If not, feature store will use the system timestamp when the data is ingested into feature store. Legacy Feature Store: For streaming ingestion, the time, aligned by days, must be no older than five years (1825 days) and no later than one year (366 days) in the future.", "format": "google-datetime", "type": "string" } @@ -28847,7 +28852,7 @@ "type": "boolean" }, "thinkingBudget": { -"description": "Optional. Indicates the thinking budget in tokens. This is only applied when enable_thinking is true.", +"description": "Optional. Indicates the thinking budget in tokens.", "format": "int32", "type": "integer" } @@ -31396,7 +31401,7 @@ false "type": "object" }, "GoogleCloudAiplatformV1Metric": { -"description": "The metric used for dataset level evaluation.", +"description": "The metric used for running evaluations.", "id": "GoogleCloudAiplatformV1Metric", "properties": { "aggregationMetrics": { @@ -32079,7 +32084,7 @@ false "type": "string" }, "invokeRoutePrefix": { -"description": "Immutable. Invoke route prefix for the custom container. \"/*\" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with [PredictionService.Invoke] eg: \"/invoke/foo/bar\". Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint.", +"description": "Immutable. Invoke route prefix for the custom container. \"/*\" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with invoke http call eg: \"/invoke/foo/bar\", however the [PredictionService.Invoke] RPC is not supported yet. Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint.", "type": "string" }, "livenessProbe": { @@ -34372,7 +34377,7 @@ false "type": "object" }, "GoogleCloudAiplatformV1PSCAutomationConfig": { -"description": "PSC config that is used to automatically create forwarding rule via ServiceConnectionMap.", +"description": "PSC config that is used to automatically create PSC endpoints in the user projects.", "id": "GoogleCloudAiplatformV1PSCAutomationConfig", "properties": { "errorMessage": { @@ -34391,7 +34396,7 @@ false "type": "string" }, "network": { -"description": "Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/insert): `projects/{project}/global/networks/{network}`. Where {project} is a project number, as in '12345', and {network} is network name.", +"description": "Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/get): `projects/{project}/global/networks/{network}`.", "type": "string" }, "projectId": { @@ -35208,6 +35213,11 @@ false "description": "Output only. The user specified name of the task that is defined in pipeline_spec.", "readOnly": true, "type": "string" +}, +"taskUniqueName": { +"description": "Output only. The unique name of a task. This field is used by rerun pipeline job. Console UI and Vertex AI SDK will support triggering pipeline job reruns. The name is constructed by concatenating all the parent tasks name with the task name. For example, if a task named \"child_task\" has a parent task named \"parent_task_1\" and parent task 1 has a parent task named \"parent_task_2\", the task unique name will be \"parent_task_2.parent_task_1.child_task\".", +"readOnly": true, +"type": "string" } }, "type": "object" @@ -35682,6 +35692,13 @@ false }, "type": "array" }, +"pscAutomationConfigs": { +"description": "Optional. List of projects and networks where the PSC endpoints will be created. This field is used by Online Inference(Prediction) only.", +"items": { +"$ref": "GoogleCloudAiplatformV1PSCAutomationConfig" +}, +"type": "array" +}, "serviceAttachment": { "description": "Output only. The name of the generated service attachment resource. This is only populated if the endpoint is deployed with PrivateServiceConnect.", "readOnly": true, @@ -46467,7 +46484,7 @@ false "type": "string" }, "model": { -"description": "Output only. The resource name of the TunedModel. Format: `projects/{project}/locations/{location}/models/{model}`.", +"description": "Output only. The resource name of the TunedModel. Format: `projects/{project}/locations/{location}/models/{model}@{version_id}` When tuning from a base model, the version_id will be 1. For continuous tuning, the version id will be incremented by 1 from the last version id in the parent model. E.g., `projects/{project}/locations/{location}/models/{model}@{last_version_id + 1}`", "readOnly": true, "type": "string" } diff --git a/googleapiclient/discovery_cache/documents/aiplatform.v1beta1.json b/googleapiclient/discovery_cache/documents/aiplatform.v1beta1.json index fcbb8f0aff..436c123580 100644 --- a/googleapiclient/discovery_cache/documents/aiplatform.v1beta1.json +++ b/googleapiclient/discovery_cache/documents/aiplatform.v1beta1.json @@ -101,6 +101,11 @@ }, { "description": "Locational Endpoint", +"endpointUrl": "https://europe-west10-aiplatform.googleapis.com/", +"location": "europe-west10" +}, +{ +"description": "Locational Endpoint", "endpointUrl": "https://europe-west2-aiplatform.googleapis.com/", "location": "europe-west2" }, @@ -20390,6 +20395,11 @@ "parent" ], "parameters": { +"filter": { +"description": "Optional. The standard list filter. Supported fields: * `timestamp` range (i.e. `timestamp>=\"2025-01-31T11:30:00-04:00\"` where the timestamp is in RFC 3339 format) More detail in [AIP-160](https://google.aip.dev/160).", +"location": "query", +"type": "string" +}, "pageSize": { "description": "Optional. The maximum number of events to return. The service may return fewer than this value. If unspecified, at most 100 events will be returned. These events are ordered by timestamp in ascending order.", "format": "int32", @@ -25213,6 +25223,11 @@ "parent" ], "parameters": { +"filter": { +"description": "Optional. The standard list filter. Supported fields: * `timestamp` range (i.e. `timestamp>=\"2025-01-31T11:30:00-04:00\"` where the timestamp is in RFC 3339 format) More detail in [AIP-160](https://google.aip.dev/160).", +"location": "query", +"type": "string" +}, "pageSize": { "description": "Optional. The maximum number of events to return. The service may return fewer than this value. If unspecified, at most 100 events will be returned. These events are ordered by timestamp in ascending order.", "format": "int32", @@ -25247,7 +25262,7 @@ } } }, -"revision": "20250613", +"revision": "20250701", "rootUrl": "https://aiplatform.googleapis.com/", "schemas": { "CloudAiLargeModelsVisionGenerateVideoResponse": { @@ -29477,6 +29492,10 @@ }, "type": "array" }, +"flexStart": { +"$ref": "GoogleCloudAiplatformV1beta1FlexStart", +"description": "Optional. Immutable. If set, use DWS resource to schedule the deployment workload. reference: (https://cloud.google.com/blog/products/compute/introducing-dynamic-workload-scheduler)" +}, "machineSpec": { "$ref": "GoogleCloudAiplatformV1beta1MachineSpec", "description": "Required. Immutable. The specification of a single machine being used." @@ -31162,6 +31181,14 @@ "description": "Optional. The branch of the event. The format is like agent_1.agent_2.agent_3, where agent_1 is the parent of agent_2, and agent_2 is the parent of agent_3. Branch is used when multiple child agents shouldn't see their siblings' conversation history.", "type": "string" }, +"customMetadata": { +"additionalProperties": { +"description": "Properties of the object.", +"type": "any" +}, +"description": "The custom metadata of the LlmResponse.", +"type": "object" +}, "groundingMetadata": { "$ref": "GoogleCloudAiplatformV1beta1GroundingMetadata", "description": "Optional. Metadata returned to client when grounding is enabled." @@ -33386,7 +33413,7 @@ "id": "GoogleCloudAiplatformV1beta1FeatureValueMetadata", "properties": { "generateTime": { -"description": "Feature generation timestamp. Typically, it is provided by user at feature ingestion time. If not, feature store will use the system timestamp when the data is ingested into feature store. For streaming ingestion, the time, aligned by days, must be no older than five years (1825 days) and no later than one year (366 days) in the future.", +"description": "Feature generation timestamp. Typically, it is provided by user at feature ingestion time. If not, feature store will use the system timestamp when the data is ingested into feature store. Legacy Feature Store: For streaming ingestion, the time, aligned by days, must be no older than five years (1825 days) and no later than one year (366 days) in the future.", "format": "google-datetime", "type": "string" } @@ -34462,6 +34489,18 @@ }, "type": "object" }, +"GoogleCloudAiplatformV1beta1FlexStart": { +"description": "FlexStart is used to schedule the deployment workload on DWS resource. It contains the max duration of the deployment.", +"id": "GoogleCloudAiplatformV1beta1FlexStart", +"properties": { +"maxRuntimeDuration": { +"description": "The max duration of the deployment is max_runtime_duration. The deployment will be terminated after the duration. The max_runtime_duration can be set up to 7 days.", +"format": "google-duration", +"type": "string" +} +}, +"type": "object" +}, "GoogleCloudAiplatformV1beta1FluencyInput": { "description": "Input for fluency metric.", "id": "GoogleCloudAiplatformV1beta1FluencyInput", @@ -35448,7 +35487,7 @@ "type": "boolean" }, "thinkingBudget": { -"description": "Optional. Indicates the thinking budget in tokens. This is only applied when enable_thinking is true.", +"description": "Optional. Indicates the thinking budget in tokens.", "format": "int32", "type": "integer" } @@ -38442,7 +38481,7 @@ false "type": "object" }, "GoogleCloudAiplatformV1beta1Metric": { -"description": "The metric used for dataset level evaluation.", +"description": "The metric used for running evaluations.", "id": "GoogleCloudAiplatformV1beta1Metric", "properties": { "aggregationMetrics": { @@ -39117,7 +39156,7 @@ false "type": "string" }, "invokeRoutePrefix": { -"description": "Immutable. Invoke route prefix for the custom container. \"/*\" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with [PredictionService.Invoke] eg: \"/invoke/foo/bar\". Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint.", +"description": "Immutable. Invoke route prefix for the custom container. \"/*\" is the only supported value right now. By setting this field, any non-root route on this model will be accessible with invoke http call eg: \"/invoke/foo/bar\", however the [PredictionService.Invoke] RPC is not supported yet. Only one of `predict_route` or `invoke_route_prefix` can be set, and we default to using `predict_route` if this field is not set. If this field is set, the Model can only be deployed to dedicated endpoint.", "type": "string" }, "livenessProbe": { @@ -42189,7 +42228,7 @@ false "type": "object" }, "GoogleCloudAiplatformV1beta1PSCAutomationConfig": { -"description": "PSC config that is used to automatically create forwarding rule via ServiceConnectionMap.", +"description": "PSC config that is used to automatically create PSC endpoints in the user projects.", "id": "GoogleCloudAiplatformV1beta1PSCAutomationConfig", "properties": { "errorMessage": { @@ -42208,7 +42247,7 @@ false "type": "string" }, "network": { -"description": "Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/insert): `projects/{project}/global/networks/{network}`. Where {project} is a project number, as in '12345', and {network} is network name.", +"description": "Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/get): `projects/{project}/global/networks/{network}`.", "type": "string" }, "projectId": { @@ -43114,6 +43153,11 @@ false "description": "Output only. The user specified name of the task that is defined in pipeline_spec.", "readOnly": true, "type": "string" +}, +"taskUniqueName": { +"description": "Output only. The unique name of a task. This field is used by rerun pipeline job. Console UI and Vertex AI SDK will support triggering pipeline job reruns. The name is constructed by concatenating all the parent tasks name with the task name. For example, if a task named \"child_task\" has a parent task named \"parent_task_1\" and parent task 1 has a parent task named \"parent_task_2\", the task unique name will be \"parent_task_2.parent_task_1.child_task\".", +"readOnly": true, +"type": "string" } }, "type": "object" @@ -43681,6 +43725,13 @@ false }, "type": "array" }, +"pscAutomationConfigs": { +"description": "Optional. List of projects and networks where the PSC endpoints will be created. This field is used by Online Inference(Prediction) only.", +"items": { +"$ref": "GoogleCloudAiplatformV1beta1PSCAutomationConfig" +}, +"type": "array" +}, "serviceAttachment": { "description": "Output only. The name of the generated service attachment resource. This is only populated if the endpoint is deployed with PrivateServiceConnect.", "readOnly": true, @@ -46074,7 +46125,7 @@ false "id": "GoogleCloudAiplatformV1beta1ReasoningEngineContextSpecMemoryBankConfigGenerationConfig", "properties": { "model": { -"description": "Required. The model used to generate memories. Format: `projects/{project}/locations/{location}/publishers/google/models/{model}` or `projects/{project}/locations/{location}/endpoints/{endpoint}`.", +"description": "Required. The model used to generate memories. Format: `projects/{project}/locations/{location}/publishers/google/models/{model}`.", "type": "string" } }, @@ -46085,7 +46136,7 @@ false "id": "GoogleCloudAiplatformV1beta1ReasoningEngineContextSpecMemoryBankConfigSimilaritySearchConfig", "properties": { "embeddingModel": { -"description": "Required. The model used to generate embeddings to lookup similar memories. Format: `projects/{project}/locations/{location}/publishers/google/models/{model}` or `projects/{project}/locations/{location}/endpoints/{endpoint}`.", +"description": "Required. The model used to generate embeddings to lookup similar memories. Format: `projects/{project}/locations/{location}/publishers/google/models/{model}`.", "type": "string" } }, @@ -55809,7 +55860,7 @@ false "type": "string" }, "model": { -"description": "Output only. The resource name of the TunedModel. Format: `projects/{project}/locations/{location}/models/{model}`.", +"description": "Output only. The resource name of the TunedModel. Format: `projects/{project}/locations/{location}/models/{model}@{version_id}` When tuning from a base model, the version_id will be 1. For continuous tuning, the version id will be incremented by 1 from the last version id in the parent model. E.g., `projects/{project}/locations/{location}/models/{model}@{last_version_id + 1}`", "readOnly": true, "type": "string" } diff --git a/googleapiclient/discovery_cache/documents/androidmanagement.v1.json b/googleapiclient/discovery_cache/documents/androidmanagement.v1.json index c356283fa9..b6a585c84a 100644 --- a/googleapiclient/discovery_cache/documents/androidmanagement.v1.json +++ b/googleapiclient/discovery_cache/documents/androidmanagement.v1.json @@ -1260,7 +1260,7 @@ } } }, -"revision": "20250626", +"revision": "20250702", "rootUrl": "https://androidmanagement.googleapis.com/", "schemas": { "AdbShellCommandEvent": { @@ -2242,7 +2242,7 @@ "USER_CONTROL_DISALLOWED" ], "enumDescriptions": [ -"Uses the default behaviour of the app to determine if user control is allowed or disallowed. For most apps, user control is allowed by default, but user control is disallowed for some critical apps such as: * extension apps (see extensionConfig for more details) * kiosk apps (see KIOSK install type for more details) * other critical system apps", +"Uses the default behaviour of the app to determine if user control is allowed or disallowed. User control is allowed by default for most apps but disallowed for following types of apps: extension apps (see extensionConfig for more details) kiosk apps (see KIOSK install type for more details) other critical system apps", "User control is allowed for the app. Kiosk apps can use this to allow user control. For extension apps (see extensionConfig for more details), user control is disallowed even if this value is set. For kiosk apps (see KIOSK install type for more details), this value can be used to allow user control.", "User control is disallowed for the app. This is supported on Android 11 and above. A NonComplianceDetail with API_LEVEL is reported if the Android version is less than 11." ], @@ -3910,7 +3910,7 @@ false "id": "ExtensionConfig", "properties": { "notificationReceiver": { -"description": "Fully qualified class name of the receiver service class for Android Device Policy to notify the extension app of any local command status updates.", +"description": "Fully qualified class name of the receiver service class for Android Device Policy to notify the extension app of any local command status updates. The service must be exported in the extension app's AndroidManifest.xml and extend NotificationReceiverService (https://developers.google.com/android/management/reference/amapi/com/google/android/managementapi/notification/NotificationReceiverService) (see Integrate with the AMAPI SDK (https://developers.google.com/android/management/sdk-integration) guide for more details).", "type": "string" }, "signingKeyFingerprintsSha256": { diff --git a/googleapiclient/discovery_cache/documents/androidpublisher.v3.json b/googleapiclient/discovery_cache/documents/androidpublisher.v3.json index 17fe859713..130db21982 100644 --- a/googleapiclient/discovery_cache/documents/androidpublisher.v3.json +++ b/googleapiclient/discovery_cache/documents/androidpublisher.v3.json @@ -4029,6 +4029,41 @@ } } }, +"productsv2": { +"methods": { +"getproductpurchasev2": { +"description": "Checks the purchase and consumption status of an inapp item.", +"flatPath": "androidpublisher/v3/applications/{packageName}/purchases/productsv2/tokens/{token}", +"httpMethod": "GET", +"id": "androidpublisher.purchases.productsv2.getproductpurchasev2", +"parameterOrder": [ +"packageName", +"token" +], +"parameters": { +"packageName": { +"description": "The package name of the application the inapp product was sold in (for example, 'com.some.thing').", +"location": "path", +"required": true, +"type": "string" +}, +"token": { +"description": "The token provided to the user's device when the inapp product was purchased.", +"location": "path", +"required": true, +"type": "string" +} +}, +"path": "androidpublisher/v3/applications/{packageName}/purchases/productsv2/tokens/{token}", +"response": { +"$ref": "ProductPurchaseV2" +}, +"scopes": [ +"https://www.googleapis.com/auth/androidpublisher" +] +} +} +}, "subscriptions": { "methods": { "acknowledge": { @@ -4796,7 +4831,7 @@ } } }, -"revision": "20250602", +"revision": "20250701", "rootUrl": "https://androidpublisher.googleapis.com/", "schemas": { "Abi": { @@ -8162,6 +8197,76 @@ false }, "type": "object" }, +"ProductLineItem": { +"description": "Contains item-level info for a ProductPurchaseV2.", +"id": "ProductLineItem", +"properties": { +"productId": { +"description": "The purchased product ID (for example, 'monthly001').", +"type": "string" +}, +"productOfferDetails": { +"$ref": "ProductOfferDetails", +"description": "The offer details for this item." +} +}, +"type": "object" +}, +"ProductOfferDetails": { +"description": "Offer details information related to a purchase line item.", +"id": "ProductOfferDetails", +"properties": { +"consumptionState": { +"description": "Output only. The consumption state of the purchase.", +"enum": [ +"CONSUMPTION_STATE_UNSPECIFIED", +"CONSUMPTION_STATE_YET_TO_BE_CONSUMED", +"CONSUMPTION_STATE_CONSUMED" +], +"enumDescriptions": [ +"Consumption state unspecified. This value should never be set.", +"Yet to be consumed.", +"Consumed already." +], +"readOnly": true, +"type": "string" +}, +"offerId": { +"description": "The offer ID. Only present for offers.", +"type": "string" +}, +"offerTags": { +"description": "The latest offer tags associated with the offer. It includes tags inherited from the purchase option.", +"items": { +"type": "string" +}, +"type": "array" +}, +"offerToken": { +"description": "The per-transaction offer token used to make this purchase line item.", +"type": "string" +}, +"purchaseOptionId": { +"description": "The purchase option ID.", +"type": "string" +}, +"quantity": { +"description": "The quantity associated with the purchase of the inapp product.", +"format": "int32", +"type": "integer" +}, +"refundableQuantity": { +"description": "The quantity eligible for refund, i.e. quantity that hasn't been refunded. The value reflects quantity-based partial refunds and full refunds.", +"format": "int32", +"type": "integer" +}, +"rentOfferDetails": { +"$ref": "RentOfferDetails", +"description": "Offer details about rent offers. This will only be set for rental line items." +} +}, +"type": "object" +}, "ProductPurchase": { "description": "A ProductPurchase resource indicates the status of a user's inapp product purchase.", "id": "ProductPurchase", @@ -8236,6 +8341,68 @@ false }, "type": "object" }, +"ProductPurchaseV2": { +"description": "A ProductPurchaseV2 resource indicates the status of a user's inapp product purchase.", +"id": "ProductPurchaseV2", +"properties": { +"acknowledgementState": { +"description": "Output only. The acknowledgement state of the purchase.", +"enum": [ +"ACKNOWLEDGEMENT_STATE_UNSPECIFIED", +"ACKNOWLEDGEMENT_STATE_PENDING", +"ACKNOWLEDGEMENT_STATE_ACKNOWLEDGED" +], +"enumDescriptions": [ +"Unspecified acknowledgement state.", +"The purchase is not acknowledged yet.", +"The purchase is acknowledged." +], +"readOnly": true, +"type": "string" +}, +"kind": { +"description": "This kind represents a ProductPurchaseV2 object in the androidpublisher service.", +"type": "string" +}, +"obfuscatedExternalAccountId": { +"description": "An obfuscated version of the id that is uniquely associated with the user's account in your app. Only present if specified using https://developer.android.com/reference/com/android/billingclient/api/BillingFlowParams.Builder#setobfuscatedaccountid when the purchase was made.", +"type": "string" +}, +"obfuscatedExternalProfileId": { +"description": "An obfuscated version of the id that is uniquely associated with the user's profile in your app. Only present if specified using https://developer.android.com/reference/com/android/billingclient/api/BillingFlowParams.Builder#setobfuscatedprofileid when the purchase was made.", +"type": "string" +}, +"orderId": { +"description": "The order id associated with the purchase of the inapp product. May not be set if there is no order associated with the purchase.", +"type": "string" +}, +"productLineItem": { +"description": "Contains item-level info for a ProductPurchaseV2.", +"items": { +"$ref": "ProductLineItem" +}, +"type": "array" +}, +"purchaseCompletionTime": { +"description": "The time when the purchase was successful, i.e., when the PurchaseState has changed to PURCHASED. This field will not be present until the payment is complete. For example, if the user initiated a pending transaction (https://developer.android.com/google/play/billing/integrate#pending), this field will not be populated until the user successfully completes the steps required to complete the transaction.", +"format": "google-datetime", +"type": "string" +}, +"purchaseStateContext": { +"$ref": "PurchaseStateContext", +"description": "Information about the purchase state of the purchase." +}, +"regionCode": { +"description": "ISO 3166-1 alpha-2 billing region code of the user at the time the product was granted.", +"type": "string" +}, +"testPurchaseContext": { +"$ref": "TestPurchaseContext", +"description": "Information related to test purchases. This will only be set for test purchases." +} +}, +"type": "object" +}, "ProductPurchasesAcknowledgeRequest": { "description": "Request for the product.purchases.acknowledge API.", "id": "ProductPurchasesAcknowledgeRequest", @@ -8247,6 +8414,30 @@ false }, "type": "object" }, +"PurchaseStateContext": { +"description": "Context about the purchase state.", +"id": "PurchaseStateContext", +"properties": { +"purchaseState": { +"description": "Output only. The purchase state of the purchase.", +"enum": [ +"PURCHASE_STATE_UNSPECIFIED", +"PURCHASED", +"CANCELLED", +"PENDING" +], +"enumDescriptions": [ +"Purchase state unspecified. This value should never be set.", +"Purchased successfully.", +"Purchase canceled.", +"The purchase is in a pending state and has not yet been completed. For more information on handling pending purchases, see https://developer.android.com/google/play/billing/integrate#pending." +], +"readOnly": true, +"type": "string" +} +}, +"type": "object" +}, "RecurringExternalTransaction": { "description": "Represents a transaction that is part of a recurring series of payments. This can be a subscription or a one-time product with multiple payments (such as preorder).", "id": "RecurringExternalTransaction", @@ -8572,6 +8763,12 @@ false }, "type": "object" }, +"RentOfferDetails": { +"description": "Offer details information related to a rental line item.", +"id": "RentOfferDetails", +"properties": {}, +"type": "object" +}, "ReplacementCancellation": { "description": "Information specific to cancellations caused by subscription replacement.", "id": "ReplacementCancellation", @@ -9763,6 +9960,25 @@ false "properties": {}, "type": "object" }, +"TestPurchaseContext": { +"description": "Context about a test purchase.", +"id": "TestPurchaseContext", +"properties": { +"fopType": { +"description": "The fop type of the test purchase.", +"enum": [ +"FOP_TYPE_UNSPECIFIED", +"TEST" +], +"enumDescriptions": [ +"Fop type unspecified. This value should never be set.", +"The purchase was made using a test card." +], +"type": "string" +} +}, +"type": "object" +}, "Testers": { "description": "The testers of an app. The resource for TestersService. Note: while it is possible in the Play Console UI to add testers via email lists, email lists are not supported by this resource.", "id": "Testers", diff --git a/googleapiclient/discovery_cache/documents/apigee.v1.json b/googleapiclient/discovery_cache/documents/apigee.v1.json index 27ead3f758..95b193b914 100644 --- a/googleapiclient/discovery_cache/documents/apigee.v1.json +++ b/googleapiclient/discovery_cache/documents/apigee.v1.json @@ -25,6 +25,11 @@ "description": "Regional Endpoint", "endpointUrl": "https://apigee.eu.rep.googleapis.com/", "location": "eu" +}, +{ +"description": "Regional Endpoint", +"endpointUrl": "https://apigee.in.rep.googleapis.com/", +"location": "in" } ], "fullyEncodeReservedExpansion": true, @@ -11122,7 +11127,7 @@ } } }, -"revision": "20250622", +"revision": "20250701", "rootUrl": "https://apigee.googleapis.com/", "schemas": { "EdgeConfigstoreBundleBadBundle": { diff --git a/googleapiclient/discovery_cache/documents/apihub.v1.json b/googleapiclient/discovery_cache/documents/apihub.v1.json index c721935825..22a4ecc345 100644 --- a/googleapiclient/discovery_cache/documents/apihub.v1.json +++ b/googleapiclient/discovery_cache/documents/apihub.v1.json @@ -2608,6 +2608,40 @@ "scopes": [ "https://www.googleapis.com/auth/cloud-platform" ] +}, +"patch": { +"description": "Updates a plugin instance in the API hub. The following fields in the plugin_instance can be updated currently: * display_name * schedule_cron_expression The update_mask should be used to specify the fields being updated. To update the auth_config and additional_config of the plugin instance, use the ApplyPluginInstanceConfig method.", +"flatPath": "v1/projects/{projectsId}/locations/{locationsId}/plugins/{pluginsId}/instances/{instancesId}", +"httpMethod": "PATCH", +"id": "apihub.projects.locations.plugins.instances.patch", +"parameterOrder": [ +"name" +], +"parameters": { +"name": { +"description": "Identifier. The unique name of the plugin instance resource. Format: `projects/{project}/locations/{location}/plugins/{plugin}/instances/{instance}`", +"location": "path", +"pattern": "^projects/[^/]+/locations/[^/]+/plugins/[^/]+/instances/[^/]+$", +"required": true, +"type": "string" +}, +"updateMask": { +"description": "Optional. The list of fields to update.", +"format": "google-fieldmask", +"location": "query", +"type": "string" +} +}, +"path": "v1/{+name}", +"request": { +"$ref": "GoogleCloudApihubV1PluginInstance" +}, +"response": { +"$ref": "GoogleCloudApihubV1PluginInstance" +}, +"scopes": [ +"https://www.googleapis.com/auth/cloud-platform" +] } } }, @@ -2780,7 +2814,7 @@ } } }, -"revision": "20250611", +"revision": "20250701", "rootUrl": "https://apihub.googleapis.com/", "schemas": { "Empty": { @@ -3941,6 +3975,10 @@ "$ref": "GoogleCloudApihubV1AttributeValues", "description": "Optional. The environment mapping to this deployment. This maps to the following system defined attribute: `projects/{project}/locations/{location}/attributes/system-environment` attribute. The number of values for this attribute will be based on the cardinality of the attribute. The same can be retrieved via GetAttribute API. All values should be from the list of allowed values defined for the attribute." }, +"managementUrl": { +"$ref": "GoogleCloudApihubV1AttributeValues", +"description": "Optional. The uri where users can navigate to for the management of the deployment. This maps to the following system defined attribute: `projects/{project}/locations/{location}/attributes/system-management-url` The number of values for this attribute will be based on the cardinality of the attribute. The same can be retrieved via GetAttribute API." +}, "name": { "description": "Identifier. The name of the deployment. Format: `projects/{project}/locations/{location}/deployments/{deployment}`", "type": "string" @@ -3953,6 +3991,10 @@ "$ref": "GoogleCloudApihubV1AttributeValues", "description": "Optional. The SLO for this deployment. This maps to the following system defined attribute: `projects/{project}/locations/{location}/attributes/system-slo` attribute. The number of values for this attribute will be based on the cardinality of the attribute. The same can be retrieved via GetAttribute API. All values should be from the list of allowed values defined for the attribute." }, +"sourceEnvironment": { +"description": "Optional. The environment at source for the deployment. For example: prod, dev, staging, etc.", +"type": "string" +}, "sourceMetadata": { "description": "Output only. The list of sources and metadata from the sources of the deployment.", "items": { @@ -3961,6 +4003,14 @@ "readOnly": true, "type": "array" }, +"sourceProject": { +"description": "Optional. The project to which the deployment belongs. For GCP gateways, this will refer to the project identifier. For others like Edge/OPDK, this will refer to the org identifier.", +"type": "string" +}, +"sourceUri": { +"$ref": "GoogleCloudApihubV1AttributeValues", +"description": "Optional. The uri where additional source specific information for this deployment can be found. This maps to the following system defined attribute: `projects/{project}/locations/{location}/attributes/system-source-uri` The number of values for this attribute will be based on the cardinality of the attribute. The same can be retrieved via GetAttribute API." +}, "updateTime": { "description": "Output only. The time at which the deployment was last updated.", "format": "google-datetime", @@ -4864,6 +4914,30 @@ "$ref": "GoogleCloudApihubV1Documentation", "description": "Optional. The documentation of the plugin, that explains how to set up and use the plugin." }, +"gatewayType": { +"description": "Optional. The type of the gateway.", +"enum": [ +"GATEWAY_TYPE_UNSPECIFIED", +"APIGEE_X_AND_HYBRID", +"APIGEE_EDGE_PUBLIC_CLOUD", +"APIGEE_EDGE_PRIVATE_CLOUD", +"CLOUD_API_GATEWAY", +"CLOUD_ENDPOINTS", +"API_DISCOVERY", +"OTHERS" +], +"enumDescriptions": [ +"The gateway type is not specified.", +"The gateway type is Apigee X and Hybrid.", +"The gateway type is Apigee Edge Public Cloud.", +"The gateway type is Apigee Edge Private Cloud.", +"The gateway type is Cloud API Gateway.", +"The gateway type is Cloud Endpoints.", +"The gateway type is API Discovery.", +"The gateway type for any other types of gateways." +], +"type": "string" +}, "hostingService": { "$ref": "GoogleCloudApihubV1HostingService", "description": "Optional. This field is optional. It is used to notify the plugin hosting service for any lifecycle changes of the plugin instance and trigger execution of plugin instance actions in case of API hub managed actions. This field should be provided if the plugin instance lifecycle of the developed plugin needs to be managed from API hub. Also, in this case the plugin hosting service interface needs to be implemented. This field should not be provided if the plugin wants to manage plugin instance lifecycle events outside of hub interface and use plugin framework for only registering of plugin and plugin instances to capture the source of data into hub. Note, in this case the plugin hosting service interface is not required to be implemented. Also, the plugin instance lifecycle actions will be disabled from API hub's UI." @@ -5005,6 +5079,10 @@ "description": "Identifier. The unique name of the plugin instance resource. Format: `projects/{project}/locations/{location}/plugins/{plugin}/instances/{instance}`", "type": "string" }, +"sourceProjectId": { +"description": "Optional. The source project id of the plugin instance. This will be the id of runtime project in case of gcp based plugins and org id in case of non gcp based plugins. This is a required field.", +"type": "string" +}, "state": { "description": "Output only. The current state of the plugin instance (e.g., enabled, disabled, provisioning).", "enum": [ @@ -5053,6 +5131,11 @@ "$ref": "GoogleCloudApihubV1ExecutionStatus", "description": "Optional. The execution information for the plugin instance action done corresponding to an API hub instance." }, +"resourceConfig": { +"$ref": "GoogleCloudApihubV1ResourceConfig", +"description": "Output only. The configuration of resources created for a given plugin instance action. Note these will be returned only in case of Non-GCP plugins like OPDK.", +"readOnly": true +}, "scheduleCronExpression": { "description": "Optional. The schedule for this plugin instance action. This can only be set if the plugin supports API_HUB_SCHEDULE_TRIGGER mode for this action.", "type": "string" @@ -5061,6 +5144,10 @@ "description": "Optional. The time zone for the schedule cron expression. If not provided, UTC will be used.", "type": "string" }, +"serviceAccount": { +"description": "Optional. The service account used to publish data. Note, the service account will only be accepted for non GCP plugins like OPDK.", +"type": "string" +}, "state": { "description": "Output only. The current state of the plugin action in the plugin instance.", "enum": [ @@ -5151,6 +5238,33 @@ }, "type": "object" }, +"GoogleCloudApihubV1ResourceConfig": { +"description": "The configuration of resources created for a given plugin instance action.", +"id": "GoogleCloudApihubV1ResourceConfig", +"properties": { +"actionType": { +"description": "Output only. The type of the action.", +"enum": [ +"ACTION_TYPE_UNSPECIFIED", +"SYNC_METADATA", +"SYNC_RUNTIME_DATA" +], +"enumDescriptions": [ +"Default unspecified action type.", +"Action type for sync metadata.", +"Action type for sync runtime data." +], +"readOnly": true, +"type": "string" +}, +"pubsubTopic": { +"description": "Output only. The pubsub topic to publish the data to. Format is projects/{project}/topics/{topic}", +"readOnly": true, +"type": "string" +} +}, +"type": "object" +}, "GoogleCloudApihubV1RuntimeProjectAttachment": { "description": "Runtime project attachment represents an attachment from the runtime project to the host project. Api Hub looks for deployments in the attached runtime projects and creates corresponding resources in Api Hub for the discovered deployments.", "id": "GoogleCloudApihubV1RuntimeProjectAttachment", diff --git a/googleapiclient/discovery_cache/documents/backupdr.v1.json b/googleapiclient/discovery_cache/documents/backupdr.v1.json index ae53f7ac32..8ba9611e61 100644 --- a/googleapiclient/discovery_cache/documents/backupdr.v1.json +++ b/googleapiclient/discovery_cache/documents/backupdr.v1.json @@ -253,6 +253,57 @@ "https://www.googleapis.com/auth/cloud-platform" ] }, +"fetchForResourceType": { +"description": "List BackupPlanAssociations for a given resource type.", +"flatPath": "v1/projects/{projectsId}/locations/{locationsId}/backupPlanAssociations:fetchForResourceType", +"httpMethod": "GET", +"id": "backupdr.projects.locations.backupPlanAssociations.fetchForResourceType", +"parameterOrder": [ +"parent" +], +"parameters": { +"filter": { +"description": "Optional. A filter expression that filters the results fetched in the response. The expression must specify the field name, a comparison operator, and the value that you want to use for filtering. Supported fields: * resource * backup_plan * state * data_source * cloud_sql_instance_backup_plan_association_properties.instance_create_time", +"location": "query", +"type": "string" +}, +"orderBy": { +"description": "Optional. A comma-separated list of fields to order by, sorted in ascending order. Use \"desc\" after a field name for descending. Supported fields: * name", +"location": "query", +"type": "string" +}, +"pageSize": { +"description": "Optional. The maximum number of BackupPlanAssociations to return. The service may return fewer than this value. If unspecified, at most 50 BackupPlanAssociations will be returned. The maximum value is 100; values above 100 will be coerced to 100.", +"format": "int32", +"location": "query", +"type": "integer" +}, +"pageToken": { +"description": "Optional. A page token, received from a previous call of `FetchBackupPlanAssociationsForResourceType`. Provide this to retrieve the subsequent page. When paginating, all other parameters provided to `FetchBackupPlanAssociationsForResourceType` must match the call that provided the page token.", +"location": "query", +"type": "string" +}, +"parent": { +"description": "Required. The parent resource name. Format: projects/{project}/locations/{location}", +"location": "path", +"pattern": "^projects/[^/]+/locations/[^/]+$", +"required": true, +"type": "string" +}, +"resourceType": { +"description": "Required. The type of the GCP resource. Ex: sql.googleapis.com/Instance", +"location": "query", +"type": "string" +} +}, +"path": "v1/{+parent}/backupPlanAssociations:fetchForResourceType", +"response": { +"$ref": "FetchBackupPlanAssociationsForResourceTypeResponse" +}, +"scopes": [ +"https://www.googleapis.com/auth/cloud-platform" +] +}, "get": { "description": "Gets details of a single BackupPlanAssociation.", "flatPath": "v1/projects/{projectsId}/locations/{locationsId}/backupPlanAssociations/{backupPlanAssociationsId}", @@ -1461,6 +1512,86 @@ } } }, +"dataSourceReferences": { +"methods": { +"fetchForResourceType": { +"description": "Fetch DataSourceReferences for a given project, location and resource type.", +"flatPath": "v1/projects/{projectsId}/locations/{locationsId}/dataSourceReferences:fetchForResourceType", +"httpMethod": "GET", +"id": "backupdr.projects.locations.dataSourceReferences.fetchForResourceType", +"parameterOrder": [ +"parent" +], +"parameters": { +"filter": { +"description": "Optional. A filter expression that filters the results fetched in the response. The expression must specify the field name, a comparison operator, and the value that you want to use for filtering. Supported fields: * data_source * data_source_gcp_resource_info.gcp_resourcename * data_source_backup_config_state * data_source_backup_count * data_source_backup_config_info.last_backup_state * data_source_gcp_resource_info.gcp_resourcename * data_source_gcp_resource_info.type * data_source_gcp_resource_info.location * data_source_gcp_resource_info.cloud_sql_instance_properties.instance_create_time", +"location": "query", +"type": "string" +}, +"orderBy": { +"description": "Optional. A comma-separated list of fields to order by, sorted in ascending order. Use \"desc\" after a field name for descending. Supported fields: * name", +"location": "query", +"type": "string" +}, +"pageSize": { +"description": "Optional. The maximum number of DataSourceReferences to return. The service may return fewer than this value. If unspecified, at most 50 DataSourceReferences will be returned. The maximum value is 100; values above 100 will be coerced to 100.", +"format": "int32", +"location": "query", +"type": "integer" +}, +"pageToken": { +"description": "Optional. A page token, received from a previous call of `FetchDataSourceReferencesForResourceType`. Provide this to retrieve the subsequent page. When paginating, all other parameters provided to `FetchDataSourceReferencesForResourceType` must match the call that provided the page token.", +"location": "query", +"type": "string" +}, +"parent": { +"description": "Required. The parent resource name. Format: projects/{project}/locations/{location}", +"location": "path", +"pattern": "^projects/[^/]+/locations/[^/]+$", +"required": true, +"type": "string" +}, +"resourceType": { +"description": "Required. The type of the GCP resource. Ex: sql.googleapis.com/Instance", +"location": "query", +"type": "string" +} +}, +"path": "v1/{+parent}/dataSourceReferences:fetchForResourceType", +"response": { +"$ref": "FetchDataSourceReferencesForResourceTypeResponse" +}, +"scopes": [ +"https://www.googleapis.com/auth/cloud-platform" +] +}, +"get": { +"description": "Gets details of a single DataSourceReference.", +"flatPath": "v1/projects/{projectsId}/locations/{locationsId}/dataSourceReferences/{dataSourceReferencesId}", +"httpMethod": "GET", +"id": "backupdr.projects.locations.dataSourceReferences.get", +"parameterOrder": [ +"name" +], +"parameters": { +"name": { +"description": "Required. The name of the DataSourceReference to retrieve. Format: projects/{project}/locations/{location}/dataSourceReferences/{data_source_reference}", +"location": "path", +"pattern": "^projects/[^/]+/locations/[^/]+/dataSourceReferences/[^/]+$", +"required": true, +"type": "string" +} +}, +"path": "v1/{+name}", +"response": { +"$ref": "DataSourceReference" +}, +"scopes": [ +"https://www.googleapis.com/auth/cloud-platform" +] +} +} +}, "managementServers": { "methods": { "create": { @@ -1901,7 +2032,7 @@ } } }, -"revision": "20250611", +"revision": "20250625", "rootUrl": "https://backupdr.googleapis.com/", "schemas": { "AbandonBackupRequest": { @@ -2274,16 +2405,23 @@ "enum": [ "BACKUP_TYPE_UNSPECIFIED", "SCHEDULED", -"ON_DEMAND" +"ON_DEMAND", +"ON_DEMAND_OPERATIONAL" ], "enumDescriptions": [ "Backup type is unspecified.", "Scheduled backup.", -"On demand backup." +"On demand backup.", +"Operational backup." ], "readOnly": true, "type": "string" }, +"cloudSqlInstanceBackupProperties": { +"$ref": "CloudSqlInstanceBackupProperties", +"description": "Output only. Cloud SQL specific backup properties.", +"readOnly": true +}, "computeInstanceBackupProperties": { "$ref": "ComputeInstanceBackupProperties", "description": "Output only. Compute Engine specific backup properties.", @@ -2373,14 +2511,16 @@ "CREATING", "ACTIVE", "DELETING", -"ERROR" +"ERROR", +"UPLOADING" ], "enumDescriptions": [ "State not set.", "The backup is being created.", "The backup has been created and is fully usable.", "The backup is being deleted.", -"The backup is experiencing an issue and might be unusable." +"The backup is experiencing an issue and might be unusable.", +"The backup is being uploaded." ], "readOnly": true, "type": "string" @@ -2773,6 +2913,11 @@ "description": "Optional. This collection of key/value pairs allows for custom labels to be supplied by the user. Example, {\"tag\": \"Weekly\"}.", "type": "object" }, +"logRetentionDays": { +"description": "Optional. Required for CloudSQL resource_type Configures how long logs will be stored. It is defined in \u201cdays\u201d. This value should be greater than or equal to minimum enforced log retention duration of the backup vault.", +"format": "int64", +"type": "string" +}, "name": { "description": "Output only. Identifier. The resource name of the `BackupPlan`. Format: `projects/{project}/locations/{location}/backupPlans/{backup_plan}`", "readOnly": true, @@ -2848,6 +2993,11 @@ "readOnly": true, "type": "string" }, +"cloudSqlInstanceBackupPlanAssociationProperties": { +"$ref": "CloudSqlInstanceBackupPlanAssociationProperties", +"description": "Output only. Cloud SQL instance's backup plan association properties.", +"readOnly": true +}, "createTime": { "description": "Output only. The time when the instance was created.", "format": "google-datetime", @@ -2961,7 +3111,7 @@ "id": "BackupRule", "properties": { "backupRetentionDays": { -"description": "Required. Configures the duration for which backup data will be kept. It is defined in \u201cdays\u201d. The value should be greater than or equal to minimum enforced retention of the backup vault. Minimum value is 1 and maximum value is 36159 for custom retention on-demand backup. Minimum value is 1 and maximum value is 90 for hourly backups. Minimum value is 1 and maximum value is 186 for daily backups. Minimum value is 7 and maximum value is 366 for weekly backups. Minimum value is 30 and maximum value is 732 for monthly backups. Minimum value is 365 and maximum value is 36159 for yearly backups.", +"description": "Required. Configures the duration for which backup data will be kept. It is defined in \u201cdays\u201d. The value should be greater than or equal to minimum enforced retention of the backup vault. Minimum value is 1 and maximum value is 36159 for custom retention on-demand backup. Minimum and maximum values are workload specific for all other rules.", "format": "int32", "type": "integer" }, @@ -3143,6 +3293,123 @@ "properties": {}, "type": "object" }, +"CloudSqlInstanceBackupPlanAssociationProperties": { +"description": "Cloud SQL instance's BPA properties.", +"id": "CloudSqlInstanceBackupPlanAssociationProperties", +"properties": { +"instanceCreateTime": { +"description": "Output only. The time when the instance was created.", +"format": "google-datetime", +"readOnly": true, +"type": "string" +} +}, +"type": "object" +}, +"CloudSqlInstanceBackupProperties": { +"description": "CloudSqlInstanceBackupProperties represents Cloud SQL Instance Backup properties. .", +"id": "CloudSqlInstanceBackupProperties", +"properties": { +"databaseInstalledVersion": { +"description": "Output only. The installed database version of the Cloud SQL instance when the backup was taken.", +"readOnly": true, +"type": "string" +}, +"finalBackup": { +"description": "Output only. Whether the backup is a final backup.", +"readOnly": true, +"type": "boolean" +}, +"instanceTier": { +"description": "Output only. The tier (or machine type) for this instance. Example: `db-custom-1-3840`", +"readOnly": true, +"type": "string" +}, +"sourceInstance": { +"description": "Output only. The source instance of the backup. Format: projects/{project}/instances/{instance}", +"readOnly": true, +"type": "string" +} +}, +"type": "object" +}, +"CloudSqlInstanceDataSourceProperties": { +"description": "CloudSqlInstanceDataSourceProperties represents the properties of a Cloud SQL resource that are stored in the DataSource. .", +"id": "CloudSqlInstanceDataSourceProperties", +"properties": { +"databaseInstalledVersion": { +"description": "Output only. The installed database version of the Cloud SQL instance.", +"readOnly": true, +"type": "string" +}, +"instanceCreateTime": { +"description": "Output only. The instance creation timestamp.", +"format": "google-datetime", +"readOnly": true, +"type": "string" +}, +"instanceTier": { +"description": "Output only. The tier (or machine type) for this instance. Example: `db-custom-1-3840`", +"readOnly": true, +"type": "string" +}, +"name": { +"description": "Output only. Name of the Cloud SQL instance backed up by the datasource. Format: projects/{project}/instances/{instance}", +"readOnly": true, +"type": "string" +} +}, +"type": "object" +}, +"CloudSqlInstanceDataSourceReferenceProperties": { +"description": "CloudSqlInstanceDataSourceReferenceProperties represents the properties of a Cloud SQL resource that are stored in the DataSourceReference. .", +"id": "CloudSqlInstanceDataSourceReferenceProperties", +"properties": { +"databaseInstalledVersion": { +"description": "Output only. The installed database version of the Cloud SQL instance.", +"readOnly": true, +"type": "string" +}, +"instanceCreateTime": { +"description": "Output only. The instance creation timestamp.", +"format": "google-datetime", +"readOnly": true, +"type": "string" +}, +"instanceTier": { +"description": "Output only. The tier (or machine type) for this instance. Example: `db-custom-1-3840`", +"readOnly": true, +"type": "string" +}, +"name": { +"description": "Output only. Name of the Cloud SQL instance backed up by the datasource. Format: projects/{project}/instances/{instance}", +"readOnly": true, +"type": "string" +} +}, +"type": "object" +}, +"CloudSqlInstanceInitializationConfig": { +"description": "CloudSqlInstanceInitializationConfig contains the configuration for initializing a Cloud SQL instance.", +"id": "CloudSqlInstanceInitializationConfig", +"properties": { +"edition": { +"description": "Required. The edition of the Cloud SQL instance.", +"enum": [ +"EDITION_UNSPECIFIED", +"ENTERPRISE", +"ENTERPRISE_PLUS" +], +"enumDescriptions": [ +"Unspecified edition.", +"Enterprise edition.", +"Enterprise Plus edition." +], +"type": "string" +} +}, +"type": "object" +}, "ComputeInstanceBackupProperties": { "description": "ComputeInstanceBackupProperties represents Compute Engine instance backup properties.", "id": "ComputeInstanceBackupProperties", @@ -3591,10 +3858,47 @@ }, "type": "object" }, +"DataSourceBackupConfigInfo": { +"description": "Information of backup configuration on the DataSource.", +"id": "DataSourceBackupConfigInfo", +"properties": { +"lastBackupState": { +"description": "Output only. The status of the last backup in this DataSource", +"enum": [ +"LAST_BACKUP_STATE_UNSPECIFIED", +"FIRST_BACKUP_PENDING", +"SUCCEEDED", +"FAILED", +"PERMISSION_DENIED" +], +"enumDescriptions": [ +"Status not set.", +"The first backup has not yet completed", +"The most recent backup was successful", +"The most recent backup failed", +"The most recent backup could not be run/failed because of the lack of permissions" +], +"readOnly": true, +"type": "string" +}, +"lastSuccessfulBackupConsistencyTime": { +"description": "Output only. Timestamp of the last successful backup to this DataSource.", +"format": "google-datetime", +"readOnly": true, +"type": "string" +} +}, +"type": "object" +}, "DataSourceGcpResource": { "description": "DataSourceGcpResource is used for protected resources that are Google Cloud Resources. This name is easeier to understand than GcpResourceDataSource or GcpDataSourceResource", "id": "DataSourceGcpResource", "properties": { +"cloudSqlInstanceDatasourceProperties": { +"$ref": "CloudSqlInstanceDataSourceProperties", +"description": "Output only. CloudSqlInstanceDataSourceProperties has a subset of Cloud SQL Instance properties that are useful at the Datasource level.", +"readOnly": true +}, "computeInstanceDatasourceProperties": { "$ref": "ComputeInstanceDataSourceProperties", "description": "ComputeInstanceDataSourceProperties has a subset of Compute Instance properties that are useful at the Datasource level." @@ -3619,6 +3923,86 @@ }, "type": "object" }, +"DataSourceGcpResourceInfo": { +"description": "The GCP resource that the DataSource is associated with.", +"id": "DataSourceGcpResourceInfo", +"properties": { +"cloudSqlInstanceProperties": { +"$ref": "CloudSqlInstanceDataSourceReferenceProperties", +"description": "Output only. The properties of the Cloud SQL instance.", +"readOnly": true +}, +"gcpResourcename": { +"description": "Output only. The resource name of the GCP resource. Ex: projects/{project}/zones/{zone}/instances/{instance}", +"readOnly": true, +"type": "string" +}, +"location": { +"description": "Output only. The location of the GCP resource. Ex: //\"global\"/\"unspecified\"", +"readOnly": true, +"type": "string" +}, +"type": { +"description": "Output only. The type of the GCP resource. Ex: compute.googleapis.com/Instance", +"readOnly": true, +"type": "string" +} +}, +"type": "object" +}, +"DataSourceReference": { +"description": "DataSourceReference is a reference to a DataSource resource.", +"id": "DataSourceReference", +"properties": { +"createTime": { +"description": "Output only. The time when the DataSourceReference was created.", +"format": "google-datetime", +"readOnly": true, +"type": "string" +}, +"dataSource": { +"description": "Output only. The resource name of the DataSource. Format: projects/{project}/locations/{location}/backupVaults/{backupVault}/dataSources/{dataSource}", +"readOnly": true, +"type": "string" +}, +"dataSourceBackupConfigInfo": { +"$ref": "DataSourceBackupConfigInfo", +"description": "Output only. Information of backup configuration on the DataSource.", +"readOnly": true +}, +"dataSourceBackupConfigState": { +"description": "Output only. The backup configuration state of the DataSource.", +"enum": [ +"BACKUP_CONFIG_STATE_UNSPECIFIED", +"ACTIVE", +"PASSIVE" +], +"enumDescriptions": [ +"The possible states of backup configuration. Status not set.", +"The data source is actively protected (i.e. there is a BackupPlanAssociation or Appliance SLA pointing to it)", +"The data source is no longer protected (but may have backups under it)" +], +"readOnly": true, +"type": "string" +}, +"dataSourceBackupCount": { +"description": "Output only. Number of backups in the DataSource.", +"format": "int64", +"readOnly": true, +"type": "string" +}, +"dataSourceGcpResourceInfo": { +"$ref": "DataSourceGcpResourceInfo", +"description": "Output only. The GCP resource that the DataSource is associated with.", +"readOnly": true +}, +"name": { +"description": "Identifier. The resource name of the DataSourceReference. Format: projects/{project}/locations/{location}/dataSourceReferences/{data_source_reference}", +"type": "string" +} +}, +"type": "object" +}, "DiskBackupProperties": { "description": "DiskBackupProperties represents the properties of a Disk backup.", "id": "DiskBackupProperties", @@ -3930,6 +4314,44 @@ }, "type": "object" }, +"FetchBackupPlanAssociationsForResourceTypeResponse": { +"description": "Response for the FetchBackupPlanAssociationsForResourceType method.", +"id": "FetchBackupPlanAssociationsForResourceTypeResponse", +"properties": { +"backupPlanAssociations": { +"description": "Output only. The BackupPlanAssociations from the specified parent.", +"items": { +"$ref": "BackupPlanAssociation" +}, +"readOnly": true, +"type": "array" +}, +"nextPageToken": { +"description": "Output only. A token, which can be sent as `page_token` to retrieve the next page. If this field is omitted, there are no subsequent pages.", +"readOnly": true, +"type": "string" +} +}, +"type": "object" +}, +"FetchDataSourceReferencesForResourceTypeResponse": { +"description": "Response for the FetchDataSourceReferencesForResourceType method.", +"id": "FetchDataSourceReferencesForResourceTypeResponse", +"properties": { +"dataSourceReferences": { +"description": "The DataSourceReferences from the specified parent.", +"items": { +"$ref": "DataSourceReference" +}, +"type": "array" +}, +"nextPageToken": { +"description": "A token, which can be sent as `page_token` to retrieve the next page. If this field is omitted, there are no subsequent pages.", +"type": "string" +} +}, +"type": "object" +}, "FetchUsableBackupVaultsResponse": { "description": "Response message for fetching usable BackupVaults.", "id": "FetchUsableBackupVaultsResponse", @@ -4137,6 +4559,10 @@ "description": "Request message for initializing the service.", "id": "InitializeServiceRequest", "properties": { +"cloudSqlInstanceInitializationConfig": { +"$ref": "CloudSqlInstanceInitializationConfig", +"description": "Optional. The configuration for initializing a Cloud SQL instance." +}, "requestId": { "description": "Optional. An optional request ID to identify requests. Specify a unique request ID so that if you must retry your request, the server will know to ignore the request if it has already been completed. The server will guarantee that for at least 60 minutes since the first request. For example, consider a situation where you make an initial request and t he request times out. If you make the request again with the same request ID, the server can check if original operation with the same request ID was received, and if so, will ignore the second request. This prevents clients from accidentally creating duplicate commitments. The request ID must be a valid UUID with the exception that zero UUID is not supported (00000000-0000-0000-0000-000000000000).", "type": "string" diff --git a/googleapiclient/discovery_cache/documents/cloudasset.v1.json b/googleapiclient/discovery_cache/documents/cloudasset.v1.json index 5d4fe4b5c3..59249f8378 100644 --- a/googleapiclient/discovery_cache/documents/cloudasset.v1.json +++ b/googleapiclient/discovery_cache/documents/cloudasset.v1.json @@ -1095,7 +1095,7 @@ } } }, -"revision": "20250307", +"revision": "20250627", "rootUrl": "https://cloudasset.googleapis.com/", "schemas": { "AccessSelector": { @@ -1485,7 +1485,7 @@ "type": "string" }, "force": { -"description": "If the destination table already exists and this flag is `TRUE`, the table will be overwritten by the contents of assets snapshot. If the flag is `FALSE` or unset and the destination table already exists, the export call returns an INVALID_ARGUMEMT error.", +"description": "If the destination table already exists and this flag is `TRUE`, the table will be overwritten by the contents of assets snapshot. If the flag is `FALSE` or unset and the destination table already exists, the export call returns an INVALID_ARGUMENT error.", "type": "boolean" }, "partitionSpec": { diff --git a/googleapiclient/discovery_cache/documents/cloudasset.v1p7beta1.json b/googleapiclient/discovery_cache/documents/cloudasset.v1p7beta1.json index 354855ff04..d337b4adef 100644 --- a/googleapiclient/discovery_cache/documents/cloudasset.v1p7beta1.json +++ b/googleapiclient/discovery_cache/documents/cloudasset.v1p7beta1.json @@ -167,7 +167,7 @@ } } }, -"revision": "20250307", +"revision": "20250627", "rootUrl": "https://cloudasset.googleapis.com/", "schemas": { "AnalyzeIamPolicyLongrunningMetadata": { @@ -349,7 +349,7 @@ "type": "string" }, "force": { -"description": "If the destination table already exists and this flag is `TRUE`, the table will be overwritten by the contents of assets snapshot. If the flag is `FALSE` or unset and the destination table already exists, the export call returns an INVALID_ARGUMEMT error.", +"description": "If the destination table already exists and this flag is `TRUE`, the table will be overwritten by the contents of assets snapshot. If the flag is `FALSE` or unset and the destination table already exists, the export call returns an INVALID_ARGUMENT error.", "type": "boolean" }, "partitionSpec": { diff --git a/googleapiclient/discovery_cache/documents/cloudidentity.v1.json b/googleapiclient/discovery_cache/documents/cloudidentity.v1.json index 91d444ae83..75471c6012 100644 --- a/googleapiclient/discovery_cache/documents/cloudidentity.v1.json +++ b/googleapiclient/discovery_cache/documents/cloudidentity.v1.json @@ -2085,7 +2085,7 @@ } } }, -"revision": "20250121", +"revision": "20250701", "rootUrl": "https://cloudidentity.googleapis.com/", "schemas": { "AddIdpCredentialOperationMetadata": { @@ -3396,7 +3396,7 @@ "additionalProperties": { "type": "string" }, -"description": "Required. One or more label entries that apply to the Group. Currently supported labels contain a key with an empty value. Google Groups are the default type of group and have a label with a key of `cloudidentity.googleapis.com/groups.discussion_forum` and an empty value. Existing Google Groups can have an additional label with a key of `cloudidentity.googleapis.com/groups.security` and an empty value added to them. **This is an immutable change and the security label cannot be removed once added.** Dynamic groups have a label with a key of `cloudidentity.googleapis.com/groups.dynamic`. Identity-mapped groups for Cloud Search have a label with a key of `system/groups/external` and an empty value.", +"description": "Required. One or more label entries that apply to the Group. Labels contain a key with an empty value. Google Groups are the default type of group and have a label with a key of `cloudidentity.googleapis.com/groups.discussion_forum` and an empty value. Existing Google Groups can have an additional label with a key of `cloudidentity.googleapis.com/groups.security` and an empty value added to them. **This is an immutable change and the security label cannot be removed once added.** Dynamic groups have a label with a key of `cloudidentity.googleapis.com/groups.dynamic`. Identity-mapped groups for Cloud Search have a label with a key of `system/groups/external` and an empty value. Google Groups can be [locked](https://support.google.com/a?p=locked-groups). To lock a group, add a label with a key of `cloudidentity.googleapis.com/groups.locked` and an empty value. Doing so locks the group. To unlock the group, remove this label.", "type": "object" }, "name": { diff --git a/googleapiclient/discovery_cache/documents/cloudidentity.v1beta1.json b/googleapiclient/discovery_cache/documents/cloudidentity.v1beta1.json index b4f6bdcf5a..23a24e349b 100644 --- a/googleapiclient/discovery_cache/documents/cloudidentity.v1beta1.json +++ b/googleapiclient/discovery_cache/documents/cloudidentity.v1beta1.json @@ -2123,7 +2123,7 @@ } } }, -"revision": "20250121", +"revision": "20250701", "rootUrl": "https://cloudidentity.googleapis.com/", "schemas": { "AddIdpCredentialOperationMetadata": { @@ -4128,7 +4128,7 @@ "additionalProperties": { "type": "string" }, -"description": "Required. One or more label entries that apply to the Group. Labels contain a key with an empty value. Google Groups are the default type of group and have a label with a key of `cloudidentity.googleapis.com/groups.discussion_forum` and an empty value. Existing Google Groups can have an additional label with a key of `cloudidentity.googleapis.com/groups.security` and an empty value added to them. **This is an immutable change and the security label cannot be removed once added.** Dynamic groups have a label with a key of `cloudidentity.googleapis.com/groups.dynamic`. Identity-mapped groups for Cloud Search have a label with a key of `system/groups/external` and an empty value. (Beta) Google Groups can be [locked](https://support.google.com/a?p=locked-groups). To lock a group, add a label with a key of `cloudidentity.googleapis.com/groups.locked` and an empty value. Doing so locks the group. To unlock the group, remove this label.", +"description": "Required. One or more label entries that apply to the Group. Labels contain a key with an empty value. Google Groups are the default type of group and have a label with a key of `cloudidentity.googleapis.com/groups.discussion_forum` and an empty value. Existing Google Groups can have an additional label with a key of `cloudidentity.googleapis.com/groups.security` and an empty value added to them. **This is an immutable change and the security label cannot be removed once added.** Dynamic groups have a label with a key of `cloudidentity.googleapis.com/groups.dynamic`. Identity-mapped groups for Cloud Search have a label with a key of `system/groups/external` and an empty value. Google Groups can be [locked](https://support.google.com/a?p=locked-groups). To lock a group, add a label with a key of `cloudidentity.googleapis.com/groups.locked` and an empty value. Doing so locks the group. To unlock the group, remove this label.", "type": "object" }, "name": { diff --git a/googleapiclient/discovery_cache/documents/composer.v1.json b/googleapiclient/discovery_cache/documents/composer.v1.json index 1c56b2a95f..67935e0281 100644 --- a/googleapiclient/discovery_cache/documents/composer.v1.json +++ b/googleapiclient/discovery_cache/documents/composer.v1.json @@ -68,6 +68,11 @@ }, { "description": "Regional Endpoint", +"endpointUrl": "https://composer.asia-southeast3.rep.googleapis.com/", +"location": "asia-southeast3" +}, +{ +"description": "Regional Endpoint", "endpointUrl": "https://composer.australia-southeast1.rep.googleapis.com/", "location": "australia-southeast1" }, @@ -103,6 +108,21 @@ }, { "description": "Regional Endpoint", +"endpointUrl": "https://composer.europe-west10.rep.googleapis.com/", +"location": "europe-west10" +}, +{ +"description": "Regional Endpoint", +"endpointUrl": "https://composer.europe-west12.rep.googleapis.com/", +"location": "europe-west12" +}, +{ +"description": "Regional Endpoint", +"endpointUrl": "https://composer.europe-west15.rep.googleapis.com/", +"location": "europe-west15" +}, +{ +"description": "Regional Endpoint", "endpointUrl": "https://composer.europe-west2.rep.googleapis.com/", "location": "europe-west2" }, @@ -133,16 +153,6 @@ }, { "description": "Regional Endpoint", -"endpointUrl": "https://composer.europe-west10.rep.googleapis.com/", -"location": "europe-west10" -}, -{ -"description": "Regional Endpoint", -"endpointUrl": "https://composer.europe-west12.rep.googleapis.com/", -"location": "europe-west12" -}, -{ -"description": "Regional Endpoint", "endpointUrl": "https://composer.me-central1.rep.googleapis.com/", "location": "me-central1" }, @@ -1216,7 +1226,7 @@ } } }, -"revision": "20250513", +"revision": "20250630", "rootUrl": "https://composer.googleapis.com/", "schemas": { "AirflowMetadataRetentionPolicyConfig": { diff --git a/googleapiclient/discovery_cache/documents/composer.v1beta1.json b/googleapiclient/discovery_cache/documents/composer.v1beta1.json index 30d91482b7..94e0f92013 100644 --- a/googleapiclient/discovery_cache/documents/composer.v1beta1.json +++ b/googleapiclient/discovery_cache/documents/composer.v1beta1.json @@ -68,6 +68,11 @@ }, { "description": "Regional Endpoint", +"endpointUrl": "https://composer.asia-southeast3.rep.googleapis.com/", +"location": "asia-southeast3" +}, +{ +"description": "Regional Endpoint", "endpointUrl": "https://composer.australia-southeast1.rep.googleapis.com/", "location": "australia-southeast1" }, @@ -103,6 +108,21 @@ }, { "description": "Regional Endpoint", +"endpointUrl": "https://composer.europe-west10.rep.googleapis.com/", +"location": "europe-west10" +}, +{ +"description": "Regional Endpoint", +"endpointUrl": "https://composer.europe-west12.rep.googleapis.com/", +"location": "europe-west12" +}, +{ +"description": "Regional Endpoint", +"endpointUrl": "https://composer.europe-west15.rep.googleapis.com/", +"location": "europe-west15" +}, +{ +"description": "Regional Endpoint", "endpointUrl": "https://composer.europe-west2.rep.googleapis.com/", "location": "europe-west2" }, @@ -133,16 +153,6 @@ }, { "description": "Regional Endpoint", -"endpointUrl": "https://composer.europe-west10.rep.googleapis.com/", -"location": "europe-west10" -}, -{ -"description": "Regional Endpoint", -"endpointUrl": "https://composer.europe-west12.rep.googleapis.com/", -"location": "europe-west12" -}, -{ -"description": "Regional Endpoint", "endpointUrl": "https://composer.me-central1.rep.googleapis.com/", "location": "me-central1" }, @@ -1216,7 +1226,7 @@ } } }, -"revision": "20250513", +"revision": "20250630", "rootUrl": "https://composer.googleapis.com/", "schemas": { "AirflowMetadataRetentionPolicyConfig": { diff --git a/googleapiclient/discovery_cache/documents/contactcenterinsights.v1.json b/googleapiclient/discovery_cache/documents/contactcenterinsights.v1.json index 4f08756c05..51eebfb697 100644 --- a/googleapiclient/discovery_cache/documents/contactcenterinsights.v1.json +++ b/googleapiclient/discovery_cache/documents/contactcenterinsights.v1.json @@ -5300,7 +5300,7 @@ } } }, -"revision": "20250624", +"revision": "20250701", "rootUrl": "https://contactcenterinsights.googleapis.com/", "schemas": { "GoogleCloudContactcenterinsightsV1Analysis": { @@ -5755,7 +5755,7 @@ true "id": "GoogleCloudContactcenterinsightsV1AuthorizedView", "properties": { "conversationFilter": { -"description": "A filter to reduce conversation results to a specific subset. The AuthorizedView's assigned permission (read/write) could be applied to the subset of conversations. If conversation_filter is empty, there is no restriction on the conversations that the AuthorizedView can access. Having *authorizedViews.get* access to the AuthorizedView means having the same read/write access to the Conversations (as well as metadata/annotations liked to the conversation) that this AuthorizedView has.", +"description": "A filter to reduce conversation results to a specific subset. The AuthorizedView's assigned permission (read/write) could be applied to the subset of conversations. If conversation_filter is empty, there is no restriction on the conversations that the AuthorizedView can access. Having *authorizedViews.get* access to the AuthorizedView means having the same read/write access to the Conversations (as well as metadata/annotations linked to the conversation) that this AuthorizedView has.", "type": "string" }, "createTime": { diff --git a/googleapiclient/discovery_cache/documents/datamigration.v1.json b/googleapiclient/discovery_cache/documents/datamigration.v1.json index a50cfd23ee..f594af21c0 100644 --- a/googleapiclient/discovery_cache/documents/datamigration.v1.json +++ b/googleapiclient/discovery_cache/documents/datamigration.v1.json @@ -2132,6 +2132,11 @@ "description": "Optional. If set to true, will skip validations.", "location": "query", "type": "boolean" +}, +"validateOnly": { +"description": "Optional. For PSC Interface only - get the tenant project before creating the resource.", +"location": "query", +"type": "boolean" } }, "path": "v1/{+parent}/privateConnections", @@ -2340,7 +2345,7 @@ } } }, -"revision": "20250619", +"revision": "20250625", "rootUrl": "https://datamigration.googleapis.com/", "schemas": { "AlloyDbConnectionProfile": { @@ -5809,6 +5814,10 @@ "description": "The name of the resource.", "type": "string" }, +"pscInterfaceConfig": { +"$ref": "PscInterfaceConfig", +"description": "PSC Interface configuration." +}, "satisfiesPzi": { "description": "Output only. Reserved for future use.", "readOnly": true, @@ -5888,6 +5897,17 @@ }, "type": "object" }, +"PscInterfaceConfig": { +"description": "The PSC Interface configuration is used to create PSC Interface between DMS's internal VPC and the consumer's PSC.", +"id": "PscInterfaceConfig", +"properties": { +"networkAttachment": { +"description": "Required. Fully qualified name of the Network Attachment that DMS will connect to. Format: `projects/{{project}}/regions/{{region}}/networkAttachments/{{name}}`", +"type": "string" +} +}, +"type": "object" +}, "RestartMigrationJobRequest": { "description": "Request message for 'RestartMigrationJob' request.", "id": "RestartMigrationJobRequest", diff --git a/googleapiclient/discovery_cache/documents/dataproc.v1.json b/googleapiclient/discovery_cache/documents/dataproc.v1.json index bcce734a60..917aa50c76 100644 --- a/googleapiclient/discovery_cache/documents/dataproc.v1.json +++ b/googleapiclient/discovery_cache/documents/dataproc.v1.json @@ -2760,6 +2760,13 @@ "name" ], "parameters": { +"jobIds": { +"description": "Optional. List of Job IDs to filter by if provided.", +"format": "int64", +"location": "query", +"repeated": true, +"type": "string" +}, "name": { "description": "Required. The fully qualified name of the session to retrieve in the format \"projects/PROJECT_ID/locations/DATAPROC_REGION/sessions/SESSION_ID/sparkApplications/APPLICATION_ID\"", "location": "path", @@ -2843,6 +2850,13 @@ "description": "Required. Parent (Session) resource reference.", "location": "query", "type": "string" +}, +"stageIds": { +"description": "Optional. List of Stage IDs to filter by if provided.", +"format": "int64", +"location": "query", +"repeated": true, +"type": "string" } }, "path": "v1/{+name}:summarizeStages", @@ -4998,7 +5012,7 @@ } } }, -"revision": "20250604", +"revision": "20250624", "rootUrl": "https://dataproc.googleapis.com/", "schemas": { "AcceleratorConfig": { @@ -8236,6 +8250,16 @@ "format": "google-duration", "type": "string" }, +"autoStopTime": { +"description": "Optional. The time when cluster will be auto-stopped (see JSON representation of Timestamp (https://developers.google.com/protocol-buffers/docs/proto3#json)).", +"format": "google-datetime", +"type": "string" +}, +"autoStopTtl": { +"description": "Optional. The lifetime duration of the cluster. The cluster will be auto-stopped at the end of this period, calculated from the time of submission of the create or update cluster request. Minimum value is 10 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)).", +"format": "google-duration", +"type": "string" +}, "idleDeleteTtl": { "description": "Optional. The duration to keep the cluster alive while idling (when no jobs are running). Passing this threshold will cause the cluster to be deleted. Minimum value is 5 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)).", "format": "google-duration", @@ -8246,6 +8270,11 @@ "format": "google-datetime", "readOnly": true, "type": "string" +}, +"idleStopTtl": { +"description": "Optional. The duration to keep the cluster started while idling (when no jobs are running). Passing this threshold will cause the cluster to be stopped. Minimum value is 5 minutes; maximum value is 14 days (see JSON representation of Duration (https://developers.google.com/protocol-buffers/docs/proto3#json)).", +"format": "google-duration", +"type": "string" } }, "type": "object" diff --git a/googleapiclient/discovery_cache/documents/datastream.v1.json b/googleapiclient/discovery_cache/documents/datastream.v1.json index ba55ecdc88..33857c7262 100644 --- a/googleapiclient/discovery_cache/documents/datastream.v1.json +++ b/googleapiclient/discovery_cache/documents/datastream.v1.json @@ -1261,7 +1261,7 @@ } } }, -"revision": "20250518", +"revision": "20250627", "rootUrl": "https://datastream.googleapis.com/", "schemas": { "AppendOnly": { @@ -3465,6 +3465,10 @@ "datasetTemplate": { "$ref": "DatasetTemplate", "description": "The dataset template to use for dynamic dataset creation." +}, +"projectId": { +"description": "Optional. The project id of the BigQuery dataset. If not specified, the project will be inferred from the stream resource.", +"type": "string" } }, "type": "object" diff --git a/googleapiclient/discovery_cache/documents/dialogflow.v2.json b/googleapiclient/discovery_cache/documents/dialogflow.v2.json index 2eae0ea69e..05cfdcfd31 100644 --- a/googleapiclient/discovery_cache/documents/dialogflow.v2.json +++ b/googleapiclient/discovery_cache/documents/dialogflow.v2.json @@ -8805,7 +8805,7 @@ } } }, -"revision": "20250625", +"revision": "20250627", "rootUrl": "https://dialogflow.googleapis.com/", "schemas": { "GoogleCloudDialogflowCxV3AdvancedSettings": { @@ -12263,7 +12263,7 @@ ], "enumDescriptions": [ "Not specified.", -"Uncompressed 16-bit signed little-endian samples (Linear PCM).", +"Uncompressed 16-bit signed little-endian samples (Linear PCM). LINT: LEGACY_NAMES", "[`FLAC`](https://xiph.org/flac/documentation.html) (Free Lossless Audio Codec) is the recommended encoding because it is lossless (therefore recognition is not compromised) and requires only about half the bandwidth of `LINEAR16`. `FLAC` stream encoding supports 16-bit and 24-bit samples, however, not all fields in `STREAMINFO` are supported.", "8-bit samples that compand 14-bit audio samples using G.711 PCMU/mu-law.", "Adaptive Multi-Rate Narrowband codec. `sample_rate_hertz` must be 8000.", @@ -15033,7 +15033,7 @@ true "type": "array" }, "sdp": { -"description": "Optional. SDP of the call. It's initially the SDP answer to the endpoint, but maybe later updated for the purpose of making the link active, etc.", +"description": "Optional. SDP of the call. It's initially the SDP answer to the incoming call, but maybe later updated for the purpose of making the link active, etc.", "type": "string" }, "sipHeaders": { @@ -16319,7 +16319,7 @@ true "id": "GoogleCloudDialogflowV2HumanAgentAssistantConfigConversationModelConfig", "properties": { "baselineModelVersion": { -"description": "Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0", +"description": "Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0", "type": "string" }, "model": { diff --git a/googleapiclient/discovery_cache/documents/dialogflow.v2beta1.json b/googleapiclient/discovery_cache/documents/dialogflow.v2beta1.json index 0f26fd0219..47287b8938 100644 --- a/googleapiclient/discovery_cache/documents/dialogflow.v2beta1.json +++ b/googleapiclient/discovery_cache/documents/dialogflow.v2beta1.json @@ -8602,7 +8602,7 @@ } } }, -"revision": "20250625", +"revision": "20250627", "rootUrl": "https://dialogflow.googleapis.com/", "schemas": { "GoogleCloudDialogflowCxV3AdvancedSettings": { @@ -12060,7 +12060,7 @@ ], "enumDescriptions": [ "Not specified.", -"Uncompressed 16-bit signed little-endian samples (Linear PCM).", +"Uncompressed 16-bit signed little-endian samples (Linear PCM). LINT: LEGACY_NAMES", "[`FLAC`](https://xiph.org/flac/documentation.html) (Free Lossless Audio Codec) is the recommended encoding because it is lossless (therefore recognition is not compromised) and requires only about half the bandwidth of `LINEAR16`. `FLAC` stream encoding supports 16-bit and 24-bit samples, however, not all fields in `STREAMINFO` are supported.", "8-bit samples that compand 14-bit audio samples using G.711 PCMU/mu-law.", "Adaptive Multi-Rate Narrowband codec. `sample_rate_hertz` must be 8000.", @@ -17472,7 +17472,7 @@ true "type": "array" }, "sdp": { -"description": "Optional. SDP of the call. It's initially the SDP answer to the endpoint, but maybe later updated for the purpose of making the link active, etc.", +"description": "Optional. SDP of the call. It's initially the SDP answer to the incoming call, but maybe later updated for the purpose of making the link active, etc.", "type": "string" }, "sipHeaders": { @@ -18566,7 +18566,7 @@ true "id": "GoogleCloudDialogflowV2beta1HumanAgentAssistantConfigConversationModelConfig", "properties": { "baselineModelVersion": { -"description": "Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0", +"description": "Version of current baseline model. It will be ignored if model is set. Valid versions are: - Article Suggestion baseline model: - 0.9 - 1.0 (default) - Summarization baseline model: - 1.0", "type": "string" }, "model": { @@ -21187,9 +21187,9 @@ true ], "enumDescriptions": [ "Not specified.", -"Uncompressed 16-bit signed little-endian samples (Linear PCM). Audio content returned as LINEAR16 also contains a WAV header.", +"Uncompressed 16-bit signed little-endian samples (Linear PCM). Audio content returned as LINEAR16 also contains a WAV header. LINT: LEGACY_NAMES", "MP3 audio at 32kbps.", -"MP3 audio at 64kbps.", +"MP3 audio at 64kbps. LINT: LEGACY_NAMES", "Opus encoded audio wrapped in an ogg container. The result will be a file which can be played natively on Android, and in browsers (at least Chrome and Firefox). The quality of the encoding is considerably higher than MP3 while using approximately the same bitrate.", "8-bit samples that compand 14-bit audio samples using G.711 PCMU/mu-law.", "8-bit samples that compand 13-bit audio samples using G.711 PCMU/a-law." @@ -23041,9 +23041,9 @@ true ], "enumDescriptions": [ "Not specified.", -"Uncompressed 16-bit signed little-endian samples (Linear PCM). Audio content returned as LINEAR16 also contains a WAV header.", +"Uncompressed 16-bit signed little-endian samples (Linear PCM). Audio content returned as LINEAR16 also contains a WAV header. LINT: LEGACY_NAMES", "MP3 audio at 32kbps.", -"MP3 audio at 64kbps.", +"MP3 audio at 64kbps. LINT: LEGACY_NAMES", "Opus encoded audio wrapped in an ogg container. The result will be a file which can be played natively on Android, and in browsers (at least Chrome and Firefox). The quality of the encoding is considerably higher than MP3 while using approximately the same bitrate.", "8-bit samples that compand 14-bit audio samples using G.711 PCMU/mu-law.", "8-bit samples that compand 13-bit audio samples using G.711 PCMU/a-law." @@ -23254,10 +23254,6 @@ true "description": "Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices).", "type": "string" }, -"secondLanguageModel": { -"description": "Optional. The name of the second voice (non-English) if language detection is supported.", -"type": "string" -}, "ssmlGender": { "description": "Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request.", "enum": [ diff --git a/googleapiclient/discovery_cache/documents/dialogflow.v3.json b/googleapiclient/discovery_cache/documents/dialogflow.v3.json index 0ea508712a..4b717dbe68 100644 --- a/googleapiclient/discovery_cache/documents/dialogflow.v3.json +++ b/googleapiclient/discovery_cache/documents/dialogflow.v3.json @@ -5309,7 +5309,7 @@ } } }, -"revision": "20250625", +"revision": "20250627", "rootUrl": "https://dialogflow.googleapis.com/", "schemas": { "GoogleCloudDialogflowCxV3Action": { @@ -6438,6 +6438,14 @@ "description": "Uri for the document. Present if specified for the document.", "type": "string" }, +"metadata": { +"additionalProperties": { +"description": "Properties of the object.", +"type": "any" +}, +"description": "Metadata associated with the document.", +"type": "object" +}, "text": { "description": "Text included in the prompt.", "type": "string" @@ -9349,9 +9357,9 @@ false ], "enumDescriptions": [ "Not specified.", -"Uncompressed 16-bit signed little-endian samples (Linear PCM). Audio content returned as LINEAR16 also contains a WAV header.", +"Uncompressed 16-bit signed little-endian samples (Linear PCM). Audio content returned as LINEAR16 also contains a WAV header. LINT: LEGACY_NAMES", "MP3 audio at 32kbps.", -"MP3 audio at 64kbps.", +"MP3 audio at 64kbps. LINT: LEGACY_NAMES", "Opus encoded audio wrapped in an ogg container. The result will be a file which can be played natively on Android, and in browsers (at least Chrome and Firefox). The quality of the encoding is considerably higher than MP3 while using approximately the same bitrate.", "8-bit samples that compand 14-bit audio samples using G.711 PCMU/mu-law.", "8-bit samples that compand 13-bit audio samples using G.711 PCMU/a-law." @@ -11917,10 +11925,6 @@ false "description": "Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices).", "type": "string" }, -"secondLanguageModel": { -"description": "Optional. The name of the second voice (non-English) if language detection is supported.", -"type": "string" -}, "ssmlGender": { "description": "Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer substitutes a voice with a different gender rather than failing the request.", "enum": [ @@ -13360,7 +13364,7 @@ false ], "enumDescriptions": [ "Not specified.", -"Uncompressed 16-bit signed little-endian samples (Linear PCM).", +"Uncompressed 16-bit signed little-endian samples (Linear PCM). LINT: LEGACY_NAMES", "[`FLAC`](https://xiph.org/flac/documentation.html) (Free Lossless Audio Codec) is the recommended encoding because it is lossless (therefore recognition is not compromised) and requires only about half the bandwidth of `LINEAR16`. `FLAC` stream encoding supports 16-bit and 24-bit samples, however, not all fields in `STREAMINFO` are supported.", "8-bit samples that compand 14-bit audio samples using G.711 PCMU/mu-law.", "Adaptive Multi-Rate Narrowband codec. `sample_rate_hertz` must be 8000.", diff --git a/googleapiclient/discovery_cache/documents/dialogflow.v3beta1.json b/googleapiclient/discovery_cache/documents/dialogflow.v3beta1.json index 14369afce6..79adb4ba4d 100644 --- a/googleapiclient/discovery_cache/documents/dialogflow.v3beta1.json +++ b/googleapiclient/discovery_cache/documents/dialogflow.v3beta1.json @@ -5436,7 +5436,7 @@ } } }, -"revision": "20250625", +"revision": "20250627", "rootUrl": "https://dialogflow.googleapis.com/", "schemas": { "GoogleCloudDialogflowCxV3AdvancedSettings": { @@ -9536,6 +9536,14 @@ "description": "Uri for the document. Present if specified for the document.", "type": "string" }, +"metadata": { +"additionalProperties": { +"description": "Properties of the object.", +"type": "any" +}, +"description": "Metadata associated with the document.", +"type": "object" +}, "text": { "description": "Text included in the prompt.", "type": "string" @@ -11705,7 +11713,7 @@ false ], "enumDescriptions": [ "Not specified.", -"Uncompressed 16-bit signed little-endian samples (Linear PCM).", +"Uncompressed 16-bit signed little-endian samples (Linear PCM). LINT: LEGACY_NAMES", "[`FLAC`](https://xiph.org/flac/documentation.html) (Free Lossless Audio Codec) is the recommended encoding because it is lossless (therefore recognition is not compromised) and requires only about half the bandwidth of `LINEAR16`. `FLAC` stream encoding supports 16-bit and 24-bit samples, however, not all fields in `STREAMINFO` are supported.", "8-bit samples that compand 14-bit audio samples using G.711 PCMU/mu-law.", "Adaptive Multi-Rate Narrowband codec. `sample_rate_hertz` must be 8000.", @@ -12725,9 +12733,9 @@ false ], "enumDescriptions": [ "Not specified.", -"Uncompressed 16-bit signed little-endian samples (Linear PCM). Audio content returned as LINEAR16 also contains a WAV header.", +"Uncompressed 16-bit signed little-endian samples (Linear PCM). Audio content returned as LINEAR16 also contains a WAV header. LINT: LEGACY_NAMES", "MP3 audio at 32kbps.", -"MP3 audio at 64kbps.", +"MP3 audio at 64kbps. LINT: LEGACY_NAMES", "Opus encoded audio wrapped in an ogg container. The result will be a file which can be played natively on Android, and in browsers (at least Chrome and Firefox). The quality of the encoding is considerably higher than MP3 while using approximately the same bitrate.", "8-bit samples that compand 14-bit audio samples using G.711 PCMU/mu-law.", "8-bit samples that compand 13-bit audio samples using G.711 PCMU/a-law." @@ -15717,10 +15725,6 @@ false "description": "Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices).", "type": "string" }, -"secondLanguageModel": { -"description": "Optional. The name of the second voice (non-English) if language detection is supported.", -"type": "string" -}, "ssmlGender": { "description": "Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request.", "enum": [ diff --git a/googleapiclient/discovery_cache/documents/discoveryengine.v1.json b/googleapiclient/discovery_cache/documents/discoveryengine.v1.json index 3b98fd4101..af9de45bf7 100644 --- a/googleapiclient/discovery_cache/documents/discoveryengine.v1.json +++ b/googleapiclient/discovery_cache/documents/discoveryengine.v1.json @@ -7052,7 +7052,7 @@ } } }, -"revision": "20250619", +"revision": "20250701", "rootUrl": "https://discoveryengine.googleapis.com/", "schemas": { "GoogleApiDistribution": { @@ -15945,7 +15945,7 @@ "id": "GoogleCloudDiscoveryengineV1alphaConnectorRunEntityRun", "properties": { "deletedRecordCount": { -"description": "The number of documents deleted.", +"description": "Optional. The number of documents deleted.", "format": "int64", "type": "string" }, @@ -15954,7 +15954,7 @@ "type": "string" }, "errorRecordCount": { -"description": "The total number of documents failed at sync at any stage (extraction, indexing, etc).", +"description": "Optional. The total number of documents failed at sync at indexing stage.", "format": "int64", "type": "string" }, @@ -15966,12 +15966,12 @@ "type": "array" }, "extractedRecordCount": { -"description": "The number of documents extracted from connector source, ready to be ingested to VAIS.", +"description": "Optional. The number of documents extracted from connector source, ready to be ingested to VAIS.", "format": "int64", "type": "string" }, "indexedRecordCount": { -"description": "The number of documents indexed.", +"description": "Optional. The number of documents indexed.", "format": "int64", "type": "string" }, @@ -15980,12 +15980,12 @@ "description": "Metadata to generate the progress bar." }, "scheduledRecordCount": { -"description": "The number of documents scheduled to be crawled/extracted from connector source. This only applies to third party connectors.", +"description": "Optional. The number of documents scheduled to be crawled/extracted from connector source. This only applies to third party connectors.", "format": "int64", "type": "string" }, "sourceApiRequestCount": { -"description": "The number of requests sent to 3p API.", +"description": "Optional. The number of requests sent to 3p API.", "format": "int64", "type": "string" }, @@ -16517,7 +16517,8 @@ "GOOGLE_DRIVE", "NATIVE_CLOUD_IDENTITY", "THIRD_PARTY_FEDERATED", -"THIRD_PARTY_EUA" +"THIRD_PARTY_EUA", +"GCNV" ], "enumDescriptions": [ "Default value.", @@ -16530,7 +16531,8 @@ "Google Drive connector.", "Native Cloud Identity connector for people search powered by People API.", "Federated connector, it is a third party connector that doesn't ingestion data, and search is powered by third party application's API.", -"Connector utilized for End User Authentication features." +"Connector utilized for End User Authentication features.", +"Google Cloud NetApp Volumes connector." ], "readOnly": true, "type": "string" @@ -16747,6 +16749,10 @@ }, "description": "Optional. Any authentication parameters specific to EUA connectors.", "type": "object" +}, +"tenant": { +"$ref": "GoogleCloudDiscoveryengineV1alphaTenant", +"description": "Optional. The tenant project the connector is connected to." } }, "type": "object" @@ -20365,6 +20371,25 @@ false }, "type": "object" }, +"GoogleCloudDiscoveryengineV1alphaTenant": { +"description": "Tenant information for a connector source. This includes some of the same information stored in the Credential message, but is limited to only what is needed to provide a list of accessible tenants to the user.", +"id": "GoogleCloudDiscoveryengineV1alphaTenant", +"properties": { +"displayName": { +"description": "Optional display name for the tenant, e.g. \"My Slack Team\".", +"type": "string" +}, +"id": { +"description": "The tenant's instance ID. Examples: Jira (\"8594f221-9797-5f78-1fa4-485e198d7cd0\"), Slack (\"T123456\").", +"type": "string" +}, +"uri": { +"description": "The URI of the tenant, if applicable. For example, the URI of a Jira instance is https://my-jira-instance.atlassian.net, and a Slack tenant does not have a URI.", +"type": "string" +} +}, +"type": "object" +}, "GoogleCloudDiscoveryengineV1alphaTrainCustomModelMetadata": { "description": "Metadata related to the progress of the TrainCustomModel operation. This is returned by the google.longrunning.Operation.metadata field.", "id": "GoogleCloudDiscoveryengineV1alphaTrainCustomModelMetadata", diff --git a/googleapiclient/discovery_cache/documents/discoveryengine.v1alpha.json b/googleapiclient/discovery_cache/documents/discoveryengine.v1alpha.json index b1faacb6fc..a0e2860ed0 100644 --- a/googleapiclient/discovery_cache/documents/discoveryengine.v1alpha.json +++ b/googleapiclient/discovery_cache/documents/discoveryengine.v1alpha.json @@ -9173,7 +9173,7 @@ } } }, -"revision": "20250619", +"revision": "20250701", "rootUrl": "https://discoveryengine.googleapis.com/", "schemas": { "ApiservingMediaRequestInfo": { @@ -15716,7 +15716,7 @@ "id": "GoogleCloudDiscoveryengineV1alphaConnectorRunEntityRun", "properties": { "deletedRecordCount": { -"description": "The number of documents deleted.", +"description": "Optional. The number of documents deleted.", "format": "int64", "type": "string" }, @@ -15725,7 +15725,7 @@ "type": "string" }, "errorRecordCount": { -"description": "The total number of documents failed at sync at any stage (extraction, indexing, etc).", +"description": "Optional. The total number of documents failed at sync at indexing stage.", "format": "int64", "type": "string" }, @@ -15737,12 +15737,12 @@ "type": "array" }, "extractedRecordCount": { -"description": "The number of documents extracted from connector source, ready to be ingested to VAIS.", +"description": "Optional. The number of documents extracted from connector source, ready to be ingested to VAIS.", "format": "int64", "type": "string" }, "indexedRecordCount": { -"description": "The number of documents indexed.", +"description": "Optional. The number of documents indexed.", "format": "int64", "type": "string" }, @@ -15751,12 +15751,12 @@ "description": "Metadata to generate the progress bar." }, "scheduledRecordCount": { -"description": "The number of documents scheduled to be crawled/extracted from connector source. This only applies to third party connectors.", +"description": "Optional. The number of documents scheduled to be crawled/extracted from connector source. This only applies to third party connectors.", "format": "int64", "type": "string" }, "sourceApiRequestCount": { -"description": "The number of requests sent to 3p API.", +"description": "Optional. The number of requests sent to 3p API.", "format": "int64", "type": "string" }, @@ -16550,7 +16550,8 @@ "GOOGLE_DRIVE", "NATIVE_CLOUD_IDENTITY", "THIRD_PARTY_FEDERATED", -"THIRD_PARTY_EUA" +"THIRD_PARTY_EUA", +"GCNV" ], "enumDescriptions": [ "Default value.", @@ -16563,7 +16564,8 @@ "Google Drive connector.", "Native Cloud Identity connector for people search powered by People API.", "Federated connector, it is a third party connector that doesn't ingestion data, and search is powered by third party application's API.", -"Connector utilized for End User Authentication features." +"Connector utilized for End User Authentication features.", +"Google Cloud NetApp Volumes connector." ], "readOnly": true, "type": "string" @@ -16780,6 +16782,10 @@ }, "description": "Optional. Any authentication parameters specific to EUA connectors.", "type": "object" +}, +"tenant": { +"$ref": "GoogleCloudDiscoveryengineV1alphaTenant", +"description": "Optional. The tenant project the connector is connected to." } }, "type": "object" @@ -23469,6 +23475,25 @@ false }, "type": "object" }, +"GoogleCloudDiscoveryengineV1alphaTenant": { +"description": "Tenant information for a connector source. This includes some of the same information stored in the Credential message, but is limited to only what is needed to provide a list of accessible tenants to the user.", +"id": "GoogleCloudDiscoveryengineV1alphaTenant", +"properties": { +"displayName": { +"description": "Optional display name for the tenant, e.g. \"My Slack Team\".", +"type": "string" +}, +"id": { +"description": "The tenant's instance ID. Examples: Jira (\"8594f221-9797-5f78-1fa4-485e198d7cd0\"), Slack (\"T123456\").", +"type": "string" +}, +"uri": { +"description": "The URI of the tenant, if applicable. For example, the URI of a Jira instance is https://my-jira-instance.atlassian.net, and a Slack tenant does not have a URI.", +"type": "string" +} +}, +"type": "object" +}, "GoogleCloudDiscoveryengineV1alphaTextInput": { "description": "Defines text input.", "id": "GoogleCloudDiscoveryengineV1alphaTextInput", @@ -24228,6 +24253,11 @@ false "description": "The name of the data source, retrieved from `Collection.data_connector.data_source`.", "type": "string" }, +"dataSourceDisplayName": { +"description": "Output only. The display name of the data source.", +"readOnly": true, +"type": "string" +}, "dataStoreComponents": { "description": "For the data store collection, list of the children data stores.", "items": { diff --git a/googleapiclient/discovery_cache/documents/discoveryengine.v1beta.json b/googleapiclient/discovery_cache/documents/discoveryengine.v1beta.json index 7d69e6ec24..6275e19e79 100644 --- a/googleapiclient/discovery_cache/documents/discoveryengine.v1beta.json +++ b/googleapiclient/discovery_cache/documents/discoveryengine.v1beta.json @@ -8004,7 +8004,7 @@ } } }, -"revision": "20250619", +"revision": "20250701", "rootUrl": "https://discoveryengine.googleapis.com/", "schemas": { "GoogleApiDistribution": { @@ -12126,7 +12126,7 @@ "id": "GoogleCloudDiscoveryengineV1alphaConnectorRunEntityRun", "properties": { "deletedRecordCount": { -"description": "The number of documents deleted.", +"description": "Optional. The number of documents deleted.", "format": "int64", "type": "string" }, @@ -12135,7 +12135,7 @@ "type": "string" }, "errorRecordCount": { -"description": "The total number of documents failed at sync at any stage (extraction, indexing, etc).", +"description": "Optional. The total number of documents failed at sync at indexing stage.", "format": "int64", "type": "string" }, @@ -12147,12 +12147,12 @@ "type": "array" }, "extractedRecordCount": { -"description": "The number of documents extracted from connector source, ready to be ingested to VAIS.", +"description": "Optional. The number of documents extracted from connector source, ready to be ingested to VAIS.", "format": "int64", "type": "string" }, "indexedRecordCount": { -"description": "The number of documents indexed.", +"description": "Optional. The number of documents indexed.", "format": "int64", "type": "string" }, @@ -12161,12 +12161,12 @@ "description": "Metadata to generate the progress bar." }, "scheduledRecordCount": { -"description": "The number of documents scheduled to be crawled/extracted from connector source. This only applies to third party connectors.", +"description": "Optional. The number of documents scheduled to be crawled/extracted from connector source. This only applies to third party connectors.", "format": "int64", "type": "string" }, "sourceApiRequestCount": { -"description": "The number of requests sent to 3p API.", +"description": "Optional. The number of requests sent to 3p API.", "format": "int64", "type": "string" }, @@ -12698,7 +12698,8 @@ "GOOGLE_DRIVE", "NATIVE_CLOUD_IDENTITY", "THIRD_PARTY_FEDERATED", -"THIRD_PARTY_EUA" +"THIRD_PARTY_EUA", +"GCNV" ], "enumDescriptions": [ "Default value.", @@ -12711,7 +12712,8 @@ "Google Drive connector.", "Native Cloud Identity connector for people search powered by People API.", "Federated connector, it is a third party connector that doesn't ingestion data, and search is powered by third party application's API.", -"Connector utilized for End User Authentication features." +"Connector utilized for End User Authentication features.", +"Google Cloud NetApp Volumes connector." ], "readOnly": true, "type": "string" @@ -12928,6 +12930,10 @@ }, "description": "Optional. Any authentication parameters specific to EUA connectors.", "type": "object" +}, +"tenant": { +"$ref": "GoogleCloudDiscoveryengineV1alphaTenant", +"description": "Optional. The tenant project the connector is connected to." } }, "type": "object" @@ -16546,6 +16552,25 @@ false }, "type": "object" }, +"GoogleCloudDiscoveryengineV1alphaTenant": { +"description": "Tenant information for a connector source. This includes some of the same information stored in the Credential message, but is limited to only what is needed to provide a list of accessible tenants to the user.", +"id": "GoogleCloudDiscoveryengineV1alphaTenant", +"properties": { +"displayName": { +"description": "Optional display name for the tenant, e.g. \"My Slack Team\".", +"type": "string" +}, +"id": { +"description": "The tenant's instance ID. Examples: Jira (\"8594f221-9797-5f78-1fa4-485e198d7cd0\"), Slack (\"T123456\").", +"type": "string" +}, +"uri": { +"description": "The URI of the tenant, if applicable. For example, the URI of a Jira instance is https://my-jira-instance.atlassian.net, and a Slack tenant does not have a URI.", +"type": "string" +} +}, +"type": "object" +}, "GoogleCloudDiscoveryengineV1alphaTrainCustomModelMetadata": { "description": "Metadata related to the progress of the TrainCustomModel operation. This is returned by the google.longrunning.Operation.metadata field.", "id": "GoogleCloudDiscoveryengineV1alphaTrainCustomModelMetadata", diff --git a/googleapiclient/discovery_cache/documents/dlp.v2.json b/googleapiclient/discovery_cache/documents/dlp.v2.json index 4a520d5fce..8525d85a8e 100644 --- a/googleapiclient/discovery_cache/documents/dlp.v2.json +++ b/googleapiclient/discovery_cache/documents/dlp.v2.json @@ -5118,7 +5118,7 @@ } } }, -"revision": "20250615", +"revision": "20250629", "rootUrl": "https://dlp.googleapis.com/", "schemas": { "GooglePrivacyDlpV2Action": { @@ -8283,6 +8283,52 @@ }, "type": "object" }, +"GooglePrivacyDlpV2Domain": { +"description": "A domain represents a thematic category that a data profile can fall under.", +"id": "GooglePrivacyDlpV2Domain", +"properties": { +"category": { +"description": "A domain category that this profile is related to.", +"enum": [ +"CATEGORY_UNSPECIFIED", +"AI", +"CODE" +], +"enumDescriptions": [ +"Category unspecified.", +"Indicates that the data profile is related to artificial intelligence. When set, all findings stored to Security Command Center will set the corresponding AI domain field of `Finding` objects.", +"Indicates that the data profile is related to code." +], +"type": "string" +}, +"signals": { +"description": "The collection of signals that influenced selection of the category.", +"items": { +"enum": [ +"SIGNAL_UNSPECIFIED", +"MODEL", +"TEXT_EMBEDDING", +"VERTEX_PLUGIN", +"VECTOR_PLUGIN", +"SOURCE_CODE", +"SERVICE" +], +"enumDescriptions": [ +"Unused.", +"One or more machine learning models are present.", +"A table appears to be a text embedding.", +"The [Cloud SQL Vertex AI](https://cloud.google.com/sql/docs/postgres/integrate-cloud-sql-with-vertex-ai) plugin is installed on the database.", +"Support for [Cloud SQL vector embeddings](https://cloud.google.com/sql/docs/mysql/enable-vector-search) is enabled on the database.", +"Source code is present.", +"If the service determines the category type. For example, Vertex AI assets would always have a `Category` of `AI`." +], +"type": "string" +}, +"type": "array" +} +}, +"type": "object" +}, "GooglePrivacyDlpV2EntityId": { "description": "An entity in a dataset is a field or set of fields that correspond to a single person. For example, in medical records the `EntityId` might be a patient identifier, or for financial records it might be an account identifier. This message is used when generalizations or analysis must take into account that multiple rows correspond to the same entity.", "id": "GooglePrivacyDlpV2EntityId", @@ -8623,6 +8669,13 @@ }, "type": "array" }, +"domains": { +"description": "Domains associated with the profile.", +"items": { +"$ref": "GooglePrivacyDlpV2Domain" +}, +"type": "array" +}, "fileClusterSummaries": { "description": "FileClusterSummary per each cluster.", "items": { @@ -11956,6 +12009,13 @@ "description": "The Google Cloud project ID that owns the resource.", "type": "string" }, +"domains": { +"description": "Domains associated with the profile.", +"items": { +"$ref": "GooglePrivacyDlpV2Domain" +}, +"type": "array" +}, "encryptionStatus": { "description": "How the table is encrypted.", "enum": [ @@ -12069,7 +12129,7 @@ "description": "The sensitivity score of this table." }, "state": { -"description": "State of a profile.", +"description": "State of a profile. This will always be set to DONE when the table data profile is written to another service like BigQuery or Pub/Sub.", "enum": [ "STATE_UNSPECIFIED", "RUNNING", diff --git a/googleapiclient/discovery_cache/documents/drive.v3.json b/googleapiclient/discovery_cache/documents/drive.v3.json index 2e4e3b7fd2..cad779be89 100644 --- a/googleapiclient/discovery_cache/documents/drive.v3.json +++ b/googleapiclient/discovery_cache/documents/drive.v3.json @@ -2806,7 +2806,7 @@ } } }, -"revision": "20250511", +"revision": "20250701", "rootUrl": "https://www.googleapis.com/", "schemas": { "About": { @@ -3458,6 +3458,21 @@ }, "type": "object" }, +"DownloadRestrictionsMetadata": { +"description": "Download restrictions applied to the file.", +"id": "DownloadRestrictionsMetadata", +"properties": { +"effectiveDownloadRestrictionWithContext": { +"$ref": "DownloadRestriction", +"description": "The effective download restriction applied to this file. This considers all restriction settings and DLP rules." +}, +"itemDownloadRestriction": { +"$ref": "DownloadRestriction", +"description": "The download restriction of the file applied directly by the owner or organizer. This does not take into account shared drive settings or DLP rules." +} +}, +"type": "object" +}, "Drive": { "description": "Representation of a shared drive. Some resource methods (such as `drives.update`) require a `driveId`. Use the `drives.list` method to retrieve the ID for a shared drive.", "id": "Drive", @@ -3506,6 +3521,10 @@ "description": "Output only. Whether the current user can change the `domainUsersOnly` restriction of this shared drive.", "type": "boolean" }, +"canChangeDownloadRestriction": { +"description": "Output only. Whether the current user can change organizer-applied download restrictions of this shared drive.", +"type": "boolean" +}, "canChangeDriveBackground": { "description": "Output only. Whether the current user can change the background of this shared drive.", "type": "boolean" @@ -3701,6 +3720,10 @@ "description": "Output only. Whether the current user can change the `copyRequiresWriterPermission` restriction of this file.", "type": "boolean" }, +"canChangeItemDownloadRestriction": { +"description": "Output only. Whether the current user can change the owner-applied download restrictions of the file.", +"type": "boolean" +}, "canChangeSecurityUpdateEnabled": { "description": "Output only. Whether the current user can change the securityUpdateEnabled field on link share metadata.", "type": "boolean" @@ -3910,6 +3933,10 @@ "description": "A short description of the file.", "type": "string" }, +"downloadRestrictions": { +"$ref": "DownloadRestrictionsMetadata", +"description": "Download restrictions applied on the file." +}, "driveId": { "description": "Output only. ID of the shared drive the file resides in. Only populated for items in shared drives.", "type": "string" @@ -5129,6 +5156,10 @@ "description": "Whether the current user can change the `domainUsersOnly` restriction of this Team Drive.", "type": "boolean" }, +"canChangeDownloadRestriction": { +"description": "Whether the current user can change organizer-applied download restrictions of this shared drive.", +"type": "boolean" +}, "canChangeSharingFoldersRequiresOrganizerPermissionRestriction": { "description": "Whether the current user can change the `sharingFoldersRequiresOrganizerPermission` restriction of this Team Drive.", "type": "boolean" diff --git a/googleapiclient/discovery_cache/documents/gmail.v1.json b/googleapiclient/discovery_cache/documents/gmail.v1.json index 669c1fb056..4c833a8bd4 100644 --- a/googleapiclient/discovery_cache/documents/gmail.v1.json +++ b/googleapiclient/discovery_cache/documents/gmail.v1.json @@ -1091,7 +1091,7 @@ "supportsMediaUpload": true }, "list": { -"description": "Lists the messages in the user's mailbox.", +"description": "Lists the messages in the user's mailbox. For example usage, see [List Gmail messages](https://developers.google.com/workspace/gmail/api/guides/list-messages).", "flatPath": "gmail/v1/users/{userId}/messages", "httpMethod": "GET", "id": "gmail.users.messages.list", @@ -3077,7 +3077,7 @@ } } }, -"revision": "20250616", +"revision": "20250630", "rootUrl": "https://gmail.googleapis.com/", "schemas": { "AutoForwarding": { diff --git a/googleapiclient/discovery_cache/documents/iam.v1.json b/googleapiclient/discovery_cache/documents/iam.v1.json index 6931acf9c5..8907fbe7e6 100644 --- a/googleapiclient/discovery_cache/documents/iam.v1.json +++ b/googleapiclient/discovery_cache/documents/iam.v1.json @@ -1529,34 +1529,6 @@ }, "workloadIdentityPools": { "methods": { -"addAttestationRule": { -"description": "Add an AttestationRule on a WorkloadIdentityPoolManagedIdentity. The total attestation rules after addition must not exceed 50.", -"flatPath": "v1/projects/{projectsId}/locations/{locationsId}/workloadIdentityPools/{workloadIdentityPoolsId}:addAttestationRule", -"httpMethod": "POST", -"id": "iam.projects.locations.workloadIdentityPools.addAttestationRule", -"parameterOrder": [ -"resource" -], -"parameters": { -"resource": { -"description": "Required. The resource name of the managed identity or namespace resource to add an attestation rule to.", -"location": "path", -"pattern": "^projects/[^/]+/locations/[^/]+/workloadIdentityPools/[^/]+$", -"required": true, -"type": "string" -} -}, -"path": "v1/{+resource}:addAttestationRule", -"request": { -"$ref": "AddAttestationRuleRequest" -}, -"response": { -"$ref": "Operation" -}, -"scopes": [ -"https://www.googleapis.com/auth/cloud-platform" -] -}, "create": { "description": "Creates a new WorkloadIdentityPool. You cannot reuse the name of a deleted pool until 30 days after deletion.", "flatPath": "v1/projects/{projectsId}/locations/{locationsId}/workloadIdentityPools", @@ -1709,47 +1681,6 @@ "https://www.googleapis.com/auth/cloud-platform" ] }, -"listAttestationRules": { -"description": "List all AttestationRule on a WorkloadIdentityPoolManagedIdentity.", -"flatPath": "v1/projects/{projectsId}/locations/{locationsId}/workloadIdentityPools/{workloadIdentityPoolsId}:listAttestationRules", -"httpMethod": "GET", -"id": "iam.projects.locations.workloadIdentityPools.listAttestationRules", -"parameterOrder": [ -"resource" -], -"parameters": { -"filter": { -"description": "Optional. A query filter. Supports the following function: * `container_ids()`: Returns only the AttestationRules under the specific container ids. The function expects a comma-delimited list with only project numbers and must use the format `projects/`. For example: `container_ids(projects/, projects/,...)`.", -"location": "query", -"type": "string" -}, -"pageSize": { -"description": "Optional. The maximum number of AttestationRules to return. If unspecified, at most 50 AttestationRules are returned. The maximum value is 100; values above 100 are truncated to 100.", -"format": "int32", -"location": "query", -"type": "integer" -}, -"pageToken": { -"description": "Optional. A page token, received from a previous `ListWorkloadIdentityPoolProviderKeys` call. Provide this to retrieve the subsequent page.", -"location": "query", -"type": "string" -}, -"resource": { -"description": "Required. The resource name of the managed identity or namespace resource to list attestation rules of.", -"location": "path", -"pattern": "^projects/[^/]+/locations/[^/]+/workloadIdentityPools/[^/]+$", -"required": true, -"type": "string" -} -}, -"path": "v1/{+resource}:listAttestationRules", -"response": { -"$ref": "ListAttestationRulesResponse" -}, -"scopes": [ -"https://www.googleapis.com/auth/cloud-platform" -] -}, "patch": { "description": "Updates an existing WorkloadIdentityPool.", "flatPath": "v1/projects/{projectsId}/locations/{locationsId}/workloadIdentityPools/{workloadIdentityPoolsId}", @@ -1784,62 +1715,6 @@ "https://www.googleapis.com/auth/cloud-platform" ] }, -"removeAttestationRule": { -"description": "Remove an AttestationRule on a WorkloadIdentityPoolManagedIdentity.", -"flatPath": "v1/projects/{projectsId}/locations/{locationsId}/workloadIdentityPools/{workloadIdentityPoolsId}:removeAttestationRule", -"httpMethod": "POST", -"id": "iam.projects.locations.workloadIdentityPools.removeAttestationRule", -"parameterOrder": [ -"resource" -], -"parameters": { -"resource": { -"description": "Required. The resource name of the managed identity or namespace resource to remove an attestation rule from.", -"location": "path", -"pattern": "^projects/[^/]+/locations/[^/]+/workloadIdentityPools/[^/]+$", -"required": true, -"type": "string" -} -}, -"path": "v1/{+resource}:removeAttestationRule", -"request": { -"$ref": "RemoveAttestationRuleRequest" -}, -"response": { -"$ref": "Operation" -}, -"scopes": [ -"https://www.googleapis.com/auth/cloud-platform" -] -}, -"setAttestationRules": { -"description": "Set all AttestationRule on a WorkloadIdentityPoolManagedIdentity. A maximum of 50 AttestationRules can be set.", -"flatPath": "v1/projects/{projectsId}/locations/{locationsId}/workloadIdentityPools/{workloadIdentityPoolsId}:setAttestationRules", -"httpMethod": "POST", -"id": "iam.projects.locations.workloadIdentityPools.setAttestationRules", -"parameterOrder": [ -"resource" -], -"parameters": { -"resource": { -"description": "Required. The resource name of the managed identity or namespace resource to add an attestation rule to.", -"location": "path", -"pattern": "^projects/[^/]+/locations/[^/]+/workloadIdentityPools/[^/]+$", -"required": true, -"type": "string" -} -}, -"path": "v1/{+resource}:setAttestationRules", -"request": { -"$ref": "SetAttestationRulesRequest" -}, -"response": { -"$ref": "Operation" -}, -"scopes": [ -"https://www.googleapis.com/auth/cloud-platform" -] -}, "setIamPolicy": { "description": "Sets the IAM policies on a WorkloadIdentityPool", "flatPath": "v1/projects/{projectsId}/locations/{locationsId}/workloadIdentityPools/{workloadIdentityPoolsId}:setIamPolicy", @@ -3898,7 +3773,7 @@ } } }, -"revision": "20250619", +"revision": "20250627", "rootUrl": "https://iam.googleapis.com/", "schemas": { "AccessRestrictions": { diff --git a/googleapiclient/discovery_cache/documents/merchantapi.accounts_v1beta.json b/googleapiclient/discovery_cache/documents/merchantapi.accounts_v1beta.json index 6d0aa3dd3d..131f71979a 100644 --- a/googleapiclient/discovery_cache/documents/merchantapi.accounts_v1beta.json +++ b/googleapiclient/discovery_cache/documents/merchantapi.accounts_v1beta.json @@ -1332,6 +1332,124 @@ "https://www.googleapis.com/auth/content" ] } +}, +"resources": { +"checkoutSettings": { +"methods": { +"create": { +"description": "Creates `CheckoutSettings` for the given merchant.", +"flatPath": "accounts/v1beta/accounts/{accountsId}/programs/{programsId}/checkoutSettings", +"httpMethod": "POST", +"id": "merchantapi.accounts.programs.checkoutSettings.create", +"parameterOrder": [ +"parent" +], +"parameters": { +"parent": { +"description": "Required. The merchant account for which the `CheckoutSettings` will be created.", +"location": "path", +"pattern": "^accounts/[^/]+/programs/[^/]+$", +"required": true, +"type": "string" +} +}, +"path": "accounts/v1beta/{+parent}/checkoutSettings", +"request": { +"$ref": "CheckoutSettings" +}, +"response": { +"$ref": "CheckoutSettings" +}, +"scopes": [ +"https://www.googleapis.com/auth/content" +] +}, +"deleteCheckoutSettings": { +"description": "Deletes `CheckoutSettings` and unenrolls merchant from `Checkout` program.", +"flatPath": "accounts/v1beta/accounts/{accountsId}/programs/{programsId}/checkoutSettings", +"httpMethod": "DELETE", +"id": "merchantapi.accounts.programs.checkoutSettings.deleteCheckoutSettings", +"parameterOrder": [ +"name" +], +"parameters": { +"name": { +"description": "Required. The name/identifier of the merchant account. Format: `accounts/{account}/programs/{program}/checkoutSettings`", +"location": "path", +"pattern": "^accounts/[^/]+/programs/[^/]+/checkoutSettings$", +"required": true, +"type": "string" +} +}, +"path": "accounts/v1beta/{+name}", +"response": { +"$ref": "Empty" +}, +"scopes": [ +"https://www.googleapis.com/auth/content" +] +}, +"getCheckoutSettings": { +"description": "Gets `CheckoutSettings` for the given merchant. This includes information about review state, enrollment state and URL settings.", +"flatPath": "accounts/v1beta/accounts/{accountsId}/programs/{programsId}/checkoutSettings", +"httpMethod": "GET", +"id": "merchantapi.accounts.programs.checkoutSettings.getCheckoutSettings", +"parameterOrder": [ +"name" +], +"parameters": { +"name": { +"description": "Required. The name/identifier of the merchant account. Format: `accounts/{account}/programs/{program}/checkoutSettings`", +"location": "path", +"pattern": "^accounts/[^/]+/programs/[^/]+/checkoutSettings$", +"required": true, +"type": "string" +} +}, +"path": "accounts/v1beta/{+name}", +"response": { +"$ref": "CheckoutSettings" +}, +"scopes": [ +"https://www.googleapis.com/auth/content" +] +}, +"updateCheckoutSettings": { +"description": "Updates `CheckoutSettings` for the given merchant.", +"flatPath": "accounts/v1beta/accounts/{accountsId}/programs/{programsId}/checkoutSettings", +"httpMethod": "PATCH", +"id": "merchantapi.accounts.programs.checkoutSettings.updateCheckoutSettings", +"parameterOrder": [ +"name" +], +"parameters": { +"name": { +"description": "Identifier. The resource name of the program configuration settings. Format: `accounts/{account}/programs/{program}/checkoutSettings`", +"location": "path", +"pattern": "^accounts/[^/]+/programs/[^/]+/checkoutSettings$", +"required": true, +"type": "string" +}, +"updateMask": { +"description": "Required. List of fields being updated. The following fields are supported (in both `snake_case` and `lowerCamelCase`): - `eligible_destinations` - `uri_settings`", +"format": "google-fieldmask", +"location": "query", +"type": "string" +} +}, +"path": "accounts/v1beta/{+name}", +"request": { +"$ref": "CheckoutSettings" +}, +"response": { +"$ref": "CheckoutSettings" +}, +"scopes": [ +"https://www.googleapis.com/auth/content" +] +} +} +} } }, "regions": { @@ -2108,7 +2226,7 @@ } } }, -"revision": "20250621", +"revision": "20250628", "rootUrl": "https://merchantapi.googleapis.com/", "schemas": { "About": { @@ -2679,6 +2797,131 @@ }, "type": "object" }, +"CheckoutSettings": { +"description": "[CheckoutSettings](https://support.google.com/merchants/answer/13945960) for a specific merchant.", +"id": "CheckoutSettings", +"properties": { +"effectiveEnrollmentState": { +"description": "Output only. The effective value of enrollment_state for a given merchant ID. If account level settings are present then this value will be a copy of the account level settings. Otherwise, it will have the value of the parent account (for only marketplace sellers).", +"enum": [ +"CHECKOUT_ENROLLMENT_STATE_UNSPECIFIED", +"INACTIVE", +"ENROLLED", +"OPTED_OUT" +], +"enumDescriptions": [ +"Default enrollment state when enrollment state is not specified.", +"Merchant has not enrolled into the program.", +"Merchant has enrolled into the program by providing either an account level URL or checkout URLs as part of their feed.", +"Merchant has previously enrolled but opted out of the program." +], +"readOnly": true, +"type": "string" +}, +"effectiveReviewState": { +"description": "Output only. The effective value of `review_state` for a given merchant ID. If account level settings are present then this value will be a copy of the account level settings. Otherwise, it will have the value of the parent account (for only marketplace sellers).", +"enum": [ +"CHECKOUT_REVIEW_STATE_UNSPECIFIED", +"IN_REVIEW", +"APPROVED", +"DISAPPROVED" +], +"enumDescriptions": [ +"Default review state when review state is not specified.", +"Merchant provided URLs are being reviewed for data quality issues.", +"Merchant account has been approved. Indicates the data quality checks have passed.", +"Merchant account has been disapproved due to data quality issues." +], +"readOnly": true, +"type": "string" +}, +"effectiveUriSettings": { +"$ref": "UriSettings", +"description": "Output only. The effective value of `uri_settings` for a given merchant. If account level settings are present then this value will be a copy of url settings. Otherwise, it will have the value of the parent account (for only marketplace sellers).", +"readOnly": true +}, +"eligibleDestinations": { +"description": "Optional. The destinations to which the checkout program applies, valid destination values are `SHOPPING_ADS`, `FREE_LISTINGS`", +"items": { +"enum": [ +"DESTINATION_ENUM_UNSPECIFIED", +"SHOPPING_ADS", +"DISPLAY_ADS", +"LOCAL_INVENTORY_ADS", +"FREE_LISTINGS", +"FREE_LOCAL_LISTINGS", +"YOUTUBE_SHOPPING", +"YOUTUBE_SHOPPING_CHECKOUT", +"YOUTUBE_AFFILIATE", +"FREE_VEHICLE_LISTINGS", +"VEHICLE_ADS", +"CLOUD_RETAIL", +"LOCAL_CLOUD_RETAIL" +], +"enumDescriptions": [ +"Not specified.", +"[Shopping ads](https://support.google.com/google-ads/answer/2454022).", +"[Display ads](https://support.google.com/merchants/answer/6069387).", +"[Local inventory ads](https://support.google.com/merchants/answer/3057972).", +"[Free listings](https://support.google.com/merchants/answer/9199328).", +"[Free local product listings](https://support.google.com/merchants/answer/9825611).", +"[YouTube Shopping](https://support.google.com/merchants/answer/12362804).", +"Youtube shopping checkout.", +"[Youtube Affiliate](https://support.google.com/youtube/answer/13376398).", +"[Free vehicle listings](https://support.google.com/merchants/answer/11189169).", +"[Vehicle ads](https://support.google.com/merchants/answer/11189169).", +"[Cloud retail](https://cloud.google.com/solutions/retail).", +"[Local cloud retail](https://cloud.google.com/solutions/retail)." +], +"type": "string" +}, +"type": "array" +}, +"enrollmentState": { +"description": "Output only. Reflects the merchant enrollment state in `Checkout` program.", +"enum": [ +"CHECKOUT_ENROLLMENT_STATE_UNSPECIFIED", +"INACTIVE", +"ENROLLED", +"OPTED_OUT" +], +"enumDescriptions": [ +"Default enrollment state when enrollment state is not specified.", +"Merchant has not enrolled into the program.", +"Merchant has enrolled into the program by providing either an account level URL or checkout URLs as part of their feed.", +"Merchant has previously enrolled but opted out of the program." +], +"readOnly": true, +"type": "string" +}, +"name": { +"description": "Identifier. The resource name of the program configuration settings. Format: `accounts/{account}/programs/{program}/checkoutSettings`", +"type": "string" +}, +"reviewState": { +"description": "Output only. Reflects the merchant review state in `Checkout` program. This is set based on the data quality reviews of the URL provided by the merchant. A merchant with enrollment state as `ENROLLED` can be in the following review states: `IN_REVIEW`, `APPROVED` or `DISAPPROVED`. A merchant must be in an `enrollment_state` of `ENROLLED` before a review can begin for the merchant.For more details, check the help center doc.", +"enum": [ +"CHECKOUT_REVIEW_STATE_UNSPECIFIED", +"IN_REVIEW", +"APPROVED", +"DISAPPROVED" +], +"enumDescriptions": [ +"Default review state when review state is not specified.", +"Merchant provided URLs are being reviewed for data quality issues.", +"Merchant account has been approved. Indicates the data quality checks have passed.", +"Merchant account has been disapproved due to data quality issues." +], +"readOnly": true, +"type": "string" +}, +"uriSettings": { +"$ref": "UriSettings", +"description": "URI settings for cart or checkout URL." +} +}, +"type": "object" +}, "ClaimHomepageRequest": { "description": "Request message for the `ClaimHomepage` method.", "id": "ClaimHomepageRequest", @@ -4910,6 +5153,21 @@ false "properties": {}, "type": "object" }, +"UriSettings": { +"description": "URL settings for cart or checkout URL.", +"id": "UriSettings", +"properties": { +"cartUriTemplate": { +"description": "Cart URL template. When the placeholders are expanded will redirect the buyer to the cart page on the merchant website with the selected item in cart. For more details, check the [help center doc](https://support.google.com/merchants/answer/13945960#method1&zippy=%2Cproduct-level-url-formatting%2Caccount-level-url-formatting)", +"type": "string" +}, +"checkoutUriTemplate": { +"description": "Checkout URL template. When the placeholders are expanded will redirect the buyer to the merchant checkout page with the item in the cart. For more details, check the [help center doc](https://support.google.com/merchants/answer/13945960#method1&zippy=%2Cproduct-level-url-formatting%2Caccount-level-url-formatting)", +"type": "string" +} +}, +"type": "object" +}, "User": { "description": "The `User` message represents a user associated with a Merchant Center account. It is used to manage user permissions and access rights within the account. For more information, see [Frequently asked questions about people and access levels](//support.google.com/merchants/answer/12160472).", "id": "User", diff --git a/googleapiclient/discovery_cache/documents/merchantapi.promotions_v1beta.json b/googleapiclient/discovery_cache/documents/merchantapi.promotions_v1beta.json index 0ee6c72a15..b5b8946e1e 100644 --- a/googleapiclient/discovery_cache/documents/merchantapi.promotions_v1beta.json +++ b/googleapiclient/discovery_cache/documents/merchantapi.promotions_v1beta.json @@ -203,7 +203,7 @@ } } }, -"revision": "20250621", +"revision": "20250628", "rootUrl": "https://merchantapi.googleapis.com/", "schemas": { "Attributes": { @@ -325,7 +325,6 @@ "description": "Optional. [Minimum purchase amount](https://support.google.com/merchants/answer/13837705?ref_topic=13773355&sjid=17642868584668136159-NC) for the promotion." }, "minimumPurchaseQuantity": { -"description": "Optional. [Minimum purchase quantity](https://support.google.com/merchants/answer/13838182?ref_topic=13773355&sjid=17642868584668136159-NC) for the promotion.", "format": "int64", "type": "string" }, diff --git a/googleapiclient/discovery_cache/documents/networkmanagement.v1.json b/googleapiclient/discovery_cache/documents/networkmanagement.v1.json index 454e5a2947..61749f584c 100644 --- a/googleapiclient/discovery_cache/documents/networkmanagement.v1.json +++ b/googleapiclient/discovery_cache/documents/networkmanagement.v1.json @@ -800,7 +800,7 @@ "vpcFlowLogsConfigs": { "methods": { "create": { -"description": "Creates a new `VpcFlowLogsConfig`. If a configuration with the exact same settings already exists (even if the ID is different), the creation fails. Notes: 1. Creating a configuration with state=DISABLED will fail 2. The following fields are not considered as `settings` for the purpose of the check mentioned above, therefore - creating another configuration with the same fields but different values for the following fields will fail as well: * name * create_time * update_time * labels * description", +"description": "Creates a new `VpcFlowLogsConfig`. If a configuration with the exact same settings already exists (even if the ID is different), the creation fails. Notes: 1. Creating a configuration with `state=DISABLED` will fail 2. The following fields are not considered as settings for the purpose of the check mentioned above, therefore - creating another configuration with the same fields but different values for the following fields will fail as well: * name * create_time * update_time * labels * description", "flatPath": "v1/projects/{projectsId}/locations/{locationsId}/vpcFlowLogsConfigs", "httpMethod": "POST", "id": "networkmanagement.projects.locations.vpcFlowLogsConfigs.create", @@ -809,7 +809,7 @@ ], "parameters": { "parent": { -"description": "Required. The parent resource of the VPC Flow Logs configuration to create: `projects/{project_id}/locations/global` `organizations/{organization_id}/locations/global`", +"description": "Required. The parent resource of the VpcFlowLogsConfig to create, in one of the following formats: - For project-level resources: `projects/{project_id}/locations/global` - For organization-level resources: `organizations/{organization_id}/locations/global`", "location": "path", "pattern": "^projects/[^/]+/locations/[^/]+$", "required": true, @@ -842,7 +842,7 @@ ], "parameters": { "name": { -"description": "Required. `VpcFlowLogsConfig` resource name using one of the form: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config} `organizations/{organization_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config}`", +"description": "Required. The resource name of the VpcFlowLogsConfig, in one of the following formats: - For a project-level resource: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` - For an organization-level resource: `organizations/{organization_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}`", "location": "path", "pattern": "^projects/[^/]+/locations/[^/]+/vpcFlowLogsConfigs/[^/]+$", "required": true, @@ -867,7 +867,7 @@ ], "parameters": { "name": { -"description": "Required. `VpcFlowLogsConfig` resource name using the form: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config}` `organizations/{organization_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config}`", +"description": "Required. The resource name of the VpcFlowLogsConfig, in one of the following formats: - For project-level resources: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` - For organization-level resources: `organizations/{organization_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}`", "location": "path", "pattern": "^projects/[^/]+/locations/[^/]+/vpcFlowLogsConfigs/[^/]+$", "required": true, @@ -913,7 +913,7 @@ "type": "string" }, "parent": { -"description": "Required. The parent resource of the VpcFlowLogsConfig: `projects/{project_id}/locations/global` `organizations/{organization_id}/locations/global`", +"description": "Required. The parent resource of the VpcFlowLogsConfig, in one of the following formats: - For project-level resourcs: `projects/{project_id}/locations/global` - For organization-level resources: `organizations/{organization_id}/locations/global`", "location": "path", "pattern": "^projects/[^/]+/locations/[^/]+$", "required": true, @@ -929,7 +929,7 @@ ] }, "patch": { -"description": "Updates an existing `VpcFlowLogsConfig`. If a configuration with the exact same settings already exists (even if the ID is different), the creation fails. Notes: 1. Updating a configuration with state=DISABLED will fail. 2. The following fields are not considered as `settings` for the purpose of the check mentioned above, therefore - updating another configuration with the same fields but different values for the following fields will fail as well: * name * create_time * update_time * labels * description", +"description": "Updates an existing `VpcFlowLogsConfig`. If a configuration with the exact same settings already exists (even if the ID is different), the creation fails. Notes: 1. Updating a configuration with `state=DISABLED` will fail. 2. The following fields are not considered as settings for the purpose of the check mentioned above, therefore - updating another configuration with the same fields but different values for the following fields will fail as well: * name * create_time * update_time * labels * description", "flatPath": "v1/projects/{projectsId}/locations/{locationsId}/vpcFlowLogsConfigs/{vpcFlowLogsConfigsId}", "httpMethod": "PATCH", "id": "networkmanagement.projects.locations.vpcFlowLogsConfigs.patch", @@ -938,14 +938,14 @@ ], "parameters": { "name": { -"description": "Identifier. Unique name of the configuration using one of the forms: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` `organizations/{organization_number}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}`", +"description": "Identifier. Unique name of the configuration. The name can have one of the following forms: - For project-level configurations: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` - For organization-level configurations: `organizations/{organization_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}`", "location": "path", "pattern": "^projects/[^/]+/locations/[^/]+/vpcFlowLogsConfigs/[^/]+$", "required": true, "type": "string" }, "updateMask": { -"description": "Required. Mask of fields to update. At least one path must be supplied in this field.", +"description": "Required. Mask of fields to update. At least one path must be supplied in this field. For example, to change the state of the configuration to ENABLED, specify `update_mask` = `\"state\"`, and the `vpc_flow_logs_config` would be: `vpc_flow_logs_config = { name = \"projects/my-project/locations/global/vpcFlowLogsConfigs/my-config\" state = \"ENABLED\" }`", "format": "google-fieldmask", "location": "query", "type": "string" @@ -969,7 +969,7 @@ } } }, -"revision": "20250611", +"revision": "20250625", "rootUrl": "https://networkmanagement.googleapis.com/", "schemas": { "AbortInfo": { @@ -3760,7 +3760,7 @@ false "type": "array" }, "name": { -"description": "Identifier. Unique name of the configuration using one of the forms: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` `organizations/{organization_number}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}`", +"description": "Identifier. Unique name of the configuration. The name can have one of the following forms: - For project-level configurations: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` - For organization-level configurations: `organizations/{organization_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}`", "type": "string" }, "state": { @@ -3778,7 +3778,7 @@ false "type": "string" }, "targetResourceState": { -"description": "Output only. A diagnostic bit - describes the state of the configured target resource for diagnostic purposes.", +"description": "Output only. Describes the state of the configured target resource for diagnostic purposes.", "enum": [ "TARGET_RESOURCE_STATE_UNSPECIFIED", "TARGET_RESOURCE_EXISTS", diff --git a/googleapiclient/discovery_cache/documents/networkmanagement.v1beta1.json b/googleapiclient/discovery_cache/documents/networkmanagement.v1beta1.json index a89e0d8f30..e83f396d4f 100644 --- a/googleapiclient/discovery_cache/documents/networkmanagement.v1beta1.json +++ b/googleapiclient/discovery_cache/documents/networkmanagement.v1beta1.json @@ -313,7 +313,7 @@ "vpcFlowLogsConfigs": { "methods": { "create": { -"description": "Creates a new `VpcFlowLogsConfig`. If a configuration with the exact same settings already exists (even if the ID is different), the creation fails. Notes: 1. Creating a configuration with state=DISABLED will fail 2. The following fields are not considered as `settings` for the purpose of the check mentioned above, therefore - creating another configuration with the same fields but different values for the following fields will fail as well: * name * create_time * update_time * labels * description", +"description": "Creates a new `VpcFlowLogsConfig`. If a configuration with the exact same settings already exists (even if the ID is different), the creation fails. Notes: 1. Creating a configuration with `state=DISABLED` will fail 2. The following fields are not considered as settings for the purpose of the check mentioned above, therefore - creating another configuration with the same fields but different values for the following fields will fail as well: * name * create_time * update_time * labels * description", "flatPath": "v1beta1/organizations/{organizationsId}/locations/{locationsId}/vpcFlowLogsConfigs", "httpMethod": "POST", "id": "networkmanagement.organizations.locations.vpcFlowLogsConfigs.create", @@ -322,7 +322,7 @@ ], "parameters": { "parent": { -"description": "Required. The parent resource of the VPC Flow Logs configuration to create: `projects/{project_id}/locations/global` `organizations/{organization_id}/locations/global`", +"description": "Required. The parent resource of the VpcFlowLogsConfig to create, in one of the following formats: - For project-level resources: `projects/{project_id}/locations/global` - For organization-level resources: `organizations/{organization_id}/locations/global`", "location": "path", "pattern": "^organizations/[^/]+/locations/[^/]+$", "required": true, @@ -355,7 +355,7 @@ ], "parameters": { "name": { -"description": "Required. `VpcFlowLogsConfig` resource name using one of the form: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config}` `organizations/{organization_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config}`", +"description": "Required. The resource name of the VpcFlowLogsConfig, in one of the following formats: - For a project-level resource: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` - For an organization-level resource: `organizations/{organization_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}`", "location": "path", "pattern": "^organizations/[^/]+/locations/[^/]+/vpcFlowLogsConfigs/[^/]+$", "required": true, @@ -380,7 +380,7 @@ ], "parameters": { "name": { -"description": "Required. `VpcFlowLogsConfig` resource name using the form: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config}` `organizations/{organization_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config}`", +"description": "Required. The resource name of the VpcFlowLogsConfig, in one of the following formats: - For project-level resources: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` - For organization-level resources: `organizations/{organization_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}`", "location": "path", "pattern": "^organizations/[^/]+/locations/[^/]+/vpcFlowLogsConfigs/[^/]+$", "required": true, @@ -426,7 +426,7 @@ "type": "string" }, "parent": { -"description": "Required. The parent resource of the VpcFlowLogsConfig: `projects/{project_id}/locations/global` `organizations/{organization_id}/locations/global`", +"description": "Required. The parent resource of the VpcFlowLogsConfig, in one of the following formats: - For project-level resourcs: `projects/{project_id}/locations/global` - For organization-level resources: `organizations/{organization_id}/locations/global`", "location": "path", "pattern": "^organizations/[^/]+/locations/[^/]+$", "required": true, @@ -442,7 +442,7 @@ ] }, "patch": { -"description": "Updates an existing `VpcFlowLogsConfig`. If a configuration with the exact same settings already exists (even if the ID is different), the creation fails. Notes: 1. Updating a configuration with state=DISABLED will fail 2. The following fields are not considered as `settings` for the purpose of the check mentioned above, therefore - updating another configuration with the same fields but different values for the following fields will fail as well: * name * create_time * update_time * labels * description", +"description": "Updates an existing `VpcFlowLogsConfig`. If a configuration with the exact same settings already exists (even if the ID is different), the creation fails. Notes: 1. Updating a configuration with `state=DISABLED` will fail 2. The following fields are not considered as settings for the purpose of the check mentioned above, therefore - updating another configuration with the same fields but different values for the following fields will fail as well: * name * create_time * update_time * labels * description", "flatPath": "v1beta1/organizations/{organizationsId}/locations/{locationsId}/vpcFlowLogsConfigs/{vpcFlowLogsConfigsId}", "httpMethod": "PATCH", "id": "networkmanagement.organizations.locations.vpcFlowLogsConfigs.patch", @@ -451,14 +451,14 @@ ], "parameters": { "name": { -"description": "Identifier. Unique name of the configuration using one of the forms: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` `organizations/{organization_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}`", +"description": "Identifier. Unique name of the configuration. The name can have one of the following forms: - For project-level configurations: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` - For organization-level configurations: `organizations/{organization_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}`", "location": "path", "pattern": "^organizations/[^/]+/locations/[^/]+/vpcFlowLogsConfigs/[^/]+$", "required": true, "type": "string" }, "updateMask": { -"description": "Required. Mask of fields to update. At least one path must be supplied in this field.", +"description": "Required. Mask of fields to update. At least one path must be supplied in this field. For example, to change the state of the configuration to ENABLED, specify `update_mask` = `\"state\"`, and the `vpc_flow_logs_config` would be: `vpc_flow_logs_config = { name = \"projects/my-project/locations/global/vpcFlowLogsConfigs/my-config\" state = \"ENABLED\" }`", "format": "google-fieldmask", "location": "query", "type": "string" @@ -971,7 +971,7 @@ "vpcFlowLogsConfigs": { "methods": { "create": { -"description": "Creates a new `VpcFlowLogsConfig`. If a configuration with the exact same settings already exists (even if the ID is different), the creation fails. Notes: 1. Creating a configuration with state=DISABLED will fail 2. The following fields are not considered as `settings` for the purpose of the check mentioned above, therefore - creating another configuration with the same fields but different values for the following fields will fail as well: * name * create_time * update_time * labels * description", +"description": "Creates a new `VpcFlowLogsConfig`. If a configuration with the exact same settings already exists (even if the ID is different), the creation fails. Notes: 1. Creating a configuration with `state=DISABLED` will fail 2. The following fields are not considered as settings for the purpose of the check mentioned above, therefore - creating another configuration with the same fields but different values for the following fields will fail as well: * name * create_time * update_time * labels * description", "flatPath": "v1beta1/projects/{projectsId}/locations/{locationsId}/vpcFlowLogsConfigs", "httpMethod": "POST", "id": "networkmanagement.projects.locations.vpcFlowLogsConfigs.create", @@ -980,7 +980,7 @@ ], "parameters": { "parent": { -"description": "Required. The parent resource of the VPC Flow Logs configuration to create: `projects/{project_id}/locations/global` `organizations/{organization_id}/locations/global`", +"description": "Required. The parent resource of the VpcFlowLogsConfig to create, in one of the following formats: - For project-level resources: `projects/{project_id}/locations/global` - For organization-level resources: `organizations/{organization_id}/locations/global`", "location": "path", "pattern": "^projects/[^/]+/locations/[^/]+$", "required": true, @@ -1013,7 +1013,7 @@ ], "parameters": { "name": { -"description": "Required. `VpcFlowLogsConfig` resource name using one of the form: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config}` `organizations/{organization_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config}`", +"description": "Required. The resource name of the VpcFlowLogsConfig, in one of the following formats: - For a project-level resource: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` - For an organization-level resource: `organizations/{organization_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}`", "location": "path", "pattern": "^projects/[^/]+/locations/[^/]+/vpcFlowLogsConfigs/[^/]+$", "required": true, @@ -1038,7 +1038,7 @@ ], "parameters": { "name": { -"description": "Required. `VpcFlowLogsConfig` resource name using the form: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config}` `organizations/{organization_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config}`", +"description": "Required. The resource name of the VpcFlowLogsConfig, in one of the following formats: - For project-level resources: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` - For organization-level resources: `organizations/{organization_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}`", "location": "path", "pattern": "^projects/[^/]+/locations/[^/]+/vpcFlowLogsConfigs/[^/]+$", "required": true, @@ -1084,7 +1084,7 @@ "type": "string" }, "parent": { -"description": "Required. The parent resource of the VpcFlowLogsConfig: `projects/{project_id}/locations/global` `organizations/{organization_id}/locations/global`", +"description": "Required. The parent resource of the VpcFlowLogsConfig, in one of the following formats: - For project-level resourcs: `projects/{project_id}/locations/global` - For organization-level resources: `organizations/{organization_id}/locations/global`", "location": "path", "pattern": "^projects/[^/]+/locations/[^/]+$", "required": true, @@ -1100,7 +1100,7 @@ ] }, "patch": { -"description": "Updates an existing `VpcFlowLogsConfig`. If a configuration with the exact same settings already exists (even if the ID is different), the creation fails. Notes: 1. Updating a configuration with state=DISABLED will fail 2. The following fields are not considered as `settings` for the purpose of the check mentioned above, therefore - updating another configuration with the same fields but different values for the following fields will fail as well: * name * create_time * update_time * labels * description", +"description": "Updates an existing `VpcFlowLogsConfig`. If a configuration with the exact same settings already exists (even if the ID is different), the creation fails. Notes: 1. Updating a configuration with `state=DISABLED` will fail 2. The following fields are not considered as settings for the purpose of the check mentioned above, therefore - updating another configuration with the same fields but different values for the following fields will fail as well: * name * create_time * update_time * labels * description", "flatPath": "v1beta1/projects/{projectsId}/locations/{locationsId}/vpcFlowLogsConfigs/{vpcFlowLogsConfigsId}", "httpMethod": "PATCH", "id": "networkmanagement.projects.locations.vpcFlowLogsConfigs.patch", @@ -1109,14 +1109,14 @@ ], "parameters": { "name": { -"description": "Identifier. Unique name of the configuration using one of the forms: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` `organizations/{organization_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}`", +"description": "Identifier. Unique name of the configuration. The name can have one of the following forms: - For project-level configurations: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` - For organization-level configurations: `organizations/{organization_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}`", "location": "path", "pattern": "^projects/[^/]+/locations/[^/]+/vpcFlowLogsConfigs/[^/]+$", "required": true, "type": "string" }, "updateMask": { -"description": "Required. Mask of fields to update. At least one path must be supplied in this field.", +"description": "Required. Mask of fields to update. At least one path must be supplied in this field. For example, to change the state of the configuration to ENABLED, specify `update_mask` = `\"state\"`, and the `vpc_flow_logs_config` would be: `vpc_flow_logs_config = { name = \"projects/my-project/locations/global/vpcFlowLogsConfigs/my-config\" state = \"ENABLED\" }`", "format": "google-fieldmask", "location": "query", "type": "string" @@ -1134,7 +1134,7 @@ ] }, "queryOrgVpcFlowLogsConfigs": { -"description": "QueryOrgVpcFlowLogsConfigs lists Organization resources for a given project.", +"description": "QueryOrgVpcFlowLogsConfigs returns a list of all organization-level VPC Flow Logs configurations applicable to the specified project.", "flatPath": "v1beta1/projects/{projectsId}/locations/{locationsId}/vpcFlowLogsConfigs:queryOrgVpcFlowLogsConfigs", "httpMethod": "GET", "id": "networkmanagement.projects.locations.vpcFlowLogsConfigs.queryOrgVpcFlowLogsConfigs", @@ -1159,7 +1159,7 @@ "type": "string" }, "parent": { -"description": "Required. The parent resource of the VpcFlowLogsConfig: `projects/{project_id}/locations/global`", +"description": "Required. The parent resource of the VpcFlowLogsConfig, specified in the following format: `projects/{project_id}/locations/global`", "location": "path", "pattern": "^projects/[^/]+/locations/[^/]+$", "required": true, @@ -1181,7 +1181,7 @@ } } }, -"revision": "20250611", +"revision": "20250625", "rootUrl": "https://networkmanagement.googleapis.com/", "schemas": { "AbortInfo": { @@ -4015,7 +4015,7 @@ false "type": "array" }, "name": { -"description": "Identifier. Unique name of the configuration using one of the forms: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` `organizations/{organization_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}`", +"description": "Identifier. Unique name of the configuration. The name can have one of the following forms: - For project-level configurations: `projects/{project_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}` - For organization-level configurations: `organizations/{organization_id}/locations/global/vpcFlowLogsConfigs/{vpc_flow_logs_config_id}`", "type": "string" }, "network": { @@ -4041,7 +4041,7 @@ false "type": "string" }, "targetResourceState": { -"description": "Output only. A diagnostic bit - describes the state of the configured target resource for diagnostic purposes.", +"description": "Output only. Describes the state of the configured target resource for diagnostic purposes.", "enum": [ "TARGET_RESOURCE_STATE_UNSPECIFIED", "TARGET_RESOURCE_EXISTS", diff --git a/googleapiclient/discovery_cache/documents/networkservices.v1beta1.json b/googleapiclient/discovery_cache/documents/networkservices.v1beta1.json index fd1977e38b..b65a34ff1c 100644 --- a/googleapiclient/discovery_cache/documents/networkservices.v1beta1.json +++ b/googleapiclient/discovery_cache/documents/networkservices.v1beta1.json @@ -2913,7 +2913,7 @@ } } }, -"revision": "20250616", +"revision": "20250626", "rootUrl": "https://networkservices.googleapis.com/", "schemas": { "AuthzExtension": { @@ -4262,7 +4262,7 @@ "type": "object" }, "LbEdgeExtension": { -"description": "`LbEdgeExtension` is a resource that lets the extension service influence the Backend Service selection or Cloud CDN cache keys by modifying the request headers.", +"description": "`LbEdgeExtension` is a resource that lets the extension service influence the selection of backend services and Cloud CDN cache keys by modifying request headers.", "id": "LbEdgeExtension", "properties": { "createTime": { diff --git a/googleapiclient/discovery_cache/documents/places.v1.json b/googleapiclient/discovery_cache/documents/places.v1.json index dd3364c891..64917e7f00 100644 --- a/googleapiclient/discovery_cache/documents/places.v1.json +++ b/googleapiclient/discovery_cache/documents/places.v1.json @@ -280,7 +280,7 @@ } } }, -"revision": "20250611", +"revision": "20250629", "rootUrl": "https://places.googleapis.com/", "schemas": { "GoogleGeoTypeViewport": { @@ -1226,12 +1226,12 @@ "description": "The price range associated with a Place." }, "primaryType": { -"description": "The primary type of the given result. This type must one of the Places API supported types. For example, \"restaurant\", \"cafe\", \"airport\", etc. A place can only have a single primary type. For the complete list of possible values, see Table A and Table B at https://developers.google.com/maps/documentation/places/web-service/place-types", +"description": "The primary type of the given result. This type must be one of the Places API supported types. For example, \"restaurant\", \"cafe\", \"airport\", etc. A place can only have a single primary type. For the complete list of possible values, see Table A and Table B at https://developers.google.com/maps/documentation/places/web-service/place-types. The primary type may be missing if the place's primary type is not a supported type. When a primary type is present, it is always one of the types in the `types` field.", "type": "string" }, "primaryTypeDisplayName": { "$ref": "GoogleTypeLocalizedText", -"description": "The display name of the primary type, localized to the request language if applicable. For the complete list of possible values, see Table A and Table B at https://developers.google.com/maps/documentation/places/web-service/place-types" +"description": "The display name of the primary type, localized to the request language if applicable. For the complete list of possible values, see Table A and Table B at https://developers.google.com/maps/documentation/places/web-service/place-types. The primary type may be missing if the place's primary type is not a supported type." }, "pureServiceAreaBusiness": { "description": "Indicates whether the place is a pure service area business. Pure service area business is a business that visits or delivers to customers directly but does not serve customers at their business address. For example, businesses like cleaning services or plumbers. Those businesses may not have a physical address or location on Google Maps.", @@ -1557,7 +1557,7 @@ "type": "boolean" }, "periods": { -"description": "NOTE: The ordering of the `periods` array is independent of the ordering of the `weekday_descriptions` array. Do not assume they will begin on the same day.", +"description": "The periods that this place is open during the week. The periods are in chronological order, in the place-local timezone. An empty (but not absent) value indicates a place that is never open, e.g. because it is closed temporarily for renovations. The starting day of `periods` is NOT fixed and should not be assumed to be Sunday. The API determines the start day based on a variety of factors. For example, for a 24/7 business, the first period may begin on the day of the request. For other businesses, it might be the first day of the week that they are open. NOTE: The ordering of the `periods` array is independent of the ordering of the `weekday_descriptions` array. Do not assume they will begin on the same day.", "items": { "$ref": "GoogleMapsPlacesV1PlaceOpeningHoursPeriod" }, diff --git a/googleapiclient/discovery_cache/documents/policysimulator.v1.json b/googleapiclient/discovery_cache/documents/policysimulator.v1.json index 3a26ed37ca..64e4cbd614 100644 --- a/googleapiclient/discovery_cache/documents/policysimulator.v1.json +++ b/googleapiclient/discovery_cache/documents/policysimulator.v1.json @@ -109,6 +109,39 @@ "resources": { "locations": { "resources": { +"accessPolicySimulations": { +"resources": { +"operations": { +"methods": { +"get": { +"description": "Gets the latest state of a long-running operation. Clients can use this method to poll the operation result at intervals as recommended by the API service.", +"flatPath": "v1/folders/{foldersId}/locations/{locationsId}/accessPolicySimulations/{accessPolicySimulationsId}/operations/{operationsId}", +"httpMethod": "GET", +"id": "policysimulator.folders.locations.accessPolicySimulations.operations.get", +"parameterOrder": [ +"name" +], +"parameters": { +"name": { +"description": "The name of the operation resource.", +"location": "path", +"pattern": "^folders/[^/]+/locations/[^/]+/accessPolicySimulations/[^/]+/operations/.*$", +"required": true, +"type": "string" +} +}, +"path": "v1/{+name}", +"response": { +"$ref": "GoogleLongrunningOperation" +}, +"scopes": [ +"https://www.googleapis.com/auth/cloud-platform" +] +} +} +} +} +}, "orgPolicyViolationsPreviews": { "resources": { "operations": { @@ -389,6 +422,39 @@ "resources": { "locations": { "resources": { +"accessPolicySimulations": { +"resources": { +"operations": { +"methods": { +"get": { +"description": "Gets the latest state of a long-running operation. Clients can use this method to poll the operation result at intervals as recommended by the API service.", +"flatPath": "v1/organizations/{organizationsId}/locations/{locationsId}/accessPolicySimulations/{accessPolicySimulationsId}/operations/{operationsId}", +"httpMethod": "GET", +"id": "policysimulator.organizations.locations.accessPolicySimulations.operations.get", +"parameterOrder": [ +"name" +], +"parameters": { +"name": { +"description": "The name of the operation resource.", +"location": "path", +"pattern": "^organizations/[^/]+/locations/[^/]+/accessPolicySimulations/[^/]+/operations/.*$", +"required": true, +"type": "string" +} +}, +"path": "v1/{+name}", +"response": { +"$ref": "GoogleLongrunningOperation" +}, +"scopes": [ +"https://www.googleapis.com/auth/cloud-platform" +] +} +} +} +} +}, "orgPolicyViolationsPreviews": { "methods": { "create": { @@ -735,6 +801,39 @@ "resources": { "locations": { "resources": { +"accessPolicySimulations": { +"resources": { +"operations": { +"methods": { +"get": { +"description": "Gets the latest state of a long-running operation. Clients can use this method to poll the operation result at intervals as recommended by the API service.", +"flatPath": "v1/projects/{projectsId}/locations/{locationsId}/accessPolicySimulations/{accessPolicySimulationsId}/operations/{operationsId}", +"httpMethod": "GET", +"id": "policysimulator.projects.locations.accessPolicySimulations.operations.get", +"parameterOrder": [ +"name" +], +"parameters": { +"name": { +"description": "The name of the operation resource.", +"location": "path", +"pattern": "^projects/[^/]+/locations/[^/]+/accessPolicySimulations/[^/]+/operations/.*$", +"required": true, +"type": "string" +} +}, +"path": "v1/{+name}", +"response": { +"$ref": "GoogleLongrunningOperation" +}, +"scopes": [ +"https://www.googleapis.com/auth/cloud-platform" +] +} +} +} +} +}, "orgPolicyViolationsPreviews": { "resources": { "operations": { @@ -942,7 +1041,7 @@ } } }, -"revision": "20250614", +"revision": "20250626", "rootUrl": "https://policysimulator.googleapis.com/", "schemas": { "GoogleCloudOrgpolicyV2AlternatePolicySpec": { diff --git a/googleapiclient/discovery_cache/documents/policysimulator.v1alpha.json b/googleapiclient/discovery_cache/documents/policysimulator.v1alpha.json index 4495553595..b2b27c9be3 100644 --- a/googleapiclient/discovery_cache/documents/policysimulator.v1alpha.json +++ b/googleapiclient/discovery_cache/documents/policysimulator.v1alpha.json @@ -109,6 +109,39 @@ "resources": { "locations": { "resources": { +"accessPolicySimulations": { +"resources": { +"operations": { +"methods": { +"get": { +"description": "Gets the latest state of a long-running operation. Clients can use this method to poll the operation result at intervals as recommended by the API service.", +"flatPath": "v1alpha/folders/{foldersId}/locations/{locationsId}/accessPolicySimulations/{accessPolicySimulationsId}/operations/{operationsId}", +"httpMethod": "GET", +"id": "policysimulator.folders.locations.accessPolicySimulations.operations.get", +"parameterOrder": [ +"name" +], +"parameters": { +"name": { +"description": "The name of the operation resource.", +"location": "path", +"pattern": "^folders/[^/]+/locations/[^/]+/accessPolicySimulations/[^/]+/operations/.*$", +"required": true, +"type": "string" +} +}, +"path": "v1alpha/{+name}", +"response": { +"$ref": "GoogleLongrunningOperation" +}, +"scopes": [ +"https://www.googleapis.com/auth/cloud-platform" +] +} +} +} +} +}, "orgPolicyViolationsPreviews": { "resources": { "operations": { @@ -294,6 +327,39 @@ "resources": { "locations": { "resources": { +"accessPolicySimulations": { +"resources": { +"operations": { +"methods": { +"get": { +"description": "Gets the latest state of a long-running operation. Clients can use this method to poll the operation result at intervals as recommended by the API service.", +"flatPath": "v1alpha/organizations/{organizationsId}/locations/{locationsId}/accessPolicySimulations/{accessPolicySimulationsId}/operations/{operationsId}", +"httpMethod": "GET", +"id": "policysimulator.organizations.locations.accessPolicySimulations.operations.get", +"parameterOrder": [ +"name" +], +"parameters": { +"name": { +"description": "The name of the operation resource.", +"location": "path", +"pattern": "^organizations/[^/]+/locations/[^/]+/accessPolicySimulations/[^/]+/operations/.*$", +"required": true, +"type": "string" +} +}, +"path": "v1alpha/{+name}", +"response": { +"$ref": "GoogleLongrunningOperation" +}, +"scopes": [ +"https://www.googleapis.com/auth/cloud-platform" +] +} +} +} +} +}, "orgPolicyViolationsPreviews": { "resources": { "operations": { @@ -409,6 +475,39 @@ "resources": { "locations": { "resources": { +"accessPolicySimulations": { +"resources": { +"operations": { +"methods": { +"get": { +"description": "Gets the latest state of a long-running operation. Clients can use this method to poll the operation result at intervals as recommended by the API service.", +"flatPath": "v1alpha/projects/{projectsId}/locations/{locationsId}/accessPolicySimulations/{accessPolicySimulationsId}/operations/{operationsId}", +"httpMethod": "GET", +"id": "policysimulator.projects.locations.accessPolicySimulations.operations.get", +"parameterOrder": [ +"name" +], +"parameters": { +"name": { +"description": "The name of the operation resource.", +"location": "path", +"pattern": "^projects/[^/]+/locations/[^/]+/accessPolicySimulations/[^/]+/operations/.*$", +"required": true, +"type": "string" +} +}, +"path": "v1alpha/{+name}", +"response": { +"$ref": "GoogleLongrunningOperation" +}, +"scopes": [ +"https://www.googleapis.com/auth/cloud-platform" +] +} +} +} +} +}, "orgPolicyViolationsPreviews": { "resources": { "operations": { @@ -521,7 +620,7 @@ } } }, -"revision": "20250614", +"revision": "20250626", "rootUrl": "https://policysimulator.googleapis.com/", "schemas": { "GoogleCloudOrgpolicyV2AlternatePolicySpec": { diff --git a/googleapiclient/discovery_cache/documents/policysimulator.v1beta.json b/googleapiclient/discovery_cache/documents/policysimulator.v1beta.json index 78d8127e00..78b23312a2 100644 --- a/googleapiclient/discovery_cache/documents/policysimulator.v1beta.json +++ b/googleapiclient/discovery_cache/documents/policysimulator.v1beta.json @@ -109,6 +109,39 @@ "resources": { "locations": { "resources": { +"accessPolicySimulations": { +"resources": { +"operations": { +"methods": { +"get": { +"description": "Gets the latest state of a long-running operation. Clients can use this method to poll the operation result at intervals as recommended by the API service.", +"flatPath": "v1beta/folders/{foldersId}/locations/{locationsId}/accessPolicySimulations/{accessPolicySimulationsId}/operations/{operationsId}", +"httpMethod": "GET", +"id": "policysimulator.folders.locations.accessPolicySimulations.operations.get", +"parameterOrder": [ +"name" +], +"parameters": { +"name": { +"description": "The name of the operation resource.", +"location": "path", +"pattern": "^folders/[^/]+/locations/[^/]+/accessPolicySimulations/[^/]+/operations/.*$", +"required": true, +"type": "string" +} +}, +"path": "v1beta/{+name}", +"response": { +"$ref": "GoogleLongrunningOperation" +}, +"scopes": [ +"https://www.googleapis.com/auth/cloud-platform" +] +} +} +} +} +}, "orgPolicyViolationsPreviews": { "resources": { "operations": { @@ -425,6 +458,39 @@ "resources": { "locations": { "resources": { +"accessPolicySimulations": { +"resources": { +"operations": { +"methods": { +"get": { +"description": "Gets the latest state of a long-running operation. Clients can use this method to poll the operation result at intervals as recommended by the API service.", +"flatPath": "v1beta/organizations/{organizationsId}/locations/{locationsId}/accessPolicySimulations/{accessPolicySimulationsId}/operations/{operationsId}", +"httpMethod": "GET", +"id": "policysimulator.organizations.locations.accessPolicySimulations.operations.get", +"parameterOrder": [ +"name" +], +"parameters": { +"name": { +"description": "The name of the operation resource.", +"location": "path", +"pattern": "^organizations/[^/]+/locations/[^/]+/accessPolicySimulations/[^/]+/operations/.*$", +"required": true, +"type": "string" +} +}, +"path": "v1beta/{+name}", +"response": { +"$ref": "GoogleLongrunningOperation" +}, +"scopes": [ +"https://www.googleapis.com/auth/cloud-platform" +] +} +} +} +} +}, "orgPolicyViolationsPreviews": { "methods": { "create": { @@ -835,6 +901,39 @@ "resources": { "locations": { "resources": { +"accessPolicySimulations": { +"resources": { +"operations": { +"methods": { +"get": { +"description": "Gets the latest state of a long-running operation. Clients can use this method to poll the operation result at intervals as recommended by the API service.", +"flatPath": "v1beta/projects/{projectsId}/locations/{locationsId}/accessPolicySimulations/{accessPolicySimulationsId}/operations/{operationsId}", +"httpMethod": "GET", +"id": "policysimulator.projects.locations.accessPolicySimulations.operations.get", +"parameterOrder": [ +"name" +], +"parameters": { +"name": { +"description": "The name of the operation resource.", +"location": "path", +"pattern": "^projects/[^/]+/locations/[^/]+/accessPolicySimulations/[^/]+/operations/.*$", +"required": true, +"type": "string" +} +}, +"path": "v1beta/{+name}", +"response": { +"$ref": "GoogleLongrunningOperation" +}, +"scopes": [ +"https://www.googleapis.com/auth/cloud-platform" +] +} +} +} +} +}, "orgPolicyViolationsPreviews": { "resources": { "operations": { @@ -1078,7 +1177,7 @@ } } }, -"revision": "20250614", +"revision": "20250626", "rootUrl": "https://policysimulator.googleapis.com/", "schemas": { "GoogleCloudOrgpolicyV2AlternatePolicySpec": { diff --git a/googleapiclient/discovery_cache/documents/redis.v1.json b/googleapiclient/discovery_cache/documents/redis.v1.json index 899ce468f6..6291556e6e 100644 --- a/googleapiclient/discovery_cache/documents/redis.v1.json +++ b/googleapiclient/discovery_cache/documents/redis.v1.json @@ -1073,7 +1073,7 @@ } } }, -"revision": "20250617", +"revision": "20250627", "rootUrl": "https://redis.googleapis.com/", "schemas": { "AOFConfig": { @@ -1457,6 +1457,10 @@ "description": "A cluster instance.", "id": "Cluster", "properties": { +"allowFewerZonesDeployment": { +"description": "Optional. Immutable. Allows customers to specify if they are okay with deploying a multi-zone cluster in less than 3 zones. Once set, if there is a zonal outage during the cluster creation, the cluster will only be deployed in 2 zones, and stay within the 2 zones for its lifecycle.", +"type": "boolean" +}, "asyncClusterEndpointsDeletionEnabled": { "description": "Optional. If true, cluster endpoints that are created and registered by customers can be deleted asynchronously. That is, such a cluster endpoint can be de-registered before the forwarding rules in the cluster endpoint are deleted.", "type": "boolean" diff --git a/googleapiclient/discovery_cache/documents/redis.v1beta1.json b/googleapiclient/discovery_cache/documents/redis.v1beta1.json index 7149c2e1ce..6dab1f5fe0 100644 --- a/googleapiclient/discovery_cache/documents/redis.v1beta1.json +++ b/googleapiclient/discovery_cache/documents/redis.v1beta1.json @@ -1073,7 +1073,7 @@ } } }, -"revision": "20250617", +"revision": "20250627", "rootUrl": "https://redis.googleapis.com/", "schemas": { "AOFConfig": { @@ -1457,6 +1457,10 @@ "description": "A cluster instance.", "id": "Cluster", "properties": { +"allowFewerZonesDeployment": { +"description": "Optional. Immutable. Allows customers to specify if they are okay with deploying a multi-zone cluster in less than 3 zones. Once set, if there is a zonal outage during the cluster creation, the cluster will only be deployed in 2 zones, and stay within the 2 zones for its lifecycle.", +"type": "boolean" +}, "asyncClusterEndpointsDeletionEnabled": { "description": "Optional. If true, cluster endpoints that are created and registered by customers can be deleted asynchronously. That is, such a cluster endpoint can be de-registered before the forwarding rules in the cluster endpoint are deleted.", "type": "boolean" diff --git a/googleapiclient/discovery_cache/documents/retail.v2.json b/googleapiclient/discovery_cache/documents/retail.v2.json index a5d57bf5da..5c22745d29 100644 --- a/googleapiclient/discovery_cache/documents/retail.v2.json +++ b/googleapiclient/discovery_cache/documents/retail.v2.json @@ -2282,7 +2282,7 @@ } } }, -"revision": "20250613", +"revision": "20250626", "rootUrl": "https://retail.googleapis.com/", "schemas": { "GoogleApiHttpBody": { @@ -3753,6 +3753,10 @@ }, "type": "array" }, +"inlineSource": { +"$ref": "GoogleCloudRetailV2IntentClassificationConfigInlineSource", +"description": "Optional. Inline source for intent classifications." +}, "modelPreamble": { "description": "Optional. Customers can use the preamble to specify any requirements for blocklisting intent classification. This preamble will be added to the blocklisting intent classification model prompt.", "type": "string" @@ -3764,6 +3768,10 @@ "description": "An example for intent classification.", "id": "GoogleCloudRetailV2IntentClassificationConfigExample", "properties": { +"classifiedPositive": { +"description": "Required. Whether the example is classified positively.", +"type": "boolean" +}, "intentType": { "description": "Optional. The intent_type must match one of the predefined intent types defined at https://cloud.google.com/retail/docs/reference/rpc/google.cloud.retail.v2alpha#querytype", "type": "string" @@ -3779,6 +3787,49 @@ }, "type": "object" }, +"GoogleCloudRetailV2IntentClassificationConfigInlineForceIntent": { +"description": "An inline force intent classification configuration.", +"id": "GoogleCloudRetailV2IntentClassificationConfigInlineForceIntent", +"properties": { +"intentType": { +"description": "Optional. The intent_type must match one of the predefined intent types defined at https://cloud.google.com/retail/docs/reference/rpc/google.cloud.retail.v2alpha#querytype", +"type": "string" +}, +"operation": { +"description": "Optional. The operation to perform for the query.", +"enum": [ +"OPERATION_UNSPECIFIED", +"EXACT_MATCH", +"CONTAINS" +], +"enumDescriptions": [ +"Unspecified match operation.", +"Exact match.", +"Contains match." +], +"type": "string" +}, +"query": { +"description": "Optional. A example query.", +"type": "string" +} +}, +"type": "object" +}, +"GoogleCloudRetailV2IntentClassificationConfigInlineSource": { +"description": "Inline source for intent classifications.", +"id": "GoogleCloudRetailV2IntentClassificationConfigInlineSource", +"properties": { +"inlineForceIntents": { +"description": "Optional. A list of inline force intent classifications.", +"items": { +"$ref": "GoogleCloudRetailV2IntentClassificationConfigInlineForceIntent" +}, +"type": "array" +} +}, +"type": "object" +}, "GoogleCloudRetailV2Interval": { "description": "A floating point interval.", "id": "GoogleCloudRetailV2Interval", @@ -5378,6 +5429,13 @@ "$ref": "GoogleCloudRetailV2SearchRequestTileNavigationSpec", "description": "Optional. This field specifies tile navigation related parameters." }, +"userAttributes": { +"additionalProperties": { +"$ref": "GoogleCloudRetailV2StringList" +}, +"description": "Optional. The user attributes that could be used for personalization of search results. * Populate at most 100 key-value pairs per query. * Only supports string keys and repeated string values. * Duplcate keys are not allowed within a single query. Example: user_attributes: [ { key: \"pets\" value { values: \"dog\" values: \"cat\" } }, { key: \"state\" value { values: \"CA\" } } ]", +"type": "object" +}, "userInfo": { "$ref": "GoogleCloudRetailV2UserInfo", "description": "User information." @@ -6134,6 +6192,20 @@ "properties": {}, "type": "object" }, +"GoogleCloudRetailV2StringList": { +"description": "A list of string values.", +"id": "GoogleCloudRetailV2StringList", +"properties": { +"values": { +"description": "String values.", +"items": { +"type": "string" +}, +"type": "array" +} +}, +"type": "object" +}, "GoogleCloudRetailV2Tile": { "description": "This field specifies the tile information including an attribute key, attribute value. More fields will be added in the future, eg: product id or product counts, etc.", "id": "GoogleCloudRetailV2Tile", diff --git a/googleapiclient/discovery_cache/documents/retail.v2alpha.json b/googleapiclient/discovery_cache/documents/retail.v2alpha.json index 47bb9fb746..ef9efeac2d 100644 --- a/googleapiclient/discovery_cache/documents/retail.v2alpha.json +++ b/googleapiclient/discovery_cache/documents/retail.v2alpha.json @@ -2859,7 +2859,7 @@ } } }, -"revision": "20250613", +"revision": "20250626", "rootUrl": "https://retail.googleapis.com/", "schemas": { "GoogleApiHttpBody": { @@ -5498,6 +5498,10 @@ }, "type": "array" }, +"inlineSource": { +"$ref": "GoogleCloudRetailV2alphaIntentClassificationConfigInlineSource", +"description": "Optional. Inline source for intent classifications." +}, "modelPreamble": { "description": "Optional. Customers can use the preamble to specify any requirements for blocklisting intent classification. This preamble will be added to the blocklisting intent classification model prompt.", "type": "string" @@ -5509,6 +5513,10 @@ "description": "An example for intent classification.", "id": "GoogleCloudRetailV2alphaIntentClassificationConfigExample", "properties": { +"classifiedPositive": { +"description": "Required. Whether the example is classified positively.", +"type": "boolean" +}, "intentType": { "description": "Optional. The intent_type must match one of the predefined intent types defined at https://cloud.google.com/retail/docs/reference/rpc/google.cloud.retail.v2alpha#querytype", "type": "string" @@ -5524,6 +5532,49 @@ }, "type": "object" }, +"GoogleCloudRetailV2alphaIntentClassificationConfigInlineForceIntent": { +"description": "An inline force intent classification configuration.", +"id": "GoogleCloudRetailV2alphaIntentClassificationConfigInlineForceIntent", +"properties": { +"intentType": { +"description": "Optional. The intent_type must match one of the predefined intent types defined at https://cloud.google.com/retail/docs/reference/rpc/google.cloud.retail.v2alpha#querytype", +"type": "string" +}, +"operation": { +"description": "Optional. The operation to perform for the query.", +"enum": [ +"OPERATION_UNSPECIFIED", +"EXACT_MATCH", +"CONTAINS" +], +"enumDescriptions": [ +"Unspecified match operation.", +"Exact match.", +"Contains match." +], +"type": "string" +}, +"query": { +"description": "Optional. A example query.", +"type": "string" +} +}, +"type": "object" +}, +"GoogleCloudRetailV2alphaIntentClassificationConfigInlineSource": { +"description": "Inline source for intent classifications.", +"id": "GoogleCloudRetailV2alphaIntentClassificationConfigInlineSource", +"properties": { +"inlineForceIntents": { +"description": "Optional. A list of inline force intent classifications.", +"items": { +"$ref": "GoogleCloudRetailV2alphaIntentClassificationConfigInlineForceIntent" +}, +"type": "array" +} +}, +"type": "object" +}, "GoogleCloudRetailV2alphaInterval": { "description": "A floating point interval.", "id": "GoogleCloudRetailV2alphaInterval", diff --git a/googleapiclient/discovery_cache/documents/retail.v2beta.json b/googleapiclient/discovery_cache/documents/retail.v2beta.json index f4cc307b73..cc66e0ca7b 100644 --- a/googleapiclient/discovery_cache/documents/retail.v2beta.json +++ b/googleapiclient/discovery_cache/documents/retail.v2beta.json @@ -2427,7 +2427,7 @@ } } }, -"revision": "20250613", +"revision": "20250626", "rootUrl": "https://retail.googleapis.com/", "schemas": { "GoogleApiHttpBody": { @@ -5492,6 +5492,10 @@ }, "type": "array" }, +"inlineSource": { +"$ref": "GoogleCloudRetailV2betaIntentClassificationConfigInlineSource", +"description": "Optional. Inline source for intent classifications." +}, "modelPreamble": { "description": "Optional. Customers can use the preamble to specify any requirements for blocklisting intent classification. This preamble will be added to the blocklisting intent classification model prompt.", "type": "string" @@ -5503,6 +5507,10 @@ "description": "An example for intent classification.", "id": "GoogleCloudRetailV2betaIntentClassificationConfigExample", "properties": { +"classifiedPositive": { +"description": "Required. Whether the example is classified positively.", +"type": "boolean" +}, "intentType": { "description": "Optional. The intent_type must match one of the predefined intent types defined at https://cloud.google.com/retail/docs/reference/rpc/google.cloud.retail.v2alpha#querytype", "type": "string" @@ -5518,6 +5526,49 @@ }, "type": "object" }, +"GoogleCloudRetailV2betaIntentClassificationConfigInlineForceIntent": { +"description": "An inline force intent classification configuration.", +"id": "GoogleCloudRetailV2betaIntentClassificationConfigInlineForceIntent", +"properties": { +"intentType": { +"description": "Optional. The intent_type must match one of the predefined intent types defined at https://cloud.google.com/retail/docs/reference/rpc/google.cloud.retail.v2alpha#querytype", +"type": "string" +}, +"operation": { +"description": "Optional. The operation to perform for the query.", +"enum": [ +"OPERATION_UNSPECIFIED", +"EXACT_MATCH", +"CONTAINS" +], +"enumDescriptions": [ +"Unspecified match operation.", +"Exact match.", +"Contains match." +], +"type": "string" +}, +"query": { +"description": "Optional. A example query.", +"type": "string" +} +}, +"type": "object" +}, +"GoogleCloudRetailV2betaIntentClassificationConfigInlineSource": { +"description": "Inline source for intent classifications.", +"id": "GoogleCloudRetailV2betaIntentClassificationConfigInlineSource", +"properties": { +"inlineForceIntents": { +"description": "Optional. A list of inline force intent classifications.", +"items": { +"$ref": "GoogleCloudRetailV2betaIntentClassificationConfigInlineForceIntent" +}, +"type": "array" +} +}, +"type": "object" +}, "GoogleCloudRetailV2betaInterval": { "description": "A floating point interval.", "id": "GoogleCloudRetailV2betaInterval", @@ -7191,6 +7242,13 @@ "$ref": "GoogleCloudRetailV2betaSearchRequestTileNavigationSpec", "description": "Optional. This field specifies tile navigation related parameters." }, +"userAttributes": { +"additionalProperties": { +"$ref": "GoogleCloudRetailV2betaStringList" +}, +"description": "Optional. The user attributes that could be used for personalization of search results. * Populate at most 100 key-value pairs per query. * Only supports string keys and repeated string values. * Duplcate keys are not allowed within a single query. Example: user_attributes: [ { key: \"pets\" value { values: \"dog\" values: \"cat\" } }, { key: \"state\" value { values: \"CA\" } } ]", +"type": "object" +}, "userInfo": { "$ref": "GoogleCloudRetailV2betaUserInfo", "description": "User information." @@ -7947,6 +8005,20 @@ "properties": {}, "type": "object" }, +"GoogleCloudRetailV2betaStringList": { +"description": "A list of string values.", +"id": "GoogleCloudRetailV2betaStringList", +"properties": { +"values": { +"description": "String values.", +"items": { +"type": "string" +}, +"type": "array" +} +}, +"type": "object" +}, "GoogleCloudRetailV2betaTile": { "description": "This field specifies the tile information including an attribute key, attribute value. More fields will be added in the future, eg: product id or product counts, etc.", "id": "GoogleCloudRetailV2betaTile", diff --git a/googleapiclient/discovery_cache/documents/script.v1.json b/googleapiclient/discovery_cache/documents/script.v1.json index 074a78dd1b..f97fd8ed4f 100644 --- a/googleapiclient/discovery_cache/documents/script.v1.json +++ b/googleapiclient/discovery_cache/documents/script.v1.json @@ -891,7 +891,7 @@ } } }, -"revision": "20250526", +"revision": "20250623", "rootUrl": "https://script.googleapis.com/", "schemas": { "Content": { @@ -913,7 +913,7 @@ "type": "object" }, "CreateProjectRequest": { -"description": "Request to create a script project. Request to create a script project.", +"description": "Request to create a script project.", "id": "CreateProjectRequest", "properties": { "parentId": { diff --git a/googleapiclient/discovery_cache/documents/secretmanager.v1.json b/googleapiclient/discovery_cache/documents/secretmanager.v1.json index 3126b2f2df..684ebde699 100644 --- a/googleapiclient/discovery_cache/documents/secretmanager.v1.json +++ b/googleapiclient/discovery_cache/documents/secretmanager.v1.json @@ -23,63 +23,68 @@ }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.me-west1.rep.googleapis.com/", -"location": "me-west1" +"endpointUrl": "https://secretmanager.africa-south1.rep.googleapis.com/", +"location": "africa-south1" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.us-central1.rep.googleapis.com/", -"location": "us-central1" +"endpointUrl": "https://secretmanager.asia-northeast1.rep.googleapis.com/", +"location": "asia-northeast1" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.us-east1.rep.googleapis.com/", -"location": "us-east1" +"endpointUrl": "https://secretmanager.asia-south1.rep.googleapis.com/", +"location": "asia-south1" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.us-central2.rep.googleapis.com/", -"location": "us-central2" +"endpointUrl": "https://secretmanager.asia-south2.rep.googleapis.com/", +"location": "asia-south2" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.us-west1.rep.googleapis.com/", -"location": "us-west1" +"endpointUrl": "https://secretmanager.asia-southeast2.rep.googleapis.com/", +"location": "asia-southeast2" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.us-west2.rep.googleapis.com/", -"location": "us-west2" +"endpointUrl": "https://secretmanager.australia-southeast1.rep.googleapis.com/", +"location": "australia-southeast1" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.us-west3.rep.googleapis.com/", -"location": "us-west3" +"endpointUrl": "https://secretmanager.australia-southeast2.rep.googleapis.com/", +"location": "australia-southeast2" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.us-west4.rep.googleapis.com/", -"location": "us-west4" +"endpointUrl": "https://secretmanager.europe-north2.rep.googleapis.com/", +"location": "europe-north2" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.us-east4.rep.googleapis.com/", -"location": "us-east4" +"endpointUrl": "https://secretmanager.europe-west1.rep.googleapis.com/", +"location": "europe-west1" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.us-east5.rep.googleapis.com/", -"location": "us-east5" +"endpointUrl": "https://secretmanager.europe-west2.rep.googleapis.com/", +"location": "europe-west2" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.us-south1.rep.googleapis.com/", -"location": "us-south1" +"endpointUrl": "https://secretmanager.europe-west3.rep.googleapis.com/", +"location": "europe-west3" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.europe-west3.rep.googleapis.com/", -"location": "europe-west3" +"endpointUrl": "https://secretmanager.europe-west4.rep.googleapis.com/", +"location": "europe-west4" +}, +{ +"description": "Regional Endpoint", +"endpointUrl": "https://secretmanager.europe-west6.rep.googleapis.com/", +"location": "europe-west6" }, { "description": "Regional Endpoint", @@ -93,68 +98,78 @@ }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.europe-west6.rep.googleapis.com/", -"location": "europe-west6" +"endpointUrl": "https://secretmanager.me-west1.rep.googleapis.com/", +"location": "me-west1" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.europe-west1.rep.googleapis.com/", -"location": "europe-west1" +"endpointUrl": "https://secretmanager.northamerica-northeast1.rep.googleapis.com/", +"location": "northamerica-northeast1" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.europe-west4.rep.googleapis.com/", -"location": "europe-west4" +"endpointUrl": "https://secretmanager.northamerica-northeast2.rep.googleapis.com/", +"location": "northamerica-northeast2" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.us-east7.rep.googleapis.com/", -"location": "us-east7" +"endpointUrl": "https://secretmanager.southamerica-east1.rep.googleapis.com/", +"location": "southamerica-east1" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.europe-north2.rep.googleapis.com/", -"location": "europe-north2" +"endpointUrl": "https://secretmanager.us-central1.rep.googleapis.com/", +"location": "us-central1" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.asia-northeast1.rep.googleapis.com/", -"location": "asia-northeast1" +"endpointUrl": "https://secretmanager.us-central2.rep.googleapis.com/", +"location": "us-central2" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.australia-southeast1.rep.googleapis.com/", -"location": "australia-southeast1" +"endpointUrl": "https://secretmanager.us-east1.rep.googleapis.com/", +"location": "us-east1" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.australia-southeast2.rep.googleapis.com/", -"location": "australia-southeast2" +"endpointUrl": "https://secretmanager.us-east4.rep.googleapis.com/", +"location": "us-east4" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.europe-west2.rep.googleapis.com/", -"location": "europe-west2" +"endpointUrl": "https://secretmanager.us-east5.rep.googleapis.com/", +"location": "us-east5" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.northamerica-northeast1.rep.googleapis.com/", -"location": "northamerica-northeast1" +"endpointUrl": "https://secretmanager.us-east7.rep.googleapis.com/", +"location": "us-east7" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.northamerica-northeast2.rep.googleapis.com/", -"location": "northamerica-northeast2" +"endpointUrl": "https://secretmanager.us-south1.rep.googleapis.com/", +"location": "us-south1" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.asia-south1.rep.googleapis.com/", -"location": "asia-south1" +"endpointUrl": "https://secretmanager.us-west1.rep.googleapis.com/", +"location": "us-west1" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.asia-south2.rep.googleapis.com/", -"location": "asia-south2" +"endpointUrl": "https://secretmanager.us-west2.rep.googleapis.com/", +"location": "us-west2" +}, +{ +"description": "Regional Endpoint", +"endpointUrl": "https://secretmanager.us-west3.rep.googleapis.com/", +"location": "us-west3" +}, +{ +"description": "Regional Endpoint", +"endpointUrl": "https://secretmanager.us-west4.rep.googleapis.com/", +"location": "us-west4" }, { "description": "Regional Endpoint", @@ -1266,7 +1281,7 @@ } } }, -"revision": "20250624", +"revision": "20250626", "rootUrl": "https://secretmanager.googleapis.com/", "schemas": { "AccessSecretVersionResponse": { diff --git a/googleapiclient/discovery_cache/documents/secretmanager.v1beta1.json b/googleapiclient/discovery_cache/documents/secretmanager.v1beta1.json index 3fa0eeabca..4b01e75b9a 100644 --- a/googleapiclient/discovery_cache/documents/secretmanager.v1beta1.json +++ b/googleapiclient/discovery_cache/documents/secretmanager.v1beta1.json @@ -23,63 +23,68 @@ }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.me-west1.rep.googleapis.com/", -"location": "me-west1" +"endpointUrl": "https://secretmanager.africa-south1.rep.googleapis.com/", +"location": "africa-south1" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.us-central1.rep.googleapis.com/", -"location": "us-central1" +"endpointUrl": "https://secretmanager.asia-northeast1.rep.googleapis.com/", +"location": "asia-northeast1" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.us-east1.rep.googleapis.com/", -"location": "us-east1" +"endpointUrl": "https://secretmanager.asia-south1.rep.googleapis.com/", +"location": "asia-south1" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.us-central2.rep.googleapis.com/", -"location": "us-central2" +"endpointUrl": "https://secretmanager.asia-south2.rep.googleapis.com/", +"location": "asia-south2" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.us-west1.rep.googleapis.com/", -"location": "us-west1" +"endpointUrl": "https://secretmanager.asia-southeast2.rep.googleapis.com/", +"location": "asia-southeast2" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.us-west2.rep.googleapis.com/", -"location": "us-west2" +"endpointUrl": "https://secretmanager.australia-southeast1.rep.googleapis.com/", +"location": "australia-southeast1" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.us-west3.rep.googleapis.com/", -"location": "us-west3" +"endpointUrl": "https://secretmanager.australia-southeast2.rep.googleapis.com/", +"location": "australia-southeast2" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.us-west4.rep.googleapis.com/", -"location": "us-west4" +"endpointUrl": "https://secretmanager.europe-north2.rep.googleapis.com/", +"location": "europe-north2" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.us-east4.rep.googleapis.com/", -"location": "us-east4" +"endpointUrl": "https://secretmanager.europe-west1.rep.googleapis.com/", +"location": "europe-west1" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.us-east5.rep.googleapis.com/", -"location": "us-east5" +"endpointUrl": "https://secretmanager.europe-west2.rep.googleapis.com/", +"location": "europe-west2" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.us-south1.rep.googleapis.com/", -"location": "us-south1" +"endpointUrl": "https://secretmanager.europe-west3.rep.googleapis.com/", +"location": "europe-west3" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.europe-west3.rep.googleapis.com/", -"location": "europe-west3" +"endpointUrl": "https://secretmanager.europe-west4.rep.googleapis.com/", +"location": "europe-west4" +}, +{ +"description": "Regional Endpoint", +"endpointUrl": "https://secretmanager.europe-west6.rep.googleapis.com/", +"location": "europe-west6" }, { "description": "Regional Endpoint", @@ -93,68 +98,78 @@ }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.europe-west6.rep.googleapis.com/", -"location": "europe-west6" +"endpointUrl": "https://secretmanager.me-west1.rep.googleapis.com/", +"location": "me-west1" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.europe-west1.rep.googleapis.com/", -"location": "europe-west1" +"endpointUrl": "https://secretmanager.northamerica-northeast1.rep.googleapis.com/", +"location": "northamerica-northeast1" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.europe-west4.rep.googleapis.com/", -"location": "europe-west4" +"endpointUrl": "https://secretmanager.northamerica-northeast2.rep.googleapis.com/", +"location": "northamerica-northeast2" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.us-east7.rep.googleapis.com/", -"location": "us-east7" +"endpointUrl": "https://secretmanager.southamerica-east1.rep.googleapis.com/", +"location": "southamerica-east1" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.europe-north2.rep.googleapis.com/", -"location": "europe-north2" +"endpointUrl": "https://secretmanager.us-central1.rep.googleapis.com/", +"location": "us-central1" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.asia-northeast1.rep.googleapis.com/", -"location": "asia-northeast1" +"endpointUrl": "https://secretmanager.us-central2.rep.googleapis.com/", +"location": "us-central2" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.australia-southeast1.rep.googleapis.com/", -"location": "australia-southeast1" +"endpointUrl": "https://secretmanager.us-east1.rep.googleapis.com/", +"location": "us-east1" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.australia-southeast2.rep.googleapis.com/", -"location": "australia-southeast2" +"endpointUrl": "https://secretmanager.us-east4.rep.googleapis.com/", +"location": "us-east4" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.europe-west2.rep.googleapis.com/", -"location": "europe-west2" +"endpointUrl": "https://secretmanager.us-east5.rep.googleapis.com/", +"location": "us-east5" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.northamerica-northeast1.rep.googleapis.com/", -"location": "northamerica-northeast1" +"endpointUrl": "https://secretmanager.us-east7.rep.googleapis.com/", +"location": "us-east7" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.northamerica-northeast2.rep.googleapis.com/", -"location": "northamerica-northeast2" +"endpointUrl": "https://secretmanager.us-south1.rep.googleapis.com/", +"location": "us-south1" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.asia-south1.rep.googleapis.com/", -"location": "asia-south1" +"endpointUrl": "https://secretmanager.us-west1.rep.googleapis.com/", +"location": "us-west1" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.asia-south2.rep.googleapis.com/", -"location": "asia-south2" +"endpointUrl": "https://secretmanager.us-west2.rep.googleapis.com/", +"location": "us-west2" +}, +{ +"description": "Regional Endpoint", +"endpointUrl": "https://secretmanager.us-west3.rep.googleapis.com/", +"location": "us-west3" +}, +{ +"description": "Regional Endpoint", +"endpointUrl": "https://secretmanager.us-west4.rep.googleapis.com/", +"location": "us-west4" }, { "description": "Regional Endpoint", @@ -786,7 +801,7 @@ } } }, -"revision": "20250624", +"revision": "20250626", "rootUrl": "https://secretmanager.googleapis.com/", "schemas": { "AccessSecretVersionResponse": { diff --git a/googleapiclient/discovery_cache/documents/secretmanager.v1beta2.json b/googleapiclient/discovery_cache/documents/secretmanager.v1beta2.json index c36bd73baf..b1873efd77 100644 --- a/googleapiclient/discovery_cache/documents/secretmanager.v1beta2.json +++ b/googleapiclient/discovery_cache/documents/secretmanager.v1beta2.json @@ -23,63 +23,68 @@ }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.me-west1.rep.googleapis.com/", -"location": "me-west1" +"endpointUrl": "https://secretmanager.africa-south1.rep.googleapis.com/", +"location": "africa-south1" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.us-central1.rep.googleapis.com/", -"location": "us-central1" +"endpointUrl": "https://secretmanager.asia-northeast1.rep.googleapis.com/", +"location": "asia-northeast1" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.us-east1.rep.googleapis.com/", -"location": "us-east1" +"endpointUrl": "https://secretmanager.asia-south1.rep.googleapis.com/", +"location": "asia-south1" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.us-central2.rep.googleapis.com/", -"location": "us-central2" +"endpointUrl": "https://secretmanager.asia-south2.rep.googleapis.com/", +"location": "asia-south2" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.us-west1.rep.googleapis.com/", -"location": "us-west1" +"endpointUrl": "https://secretmanager.asia-southeast2.rep.googleapis.com/", +"location": "asia-southeast2" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.us-west2.rep.googleapis.com/", -"location": "us-west2" +"endpointUrl": "https://secretmanager.australia-southeast1.rep.googleapis.com/", +"location": "australia-southeast1" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.us-west3.rep.googleapis.com/", -"location": "us-west3" +"endpointUrl": "https://secretmanager.australia-southeast2.rep.googleapis.com/", +"location": "australia-southeast2" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.us-west4.rep.googleapis.com/", -"location": "us-west4" +"endpointUrl": "https://secretmanager.europe-north2.rep.googleapis.com/", +"location": "europe-north2" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.us-east4.rep.googleapis.com/", -"location": "us-east4" +"endpointUrl": "https://secretmanager.europe-west1.rep.googleapis.com/", +"location": "europe-west1" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.us-east5.rep.googleapis.com/", -"location": "us-east5" +"endpointUrl": "https://secretmanager.europe-west2.rep.googleapis.com/", +"location": "europe-west2" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.us-south1.rep.googleapis.com/", -"location": "us-south1" +"endpointUrl": "https://secretmanager.europe-west3.rep.googleapis.com/", +"location": "europe-west3" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.europe-west3.rep.googleapis.com/", -"location": "europe-west3" +"endpointUrl": "https://secretmanager.europe-west4.rep.googleapis.com/", +"location": "europe-west4" +}, +{ +"description": "Regional Endpoint", +"endpointUrl": "https://secretmanager.europe-west6.rep.googleapis.com/", +"location": "europe-west6" }, { "description": "Regional Endpoint", @@ -93,68 +98,78 @@ }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.europe-west6.rep.googleapis.com/", -"location": "europe-west6" +"endpointUrl": "https://secretmanager.me-west1.rep.googleapis.com/", +"location": "me-west1" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.europe-west1.rep.googleapis.com/", -"location": "europe-west1" +"endpointUrl": "https://secretmanager.northamerica-northeast1.rep.googleapis.com/", +"location": "northamerica-northeast1" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.europe-west4.rep.googleapis.com/", -"location": "europe-west4" +"endpointUrl": "https://secretmanager.northamerica-northeast2.rep.googleapis.com/", +"location": "northamerica-northeast2" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.us-east7.rep.googleapis.com/", -"location": "us-east7" +"endpointUrl": "https://secretmanager.southamerica-east1.rep.googleapis.com/", +"location": "southamerica-east1" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.europe-north2.rep.googleapis.com/", -"location": "europe-north2" +"endpointUrl": "https://secretmanager.us-central1.rep.googleapis.com/", +"location": "us-central1" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.asia-northeast1.rep.googleapis.com/", -"location": "asia-northeast1" +"endpointUrl": "https://secretmanager.us-central2.rep.googleapis.com/", +"location": "us-central2" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.australia-southeast1.rep.googleapis.com/", -"location": "australia-southeast1" +"endpointUrl": "https://secretmanager.us-east1.rep.googleapis.com/", +"location": "us-east1" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.australia-southeast2.rep.googleapis.com/", -"location": "australia-southeast2" +"endpointUrl": "https://secretmanager.us-east4.rep.googleapis.com/", +"location": "us-east4" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.europe-west2.rep.googleapis.com/", -"location": "europe-west2" +"endpointUrl": "https://secretmanager.us-east5.rep.googleapis.com/", +"location": "us-east5" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.northamerica-northeast1.rep.googleapis.com/", -"location": "northamerica-northeast1" +"endpointUrl": "https://secretmanager.us-east7.rep.googleapis.com/", +"location": "us-east7" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.northamerica-northeast2.rep.googleapis.com/", -"location": "northamerica-northeast2" +"endpointUrl": "https://secretmanager.us-south1.rep.googleapis.com/", +"location": "us-south1" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.asia-south1.rep.googleapis.com/", -"location": "asia-south1" +"endpointUrl": "https://secretmanager.us-west1.rep.googleapis.com/", +"location": "us-west1" }, { "description": "Regional Endpoint", -"endpointUrl": "https://secretmanager.asia-south2.rep.googleapis.com/", -"location": "asia-south2" +"endpointUrl": "https://secretmanager.us-west2.rep.googleapis.com/", +"location": "us-west2" +}, +{ +"description": "Regional Endpoint", +"endpointUrl": "https://secretmanager.us-west3.rep.googleapis.com/", +"location": "us-west3" +}, +{ +"description": "Regional Endpoint", +"endpointUrl": "https://secretmanager.us-west4.rep.googleapis.com/", +"location": "us-west4" }, { "description": "Regional Endpoint", @@ -1266,7 +1281,7 @@ } } }, -"revision": "20250624", +"revision": "20250626", "rootUrl": "https://secretmanager.googleapis.com/", "schemas": { "AccessSecretVersionResponse": { diff --git a/googleapiclient/discovery_cache/documents/securitycenter.v1.json b/googleapiclient/discovery_cache/documents/securitycenter.v1.json index fee6989bd8..38982cae16 100644 --- a/googleapiclient/discovery_cache/documents/securitycenter.v1.json +++ b/googleapiclient/discovery_cache/documents/securitycenter.v1.json @@ -5938,7 +5938,7 @@ } } }, -"revision": "20250616", +"revision": "20250625", "rootUrl": "https://securitycenter.googleapis.com/", "schemas": { "Access": { @@ -12820,6 +12820,11 @@ "script": { "$ref": "GoogleCloudSecuritycenterV2File", "description": "When the process represents the invocation of a script, `binary` provides information about the interpreter, while `script` provides information about the script file provided to the interpreter." +}, +"userId": { +"description": "The ID of the user that executed the process. E.g. If this is the root user this will always be 0.", +"format": "int64", +"type": "string" } }, "type": "object" @@ -15442,6 +15447,11 @@ "script": { "$ref": "File", "description": "When the process represents the invocation of a script, `binary` provides information about the interpreter, while `script` provides information about the script file provided to the interpreter." +}, +"userId": { +"description": "The ID of the user that executed the process. E.g. If this is the root user this will always be 0.", +"format": "int64", +"type": "string" } }, "type": "object" diff --git a/googleapiclient/discovery_cache/documents/securitycenter.v1beta1.json b/googleapiclient/discovery_cache/documents/securitycenter.v1beta1.json index 39e6d3f581..badfba4a29 100644 --- a/googleapiclient/discovery_cache/documents/securitycenter.v1beta1.json +++ b/googleapiclient/discovery_cache/documents/securitycenter.v1beta1.json @@ -913,7 +913,7 @@ } } }, -"revision": "20250616", +"revision": "20250625", "rootUrl": "https://securitycenter.googleapis.com/", "schemas": { "Access": { @@ -7490,6 +7490,11 @@ "script": { "$ref": "GoogleCloudSecuritycenterV2File", "description": "When the process represents the invocation of a script, `binary` provides information about the interpreter, while `script` provides information about the script file provided to the interpreter." +}, +"userId": { +"description": "The ID of the user that executed the process. E.g. If this is the root user this will always be 0.", +"format": "int64", +"type": "string" } }, "type": "object" @@ -9767,6 +9772,11 @@ "script": { "$ref": "File", "description": "When the process represents the invocation of a script, `binary` provides information about the interpreter, while `script` provides information about the script file provided to the interpreter." +}, +"userId": { +"description": "The ID of the user that executed the process. E.g. If this is the root user this will always be 0.", +"format": "int64", +"type": "string" } }, "type": "object" diff --git a/googleapiclient/discovery_cache/documents/securitycenter.v1beta2.json b/googleapiclient/discovery_cache/documents/securitycenter.v1beta2.json index 4e7a30b89d..0f2a57b35c 100644 --- a/googleapiclient/discovery_cache/documents/securitycenter.v1beta2.json +++ b/googleapiclient/discovery_cache/documents/securitycenter.v1beta2.json @@ -2003,7 +2003,7 @@ } } }, -"revision": "20250616", +"revision": "20250625", "rootUrl": "https://securitycenter.googleapis.com/", "schemas": { "Access": { @@ -8483,6 +8483,11 @@ "script": { "$ref": "GoogleCloudSecuritycenterV2File", "description": "When the process represents the invocation of a script, `binary` provides information about the interpreter, while `script` provides information about the script file provided to the interpreter." +}, +"userId": { +"description": "The ID of the user that executed the process. E.g. If this is the root user this will always be 0.", +"format": "int64", +"type": "string" } }, "type": "object" @@ -10426,6 +10431,11 @@ "script": { "$ref": "File", "description": "When the process represents the invocation of a script, `binary` provides information about the interpreter, while `script` provides information about the script file provided to the interpreter." +}, +"userId": { +"description": "The ID of the user that executed the process. E.g. If this is the root user this will always be 0.", +"format": "int64", +"type": "string" } }, "type": "object" diff --git a/googleapiclient/discovery_cache/documents/spanner.v1.json b/googleapiclient/discovery_cache/documents/spanner.v1.json index 5a8a4dd24f..a588be2431 100644 --- a/googleapiclient/discovery_cache/documents/spanner.v1.json +++ b/googleapiclient/discovery_cache/documents/spanner.v1.json @@ -3406,7 +3406,7 @@ } } }, -"revision": "20250523", +"revision": "20250625", "rootUrl": "https://spanner.googleapis.com/", "schemas": { "AdaptMessageRequest": { @@ -3966,6 +3966,33 @@ }, "type": "object" }, +"ChangeStreamRecord": { +"description": "Spanner Change Streams enable customers to capture and stream out changes to their Spanner databases in real-time. A change stream can be created with option partition_mode='IMMUTABLE_KEY_RANGE' or partition_mode='MUTABLE_KEY_RANGE'. This message is only used in Change Streams created with the option partition_mode='MUTABLE_KEY_RANGE'. Spanner automatically creates a special Table-Valued Function (TVF) along with each Change Streams. The function provides access to the change stream's records. The function is named READ_ (where is the name of the change stream), and it returns a table with only one column called ChangeRecord.", +"id": "ChangeStreamRecord", +"properties": { +"dataChangeRecord": { +"$ref": "DataChangeRecord", +"description": "Data change record describing a data change for a change stream partition." +}, +"heartbeatRecord": { +"$ref": "HeartbeatRecord", +"description": "Heartbeat record describing a heartbeat for a change stream partition." +}, +"partitionEndRecord": { +"$ref": "PartitionEndRecord", +"description": "Partition end record describing a terminated change stream partition." +}, +"partitionEventRecord": { +"$ref": "PartitionEventRecord", +"description": "Partition event record describing key range changes for a change stream partition." +}, +"partitionStartRecord": { +"$ref": "PartitionStartRecord", +"description": "Partition start record describing a new change stream partition." +} +}, +"type": "object" +}, "ChildLink": { "description": "Metadata associated with a parent-child relationship appearing in a PlanNode.", "id": "ChildLink", @@ -3986,6 +4013,30 @@ }, "type": "object" }, +"ColumnMetadata": { +"description": "Metadata for a column.", +"id": "ColumnMetadata", +"properties": { +"isPrimaryKey": { +"description": "Indicates whether the column is a primary key column.", +"type": "boolean" +}, +"name": { +"description": "Name of the column.", +"type": "string" +}, +"ordinalPosition": { +"description": "Ordinal position of the column based on the original table definition in the schema starting with a value of 1.", +"format": "int64", +"type": "string" +}, +"type": { +"$ref": "Type", +"description": "Type of the column." +} +}, +"type": "object" +}, "CommitRequest": { "description": "The request for Commit.", "id": "CommitRequest", @@ -4456,6 +4507,100 @@ }, "type": "object" }, +"DataChangeRecord": { +"description": "A data change record contains a set of changes to a table with the same modification type (insert, update, or delete) committed at the same commit timestamp in one change stream partition for the same transaction. Multiple data change records can be returned for the same transaction across multiple change stream partitions.", +"id": "DataChangeRecord", +"properties": { +"columnMetadata": { +"description": "Provides metadata describing the columns associated with the mods listed below.", +"items": { +"$ref": "ColumnMetadata" +}, +"type": "array" +}, +"commitTimestamp": { +"description": "Indicates the timestamp in which the change was committed. DataChangeRecord.commit_timestamps, PartitionStartRecord.start_timestamps, PartitionEventRecord.commit_timestamps, and PartitionEndRecord.end_timestamps can have the same value in the same partition.", +"format": "google-datetime", +"type": "string" +}, +"isLastRecordInTransactionInPartition": { +"description": "Indicates whether this is the last record for a transaction in the current partition. Clients can use this field to determine when all records for a transaction in the current partition have been received.", +"type": "boolean" +}, +"isSystemTransaction": { +"description": "Indicates whether the transaction is a system transaction. System transactions include those issued by time-to-live (TTL), column backfill, etc.", +"type": "boolean" +}, +"modType": { +"description": "Describes the type of change.", +"enum": [ +"MOD_TYPE_UNSPECIFIED", +"INSERT", +"UPDATE", +"DELETE" +], +"enumDescriptions": [ +"Not specified.", +"Indicates data was inserted.", +"Indicates existing data was updated.", +"Indicates existing data was deleted." +], +"type": "string" +}, +"mods": { +"description": "Describes the changes that were made.", +"items": { +"$ref": "Mod" +}, +"type": "array" +}, +"numberOfPartitionsInTransaction": { +"description": "Indicates the number of partitions that return data change records for this transaction. This value can be helpful in assembling all records associated with a particular transaction.", +"format": "int32", +"type": "integer" +}, +"numberOfRecordsInTransaction": { +"description": "Indicates the number of data change records that are part of this transaction across all change stream partitions. This value can be used to assemble all the records associated with a particular transaction.", +"format": "int32", +"type": "integer" +}, +"recordSequence": { +"description": "Record sequence numbers are unique and monotonically increasing (but not necessarily contiguous) for a specific timestamp across record types in the same partition. To guarantee ordered processing, the reader should process records (of potentially different types) in record_sequence order for a specific timestamp in the same partition. The record sequence number ordering across partitions is only meaningful in the context of a specific transaction. Record sequence numbers are unique across partitions for a specific transaction. Sort the DataChangeRecords for the same server_transaction_id by record_sequence to reconstruct the ordering of the changes within the transaction.", +"type": "string" +}, +"serverTransactionId": { +"description": "Provides a globally unique string that represents the transaction in which the change was committed. Multiple transactions can have the same commit timestamp, but each transaction has a unique server_transaction_id.", +"type": "string" +}, +"table": { +"description": "Name of the table affected by the change.", +"type": "string" +}, +"transactionTag": { +"description": "Indicates the transaction tag associated with this transaction.", +"type": "string" +}, +"valueCaptureType": { +"description": "Describes the value capture type that was specified in the change stream configuration when this change was captured.", +"enum": [ +"VALUE_CAPTURE_TYPE_UNSPECIFIED", +"OLD_AND_NEW_VALUES", +"NEW_VALUES", +"NEW_ROW", +"NEW_ROW_AND_OLD_VALUES" +], +"enumDescriptions": [ +"Not specified.", +"Records both old and new values of the modified watched columns.", +"Records only new values of the modified watched columns.", +"Records new values of all watched columns, including modified and unmodified columns.", +"Records the new values of all watched columns, including modified and unmodified columns. Also records the old values of the modified columns." +], +"type": "string" +} +}, +"type": "object" +}, "Database": { "description": "A Cloud Spanner database.", "id": "Database", @@ -5018,6 +5163,18 @@ }, "type": "object" }, +"HeartbeatRecord": { +"description": "A heartbeat record is returned as a progress indicator, when there are no data changes or any other partition record types in the change stream partition.", +"id": "HeartbeatRecord", +"properties": { +"timestamp": { +"description": "Indicates the timestamp at which the query has returned all the records in the change stream partition with timestamp <= heartbeat timestamp. The heartbeat timestamp will not be the same as the timestamps of other record types in the same partition.", +"format": "google-datetime", +"type": "string" +} +}, +"type": "object" +}, "IncludeReplicas": { "description": "An `IncludeReplicas` contains a repeated set of `ReplicaSelection` which indicates the order in which replicas should be considered.", "id": "IncludeReplicas", @@ -6020,6 +6177,61 @@ }, "type": "object" }, +"Mod": { +"description": "A mod describes all data changes in a watched table row.", +"id": "Mod", +"properties": { +"keys": { +"description": "Returns the value of the primary key of the modified row.", +"items": { +"$ref": "ModValue" +}, +"type": "array" +}, +"newValues": { +"description": "Returns the new values after the change for the modified columns. Always empty for DELETE.", +"items": { +"$ref": "ModValue" +}, +"type": "array" +}, +"oldValues": { +"description": "Returns the old values before the change for the modified columns. Always empty for INSERT, or if old values are not being captured specified by value_capture_type.", +"items": { +"$ref": "ModValue" +}, +"type": "array" +} +}, +"type": "object" +}, +"ModValue": { +"description": "Returns the value and associated metadata for a particular field of the Mod.", +"id": "ModValue", +"properties": { +"columnMetadataIndex": { +"description": "Index within the repeated column_metadata field, to obtain the column metadata for the column that was modified.", +"format": "int32", +"type": "integer" +}, +"value": { +"description": "The value of the column.", +"type": "any" +} +}, +"type": "object" +}, +"MoveInEvent": { +"description": "Describes move-in of the key ranges into the change stream partition identified by partition_token. To maintain processing the changes for a particular key in timestamp order, the query processing the change stream partition identified by partition_token should not advance beyond the partition event record commit timestamp until the queries processing the source change stream partitions have processed all change stream records with timestamps <= the partition event record commit timestamp.", +"id": "MoveInEvent", +"properties": { +"sourcePartitionToken": { +"description": "An unique partition identifier describing the source change stream partition that recorded changes for the key range that is moving into this partition.", +"type": "string" +} +}, +"type": "object" +}, "MoveInstanceRequest": { "description": "The request for MoveInstance.", "id": "MoveInstanceRequest", @@ -6038,6 +6250,17 @@ }, "type": "object" }, +"MoveOutEvent": { +"description": "Describes move-out of the key ranges out of the change stream partition identified by partition_token. To maintain processing the changes for a particular key in timestamp order, the query processing the MoveOutEvent in the partition identified by partition_token should inform the queries processing the destination partitions that they can unblock and proceed processing records past the commit_timestamp.", +"id": "MoveOutEvent", +"properties": { +"destinationPartitionToken": { +"description": "An unique partition identifier describing the destination change stream partition that will record changes for the key range that is moving out of this partition.", +"type": "string" +} +}, +"type": "object" +}, "MultiplexedSessionPrecommitToken": { "description": "When a read-write transaction is executed on a multiplexed session, this precommit token is sent back to the client as a part of the Transaction message in the BeginTransaction response and also as a part of the ResultSet and PartialResultSet responses.", "id": "MultiplexedSessionPrecommitToken", @@ -6219,6 +6442,60 @@ }, "type": "object" }, +"PartitionEndRecord": { +"description": "A partition end record serves as a notification that the client should stop reading the partition. No further records are expected to be retrieved on it.", +"id": "PartitionEndRecord", +"properties": { +"endTimestamp": { +"description": "End timestamp at which the change stream partition is terminated. All changes generated by this partition will have timestamps <= end_timestamp. DataChangeRecord.commit_timestamps, PartitionStartRecord.start_timestamps, PartitionEventRecord.commit_timestamps, and PartitionEndRecord.end_timestamps can have the same value in the same partition. PartitionEndRecord is the last record returned for a partition.", +"format": "google-datetime", +"type": "string" +}, +"partitionToken": { +"description": "Unique partition identifier describing the terminated change stream partition. partition_token is equal to the partition token of the change stream partition currently queried to return this PartitionEndRecord.", +"type": "string" +}, +"recordSequence": { +"description": "Record sequence numbers are unique and monotonically increasing (but not necessarily contiguous) for a specific timestamp across record types in the same partition. To guarantee ordered processing, the reader should process records (of potentially different types) in record_sequence order for a specific timestamp in the same partition.", +"type": "string" +} +}, +"type": "object" +}, +"PartitionEventRecord": { +"description": "A partition event record describes key range changes for a change stream partition. The changes to a row defined by its primary key can be captured in one change stream partition for a specific time range, and then be captured in a different change stream partition for a different time range. This movement of key ranges across change stream partitions is a reflection of activities, such as Spanner's dynamic splitting and load balancing, etc. Processing this event is needed if users want to guarantee processing of the changes for any key in timestamp order. If time ordered processing of changes for a primary key is not needed, this event can be ignored. To guarantee time ordered processing for each primary key, if the event describes move-ins, the reader of this partition needs to wait until the readers of the source partitions have processed all records with timestamps <= this PartitionEventRecord.commit_timestamp, before advancing beyond this PartitionEventRecord. If the event describes move-outs, the reader can notify the readers of the destination partitions that they can continue processing.", +"id": "PartitionEventRecord", +"properties": { +"commitTimestamp": { +"description": "Indicates the commit timestamp at which the key range change occurred. DataChangeRecord.commit_timestamps, PartitionStartRecord.start_timestamps, PartitionEventRecord.commit_timestamps, and PartitionEndRecord.end_timestamps can have the same value in the same partition.", +"format": "google-datetime", +"type": "string" +}, +"moveInEvents": { +"description": "Set when one or more key ranges are moved into the change stream partition identified by partition_token. Example: Two key ranges are moved into partition (P1) from partition (P2) and partition (P3) in a single transaction at timestamp T. The PartitionEventRecord returned in P1 will reflect the move as: PartitionEventRecord { commit_timestamp: T partition_token: \"P1\" move_in_events { source_partition_token: \"P2\" } move_in_events { source_partition_token: \"P3\" } } The PartitionEventRecord returned in P2 will reflect the move as: PartitionEventRecord { commit_timestamp: T partition_token: \"P2\" move_out_events { destination_partition_token: \"P1\" } } The PartitionEventRecord returned in P3 will reflect the move as: PartitionEventRecord { commit_timestamp: T partition_token: \"P3\" move_out_events { destination_partition_token: \"P1\" } }", +"items": { +"$ref": "MoveInEvent" +}, +"type": "array" +}, +"moveOutEvents": { +"description": "Set when one or more key ranges are moved out of the change stream partition identified by partition_token. Example: Two key ranges are moved out of partition (P1) to partition (P2) and partition (P3) in a single transaction at timestamp T. The PartitionEventRecord returned in P1 will reflect the move as: PartitionEventRecord { commit_timestamp: T partition_token: \"P1\" move_out_events { destination_partition_token: \"P2\" } move_out_events { destination_partition_token: \"P3\" } } The PartitionEventRecord returned in P2 will reflect the move as: PartitionEventRecord { commit_timestamp: T partition_token: \"P2\" move_in_events { source_partition_token: \"P1\" } } The PartitionEventRecord returned in P3 will reflect the move as: PartitionEventRecord { commit_timestamp: T partition_token: \"P3\" move_in_events { source_partition_token: \"P1\" } }", +"items": { +"$ref": "MoveOutEvent" +}, +"type": "array" +}, +"partitionToken": { +"description": "Unique partition identifier describing the partition this event occurred on. partition_token is equal to the partition token of the change stream partition currently queried to return this PartitionEventRecord.", +"type": "string" +}, +"recordSequence": { +"description": "Record sequence numbers are unique and monotonically increasing (but not necessarily contiguous) for a specific timestamp across record types in the same partition. To guarantee ordered processing, the reader should process records (of potentially different types) in record_sequence order for a specific timestamp in the same partition.", +"type": "string" +} +}, +"type": "object" +}, "PartitionOptions": { "description": "Options for a `PartitionQueryRequest` and `PartitionReadRequest`.", "id": "PartitionOptions", @@ -6322,6 +6599,29 @@ }, "type": "object" }, +"PartitionStartRecord": { +"description": "A partition start record serves as a notification that the client should schedule the partitions to be queried. PartitionStartRecord returns information about one or more partitions.", +"id": "PartitionStartRecord", +"properties": { +"partitionTokens": { +"description": "Unique partition identifiers to be used in queries.", +"items": { +"type": "string" +}, +"type": "array" +}, +"recordSequence": { +"description": "Record sequence numbers are unique and monotonically increasing (but not necessarily contiguous) for a specific timestamp across record types in the same partition. To guarantee ordered processing, the reader should process records (of potentially different types) in record_sequence order for a specific timestamp in the same partition.", +"type": "string" +}, +"startTimestamp": { +"description": "Start timestamp at which the partitions should be queried to return change stream records with timestamps >= start_timestamp. DataChangeRecord.commit_timestamps, PartitionStartRecord.start_timestamps, PartitionEventRecord.commit_timestamps, and PartitionEndRecord.end_timestamps can have the same value in the same partition.", +"format": "google-datetime", +"type": "string" +} +}, +"type": "object" +}, "PartitionedDml": { "description": "Message type to initiate a Partitioned DML transaction.", "id": "PartitionedDml", @@ -7276,7 +7576,7 @@ "type": "object" }, "TransactionOptions": { -"description": "Transactions: Each session can have at most one active transaction at a time (note that standalone reads and queries use a transaction internally and do count towards the one transaction limit). After the active transaction is completed, the session can immediately be re-used for the next transaction. It is not necessary to create a new session for each transaction. Transaction modes: Cloud Spanner supports three transaction modes: 1. Locking read-write. This type of transaction is the only way to write data into Cloud Spanner. These transactions rely on pessimistic locking and, if necessary, two-phase commit. Locking read-write transactions may abort, requiring the application to retry. 2. Snapshot read-only. Snapshot read-only transactions provide guaranteed consistency across several reads, but do not allow writes. Snapshot read-only transactions can be configured to read at timestamps in the past, or configured to perform a strong read (where Spanner selects a timestamp such that the read is guaranteed to see the effects of all transactions that have committed before the start of the read). Snapshot read-only transactions do not need to be committed. Queries on change streams must be performed with the snapshot read-only transaction mode, specifying a strong read. See TransactionOptions.ReadOnly.strong for more details. 3. Partitioned DML. This type of transaction is used to execute a single Partitioned DML statement. Partitioned DML partitions the key space and runs the DML statement over each partition in parallel using separate, internal transactions that commit independently. Partitioned DML transactions do not need to be committed. For transactions that only read, snapshot read-only transactions provide simpler semantics and are almost always faster. In particular, read-only transactions do not take locks, so they do not conflict with read-write transactions. As a consequence of not taking locks, they also do not abort, so retry loops are not needed. Transactions may only read-write data in a single database. They may, however, read-write data in different tables within that database. Locking read-write transactions: Locking transactions may be used to atomically read-modify-write data anywhere in a database. This type of transaction is externally consistent. Clients should attempt to minimize the amount of time a transaction is active. Faster transactions commit with higher probability and cause less contention. Cloud Spanner attempts to keep read locks active as long as the transaction continues to do reads, and the transaction has not been terminated by Commit or Rollback. Long periods of inactivity at the client may cause Cloud Spanner to release a transaction's locks and abort it. Conceptually, a read-write transaction consists of zero or more reads or SQL statements followed by Commit. At any time before Commit, the client can send a Rollback request to abort the transaction. Semantics: Cloud Spanner can commit the transaction if all read locks it acquired are still valid at commit time, and it is able to acquire write locks for all writes. Cloud Spanner can abort the transaction for any reason. If a commit attempt returns `ABORTED`, Cloud Spanner guarantees that the transaction has not modified any user data in Cloud Spanner. Unless the transaction commits, Cloud Spanner makes no guarantees about how long the transaction's locks were held for. It is an error to use Cloud Spanner locks for any sort of mutual exclusion other than between Cloud Spanner transactions themselves. Retrying aborted transactions: When a transaction aborts, the application can choose to retry the whole transaction again. To maximize the chances of successfully committing the retry, the client should execute the retry in the same session as the original attempt. The original session's lock priority increases with each consecutive abort, meaning that each attempt has a slightly better chance of success than the previous. Note that the lock priority is preserved per session (not per transaction). Lock priority is set by the first read or write in the first attempt of a read-write transaction. If the application starts a new session to retry the whole transaction, the transaction loses its original lock priority. Moreover, the lock priority is only preserved if the transaction fails with an `ABORTED` error. Under some circumstances (for example, many transactions attempting to modify the same row(s)), a transaction can abort many times in a short period before successfully committing. Thus, it is not a good idea to cap the number of retries a transaction can attempt; instead, it is better to limit the total amount of time spent retrying. Idle transactions: A transaction is considered idle if it has no outstanding reads or SQL queries and has not started a read or SQL query within the last 10 seconds. Idle transactions can be aborted by Cloud Spanner so that they don't hold on to locks indefinitely. If an idle transaction is aborted, the commit fails with error `ABORTED`. If this behavior is undesirable, periodically executing a simple SQL query in the transaction (for example, `SELECT 1`) prevents the transaction from becoming idle. Snapshot read-only transactions: Snapshot read-only transactions provides a simpler method than locking read-write transactions for doing several consistent reads. However, this type of transaction does not support writes. Snapshot transactions do not take locks. Instead, they work by choosing a Cloud Spanner timestamp, then executing all reads at that timestamp. Since they do not acquire locks, they do not block concurrent read-write transactions. Unlike locking read-write transactions, snapshot read-only transactions never abort. They can fail if the chosen read timestamp is garbage collected; however, the default garbage collection policy is generous enough that most applications do not need to worry about this in practice. Snapshot read-only transactions do not need to call Commit or Rollback (and in fact are not permitted to do so). To execute a snapshot transaction, the client specifies a timestamp bound, which tells Cloud Spanner how to choose a read timestamp. The types of timestamp bound are: - Strong (the default). - Bounded staleness. - Exact staleness. If the Cloud Spanner database to be read is geographically distributed, stale read-only transactions can execute more quickly than strong or read-write transactions, because they are able to execute far from the leader replica. Each type of timestamp bound is discussed in detail below. Strong: Strong reads are guaranteed to see the effects of all transactions that have committed before the start of the read. Furthermore, all rows yielded by a single read are consistent with each other -- if any part of the read observes a transaction, all parts of the read see the transaction. Strong reads are not repeatable: two consecutive strong read-only transactions might return inconsistent results if there are concurrent writes. If consistency across reads is required, the reads should be executed within a transaction or at an exact read timestamp. Queries on change streams (see below for more details) must also specify the strong read timestamp bound. See TransactionOptions.ReadOnly.strong. Exact staleness: These timestamp bounds execute reads at a user-specified timestamp. Reads at a timestamp are guaranteed to see a consistent prefix of the global transaction history: they observe modifications done by all transactions with a commit timestamp less than or equal to the read timestamp, and observe none of the modifications done by transactions with a larger commit timestamp. They block until all conflicting transactions that can be assigned commit timestamps <= the read timestamp have finished. The timestamp can either be expressed as an absolute Cloud Spanner commit timestamp or a staleness relative to the current time. These modes do not require a \"negotiation phase\" to pick a timestamp. As a result, they execute slightly faster than the equivalent boundedly stale concurrency modes. On the other hand, boundedly stale reads usually return fresher results. See TransactionOptions.ReadOnly.read_timestamp and TransactionOptions.ReadOnly.exact_staleness. Bounded staleness: Bounded staleness modes allow Cloud Spanner to pick the read timestamp, subject to a user-provided staleness bound. Cloud Spanner chooses the newest timestamp within the staleness bound that allows execution of the reads at the closest available replica without blocking. All rows yielded are consistent with each other -- if any part of the read observes a transaction, all parts of the read see the transaction. Boundedly stale reads are not repeatable: two stale reads, even if they use the same staleness bound, can execute at different timestamps and thus return inconsistent results. Boundedly stale reads execute in two phases: the first phase negotiates a timestamp among all replicas needed to serve the read. In the second phase, reads are executed at the negotiated timestamp. As a result of the two phase execution, bounded staleness reads are usually a little slower than comparable exact staleness reads. However, they are typically able to return fresher results, and are more likely to execute at the closest replica. Because the timestamp negotiation requires up-front knowledge of which rows are read, it can only be used with single-use read-only transactions. See TransactionOptions.ReadOnly.max_staleness and TransactionOptions.ReadOnly.min_read_timestamp. Old read timestamps and garbage collection: Cloud Spanner continuously garbage collects deleted and overwritten data in the background to reclaim storage space. This process is known as \"version GC\". By default, version GC reclaims versions after they are one hour old. Because of this, Cloud Spanner can't perform reads at read timestamps more than one hour in the past. This restriction also applies to in-progress reads and/or SQL queries whose timestamp become too old while executing. Reads and SQL queries with too-old read timestamps fail with the error `FAILED_PRECONDITION`. You can configure and extend the `VERSION_RETENTION_PERIOD` of a database up to a period as long as one week, which allows Cloud Spanner to perform reads up to one week in the past. Querying change Streams: A Change Stream is a schema object that can be configured to watch data changes on the entire database, a set of tables, or a set of columns in a database. When a change stream is created, Spanner automatically defines a corresponding SQL Table-Valued Function (TVF) that can be used to query the change records in the associated change stream using the ExecuteStreamingSql API. The name of the TVF for a change stream is generated from the name of the change stream: READ_. All queries on change stream TVFs must be executed using the ExecuteStreamingSql API with a single-use read-only transaction with a strong read-only timestamp_bound. The change stream TVF allows users to specify the start_timestamp and end_timestamp for the time range of interest. All change records within the retention period is accessible using the strong read-only timestamp_bound. All other TransactionOptions are invalid for change stream queries. In addition, if TransactionOptions.read_only.return_read_timestamp is set to true, a special value of 2^63 - 2 is returned in the Transaction message that describes the transaction, instead of a valid read timestamp. This special value should be discarded and not used for any subsequent queries. Please see https://cloud.google.com/spanner/docs/change-streams for more details on how to query the change stream TVFs. Partitioned DML transactions: Partitioned DML transactions are used to execute DML statements with a different execution strategy that provides different, and often better, scalability properties for large, table-wide operations than DML in a ReadWrite transaction. Smaller scoped statements, such as an OLTP workload, should prefer using ReadWrite transactions. Partitioned DML partitions the keyspace and runs the DML statement on each partition in separate, internal transactions. These transactions commit automatically when complete, and run independently from one another. To reduce lock contention, this execution strategy only acquires read locks on rows that match the WHERE clause of the statement. Additionally, the smaller per-partition transactions hold locks for less time. That said, Partitioned DML is not a drop-in replacement for standard DML used in ReadWrite transactions. - The DML statement must be fully-partitionable. Specifically, the statement must be expressible as the union of many statements which each access only a single row of the table. - The statement is not applied atomically to all rows of the table. Rather, the statement is applied atomically to partitions of the table, in independent transactions. Secondary index rows are updated atomically with the base table rows. - Partitioned DML does not guarantee exactly-once execution semantics against a partition. The statement is applied at least once to each partition. It is strongly recommended that the DML statement should be idempotent to avoid unexpected results. For instance, it is potentially dangerous to run a statement such as `UPDATE table SET column = column + 1` as it could be run multiple times against some rows. - The partitions are committed automatically - there is no support for Commit or Rollback. If the call returns an error, or if the client issuing the ExecuteSql call dies, it is possible that some rows had the statement executed on them successfully. It is also possible that statement was never executed against other rows. - Partitioned DML transactions may only contain the execution of a single DML statement via ExecuteSql or ExecuteStreamingSql. - If any error is encountered during the execution of the partitioned DML operation (for instance, a UNIQUE INDEX violation, division by zero, or a value that can't be stored due to schema constraints), then the operation is stopped at that point and an error is returned. It is possible that at this point, some partitions have been committed (or even committed multiple times), and other partitions have not been run at all. Given the above, Partitioned DML is good fit for large, database-wide, operations that are idempotent, such as deleting old rows from a very large table.", +"description": "Options to use for transactions.", "id": "TransactionOptions", "properties": { "excludeTxnFromChangeStreams": { diff --git a/googleapiclient/discovery_cache/documents/storage.v1.json b/googleapiclient/discovery_cache/documents/storage.v1.json index 4ec902ba20..907b528155 100644 --- a/googleapiclient/discovery_cache/documents/storage.v1.json +++ b/googleapiclient/discovery_cache/documents/storage.v1.json @@ -253,7 +253,7 @@ "location": "northamerica-south1" } ], -"etag": "\"36383730363437323837383838393538333732\"", +"etag": "\"31353834363234393531343533373938353433\"", "icons": { "x16": "https://www.google.com/images/icons/product/cloud_storage-16.png", "x32": "https://www.google.com/images/icons/product/cloud_storage-32.png" @@ -3306,6 +3306,11 @@ "location": "query", "type": "string" }, +"filter": { +"description": "Filter the returned objects. Currently only supported for the contexts field. If delimiter is set, the returned prefixes are exempt from this filter.", +"location": "query", +"type": "string" +}, "includeFoldersAsPrefixes": { "description": "Only applicable if delimiter is set to '/'. If true, will also include folders and managed folders (besides objects) in the returned prefixes.", "location": "query", @@ -4539,7 +4544,7 @@ } } }, -"revision": "20250605", +"revision": "20250629", "rootUrl": "https://storage.googleapis.com/", "schemas": { "AdvanceRelocateBucketOperationRequest": { @@ -5932,6 +5937,20 @@ "description": "Content-Type of the object data. If an object is stored without a Content-Type, it is served as application/octet-stream.", "type": "string" }, +"contexts": { +"description": "User-defined or system-defined object contexts. Each object context is a key-payload pair, where the key provides the identification and the payload holds the associated value and additional metadata.", +"properties": { +"custom": { +"additionalProperties": { +"$ref": "ObjectCustomContextPayload", +"description": "A single user-defined object context." +}, +"description": "User-defined object contexts.", +"type": "object" +} +}, +"type": "object" +}, "crc32c": { "description": "CRC32c checksum, as described in RFC 4960, Appendix B; encoded using base64 in big-endian byte order. For more information about using the CRC32c checksum, see [Data Validation and Change Detection](https://cloud.google.com/storage/docs/data-validation).", "type": "string" @@ -6201,6 +6220,27 @@ }, "type": "object" }, +"ObjectCustomContextPayload": { +"description": "The payload of a single user-defined object context.", +"id": "ObjectCustomContextPayload", +"properties": { +"createTime": { +"description": "The time at which the object context was created in RFC 3339 format.", +"format": "date-time", +"type": "string" +}, +"updateTime": { +"description": "The time at which the object context was last updated in RFC 3339 format.", +"format": "date-time", +"type": "string" +}, +"value": { +"description": "The value of the object context.", +"type": "string" +} +}, +"type": "object" +}, "Objects": { "description": "A list of objects.", "id": "Objects", diff --git a/googleapiclient/discovery_cache/documents/workspaceevents.v1.json b/googleapiclient/discovery_cache/documents/workspaceevents.v1.json index 816afc27f9..0af9aa6429 100644 --- a/googleapiclient/discovery_cache/documents/workspaceevents.v1.json +++ b/googleapiclient/discovery_cache/documents/workspaceevents.v1.json @@ -401,7 +401,7 @@ "type": "string" }, "updateMask": { -"description": "Optional. The field to update. If omitted, updates any fields included in the request. You can update one of the following fields in a subscription: * `expire_time`: The timestamp when the subscription expires. * `ttl`: The time-to-live (TTL) or duration of the subscription. * `event_types`: The list of event types to receive about the target resource. To fully replace the subscription (the equivalent of `PUT`), use `*`. Any omitted fields are updated with empty values.", +"description": "Optional. The field to update. If omitted, updates any fields included in the request. You can update one of the following fields in a subscription: * `expire_time`: The timestamp when the subscription expires. * `ttl`: The time-to-live (TTL) or duration of the subscription. * `event_types`: The list of event types to receive about the target resource. When using the `*` wildcard (equivalent to `PUT`), omitted fields are set to empty values and rejected if they're invalid.", "format": "google-fieldmask", "location": "query", "type": "string" @@ -486,7 +486,7 @@ } } }, -"revision": "20250615", +"revision": "20250629", "rootUrl": "https://workspaceevents.googleapis.com/", "schemas": { "ListSubscriptionsResponse": { diff --git a/googleapiclient/discovery_cache/documents/workstations.v1beta.json b/googleapiclient/discovery_cache/documents/workstations.v1beta.json index a6688a6c39..293000ac31 100644 --- a/googleapiclient/discovery_cache/documents/workstations.v1beta.json +++ b/googleapiclient/discovery_cache/documents/workstations.v1beta.json @@ -1142,7 +1142,7 @@ } } }, -"revision": "20250604", +"revision": "20250625", "rootUrl": "https://workstations.googleapis.com/", "schemas": { "Accelerator": { @@ -1263,6 +1263,10 @@ "description": "Optional. The number of boost VMs that the system should keep idle so that workstations can be boosted quickly. Defaults to `0`.", "format": "int32", "type": "integer" +}, +"reservationAffinity": { +"$ref": "ReservationAffinity", +"description": "Optional. [ReservationAffinity](https://cloud.google.com/compute/docs/instances/reserving-zonal-resources) specifies a reservation that can be consumed to create boost VM instances. If SPECIFIC_RESERVATION is specified, Cloud Workstations will only create VMs in the zone where the reservation is located. This would affect availability since the service will no longer be resilient to zonal outages. If ANY_RESERVATION is specified, creating reservations in both zones that the config creates VMs in will ensure higher availability. **Important Considerations for Reservation Affinity:** * This feature is intended for advanced users and requires familiarity with Google Compute Engine reservations. * Using reservations incurs charges, regardless of utilization. * The resources in the pool will consume the specified reservation. Take this into account when setting the pool size." } }, "type": "object" @@ -1454,6 +1458,10 @@ "readOnly": true, "type": "integer" }, +"reservationAffinity": { +"$ref": "ReservationAffinity", +"description": "Optional. [ReservationAffinity](https://cloud.google.com/compute/docs/instances/reserving-zonal-resources) specifies a reservation that can be consumed to create boost VM instances. If SPECIFIC_RESERVATION is specified, Cloud Workstations will only create VMs in the zone where the reservation is located. This would affect availability since the service will no longer be resilient to zonal outages. If ANY_RESERVATION is specified, creating reservations in both zones that the config creates VMs in will ensure higher availability. **Important Considerations for Reservation Affinity:** * This feature is intended for advanced users and requires familiarity with Google Compute Engine reservations. * Using reservations incurs charges, regardless of utilization. * The resources in the pool will consume the specified reservation. Take this into account when setting the pool size." +}, "serviceAccount": { "description": "Optional. The email address of the service account for Cloud Workstations VMs created with this configuration. When specified, be sure that the service account has `logging.logEntries.create` and `monitoring.timeSeries.create` permissions on the project so it can write logs out to Cloud Logging. If using a custom container image, the service account must have [Artifact Registry Reader](https://cloud.google.com/artifact-registry/docs/access-control#roles) permission to pull the specified image. If you as the administrator want to be able to `ssh` into the underlying VM, you need to set this value to a service account for which you have the `iam.serviceAccounts.actAs` permission. Conversely, if you don't want anyone to be able to `ssh` into the underlying VM, use a service account where no one has that permission. If not set, VMs run with a service account provided by the Cloud Workstations service, and the image must be publicly accessible.", "type": "string" @@ -1987,6 +1995,40 @@ }, "type": "object" }, +"ReservationAffinity": { +"description": "ReservationAffinity is the configuration of the desired reservation from which instances can consume resources.", +"id": "ReservationAffinity", +"properties": { +"consumeReservationType": { +"description": "Optional. Corresponds to the type of reservation consumption.", +"enum": [ +"TYPE_UNSPECIFIED", +"NO_RESERVATION", +"ANY_RESERVATION", +"SPECIFIC_RESERVATION" +], +"enumDescriptions": [ +"Default value. This should not be used.", +"Do not consume from any reserved capacity.", +"Consume any reservation available.", +"Must consume from a specific reservation. Must specify key value fields for specifying the reservations." +], +"type": "string" +}, +"key": { +"description": "Optional. Corresponds to the label key of reservation resource.", +"type": "string" +}, +"values": { +"description": "Optional. Corresponds to the label values of reservation resources. Valid values are either a name to a reservation in the same project or \"projects/{project}/reservations/{reservation}\" to target a shared reservation in the same zone but in a different project.", +"items": { +"type": "string" +}, +"type": "array" +} +}, +"type": "object" +}, "RuntimeHost": { "description": "Runtime host for the workstation.", "id": "RuntimeHost", diff --git a/googleapiclient/version.py b/googleapiclient/version.py index f59c8403c0..52ed41310d 100644 --- a/googleapiclient/version.py +++ b/googleapiclient/version.py @@ -12,4 +12,4 @@ # See the License for the specific language governing permissions and # limitations under the License. -__version__ = "2.175.0" +__version__ = "2.176.0"