From 15a849dd60e824337e45e695ae2148413f240e73 Mon Sep 17 00:00:00 2001 From: "gcf-owl-bot[bot]" <78513119+gcf-owl-bot[bot]@users.noreply.github.com> Date: Fri, 13 Sep 2024 22:15:51 +0000 Subject: [PATCH] feat: [Dataproc] add resource reference for KMS keys and fix comments (#7671) PiperOrigin-RevId: 674408200 Source-Link: https://github.com/googleapis/googleapis/commit/964f6c9ab536dd33a99bbba8d9d8071d0cf39f89 Source-Link: https://github.com/googleapis/googleapis-gen/commit/2979ec5cbcf2abae857f9a4cbe5cca09f040cb58 Copy-Tag: eyJwIjoiRGF0YXByb2MvLk93bEJvdC55YW1sIiwiaCI6IjI5NzllYzVjYmNmMmFiYWU4NTdmOWE0Y2JlNWNjYTA5ZjA0MGNiNTgifQ== --- Dataproc/metadata/V1/Clusters.php | Bin 15156 -> 15194 bytes Dataproc/metadata/V1/Shared.php | Bin 5473 -> 5596 bytes Dataproc/metadata/V1/WorkflowTemplates.php | Bin 9255 -> 9293 bytes Dataproc/src/V1/AcceleratorConfig.php | 32 +++---- .../src/V1/Client/ClusterControllerClient.php | 22 +++++ .../Client/WorkflowTemplateServiceClient.php | 22 +++++ Dataproc/src/V1/Component.php | 9 +- Dataproc/src/V1/DiagnoseClusterRequest.php | 8 +- Dataproc/src/V1/EncryptionConfig.php | 34 ++++--- Dataproc/src/V1/FlinkJob.php | 8 +- Dataproc/src/V1/GceClusterConfig.php | 88 +++++++++++------- Dataproc/src/V1/HadoopJob.php | 16 ++-- Dataproc/src/V1/HiveJob.php | 8 +- Dataproc/src/V1/Job.php | 48 +++++----- Dataproc/src/V1/JobScheduling.php | 32 +++---- Dataproc/src/V1/JobStatus.php | 8 +- Dataproc/src/V1/JobStatus/Substate.php | 4 +- Dataproc/src/V1/KerberosConfig.php | 8 +- Dataproc/src/V1/ListClustersRequest.php | 60 ++++++------ Dataproc/src/V1/LoggingConfig.php | 8 +- Dataproc/src/V1/PigJob.php | 8 +- Dataproc/src/V1/PySparkJob.php | 8 +- Dataproc/src/V1/SoftwareConfig.php | 8 +- Dataproc/src/V1/SparkJob.php | 17 ++-- Dataproc/src/V1/SparkRJob.php | 8 +- Dataproc/src/V1/SparkSqlJob.php | 8 +- .../V1/WorkflowTemplate/EncryptionConfig.php | 6 +- .../cluster_controller_descriptor_config.php | 1 + ...low_template_service_descriptor_config.php | 1 + 29 files changed, 283 insertions(+), 197 deletions(-) diff --git a/Dataproc/metadata/V1/Clusters.php b/Dataproc/metadata/V1/Clusters.php index 5b2596d272e88198f6fc5bc159ef5827a66eb26e..3962d2eb574def42308956d327997996057cce11 100644 GIT binary patch delta 93 zcmdl|cB^c|Ll)MRIb1R7lj~R|H~(i5<7O7@C< wk@1(KGM8dvD{;W6AQ~sdoII?-iDjYS)3RdJ2&rQO=Z&5;1Y1nOD?J`D9OyvbI#98%S;y% e;bPCuEsoDlt(0I?VASArVt(MrxH*sGzBB-Sv=;sV diff --git a/Dataproc/src/V1/AcceleratorConfig.php b/Dataproc/src/V1/AcceleratorConfig.php index caa0cd0d3675..ddc545cd74ed 100644 --- a/Dataproc/src/V1/AcceleratorConfig.php +++ b/Dataproc/src/V1/AcceleratorConfig.php @@ -23,14 +23,14 @@ class AcceleratorConfig extends \Google\Protobuf\Internal\Message * [Compute Engine * AcceleratorTypes](https://cloud.google.com/compute/docs/reference/v1/acceleratorTypes). * Examples: - * * `https://www.googleapis.com/compute/v1/projects/[project_id]/zones/[zone]/acceleratorTypes/nvidia-tesla-k80` - * * `projects/[project_id]/zones/[zone]/acceleratorTypes/nvidia-tesla-k80` - * * `nvidia-tesla-k80` + * * `https://www.googleapis.com/compute/v1/projects/[project_id]/zones/[zone]/acceleratorTypes/nvidia-tesla-t4` + * * `projects/[project_id]/zones/[zone]/acceleratorTypes/nvidia-tesla-t4` + * * `nvidia-tesla-t4` * **Auto Zone Exception**: If you are using the Dataproc * [Auto Zone * Placement](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) * feature, you must use the short name of the accelerator type - * resource, for example, `nvidia-tesla-k80`. + * resource, for example, `nvidia-tesla-t4`. * * Generated from protobuf field string accelerator_type_uri = 1; */ @@ -54,14 +54,14 @@ class AcceleratorConfig extends \Google\Protobuf\Internal\Message * [Compute Engine * AcceleratorTypes](https://cloud.google.com/compute/docs/reference/v1/acceleratorTypes). * Examples: - * * `https://www.googleapis.com/compute/v1/projects/[project_id]/zones/[zone]/acceleratorTypes/nvidia-tesla-k80` - * * `projects/[project_id]/zones/[zone]/acceleratorTypes/nvidia-tesla-k80` - * * `nvidia-tesla-k80` + * * `https://www.googleapis.com/compute/v1/projects/[project_id]/zones/[zone]/acceleratorTypes/nvidia-tesla-t4` + * * `projects/[project_id]/zones/[zone]/acceleratorTypes/nvidia-tesla-t4` + * * `nvidia-tesla-t4` * **Auto Zone Exception**: If you are using the Dataproc * [Auto Zone * Placement](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) * feature, you must use the short name of the accelerator type - * resource, for example, `nvidia-tesla-k80`. + * resource, for example, `nvidia-tesla-t4`. * @type int $accelerator_count * The number of the accelerator cards of this type exposed to this instance. * } @@ -77,14 +77,14 @@ public function __construct($data = NULL) { * [Compute Engine * AcceleratorTypes](https://cloud.google.com/compute/docs/reference/v1/acceleratorTypes). * Examples: - * * `https://www.googleapis.com/compute/v1/projects/[project_id]/zones/[zone]/acceleratorTypes/nvidia-tesla-k80` - * * `projects/[project_id]/zones/[zone]/acceleratorTypes/nvidia-tesla-k80` - * * `nvidia-tesla-k80` + * * `https://www.googleapis.com/compute/v1/projects/[project_id]/zones/[zone]/acceleratorTypes/nvidia-tesla-t4` + * * `projects/[project_id]/zones/[zone]/acceleratorTypes/nvidia-tesla-t4` + * * `nvidia-tesla-t4` * **Auto Zone Exception**: If you are using the Dataproc * [Auto Zone * Placement](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) * feature, you must use the short name of the accelerator type - * resource, for example, `nvidia-tesla-k80`. + * resource, for example, `nvidia-tesla-t4`. * * Generated from protobuf field string accelerator_type_uri = 1; * @return string @@ -100,14 +100,14 @@ public function getAcceleratorTypeUri() * [Compute Engine * AcceleratorTypes](https://cloud.google.com/compute/docs/reference/v1/acceleratorTypes). * Examples: - * * `https://www.googleapis.com/compute/v1/projects/[project_id]/zones/[zone]/acceleratorTypes/nvidia-tesla-k80` - * * `projects/[project_id]/zones/[zone]/acceleratorTypes/nvidia-tesla-k80` - * * `nvidia-tesla-k80` + * * `https://www.googleapis.com/compute/v1/projects/[project_id]/zones/[zone]/acceleratorTypes/nvidia-tesla-t4` + * * `projects/[project_id]/zones/[zone]/acceleratorTypes/nvidia-tesla-t4` + * * `nvidia-tesla-t4` * **Auto Zone Exception**: If you are using the Dataproc * [Auto Zone * Placement](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) * feature, you must use the short name of the accelerator type - * resource, for example, `nvidia-tesla-k80`. + * resource, for example, `nvidia-tesla-t4`. * * Generated from protobuf field string accelerator_type_uri = 1; * @param string $var diff --git a/Dataproc/src/V1/Client/ClusterControllerClient.php b/Dataproc/src/V1/Client/ClusterControllerClient.php index 0085f196ecc4..bddd5b3913bd 100644 --- a/Dataproc/src/V1/Client/ClusterControllerClient.php +++ b/Dataproc/src/V1/Client/ClusterControllerClient.php @@ -214,6 +214,27 @@ public static function clusterRegionName(string $project, string $region, string ]); } + /** + * Formats a string containing the fully-qualified path to represent a crypto_key + * resource. + * + * @param string $project + * @param string $location + * @param string $keyRing + * @param string $cryptoKey + * + * @return string The formatted crypto_key resource. + */ + public static function cryptoKeyName(string $project, string $location, string $keyRing, string $cryptoKey): string + { + return self::getPathTemplate('cryptoKey')->render([ + 'project' => $project, + 'location' => $location, + 'key_ring' => $keyRing, + 'crypto_key' => $cryptoKey, + ]); + } + /** * Formats a string containing the fully-qualified path to represent a node_group * resource. @@ -260,6 +281,7 @@ public static function serviceName(string $project, string $location, string $se * Template: Pattern * - cluster: projects/{project}/locations/{location}/clusters/{cluster} * - clusterRegion: projects/{project}/regions/{region}/clusters/{cluster} + * - cryptoKey: projects/{project}/locations/{location}/keyRings/{key_ring}/cryptoKeys/{crypto_key} * - nodeGroup: projects/{project}/regions/{region}/clusters/{cluster}/nodeGroups/{node_group} * - service: projects/{project}/locations/{location}/services/{service} * diff --git a/Dataproc/src/V1/Client/WorkflowTemplateServiceClient.php b/Dataproc/src/V1/Client/WorkflowTemplateServiceClient.php index e879ce3936a6..f430890f58e0 100644 --- a/Dataproc/src/V1/Client/WorkflowTemplateServiceClient.php +++ b/Dataproc/src/V1/Client/WorkflowTemplateServiceClient.php @@ -193,6 +193,27 @@ public static function clusterRegionName(string $project, string $region, string ]); } + /** + * Formats a string containing the fully-qualified path to represent a crypto_key + * resource. + * + * @param string $project + * @param string $location + * @param string $keyRing + * @param string $cryptoKey + * + * @return string The formatted crypto_key resource. + */ + public static function cryptoKeyName(string $project, string $location, string $keyRing, string $cryptoKey): string + { + return self::getPathTemplate('cryptoKey')->render([ + 'project' => $project, + 'location' => $location, + 'key_ring' => $keyRing, + 'crypto_key' => $cryptoKey, + ]); + } + /** * Formats a string containing the fully-qualified path to represent a location * resource. @@ -335,6 +356,7 @@ public static function workflowTemplateName(string $project, string $region, str * The following name formats are supported: * Template: Pattern * - clusterRegion: projects/{project}/regions/{region}/clusters/{cluster} + * - cryptoKey: projects/{project}/locations/{location}/keyRings/{key_ring}/cryptoKeys/{crypto_key} * - location: projects/{project}/locations/{location} * - nodeGroup: projects/{project}/regions/{region}/clusters/{cluster}/nodeGroups/{node_group} * - projectLocationWorkflowTemplate: projects/{project}/locations/{location}/workflowTemplates/{workflow_template} diff --git a/Dataproc/src/V1/Component.php b/Dataproc/src/V1/Component.php index 077fa77a16d7..765ac34f3243 100644 --- a/Dataproc/src/V1/Component.php +++ b/Dataproc/src/V1/Component.php @@ -20,10 +20,11 @@ class Component */ const COMPONENT_UNSPECIFIED = 0; /** - * The Anaconda python distribution. The Anaconda component is not supported - * in the Dataproc [2.0 image] - * (/https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-release-2.0). - * The 2.0 image is pre-installed with Miniconda. + * The Anaconda component is no longer supported or applicable to + * [supported Dataproc on Compute Engine image versions] + * (https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-version-clusters#supported-dataproc-image-versions). + * It cannot be activated on clusters created with supported Dataproc on + * Compute Engine image versions. * * Generated from protobuf enum ANACONDA = 5; */ diff --git a/Dataproc/src/V1/DiagnoseClusterRequest.php b/Dataproc/src/V1/DiagnoseClusterRequest.php index 2b1185b7a6f9..27c9e44900e7 100644 --- a/Dataproc/src/V1/DiagnoseClusterRequest.php +++ b/Dataproc/src/V1/DiagnoseClusterRequest.php @@ -35,7 +35,7 @@ class DiagnoseClusterRequest extends \Google\Protobuf\Internal\Message */ protected $cluster_name = ''; /** - * Optional. The output Cloud Storage directory for the diagnostic + * Optional. (Optional) The output Cloud Storage directory for the diagnostic * tarball. If not specified, a task-specific directory in the cluster's * staging bucket will be used. * @@ -103,7 +103,7 @@ public static function build(string $projectId, string $region, string $clusterN * @type string $cluster_name * Required. The cluster name. * @type string $tarball_gcs_dir - * Optional. The output Cloud Storage directory for the diagnostic + * Optional. (Optional) The output Cloud Storage directory for the diagnostic * tarball. If not specified, a task-specific directory in the cluster's * staging bucket will be used. * @type int $tarball_access @@ -206,7 +206,7 @@ public function setClusterName($var) } /** - * Optional. The output Cloud Storage directory for the diagnostic + * Optional. (Optional) The output Cloud Storage directory for the diagnostic * tarball. If not specified, a task-specific directory in the cluster's * staging bucket will be used. * @@ -219,7 +219,7 @@ public function getTarballGcsDir() } /** - * Optional. The output Cloud Storage directory for the diagnostic + * Optional. (Optional) The output Cloud Storage directory for the diagnostic * tarball. If not specified, a task-specific directory in the cluster's * staging bucket will be used. * diff --git a/Dataproc/src/V1/EncryptionConfig.php b/Dataproc/src/V1/EncryptionConfig.php index 4cd725c08ba1..cbd06d6f50a0 100644 --- a/Dataproc/src/V1/EncryptionConfig.php +++ b/Dataproc/src/V1/EncryptionConfig.php @@ -16,8 +16,11 @@ class EncryptionConfig extends \Google\Protobuf\Internal\Message { /** - * Optional. The Cloud KMS key name to use for PD disk encryption for all - * instances in the cluster. + * Optional. The Cloud KMS key resource name to use for persistent disk + * encryption for all instances in the cluster. See [Use CMEK with cluster + * data] + * (https://cloud.google.com//dataproc/docs/concepts/configuring-clusters/customer-managed-encryption#use_cmek_with_cluster_data) + * for more information. * * Generated from protobuf field string gce_pd_kms_key_name = 1 [(.google.api.field_behavior) = OPTIONAL]; */ @@ -48,7 +51,7 @@ class EncryptionConfig extends \Google\Protobuf\Internal\Message * * [PrestoJob](https://cloud.google.com/dataproc/docs/reference/rest/v1/PrestoJob) * scriptVariables and queryList.queries * - * Generated from protobuf field string kms_key = 2 [(.google.api.field_behavior) = OPTIONAL]; + * Generated from protobuf field string kms_key = 2 [(.google.api.field_behavior) = OPTIONAL, (.google.api.resource_reference) = { */ protected $kms_key = ''; @@ -59,8 +62,11 @@ class EncryptionConfig extends \Google\Protobuf\Internal\Message * Optional. Data for populating the Message object. * * @type string $gce_pd_kms_key_name - * Optional. The Cloud KMS key name to use for PD disk encryption for all - * instances in the cluster. + * Optional. The Cloud KMS key resource name to use for persistent disk + * encryption for all instances in the cluster. See [Use CMEK with cluster + * data] + * (https://cloud.google.com//dataproc/docs/concepts/configuring-clusters/customer-managed-encryption#use_cmek_with_cluster_data) + * for more information. * @type string $kms_key * Optional. The Cloud KMS key resource name to use for cluster persistent * disk and job argument encryption. See [Use CMEK with cluster data] @@ -94,8 +100,11 @@ public function __construct($data = NULL) { } /** - * Optional. The Cloud KMS key name to use for PD disk encryption for all - * instances in the cluster. + * Optional. The Cloud KMS key resource name to use for persistent disk + * encryption for all instances in the cluster. See [Use CMEK with cluster + * data] + * (https://cloud.google.com//dataproc/docs/concepts/configuring-clusters/customer-managed-encryption#use_cmek_with_cluster_data) + * for more information. * * Generated from protobuf field string gce_pd_kms_key_name = 1 [(.google.api.field_behavior) = OPTIONAL]; * @return string @@ -106,8 +115,11 @@ public function getGcePdKmsKeyName() } /** - * Optional. The Cloud KMS key name to use for PD disk encryption for all - * instances in the cluster. + * Optional. The Cloud KMS key resource name to use for persistent disk + * encryption for all instances in the cluster. See [Use CMEK with cluster + * data] + * (https://cloud.google.com//dataproc/docs/concepts/configuring-clusters/customer-managed-encryption#use_cmek_with_cluster_data) + * for more information. * * Generated from protobuf field string gce_pd_kms_key_name = 1 [(.google.api.field_behavior) = OPTIONAL]; * @param string $var @@ -147,7 +159,7 @@ public function setGcePdKmsKeyName($var) * * [PrestoJob](https://cloud.google.com/dataproc/docs/reference/rest/v1/PrestoJob) * scriptVariables and queryList.queries * - * Generated from protobuf field string kms_key = 2 [(.google.api.field_behavior) = OPTIONAL]; + * Generated from protobuf field string kms_key = 2 [(.google.api.field_behavior) = OPTIONAL, (.google.api.resource_reference) = { * @return string */ public function getKmsKey() @@ -181,7 +193,7 @@ public function getKmsKey() * * [PrestoJob](https://cloud.google.com/dataproc/docs/reference/rest/v1/PrestoJob) * scriptVariables and queryList.queries * - * Generated from protobuf field string kms_key = 2 [(.google.api.field_behavior) = OPTIONAL]; + * Generated from protobuf field string kms_key = 2 [(.google.api.field_behavior) = OPTIONAL, (.google.api.resource_reference) = { * @param string $var * @return $this */ diff --git a/Dataproc/src/V1/FlinkJob.php b/Dataproc/src/V1/FlinkJob.php index ba2e179446aa..2ff8f95df78c 100644 --- a/Dataproc/src/V1/FlinkJob.php +++ b/Dataproc/src/V1/FlinkJob.php @@ -41,7 +41,7 @@ class FlinkJob extends \Google\Protobuf\Internal\Message * Optional. A mapping of property names to values, used to configure Flink. * Properties that conflict with values set by the Dataproc API might be * overwritten. Can include properties set in - * /etc/flink/conf/flink-defaults.conf and classes in user code. + * `/etc/flink/conf/flink-defaults.conf` and classes in user code. * * Generated from protobuf field map properties = 7 [(.google.api.field_behavior) = OPTIONAL]; */ @@ -80,7 +80,7 @@ class FlinkJob extends \Google\Protobuf\Internal\Message * Optional. A mapping of property names to values, used to configure Flink. * Properties that conflict with values set by the Dataproc API might be * overwritten. Can include properties set in - * /etc/flink/conf/flink-defaults.conf and classes in user code. + * `/etc/flink/conf/flink-defaults.conf` and classes in user code. * @type \Google\Cloud\Dataproc\V1\LoggingConfig $logging_config * Optional. The runtime log config for job execution. * } @@ -246,7 +246,7 @@ public function setSavepointUri($var) * Optional. A mapping of property names to values, used to configure Flink. * Properties that conflict with values set by the Dataproc API might be * overwritten. Can include properties set in - * /etc/flink/conf/flink-defaults.conf and classes in user code. + * `/etc/flink/conf/flink-defaults.conf` and classes in user code. * * Generated from protobuf field map properties = 7 [(.google.api.field_behavior) = OPTIONAL]; * @return \Google\Protobuf\Internal\MapField @@ -260,7 +260,7 @@ public function getProperties() * Optional. A mapping of property names to values, used to configure Flink. * Properties that conflict with values set by the Dataproc API might be * overwritten. Can include properties set in - * /etc/flink/conf/flink-defaults.conf and classes in user code. + * `/etc/flink/conf/flink-defaults.conf` and classes in user code. * * Generated from protobuf field map properties = 7 [(.google.api.field_behavior) = OPTIONAL]; * @param array|\Google\Protobuf\Internal\MapField $var diff --git a/Dataproc/src/V1/GceClusterConfig.php b/Dataproc/src/V1/GceClusterConfig.php index f845773efb02..a3a2964165e4 100644 --- a/Dataproc/src/V1/GceClusterConfig.php +++ b/Dataproc/src/V1/GceClusterConfig.php @@ -55,12 +55,18 @@ class GceClusterConfig extends \Google\Protobuf\Internal\Message */ protected $subnetwork_uri = ''; /** - * Optional. If true, all instances in the cluster will only have internal IP - * addresses. By default, clusters are not restricted to internal IP - * addresses, and will have ephemeral external IP addresses assigned to each - * instance. This `internal_ip_only` restriction can only be enabled for - * subnetwork enabled networks, and all off-cluster dependencies must be - * configured to be accessible without external IP addresses. + * Optional. This setting applies to subnetwork-enabled networks. It is set to + * `true` by default in clusters created with image versions 2.2.x. + * When set to `true`: + * * All cluster VMs have internal IP addresses. + * * [Google Private Access] + * (https://cloud.google.com/vpc/docs/private-google-access) + * must be enabled to access Dataproc and other Google Cloud APIs. + * * Off-cluster dependencies must be configured to be accessible + * without external IP addresses. + * When set to `false`: + * * Cluster VMs are not restricted to internal IP addresses. + * * Ephemeral external IP addresses are assigned to each cluster VM. * * Generated from protobuf field optional bool internal_ip_only = 7 [(.google.api.field_behavior) = OPTIONAL]; */ @@ -103,8 +109,8 @@ class GceClusterConfig extends \Google\Protobuf\Internal\Message */ private $service_account_scopes; /** - * The Compute Engine tags to add to all instances (see [Tagging - * instances](https://cloud.google.com/compute/docs/label-or-tag-resources#tags)). + * The Compute Engine network tags to add to all instances (see [Tagging + * instances](https://cloud.google.com/vpc/docs/add-remove-network-tags)). * * Generated from protobuf field repeated string tags = 4; */ @@ -178,12 +184,18 @@ class GceClusterConfig extends \Google\Protobuf\Internal\Message * * `projects/[project_id]/regions/[region]/subnetworks/sub0` * * `sub0` * @type bool $internal_ip_only - * Optional. If true, all instances in the cluster will only have internal IP - * addresses. By default, clusters are not restricted to internal IP - * addresses, and will have ephemeral external IP addresses assigned to each - * instance. This `internal_ip_only` restriction can only be enabled for - * subnetwork enabled networks, and all off-cluster dependencies must be - * configured to be accessible without external IP addresses. + * Optional. This setting applies to subnetwork-enabled networks. It is set to + * `true` by default in clusters created with image versions 2.2.x. + * When set to `true`: + * * All cluster VMs have internal IP addresses. + * * [Google Private Access] + * (https://cloud.google.com/vpc/docs/private-google-access) + * must be enabled to access Dataproc and other Google Cloud APIs. + * * Off-cluster dependencies must be configured to be accessible + * without external IP addresses. + * When set to `false`: + * * Cluster VMs are not restricted to internal IP addresses. + * * Ephemeral external IP addresses are assigned to each cluster VM. * @type int $private_ipv6_google_access * Optional. The type of IPv6 access for a cluster. * @type string $service_account @@ -210,8 +222,8 @@ class GceClusterConfig extends \Google\Protobuf\Internal\Message * * https://www.googleapis.com/auth/bigtable.data * * https://www.googleapis.com/auth/devstorage.full_control * @type array|\Google\Protobuf\Internal\RepeatedField $tags - * The Compute Engine tags to add to all instances (see [Tagging - * instances](https://cloud.google.com/compute/docs/label-or-tag-resources#tags)). + * The Compute Engine network tags to add to all instances (see [Tagging + * instances](https://cloud.google.com/vpc/docs/add-remove-network-tags)). * @type array|\Google\Protobuf\Internal\MapField $metadata * Optional. The Compute Engine metadata entries to add to all instances (see * [Project and instance @@ -353,12 +365,18 @@ public function setSubnetworkUri($var) } /** - * Optional. If true, all instances in the cluster will only have internal IP - * addresses. By default, clusters are not restricted to internal IP - * addresses, and will have ephemeral external IP addresses assigned to each - * instance. This `internal_ip_only` restriction can only be enabled for - * subnetwork enabled networks, and all off-cluster dependencies must be - * configured to be accessible without external IP addresses. + * Optional. This setting applies to subnetwork-enabled networks. It is set to + * `true` by default in clusters created with image versions 2.2.x. + * When set to `true`: + * * All cluster VMs have internal IP addresses. + * * [Google Private Access] + * (https://cloud.google.com/vpc/docs/private-google-access) + * must be enabled to access Dataproc and other Google Cloud APIs. + * * Off-cluster dependencies must be configured to be accessible + * without external IP addresses. + * When set to `false`: + * * Cluster VMs are not restricted to internal IP addresses. + * * Ephemeral external IP addresses are assigned to each cluster VM. * * Generated from protobuf field optional bool internal_ip_only = 7 [(.google.api.field_behavior) = OPTIONAL]; * @return bool @@ -379,12 +397,18 @@ public function clearInternalIpOnly() } /** - * Optional. If true, all instances in the cluster will only have internal IP - * addresses. By default, clusters are not restricted to internal IP - * addresses, and will have ephemeral external IP addresses assigned to each - * instance. This `internal_ip_only` restriction can only be enabled for - * subnetwork enabled networks, and all off-cluster dependencies must be - * configured to be accessible without external IP addresses. + * Optional. This setting applies to subnetwork-enabled networks. It is set to + * `true` by default in clusters created with image versions 2.2.x. + * When set to `true`: + * * All cluster VMs have internal IP addresses. + * * [Google Private Access] + * (https://cloud.google.com/vpc/docs/private-google-access) + * must be enabled to access Dataproc and other Google Cloud APIs. + * * Off-cluster dependencies must be configured to be accessible + * without external IP addresses. + * When set to `false`: + * * Cluster VMs are not restricted to internal IP addresses. + * * Ephemeral external IP addresses are assigned to each cluster VM. * * Generated from protobuf field optional bool internal_ip_only = 7 [(.google.api.field_behavior) = OPTIONAL]; * @param bool $var @@ -515,8 +539,8 @@ public function setServiceAccountScopes($var) } /** - * The Compute Engine tags to add to all instances (see [Tagging - * instances](https://cloud.google.com/compute/docs/label-or-tag-resources#tags)). + * The Compute Engine network tags to add to all instances (see [Tagging + * instances](https://cloud.google.com/vpc/docs/add-remove-network-tags)). * * Generated from protobuf field repeated string tags = 4; * @return \Google\Protobuf\Internal\RepeatedField @@ -527,8 +551,8 @@ public function getTags() } /** - * The Compute Engine tags to add to all instances (see [Tagging - * instances](https://cloud.google.com/compute/docs/label-or-tag-resources#tags)). + * The Compute Engine network tags to add to all instances (see [Tagging + * instances](https://cloud.google.com/vpc/docs/add-remove-network-tags)). * * Generated from protobuf field repeated string tags = 4; * @param array|\Google\Protobuf\Internal\RepeatedField $var diff --git a/Dataproc/src/V1/HadoopJob.php b/Dataproc/src/V1/HadoopJob.php index 047ef763aefa..71599754da5a 100644 --- a/Dataproc/src/V1/HadoopJob.php +++ b/Dataproc/src/V1/HadoopJob.php @@ -22,7 +22,7 @@ class HadoopJob extends \Google\Protobuf\Internal\Message /** * Optional. The arguments to pass to the driver. Do not * include arguments, such as `-libjars` or `-Dfoo=bar`, that can be set as - * job properties, since a collision may occur that causes an incorrect job + * job properties, since a collision might occur that causes an incorrect job * submission. * * Generated from protobuf field repeated string args = 3 [(.google.api.field_behavior) = OPTIONAL]; @@ -53,7 +53,7 @@ class HadoopJob extends \Google\Protobuf\Internal\Message private $archive_uris; /** * Optional. A mapping of property names to values, used to configure Hadoop. - * Properties that conflict with values set by the Dataproc API may be + * Properties that conflict with values set by the Dataproc API might be * overwritten. Can include properties set in `/etc/hadoop/conf/*-site` and * classes in user code. * @@ -86,7 +86,7 @@ class HadoopJob extends \Google\Protobuf\Internal\Message * @type array|\Google\Protobuf\Internal\RepeatedField $args * Optional. The arguments to pass to the driver. Do not * include arguments, such as `-libjars` or `-Dfoo=bar`, that can be set as - * job properties, since a collision may occur that causes an incorrect job + * job properties, since a collision might occur that causes an incorrect job * submission. * @type array|\Google\Protobuf\Internal\RepeatedField $jar_file_uris * Optional. Jar file URIs to add to the CLASSPATHs of the @@ -101,7 +101,7 @@ class HadoopJob extends \Google\Protobuf\Internal\Message * .jar, .tar, .tar.gz, .tgz, or .zip. * @type array|\Google\Protobuf\Internal\MapField $properties * Optional. A mapping of property names to values, used to configure Hadoop. - * Properties that conflict with values set by the Dataproc API may be + * Properties that conflict with values set by the Dataproc API might be * overwritten. Can include properties set in `/etc/hadoop/conf/*-site` and * classes in user code. * @type \Google\Cloud\Dataproc\V1\LoggingConfig $logging_config @@ -188,7 +188,7 @@ public function setMainClass($var) /** * Optional. The arguments to pass to the driver. Do not * include arguments, such as `-libjars` or `-Dfoo=bar`, that can be set as - * job properties, since a collision may occur that causes an incorrect job + * job properties, since a collision might occur that causes an incorrect job * submission. * * Generated from protobuf field repeated string args = 3 [(.google.api.field_behavior) = OPTIONAL]; @@ -202,7 +202,7 @@ public function getArgs() /** * Optional. The arguments to pass to the driver. Do not * include arguments, such as `-libjars` or `-Dfoo=bar`, that can be set as - * job properties, since a collision may occur that causes an incorrect job + * job properties, since a collision might occur that causes an incorrect job * submission. * * Generated from protobuf field repeated string args = 3 [(.google.api.field_behavior) = OPTIONAL]; @@ -307,7 +307,7 @@ public function setArchiveUris($var) /** * Optional. A mapping of property names to values, used to configure Hadoop. - * Properties that conflict with values set by the Dataproc API may be + * Properties that conflict with values set by the Dataproc API might be * overwritten. Can include properties set in `/etc/hadoop/conf/*-site` and * classes in user code. * @@ -321,7 +321,7 @@ public function getProperties() /** * Optional. A mapping of property names to values, used to configure Hadoop. - * Properties that conflict with values set by the Dataproc API may be + * Properties that conflict with values set by the Dataproc API might be * overwritten. Can include properties set in `/etc/hadoop/conf/*-site` and * classes in user code. * diff --git a/Dataproc/src/V1/HiveJob.php b/Dataproc/src/V1/HiveJob.php index 0cc4f8a2d888..e607c9431397 100644 --- a/Dataproc/src/V1/HiveJob.php +++ b/Dataproc/src/V1/HiveJob.php @@ -33,7 +33,7 @@ class HiveJob extends \Google\Protobuf\Internal\Message private $script_variables; /** * Optional. A mapping of property names and values, used to configure Hive. - * Properties that conflict with values set by the Dataproc API may be + * Properties that conflict with values set by the Dataproc API might be * overwritten. Can include properties set in `/etc/hadoop/conf/*-site.xml`, * /etc/hive/conf/hive-site.xml, and classes in user code. * @@ -69,7 +69,7 @@ class HiveJob extends \Google\Protobuf\Internal\Message * Hive command: `SET name="value";`). * @type array|\Google\Protobuf\Internal\MapField $properties * Optional. A mapping of property names and values, used to configure Hive. - * Properties that conflict with values set by the Dataproc API may be + * Properties that conflict with values set by the Dataproc API might be * overwritten. Can include properties set in `/etc/hadoop/conf/*-site.xml`, * /etc/hive/conf/hive-site.xml, and classes in user code. * @type array|\Google\Protobuf\Internal\RepeatedField $jar_file_uris @@ -205,7 +205,7 @@ public function setScriptVariables($var) /** * Optional. A mapping of property names and values, used to configure Hive. - * Properties that conflict with values set by the Dataproc API may be + * Properties that conflict with values set by the Dataproc API might be * overwritten. Can include properties set in `/etc/hadoop/conf/*-site.xml`, * /etc/hive/conf/hive-site.xml, and classes in user code. * @@ -219,7 +219,7 @@ public function getProperties() /** * Optional. A mapping of property names and values, used to configure Hive. - * Properties that conflict with values set by the Dataproc API may be + * Properties that conflict with values set by the Dataproc API might be * overwritten. Can include properties set in `/etc/hadoop/conf/*-site.xml`, * /etc/hive/conf/hive-site.xml, and classes in user code. * diff --git a/Dataproc/src/V1/Job.php b/Dataproc/src/V1/Job.php index a39dbbfe682b..e9ce2e8354f5 100644 --- a/Dataproc/src/V1/Job.php +++ b/Dataproc/src/V1/Job.php @@ -33,7 +33,7 @@ class Job extends \Google\Protobuf\Internal\Message protected $placement = null; /** * Output only. The job status. Additional application-specific - * status information may be contained in the type_job + * status information might be contained in the type_job * and yarn_applications fields. * * Generated from protobuf field .google.cloud.dataproc.v1.JobStatus status = 8 [(.google.api.field_behavior) = OUTPUT_ONLY]; @@ -48,7 +48,7 @@ class Job extends \Google\Protobuf\Internal\Message /** * Output only. The collection of YARN applications spun up by this job. * **Beta** Feature: This report is available for testing purposes only. It - * may be changed before final release. + * might be changed before final release. * * Generated from protobuf field repeated .google.cloud.dataproc.v1.YarnApplication yarn_applications = 9 [(.google.api.field_behavior) = OUTPUT_ONLY]; */ @@ -62,8 +62,8 @@ class Job extends \Google\Protobuf\Internal\Message protected $driver_output_resource_uri = ''; /** * Output only. If present, the location of miscellaneous control files - * which may be used as part of job setup and handling. If not present, - * control files may be placed in the same location as `driver_output_uri`. + * which can be used as part of job setup and handling. If not present, + * control files might be placed in the same location as `driver_output_uri`. * * Generated from protobuf field string driver_control_files_uri = 15 [(.google.api.field_behavior) = OUTPUT_ONLY]; */ @@ -72,7 +72,7 @@ class Job extends \Google\Protobuf\Internal\Message * Optional. The labels to associate with this job. * Label **keys** must contain 1 to 63 characters, and must conform to * [RFC 1035](https://www.ietf.org/rfc/rfc1035.txt). - * Label **values** may be empty, but, if present, must contain 1 to 63 + * Label **values** can be empty, but, if present, must contain 1 to 63 * characters, and must conform to [RFC * 1035](https://www.ietf.org/rfc/rfc1035.txt). No more than 32 labels can be * associated with a job. @@ -89,7 +89,7 @@ class Job extends \Google\Protobuf\Internal\Message /** * Output only. A UUID that uniquely identifies a job within the project * over time. This is in contrast to a user-settable reference.job_id that - * may be reused over time. + * might be reused over time. * * Generated from protobuf field string job_uuid = 22 [(.google.api.field_behavior) = OUTPUT_ONLY]; */ @@ -147,26 +147,26 @@ class Job extends \Google\Protobuf\Internal\Message * Optional. Job is a Flink job. * @type \Google\Cloud\Dataproc\V1\JobStatus $status * Output only. The job status. Additional application-specific - * status information may be contained in the type_job + * status information might be contained in the type_job * and yarn_applications fields. * @type array<\Google\Cloud\Dataproc\V1\JobStatus>|\Google\Protobuf\Internal\RepeatedField $status_history * Output only. The previous job status. * @type array<\Google\Cloud\Dataproc\V1\YarnApplication>|\Google\Protobuf\Internal\RepeatedField $yarn_applications * Output only. The collection of YARN applications spun up by this job. * **Beta** Feature: This report is available for testing purposes only. It - * may be changed before final release. + * might be changed before final release. * @type string $driver_output_resource_uri * Output only. A URI pointing to the location of the stdout of the job's * driver program. * @type string $driver_control_files_uri * Output only. If present, the location of miscellaneous control files - * which may be used as part of job setup and handling. If not present, - * control files may be placed in the same location as `driver_output_uri`. + * which can be used as part of job setup and handling. If not present, + * control files might be placed in the same location as `driver_output_uri`. * @type array|\Google\Protobuf\Internal\MapField $labels * Optional. The labels to associate with this job. * Label **keys** must contain 1 to 63 characters, and must conform to * [RFC 1035](https://www.ietf.org/rfc/rfc1035.txt). - * Label **values** may be empty, but, if present, must contain 1 to 63 + * Label **values** can be empty, but, if present, must contain 1 to 63 * characters, and must conform to [RFC * 1035](https://www.ietf.org/rfc/rfc1035.txt). No more than 32 labels can be * associated with a job. @@ -175,7 +175,7 @@ class Job extends \Google\Protobuf\Internal\Message * @type string $job_uuid * Output only. A UUID that uniquely identifies a job within the project * over time. This is in contrast to a user-settable reference.job_id that - * may be reused over time. + * might be reused over time. * @type bool $done * Output only. Indicates whether the job is completed. If the value is * `false`, the job is still in progress. If `true`, the job is completed, and @@ -582,7 +582,7 @@ public function setFlinkJob($var) /** * Output only. The job status. Additional application-specific - * status information may be contained in the type_job + * status information might be contained in the type_job * and yarn_applications fields. * * Generated from protobuf field .google.cloud.dataproc.v1.JobStatus status = 8 [(.google.api.field_behavior) = OUTPUT_ONLY]; @@ -605,7 +605,7 @@ public function clearStatus() /** * Output only. The job status. Additional application-specific - * status information may be contained in the type_job + * status information might be contained in the type_job * and yarn_applications fields. * * Generated from protobuf field .google.cloud.dataproc.v1.JobStatus status = 8 [(.google.api.field_behavior) = OUTPUT_ONLY]; @@ -649,7 +649,7 @@ public function setStatusHistory($var) /** * Output only. The collection of YARN applications spun up by this job. * **Beta** Feature: This report is available for testing purposes only. It - * may be changed before final release. + * might be changed before final release. * * Generated from protobuf field repeated .google.cloud.dataproc.v1.YarnApplication yarn_applications = 9 [(.google.api.field_behavior) = OUTPUT_ONLY]; * @return \Google\Protobuf\Internal\RepeatedField @@ -662,7 +662,7 @@ public function getYarnApplications() /** * Output only. The collection of YARN applications spun up by this job. * **Beta** Feature: This report is available for testing purposes only. It - * may be changed before final release. + * might be changed before final release. * * Generated from protobuf field repeated .google.cloud.dataproc.v1.YarnApplication yarn_applications = 9 [(.google.api.field_behavior) = OUTPUT_ONLY]; * @param array<\Google\Cloud\Dataproc\V1\YarnApplication>|\Google\Protobuf\Internal\RepeatedField $var @@ -706,8 +706,8 @@ public function setDriverOutputResourceUri($var) /** * Output only. If present, the location of miscellaneous control files - * which may be used as part of job setup and handling. If not present, - * control files may be placed in the same location as `driver_output_uri`. + * which can be used as part of job setup and handling. If not present, + * control files might be placed in the same location as `driver_output_uri`. * * Generated from protobuf field string driver_control_files_uri = 15 [(.google.api.field_behavior) = OUTPUT_ONLY]; * @return string @@ -719,8 +719,8 @@ public function getDriverControlFilesUri() /** * Output only. If present, the location of miscellaneous control files - * which may be used as part of job setup and handling. If not present, - * control files may be placed in the same location as `driver_output_uri`. + * which can be used as part of job setup and handling. If not present, + * control files might be placed in the same location as `driver_output_uri`. * * Generated from protobuf field string driver_control_files_uri = 15 [(.google.api.field_behavior) = OUTPUT_ONLY]; * @param string $var @@ -738,7 +738,7 @@ public function setDriverControlFilesUri($var) * Optional. The labels to associate with this job. * Label **keys** must contain 1 to 63 characters, and must conform to * [RFC 1035](https://www.ietf.org/rfc/rfc1035.txt). - * Label **values** may be empty, but, if present, must contain 1 to 63 + * Label **values** can be empty, but, if present, must contain 1 to 63 * characters, and must conform to [RFC * 1035](https://www.ietf.org/rfc/rfc1035.txt). No more than 32 labels can be * associated with a job. @@ -755,7 +755,7 @@ public function getLabels() * Optional. The labels to associate with this job. * Label **keys** must contain 1 to 63 characters, and must conform to * [RFC 1035](https://www.ietf.org/rfc/rfc1035.txt). - * Label **values** may be empty, but, if present, must contain 1 to 63 + * Label **values** can be empty, but, if present, must contain 1 to 63 * characters, and must conform to [RFC * 1035](https://www.ietf.org/rfc/rfc1035.txt). No more than 32 labels can be * associated with a job. @@ -811,7 +811,7 @@ public function setScheduling($var) /** * Output only. A UUID that uniquely identifies a job within the project * over time. This is in contrast to a user-settable reference.job_id that - * may be reused over time. + * might be reused over time. * * Generated from protobuf field string job_uuid = 22 [(.google.api.field_behavior) = OUTPUT_ONLY]; * @return string @@ -824,7 +824,7 @@ public function getJobUuid() /** * Output only. A UUID that uniquely identifies a job within the project * over time. This is in contrast to a user-settable reference.job_id that - * may be reused over time. + * might be reused over time. * * Generated from protobuf field string job_uuid = 22 [(.google.api.field_behavior) = OUTPUT_ONLY]; * @param string $var diff --git a/Dataproc/src/V1/JobScheduling.php b/Dataproc/src/V1/JobScheduling.php index 84563b4503ca..a7499e413e7e 100644 --- a/Dataproc/src/V1/JobScheduling.php +++ b/Dataproc/src/V1/JobScheduling.php @@ -16,11 +16,11 @@ class JobScheduling extends \Google\Protobuf\Internal\Message { /** - * Optional. Maximum number of times per hour a driver may be restarted as + * Optional. Maximum number of times per hour a driver can be restarted as * a result of driver exiting with non-zero code before job is * reported failed. - * A job may be reported as thrashing if the driver exits with a non-zero code - * four times within a 10-minute window. + * A job might be reported as thrashing if the driver exits with a non-zero + * code four times within a 10-minute window. * Maximum value is 10. * **Note:** This restartable job option is not supported in Dataproc * [workflow templates] @@ -30,7 +30,7 @@ class JobScheduling extends \Google\Protobuf\Internal\Message */ protected $max_failures_per_hour = 0; /** - * Optional. Maximum total number of times a driver may be restarted as a + * Optional. Maximum total number of times a driver can be restarted as a * result of the driver exiting with a non-zero code. After the maximum number * is reached, the job will be reported as failed. * Maximum value is 240. @@ -50,17 +50,17 @@ class JobScheduling extends \Google\Protobuf\Internal\Message * Optional. Data for populating the Message object. * * @type int $max_failures_per_hour - * Optional. Maximum number of times per hour a driver may be restarted as + * Optional. Maximum number of times per hour a driver can be restarted as * a result of driver exiting with non-zero code before job is * reported failed. - * A job may be reported as thrashing if the driver exits with a non-zero code - * four times within a 10-minute window. + * A job might be reported as thrashing if the driver exits with a non-zero + * code four times within a 10-minute window. * Maximum value is 10. * **Note:** This restartable job option is not supported in Dataproc * [workflow templates] * (https://cloud.google.com/dataproc/docs/concepts/workflows/using-workflows#adding_jobs_to_a_template). * @type int $max_failures_total - * Optional. Maximum total number of times a driver may be restarted as a + * Optional. Maximum total number of times a driver can be restarted as a * result of the driver exiting with a non-zero code. After the maximum number * is reached, the job will be reported as failed. * Maximum value is 240. @@ -76,11 +76,11 @@ public function __construct($data = NULL) { } /** - * Optional. Maximum number of times per hour a driver may be restarted as + * Optional. Maximum number of times per hour a driver can be restarted as * a result of driver exiting with non-zero code before job is * reported failed. - * A job may be reported as thrashing if the driver exits with a non-zero code - * four times within a 10-minute window. + * A job might be reported as thrashing if the driver exits with a non-zero + * code four times within a 10-minute window. * Maximum value is 10. * **Note:** This restartable job option is not supported in Dataproc * [workflow templates] @@ -95,11 +95,11 @@ public function getMaxFailuresPerHour() } /** - * Optional. Maximum number of times per hour a driver may be restarted as + * Optional. Maximum number of times per hour a driver can be restarted as * a result of driver exiting with non-zero code before job is * reported failed. - * A job may be reported as thrashing if the driver exits with a non-zero code - * four times within a 10-minute window. + * A job might be reported as thrashing if the driver exits with a non-zero + * code four times within a 10-minute window. * Maximum value is 10. * **Note:** This restartable job option is not supported in Dataproc * [workflow templates] @@ -118,7 +118,7 @@ public function setMaxFailuresPerHour($var) } /** - * Optional. Maximum total number of times a driver may be restarted as a + * Optional. Maximum total number of times a driver can be restarted as a * result of the driver exiting with a non-zero code. After the maximum number * is reached, the job will be reported as failed. * Maximum value is 240. @@ -136,7 +136,7 @@ public function getMaxFailuresTotal() } /** - * Optional. Maximum total number of times a driver may be restarted as a + * Optional. Maximum total number of times a driver can be restarted as a * result of the driver exiting with a non-zero code. After the maximum number * is reached, the job will be reported as failed. * Maximum value is 240. diff --git a/Dataproc/src/V1/JobStatus.php b/Dataproc/src/V1/JobStatus.php index a543d6cba10a..aaabdd201ad0 100644 --- a/Dataproc/src/V1/JobStatus.php +++ b/Dataproc/src/V1/JobStatus.php @@ -23,7 +23,7 @@ class JobStatus extends \Google\Protobuf\Internal\Message protected $state = 0; /** * Optional. Output only. Job state details, such as an error - * description if the state is ERROR. + * description if the state is `ERROR`. * * Generated from protobuf field string details = 2 [(.google.api.field_behavior) = OUTPUT_ONLY, (.google.api.field_behavior) = OPTIONAL]; */ @@ -52,7 +52,7 @@ class JobStatus extends \Google\Protobuf\Internal\Message * Output only. A state message specifying the overall job state. * @type string $details * Optional. Output only. Job state details, such as an error - * description if the state is ERROR. + * description if the state is `ERROR`. * @type \Google\Protobuf\Timestamp $state_start_time * Output only. The time when this state was entered. * @type int $substate @@ -93,7 +93,7 @@ public function setState($var) /** * Optional. Output only. Job state details, such as an error - * description if the state is ERROR. + * description if the state is `ERROR`. * * Generated from protobuf field string details = 2 [(.google.api.field_behavior) = OUTPUT_ONLY, (.google.api.field_behavior) = OPTIONAL]; * @return string @@ -105,7 +105,7 @@ public function getDetails() /** * Optional. Output only. Job state details, such as an error - * description if the state is ERROR. + * description if the state is `ERROR`. * * Generated from protobuf field string details = 2 [(.google.api.field_behavior) = OUTPUT_ONLY, (.google.api.field_behavior) = OPTIONAL]; * @param string $var diff --git a/Dataproc/src/V1/JobStatus/Substate.php b/Dataproc/src/V1/JobStatus/Substate.php index 83985a336960..80c582d55f89 100644 --- a/Dataproc/src/V1/JobStatus/Substate.php +++ b/Dataproc/src/V1/JobStatus/Substate.php @@ -27,7 +27,7 @@ class Substate */ const SUBMITTED = 1; /** - * The Job has been received and is awaiting execution (it may be waiting + * The Job has been received and is awaiting execution (it might be waiting * for a condition to be met). See the "details" field for the reason for * the delay. * Applies to RUNNING state. @@ -36,7 +36,7 @@ class Substate */ const QUEUED = 2; /** - * The agent-reported status is out of date, which may be caused by a + * The agent-reported status is out of date, which can be caused by a * loss of communication between the agent and Dataproc. If the * agent does not send a timely update, the job will fail. * Applies to RUNNING state. diff --git a/Dataproc/src/V1/KerberosConfig.php b/Dataproc/src/V1/KerberosConfig.php index af6293590870..2494c33cae09 100644 --- a/Dataproc/src/V1/KerberosConfig.php +++ b/Dataproc/src/V1/KerberosConfig.php @@ -30,7 +30,7 @@ class KerberosConfig extends \Google\Protobuf\Internal\Message */ protected $root_principal_password_uri = ''; /** - * Optional. The uri of the KMS key used to encrypt various sensitive + * Optional. The URI of the KMS key used to encrypt sensitive * files. * * Generated from protobuf field string kms_key_uri = 3 [(.google.api.field_behavior) = OPTIONAL]; @@ -141,7 +141,7 @@ class KerberosConfig extends \Google\Protobuf\Internal\Message * Optional. The Cloud Storage URI of a KMS encrypted file containing the root * principal password. * @type string $kms_key_uri - * Optional. The uri of the KMS key used to encrypt various sensitive + * Optional. The URI of the KMS key used to encrypt sensitive * files. * @type string $keystore_uri * Optional. The Cloud Storage URI of the keystore file used for SSL @@ -250,7 +250,7 @@ public function setRootPrincipalPasswordUri($var) } /** - * Optional. The uri of the KMS key used to encrypt various sensitive + * Optional. The URI of the KMS key used to encrypt sensitive * files. * * Generated from protobuf field string kms_key_uri = 3 [(.google.api.field_behavior) = OPTIONAL]; @@ -262,7 +262,7 @@ public function getKmsKeyUri() } /** - * Optional. The uri of the KMS key used to encrypt various sensitive + * Optional. The URI of the KMS key used to encrypt sensitive * files. * * Generated from protobuf field string kms_key_uri = 3 [(.google.api.field_behavior) = OPTIONAL]; diff --git a/Dataproc/src/V1/ListClustersRequest.php b/Dataproc/src/V1/ListClustersRequest.php index f5e24e9240c2..e09e47dd76cd 100644 --- a/Dataproc/src/V1/ListClustersRequest.php +++ b/Dataproc/src/V1/ListClustersRequest.php @@ -35,12 +35,12 @@ class ListClustersRequest extends \Google\Protobuf\Internal\Message * where **field** is one of `status.state`, `clusterName`, or `labels.[KEY]`, * and `[KEY]` is a label key. **value** can be `*` to match all values. * `status.state` can be one of the following: `ACTIVE`, `INACTIVE`, - * `CREATING`, `RUNNING`, `ERROR`, `DELETING`, or `UPDATING`. `ACTIVE` - * contains the `CREATING`, `UPDATING`, and `RUNNING` states. `INACTIVE` - * contains the `DELETING` and `ERROR` states. - * `clusterName` is the name of the cluster provided at creation time. - * Only the logical `AND` operator is supported; space-separated items are - * treated as having an implicit `AND` operator. + * `CREATING`, `RUNNING`, `ERROR`, `DELETING`, `UPDATING`, `STOPPING`, or + * `STOPPED`. `ACTIVE` contains the `CREATING`, `UPDATING`, and `RUNNING` + * states. `INACTIVE` contains the `DELETING`, `ERROR`, `STOPPING`, and + * `STOPPED` states. `clusterName` is the name of the cluster provided at + * creation time. Only the logical `AND` operator is supported; + * space-separated items are treated as having an implicit `AND` operator. * Example filter: * status.state = ACTIVE AND clusterName = mycluster * AND labels.env = staging AND labels.starred = * @@ -89,12 +89,12 @@ public static function build(string $projectId, string $region): self * where **field** is one of `status.state`, `clusterName`, or `labels.[KEY]`, * and `[KEY]` is a label key. **value** can be `*` to match all values. * `status.state` can be one of the following: `ACTIVE`, `INACTIVE`, - * `CREATING`, `RUNNING`, `ERROR`, `DELETING`, or `UPDATING`. `ACTIVE` - * contains the `CREATING`, `UPDATING`, and `RUNNING` states. `INACTIVE` - * contains the `DELETING` and `ERROR` states. - * `clusterName` is the name of the cluster provided at creation time. - * Only the logical `AND` operator is supported; space-separated items are - * treated as having an implicit `AND` operator. + * `CREATING`, `RUNNING`, `ERROR`, `DELETING`, `UPDATING`, `STOPPING`, or + * `STOPPED`. `ACTIVE` contains the `CREATING`, `UPDATING`, and `RUNNING` + * states. `INACTIVE` contains the `DELETING`, `ERROR`, `STOPPING`, and + * `STOPPED` states. `clusterName` is the name of the cluster provided at + * creation time. Only the logical `AND` operator is supported; + * space-separated items are treated as having an implicit `AND` operator. * * Example filter: * @@ -131,12 +131,12 @@ public static function buildFromProjectIdRegionFilter(string $projectId, string * where **field** is one of `status.state`, `clusterName`, or `labels.[KEY]`, * and `[KEY]` is a label key. **value** can be `*` to match all values. * `status.state` can be one of the following: `ACTIVE`, `INACTIVE`, - * `CREATING`, `RUNNING`, `ERROR`, `DELETING`, or `UPDATING`. `ACTIVE` - * contains the `CREATING`, `UPDATING`, and `RUNNING` states. `INACTIVE` - * contains the `DELETING` and `ERROR` states. - * `clusterName` is the name of the cluster provided at creation time. - * Only the logical `AND` operator is supported; space-separated items are - * treated as having an implicit `AND` operator. + * `CREATING`, `RUNNING`, `ERROR`, `DELETING`, `UPDATING`, `STOPPING`, or + * `STOPPED`. `ACTIVE` contains the `CREATING`, `UPDATING`, and `RUNNING` + * states. `INACTIVE` contains the `DELETING`, `ERROR`, `STOPPING`, and + * `STOPPED` states. `clusterName` is the name of the cluster provided at + * creation time. Only the logical `AND` operator is supported; + * space-separated items are treated as having an implicit `AND` operator. * Example filter: * status.state = ACTIVE AND clusterName = mycluster * AND labels.env = staging AND labels.starred = * @@ -212,12 +212,12 @@ public function setRegion($var) * where **field** is one of `status.state`, `clusterName`, or `labels.[KEY]`, * and `[KEY]` is a label key. **value** can be `*` to match all values. * `status.state` can be one of the following: `ACTIVE`, `INACTIVE`, - * `CREATING`, `RUNNING`, `ERROR`, `DELETING`, or `UPDATING`. `ACTIVE` - * contains the `CREATING`, `UPDATING`, and `RUNNING` states. `INACTIVE` - * contains the `DELETING` and `ERROR` states. - * `clusterName` is the name of the cluster provided at creation time. - * Only the logical `AND` operator is supported; space-separated items are - * treated as having an implicit `AND` operator. + * `CREATING`, `RUNNING`, `ERROR`, `DELETING`, `UPDATING`, `STOPPING`, or + * `STOPPED`. `ACTIVE` contains the `CREATING`, `UPDATING`, and `RUNNING` + * states. `INACTIVE` contains the `DELETING`, `ERROR`, `STOPPING`, and + * `STOPPED` states. `clusterName` is the name of the cluster provided at + * creation time. Only the logical `AND` operator is supported; + * space-separated items are treated as having an implicit `AND` operator. * Example filter: * status.state = ACTIVE AND clusterName = mycluster * AND labels.env = staging AND labels.starred = * @@ -237,12 +237,12 @@ public function getFilter() * where **field** is one of `status.state`, `clusterName`, or `labels.[KEY]`, * and `[KEY]` is a label key. **value** can be `*` to match all values. * `status.state` can be one of the following: `ACTIVE`, `INACTIVE`, - * `CREATING`, `RUNNING`, `ERROR`, `DELETING`, or `UPDATING`. `ACTIVE` - * contains the `CREATING`, `UPDATING`, and `RUNNING` states. `INACTIVE` - * contains the `DELETING` and `ERROR` states. - * `clusterName` is the name of the cluster provided at creation time. - * Only the logical `AND` operator is supported; space-separated items are - * treated as having an implicit `AND` operator. + * `CREATING`, `RUNNING`, `ERROR`, `DELETING`, `UPDATING`, `STOPPING`, or + * `STOPPED`. `ACTIVE` contains the `CREATING`, `UPDATING`, and `RUNNING` + * states. `INACTIVE` contains the `DELETING`, `ERROR`, `STOPPING`, and + * `STOPPED` states. `clusterName` is the name of the cluster provided at + * creation time. Only the logical `AND` operator is supported; + * space-separated items are treated as having an implicit `AND` operator. * Example filter: * status.state = ACTIVE AND clusterName = mycluster * AND labels.env = staging AND labels.starred = * diff --git a/Dataproc/src/V1/LoggingConfig.php b/Dataproc/src/V1/LoggingConfig.php index fc0e3ce7dd92..234b9ee24744 100644 --- a/Dataproc/src/V1/LoggingConfig.php +++ b/Dataproc/src/V1/LoggingConfig.php @@ -16,7 +16,7 @@ class LoggingConfig extends \Google\Protobuf\Internal\Message { /** - * The per-package log levels for the driver. This may include + * The per-package log levels for the driver. This can include * "root" package name to configure rootLogger. * Examples: * - 'com.google = FATAL' @@ -34,7 +34,7 @@ class LoggingConfig extends \Google\Protobuf\Internal\Message * Optional. Data for populating the Message object. * * @type array|\Google\Protobuf\Internal\MapField $driver_log_levels - * The per-package log levels for the driver. This may include + * The per-package log levels for the driver. This can include * "root" package name to configure rootLogger. * Examples: * - 'com.google = FATAL' @@ -48,7 +48,7 @@ public function __construct($data = NULL) { } /** - * The per-package log levels for the driver. This may include + * The per-package log levels for the driver. This can include * "root" package name to configure rootLogger. * Examples: * - 'com.google = FATAL' @@ -64,7 +64,7 @@ public function getDriverLogLevels() } /** - * The per-package log levels for the driver. This may include + * The per-package log levels for the driver. This can include * "root" package name to configure rootLogger. * Examples: * - 'com.google = FATAL' diff --git a/Dataproc/src/V1/PigJob.php b/Dataproc/src/V1/PigJob.php index 91bf4c7fb122..e8b4115ec40a 100644 --- a/Dataproc/src/V1/PigJob.php +++ b/Dataproc/src/V1/PigJob.php @@ -33,7 +33,7 @@ class PigJob extends \Google\Protobuf\Internal\Message private $script_variables; /** * Optional. A mapping of property names to values, used to configure Pig. - * Properties that conflict with values set by the Dataproc API may be + * Properties that conflict with values set by the Dataproc API might be * overwritten. Can include properties set in `/etc/hadoop/conf/*-site.xml`, * /etc/pig/conf/pig.properties, and classes in user code. * @@ -74,7 +74,7 @@ class PigJob extends \Google\Protobuf\Internal\Message * command: `name=[value]`). * @type array|\Google\Protobuf\Internal\MapField $properties * Optional. A mapping of property names to values, used to configure Pig. - * Properties that conflict with values set by the Dataproc API may be + * Properties that conflict with values set by the Dataproc API might be * overwritten. Can include properties set in `/etc/hadoop/conf/*-site.xml`, * /etc/pig/conf/pig.properties, and classes in user code. * @type array|\Google\Protobuf\Internal\RepeatedField $jar_file_uris @@ -211,7 +211,7 @@ public function setScriptVariables($var) /** * Optional. A mapping of property names to values, used to configure Pig. - * Properties that conflict with values set by the Dataproc API may be + * Properties that conflict with values set by the Dataproc API might be * overwritten. Can include properties set in `/etc/hadoop/conf/*-site.xml`, * /etc/pig/conf/pig.properties, and classes in user code. * @@ -225,7 +225,7 @@ public function getProperties() /** * Optional. A mapping of property names to values, used to configure Pig. - * Properties that conflict with values set by the Dataproc API may be + * Properties that conflict with values set by the Dataproc API might be * overwritten. Can include properties set in `/etc/hadoop/conf/*-site.xml`, * /etc/pig/conf/pig.properties, and classes in user code. * diff --git a/Dataproc/src/V1/PySparkJob.php b/Dataproc/src/V1/PySparkJob.php index 12c7dfd00b35..f4572edb2ddb 100644 --- a/Dataproc/src/V1/PySparkJob.php +++ b/Dataproc/src/V1/PySparkJob.php @@ -64,7 +64,7 @@ class PySparkJob extends \Google\Protobuf\Internal\Message private $archive_uris; /** * Optional. A mapping of property names to values, used to configure PySpark. - * Properties that conflict with values set by the Dataproc API may be + * Properties that conflict with values set by the Dataproc API might be * overwritten. Can include properties set in * /etc/spark/conf/spark-defaults.conf and classes in user code. * @@ -106,7 +106,7 @@ class PySparkJob extends \Google\Protobuf\Internal\Message * .jar, .tar, .tar.gz, .tgz, and .zip. * @type array|\Google\Protobuf\Internal\MapField $properties * Optional. A mapping of property names to values, used to configure PySpark. - * Properties that conflict with values set by the Dataproc API may be + * Properties that conflict with values set by the Dataproc API might be * overwritten. Can include properties set in * /etc/spark/conf/spark-defaults.conf and classes in user code. * @type \Google\Cloud\Dataproc\V1\LoggingConfig $logging_config @@ -292,7 +292,7 @@ public function setArchiveUris($var) /** * Optional. A mapping of property names to values, used to configure PySpark. - * Properties that conflict with values set by the Dataproc API may be + * Properties that conflict with values set by the Dataproc API might be * overwritten. Can include properties set in * /etc/spark/conf/spark-defaults.conf and classes in user code. * @@ -306,7 +306,7 @@ public function getProperties() /** * Optional. A mapping of property names to values, used to configure PySpark. - * Properties that conflict with values set by the Dataproc API may be + * Properties that conflict with values set by the Dataproc API might be * overwritten. Can include properties set in * /etc/spark/conf/spark-defaults.conf and classes in user code. * diff --git a/Dataproc/src/V1/SoftwareConfig.php b/Dataproc/src/V1/SoftwareConfig.php index a319987b0f92..a8aa8664a399 100644 --- a/Dataproc/src/V1/SoftwareConfig.php +++ b/Dataproc/src/V1/SoftwareConfig.php @@ -18,7 +18,7 @@ class SoftwareConfig extends \Google\Protobuf\Internal\Message /** * Optional. The version of software inside the cluster. It must be one of the * supported [Dataproc - * Versions](https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#supported_dataproc_versions), + * Versions](https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#supported-dataproc-image-versions), * such as "1.2" (including a subminor version, such as "1.2.29"), or the * ["preview" * version](https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#other_versions). @@ -63,7 +63,7 @@ class SoftwareConfig extends \Google\Protobuf\Internal\Message * @type string $image_version * Optional. The version of software inside the cluster. It must be one of the * supported [Dataproc - * Versions](https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#supported_dataproc_versions), + * Versions](https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#supported-dataproc-image-versions), * such as "1.2" (including a subminor version, such as "1.2.29"), or the * ["preview" * version](https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#other_versions). @@ -96,7 +96,7 @@ public function __construct($data = NULL) { /** * Optional. The version of software inside the cluster. It must be one of the * supported [Dataproc - * Versions](https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#supported_dataproc_versions), + * Versions](https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#supported-dataproc-image-versions), * such as "1.2" (including a subminor version, such as "1.2.29"), or the * ["preview" * version](https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#other_versions). @@ -113,7 +113,7 @@ public function getImageVersion() /** * Optional. The version of software inside the cluster. It must be one of the * supported [Dataproc - * Versions](https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#supported_dataproc_versions), + * Versions](https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#supported-dataproc-image-versions), * such as "1.2" (including a subminor version, such as "1.2.29"), or the * ["preview" * version](https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#other_versions). diff --git a/Dataproc/src/V1/SparkJob.php b/Dataproc/src/V1/SparkJob.php index 64487c6910b2..f80b2d022e0d 100644 --- a/Dataproc/src/V1/SparkJob.php +++ b/Dataproc/src/V1/SparkJob.php @@ -48,7 +48,7 @@ class SparkJob extends \Google\Protobuf\Internal\Message private $archive_uris; /** * Optional. A mapping of property names to values, used to configure Spark. - * Properties that conflict with values set by the Dataproc API may be + * Properties that conflict with values set by the Dataproc API might be * overwritten. Can include properties set in * /etc/spark/conf/spark-defaults.conf and classes in user code. * @@ -73,7 +73,8 @@ class SparkJob extends \Google\Protobuf\Internal\Message * The HCFS URI of the jar file that contains the main class. * @type string $main_class * The name of the driver's main class. The jar file that contains the class - * must be in the default CLASSPATH or specified in `jar_file_uris`. + * must be in the default CLASSPATH or specified in + * SparkJob.jar_file_uris. * @type array|\Google\Protobuf\Internal\RepeatedField $args * Optional. The arguments to pass to the driver. Do not include arguments, * such as `--conf`, that can be set as job properties, since a collision may @@ -90,7 +91,7 @@ class SparkJob extends \Google\Protobuf\Internal\Message * .jar, .tar, .tar.gz, .tgz, and .zip. * @type array|\Google\Protobuf\Internal\MapField $properties * Optional. A mapping of property names to values, used to configure Spark. - * Properties that conflict with values set by the Dataproc API may be + * Properties that conflict with values set by the Dataproc API might be * overwritten. Can include properties set in * /etc/spark/conf/spark-defaults.conf and classes in user code. * @type \Google\Cloud\Dataproc\V1\LoggingConfig $logging_config @@ -135,7 +136,8 @@ public function setMainJarFileUri($var) /** * The name of the driver's main class. The jar file that contains the class - * must be in the default CLASSPATH or specified in `jar_file_uris`. + * must be in the default CLASSPATH or specified in + * SparkJob.jar_file_uris. * * Generated from protobuf field string main_class = 2; * @return string @@ -152,7 +154,8 @@ public function hasMainClass() /** * The name of the driver's main class. The jar file that contains the class - * must be in the default CLASSPATH or specified in `jar_file_uris`. + * must be in the default CLASSPATH or specified in + * SparkJob.jar_file_uris. * * Generated from protobuf field string main_class = 2; * @param string $var @@ -284,7 +287,7 @@ public function setArchiveUris($var) /** * Optional. A mapping of property names to values, used to configure Spark. - * Properties that conflict with values set by the Dataproc API may be + * Properties that conflict with values set by the Dataproc API might be * overwritten. Can include properties set in * /etc/spark/conf/spark-defaults.conf and classes in user code. * @@ -298,7 +301,7 @@ public function getProperties() /** * Optional. A mapping of property names to values, used to configure Spark. - * Properties that conflict with values set by the Dataproc API may be + * Properties that conflict with values set by the Dataproc API might be * overwritten. Can include properties set in * /etc/spark/conf/spark-defaults.conf and classes in user code. * diff --git a/Dataproc/src/V1/SparkRJob.php b/Dataproc/src/V1/SparkRJob.php index 702f31a6b658..f1f76ab1504d 100644 --- a/Dataproc/src/V1/SparkRJob.php +++ b/Dataproc/src/V1/SparkRJob.php @@ -49,7 +49,7 @@ class SparkRJob extends \Google\Protobuf\Internal\Message private $archive_uris; /** * Optional. A mapping of property names to values, used to configure SparkR. - * Properties that conflict with values set by the Dataproc API may be + * Properties that conflict with values set by the Dataproc API might be * overwritten. Can include properties set in * /etc/spark/conf/spark-defaults.conf and classes in user code. * @@ -85,7 +85,7 @@ class SparkRJob extends \Google\Protobuf\Internal\Message * .jar, .tar, .tar.gz, .tgz, and .zip. * @type array|\Google\Protobuf\Internal\MapField $properties * Optional. A mapping of property names to values, used to configure SparkR. - * Properties that conflict with values set by the Dataproc API may be + * Properties that conflict with values set by the Dataproc API might be * overwritten. Can include properties set in * /etc/spark/conf/spark-defaults.conf and classes in user code. * @type \Google\Cloud\Dataproc\V1\LoggingConfig $logging_config @@ -215,7 +215,7 @@ public function setArchiveUris($var) /** * Optional. A mapping of property names to values, used to configure SparkR. - * Properties that conflict with values set by the Dataproc API may be + * Properties that conflict with values set by the Dataproc API might be * overwritten. Can include properties set in * /etc/spark/conf/spark-defaults.conf and classes in user code. * @@ -229,7 +229,7 @@ public function getProperties() /** * Optional. A mapping of property names to values, used to configure SparkR. - * Properties that conflict with values set by the Dataproc API may be + * Properties that conflict with values set by the Dataproc API might be * overwritten. Can include properties set in * /etc/spark/conf/spark-defaults.conf and classes in user code. * diff --git a/Dataproc/src/V1/SparkSqlJob.php b/Dataproc/src/V1/SparkSqlJob.php index 1945abd33ad0..871a49648475 100644 --- a/Dataproc/src/V1/SparkSqlJob.php +++ b/Dataproc/src/V1/SparkSqlJob.php @@ -26,7 +26,7 @@ class SparkSqlJob extends \Google\Protobuf\Internal\Message /** * Optional. A mapping of property names to values, used to configure * Spark SQL's SparkConf. Properties that conflict with values set by the - * Dataproc API may be overwritten. + * Dataproc API might be overwritten. * * Generated from protobuf field map properties = 4 [(.google.api.field_behavior) = OPTIONAL]; */ @@ -61,7 +61,7 @@ class SparkSqlJob extends \Google\Protobuf\Internal\Message * @type array|\Google\Protobuf\Internal\MapField $properties * Optional. A mapping of property names to values, used to configure * Spark SQL's SparkConf. Properties that conflict with values set by the - * Dataproc API may be overwritten. + * Dataproc API might be overwritten. * @type array|\Google\Protobuf\Internal\RepeatedField $jar_file_uris * Optional. HCFS URIs of jar files to be added to the Spark CLASSPATH. * @type \Google\Cloud\Dataproc\V1\LoggingConfig $logging_config @@ -166,7 +166,7 @@ public function setScriptVariables($var) /** * Optional. A mapping of property names to values, used to configure * Spark SQL's SparkConf. Properties that conflict with values set by the - * Dataproc API may be overwritten. + * Dataproc API might be overwritten. * * Generated from protobuf field map properties = 4 [(.google.api.field_behavior) = OPTIONAL]; * @return \Google\Protobuf\Internal\MapField @@ -179,7 +179,7 @@ public function getProperties() /** * Optional. A mapping of property names to values, used to configure * Spark SQL's SparkConf. Properties that conflict with values set by the - * Dataproc API may be overwritten. + * Dataproc API might be overwritten. * * Generated from protobuf field map properties = 4 [(.google.api.field_behavior) = OPTIONAL]; * @param array|\Google\Protobuf\Internal\MapField $var diff --git a/Dataproc/src/V1/WorkflowTemplate/EncryptionConfig.php b/Dataproc/src/V1/WorkflowTemplate/EncryptionConfig.php index a1bcea7f018b..08fe57dbb5e7 100644 --- a/Dataproc/src/V1/WorkflowTemplate/EncryptionConfig.php +++ b/Dataproc/src/V1/WorkflowTemplate/EncryptionConfig.php @@ -43,7 +43,7 @@ class EncryptionConfig extends \Google\Protobuf\Internal\Message * * [PrestoJob](https://cloud.google.com/dataproc/docs/reference/rest/v1/PrestoJob) * scriptVariables and queryList.queries * - * Generated from protobuf field string kms_key = 1 [(.google.api.field_behavior) = OPTIONAL]; + * Generated from protobuf field string kms_key = 1 [(.google.api.field_behavior) = OPTIONAL, (.google.api.resource_reference) = { */ protected $kms_key = ''; @@ -115,7 +115,7 @@ public function __construct($data = NULL) { * * [PrestoJob](https://cloud.google.com/dataproc/docs/reference/rest/v1/PrestoJob) * scriptVariables and queryList.queries * - * Generated from protobuf field string kms_key = 1 [(.google.api.field_behavior) = OPTIONAL]; + * Generated from protobuf field string kms_key = 1 [(.google.api.field_behavior) = OPTIONAL, (.google.api.resource_reference) = { * @return string */ public function getKmsKey() @@ -151,7 +151,7 @@ public function getKmsKey() * * [PrestoJob](https://cloud.google.com/dataproc/docs/reference/rest/v1/PrestoJob) * scriptVariables and queryList.queries * - * Generated from protobuf field string kms_key = 1 [(.google.api.field_behavior) = OPTIONAL]; + * Generated from protobuf field string kms_key = 1 [(.google.api.field_behavior) = OPTIONAL, (.google.api.resource_reference) = { * @param string $var * @return $this */ diff --git a/Dataproc/src/V1/resources/cluster_controller_descriptor_config.php b/Dataproc/src/V1/resources/cluster_controller_descriptor_config.php index 4e2d3eeb3557..0f22e5235cd4 100644 --- a/Dataproc/src/V1/resources/cluster_controller_descriptor_config.php +++ b/Dataproc/src/V1/resources/cluster_controller_descriptor_config.php @@ -295,6 +295,7 @@ 'templateMap' => [ 'cluster' => 'projects/{project}/locations/{location}/clusters/{cluster}', 'clusterRegion' => 'projects/{project}/regions/{region}/clusters/{cluster}', + 'cryptoKey' => 'projects/{project}/locations/{location}/keyRings/{key_ring}/cryptoKeys/{crypto_key}', 'nodeGroup' => 'projects/{project}/regions/{region}/clusters/{cluster}/nodeGroups/{node_group}', 'service' => 'projects/{project}/locations/{location}/services/{service}', ], diff --git a/Dataproc/src/V1/resources/workflow_template_service_descriptor_config.php b/Dataproc/src/V1/resources/workflow_template_service_descriptor_config.php index dc59aae5180f..af73a590df7a 100644 --- a/Dataproc/src/V1/resources/workflow_template_service_descriptor_config.php +++ b/Dataproc/src/V1/resources/workflow_template_service_descriptor_config.php @@ -171,6 +171,7 @@ ], 'templateMap' => [ 'clusterRegion' => 'projects/{project}/regions/{region}/clusters/{cluster}', + 'cryptoKey' => 'projects/{project}/locations/{location}/keyRings/{key_ring}/cryptoKeys/{crypto_key}', 'location' => 'projects/{project}/locations/{location}', 'nodeGroup' => 'projects/{project}/regions/{region}/clusters/{cluster}/nodeGroups/{node_group}', 'projectLocationWorkflowTemplate' => 'projects/{project}/locations/{location}/workflowTemplates/{workflow_template}',