---
title: Getting Started with Datadog
description: Datadog, the leading service for cloud-scale monitoring.
breadcrumbs: Docs > Infrastructure > Datadog Resource Catalog
---

# gcp_kubernetes_engine_node_pool{% #gcp_kubernetes_engine_node_pool %}

## `ancestors`{% #ancestors %}

**Type**: `UNORDERED_LIST_STRING`

## `autoscaling`{% #autoscaling %}

**Type**: `STRUCT`**Provider name**: `autoscaling`**Description**: Autoscaler configuration for this NodePool. Autoscaler is enabled only if a valid configuration is present.

- `autoprovisioned`**Type**: `BOOLEAN`**Provider name**: `autoprovisioned`**Description**: Can this node pool be deleted automatically.
- `enabled`**Type**: `BOOLEAN`**Provider name**: `enabled`**Description**: Is autoscaling enabled for this node pool.
- `location_policy`**Type**: `STRING`**Provider name**: `locationPolicy`**Description**: Location policy used when scaling up a nodepool.**Possible values**:
  - `LOCATION_POLICY_UNSPECIFIED` - Not set.
  - `BALANCED` - BALANCED is a best effort policy that aims to balance the sizes of different zones.
  - `ANY` - ANY policy picks zones that have the highest capacity available.
- `max_node_count`**Type**: `INT32`**Provider name**: `maxNodeCount`**Description**: Maximum number of nodes for one location in the NodePool. Must be >= min_node_count. There has to be enough quota to scale up the cluster.
- `min_node_count`**Type**: `INT32`**Provider name**: `minNodeCount`**Description**: Minimum number of nodes for one location in the NodePool. Must be >= 1 and <= max_node_count.
- `total_max_node_count`**Type**: `INT32`**Provider name**: `totalMaxNodeCount`**Description**: Maximum number of nodes in the node pool. Must be greater than total_min_node_count. There has to be enough quota to scale up the cluster. The total_*_node_count fields are mutually exclusive with the *_node_count fields.
- `total_min_node_count`**Type**: `INT32`**Provider name**: `totalMinNodeCount`**Description**: Minimum number of nodes in the node pool. Must be greater than 1 less than total_max_node_count. The total_*_node_count fields are mutually exclusive with the *_node_count fields.

## `best_effort_provisioning`{% #best_effort_provisioning %}

**Type**: `STRUCT`**Provider name**: `bestEffortProvisioning`**Description**: Enable best effort provisioning for nodes

- `enabled`**Type**: `BOOLEAN`**Provider name**: `enabled`**Description**: When this is enabled, cluster/node pool creations will ignore non-fatal errors like stockout to best provision as many nodes as possible right now and eventually bring up all target number of nodes
- `min_provision_nodes`**Type**: `INT32`**Provider name**: `minProvisionNodes`**Description**: Minimum number of nodes to be provisioned to be considered as succeeded, and the rest of nodes will be provisioned gradually and eventually when stockout issue has been resolved.

## `conditions`{% #conditions %}

**Type**: `UNORDERED_LIST_STRUCT`**Provider name**: `conditions`**Description**: Which conditions caused the current node pool state.

- `canonical_code`**Type**: `STRING`**Provider name**: `canonicalCode`**Description**: Canonical code of the condition.**Possible values**:
  - `OK` - Not an error; returned on success. HTTP Mapping: 200 OK
  - `CANCELLED` - The operation was cancelled, typically by the caller. HTTP Mapping: 499 Client Closed Request
  - `UNKNOWN` - Unknown error. For example, this error may be returned when a `Status` value received from another address space belongs to an error space that is not known in this address space. Also errors raised by APIs that do not return enough error information may be converted to this error. HTTP Mapping: 500 Internal Server Error
  - `INVALID_ARGUMENT` - The client specified an invalid argument. Note that this differs from `FAILED_PRECONDITION`. `INVALID_ARGUMENT` indicates arguments that are problematic regardless of the state of the system (e.g., a malformed file name). HTTP Mapping: 400 Bad Request
  - `DEADLINE_EXCEEDED` - The deadline expired before the operation could complete. For operations that change the state of the system, this error may be returned even if the operation has completed successfully. For example, a successful response from a server could have been delayed long enough for the deadline to expire. HTTP Mapping: 504 Gateway Timeout
  - `NOT_FOUND` - Some requested entity (e.g., file or directory) was not found. Note to server developers: if a request is denied for an entire class of users, such as gradual feature rollout or undocumented allowlist, `NOT_FOUND` may be used. If a request is denied for some users within a class of users, such as user-based access control, `PERMISSION_DENIED` must be used. HTTP Mapping: 404 Not Found
  - `ALREADY_EXISTS` - The entity that a client attempted to create (e.g., file or directory) already exists. HTTP Mapping: 409 Conflict
  - `PERMISSION_DENIED` - The caller does not have permission to execute the specified operation. `PERMISSION_DENIED` must not be used for rejections caused by exhausting some resource (use `RESOURCE_EXHAUSTED` instead for those errors). `PERMISSION_DENIED` must not be used if the caller can not be identified (use `UNAUTHENTICATED` instead for those errors). This error code does not imply the request is valid or the requested entity exists or satisfies other pre-conditions. HTTP Mapping: 403 Forbidden
  - `UNAUTHENTICATED` - The request does not have valid authentication credentials for the operation. HTTP Mapping: 401 Unauthorized
  - `RESOURCE_EXHAUSTED` - Some resource has been exhausted, perhaps a per-user quota, or perhaps the entire file system is out of space. HTTP Mapping: 429 Too Many Requests
  - `FAILED_PRECONDITION` - The operation was rejected because the system is not in a state required for the operation's execution. For example, the directory to be deleted is non-empty, an rmdir operation is applied to a non-directory, etc. Service implementors can use the following guidelines to decide between `FAILED_PRECONDITION`, `ABORTED`, and `UNAVAILABLE`: (a) Use `UNAVAILABLE` if the client can retry just the failing call. (b) Use `ABORTED` if the client should retry at a higher level. For example, when a client-specified test-and-set fails, indicating the client should restart a read-modify-write sequence. (c) Use `FAILED_PRECONDITION` if the client should not retry until the system state has been explicitly fixed. For example, if an 'rmdir' fails because the directory is non-empty, `FAILED_PRECONDITION` should be returned since the client should not retry unless the files are deleted from the directory. HTTP Mapping: 400 Bad Request
  - `ABORTED` - The operation was aborted, typically due to a concurrency issue such as a sequencer check failure or transaction abort. See the guidelines above for deciding between `FAILED_PRECONDITION`, `ABORTED`, and `UNAVAILABLE`. HTTP Mapping: 409 Conflict
  - `OUT_OF_RANGE` - The operation was attempted past the valid range. E.g., seeking or reading past end-of-file. Unlike `INVALID_ARGUMENT`, this error indicates a problem that may be fixed if the system state changes. For example, a 32-bit file system will generate `INVALID_ARGUMENT` if asked to read at an offset that is not in the range [0,2^32-1], but it will generate `OUT_OF_RANGE` if asked to read from an offset past the current file size. There is a fair bit of overlap between `FAILED_PRECONDITION` and `OUT_OF_RANGE`. We recommend using `OUT_OF_RANGE` (the more specific error) when it applies so that callers who are iterating through a space can easily look for an `OUT_OF_RANGE` error to detect when they are done. HTTP Mapping: 400 Bad Request
  - `UNIMPLEMENTED` - The operation is not implemented or is not supported/enabled in this service. HTTP Mapping: 501 Not Implemented
  - `INTERNAL` - Internal errors. This means that some invariants expected by the underlying system have been broken. This error code is reserved for serious errors. HTTP Mapping: 500 Internal Server Error
  - `UNAVAILABLE` - The service is currently unavailable. This is most likely a transient condition, which can be corrected by retrying with a backoff. Note that it is not always safe to retry non-idempotent operations. See the guidelines above for deciding between `FAILED_PRECONDITION`, `ABORTED`, and `UNAVAILABLE`. HTTP Mapping: 503 Service Unavailable
  - `DATA_LOSS` - Unrecoverable data loss or corruption. HTTP Mapping: 500 Internal Server Error
- `code`**Type**: `STRING`**Provider name**: `code`**Description**: Machine-friendly representation of the condition Deprecated. Use canonical_code instead.**Possible values**:
  - `UNKNOWN` - UNKNOWN indicates a generic condition.
  - `GCE_STOCKOUT` - GCE_STOCKOUT indicates that Google Compute Engine resources are temporarily unavailable.
  - `GKE_SERVICE_ACCOUNT_DELETED` - GKE_SERVICE_ACCOUNT_DELETED indicates that the user deleted their robot service account.
  - `GCE_QUOTA_EXCEEDED` - Google Compute Engine quota was exceeded.
  - `SET_BY_OPERATOR` - Cluster state was manually changed by an SRE due to a system logic error.
  - `CLOUD_KMS_KEY_ERROR` - Unable to perform an encrypt operation against the CloudKMS key used for etcd level encryption.
  - `CA_EXPIRING` - Cluster CA is expiring soon. More codes TBA
- `message`**Type**: `STRING`**Provider name**: `message`**Description**: Human-friendly representation of the condition

## `config`{% #config %}

**Type**: `STRUCT`**Provider name**: `config`**Description**: The node configuration of the pool.

- `accelerators`**Type**: `UNORDERED_LIST_STRUCT`**Provider name**: `accelerators`**Description**: A list of hardware accelerators to be attached to each node. See [https://cloud.google.com/compute/docs/gpus](https://cloud.google.com/compute/docs/gpus) for more information about support for GPUs.

  - `accelerator_count`**Type**: `INT64`**Provider name**: `acceleratorCount`**Description**: The number of the accelerator cards exposed to an instance.
  - `accelerator_type`**Type**: `STRING`**Provider name**: `acceleratorType`**Description**: The accelerator type resource name. List of supported accelerators [here](https://cloud.google.com/compute/docs/gpus)
  - `gpu_driver_installation_config`**Type**: `STRUCT`**Provider name**: `gpuDriverInstallationConfig`**Description**: The configuration for auto installation of GPU driver.
    - `gpu_driver_version`**Type**: `STRING`**Provider name**: `gpuDriverVersion`**Description**: Mode for how the GPU driver is installed.**Possible values**:
      - `GPU_DRIVER_VERSION_UNSPECIFIED` - Default value is to not install any GPU driver.
      - `INSTALLATION_DISABLED` - Disable GPU driver auto installation and needs manual installation
      - `DEFAULT` - Default' GPU driver in COS and Ubuntu.
      - `LATEST` - Latest' GPU driver in COS.
  - `gpu_partition_size`**Type**: `STRING`**Provider name**: `gpuPartitionSize`**Description**: Size of partitions to create on the GPU. Valid values are described in the NVIDIA [mig user guide](https://docs.nvidia.com/datacenter/tesla/mig-user-guide/#partitioning).
  - `gpu_sharing_config`**Type**: `STRUCT`**Provider name**: `gpuSharingConfig`**Description**: The configuration for GPU sharing options.
    - `gpu_sharing_strategy`**Type**: `STRING`**Provider name**: `gpuSharingStrategy`**Description**: The type of GPU sharing strategy to enable on the GPU node.**Possible values**:
      - `GPU_SHARING_STRATEGY_UNSPECIFIED` - Default value.
      - `TIME_SHARING` - GPUs are time-shared between containers.
      - `MPS` - GPUs are shared between containers with NVIDIA MPS.
    - `max_shared_clients_per_gpu`**Type**: `INT64`**Provider name**: `maxSharedClientsPerGpu`**Description**: The max number of containers that can share a physical GPU.
  - `max_time_shared_clients_per_gpu`**Type**: `INT64`**Provider name**: `maxTimeSharedClientsPerGpu`**Description**: The number of time-shared GPU resources to expose for each physical GPU.

- `advanced_machine_features`**Type**: `STRUCT`**Provider name**: `advancedMachineFeatures`**Description**: Advanced features for the Compute Engine VM.

  - `enable_nested_virtualization`**Type**: `BOOLEAN`**Provider name**: `enableNestedVirtualization`**Description**: Whether or not to enable nested virtualization (defaults to false).
  - `threads_per_core`**Type**: `INT64`**Provider name**: `threadsPerCore`**Description**: The number of threads per physical core. To disable simultaneous multithreading (SMT) set this to 1. If unset, the maximum number of threads supported per core by the underlying processor is assumed.

- `boot_disk_kms_key`**Type**: `STRING`**Provider name**: `bootDiskKmsKey`**Description**: The Customer Managed Encryption Key used to encrypt the boot disk attached to each node in the node pool. This should be of the form projects/[KEY_PROJECT_ID]/locations/[LOCATION]/keyRings/[RING_NAME]/cryptoKeys/[KEY_NAME]. For more information about protecting resources with Cloud KMS Keys please see: [https://cloud.google.com/compute/docs/disks/customer-managed-encryption](https://cloud.google.com/compute/docs/disks/customer-managed-encryption)

- `confidential_nodes`**Type**: `STRUCT`**Provider name**: `confidentialNodes`**Description**: Confidential nodes config. All the nodes in the node pool will be Confidential VM once enabled.

  - `enabled`**Type**: `BOOLEAN`**Provider name**: `enabled`**Description**: Whether Confidential Nodes feature is enabled.

- `containerd_config`**Type**: `STRUCT`**Provider name**: `containerdConfig`**Description**: Parameters for containerd customization.

  - `private_registry_access_config`**Type**: `STRUCT`**Provider name**: `privateRegistryAccessConfig`**Description**: PrivateRegistryAccessConfig is used to configure access configuration for private container registries.
    - `certificate_authority_domain_config`**Type**: `UNORDERED_LIST_STRUCT`**Provider name**: `certificateAuthorityDomainConfig`**Description**: Private registry access configuration.
      - `fqdns`**Type**: `UNORDERED_LIST_STRING`**Provider name**: `fqdns`**Description**: List of fully qualified domain names (FQDN). Specifying port is supported. Wilcards are NOT supported. Examples: - my.customdomain.com - 10.0.1.2:5000
      - `gcp_secret_manager_certificate_config`**Type**: `STRUCT`**Provider name**: `gcpSecretManagerCertificateConfig`**Description**: Google Secret Manager (GCP) certificate configuration.
        - `secret_uri`**Type**: `STRING`**Provider name**: `secretUri`**Description**: Secret URI, in the form "projects/$PROJECT_ID/secrets/$SECRET_NAME/versions/$VERSION". Version can be fixed (e.g. "2") or "latest"
    - `enabled`**Type**: `BOOLEAN`**Provider name**: `enabled`**Description**: Private registry access is enabled.

- `disk_size_gb`**Type**: `INT32`**Provider name**: `diskSizeGb`**Description**: Size of the disk attached to each node, specified in GB. The smallest allowed disk size is 10GB. If unspecified, the default disk size is 100GB.

- `disk_type`**Type**: `STRING`**Provider name**: `diskType`**Description**: Type of the disk attached to each node (e.g. 'pd-standard', 'pd-ssd' or 'pd-balanced') If unspecified, the default disk type is 'pd-standard'

- `enable_confidential_storage`**Type**: `BOOLEAN`**Provider name**: `enableConfidentialStorage`**Description**: Optional. Reserved for future use.

- `ephemeral_storage_config`**Type**: `STRUCT`**Provider name**: `ephemeralStorageConfig`**Description**: Parameters for the ephemeral storage filesystem. If unspecified, ephemeral storage is backed by the boot disk.

  - `local_ssd_count`**Type**: `INT32`**Provider name**: `localSsdCount`**Description**: Number of local SSDs to use to back ephemeral storage. Uses NVMe interfaces. The limit for this value is dependent upon the maximum number of disk available on a machine per zone. See: [https://cloud.google.com/compute/docs/disks/local-ssd](https://cloud.google.com/compute/docs/disks/local-ssd) for more information. A zero (or unset) value has different meanings depending on machine type being used: 1. For pre-Gen3 machines, which support flexible numbers of local ssds, zero (or unset) means to disable using local SSDs as ephemeral storage. 2. For Gen3 machines which dictate a specific number of local ssds, zero (or unset) means to use the default number of local ssds that goes with that machine type. For example, for a c3-standard-8-lssd machine, 2 local ssds would be provisioned. For c3-standard-8 (which doesn't support local ssds), 0 will be provisioned. See [https://cloud.google.com/compute/docs/disks/local-ssd#choose_number_local_ssds](https://cloud.google.com/compute/docs/disks/local-ssd#choose_number_local_ssds) for more info.

- `ephemeral_storage_local_ssd_config`**Type**: `STRUCT`**Provider name**: `ephemeralStorageLocalSsdConfig`**Description**: Parameters for the node ephemeral storage using Local SSDs. If unspecified, ephemeral storage is backed by the boot disk. This field is functionally equivalent to the ephemeral_storage_config

  - `local_ssd_count`**Type**: `INT32`**Provider name**: `localSsdCount`**Description**: Number of local SSDs to use to back ephemeral storage. Uses NVMe interfaces. A zero (or unset) value has different meanings depending on machine type being used: 1. For pre-Gen3 machines, which support flexible numbers of local ssds, zero (or unset) means to disable using local SSDs as ephemeral storage. The limit for this value is dependent upon the maximum number of disk available on a machine per zone. See: [https://cloud.google.com/compute/docs/disks/local-ssd](https://cloud.google.com/compute/docs/disks/local-ssd) for more information. 2. For Gen3 machines which dictate a specific number of local ssds, zero (or unset) means to use the default number of local ssds that goes with that machine type. For example, for a c3-standard-8-lssd machine, 2 local ssds would be provisioned. For c3-standard-8 (which doesn't support local ssds), 0 will be provisioned. See [https://cloud.google.com/compute/docs/disks/local-ssd#choose_number_local_ssds](https://cloud.google.com/compute/docs/disks/local-ssd#choose_number_local_ssds) for more info.

- `fast_socket`**Type**: `STRUCT`**Provider name**: `fastSocket`**Description**: Enable or disable NCCL fast socket for the node pool.

  - `enabled`**Type**: `BOOLEAN`**Provider name**: `enabled`**Description**: Whether Fast Socket features are enabled in the node pool.

- `gcfs_config`**Type**: `STRUCT`**Provider name**: `gcfsConfig`**Description**: GCFS (Google Container File System) configs.

  - `enabled`**Type**: `BOOLEAN`**Provider name**: `enabled`**Description**: Whether to use GCFS.

- `gvnic`**Type**: `STRUCT`**Provider name**: `gvnic`**Description**: Enable or disable gvnic on the node pool.

  - `enabled`**Type**: `BOOLEAN`**Provider name**: `enabled`**Description**: Whether gVNIC features are enabled in the node pool.

- `host_maintenance_policy`**Type**: `STRUCT`**Provider name**: `hostMaintenancePolicy`**Description**: HostMaintenancePolicy contains the desired maintenance policy for the Google Compute Engine hosts.

  - `maintenance_interval`**Type**: `STRING`**Provider name**: `maintenanceInterval`**Description**: Specifies the frequency of planned maintenance events.**Possible values**:
    - `MAINTENANCE_INTERVAL_UNSPECIFIED` - The maintenance interval is not explicitly specified.
    - `AS_NEEDED` - Nodes are eligible to receive infrastructure and hypervisor updates as they become available. This may result in more maintenance operations (live migrations or terminations) for the node than the PERIODIC option.
    - `PERIODIC` - Nodes receive infrastructure and hypervisor updates on a periodic basis, minimizing the number of maintenance operations (live migrations or terminations) on an individual VM. This may mean underlying VMs will take longer to receive an update than if it was configured for AS_NEEDED. Security updates will still be applied as soon as they are available.
  - `opportunistic_maintenance_strategy`**Type**: `STRUCT`**Provider name**: `opportunisticMaintenanceStrategy`**Description**: Strategy that will trigger maintenance on behalf of the customer.
    - `maintenance_availability_window`**Type**: `STRING`**Provider name**: `maintenanceAvailabilityWindow`**Description**: The window of time that opportunistic maintenance can run. Example: A setting of 14 days implies that opportunistic maintenance can only be ran in the 2 weeks leading up to the scheduled maintenance date. Setting 28 days allows opportunistic maintenance to run at any time in the scheduled maintenance window (all `PERIODIC` maintenance is set 28 days in advance).
    - `min_nodes_per_pool`**Type**: `INT64`**Provider name**: `minNodesPerPool`**Description**: The minimum nodes required to be available in a pool. Blocks maintenance if it would cause the number of running nodes to dip below this value.
    - `node_idle_time_window`**Type**: `STRING`**Provider name**: `nodeIdleTimeWindow`**Description**: The amount of time that a node can remain idle (no customer owned workloads running), before triggering maintenance.

- `image_type`**Type**: `STRING`**Provider name**: `imageType`**Description**: The image type to use for this node. Note that for a given image type, the latest version of it will be used. Please see [https://cloud.google.com/kubernetes-engine/docs/concepts/node-images](https://cloud.google.com/kubernetes-engine/docs/concepts/node-images) for available image types.

- `kubelet_config`**Type**: `STRUCT`**Provider name**: `kubeletConfig`**Description**: Node kubelet configs.

  - `cpu_cfs_quota`**Type**: `BOOLEAN`**Provider name**: `cpuCfsQuota`**Description**: Enable CPU CFS quota enforcement for containers that specify CPU limits. This option is enabled by default which makes kubelet use CFS quota ([https://www.kernel.org/doc/Documentation/scheduler/sched-bwc.txt](https://www.kernel.org/doc/Documentation/scheduler/sched-bwc.txt)) to enforce container CPU limits. Otherwise, CPU limits will not be enforced at all. Disable this option to mitigate CPU throttling problems while still having your pods to be in Guaranteed QoS class by specifying the CPU limits. The default value is 'true' if unspecified.
  - `cpu_cfs_quota_period`**Type**: `STRING`**Provider name**: `cpuCfsQuotaPeriod`**Description**: Set the CPU CFS quota period value 'cpu.cfs_period_us'. The string must be a sequence of decimal numbers, each with optional fraction and a unit suffix, such as "300ms". Valid time units are "ns", "us" (or "µs"), "ms", "s", "m", "h". The value must be a positive duration.
  - `cpu_manager_policy`**Type**: `STRING`**Provider name**: `cpuManagerPolicy`**Description**: Control the CPU management policy on the node. See [https://kubernetes.io/docs/tasks/administer-cluster/cpu-management-policies/](https://kubernetes.io/docs/tasks/administer-cluster/cpu-management-policies/) The following values are allowed. * "none": the default, which represents the existing scheduling behavior. * "static": allows pods with certain resource characteristics to be granted increased CPU affinity and exclusivity on the node. The default value is 'none' if unspecified.
  - `insecure_kubelet_readonly_port_enabled`**Type**: `BOOLEAN`**Provider name**: `insecureKubeletReadonlyPortEnabled`**Description**: Enable or disable Kubelet read only port.
  - `pod_pids_limit`**Type**: `INT64`**Provider name**: `podPidsLimit`**Description**: Set the Pod PID limits. See [https://kubernetes.io/docs/concepts/policy/pid-limiting/#pod-pid-limits](https://kubernetes.io/docs/concepts/policy/pid-limiting/#pod-pid-limits) Controls the maximum number of processes allowed to run in a pod. The value must be greater than or equal to 1024 and less than 4194304.

- `linux_node_config`**Type**: `STRUCT`**Provider name**: `linuxNodeConfig`**Description**: Parameters that can be configured on Linux nodes.

  - `cgroup_mode`**Type**: `STRING`**Provider name**: `cgroupMode`**Description**: cgroup_mode specifies the cgroup mode to be used on the node.**Possible values**:
    - `CGROUP_MODE_UNSPECIFIED` - CGROUP_MODE_UNSPECIFIED is when unspecified cgroup configuration is used. The default for the GKE node OS image will be used.
    - `CGROUP_MODE_V1` - CGROUP_MODE_V1 specifies to use cgroupv1 for the cgroup configuration on the node image.
    - `CGROUP_MODE_V2` - CGROUP_MODE_V2 specifies to use cgroupv2 for the cgroup configuration on the node image.
  - `hugepages`**Type**: `STRUCT`**Provider name**: `hugepages`**Description**: Optional. Amounts for 2M and 1G hugepages
    - `hugepage_size1g`**Type**: `INT32`**Provider name**: `hugepageSize1g`**Description**: Optional. Amount of 1G hugepages
    - `hugepage_size2m`**Type**: `INT32`**Provider name**: `hugepageSize2m`**Description**: Optional. Amount of 2M hugepages

- `local_nvme_ssd_block_config`**Type**: `STRUCT`**Provider name**: `localNvmeSsdBlockConfig`**Description**: Parameters for using raw-block Local NVMe SSDs.

  - `local_ssd_count`**Type**: `INT32`**Provider name**: `localSsdCount`**Description**: Number of local NVMe SSDs to use. The limit for this value is dependent upon the maximum number of disk available on a machine per zone. See: [https://cloud.google.com/compute/docs/disks/local-ssd](https://cloud.google.com/compute/docs/disks/local-ssd) for more information. A zero (or unset) value has different meanings depending on machine type being used: 1. For pre-Gen3 machines, which support flexible numbers of local ssds, zero (or unset) means to disable using local SSDs as ephemeral storage. 2. For Gen3 machines which dictate a specific number of local ssds, zero (or unset) means to use the default number of local ssds that goes with that machine type. For example, for a c3-standard-8-lssd machine, 2 local ssds would be provisioned. For c3-standard-8 (which doesn't support local ssds), 0 will be provisioned. See [https://cloud.google.com/compute/docs/disks/local-ssd#choose_number_local_ssds](https://cloud.google.com/compute/docs/disks/local-ssd#choose_number_local_ssds) for more info.

- `local_ssd_count`**Type**: `INT32`**Provider name**: `localSsdCount`**Description**: The number of local SSD disks to be attached to the node. The limit for this value is dependent upon the maximum number of disks available on a machine per zone. See: [https://cloud.google.com/compute/docs/disks/local-ssd](https://cloud.google.com/compute/docs/disks/local-ssd) for more information.

- `logging_config`**Type**: `STRUCT`**Provider name**: `loggingConfig`**Description**: Logging configuration.

  - `variant_config`**Type**: `STRUCT`**Provider name**: `variantConfig`**Description**: Logging variant configuration.
    - `variant`**Type**: `STRING`**Provider name**: `variant`**Description**: Logging variant deployed on nodes.**Possible values**:
      - `VARIANT_UNSPECIFIED` - Default value. This shouldn't be used.
      - `DEFAULT` - default logging variant.
      - `MAX_THROUGHPUT` - maximum logging throughput variant.

- `machine_type`**Type**: `STRING`**Provider name**: `machineType`**Description**: The name of a Google Compute Engine [machine type](https://cloud.google.com/compute/docs/machine-types). If unspecified, the default machine type is `e2-medium`.

- `min_cpu_platform`**Type**: `STRING`**Provider name**: `minCpuPlatform`**Description**: Minimum CPU platform to be used by this instance. The instance may be scheduled on the specified or newer CPU platform. Applicable values are the friendly names of CPU platforms, such as `minCpuPlatform: "Intel Haswell"` or `minCpuPlatform: "Intel Sandy Bridge"`. For more information, read [how to specify min CPU platform](https://cloud.google.com/compute/docs/instances/specify-min-cpu-platform).

- `node_group`**Type**: `STRING`**Provider name**: `nodeGroup`**Description**: Setting this field will assign instances of this pool to run on the specified node group. This is useful for running workloads on [sole tenant nodes](https://cloud.google.com/compute/docs/nodes/sole-tenant-nodes).

- `oauth_scopes`**Type**: `UNORDERED_LIST_STRING`**Provider name**: `oauthScopes`**Description**: The set of Google API scopes to be made available on all of the node VMs under the "default" service account. The following scopes are recommended, but not required, and by default are not included: * `https://www.googleapis.com/auth/compute` is required for mounting persistent storage on your nodes. * `https://www.googleapis.com/auth/devstorage.read_only` is required for communicating with **gcr.io** (the [Google Container Registry](https://cloud.google.com/container-registry/)). If unspecified, no scopes are added, unless Cloud Logging or Cloud Monitoring are enabled, in which case their required scopes will be added.

- `preemptible`**Type**: `BOOLEAN`**Provider name**: `preemptible`**Description**: Whether the nodes are created as preemptible VM instances. See: [https://cloud.google.com/compute/docs/instances/preemptible](https://cloud.google.com/compute/docs/instances/preemptible) for more information about preemptible VM instances.

- `reservation_affinity`**Type**: `STRUCT`**Provider name**: `reservationAffinity`**Description**: The optional reservation affinity. Setting this field will apply the specified [Zonal Compute Reservation](https://cloud.google.com/compute/docs/instances/reserving-zonal-resources) to this node pool.

  - `consume_reservation_type`**Type**: `STRING`**Provider name**: `consumeReservationType`**Description**: Corresponds to the type of reservation consumption.**Possible values**:
    - `UNSPECIFIED` - Default value. This should not be used.
    - `NO_RESERVATION` - Do not consume from any reserved capacity.
    - `ANY_RESERVATION` - Consume any reservation available.
    - `SPECIFIC_RESERVATION` - Must consume from a specific reservation. Must specify key value fields for specifying the reservations.
  - `key`**Type**: `STRING`**Provider name**: `key`**Description**: Corresponds to the label key of a reservation resource. To target a SPECIFIC_RESERVATION by name, specify "compute.googleapis.com/reservation-name" as the key and specify the name of your reservation as its value.
  - `values`**Type**: `UNORDERED_LIST_STRING`**Provider name**: `values`**Description**: Corresponds to the label value(s) of reservation resource(s).

- `resource_labels`**Type**: `MAP_STRING_STRING`**Provider name**: `resourceLabels`**Description**: The resource labels for the node pool to use to annotate any related Google Compute Engine resources.

- `resource_manager_tags`**Type**: `STRUCT`**Provider name**: `resourceManagerTags`**Description**: A map of resource manager tag keys and values to be attached to the nodes.

- `sandbox_config`**Type**: `STRUCT`**Provider name**: `sandboxConfig`**Description**: Sandbox configuration for this node.

  - `sandbox_type`**Type**: `STRING`**Provider name**: `sandboxType`**Description**: Type of the sandbox to use for the node (e.g. 'gvisor')
  - `type`**Type**: `STRING`**Provider name**: `type`**Description**: Type of the sandbox to use for the node.**Possible values**:
    - `UNSPECIFIED` - Default value. This should not be used.
    - `GVISOR` - Run sandbox using gvisor.

- `secondary_boot_disk_update_strategy`**Type**: `STRUCT`**Provider name**: `secondaryBootDiskUpdateStrategy`**Description**: Secondary boot disk update strategy.

- `secondary_boot_disks`**Type**: `UNORDERED_LIST_STRUCT`**Provider name**: `secondaryBootDisks`**Description**: List of secondary boot disks attached to the nodes.

  - `disk_image`**Type**: `STRING`**Provider name**: `diskImage`**Description**: Fully-qualified resource ID for an existing disk image.
  - `mode`**Type**: `STRING`**Provider name**: `mode`**Description**: Disk mode (container image cache, etc.)**Possible values**:
    - `MODE_UNSPECIFIED` - MODE_UNSPECIFIED is when mode is not set.
    - `CONTAINER_IMAGE_CACHE` - CONTAINER_IMAGE_CACHE is for using the secondary boot disk as a container image cache.

- `service_account`**Type**: `STRING`**Provider name**: `serviceAccount`**Description**: The Google Cloud Platform Service Account to be used by the node VMs. Specify the email address of the Service Account; otherwise, if no Service Account is specified, the "default" service account is used.

- `shielded_instance_config`**Type**: `STRUCT`**Provider name**: `shieldedInstanceConfig`**Description**: Shielded Instance options.

  - `enable_integrity_monitoring`**Type**: `BOOLEAN`**Provider name**: `enableIntegrityMonitoring`**Description**: Defines whether the instance has integrity monitoring enabled. Enables monitoring and attestation of the boot integrity of the instance. The attestation is performed against the integrity policy baseline. This baseline is initially derived from the implicitly trusted boot image when the instance is created.
  - `enable_secure_boot`**Type**: `BOOLEAN`**Provider name**: `enableSecureBoot`**Description**: Defines whether the instance has Secure Boot enabled. Secure Boot helps ensure that the system only runs authentic software by verifying the digital signature of all boot components, and halting the boot process if signature verification fails.

- `sole_tenant_config`**Type**: `STRUCT`**Provider name**: `soleTenantConfig`**Description**: Parameters for node pools to be backed by shared sole tenant node groups.

  - `node_affinities`**Type**: `UNORDERED_LIST_STRUCT`**Provider name**: `nodeAffinities`**Description**: NodeAffinities used to match to a shared sole tenant node group.
    - `key`**Type**: `STRING`**Provider name**: `key`**Description**: Key for NodeAffinity.
    - `operator`**Type**: `STRING`**Provider name**: `operator`**Description**: Operator for NodeAffinity.**Possible values**:
      - `OPERATOR_UNSPECIFIED` - Invalid or unspecified affinity operator.
      - `IN` - Affinity operator.
      - `NOT_IN` - Anti-affinity operator.
    - `values`**Type**: `UNORDERED_LIST_STRING`**Provider name**: `values`**Description**: Values for NodeAffinity.

- `spot`**Type**: `BOOLEAN`**Provider name**: `spot`**Description**: Spot flag for enabling Spot VM, which is a rebrand of the existing preemptible flag.

- `taints`**Type**: `UNORDERED_LIST_STRUCT`**Provider name**: `taints`**Description**: List of kubernetes taints to be applied to each node. For more information, including usage and the valid values, see: [https://kubernetes.io/docs/concepts/configuration/taint-and-toleration/](https://kubernetes.io/docs/concepts/configuration/taint-and-toleration/)

  - `effect`**Type**: `STRING`**Provider name**: `effect`**Description**: Effect for taint.**Possible values**:
    - `EFFECT_UNSPECIFIED` - Not set
    - `NO_SCHEDULE` - NoSchedule
    - `PREFER_NO_SCHEDULE` - PreferNoSchedule
    - `NO_EXECUTE` - NoExecute
  - `key`**Type**: `STRING`**Provider name**: `key`**Description**: Key for taint.
  - `value`**Type**: `STRING`**Provider name**: `value`**Description**: Value for taint.

- `windows_node_config`**Type**: `STRUCT`**Provider name**: `windowsNodeConfig`**Description**: Parameters that can be configured on Windows nodes.

  - `os_version`**Type**: `STRING`**Provider name**: `osVersion`**Description**: OSVersion specifies the Windows node config to be used on the node**Possible values**:
    - `OS_VERSION_UNSPECIFIED` - When OSVersion is not specified
    - `OS_VERSION_LTSC2019` - LTSC2019 specifies to use LTSC2019 as the Windows Servercore Base Image
    - `OS_VERSION_LTSC2022` - LTSC2022 specifies to use LTSC2022 as the Windows Servercore Base Image

- `workload_metadata_config`**Type**: `STRUCT`**Provider name**: `workloadMetadataConfig`**Description**: The workload metadata configuration for this node.

  - `mode`**Type**: `STRING`**Provider name**: `mode`**Description**: Mode is the configuration for how to expose metadata to workloads running on the node pool.**Possible values**:
    - `MODE_UNSPECIFIED` - Not set.
    - `GCE_METADATA` - Expose all Compute Engine metadata to pods.
    - `GKE_METADATA` - Run the GKE Metadata Server on this node. The GKE Metadata Server exposes a metadata API to workloads that is compatible with the V1 Compute Metadata APIs exposed by the Compute Engine and App Engine Metadata Servers. This feature can only be enabled if Workload Identity is enabled at the cluster level.
  - `node_metadata`**Type**: `STRING`**Provider name**: `nodeMetadata`**Description**: NodeMetadata is the configuration for how to expose metadata to the workloads running on the node.**Possible values**:
    - `UNSPECIFIED` - Not set.
    - `SECURE` - Prevent workloads not in hostNetwork from accessing certain VM metadata, specifically kube-env, which contains Kubelet credentials, and the instance identity token. Metadata concealment is a temporary security solution available while the bootstrapping process for cluster nodes is being redesigned with significant security improvements. This feature is scheduled to be deprecated in the future and later removed.
    - `EXPOSE` - Expose all VM metadata to pods.
    - `GKE_METADATA_SERVER` - Run the GKE Metadata Server on this node. The GKE Metadata Server exposes a metadata API to workloads that is compatible with the V1 Compute Metadata APIs exposed by the Compute Engine and App Engine Metadata Servers. This feature can only be enabled if Workload Identity is enabled at the cluster level.

## `etag`{% #etag %}

**Type**: `STRING`**Provider name**: `etag`**Description**: This checksum is computed by the server based on the value of node pool fields, and may be sent on update requests to ensure the client has an up-to-date value before proceeding.

## `gcp_status`{% #gcp_status %}

**Type**: `STRING`**Provider name**: `status`**Description**: Output only. The status of the nodes in this pool instance.**Possible values**:

- `STATUS_UNSPECIFIED` - Not set.
- `PROVISIONING` - The PROVISIONING state indicates the node pool is being created.
- `RUNNING` - The RUNNING state indicates the node pool has been created and is fully usable.
- `RUNNING_WITH_ERROR` - The RUNNING_WITH_ERROR state indicates the node pool has been created and is partially usable. Some error state has occurred and some functionality may be impaired. Customer may need to reissue a request or trigger a new update.
- `RECONCILING` - The RECONCILING state indicates that some work is actively being done on the node pool, such as upgrading node software. Details can be found in the `statusMessage` field.
- `STOPPING` - The STOPPING state indicates the node pool is being deleted.
- `ERROR` - The ERROR state indicates the node pool may be unusable. Details can be found in the `statusMessage` field.

## `initial_node_count`{% #initial_node_count %}

**Type**: `INT32`**Provider name**: `initialNodeCount`**Description**: The initial node count for the pool. You must ensure that your Compute Engine [resource quota](https://cloud.google.com/compute/quotas) is sufficient for this number of instances. You must also have available firewall and routes quota.

## `instance_group_urls`{% #instance_group_urls %}

**Type**: `UNORDERED_LIST_STRING`**Provider name**: `instanceGroupUrls`**Description**: Output only. The resource URLs of the [managed instance groups](https://cloud.google.com/compute/docs/instance-groups/creating-groups-of-managed-instances) associated with this node pool. During the node pool blue-green upgrade operation, the URLs contain both blue and green resources.

## `labels`{% #labels %}

**Type**: `UNORDERED_LIST_STRING`

## `locations`{% #locations %}

**Type**: `UNORDERED_LIST_STRING`**Provider name**: `locations`**Description**: The list of Google Compute Engine [zones](https://cloud.google.com/compute/docs/zones#available) in which the NodePool's nodes should be located. If this value is unspecified during node pool creation, the [Cluster.Locations](https://cloud.google.com/kubernetes-engine/docs/reference/rest/v1/projects.locations.clusters#Cluster.FIELDS.locations) value will be used, instead. Warning: changing node pool locations will result in nodes being added and/or removed.

## `management`{% #management %}

**Type**: `STRUCT`**Provider name**: `management`**Description**: NodeManagement configuration for this NodePool.

- `auto_repair`**Type**: `BOOLEAN`**Provider name**: `autoRepair`**Description**: Whether the nodes will be automatically repaired.
- `auto_upgrade`**Type**: `BOOLEAN`**Provider name**: `autoUpgrade`**Description**: Whether the nodes will be automatically upgraded.
- `upgrade_options`**Type**: `STRUCT`**Provider name**: `upgradeOptions`**Description**: Specifies the Auto Upgrade knobs for the node pool.
  - `auto_upgrade_start_time`**Type**: `STRING`**Provider name**: `autoUpgradeStartTime`**Description**: Output only. This field is set when upgrades are about to commence with the approximate start time for the upgrades, in [RFC3339](https://www.ietf.org/rfc/rfc3339.txt) text format.
  - `description`**Type**: `STRING`**Provider name**: `description`**Description**: Output only. This field is set when upgrades are about to commence with the description of the upgrade.

## `max_pods_constraint`{% #max_pods_constraint %}

**Type**: `STRUCT`**Provider name**: `maxPodsConstraint`**Description**: The constraint on the maximum number of pods that can be run simultaneously on a node in the node pool.

- `max_pods_per_node`**Type**: `INT64`**Provider name**: `maxPodsPerNode`**Description**: Constraint enforced on the max num of pods per node.

## `name`{% #name %}

**Type**: `STRING`**Provider name**: `name`**Description**: The name of the node pool.

## `network_config`{% #network_config %}

**Type**: `STRUCT`**Provider name**: `networkConfig`**Description**: Networking configuration for this NodePool. If specified, it overrides the cluster-level defaults.

- `additional_node_network_configs`**Type**: `UNORDERED_LIST_STRUCT`**Provider name**: `additionalNodeNetworkConfigs`**Description**: We specify the additional node networks for this node pool using this list. Each node network corresponds to an additional interface
  - `network`**Type**: `STRING`**Provider name**: `network`**Description**: Name of the VPC where the additional interface belongs
  - `subnetwork`**Type**: `STRING`**Provider name**: `subnetwork`**Description**: Name of the subnetwork where the additional interface belongs
- `additional_pod_network_configs`**Type**: `UNORDERED_LIST_STRUCT`**Provider name**: `additionalPodNetworkConfigs`**Description**: We specify the additional pod networks for this node pool using this list. Each pod network corresponds to an additional alias IP range for the node
  - `max_pods_per_node`**Type**: `STRUCT`**Provider name**: `maxPodsPerNode`**Description**: The maximum number of pods per node which use this pod network.
    - `max_pods_per_node`**Type**: `INT64`**Provider name**: `maxPodsPerNode`**Description**: Constraint enforced on the max num of pods per node.
  - `secondary_pod_range`**Type**: `STRING`**Provider name**: `secondaryPodRange`**Description**: The name of the secondary range on the subnet which provides IP address for this pod range.
  - `subnetwork`**Type**: `STRING`**Provider name**: `subnetwork`**Description**: Name of the subnetwork where the additional pod network belongs.
- `create_pod_range`**Type**: `BOOLEAN`**Provider name**: `createPodRange`**Description**: Input only. Whether to create a new range for pod IPs in this node pool. Defaults are provided for `pod_range` and `pod_ipv4_cidr_block` if they are not specified. If neither `create_pod_range` or `pod_range` are specified, the cluster-level default (`ip_allocation_policy.cluster_ipv4_cidr_block`) is used. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created.
- `enable_private_nodes`**Type**: `BOOLEAN`**Provider name**: `enablePrivateNodes`**Description**: Whether nodes have internal IP addresses only. If enable_private_nodes is not specified, then the value is derived from cluster.privateClusterConfig.enablePrivateNodes
- `network_performance_config`**Type**: `STRUCT`**Provider name**: `networkPerformanceConfig`**Description**: Network bandwidth tier configuration.
  - `external_ip_egress_bandwidth_tier`**Type**: `STRING`**Provider name**: `externalIpEgressBandwidthTier`**Description**: Specifies the network bandwidth tier for the NodePool for traffic to external/public IP addresses.**Possible values**:
    - `TIER_UNSPECIFIED` - Default value
    - `TIER_1` - Higher bandwidth, actual values based on VM size.
  - `total_egress_bandwidth_tier`**Type**: `STRING`**Provider name**: `totalEgressBandwidthTier`**Description**: Specifies the total network bandwidth tier for the NodePool.**Possible values**:
    - `TIER_UNSPECIFIED` - Default value
    - `TIER_1` - Higher bandwidth, actual values based on VM size.
- `pod_cidr_overprovision_config`**Type**: `STRUCT`**Provider name**: `podCidrOverprovisionConfig`**Description**: [PRIVATE FIELD] Pod CIDR size overprovisioning config for the nodepool. Pod CIDR size per node depends on max_pods_per_node. By default, the value of max_pods_per_node is rounded off to next power of 2 and we then double that to get the size of pod CIDR block per node. Example: max_pods_per_node of 30 would result in 64 IPs (/26). This config can disable the doubling of IPs (we still round off to next power of 2) Example: max_pods_per_node of 30 will result in 32 IPs (/27) when overprovisioning is disabled.
  - `disable`**Type**: `BOOLEAN`**Provider name**: `disable`**Description**: Whether Pod CIDR overprovisioning is disabled. Note: Pod CIDR overprovisioning is enabled by default.
- `pod_ipv4_cidr_block`**Type**: `STRING`**Provider name**: `podIpv4CidrBlock`**Description**: The IP address range for pod IPs in this node pool. Only applicable if `create_pod_range` is true. Set to blank to have a range chosen with the default size. Set to /netmask (e.g. `/14`) to have a range chosen with a specific netmask. Set to a [CIDR](https://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) to pick a specific range to use. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created.
- `pod_ipv4_range_utilization`**Type**: `DOUBLE`**Provider name**: `podIpv4RangeUtilization`**Description**: Output only. The utilization of the IPv4 range for the pod. The ratio is Usage/[Total number of IPs in the secondary range], Usage=numNodes*numZones*podIPsPerNode.
- `pod_range`**Type**: `STRING`**Provider name**: `podRange`**Description**: The ID of the secondary range for pod IPs. If `create_pod_range` is true, this ID is used for the new range. If `create_pod_range` is false, uses an existing secondary range with this ID. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created.

## `organization_id`{% #organization_id %}

**Type**: `STRING`

## `parent`{% #parent %}

**Type**: `STRING`

## `placement_policy`{% #placement_policy %}

**Type**: `STRUCT`**Provider name**: `placementPolicy`**Description**: Specifies the node placement policy.

- `policy_name`**Type**: `STRING`**Provider name**: `policyName`**Description**: If set, refers to the name of a custom resource policy supplied by the user. The resource policy must be in the same project and region as the node pool. If not found, InvalidArgument error is returned.
- `tpu_topology`**Type**: `STRING`**Provider name**: `tpuTopology`**Description**: TPU placement topology for pod slice node pool. [https://cloud.google.com/tpu/docs/types-topologies#tpu_topologies](https://cloud.google.com/tpu/docs/types-topologies#tpu_topologies)
- `type`**Type**: `STRING`**Provider name**: `type`**Description**: The type of placement.**Possible values**:
  - `TYPE_UNSPECIFIED` - TYPE_UNSPECIFIED specifies no requirements on nodes placement.
  - `COMPACT` - COMPACT specifies node placement in the same availability domain to ensure low communication latency.

## `pod_ipv4_cidr_size`{% #pod_ipv4_cidr_size %}

**Type**: `INT32`**Provider name**: `podIpv4CidrSize`**Description**: Output only. The pod CIDR block size per node in this node pool.

## `project_id`{% #project_id %}

**Type**: `STRING`

## `project_number`{% #project_number %}

**Type**: `STRING`

## `queued_provisioning`{% #queued_provisioning %}

**Type**: `STRUCT`**Provider name**: `queuedProvisioning`**Description**: Specifies the configuration of queued provisioning.

- `enabled`**Type**: `BOOLEAN`**Provider name**: `enabled`**Description**: Denotes that this nodepool is QRM specific, meaning nodes can be only obtained through queuing via the Cluster Autoscaler ProvisioningRequest API.

## `region_id`{% #region_id %}

**Type**: `STRING`

## `resource_name`{% #resource_name %}

**Type**: `STRING`

## `self_link`{% #self_link %}

**Type**: `STRING`**Provider name**: `selfLink`**Description**: Output only. Server-defined URL for the resource.

## `status_message`{% #status_message %}

**Type**: `STRING`**Provider name**: `statusMessage`**Description**: Output only. Deprecated. Use conditions instead. Additional information about the current status of this node pool instance, if available.

## `tags`{% #tags %}

**Type**: `UNORDERED_LIST_STRING`

## `update_info`{% #update_info %}

**Type**: `STRUCT`**Provider name**: `updateInfo`**Description**: Output only. Update info contains relevant information during a node pool update.

- `blue_green_info`**Type**: `STRUCT`**Provider name**: `blueGreenInfo`**Description**: Information of a blue-green upgrade.
  - `blue_instance_group_urls`**Type**: `UNORDERED_LIST_STRING`**Provider name**: `blueInstanceGroupUrls`**Description**: The resource URLs of the [managed instance groups] (/compute/docs/instance-groups/creating-groups-of-managed-instances) associated with blue pool.
  - `blue_pool_deletion_start_time`**Type**: `STRING`**Provider name**: `bluePoolDeletionStartTime`**Description**: Time to start deleting blue pool to complete blue-green upgrade, in [RFC3339](https://www.ietf.org/rfc/rfc3339.txt) text format.
  - `green_instance_group_urls`**Type**: `UNORDERED_LIST_STRING`**Provider name**: `greenInstanceGroupUrls`**Description**: The resource URLs of the [managed instance groups] (/compute/docs/instance-groups/creating-groups-of-managed-instances) associated with green pool.
  - `green_pool_version`**Type**: `STRING`**Provider name**: `greenPoolVersion`**Description**: Version of green pool.
  - `phase`**Type**: `STRING`**Provider name**: `phase`**Description**: Current blue-green upgrade phase.**Possible values**:
    - `PHASE_UNSPECIFIED` - Unspecified phase.
    - `UPDATE_STARTED` - blue-green upgrade has been initiated.
    - `CREATING_GREEN_POOL` - Start creating green pool nodes.
    - `CORDONING_BLUE_POOL` - Start cordoning blue pool nodes.
    - `WAITING_TO_DRAIN_BLUE_POOL` - Start waiting after cordoning the blue pool and before draining it.
    - `DRAINING_BLUE_POOL` - Start draining blue pool nodes.
    - `NODE_POOL_SOAKING` - Start soaking time after draining entire blue pool.
    - `DELETING_BLUE_POOL` - Start deleting blue nodes.
    - `ROLLBACK_STARTED` - Rollback has been initiated.

## `upgrade_settings`{% #upgrade_settings %}

**Type**: `STRUCT`**Provider name**: `upgradeSettings`**Description**: Upgrade settings control disruption and speed of the upgrade.

- `blue_green_settings`**Type**: `STRUCT`**Provider name**: `blueGreenSettings`**Description**: Settings for blue-green upgrade strategy.
  - `autoscaled_rollout_policy`**Type**: `STRUCT`**Provider name**: `autoscaledRolloutPolicy`**Description**: Autoscaled policy for cluster autoscaler enabled blue-green upgrade.

  - `node_pool_soak_duration`**Type**: `STRING`**Provider name**: `nodePoolSoakDuration`**Description**: Time needed after draining entire blue pool. After this period, blue pool will be cleaned up.

  - `standard_rollout_policy`**Type**: `STRUCT`**Provider name**: `standardRolloutPolicy`**Description**: Standard policy for the blue-green upgrade.

    - `batch_node_count`**Type**: `INT32`**Provider name**: `batchNodeCount`**Description**: Number of blue nodes to drain in a batch.
    - `batch_percentage`**Type**: `FLOAT`**Provider name**: `batchPercentage`**Description**: Percentage of the blue pool nodes to drain in a batch. The range of this field should be (0.0, 1.0].
    - `batch_soak_duration`**Type**: `STRING`**Provider name**: `batchSoakDuration`**Description**: Soak time after each batch gets drained. Default to zero.
- `max_surge`**Type**: `INT32`**Provider name**: `maxSurge`**Description**: The maximum number of nodes that can be created beyond the current size of the node pool during the upgrade process.
- `max_unavailable`**Type**: `INT32`**Provider name**: `maxUnavailable`**Description**: The maximum number of nodes that can be simultaneously unavailable during the upgrade process. A node is considered available if its status is Ready.
- `strategy`**Type**: `STRING`**Provider name**: `strategy`**Description**: Update strategy of the node pool.**Possible values**:
  - `NODE_POOL_UPDATE_STRATEGY_UNSPECIFIED` - Default value if unset. GKE internally defaults the update strategy to SURGE for unspecified strategies.
  - `BLUE_GREEN` - blue-green upgrade.
  - `SURGE` - SURGE is the traditional way of upgrading a node pool. max_surge and max_unavailable determines the level of upgrade parallelism.

## `version`{% #version %}

**Type**: `STRING`**Provider name**: `version`**Description**: The version of Kubernetes running on this NodePool's nodes. If unspecified, it defaults as described [here](https://cloud.google.com/kubernetes-engine/versioning#specifying_node_version).

## `zone_id`{% #zone_id %}

**Type**: `STRING`
