This page shows you how to investigate and resolve GKE logging related issues.
If you need additional assistance, reach out to Cloud Customer Care.Missing cluster logs in Cloud Logging
Verify logging is enabled in the project
List enabled services:
gcloud services list --enabled --filter="NAME=logging.googleapis.com"
The following output indicates that logging is enabled for the project:
NAME TITLE logging.googleapis.com Cloud Logging API
Optional: Check the logs in Logs Viewer to determine who disabled the API and when they disabled the API:
protoPayload.methodName="google.api.serviceusage.v1.ServiceUsage.DisableService" protoPayload.response.services="logging.googleapis.com"
If logging is disabled, enable logging:
gcloud services enable logging.googleapis.com
Verify logging is enabled on the cluster
List the clusters:
gcloud container clusters list \ --project=PROJECT_ID \ '--format=value(name,loggingConfig.componentConfig.enableComponents)' \ --sort-by=name | column -t
Replace the following:
PROJECT_ID
: your Google Cloud project ID.
The output is similar to the following:
cluster-1 SYSTEM_COMPONENTS cluster-2 SYSTEM_COMPONENTS;WORKLOADS cluster-3
If the value for your cluster is empty, logging is disabled. For example,
cluster-3
in this output has logging disabled.Enable cluster logging if set to
NONE
:gcloud container clusters update CLUSTER_NAME \ --logging=SYSTEM,WORKLOAD \ --location=COMPUTE_LOCATION
Replace the following:
CLUSTER_NAME
: the name of your cluster.COMPUTE_LOCATION
: the Compute Engine location for your cluster.
Verify nodes in the node pools have Cloud Logging access scope
One of the following scopes is required for nodes to write logs to Cloud Logging:
https://www.googleapis.com/auth/logging.write
https://www.googleapis.com/auth/cloud-platform
https://www.googleapis.com/auth/logging.admin
Check the scopes configured on each node pool in the cluster:
gcloud container node-pools list --cluster=CLUSTER_NAME \ --format="table(name,config.oauthScopes)" \ --location COMPUTE_LOCATION
Replace the following:
CLUSTER_NAME
: the name of your cluster.COMPUTE_LOCATION
: the Compute Engine location for your cluster.
Migrate your workloads from the old node pool to the newly created node pool and monitor the progress.
Create new node pools with the correct logging scope:
gcloud container node-pools create NODE_POOL_NAME \ --cluster=CLUSTER_NAME \ --location=COMPUTE_LOCATION \ --scopes="gke-default"
Replace the following:
CLUSTER_NAME
: the name of your cluster.COMPUTE_LOCATION
: the Compute Engine location for your cluster.
Verify the node pool's service account has a role with the correct IAM permissions
The service account must have a role which contains the permission
logging.logEntries.create
to create logs.
Find the service account for each node pool:
gcloud container node-pools list \ --cluster=CLUSTER_NAME \ --format="table(name,config.serviceAccount)" \ --location=COMPUTE_LOCATION
Replace the following:
CLUSTER_NAME
: the name of your cluster.COMPUTE_LOCATION
: the Compute Engine location for your cluster.
The output is similar to the following:
NAME SERVICE_ACCOUNT default-pool gke-cluster-sa@developer.gserviceaccount.com
If the node pool is using the default Compute Engine service account, you can describe it using the following command. As a best practice, use a minimally-privileged custom service account for your node pools. The default Compute Engine service account contains more than the minimum permissions needed to run your clusters.
gcloud compute project-info describe --format="table(defaultServiceAccount)"
Verify IAM roles have sufficient permissions.
View the permissions contained in roles that are granted to a specific service account:
gcloud projects get-iam-policy PROJECT_ID \ --flatten="bindings[]" \ --filter="bindings.members=serviceAccount:SERVICE_ACCOUNT" \ --format="table[box](bindings.role)"
Grant a role that contains the
logging.logEntries.create
permission to the service account. You can use a predefined role or create a custom role.
Verify that Cloud Logging write API quotas have not been reached
Confirm that you have not reached API write quotas for Cloud Logging.
Go to the Quotas page in the Google Cloud console.
Filter the table by "Cloud Logging API".
Confirm that you have not reached any of the quotas.
Debugging GKE logging issues with gcpdiag
If you are missing or getting incomplete logs from your GKE cluster, use thegcpdiag
tool for troubleshooting.
gcpdiag
is an open source tool. It is not an officially supported Google Cloud product.
You can use the gcpdiag
tool to help you identify and fix Google Cloud
project issues. For more information, see the
gcpdiag project on GitHub.
- Project-Level Logging: Ensures that the Google Cloud project housing the GKE cluster has the Cloud Logging API enabled.
- Cluster-Level Logging: Verifies that logging is explicitly enabled within the configuration of the GKE cluster.
- Node Pool Permissions: Confirms that the nodes within the cluster's node pools have the 'Cloud Logging Write' scope enabled, allowing them to send log data.
- Service Account Permissions: Validates that the service account used by the node pools possesses the necessary IAM permissions to interact with Cloud Logging. Specifically, the 'roles/logging.logWriter' role is typically required.
- Cloud Logging API Write Quotas: Verifies that Cloud Logging API Write quotas have not been exceeded within the specified timeframe.
Google Cloud console
- Complete and then copy the following command.
- Open the Google Cloud console and activate Cloud Shell. Open Cloud console
- Paste the copied command.
- Run the
gcpdiag
command, which downloads thegcpdiag
docker image, and then performs diagnostic checks. If applicable, follow the output instructions to fix failed checks.
GOOGLE_AUTH_TOKEN=GOOGLE_AUTH_TOKEN \
gcpdiag runbook gke/logs \
--parameter project_id=PROJECT_ID \
--parameter name=GKE_NAME \
--parameter location=LOCATION \
--auto --reason=REASON
Docker
You can
run gcpdiag
using a wrapper that starts gcpdiag
in a
Docker container. Docker or
Podman must be installed.
- Copy and run the following command on your local workstation.
curl https://gcpdiag.dev/gcpdiag.sh >gcpdiag && chmod +x gcpdiag
- Execute the
gcpdiag
command../gcpdiag runbook gke/logs \ --parameter project_id=PROJECT_ID \ --parameter name=GKE_NAME \ --parameter location=LOCATION
View available parameters for this runbook.
Replace the following:
- PROJECT_ID: The ID of the project containing the resource.
- GKE_NAME: The name of the GKE cluster.
- LOCATION: The zone or region of the GKE cluster.
Useful flags:
--universe-domain
: If applicable, the Trusted Partner Sovereign Cloud domain hosting the resource--parameter
or-p
: Runbook parameters
For a list and description of all gcpdiag
tool flags, see the
gcpdiag
usage instructions.