This page explains how Google Distributed Cloud Edge logs various types of information about its operation and how to view that information. The collection of some types of logs and metrics incur additional charges. For more information, see Billing for logs and metrics.
Configure logging and monitoring
Before you can start gathering logs and metrics, you must do the following:
Enable the logging APIs by using the following commands:
gcloud services enable opsconfigmonitoring.googleapis.com --project PROJECT_ID gcloud services enable logging.googleapis.com --project PROJECT_ID gcloud services enable monitoring.googleapis.com --project PROJECT_ID
Replace
PROJECT_ID
with the ID of the target Google Cloud project.Grant the roles required to write logs and metrics:
gcloud projects add-iam-policy-binding PROJECT_ID \ --role roles/opsconfigmonitoring.resourceMetadata.writer \ --member "serviceAccount:PROJECT_ID.svc.id.goog[kube-system/metadata-agent]" gcloud projects add-iam-policy-binding PROJECT_ID \ --role roles/logging.logWriter \ --member "serviceAccount:PROJECT_ID.svc.id.goog[kube-system/stackdriver-log-forwarder]" gcloud projects add-iam-policy-binding PROJECT_ID \ --role roles/monitoring.metricWriter \ --member "serviceAccount:PROJECT_ID.svc.id.goog[kube-system/gke-metrics-agent]"
Replace
PROJECT_ID
with the ID of the target Google Cloud project.
Logs
This section lists the Cloud Logging resource types supported by Distributed Cloud Edge. To view Distributed Cloud Edge logs, use the Logs Explorer in the Google Cloud console. Distributed Cloud Edge logging is always enabled.
The Distributed Cloud Edge logged resource types are the following standard Kubernetes resources:
k8s_container
k8s_node
You can also capture and retrieve Distributed Cloud Edge logs by using the Cloud Logging API. For information about how to configure this logging mechanism, see the documentation for Cloud Logging client libraries.
Metrics
This section lists the Cloud Monitoring metrics supported by Distributed Cloud Edge. To view Distributed Cloud Edge metrics, use the Metrics explorer in the Google Cloud console.
Distributed Cloud Edge cluster metrics
For Distributed Cloud Edge clusters, Distributed Cloud Edge provides the following types of metrics generated by Distributed Cloud Edge nodes:
- Resource metrics provide information about Distributed Cloud Edge node and Pod performance, such as CPU load and memory usage.
- System application metrics provide information about
Distributed Cloud Edge system workloads, such as
coredns
.
For a list of these metrics, see Google Distributed Cloud and Google Distributed Cloud metrics.
Distributed Cloud Edge does not provide metrics generated by the Kubernetes control planes associated with Distributed Cloud Edge clusters.
Distributed Cloud Edge hardware metrics
Distributed Cloud Edge provides metrics for Distributed Cloud Edge hardware by using the following resource types:
edgecontainer.googleapis.com/Machine
edgecontainer.googleapis.com/Rack
Machine
resource metrics
Distributed Cloud Edge writes the following Cloud Monitoring API
metrics for the edgecontainer.googleapis.com/Machine
resource:
Metric | Description |
---|---|
/machine/cpu/total_cores
|
Total count of physical processor cores present on the machine. |
/machine/cpu/usage_time
|
Cumulative CPU usage time for all cores on the machine. Type can be
workload (customer workloads)
or system (everything else). |
/machine/cpu/utilization
|
CPU utilization percentage on the machine. Range is 0 to 1. Type can be
workload (customer workloads)
or system (everything else). |
/machine/memory/total_bytes
|
Byte count of total memory on the machine. |
/machine/memory/used_bytes
|
Byte count of used memory on the machine. memory_type is
either evictable (reclaimable by the kernel) or
non-evictable (not reclaimable). |
/machine/memory/utilization
|
Memory utilization percentage on the machine. Range is 0 to 1.
memory_type is either evictable (reclaimable
by the kernel) or non-evictable (not reclaimable). |
/machine/network/up
|
Indicates whether the network interface is up and running. Includes primary cards, secondary cards, and ports. |
/machine/network/link_speed
|
Link speed of the primary network interface card. |
/machine/network/received_bytes_count
|
Received byte count for the primary network interface card. |
/machine/network/sent_bytes_count
|
Sent byte count for the primary network interface card. |
/machine/network/connectivity
|
Indicates whether the primary network interface card has internet connectivity. |
/machine/disk/total_bytes
|
Byte count of total disk space on the machine. |
/machine/disk/used_bytes
|
Byte count of used disk space on the machine. |
/machine/disk/utilization
|
Disk space utilization percentage on the machine. Range is 0 to 1. |
/machine/restart_count
|
Number of restarts that the machine has undergone. |
/machine/uptime
|
Machine uptime since the last restart. |
/machine/connected
|
Indicates whether the machine is connected to Google Cloud. |
Rack
resource metrics
Distributed Cloud Edge writes the following Cloud Monitoring API
metrics for the edgecontainer.googleapis.com/Rack
resource:
Metric | Description |
---|---|
/router/bgp_up
|
Indicates whether the BGP peering session on the router is up and
healthy.
router_id identifies the specific router (up to 2 per rack). |
/router/connected
|
Indicates whether the BGP router is connected to Google Cloud.
router_id identifies the specific router (up to 2 per rack). |
Export custom application logs and metrics
Distributed Cloud Edge automatically exports logs for applications running on Distributed Cloud Edge workloads. To export metrics for an application running on Distributed Cloud Edge workloads, you must annotate it as described in the next section.
Annotate the workload to enable metrics export
To enable the collection of custom metrics from an application, add the following annotations to the application's Service or Deployment manifest:
prometheus.io/scrape: "true"
prometheus.io/path: "ENDPOINT_PATH"
. ReplaceENDPOINT_PATH
with the full path to the target application's metric endpoint.prometheus.io/port: "PORT_NUMBER"
: the port on which the application's metric endpoint listens for connections.
Run an example application
In this section, you create an application that writes custom logs and exposes a custom metric endpoint.
Save the following Service and Deployment manifests to a file named
my-app.yaml
. Notice that the Service has the annotationprometheus.io/scrape: "true"
:kind: Service apiVersion: v1 metadata: name: "monitoring-example" namespace: "default" annotations: prometheus.io/scrape: "true" spec: selector: app: "monitoring-example" ports: - name: http port: 9090 --- apiVersion: apps/v1 kind: Deployment metadata: name: "monitoring-example" namespace: "default" labels: app: "monitoring-example" spec: replicas: 1 selector: matchLabels: app: "monitoring-example" template: metadata: labels: app: "monitoring-example" spec: containers: - image: gcr.io/google-samples/prometheus-dummy-exporter:latest name: prometheus-example-exporter imagePullPolicy: Always command: - /bin/sh - -c - ./prometheus-dummy-exporter --metric-name=example_monitoring_up --metric-value=1 --port=9090 resources: requests: cpu: 100m
Create the Deployment and the Service:
kubectl --kubeconfig apply -f my-app.yaml
View application logs
Console
In the Google Cloud console, go to the Logs Explorer page.
Click Resource.
In the All resource types list, select
Kubernetes Container
.For Cluster name, select the name of your user cluster.
For Namespace name, select
default
.Click Add, and then click Run query.
In the Query results section, you can see log entries from the
monitoring-example
Deployment. For example:{ "textPayload": "2020/11/14 01:24:24 Starting to listen on :9090\n", "insertId": "1oa4vhg3qfxidt", "resource": { "type": "k8s_container", "labels": { "pod_name": "monitoring-example-7685d96496-xqfsf", "cluster_name": ..., "namespace_name": "default", "project_id": ..., "location": "us-west1", "container_name": "prometheus-example-exporter" } }, "timestamp": "2020-11-14T01:24:24.358600252Z", "labels": { "k8s-pod/pod-template-hash": "7685d96496", "k8s-pod/app": "monitoring-example" }, "logName": "projects/.../logs/stdout", "receiveTimestamp": "2020-11-14T01:24:39.562864735Z" }
gcloud
Use the
gcloud logging read
command:gcloud logging read 'resource.labels.project_id="PROJECT_ID" AND \ resource.type="k8s_container" AND resource.labels.namespace_name="default"'
Replace
PROJECT_ID
with the ID of your project.In the output, you can see log entries from the
monitoring-example
Deployment. For example:insertId: 1oa4vhg3qfxidt labels: k8s-pod/app: monitoring-example k8s- pod/pod-template-hash: 7685d96496 logName: projects/.../logs/stdout receiveTimestamp: '2020-11-14T01:24:39.562864735Z' resource: labels: cluster_name: ... container_name: prometheus-example-exporter location: us-west1 namespace_name: default pod_name: monitoring-example-7685d96496-xqfsf project_id: ... type: k8s_container textPayload: | 2020/11/14 01:24:24 Starting to listen on :9090 timestamp: '2020-11-14T01:24:24.358600252Z'
View application metrics
Your example application exposes a custom metric named example_monitoring_up
.
You can view the values of that metric in the Google Cloud console.
In the Google Cloud console, go to the Metrics explorer page.
For Resource type, select
Kubernetes Pod
.For Metric, select
external/prometheus/example_monitoring_up
.In the chart, you can see that
example_monitoring_up
has a repeated value of 1.