In many cases, elevated latency in your application eventually results in 5xx server errors. Because the root cause of both the error and the latency spikes might be the same, apply the following strategies for troubleshooting latency issues:
Scope the latency issue
Define the scope of the issue by asking the following questions:
- Which applications, services, and versions does this issue impact?
- Which specific endpoints on the service does this issue impact?
- Does this impact all clients globally, or a specific subset of clients?
- What is the start and end time of the incident? Consider specifying the timezone.
- What are the specific errors?
- What is the observed latency delta, which is usually specified as an increase at a specific percentile? For example, latency increased by 2 seconds at the 90th percentile.
- How did you measure the latency? In particular, did you measure it at the client, or is it visible in Cloud Logging or in the Cloud Monitoring latency data that the App Engine serving infrastructure provides?
- What are the dependencies of your service, and do any of them experience incidents?
- Did you make any recent code, configuration, or workload changes that triggered this issue?
A service might have its own custom monitoring and logging that you can use to further narrow down the scope of the issue. Defining the scope of the problem will guide you towards the likely root cause and determine your next troubleshooting steps.
Identify the cause
Determine which component in the request path is mostly likely to be causing the latency or errors. The main components in the request path are as follows:
Client --> Internet --> Google Front End (GFE) --> App Engine serving infrastructure --> Service instance
If the previous information doesn't point you to the source of the failure, apply the following strategies while reviewing your service instance's health and performance:
Monitor the App Engine request logs. If you see HTTP status code errors or elevated latency in those logs, the issue likely lies in the instance that runs your service.
If the number of service instances hasn't scaled up to match traffic levels, your instances might be overloaded, resulting in elevated errors and latency.
If you see elevated errors or latency in Cloud Monitoring, the problem might be from the upstream of the load balancer, which records the App Engine metrics. In most cases, this points to a problem in the service instances.
If you see elevated latency or errors in monitoring metrics but not in the request logs, it either indicates a load balancing failure, or a severe instance failure that's preventing the load balancer from routing requests. To distinguish between these cases, look at the request logs before the incident starts. If the request logs show increasing latency before failure, the application instances were beginning to fail before the load balancer stopped routing requests to them.
Troubleshoot
This section describes troubleshooting strategies for elevated latency issues from the following components in the request path:
- Internet
- Google Front End (GFE)
- App Engine serving infrastructure
- Application instance
- Application dependencies
Internet
Your application might encounter latency issues due to poor connectivity or lower bandwidth.
Poor internet connectivity
To determine if the issue is poor internet connectivity, run the following command on your client:
$ curl -s -o /dev/null -w '%{time_connect}\n' <hostname>
The value for time_connect
represents the latency of the client's
connection to the nearest Google Front End.
For slow connections, troubleshoot further using traceroute
to determine which hop on the network causes the delay.
Run tests from clients in different geographical locations. App Engine automatically routes requests to the closest Google data center, which varies based on the client's location.
Low bandwidth
The application might be responding quickly; however, network bottlenecks delay the App Engine serving infrastructure from sending packets across the network quickly, slowing down responses.
Google Front End (GFE)
Your application might encounter latency issues due to incorrect routing, parallel requests sent from HTTP/2 clients, or termination of SSL connections.
Map a client IP to a geographical region
Google resolves the App Engine application's hostname to the closest GFE to the client, based on the client IP address it uses in the DNS lookup. If the client's DNS resolver isn't using the EDNS0 protocol, Google might not route the client requests to the closest GFE.
HTTP/2 head-of-line blocking
HTTP/2 clients sending multiple requests in parallel might see elevated latency due to head-of-line blocking at the GFE. To resolve this issue, the clients must use the QUIC protocol.
SSL termination for custom domains
The GFE might terminate the SSL connection. If you're using a custom domain instead of an appspot.com
domain, SSL
termination requires an extra hop. This might add latency for applications running in some regions. For
more information, see Mapping custom domains.
App Engine serving infrastructure
You might see elevated latency in your application due to service-wide incidents or autoscaling.
Service-wide incidents
Google posts details of a severe service-wide in the Service Health dashboard. However, Google does rollouts gradually, so a service-wide incident is unlikely to affect all your instances at once.
Autoscaling
Elevated latency or errors can result from the following autoscaling scenarios:
Scale up traffic too fast: App Engine autoscaling might not scale your instances as fast as traffic increases, leading to temporary overloading. Typically, overloading occurs when traffic is generated by a computer program instead of end users. To resolve this issue, throttle the system that generates the traffic.
Spikes in traffic: spikes in traffic might cause elevated latency in cases where an autoscaled service needs to scale up more quickly than is possible, without affecting latency. End user traffic doesn't usually cause frequent traffic spikes. If you see traffic spikes, then you should investigate the cause. If a batch system is running at intervals, you might be able to smooth out the traffic or use different scaling settings.
Autoscaler settings: the autoscaler can be configured based on the scaling characteristics of your service. The scaling parameters might become non-optimal during the following scenarios:
App Engine standard environment scaling settings might cause latency if set too aggressively. If you see server responses with the status code
500
and the message "Request was aborted after waiting too long to attempt to service your request" in your logs, it means that the request timed out on the pending queue while waiting for an idle instance.You might see increased pending time with manual scaling even when you have provisioned sufficient instances. We recommend that you don't use manual scaling if your application serves end-user traffic. Manual scaling is better for workloads such as task queues.
Basic scaling minimizes costs at the expense of latency. We recommend that you don't use basic scaling for latency sensitive services.
App Engine's default scaling setting provides optimal latency for most services. If you still see requests with high pending time, specify a minimum number of instances. If you tune the scaling settings to reduce costs by minimizing idle instances, you run the risk of latency spikes if the load increases suddenly.
We recommend that you benchmark performance with the default scaling settings, and then run a new benchmark after each change to these settings.
Deployments
Elevated latency shortly after a deployment indicates that you haven't sufficiently scaled up before migrating traffic. Newer instances might not have warmed up local caches, and they serve more slowly than older instances.
To avoid latency spikes, don't deploy an App Engine service using the same version name as an existing version of the service. If you reuse an existing version name, you won't be able to slowly migrate traffic to the new version. Requests might be slower because App Engine restarts every instance within a short period of time. You also have to redeploy if you want to revert to the previous version.
Application instance
This section describes the common strategies you can apply to your application instances, and source code to optimize performance and reduce latency.
Application code
Issues in application code can be challenging to debug, particularly if they are intermittent or not reproducible.
To resolve issues, do the following:
For diagnosing your issues, we recommend instrumenting your application using logging, monitoring, and tracing. You can also use Cloud Profiler.
Try to reproduce the issue in a local development environment which might allow you to run language-specific debugging tools that might not be possible to run within App Engine.
To better understand how your application fails and what bottlenecks occur, load test your application until failure. Set a maximum instance count, and then gradually increase load until the application fails.
If the latency issue correlates with deployment of a new version of your application code, roll back to determine whether the new version caused the incident. However, if you deploy continuously, then the frequent deployments make it hard to determine whether or not the deployment caused the incident based on time of onset.
Your application might store configuration settings within the Datastore or elsewhere. Create a timeline of configuration changes to determine whether any of these line up with the onset of elevated latency.
Workload change
A workload change might cause elevated latency. Some monitoring metrics that
indicate workload changes include qps
, API usage, and
latency. Check also for changes in request and response sizes.
Memory pressure
If monitoring shows either a saw-tooth pattern in memory usage, or a drop in memory usage that correlates to deployments, a memory leak might be the cause of performance issues. A memory leak might also cause frequent garbage collection that leads to higher latency. If you aren't able to trace this issue to a problem in the code, try provisioning larger instances with more memory.
Resource leak
If an instance of your application shows rising latency that correlates with instance age, then you might have a resource leak that causes performance issues. The latency drops after a deployment is complete. For example, a data structure that gets slower over time due to higher CPU usage might cause any CPU-bound workload to get slower.
Code optimization
To reduce latency on App Engine, optimize code by using the following methods:
Offline work: use Cloud Tasks to prevent user requests from blocking the application waiting for completion of work, such as sending mail.
Asynchronous API calls: ensure that your code isn't blocked waiting for an API call to complete.
Batch API calls: the batch version of API calls is usually faster than sending individual calls.
Denormalize data models: reduce the latency of calls made to the data persistence layer by denormalizing your data models.
Application dependencies
Monitor dependencies of your application to detect if latency spikes correlate with a dependency failure.
A workload change and an increase in traffic might cause a dependency's latency to increase.
Non-scaling dependency
If your application's dependency doesn't scale as the number of App Engine instances scales up, the dependency might overload when traffic increases. An example of a dependency that might not scale is a SQL database. A higher number of application instances leads to a higher number of database connections, which might cause a cascading failure by preventing the database from starting up. To resolve this issue, do the following:
- Deploy a new default version that doesn't connect to the database.
- Shut down the previous default version.
- Deploy a new non-default version that connects to the database.
- Slowly migrate traffic to the new version.
As a preventive measure, design your application to drop requests to the dependency using adaptive throttling.
Caching layer failure
To speed up requests, use multiple caching layers, such as edge caching, Memcache, and in-instance memory. A failure in one of these caching layers might cause a sudden latency increase. For example, a Memcache flush might cause more requests to go to a slower Datastore.