Google Cloud global external proxy Network Load Balancers let you use a single IP address for all users around the world. Global external proxy Network Load Balancers automatically route traffic to backend instances that are closest to the user.
This page contains instructions for setting up a global external proxy Network Load Balancer with a target TCP proxy and VM instance group backends. Before you start, read the External proxy Network Load Balancer overview for detailed information about how these load balancers work.
Setup overview
This example demonstrates how to set up an external proxy Network Load Balancer for
a service that exists in two regions: Region A and Region B.
For purposes of the example, the service is a set of Apache
servers configured to respond on port 110
. Many browsers don't
allow port 110
, so the testing section uses curl
.
In this example, you configure the following:
- Four instances distributed between two regions
- Instance groups, which contain the instances
- A health check for verifying instance health
- A backend service, which monitors the instances and prevents them from exceeding configured usage
- The target TCP proxy
- An external static IPv4 address and forwarding rule that sends user traffic to the proxy
- An external static IPv6 address and forwarding rule that sends user traffic to the proxy
- A firewall rule that allows traffic from the load balancer and health checker to reach the instances
After the load balancer is configured, you test the configuration.
Permissions
To follow this guide, you must be able to create instances and modify a network in a project. You must be either a project owner or editor, or you must have all of the following Compute Engine IAM roles:
Task | Required Role |
---|---|
Create networks, subnets, and load balancer components | Network Admin |
Add and remove firewall rules | Security Admin |
Create instances | Compute Instance Admin |
For more information, see the following guides:
Configure the network and subnets
To create the example network and subnet, follow these steps.
Console
To support both IPv4 and IPv6 traffic, use the following steps:
In the Google Cloud console, go to the VPC networks page.
Click Create VPC network.
Enter a Name for the network.
Optional: If you want to configure internal IPv6 address ranges on subnets in this network, complete these steps:
- For VPC network ULA internal IPv6 range, select Enabled.
For Allocate internal IPv6 range, select Automatically or Manually.
If you select Manually, enter a
/48
range from within thefd20::/20
range. If the range is in use, you are prompted to provide a different range.
For the Subnet creation mode, choose Custom.
In the New subnet section, configure the following fields:
- In the Name field, provide a name for the subnet.
- In the Region field, select a region.
- For IP stack type, select IPv4 and IPv6 (dual-stack).
In the IP address range field, enter an IP address range. This is the primary IPv4 range for the subnet.
Although you can configure an IPv4 range of addresses for the subnet, you cannot choose the range of the IPv6 addresses for the subnet. Google provides a fixed size (
/64
) IPv6 CIDR block.For IPv6 access type, select External.
Click Done.
To add a subnet in a different region, click Add subnet and repeat the previous steps.
Click Create.
To support IPv4 traffic only, use the following steps:
In the Google Cloud console, go to the VPC networks page.
Click Create VPC network.
In the Name field, enter a name for the network.
For the Subnet creation mode, choose Custom.
In the New subnet section, configure the following:
- In the Name field, provide a name for the subnet.
- In the Region field, select a region.
- For IP stack type, select IPv4 (single-stack).
- In the IP address range field, enter the primary IPv4 range for the subnet.
Click Done.
To add a subnet in a different region, click Add subnet and repeat the previous steps.
Click Create.
gcloud
Create the custom mode VPC network:
gcloud compute networks update NETWORK \ [ --enable-ula-internal-ipv6 [ --internal-ipv6-range=ULA_IPV6_RANGE ]] \ --switch-to-custom-subnet-mode
Within the network, create a subnet for backends.
For IPv4 and IPv6 traffic, use the following command to update a subnet:
gcloud compute networks subnets update SUBNET \ --stack-type=IPV4_IPv6 \ --ipv6-access-type=EXTERNAL \ --network=NETWORK \ --region=REGION_A
gcloud compute networks subnets update SUBNET_B \ --stack-type=IPV4_IPv6 \ --ipv6-access-type=EXTERNAL \ --network=NETWORK \ --region=REGION_B
For IPv4 traffic only, use the following command:
gcloud compute networks subnets update SUBNET \ --network=NETWORK \ --stack-type=IPV4_ONLY \ --range=10.1.2.0/24 \ --region=REGION_A
gcloud compute networks subnets update SUBNET_B \ --stack-type=IPV4_ONLY \ --ipv6-access-type=EXTERNAL \ --network=NETWORK \ --region=REGION_B
Replace the following:
NETWORK
: a name for the VPC networkULA_IPV6_RANGE
: a/48
prefix from within thefd20::/20
range used by Google for internal IPv6 subnet ranges. If you don't use the--internal-ipv6-range
flag, Google selects a/48
prefix for the networkSUBNET
: a name for the subnet
REGION_A
orREGION_B
: the name of the region
Configure instance group backends
This section shows how to create basic instance groups, add instances to them, and then add those instances to a backend service with a health check. A production system would normally use managed instance groups based on instance templates, but this configuration is quicker for initial testing.
Configure instances
For testing purposes, install Apache on four instances, two in each of two instance groups. Typically, external proxy Network Load Balancers aren't used for HTTP traffic, but Apache software is commonly used for testing.
In this example, the instances are created with the tag tcp-lb
. This tag is
used later by the firewall rule.
Console
Create instances
In the Google Cloud console, go to the VM instances page.
Click Create instance.
Set Name to
vm-a1
.Set the Region to
REGION_A
.Set the Zone to
ZONE_A
.Click Advanced options.
Click Networking and configure the following field:
- For Network tags, enter
tcp-lb
,allow-health-check-ipv6
.
To support both IPv4 and IPv6 traffic, use the following steps:
- In the Network interfaces section, click
- Network:
NETWORK
- Subnet:
SUBNET
- IP stack type: IPv4 and IPv6 (dual-stack)
Edit, and make the
following changes:
- Network:
- Click Done
- For Network tags, enter
Click Management. Enter the following script into the Startup script field.
sudo apt-get update sudo apt-get install apache2 -y sudo sed -i '/Listen 80/c\Listen 110' /etc/apache2/ports.conf sudo service apache2 restart echo '<!doctype html><html><body><h1>vm-a1</h1></body></html>' | sudo tee /var/www/html/index.html
Click Create.
Create
vm-a2
with the same settings, except with the following script in the Startup script field:sudo apt-get update sudo apt-get install apache2 -y sudo sed -i '/Listen 80/c\Listen 110' /etc/apache2/ports.conf sudo service apache2 restart echo '<!doctype html><html><body><h1>vm-a2</h1></body></html>' | sudo tee /var/www/html/index.html
Create
vm-b1
with the same settings, except with Region set toREGION_B
and Zone set toZONE_B
. Enter the following script in the Startup script field:sudo apt-get update sudo apt-get install apache2 -y sudo sed -i '/Listen 80/c\Listen 110' /etc/apache2/ports.conf sudo service apache2 restart echo '<!doctype html><html><body><h1>vm-b1</h1></body></html>' | sudo tee /var/www/html/index.html
Create
vm-b2
with the same settings, except with Region set toREGION_B
and Zone set toZONE_B
. Enter the following script in the Startup script field:sudo apt-get update sudo apt-get install apache2 -y sudo sed -i '/Listen 80/c\Listen 110' /etc/apache2/ports.conf sudo service apache2 restart echo '<!doctype html><html><body><h1>vm-b2</h1></body></html>' | sudo tee /var/www/html/index.html
gcloud
Create
vm-a1
in zoneZONE_A
gcloud compute instances create vm-a1 \ --image-family debian-12 \ --image-project debian-cloud \ --tags tcp-lb \ --zone ZONE_A \ --metadata startup-script="#! /bin/bash sudo apt-get update sudo apt-get install apache2 -y sudo sed -i '/Listen 80/c\Listen 110' /etc/apache2/ports.conf sudo service apache2 restart echo '<!doctype html><html><body><h1>vm-a1</h1></body></html>' | tee /var/www/html/index.html EOF"
Create
vm-a2
in zoneZONE_A
gcloud compute instances create vm-a2 \ --image-family debian-12 \ --image-project debian-cloud \ --tags tcp-lb \ --zone ZONE_A \ --metadata startup-script="#! /bin/bash sudo apt-get update sudo apt-get install apache2 -y sudo sed -i '/Listen 80/c\Listen 110' /etc/apache2/ports.conf sudo service apache2 restart echo '<!doctype html><html><body><h1>vm-a2</h1></body></html>' | tee /var/www/html/index.html EOF"
Create
vm-b1
in zoneZONE_B
gcloud compute instances create vm-b1 \ --image-family debian-12 \ --image-project debian-cloud \ --tags tcp-lb \ --zone ZONE_B \ --metadata startup-script="#! /bin/bash sudo apt-get update sudo apt-get install apache2 -y sudo sed -i '/Listen 80/c\Listen 110' /etc/apache2/ports.conf sudo service apache2 restart echo '<!doctype html><html><body><h1>vm-b1</h1></body></html>' | tee /var/www/html/index.html EOF"
Create
vm-b2
in zoneZONE_B
gcloud compute instances create vm-b2 \ --image-family debian-12 \ --image-project debian-cloud \ --tags tcp-lb \ --zone ZONE_B \ --metadata startup-script="#! /bin/bash sudo apt-get update sudo apt-get install apache2 -y sudo sed -i '/Listen 80/c\Listen 110' /etc/apache2/ports.conf sudo service apache2 restart echo '<!doctype html><html><body><h1>vm-b2</h1></body></html>' | tee /var/www/html/index.html EOF"
Create instance groups
In this section you create an instance group in each zone and add the instances.
Console
In the Google Cloud console, go to the Instance groups page.
Click Create instance group.
Click New unmanaged instance group.
Set the Name to
instance-group-a
.Set the Zone to
ZONE_A
.Under Port mapping, click Add port. A load balancer sends traffic to an instance group through a named port. Create a named port to map the incoming traffic to a specific port number.
- Set Port name to
tcp110
. - Set Port numbers to
110
.
- Set Port name to
Under VM instances, select
vm-a1
andvm-a2
.Leave the other settings as they are.
Click Create.
Repeat the steps, but set the following values:
- Name:
instance-group-b
- Region:
REGION_B
- Zone:
ZONE_B
- Port name:
tcp110
- Port numbers:
110
- Instances: vm-b1 and vm-b2.
- Name:
gcloud
Create the
instance-group-a
instance group.gcloud compute instance-groups unmanaged create instance-group-a \ --zone ZONE_A
Create a named port for the instance group.
gcloud compute instance-groups set-named-ports instance-group-a \ --named-ports tcp110:110 \ --zone ZONE_A
Add
vm-a1
andvm-a2
toinstance-group-a
.gcloud compute instance-groups unmanaged add-instances instance-group-a \ --instances vm-a1,vm-a2 \ --zone ZONE_A
Create the
us-ig2
instance group.gcloud compute instance-groups unmanaged create instance-group-b \ --zone ZONE_B
Create a named port for the instance group.
gcloud compute instance-groups set-named-ports instance-group-b \ --named-ports tcp110:110 \ --zone ZONE_B
Add
vm-b1
andvm-b2
to instance-group-bgcloud compute instance-groups unmanaged add-instances instance-group-b \ --instances vm-b1,vm-b2 \ --zone ZONE_B
You now have one instance group per region. Each instance group has two VM instances.
Create a firewall rule for the external proxy Network Load Balancer
Configure the firewall to allow traffic from the load balancer and health checker to the instances. In this case, we will open TCP port 110. The health check will use the same port. Since the traffic between the load balancer and your instances uses IPv4, only IPv4 ranges need be opened.
Console
In the Google Cloud console, go to the Firewall policies page.
Click Create firewall rule.
In the Name field, enter
allow-tcp-lb-and-health
.Select a network.
Under Targets, select Specified target tags.
Set Target tags to
tcp-lb
.Set Source filter to IPv4 ranges.
Set Source IPv4 ranges to
130.211.0.0/22
,35.191.0.0/16
.Under Protocols and ports, set Specified protocols and ports to
tcp:110
.Click Create.
gcloud
gcloud compute firewall-rules create allow-tcp-lb-and-health \ --source-ranges 130.211.0.0/22,35.191.0.0/16 \ --target-tags tcp-lb \ --allow tcp:110
Create an IPv6 health check firewall rule
Ensure that you have an ingress rule that is applicable to the instances
being load balanced and that allows traffic from the Google Cloud
health checking systems (2600:2d00:1:b029::/64
). This example uses the
target tag allow-health-check-ipv6
to identify the VM instances to which
it applies.
Without this firewall rule, the default deny ingress rule blocks incoming IPv6 traffic to the backend instances.
Console
In the Google Cloud console, go to the Firewall policies page.
To allow IPv6 subnet traffic, click Create firewall rule again and enter the following information:
- Name:
fw-allow-lb-access-ipv6
- Select the network.
- Priority:
1000
- Direction of traffic: ingress
- Targets: Specified target tags
- Target tags:
allow-health-check-ipv6
- Source filter: IPv6 ranges
- Source IPv6 ranges:
2600:2d00:1:b029::/64
,2600:2d00:1:1::/64
- Protocols and ports: Allow all
- Name:
Click Create.
gcloud
Create the fw-allow-lb-access-ipv6
firewall rule to allow communication
with the subnet:
gcloud compute firewall-rules create fw-allow-lb-access-ipv6 \ --network=NETWORK \ --action=allow \ --direction=ingress \ --target-tags=allow-health-check-ipv6 \ --source-ranges=2600:2d00:1:b029::/64,2600:2d00:1:1::/64 \ --rules=all
Configure the load balancer
Console
Start your configuration
In the Google Cloud console, go to the Load balancing page.
- Click Create load balancer.
- For Type of load balancer, select Network Load Balancer (TCP/UDP/SSL) and click Next.
- For Proxy or passthrough, select Proxy load balancer and click Next.
- For Public facing or internal, select Public facing (external) and click Next.
- For Global or single region deployment, select Best for global workloads and click Next.
- For Load balancer generation, select Global external proxy Network Load Balancer and click Next.
- Click Configure.
Basic configuration
Set the Name to my-tcp-lb
.
Backend configuration
- Click Backend configuration.
- Under Backend type, select Instance groups.
- Under Protocol, select TCP.
- In the IP address selection policy list, select Prefer IPv6.
- Configure the first backend:
- Under New backend, select instance group
instance-group-a
. - Retain the remaining default values.
- Under New backend, select instance group
- Configure the second backend:
- For IP stack type, select
IPv4 and IPv6 (dual-stack)
. - Click Add backend.
- Select instance group
instance-group-b
. - Under Port numbers, delete
80
and add110
.
To support both IPv4 and IPv6 traffic:
- For IP stack type, select
- Configure the health check:
- Under Health check, select Create health check.
- Set the health check Name to
my-tcp-health-check
. - Under Protocol, select TCP.
- Set Port to
110
. - Retain the remaining default values.
- Click Save and continue.
- In the Google Cloud console, verify that there is a check mark next to Backend configuration. If not, double-check that you have completed all of the steps.
Frontend configuration
- Click Frontend configuration.
- Add the first forwarding rule:
- Enter a Name of
my-tcp-lb-forwarding-rule
. - Under Protocol, select TCP.
- Under IP address, select Create IP address:
- Enter a Name of
tcp-lb-static-ip
. - Click Reserve.
- Enter a Name of
- Set Port to
110
. - In this example, don't enable the Proxy Protocol because it doesn't work with the Apache HTTP Server software. For more information, see Proxy protocol.
- Click Done.
- Enter a Name of
In the Google Cloud console, verify that there is a check mark next to Frontend configuration. If not, double-check that you have completed all the previous steps.
Review and finalize
- Click Review and finalize.
- Review your load balancer configuration settings.
- Optional: Click Equivalent code to view the REST API request that will be used to create the load balancer.
- Click Create.
gcloud
- Create a health check.
gcloud compute health-checks create tcp my-tcp-health-check --port 110
- Create a backend service.
gcloud beta compute backend-services create my-tcp-lb \ --load-balancing-scheme EXTERNAL_MANAGED \ --global-health-checks \ --global \ --protocol TCP \ --ip-address-selection-policy=PREFER_IPV6 \ --health-checks my-tcp-health-check \ --timeout 5m \ --port-name tcp110
Alternatively, you can configure encrypted communication from the load balancer to the instances with
--protocol SSL
. Add instance groups to your backend service.
gcloud beta compute backend-services add-backend my-tcp-lb \ --global \ --instance-group instance-group-a \ --instance-group-zone ZONE_A \ --balancing-mode UTILIZATION \ --max-utilization 0.8
gcloud beta compute backend-services add-backend my-tcp-lb \ --global \ --instance-group instance-group-b \ --instance-group-zone ZONE_B \ --balancing-mode UTILIZATION \ --max-utilization 0.8
- Configure a target TCP proxy. If you want to turn on the
proxy header, set it to
PROXY_V1
instead ofNONE
.gcloud beta compute target-tcp-proxies create my-tcp-lb-target-proxy \ --backend-service my-tcp-lb \ --proxy-header NONE
- Reserve global static IPv4 and IPv6 addresses.
Your customers can use these IP addresses to reach your load balanced service.
gcloud compute addresses create tcp-lb-static-ipv4 \ --ip-version=IPV4 \ --global
gcloud compute addresses create tcp-lb-static-ipv6 \ --ip-version=IPV6 \ --global
- Configure global forwarding rules for the two addresses.
gcloud beta compute forwarding-rules create my-tcp-lb-ipv4-forwarding-rule \ --load-balancing-scheme EXTERNAL_MANAGED \ --global \ --target-tcp-proxy my-tcp-lb-target-proxy \ --address tcp-lb-static-ipv4 \ --ports 110
Test the load balancer
Get the load balancer's IP address.
To get the IPv4 address, run the following command:
gcloud compute addresses describe tcp-lb-static-ipv4
To get the IPv6 address, run the following command:
gcloud compute addresses describe tcp-lb-static-ipv6
Send traffic to your load balancer by running the following command. Replace
LB_IP_ADDRESS
with your load balancer's IPv4 or IPv6 address.curl -m1 LB_IP_ADDRESS:110
For example, if the assigned IPv6 address is
[2001:db8:1:1:1:1:1:1/96]:110
, the command should look like:curl -m1 http://[2001:db8:1:1:1:1:1:1]:110
If you can't reach the load balancer, try the steps described under Troubleshooting your setup.
Additional configuration options
This section expands on the configuration example to provide alternative and additional configuration options. All of the tasks are optional. You can perform them in any order.
PROXY protocol for retaining client connection information
The proxy Network Load Balancer ends TCP connections from the client and creates new connections to the instances. By default, the original client IP and port information is not preserved.
To preserve and send the original connection information to your instances, enable PROXY protocol version 1. This protocol sends an additional header that contains the source IP address, destination IP address, and port numbers to the instance as a part of the request.
Make sure that the proxy Network Load Balancer's backend instances are running servers that support PROXY protocol headers. If the servers are not configured to support PROXY protocol headers, the backend instances return empty responses.
If you set the PROXY protocol for user traffic, you can also set it for your
health checks. If you are checking health and serving
content on the same port, set the health check's --proxy-header
to match your
load balancer setting.
The PROXY protocol header is typically a single line of user-readable text in the following format:
PROXY TCP4 <client IP> <load balancing IP> <source port> <dest port>\r\n
The following example shows a PROXY protocol:
PROXY TCP4 192.0.2.1 198.51.100.1 15221 110\r\n
In the preceding example, the client IP is 192.0.2.1
, the load balancing IP is
198.51.100.1
, the client port is 15221
, and the destination port is 110
.
When the client IP is not known, the load balancer generates a PROXY protocol header in the following format:
PROXY UNKNOWN\r\n
Update PROXY protocol header for target proxy
The example load balancer setup on this page shows you how to enable the PROXY protocol header while creating the proxy Network Load Balancer. Use these steps to change the PROXY protocol header for an existing target proxy.
Console
In the Google Cloud console, go to the Load balancing page.
- Click Edit for your load balancer.
- Click Frontend configuration.
- Change the value of the Proxy protocol field to On.
- Click Update to save your changes.
gcloud
In the following command, edit the --proxy-header
field and set it to
either NONE
or PROXY_V1
depending on your requirement.
gcloud compute target-tcp-proxies update TARGET_PROXY_NAME \ --proxy-header=[NONE | PROXY_V1]
Configure session affinity
The example configuration creates a backend service without session affinity.
These procedures show you how to update a backend service for the example load balancer so that the backend service uses client IP affinity or generated cookie affinity.
When client IP affinity is enabled, the load balancer directs a particular client's requests to the same backend VM based on a hash created from the client's IP address and the load balancer's IP address (the external IP address of an external forwarding rule).
Console
To enable client IP session affinity:
In the Google Cloud console, go to the Load balancing page.
Click Backends.
Click my-tcp-lb (the name of the backend service you created for this example) and click Edit.
On the Backend service details page, click Advanced configuration.
Under Session affinity, select Client IP from the menu.
Click Update.
gcloud
Use the following Google Cloud CLI command to update the my-tcp-lb
backend
service, specifying client IP session affinity:
gcloud compute backend-services update my-tcp-lb \ --global \ --session-affinity=CLIENT_IP
API
To set client IP session affinity, make a PATCH
request to the
backendServices/patch
method.
PATCH https://www.googleapis.com/compute/v1/projects/[PROJECT_ID]/global/us-west1/backendServices/my-tcp-lb
{
"sessionAffinity": "CLIENT_IP"
}
Enable connection draining
You can enable connection draining on backend services to ensure minimal interruption to your users when an instance that is serving traffic is terminated, removed manually, or removed by an autoscaler. To learn more about connection draining, read the Enabling connection draining documentation.
What's next
- Convert proxy Network Load Balancer to IPv6
- External proxy Network Load Balancer overview
- Proxy Network Load Balancer logging and monitoring
- Convert to dual-stack backends
- Clean up a load balancing setup