Networking for hybrid and multi-cloud workloads: Reference architectures

Last reviewed 2023-11-13 UTC

This document is part of a series that describes networking and security architectures for enterprises that are migrating data center workloads to Google Cloud.

The series consists of the following documents:

This document discusses networking for a scenario where workloads run in more than one place, such as on-premises and the cloud, or in multiple cloud environments.

Lift-and-shift architecture

The first hybrid workload access scenario is a lift-and-shift architecture.

Establishing private connectivity

You can establish connectivity to on-premises networks using either Dedicated Interconnect or Partner Interconnect. The topology illustrated in figure 1 shows how you can use four Interconnect connections in two different metros and different edge availability domains to achieve 99.99% availability using Dedicated Interconnect. For more information and detailed recommendations, see Hybrid connectivity between an on-premises environment and Google Cloud in the enterprise foundations blueprint.

Configuration of redundant Cloud Interconnect connections for 99.99% availability.

Figure 1. Configuration of redundant Cloud Interconnect connections for 99.99% availability.

Network Connectivity Center lets you use Google's network for data transfer between multiple on-premises or cloud-hosted sites. This approach lets you take advantage of the reach and reliability of Google's network when you need to move data. You can use your existing Cloud VPN, Cloud Interconnect, and SD-WAN router appliances as Network Connectivity Center spokes to support data transfer between the on-premises networks and branch sites, as shown in figure 2.

Network Connectivity Center configuration connecting different on-premises enterprise networks outside of Google Cloud using the Google backbone network.

Figure 2. Network Connectivity Center configuration connecting different on-premises enterprise and other cloud networks outside of Google Cloud using the Google backbone network.

For more information about setting up Network Connectivity Center, see Considerations in the Network Connectivity Center documentation.

SD-WAN appliances

Network Connectivity Center lets you use a third-party network virtual appliance as a Network Connectivity Center spoke to establish connectivity between an external site and your VPC network resources. A router appliance could be a third-party SD-WAN router supported by one of our partners or another virtual appliance that lets you exchange routes with the Cloud Router instance. These appliance-based solutions are in addition to the current site-to-cloud connectivity options that are available with Cloud VPN and Cloud Interconnect as spokes. Figure 3 shows a topology that uses SD-WAN appliances.

Network Connectivity Center configuration using router appliance for integrating your SD-WAN implementation with Google's network.

Figure 3. Network Connectivity Center configuration using router appliance for integrating your SD-WAN implementation with Google's network.

You can use third-party appliances to perform security functions. The security capabilities of the appliance can be integrated in the router appliance as shown in figure 3. It's also a common pattern to use a networking virtual appliance, where traffic from on-premises lands in a transit VPC network, and the appliance establishes the connectivity with the workload VPC network, as shown in figure 4.

For more information about setting up Network Connectivity Center, see Considerations in the Network Connectivity Center documentation.

Hybrid services architecture

As in the intra-cloud use case discussed in Networking for secure intra-cloud access: Reference architectures, Network Connectivity Center enables connectivity from your branch sites and on-premises networks to Google Cloud. Private Service Connect provides private access to Google-managed services or lets you consume other services that are built and deployed in the cloud.

You can also implement network security by using a combination of VPC Service Controls, Google Cloud firewalls, and network virtual appliances, as shown in figure 4.

Networks with an architecture that uses both a lift-and-shift pattern and a hybrid services design pattern that is designed to provide a secure data plane.

Figure 4. Networks with an architecture that uses both a lift-and-shift pattern and a hybrid services design pattern that is designed to provide a secure data plane.

Zero Trust Distributed Architecture

In a hybrid environment, microservices run in service meshes that are deployed across different cloud providers and on-premises environments. You can help secure communication between the microservices by using mTLS and authorization policies. It's a common practice for enterprises to build service meshes in the cloud and to extend the meshes to on-premises. Figure 5 shows an example in which services that are deployed on-premises access the services in the cloud. End-to-end mTLS between the services is enabled using the east-west gateway and SNI-based routing. Anthos Service Mesh helps you secure service-to-service communications, letting you configure authorization policies for the services and deploy certificates and keys that are provided by a managed certificate authority.

Hybrid environments typically feature multiple mesh deployments, such as multiple GKE clusters. An important component in this flow is SNI routing, which is used at the GKE east-west gateway for each cluster. This configuration allows direct-to-workload mTLS routing by the gateway while preserving end-to-end mTLS connectivity.

Zero-trust service mesh deployed across an on-premises environment and Google Cloud.

Figure 5. Zero-trust service mesh deployed across an on-premises environment and Google Cloud.

Enterprises can use Anthos Service Mesh to deploy across clouds. To address challenges in managing identity and certificates across cloud providers, Anthos Service Mesh provides workload identity and an intermediate in-cluster certificate authority (CA), using CA Service (CA Service). The intermediate CA can be chained to an external CA or can be hosted in Google. You can customize CA attributes like region and the signature algorithm, using both enterprise-owned HSMs and Cloud HSM.

Workload identity lets you assign distinct, fine-grained identities and authorization for each microservice in your cluster. Anthos Service Mesh manages the process of issuing certificates and of automatically rotating keys and certificates, without disrupting communications. It also provides a single root of trust across GKE clusters.

Figure 6 shows an architecture that uses Anthos Service Mesh to manage identity and authorization.

Services in the mesh can access Google Cloud services using workload identity federation. This feature lets services act with the authority of a Google service account when they invoke Google Cloud APIs. Workload identity federation also lets the service mesh that's installed in other cloud providers access the Google cloud APIs.

Zero-trust service mesh deployed across clouds.

Figure 6. Zero-trust service mesh deployed across clouds.

You can use Traffic Director to route traffic from the mesh to on-premises or to any other cloud.

For example, you can create services in Traffic Director called on-prem-service and other-cloud-service and add hybrid connectivity network endpoint groups (NEGs) that have endpoints 10.1.0.1:80 and 10.2.0.1:80. Traffic Director then sends the traffic to its clients, which are mesh sidecar proxies that run alongside your applications. Thus, when your application sends a request to the on-prem-service service, the Traffic Director client inspects the request and directs it to the 10.0.0.1:80 endpoint. Figure 7 illustrates this configuration.

Traffic steered from a service mesh using Traffic Director.

Figure 7. Traffic steered from a service mesh using Traffic Director.

You can also incorporate advanced functionality such as weight-based traffic steering, as shown in figure 8. This capability lets you enable critical enterprise needs such as cloud migration. Traffic Director serves as a versatile, globally managed control plane for your service meshes.

Weighted traffic steered using Traffic Director.

Figure 8. Weighted traffic steered using Traffic Director.

What's next