Troubleshooting

This troubleshooting guide can help you solve common issues that you might encounter when using Cloud Interconnect:

For answers to common questions about Cloud Interconnect architecture and features, see the Cloud Interconnect FAQ.

For definitions of terms used on this page, see Cloud Interconnect key terms.

To find logging and monitoring information and view Cloud Interconnect metrics, see Monitoring connections.

General troubleshooting

Check for Cloud Interconnect service disruptions

  • You can check known disruptions on the Google Cloud Status Dashboard. You can also subscribe to the Cloud Incidents JSON feed or RSS feed for push updates.

  • You are notified of maintenance events on that impact your Dedicated Interconnect connections. For details, see Infrastructure maintenance events.

  • You are notified of maintenance events that impact your Partner Interconnect VLAN attachments. Notifications for Partner Interconnect are sent in a similar fashion as the notifications for Dedicated Interconnect connections with some minor differences. For details, see Infrastructure maintenance events.

Can't connect to resources in other regions

By default, Virtual Private Cloud (VPC) networks are regional, meaning that Cloud Router advertises only the subnets in its region. To connect to other regions, set the dynamic routing mode of your VPC network to global so that Cloud Router can advertise all subnets.

For more information, see Dynamic routing mode in the Cloud Router documentation.

Can't reach VMs in a peered VPC network

In this scenario, you have set up a Cloud Interconnect connection between your on-premises network and a VPC network, network A. You have also set up VPC Network Peering between network A and another VPC network, network B. However, you are unable to reach VMs in network B from your on-premises network.

This configuration is unsupported as described in the restrictions of the VPC Network Peering overview.

However, you can use custom IP range advertisements from Cloud Routers in your VPC network to share routes to destinations in the peer network. In addition, you must configure your VPC Network Peering connections to import and export custom routes.

For more information about advertising routes between on-premises networks and VPC peered networks, see the following resources:

Missing subnets in a connection

To advertise all of the available subnets, specify the missing routes with custom route advertisements and advertise all subnet routes between regions with global dynamic routing. To do this, follow these steps:

  1. Specify custom route advertisements on both a Cloud Router and the Border Gateway Protocol (BGP) session. To enter the missing routes, set the following parameters:

    --set-advertisement-groups = ADVERTISED_GROUPS
    --set-advertisement-ranges = ADVERTISED_IP_RANGES
    

    Replace the following:

    • ADVERTISED_GROUPS: a Google-defined group that Cloud Router dynamically advertises; it can have a value of all_subnets, which mimics the default behavior of a Cloud Router
    • ADVERTISED_IP_RANGES: the contents of the new array of IP address ranges; it can have one or more values of your choice

    For more details and examples, see Advertising custom IP ranges.

  2. Enable global dynamic routing mode.

Can't ping Cloud Router

If you can't ping Cloud Router from your on-premises router, find your product in the following table and perform the troubleshooting steps for that product. VMs cannot reach 169.254.0.0/16.

Troubleshooting steps to follow Dedicated Interconnect Partner Interconnect with L3 partner Partner Interconnect with L2 partner
For Partner Interconnect, the Cloud Router might never be pingable because some partners filter traffic to the IP range (169.254.0.0/16) for Cloud Router. For L3 partners, the partner automatically configures BGP. If BGP doesn't come up, contact your partner. Not applicable Yes Not applicable
Verify that your local device has learned the correct MAC address for the Google Cloud side of the connection. For more information, see Troubleshooting ARP. Yes Not applicable Not applicable
Verify that your Cloud Router has an interface and a BGP peer. Cloud Router is not pingable unless the interface and BGP peer are fully configured, including the remote ASN.
  • For Dedicated Interconnect, see BGP session not working.
  • For L2 Partner Interconnect, Google has added the interface and BGP peer for Cloud Router automatically, but you must configure the remote ASN.
Yes Not applicable Yes

Troubleshooting ARP

For Dedicated Interconnect, to find the correct MAC address, run the following gcloud command:

  gcloud compute interconnects get-diagnostics INTERCONNECT_NAME

The googleSystemID contains the MAC address that should be present in your device's ARP table for IP addresses assigned to Cloud Router.

  result:
    links:
    — circuitId: SAMPLE-0
      googleDemarc: sample-local-demarc-0
      lacpStatus:
        googleSystemId: ''
        neighborSystemId: ''
        state: DETACHED
      receivingOpticalPower:
        value: 0.0
      transmittingOpticalPower:
        value: 0.0
    macAddress: 00:00:00:00:00:00

If your device has not learned a MAC address, verify that the correct VLAN ID and IP address are configured on the subinterface.

For Partner Interconnect, if you see the wrong MAC address on your device, verify that you have not bridged the Layer 2 segments of two VLAN attachments. The Google Cloud side of the Cloud Interconnect connection is configured with ip proxy-arp, which replies to all ARP requests and can cause your on-premises router to learn incorrect ARP entries.

Can't create VLAN attachment

If you attempt to create a VLAN attachment for Dedicated Interconnect or Partner Interconnect that violates an organization policy, you see an error message. See the following example error message from running gcloud compute interconnects attachments partner create:

ERROR: (gcloud.compute.interconnects.attachments.partner.create) Could not fetch resource:
- Constraint constraints/compute.restrictPartnerInterconnectUsage violated for projects/example-project. projects/example-project/global/networks/example-network is not allowed to use the Partner Interconnect.

For more information, see Restrict Cloud Interconnect usage and contact your organization administrator.

Share connections with other projects in my organization

Use Shared VPC to share a connection, such as a VLAN attachment or a Dedicated Interconnect connection in a host project.

For more information about setting up a Shared VPC network, see Provisioning Shared VPC.

For more information about configuring attachments in a Shared VPC network, see Options for connecting to multiple VPC networks.

Dedicated Interconnect

Google can't ping you during the connection provisioning process

Check that you're using the correct IP and LACP configuration. During the testing process, Google sends you different test IP configurations for your on-premises router, depending on whether you order a multiple-link bundle or a single-link bundle. Don't configure VLAN attachments for either of these tests.

  • The first set of IP addresses that Google sends is for testing multi-circuit connectivity on each individual link. Configure the test IP addresses on every physical link (without LACP configured), as instructed in the emails that Google sent you. Google must ping all the IP addresses successfully before this first test passes.
  • For the second test, remove all the IP addresses from the first test. Configure the port channel with LACP even if your connection has only one link. Google pings the port channel address. Don't modify the LACP configuration of your port channel after the connection has passed the final test. However, you must remove the test IP address from the port channel interface.
  • Google sends the final production IP address for testing single-circuit connectivity. Configure the IP address on the bundle interface (with LACP configured, either active or passive mode), as instructed in the email that Google sent you. Google must ping the bundle interface IP address successfully before this test passes. Configure the port channel with LACP even if your connection has only one link.

Can't ping Cloud Router

  • Check that you can ping Google's port channel IP address. The IP address is the googleIpAddress value when you view connection details.
  • Check that you have the correct VLAN on your on-premises router's subinterface. The VLAN information should match the information provided by the VLAN attachment.
  • Check that you have the right IP address on your on-premises router's subinterface. When you create a VLAN attachment, it allocates a pair of link-local IP addresses. One is for an interface on a Cloud Router (cloudRouterIpAddress), and the other is for a subinterface on your on-premises router's port channel, not the port channel itself (customerRouterIpAddress).
  • If you're testing the performance of your VLAN attachments, don't ping Cloud Router. Instead, create and then use a Compute Engine virtual machine (VM) instance in your VPC network. For more information, see Performance testing.

BGP session not working

  • Enable multi-hop BGP on your on-premises router with at least two hops.
  • Check that you have the correct neighbor IP address configured on your on-premises router. Use the BGP peer IP address (cloudRouterIpAddress) that the VLAN attachment allocated.
  • Check that the local ASN configuration on your on-premises router matches the peer ASN on the Cloud Router. In addition, check that the local ASN configuration on the Cloud Router matches the peer ASN on your on-premises router.
  • Each attachment is allocated a unique /29 CIDR from 169.254.0.0/16 within your VPC network. One IP address in the /29 CIDR is allocated for the Cloud Router and the other for your on-premises router.

    Check that the correct IP addresses are allocated for your on-premises router interface and its BGP neighbor. A common mistake is to configure a /30 on your on-premises router interface instead of a /29. Google Cloud reserves all other addresses in the /29 CIDR.

    Make sure that you have not allocated any other IP addresses from this CIDR to the VLAN attachment interface on your router.

Can't reach VMs in your VPC network

  • Check that you can ping the port channel and VLAN attachment.
  • Check that your BGP session is active.
  • Check that your on-premises router is advertising and receiving routes.
  • Check that there are no overlaps between your on-premises route advertisements and Google Cloud network ranges.
  • Set the MTU size to the same values for your on-premises router, VPC, and VLAN attachment.

IPv6 traffic is not passing over the VLAN attachment

If you are experiencing difficulty connecting to IPv6 hosts, do the following:

  1. Verify that IPv6 routes are being correctly advertised. If IPv6 routes are not being advertised, see Troubleshoot BGP routes and route selection.
  2. Inspect firewall rules to ensure that you are allowing IPv6 traffic.
  3. Verify that you do not have overlapping IPv6 subnet ranges in your VPC network and your on-premises network. See Check overlapping subnet ranges.
  4. Determine whether you have exceeded any quotas and limits for your learned routes in Cloud Router. If you have exceeded your quota for learned routes, IPv6 prefixes are dropped before IPv4 prefixes. See Check quotas and limits.
  5. Verify that all components related to IPv6 configuration have been configured correctly.
    • The VPC network has enabled the use of internal IPv6 addresses with the --enable-ula-internal-ipv6 flag.
    • The VPC subnet is configured to use the IPV4_IPV6 stack type.
    • The VPC subnet has--ipv6-access-type set to INTERNAL.
    • The Compute Engine VMs on the subnet are configured with IPv6 addresses.
    • The VLAN attachment is configured to use the IPV4_IPV6 stack type.
    • The BGP peer has enabled IPv6 and correct IPv6 next hop addresses are configured for the BGP session.
  6. Run gcloud compute interconnects attachments describe for the IPv6-enabled VLAN attachment.
    • From the output, obtain the IPv6 address listed in cloudRouterIpv6Address.
    • Using ping, verify that the IPv6 address listed in cloudRouterIpv6Address is reachable from your on-premises network.

Can't create a VLAN attachment with the IPv4 and IPv6 (dual-stack) stack type

The creation of the VLAN attachment with the IPv4 and IPv6 (dual-stack) stack-type fails with the following error message:

Cannot create a dual stack interconnect attachment. Dual stack attachment can only be used
with a provisioned interconnect attachments that have Dataplane version 2.

To fix this issue:

  1. Check the All colocation facilities table to see which regions support the creation of attachments on Dataplane v2.

  2. If the region is not listed, recreate the attachment in a supported region.

Can't modify an existing VLAN attachment to use the IPv4 and IPv6 (dual-stack) stack type

The update of an existing VLAN attachment to use the IPv4 and IPv6 (dual-stack) stack-type fails with the following error message:

Cannot create a dual stack interconnect attachment. Dual stack attachment can only
be used with a provisioned interconnect attachments that have Dataplane version 2.

To fix this issue:

  1. Check the Dataplane version of the attachment, and verify that the attachment is using Dataplane version 1. See Check the Dataplane version of an attachment.

  2. Recreate the attachment in a region that supports the creation of all new attachments on Dataplane v2. For a list of regions, see All colocation facilities.

Performance testing over your VLAN attachments

If you need to test the performance of your VLAN attachments, use a VM in your VPC network. Add the performance tools that you require on the VM. Don't use the Cloud Router link-local IP address to test for latency, such as ICMP ping or path MTU. Using Cloud Router can give unpredictable results.

Get diagnostics

To get current, detailed technical information about the Google Cloud side of a Dedicated Interconnect connection on demand, see Get diagnostics.

Partner Interconnect

BGP session not working (Layer 2 connections)

  • Check that your on-premises router has been configured with a BGP session to your Cloud Routers. For more information, see Configuring on-premises routers for Partner Interconnect.
  • Enable multi-hop BGP on your on-premises router with at least two hops.
  • Check that you have the correct neighbor IP address configured on your on-premises router. Use the BGP peer IP address (cloudRouterIpAddress) that the VLAN attachment allocated.
  • Check that the local ASN configuration on your on-premises router matches the peer ASN on the Cloud Router (16550). In addition, check that the local ASN configuration on the Cloud Router matches the peer ASN on your on-premises router.

BGP session not working (Layer 3 connections)

  • Your Cloud Router must be configured with your service provider's ASN. Contact your service provider for assistance.

VLAN attachment down for a Partner Interconnect connection

The status for a VLAN attachment can show as down even if there are no issues with the Google Cloud configuration and the Partner Interconnect connection.

Make sure that you have configured EBGP multihop on your on-premises router to have at least four hops. You can see a sample configuration in Configuring on-premises routers.

Pairing key issue in a Partner Interconnect connection

When you try to set up a Partner Interconnect connection, you might encounter an error message such as "Google - Provider status not available." To fix this issue, follow these steps:

  1. Ensure that the pairing key was generated by the customer-side VLAN attachment (PARTNER type). The key is a long random string that Google uses to identify the attachment. The destination Google Cloud region and edge availability domain are encoded in the pairing key in the following format:

    <random_string>/<region_name>/<domain>
    

    The domain field contains the string any if the VLAN attachment is not restricted to a particular domain or if you don't specify the edge availability domain. For more information about the pairing keys, see pairing key.

  2. Ensure that the edge availability domain of the Partner Interconnect connection matches the domain specified by the pairing key.

Can't ping Cloud Router (Layer 2 connections)

  • Check that you have the correct VLAN attachment on your on-premises router's subinterface. The VLAN attachment information should match the information provided by your service provider.
  • Check that you have the right IP address on your on-premises router's subinterface. After your service provider configures your VLAN attachment, the attachment allocates a pair of link-local IP addresses. One is for an interface on the associated Cloud Router (cloudRouterIpAddress), and the other is for a subinterface on your on-premises router's port channel, not the port channel itself (customerRouterIpAddress).
  • If you're testing the performance of your attachments, don't ping Cloud Router. Instead, create and then use a VM in your VPC network. For more information, see Performance testing.

Loss of optic power on Partner Interconnect connection's port

If there is a loss of optic power on a port, you might encounter one of the following issues:

  • Loss of Layer 3 connectivity (loss of a BGP session) or inability to access your Google Cloud VM instances.
  • Degraded performance of your link bundle. This issue happens if multiple 10GE ports are bundled together and only some of the links in a bundle are functioning.

Loss of optic power on a port means that the hardware is unable to detect a signal from the other end. This can be caused by one of the following issues:

  • A faulty transceiver
  • A faulty transport system
  • A physical fiber issue

To fix this issue, contact your Partner Interconnect and/or circuit provider. They can perform the following steps:

  1. Check that their transceiver is functioning.
  2. Run a hard loop to the Meet-Me Room (MMR) to check if the light levels on their device are working as expected.
  3. Check whether the issues are on their side, or on the Google side. The best way to isolate the interface is to put a bidirectional loop at the demarc. The interfaces on each side will transmit light down to the demarc where it will be looped back to itself. The faulty side will be the side of the demarc that does not come up cleanly.
  4. Clean and reseat all the fiber on their side.

Can't send and learn MED values over an L3 Partner Interconnect connection

If you are using a Partner Interconnect connection where a Layer 3 service provider handles BGP for you, Cloud Router can't learn MED values from your on-premises router or send MED values to that router. This is because MED values can't pass through autonomous systems. Over this type of connection, you can't set route priorities for routes advertised by Cloud Router to your on-premises router. In addition, you can't set route priorities for routes advertised by your on-premises router to your VPC network.

IPv6 traffic isn't working after changing an attachment's stack type to dual stack

  1. View Cloud Router status and verify that status: UP is displayed.

    If BGP isn't up, do the following:

    1. Confirm that your on-premises router (or your partner's router if you are using a Layer 3 partner) is configured with an IPv6 BGP session, and that the session is using the correct IPv6 addresses.

    2. View the BGP session configuration and verify that bgpPeers.enable displays 'TRUE' for your Cloud Router.

    If BGP is up, view Cloud Router routes to verify that the expected IPv6 best_routes are displayed.

    If the expected best_routes aren't displayed, confirm that your on-premises router (or your partner's router if you are using a Layer 3 partner) is configured with the correct IPv6 routes.

All other issues

For additional assistance, contact your service provider. If needed, your service provider will contact Google to fix issues related to the Google side of the network.

HA VPN over Cloud Interconnect

When you deploy HA VPN over Cloud Interconnect, you create two operational tiers:

  • The Cloud Interconnect tier, which includes the VLAN attachments and the Cloud Router for Cloud Interconnect.
  • The HA VPN tier, which includes the HA VPN gateways and tunnels and the Cloud Router for HA VPN.

Each tier requires its own Cloud Router:

  • The Cloud Router for Cloud Interconnect is used exclusively to exchange VPN gateway prefixes between the VLAN attachments. This Cloud Router is used only by the VLAN attachments of the Cloud Interconnect tier. It cannot be used in the HA VPN tier.
  • The Cloud Router for HA VPN exchanges prefixes between your VPC network and your on-premises network. You configure the Cloud Router for HA VPN and its BGP sessions in the same way you would for a regular HA VPN deployment.

You build the HA VPN tier on top of the Cloud Interconnect tier. Therefore, the HA VPN tier requires that the Cloud Interconnect tier, based on either Dedicated Interconnect or Partner Interconnect, is properly configured and operational.

To troubleshoot an HA VPN over Cloud Interconnect deployment, troubleshoot the Cloud Interconnect tier first. After you verify that Cloud Interconnect is functioning correctly, troubleshoot the HA VPN tier.

Can't create an encrypted VLAN attachment

The creation of the encrypted VLAN attachment fails with the following error message:

  Cannot create an interconnect attachment with IPSEC encryption.
  IPSEC encryption can only be used with a provisioned interconnect attachments
  that have Dataplane version 2.
  

To fix this issue, follow these steps:

  1. Check the All colocation facilities table to see which regions support the creation of attachments on Dataplane v2.

  2. If the region is not listed, recreate the attachment in a supported region.

Can't establish BGP session for the Cloud Router for Cloud Interconnect

To detect whether the BGP session associated with the VLAN attachment is down, run the following command:

   gcloud compute routers get-status INTERCONNECT_ROUTER_NAME

Replace INTERCONNECT_ROUTER_NAME with the name of the Cloud Router that you created for the Cloud Interconnect tier of your HA VPN over Cloud Interconnect deployment.

To fix this issue, follow these steps:

  1. Follow the steps in Testing connections and Get diagnostics to make sure that the underlying Cloud Interconnect connection is healthy.
  2. Check that the BGP session interface is pointing to the correct attachment.
  3. Check the IP addresses that are configured for the BGP session interface on both the Cloud Router and the on-premises router.
  4. Check that the ASN numbers are correctly configured on both the Cloud Router and the on-premises router.
  5. Check that the Cloud Interconnect connection and the VLAN attachment are in an admin-enabled state.

Can't establish an HA VPN tunnel

To check the tunnel state, run the command:

gcloud compute vpn-tunnels describe VPN_TUNNEL_NAME

Replace VPN_TUNNEL_NAME with the name of the HA VPN tunnel.

To fix this issue, follow these steps:

  1. Because the VPN tunnel is routed over a VLAN attachment, check that the BGP session associated with the VLAN attachment is established. If not, see Can't establish BGP session for the Cloud Router for Cloud Interconnect.
  2. Check whether the tunnel PSK and ciphers are configured correctly on both the Cloud VPN gateways and the on-premises VPN gateways.
  3. Check that the on-premises router's BGP announcement includes the on-premises VPN gateway addresses. If not, adjust the BGP configuration of the on-premises router to advertise the addresses.
  4. Check that the routes to on-premises VPN gateways have not been dropped due to conflicts with existing BGP routes. If there are conflicts, adjust the VPN gateway addresses or advertised routes.
  5. Check that the BGP announcement from Cloud Router includes the HA VPN gateway addresses. Check this from the on-premises router or by inspecting the advertisedRoutes field of the BGP peer. To view the advertisedRoutes field, run the following command:

    gcloud compute routers get-status INTERCONNECT_ROUTER_NAME
    

    Replace INTERCONNECT_ROUTER_NAME with the name of the Cloud Router that you created for the Cloud Interconnect tier of your HA VPN over Cloud Interconnect deployment.

    If the HA VPN gateway addresses are not advertised, ensure that the VLAN attachments are associated with the encrypted Cloud Interconnect router. Check that each VLAN attachment is configured with the expected regional internal IPv4 addresses (--ipsec-internal-addresses).

Can't establish BGP session for the Cloud Router for HA VPN

To check if the BGP session associated with the VLAN attachment is down, run the command:

gcloud compute routers get-status VPN_ROUTER_NAME

Replace VPN_ROUTER_NAME with the name of the Cloud Router that you created for the HA VPN tier of your HA VPN over Cloud Interconnect deployment.

To fix this issue, follow these steps:

  1. Because the BGP traffic is routed over the VPN tunnel, check that the VPN tunnel is established. If not, follow the steps in Can't establish an HA VPN tunnel.
  2. Check that the BGP session interface for the Cloud Router is pointing to the correct VPN tunnel.
  3. Check that the IP addresses of the BGP session interface are configured correctly on both the Cloud Router and the on-premises VPN device.
  4. Check that the ASN numbers are configured correctly on both the Cloud Router and the on-premises router.

VPC traffic is not reaching on-premises networks or vice versa

The traffic generated from a VM, such as from ping or iperf, can't reach the on-premises network, or the on-premises network can't reach the traffic generated from a VM.

To fix this issue, follow these steps:

  1. Because the VM traffic is routed over the VPN tunnel, make sure that the route from the VM to the VPN tunnel is being sent by the Cloud Router.

  2. Check that the Cloud Router session for HA VPN is established. If not, see Can't establish BGP session for the Cloud Router for HA VPN.

Packet loss or low throughput

Traffic from VMs on VPC networks to on-premises networks or traffic from on-premises networks to VPC networks is dropped.

You observe packet loss or low throughput through ping, iperf, and other network monitoring tools.

To fix this issue, follow these steps:

  1. Check to see if the VLAN attachment is overloaded with traffic. If so, spread the traffic over more VLAN attachments or update the attachment's capacity.
  2. Check whether HA VPN is overloaded with traffic. If so, create additional VPN tunnels over the VLAN attachment to redistribute the traffic.
  3. Make sure that there are no unexpected or sudden spikes in traffic or bursty traffic. TCP streams might be affected by other streams, such as bursty UDP traffic.
  4. Check whether packets exceeding the tunnel MTU are fragmented. Ensure that the MTU is set properly with your VLAN attachments, and check that UDP traffic isn't being MSS clamped.

Unable to delete an encrypted VLAN attachment

You receive the following error when you try to delete an encrypted VLAN attachment for Dedicated Interconnect or Partner Interconnect:

ResourceInUseByAnotherResourceException

To fix this issue, make sure that you have first deleted all the HA VPN gateways and tunnels associated with the encrypted VLAN attachment. For more information, see Delete HA VPN over Cloud Interconnect.

Mismatched IP address types across encrypted VLAN attachments

When you try to create an HA VPN gateway for use in an HA VPN over Cloud Interconnect deployment, you get the following error:

One of the VLAN attachments has private IP address type, while the other one
has public IP address type; they must have same IP address type.

This error occurs when you specify two encrypted VLAN attachments for an HA VPN gateway, and they have different IP addressing schemes for the HA VPN tunnel interfaces. The IP address types must match across both VLAN attachments.

During VPN gateway creation, specify VLAN attachments that use either both private IP addresses or both public IP addresses. If you get this error when creating a VPN gateway, retry recreating the VLAN attachment with the correct address type.

Missing VLAN attachment from HA VPN gateway interface

If being deployed for HA VPN over Cloud Interconnect, an HA VPN gateway must have an encrypted VLAN attachment specified for both its interfaces.

If you specify only one VLAN attachment, the following error appears:

VPN Gateway should have VLAN attachment specified in both interfaces or in none.

To fix this issue, create the HA VPN gateway and specify two VLAN attachments.

Invalid VLAN attachment type

Encrypted VLAN attachments must have the type DEDICATED or PARTNER.

If you specify an invalid type for an encrypted VLAN attachment, the following error message appears:

VLAN attachment should have type DEDICATED or PARTNER.

To fix this issue, only create encrypted VLAN attachments with the type DEDICATED or PARTNER.

Wrong MTU value for VLAN attachment

When creating an encrypted VLAN attachment for Dedicated Interconnect, the following error message appears:

Wrong MTU value [mtuValue] for VLAN attachment.
The supported MTU for IPsec packets for HA VPN over Cloud Interconnect is 1440.

To fix this issue, recreate the attachment with the correct value of 1440, which is the default value.

VLAN attachments have a different type

When specifying encrypted VLAN attachments for your HA VPN gateway interfaces, the following error message appears:

VLAN attachments should both have same type DEDICATED or PARTNER.
But found one DEDICATED and one PARTNER.

To fix this issue, specify two VLAN attachments that are the same type, either DEDICATED or PARTNER.

Dedicated VLAN attachments are not in the same metropolitan area

When specifying encrypted VLAN attachments for your HA VPN gateway interfaces, the following error message appears:

Dedicated VLAN attachments should be in the same metropolitan area.

To fix this issue, create two encrypted VLAN attachments for Dedicated Interconnect in the same metropolitan area.

HA VPN gateway is not in the same network as VLAN attachment

When specifying encrypted VLAN attachments for your HA VPN gateway interfaces, the following error message appears:

VPN Gateway should be in the same network as VLAN attachment.
VLAN attachment network: [networkName], VPN gateway network: [networkName]

To fix this issue, create the HA VPN gateway in the same network as the encrypted VLAN attachments.

Wrong encryption type for VLAN attachment

When specifying encrypted VLAN attachments for your HA VPN gateway interfaces, the following error message appears:

Wrong encryption type for VLAN attachment [interconnectAttachmentName],
required IPSEC.

To fix this issue, specify only encrypted VLAN attachments that are configured with the encryption type IPSEC. If necessary, create encrypted VLAN attachments.

VLAN attachment zone does not match interfaceId

When specifying encrypted VLAN attachments for your HA VPN gateway interfaces, the following error message appears:

VLAN attachment zone should match interfaceId: interface 0 - zone1,
interface 1 - zone2, but found interface [interfaceId] - [zone] for
[interconnectAttachmentName].

The first HA VPN gateway interface (interface 0) must match the encrypted VLAN attachment from zone 1, and the second interface (interface 1) must match the encrypted VLAN attachment from zone 2.

To fix this issue, specify encrypted VLAN attachments from the zones matched correctly to the HA VPN gateway interfaces.

VPN gateway is not in the same region as VLAN attachment

When specifying encrypted VLAN attachments for your HA VPN gateway interfaces, the following error message appears:

VPN Gateway should be in the same region as VLAN attachment.
VLAN attachment region: [region], VPN gateway region: [region].

To fix this issue, create HA VPN gateways and encrypted VLAN attachments in the same region.

Partner VLAN attachment is not in active state

When specifying encrypted VLAN attachments for Partner Interconnect for your HA VPN gateway interfaces, the following error message appears:

Interconnect Attachments [name] must be in active state.

You must activate the VLAN attachments for Partner Interconnect before you associate them with HA VPN gateway interfaces.

For more information, see Activate connections.

Wrong type of Cloud Router specified for VLAN attachment

When you try to create an encrypted VLAN attachment, the following error message appears:

Router must be an encrypted interconnect router.

To fix this issue, create a Cloud Router with the --encrypted-interconnect-router flag. This Cloud Router is used exclusively for HA VPN over Cloud Interconnect.

Then, create the encrypted VLAN attachment and provide the encrypted Cloud Router.

Cross-Cloud Interconnect

This section describes issues that can come up with Cross-Cloud Interconnect.

Google supports the connection up to the point where it reaches the network of your other cloud service provider. Google does not guarantee uptime from the other cloud service provider and cannot create a support ticket on your behalf. However, with your permission Google Cloud Support can communicate directly with your other cloud provider's support team to expedite issue resolution.

Mismatches between redundant ports

After you order Cross-Cloud Interconnect ports, you give Google information about how you want the ports connected to your remote cloud ports. You also use this information later, when you configure your resources. If you have connectivity issues, one problem might be that you didn't use the correct connectivity details.

For example, you might provide instructions like the following:

  • Connect gcp-1 to azure-1.

  • Connect gcp-2 to azure-2.

However, when configuring your resources, you might assume that the ports are connected as follows:

  • Connect gcp-1 to azure-2.

  • Connect gcp-2 to azure-1.

This condition might be detectable through examination of the ARP cache. However, a simpler solution is to go to the remote cloud and try swapping the IP address ranges assigned to the two BGP peers.

For example, suppose azure-1 has a VLAN attachment peering on 169.254.0.2 and azure-2 has a VLAN attachment peering on 169.254.99.2. Swap the IP address ranges so that the azure-1 attachment uses 169.254.99.2 and the azure-2 attachment uses 169.254.0.2.

If you used different VLAN IDs for the attachment on each port, you would also need to swap the VLAN IDs used by the attachments. For Azure, this scenario is impossible because it requires using the same VLAN ID on both ports for each attachment.

VLAN IDs

Sometimes connectivity problems can be traced to mistakes with VLAN ID values. VLAN ID is a field on your Cross-Cloud Interconnect VLAN attachment.

Azure

Azure requires that VLAN IDs be allocated identically on both ports of a pair. When you create your VLAN attachment, the Google Cloud console enforces this requirement. However, if you set up the attachments by using the Google Cloud CLI or the API, it's possible to allocate different VLAN IDs. This risk is especially great if you let VLAN IDs be automatically assigned when you create the attachment. If you don't explicitly set the ID, it's automatically assigned.

AWS

When connecting to AWS, it's OK to use automatic assignment of VLAN IDs. However, when configuring your AWS resources, you must manually configure the VLAN IDs to match the ones that Google Cloud automatically assigned. Also, it's important to not confuse which VLAN ID should be used for each port. If the wrong VLAN ID is configured on a port, the virtual routers are unable to communicate.

Verify connectivity

If you haven't already, follow the steps for verifying connectivity between Google Cloud and your remote cloud network: