Kf integrates tightly with Kubernetes and Istio to provide robust network policy enforcement.
By default, Kf workloads are run in the Kubernetes cluster and resolve addresses using Kubernetes DNS. This DNS resolver will first attempt to resolve addresses within the cluster, and only if none are found will attempt external resolution.
Each Kubernetes Pod is executed on a Node, a physical or virtual machine responsible for managing the container images that make up a Pod. Nodes exist on a physical or virtual network.
Together, these form a hierarchy of systems you can apply network policies. These are listed below from least to most granular.
Network level policies
Workload protection starts with the network your GKE cluster is installed on.
If you're running Kf on a GKE cluster on GCP, Kf recommends:
- Placing your GKE cluster on a Virtual Private Cloud (VPC) network.
- With Private Google Access enabled.
- Using Cloud NAT to control egress.
Node level policies
You can set up policies for containers running on the Node using Kubernetes NetworkPolicies. These are the closest mapping to Cloud Foundry network policies that exist in Kubernetes.
NetworkPolicies are backed by a Kubernetes add-on. If you set up your own GKE cluster, you will need to enable NetworkPolicy enforcement.
Kf labels Apps with
kf.dev/networkpolicy=app and builds with
This allows you to target NetworkPolicies directly at Pods running Apps or Builds.
Each Kf Space creates two NetworkPolicies to start with, one targeting Apps and
one targeting Builds. You can change the configuration on the Space's
These fields can be set to one of the following values:
||Allows all traffic.|
||Denies all traffic.|
By default Kf uses a permissive network policy. This allows the following functionality that Kf uses:
- North/South routing to the cluster ingress gateway
- Egress to the Internet e.g. to fetch Buildpacks
- East/West routing between Apps
- Access to the Kubernetes DNS server
- Access to container registries
- Direct access to the VPC network
- Access to Google services like Cloud Logging
- Access to the Workload Identity server for automatic rotating credentials
Service mesh policies
If you need fine-grained networking control, authentication, authorization, and observability you can apply policies using Anthos Service Mesh.
A service mesh is an infrastructure layer that enables managed, observable and secure communication across your services, letting you create robust enterprise applications made up of many microservices on your chosen infrastructure.
You can see the list of supported features here.