Now that you have a continuously running, replicated application you can expose it on a network. Before discussing the Kubernetes approach to networking, it is worthwhile to contrast it with the “normal” way networking works with Docker.
By default, Docker uses host-private networking, so containers can talk to other containers only if they are on the same machine. In order for Docker containers to communicate across nodes, they must be allocated ports on the machine’s own IP address, which are then forwarded or proxied to the containers. This obviously means that containers must either coordinate which ports they use very carefully or else be allocated ports dynamically.
Coordinating ports across multiple developers is very difficult to do at scale and exposes users to cluster-level issues outside of their control. Kubernetes assumes that pods can communicate with other pods, regardless of which host they land on. We give every pod its own cluster-private-IP address so you do not need to explicitly create links between pods or mapping container ports to host ports. This means that containers within a Pod can all reach each other’s ports on localhost, and all pods in a cluster can see each other without NAT. The rest of this document will elaborate on how you can run reliable services on such a networking model.
This guide uses a simple nginx server to demonstrate proof of concept. The same principles are embodied in a more complete Jenkins CI application.
We did this in a previous example, but lets do it once again and focus on the networking perspective. Create an nginx pod, and note that it has a container port specification:
This makes it accessible from any node in your cluster. Check the nodes the pod is running on:
$ kubectl create -f ./run-my-nginx.yaml $ kubectl get pods -l run=my-nginx -o wide NAME READY STATUS RESTARTS AGE NODE my-nginx-3800858182-jr4a2 1/1 Running 0 13s kubernetes-minion-905m my-nginx-3800858182-kna2y 1/1 Running 0 13s kubernetes-minion-ljyd
Check your pods’ IPs:
$ kubectl get pods -l run=my-nginx -o yaml | grep podIP podIP: 10.244.3.4 podIP: 10.244.2.5
You should be able to ssh into any node in your cluster and curl both IPs. Note that the containers are not using port 80 on the node, nor are there any special NAT rules to route traffic to the pod. This means you can run multiple nginx pods on the same node all using the same containerPort and access them from any other pod or node in your cluster using IP. Like Docker, ports can still be published to the host node’s interface(s), but the need for this is radically diminished because of the networking model.
You can read more about how we achieve this if you’re curious.
So we have pods running nginx in a flat, cluster wide, address space. In theory, you could talk to these pods directly, but what happens when a node dies? The pods die with it, and the Deployment will create new ones, with different IPs. This is the problem a Service solves.
A Kubernetes Service is an abstraction which defines a logical set of Pods running somewhere in your cluster, that all provide the same functionality. When created, each Service is assigned a unique IP address (also called clusterIP). This address is tied to the lifespan of the Service, and will not change while the Service is alive. Pods can be configured to talk to the Service, and know that communication to the Service will be automatically load-balanced out to some pod that is a member of the Service.
You can create a Service for your 2 nginx replicas with
$ kubectl expose deployment/my-nginx service "my-nginx" exposed
This is equivalent to
kubectl create -f the following yaml:
This specification will create a Service which targets TCP port 80 on any Pod with the
run: my-nginx label, and expose it on an abstracted Service port (
targetPort: is the port the container accepts traffic on,
port: is the abstracted Service port, which can be any port other pods use to access the Service). View service API object to see the list of supported fields in service definition.
Check your Service:
$ kubectl get svc my-nginx NAME CLUSTER-IP EXTERNAL-IP PORT(S) AGE my-nginx 10.0.162.149 <none> 80/TCP 21s
As mentioned previously, a Service is backed by a group of pods. These pods are exposed through
endpoints. The Service’s selector will be evaluated continuously and the results will be POSTed to an Endpoints object also named
my-nginx. When a pod dies, it is automatically removed from the endpoints, and new pods matching the Service’s selector will automatically get added to the endpoints. Check the endpoints, and note that the IPs are the same as the pods created in the first step:
$ kubectl describe svc my-nginx Name: my-nginx Namespace: default Labels: run=my-nginx Selector: run=my-nginx Type: ClusterIP IP: 10.0.162.149 Port: <unset> 80/TCP Endpoints: 10.244.2.5:80,10.244.3.4:80 Session Affinity: None No events. $ kubectl get ep my-nginx NAME ENDPOINTS AGE my-nginx 10.244.2.5:80,10.244.3.4:80 1m
You should now be able to curl the nginx Service on
<CLUSTER-IP>:<PORT> from any node in your cluster. Note that the Service IP is completely virtual, it never hits the wire, if you’re curious about how this works you can read more about the service proxy.
Kubernetes supports 2 primary modes of finding a Service - environment variables and DNS. The former works out of the box while the latter requires the kube-dns cluster addon.
When a Pod is run on a Node, the kubelet adds a set of environment variables for each active Service. This introduces an ordering problem. To see why, inspect the environment of your running nginx pods (your pod name will be different):
$ kubectl exec my-nginx-3800858182-jr4a2 -- printenv | grep SERVICE KUBERNETES_SERVICE_HOST=10.0.0.1 KUBERNETES_SERVICE_PORT=443 KUBERNETES_SERVICE_PORT_HTTPS=443
Note there’s no mention of your Service. This is because you created the replicas before the Service. Another disadvantage of doing this is that the scheduler might put both pods on the same machine, which will take your entire Service down if it dies. We can do this the right way by killing the 2 pods and waiting for the Deployment to recreate them. This time around the Service exists before the replicas. This will give you scheduler-level Service spreading of your pods (provided all your nodes have equal capacity), as well as the right environment variables:
$ kubectl scale deployment my-nginx --replicas=0; kubectl scale deployment my-nginx --replicas=2; $ kubectl get pods -l run=my-nginx -o wide NAME READY STATUS RESTARTS AGE NODE my-nginx-3800858182-e9ihh 1/1 Running 0 5s kubernetes-minion-ljyd my-nginx-3800858182-j4rm4 1/1 Running 0 5s kubernetes-minion-905m
You may notice that the pods have different names, since they are killed and recreated.
$ kubectl exec my-nginx-3800858182-e9ihh -- printenv | grep SERVICE KUBERNETES_SERVICE_PORT=443 MY_NGINX_SERVICE_HOST=10.0.162.149 KUBERNETES_SERVICE_HOST=10.0.0.1 MY_NGINX_SERVICE_PORT=80 KUBERNETES_SERVICE_PORT_HTTPS=443
Kubernetes offers a DNS cluster addon Service that uses skydns to automatically assign dns names to other Services. You can check if it’s running on your cluster:
$ kubectl get services kube-dns --namespace=kube-system NAME CLUSTER-IP EXTERNAL-IP PORT(S) AGE kube-dns 10.0.0.10 <none> 53/UDP,53/TCP 8m
If it isn’t running, you can enable it. The rest of this section will assume you have a Service with a long lived IP (my-nginx), and a dns server that has assigned a name to that IP (the kube-dns cluster addon), so you can talk to the Service from any pod in your cluster using standard methods (e.g. gethostbyname). Let’s run another curl application to test this:
$ kubectl run curl --image=radial/busyboxplus:curl -i --tty Waiting for pod default/curl-131556218-9fnch to be running, status is Pending, pod ready: false Hit enter for command prompt
Then, hit enter and run
[ root@curl-131556218-9fnch:/ ]$ nslookup my-nginx Server: 10.0.0.10 Address 1: 10.0.0.10 Name: my-nginx Address 1: 10.0.162.149
Till now we have only accessed the nginx server from within the cluster. Before exposing the Service to the internet, you want to make sure the communication channel is secure. For this, you will need:
You can acquire all these from the nginx https example, in short:
$ make keys secret KEY=/tmp/nginx.key CERT=/tmp/nginx.crt SECRET=/tmp/secret.json $ kubectl create -f /tmp/secret.json secrets/nginxsecret $ kubectl get secrets NAME TYPE DATA default-token-il9rc kubernetes.io/service-account-token 1 nginxsecret Opaque 2
Now modify your nginx replicas to start a https server using the certificate in the secret, and the Service, to expose both ports (80 and 443):
Noteworthy points about the nginx-secure-app manifest:
$ kubectl apply -f ./nginx-secure-app.yaml $ kubectl delete rc,svc -l app=nginx; kubectl create -f ./nginx-app.yaml service "my-nginx" configured deployment "my-nginx" configured
At this point you can reach the nginx server from any node.
$ kubectl get pods -o yaml | grep -i podip podIP: 10.244.3.5 node $ curl -k https://10.244.3.5 ... <h1>Welcome to nginx!</h1>
Note how we supplied the
-k parameter to curl in the last step, this is because we don’t know anything about the pods running nginx at certificate generation time,
so we have to tell curl to ignore the CName mismatch. By creating a Service we linked the CName used in the certificate with the actual DNS name used by pods during Service lookup.
Lets test this from a pod (the same secret is being reused for simplicity, the pod only needs nginx.crt to access the Service):
$ kubectl create -f ./curlpod.yaml $ kubectl get pods NAME READY STATUS RESTARTS AGE curlpod 1/1 Running 0 2m $ kubectl exec curlpod -- curl https://my-nginx --cacert /etc/nginx/ssl/nginx.crt ... <title>Welcome to nginx!</title> ...
For some parts of your applications you may want to expose a Service onto an external IP address. Kubernetes supports two ways of doing this: NodePorts and LoadBalancers. The Service created in the last section already used
NodePort, so your nginx https replica is ready to serve traffic on the internet if your node has a public IP.
$ kubectl get svc my-nginx -o yaml | grep nodePort -C 5 uid: 07191fb3-f61a-11e5-8ae5-42010af00002 spec: clusterIP: 10.0.162.149 ports: - name: http nodePort: 31704 port: 8080 protocol: TCP targetPort: 80 - name: https nodePort: 32453 port: 443 protocol: TCP targetPort: 443 selector: run: my-nginx $ kubectl get nodes -o yaml | grep ExternalIP -C 1 - address: 184.108.40.206 type: ExternalIP allocatable: -- - address: 220.127.116.11 type: ExternalIP allocatable: ... $ curl https://<EXTERNAL-IP>:<NODE-PORT> -k ... <h1>Welcome to nginx!</h1>
Lets now recreate the Service to use a cloud load balancer, just change the
my-nginx Service from
$ kubectl edit svc my-nginx $ kubectl get svc my-nginx NAME CLUSTER-IP EXTERNAL-IP PORT(S) AGE my-nginx 10.0.162.149 18.104.22.168 80/TCP,81/TCP,82/TCP 21s $ curl https://<EXTERNAL-IP> -k ... <title>Welcome to nginx!</title>
The IP address in the
EXTERNAL-IP column is the one that is available on the public internet. The
CLUSTER-IP is only available inside your
cluster/private cloud network.
Note that on AWS, type
LoadBalancer creates an ELB, which uses a (long)
hostname, not an IP. It’s too long to fit in the standard
kubectl get svc
output, in fact, so you’ll need to do
kubectl describe service my-nginx to
see it. You’ll see something like this:
$ kubectl describe service my-nginx ... LoadBalancer Ingress: a320587ffd19711e5a37606cf4a74574-1142138393.us-east-1.elb.amazonaws.com ...
Kubernetes also supports Federated Services, which can span multiple clusters and cloud providers, to provide increased availability, bettern fault tolerance and greater scalability for your services. See the Federated Services User Guide for further information.