2016-05-06 10:34:30 +00:00
# Kubernetes Ingress Controller
2018-01-15 08:40:05 +00:00
This guide explains how to use Træfik as an Ingress controller for a Kubernetes cluster.
2017-09-11 17:10:04 +00:00
2017-09-01 17:44:03 +00:00
If you are not familiar with Ingresses in Kubernetes you might want to read the [Kubernetes user guide ](https://kubernetes.io/docs/concepts/services-networking/ingress/ )
2016-05-06 10:34:30 +00:00
The config files used in this guide can be found in the [examples directory ](https://github.com/containous/traefik/tree/master/examples/k8s )
## Prerequisites
2018-01-15 08:40:05 +00:00
1. A working Kubernetes cluster. If you want to follow along with this guide, you should setup [minikube ](https://kubernetes.io/docs/getting-started-guides/minikube/ ) on your machine, as it is the quickest way to get a local Kubernetes cluster setup for experimentation and development.
!!! note
The guide is likely not fully adequate for a production-ready setup.
2016-05-06 10:34:30 +00:00
2017-09-01 17:44:03 +00:00
2. The `kubectl` binary should be [installed on your workstation ](https://kubernetes.io/docs/getting-started-guides/minikube/#download-kubectl ).
2016-05-06 10:34:30 +00:00
2017-05-15 17:59:57 +00:00
### Role Based Access Control configuration (Kubernetes 1.6+ only)
2018-07-25 15:58:02 +00:00
Kubernetes introduces [Role Based Access Control (RBAC) ](https://kubernetes.io/docs/reference/access-authn-authz/rbac/ ) in 1.6+ to allow fine-grained control of Kubernetes resources and API.
2017-05-15 17:59:57 +00:00
2017-10-10 14:26:03 +00:00
If your cluster is configured with RBAC, you will need to authorize Træfik to use the Kubernetes API. There are two ways to set up the proper permission: Via namespace-specific RoleBindings or a single, global ClusterRoleBinding.
2017-05-15 17:59:57 +00:00
2017-10-10 14:26:03 +00:00
RoleBindings per namespace enable to restrict granted permissions to the very namespaces only that Træfik is watching over, thereby following the least-privileges principle. This is the preferred approach if Træfik is not supposed to watch all namespaces, and the set of namespaces does not change dynamically. Otherwise, a single ClusterRoleBinding must be employed.
2017-12-05 01:46:03 +00:00
!!! note
RoleBindings per namespace are available in Træfik 1.5 and later. Please use ClusterRoleBindings for older versions.
2017-10-10 14:26:03 +00:00
For the sake of simplicity, this guide will use a ClusterRoleBinding:
2017-05-15 17:59:57 +00:00
```yaml
---
kind: ClusterRole
apiVersion: rbac.authorization.k8s.io/v1beta1
metadata:
name: traefik-ingress-controller
rules:
- apiGroups:
- ""
resources:
- services
- endpoints
2017-07-24 09:30:36 +00:00
- secrets
2017-05-15 17:59:57 +00:00
verbs:
- get
- list
- watch
- apiGroups:
- extensions
resources:
- ingresses
verbs:
- get
- list
- watch
---
kind: ClusterRoleBinding
apiVersion: rbac.authorization.k8s.io/v1beta1
metadata:
name: traefik-ingress-controller
roleRef:
apiGroup: rbac.authorization.k8s.io
kind: ClusterRole
name: traefik-ingress-controller
subjects:
- kind: ServiceAccount
name: traefik-ingress-controller
namespace: kube-system
```
[examples/k8s/traefik-rbac.yaml ](https://github.com/containous/traefik/tree/master/examples/k8s/traefik-rbac.yaml )
```shell
kubectl apply -f https://raw.githubusercontent.com/containous/traefik/master/examples/k8s/traefik-rbac.yaml
```
2017-10-10 14:26:03 +00:00
For namespaced restrictions, one RoleBinding is required per watched namespace along with a corresponding configuration of Træfik's `kubernetes.namespaces` parameter.
2017-07-29 10:50:04 +00:00
## Deploy Træfik using a Deployment or DaemonSet
2016-05-06 10:34:30 +00:00
2017-08-26 10:12:44 +00:00
It is possible to use Træfik with a [Deployment ](https://kubernetes.io/docs/concepts/workloads/controllers/deployment/ ) or a [DaemonSet ](https://kubernetes.io/docs/concepts/workloads/controllers/daemonset/ ) object,
whereas both options have their own pros and cons:
2017-12-05 01:46:03 +00:00
2018-05-25 15:26:05 +00:00
- The scalability can be much better when using a Deployment, because you will have a Single-Pod-per-Node model when using a DaemonSet, whereas you may need less replicas based on your environment when using a Deployment.
- DaemonSets automatically scale to new nodes, when the nodes join the cluster, whereas Deployment pods are only scheduled on new nodes if required.
- DaemonSets ensure that only one replica of pods run on any single node. Deployments require affinity settings if you want to ensure that two pods don't end up on the same node.
- DaemonSets can be run with the `NET_BIND_SERVICE` capability, which will allow it to bind to port 80/443/etc on each host. This will allow bypassing the kube-proxy, and reduce traffic hops. Note that this is against the Kubernetes Best Practices [Guidelines ](https://kubernetes.io/docs/concepts/configuration/overview/#services ), and raises the potential for scheduling/scaling issues. Despite potential issues, this remains the choice for most ingress controllers.
- If you are unsure which to choose, start with the Daemonset.
2017-07-29 10:50:04 +00:00
The Deployment objects looks like this:
2016-05-06 10:34:30 +00:00
2017-11-09 09:52:03 +00:00
```yaml
2017-05-15 17:59:57 +00:00
---
apiVersion: v1
kind: ServiceAccount
metadata:
name: traefik-ingress-controller
namespace: kube-system
---
2016-05-06 10:34:30 +00:00
kind: Deployment
apiVersion: extensions/v1beta1
metadata:
name: traefik-ingress-controller
namespace: kube-system
labels:
k8s-app: traefik-ingress-lb
spec:
replicas: 1
selector:
matchLabels:
k8s-app: traefik-ingress-lb
template:
metadata:
labels:
k8s-app: traefik-ingress-lb
name: traefik-ingress-lb
spec:
2017-05-15 17:59:57 +00:00
serviceAccountName: traefik-ingress-controller
2016-05-06 10:34:30 +00:00
terminationGracePeriodSeconds: 60
containers:
2016-11-10 18:14:53 +00:00
- image: traefik
2016-05-06 10:34:30 +00:00
name: traefik-ingress-lb
2018-05-25 15:26:05 +00:00
ports:
- name: http
containerPort: 80
- name: admin
containerPort: 8080
2017-07-29 10:50:04 +00:00
args:
2018-01-11 08:46:03 +00:00
- --api
2017-07-29 10:50:04 +00:00
- --kubernetes
2018-03-22 15:44:05 +00:00
- --logLevel=INFO
2017-07-29 10:50:04 +00:00
---
kind: Service
apiVersion: v1
metadata:
name: traefik-ingress-service
2017-10-25 14:58:04 +00:00
namespace: kube-system
2017-07-29 10:50:04 +00:00
spec:
selector:
k8s-app: traefik-ingress-lb
ports:
- protocol: TCP
port: 80
2017-09-05 09:42:03 +00:00
name: web
2017-07-29 10:50:04 +00:00
- protocol: TCP
port: 8080
2017-09-05 09:42:03 +00:00
name: admin
2017-07-29 10:50:04 +00:00
type: NodePort
```
2018-01-15 08:40:05 +00:00
2017-07-29 10:50:04 +00:00
[examples/k8s/traefik-deployment.yaml ](https://github.com/containous/traefik/tree/master/examples/k8s/traefik-deployment.yaml )
2017-09-11 17:10:04 +00:00
!!! note
The Service will expose two NodePorts which allow access to the ingress and the web interface.
2017-07-29 10:50:04 +00:00
The DaemonSet objects looks not much different:
```yaml
---
apiVersion: v1
kind: ServiceAccount
metadata:
name: traefik-ingress-controller
namespace: kube-system
---
kind: DaemonSet
apiVersion: extensions/v1beta1
metadata:
name: traefik-ingress-controller
namespace: kube-system
labels:
k8s-app: traefik-ingress-lb
spec:
template:
metadata:
labels:
k8s-app: traefik-ingress-lb
name: traefik-ingress-lb
spec:
serviceAccountName: traefik-ingress-controller
terminationGracePeriodSeconds: 60
containers:
- image: traefik
name: traefik-ingress-lb
2016-05-06 10:34:30 +00:00
ports:
2017-07-29 10:50:04 +00:00
- name: http
containerPort: 80
2016-05-06 10:34:30 +00:00
hostPort: 80
2017-07-29 10:50:04 +00:00
- name: admin
containerPort: 8080
securityContext:
2018-03-22 15:44:05 +00:00
capabilities:
drop:
- ALL
add:
- NET_BIND_SERVICE
2016-05-06 10:34:30 +00:00
args:
2018-01-11 08:46:03 +00:00
- --api
2016-05-06 10:34:30 +00:00
- --kubernetes
2018-03-19 09:38:04 +00:00
- --logLevel=INFO
2017-07-29 10:50:04 +00:00
---
kind: Service
apiVersion: v1
metadata:
name: traefik-ingress-service
2017-10-25 14:58:04 +00:00
namespace: kube-system
2017-07-29 10:50:04 +00:00
spec:
selector:
k8s-app: traefik-ingress-lb
ports:
- protocol: TCP
port: 80
2017-09-05 09:42:03 +00:00
name: web
2017-07-29 10:50:04 +00:00
- protocol: TCP
port: 8080
2017-09-05 09:42:03 +00:00
name: admin
2016-05-06 10:34:30 +00:00
```
2017-07-29 10:50:04 +00:00
[examples/k8s/traefik-ds.yaml ](https://github.com/containous/traefik/tree/master/examples/k8s/traefik-ds.yaml )
2018-05-25 15:26:05 +00:00
!!! note
This will create a Daemonset that uses privileged ports 80/8080 on the host. This may not work on all providers, but illustrates the static (non-NodePort) hostPort binding. The `traefik-ingress-service` can still be used inside the cluster to access the DaemonSet pods.
2017-07-29 10:50:04 +00:00
To deploy Træfik to your cluster start by submitting one of the YAML files to the cluster with `kubectl` :
2016-05-06 10:34:30 +00:00
2017-07-29 10:50:04 +00:00
```shell
2017-09-11 17:10:04 +00:00
kubectl apply -f https://raw.githubusercontent.com/containous/traefik/master/examples/k8s/traefik-deployment.yaml
2017-07-29 10:50:04 +00:00
```
2016-05-06 10:34:30 +00:00
2017-05-15 17:59:57 +00:00
```shell
2017-09-11 17:10:04 +00:00
kubectl apply -f https://raw.githubusercontent.com/containous/traefik/master/examples/k8s/traefik-ds.yaml
2017-04-08 15:56:41 +00:00
```
2017-09-11 17:10:04 +00:00
There are some significant differences between using Deployments and DaemonSets:
2017-07-29 10:50:04 +00:00
2017-09-11 17:10:04 +00:00
- The Deployment has easier up and down scaling possibilities.
It can implement full pod lifecycle and supports rolling updates from Kubernetes 1.2.
At least one Pod is needed to run the Deployment.
- The DaemonSet automatically scales to all nodes that meets a specific selector and guarantees to fill nodes one at a time.
Rolling updates are fully supported from Kubernetes 1.7 for DaemonSets as well.
2017-07-29 10:50:04 +00:00
### Check the Pods
2016-05-06 10:34:30 +00:00
2017-07-29 10:50:04 +00:00
Now lets check if our command was successful.
2016-05-06 10:34:30 +00:00
Start by listing the pods in the `kube-system` namespace:
2017-05-15 17:59:57 +00:00
```shell
2017-09-11 17:10:04 +00:00
kubectl --namespace=kube-system get pods
```
2016-05-06 10:34:30 +00:00
2018-01-15 08:40:05 +00:00
```shell
2016-05-06 10:34:30 +00:00
NAME READY STATUS RESTARTS AGE
kube-addon-manager-minikubevm 1/1 Running 0 4h
kubernetes-dashboard-s8krj 1/1 Running 0 4h
traefik-ingress-controller-678226159-eqseo 1/1 Running 0 7m
```
2017-08-26 10:12:44 +00:00
You should see that after submitting the Deployment or DaemonSet to Kubernetes it has launched a Pod, and it is now running.
2018-02-21 16:34:04 +00:00
_It might take a few moments for Kubernetes to pull the Træfik image and start the container._
2016-05-06 10:34:30 +00:00
2017-09-11 17:10:04 +00:00
!!! note
You could also check the deployment with the Kubernetes dashboard, run
`minikube dashboard` to open it in your browser, then choose the `kube-system`
namespace from the menu at the top right of the screen.
2016-05-06 10:34:30 +00:00
2017-08-26 10:12:44 +00:00
You should now be able to access Træfik on port 80 of your Minikube instance when using the DaemonSet:
2016-05-06 10:34:30 +00:00
2018-01-15 08:40:05 +00:00
```shell
2016-05-06 10:34:30 +00:00
curl $(minikube ip)
2017-09-11 17:10:04 +00:00
```
2018-01-15 08:40:05 +00:00
```shell
2016-05-06 10:34:30 +00:00
404 page not found
```
2018-01-03 08:12:03 +00:00
If you decided to use the deployment, then you need to target the correct NodePort, which can be seen when you execute `kubectl get services --namespace=kube-system` .
2017-07-29 10:50:04 +00:00
2018-01-15 08:40:05 +00:00
```shell
2017-07-29 10:50:04 +00:00
curl $(minikube ip):< NODEPORT >
2017-09-11 17:10:04 +00:00
```
2018-01-15 08:40:05 +00:00
```shell
2017-07-29 10:50:04 +00:00
404 page not found
```
2017-09-11 17:10:04 +00:00
!!! note
We expect to see a 404 response here as we haven't yet given Træfik any configuration.
2016-05-06 10:34:30 +00:00
2018-01-03 08:12:03 +00:00
All further examples below assume a DaemonSet installation. Deployment users will need to append the NodePort when constructing requests.
2017-03-31 07:57:06 +00:00
## Deploy Træfik using Helm Chart
2018-01-15 08:40:05 +00:00
!!! note
2018-02-21 16:34:04 +00:00
The Helm Chart is maintained by the community, not the Træfik project maintainers.
2017-09-11 17:10:04 +00:00
2018-01-15 08:40:05 +00:00
Instead of installing Træfik via Kubernetes object directly, you can also use the Træfik Helm chart.
2017-03-31 07:57:06 +00:00
2018-01-15 08:40:05 +00:00
Install the Træfik chart by:
2017-03-31 07:57:06 +00:00
2017-07-29 10:50:04 +00:00
```shell
2017-09-11 17:10:04 +00:00
helm install stable/traefik
2017-03-31 07:57:06 +00:00
```
2018-05-30 15:38:04 +00:00
Install the Træfik chart using a values.yaml file.
2017-03-31 07:57:06 +00:00
2018-05-30 15:38:04 +00:00
```shell
helm install --values values.yaml stable/traefik
```
```yaml
dashboard:
enabled: true
domain: traefik-ui.minikube
kubernetes:
namespaces:
- default
- kube-system
```
2018-01-15 08:40:05 +00:00
For more information, check out [the documentation ](https://github.com/kubernetes/charts/tree/master/stable/traefik ).
2017-03-31 07:57:06 +00:00
2018-01-15 08:40:05 +00:00
## Submitting an Ingress to the Cluster
2016-05-06 10:34:30 +00:00
2017-08-26 10:12:44 +00:00
Lets start by creating a Service and an Ingress that will expose the [Træfik Web UI ](https://github.com/containous/traefik#web-ui ).
2016-05-06 10:34:30 +00:00
```yaml
apiVersion: v1
kind: Service
metadata:
name: traefik-web-ui
namespace: kube-system
spec:
selector:
2017-02-03 16:47:48 +00:00
k8s-app: traefik-ingress-lb
2016-05-06 10:34:30 +00:00
ports:
2018-06-28 21:10:03 +00:00
- name: web
port: 80
2016-05-06 10:34:30 +00:00
targetPort: 8080
---
apiVersion: extensions/v1beta1
kind: Ingress
metadata:
name: traefik-web-ui
namespace: kube-system
spec:
rules:
2017-05-04 12:31:38 +00:00
- host: traefik-ui.minikube
2016-05-06 10:34:30 +00:00
http:
paths:
2018-06-28 21:10:03 +00:00
- path: /
backend:
2016-05-06 10:34:30 +00:00
serviceName: traefik-web-ui
2018-06-28 21:10:03 +00:00
servicePort: web
2016-05-06 10:34:30 +00:00
```
2018-01-15 08:40:05 +00:00
2016-05-06 10:34:30 +00:00
[examples/k8s/ui.yaml ](https://github.com/containous/traefik/tree/master/examples/k8s/ui.yaml )
2017-04-30 18:17:57 +00:00
```shell
2017-05-15 17:59:57 +00:00
kubectl apply -f https://raw.githubusercontent.com/containous/traefik/master/examples/k8s/ui.yaml
2016-05-06 10:34:30 +00:00
```
2018-01-15 08:40:05 +00:00
Now lets setup an entry in our `/etc/hosts` file to route `traefik-ui.minikube` to our cluster.
2016-05-06 10:34:30 +00:00
2018-01-15 08:40:05 +00:00
In production you would want to set up real DNS entries.
You can get the IP address of your minikube instance by running `minikube ip` :
2016-05-06 10:34:30 +00:00
2017-04-30 18:17:57 +00:00
```shell
2017-05-04 12:31:38 +00:00
echo "$(minikube ip) traefik-ui.minikube" | sudo tee -a /etc/hosts
2016-05-06 10:34:30 +00:00
```
2018-01-15 08:40:05 +00:00
We should now be able to visit [traefik-ui.minikube ](http://traefik-ui.minikube ) in the browser and view the Træfik web UI.
2016-05-06 10:34:30 +00:00
2018-01-07 23:36:03 +00:00
### Add a TLS Certificate to the Ingress
!!! note
2018-04-30 18:28:03 +00:00
For this example to work you need a TLS entrypoint. You don't have to provide a TLS certificate at this point.
For more details see [here ](/configuration/entrypoints/ ).
2018-01-07 23:36:03 +00:00
To setup an HTTPS-protected ingress, you can leverage the TLS feature of the ingress resource.
```yaml
apiVersion: extensions/v1beta1
kind: Ingress
metadata:
name: traefik-web-ui
namespace: kube-system
annotations:
kubernetes.io/ingress.class: traefik
spec:
rules:
- host: traefik-ui.minikube
http:
paths:
- backend:
serviceName: traefik-web-ui
servicePort: 80
tls:
2018-04-03 08:30:03 +00:00
- secretName: traefik-ui-tls-cert
2018-01-07 23:36:03 +00:00
```
2018-04-30 18:28:03 +00:00
In addition to the modified ingress you need to provide the TLS certificate via a Kubernetes secret in the same namespace as the ingress.
The following two commands will generate a new certificate and create a secret containing the key and cert files.
2018-01-07 23:36:03 +00:00
```shell
openssl req -x509 -nodes -days 365 -newkey rsa:2048 -keyout tls.key -out tls.crt -subj "/CN=traefik-ui.minikube"
kubectl -n kube-system create secret tls traefik-ui-tls-cert --key=tls.key --cert=tls.crt
```
If there are any errors while loading the TLS section of an ingress, the whole ingress will be skipped.
!!! note
2018-04-30 18:28:03 +00:00
The secret must have two entries named `tls.key` and `tls.crt` .
See the [Kubernetes documentation ](https://kubernetes.io/docs/concepts/services-networking/ingress/#tls ) for more details.
2018-01-07 23:36:03 +00:00
!!! note
2018-04-30 18:28:03 +00:00
The TLS certificates will be added to all entrypoints defined by the ingress annotation `traefik.frontend.entryPoints` .
If no such annotation is provided, the TLS certificates will be added to all TLS-enabled `defaultEntryPoints` .
2018-01-07 23:36:03 +00:00
!!! note
2018-04-30 18:28:03 +00:00
The field `hosts` in the TLS configuration is ignored. Instead, the domains provided by the certificate are used for this purpose.
It is recommended to not use wildcard certificates as they will match globally.
2018-01-07 23:36:03 +00:00
2017-11-09 09:52:03 +00:00
## Basic Authentication
2018-02-21 16:34:04 +00:00
It's possible to protect access to Træfik through basic authentication. (See the [Kubernetes Ingress ](/configuration/backends/kubernetes ) configuration page for syntactical details and restrictions.)
2017-11-09 09:52:03 +00:00
2018-01-15 08:40:05 +00:00
### Creating the Secret
2017-11-09 09:52:03 +00:00
2018-04-11 08:34:04 +00:00
A. Use `htpasswd` to create a file containing the username and the MD5-encoded password:
2017-11-09 09:52:03 +00:00
```shell
htpasswd -c ./auth myusername
```
You will be prompted for a password which you will have to enter twice.
`htpasswd` will create a file with the following:
```shell
cat auth
```
2018-01-15 08:40:05 +00:00
```shell
2017-11-09 09:52:03 +00:00
myusername:$apr1$78Jyn/1K$ERHKVRPPlzAX8eBtLuvRZ0
```
2018-01-15 08:40:05 +00:00
B. Now use `kubectl` to create a secret in the `monitoring` namespace using the file created by `htpasswd` .
2017-11-09 09:52:03 +00:00
```shell
kubectl create secret generic mysecret --from-file auth --namespace=monitoring
```
!!! note
2018-01-15 08:40:05 +00:00
Secret must be in same namespace as the Ingress object.
2017-11-09 09:52:03 +00:00
2018-01-15 08:40:05 +00:00
C. Attach the following annotations to the Ingress object:
2017-11-09 09:52:03 +00:00
2018-07-31 15:58:04 +00:00
- `traefik.ingress.kubernetes.io/auth-type: "basic"`
- `traefik.ingress.kubernetes.io/auth-secret: "mysecret"`
2017-11-09 09:52:03 +00:00
2018-01-15 08:40:05 +00:00
They specify basic authentication and reference the Secret `mysecret` containing the credentials.
Following is a full Ingress example based on Prometheus:
2017-11-09 09:52:03 +00:00
```yaml
apiVersion: extensions/v1beta1
kind: Ingress
metadata:
name: prometheus-dashboard
namespace: monitoring
annotations:
kubernetes.io/ingress.class: traefik
2018-07-31 15:58:04 +00:00
traefik.ingress.kubernetes.io/auth-type: "basic"
traefik.ingress.kubernetes.io/auth-secret: "mysecret"
2017-11-09 09:52:03 +00:00
spec:
rules:
- host: dashboard.prometheus.example.com
http:
paths:
- backend:
serviceName: prometheus
servicePort: 9090
```
2018-01-15 08:40:05 +00:00
You can apply the example as following:
2017-11-09 09:52:03 +00:00
```shell
kubectl create -f prometheus-ingress.yaml -n monitoring
```
2018-01-15 08:40:05 +00:00
## Name-based Routing
2016-05-06 10:34:30 +00:00
2018-01-15 08:40:05 +00:00
In this example we are going to setup websites for three of the United Kingdoms best loved cheeses: Cheddar, Stilton, and Wensleydale.
2016-05-06 10:34:30 +00:00
2018-01-15 08:40:05 +00:00
First lets start by launching the pods for the cheese websites.
2016-05-06 10:34:30 +00:00
```yaml
---
kind: Deployment
apiVersion: extensions/v1beta1
metadata:
name: stilton
labels:
app: cheese
cheese: stilton
spec:
replicas: 2
selector:
matchLabels:
app: cheese
task: stilton
template:
metadata:
labels:
app: cheese
task: stilton
version: v0.0.1
spec:
containers:
- name: cheese
image: errm/cheese:stilton
ports:
- containerPort: 80
---
kind: Deployment
apiVersion: extensions/v1beta1
metadata:
name: cheddar
labels:
app: cheese
cheese: cheddar
spec:
replicas: 2
selector:
matchLabels:
app: cheese
task: cheddar
template:
metadata:
labels:
app: cheese
task: cheddar
version: v0.0.1
spec:
containers:
- name: cheese
image: errm/cheese:cheddar
ports:
- containerPort: 80
---
kind: Deployment
apiVersion: extensions/v1beta1
metadata:
name: wensleydale
labels:
app: cheese
cheese: wensleydale
spec:
replicas: 2
selector:
matchLabels:
app: cheese
task: wensleydale
template:
metadata:
labels:
app: cheese
task: wensleydale
version: v0.0.1
spec:
containers:
- name: cheese
image: errm/cheese:wensleydale
ports:
- containerPort: 80
```
2018-01-15 08:40:05 +00:00
2016-05-06 10:34:30 +00:00
[examples/k8s/cheese-deployments.yaml ](https://github.com/containous/traefik/tree/master/examples/k8s/cheese-deployments.yaml )
2017-04-30 18:17:57 +00:00
```shell
2017-05-15 17:59:57 +00:00
kubectl apply -f https://raw.githubusercontent.com/containous/traefik/master/examples/k8s/cheese-deployments.yaml
2016-05-06 10:34:30 +00:00
```
2018-01-15 08:40:05 +00:00
Next we need to setup a Service for each of the cheese pods.
2016-05-06 10:34:30 +00:00
```yaml
---
apiVersion: v1
kind: Service
metadata:
name: stilton
spec:
ports:
- name: http
targetPort: 80
port: 80
selector:
app: cheese
task: stilton
---
apiVersion: v1
kind: Service
metadata:
name: cheddar
spec:
ports:
- name: http
targetPort: 80
port: 80
selector:
app: cheese
task: cheddar
---
apiVersion: v1
kind: Service
metadata:
name: wensleydale
2017-02-03 16:47:48 +00:00
annotations:
traefik.backend.circuitbreaker: "NetworkErrorRatio() > 0.5"
2016-05-06 10:34:30 +00:00
spec:
ports:
- name: http
targetPort: 80
port: 80
selector:
app: cheese
task: wensleydale
```
2017-02-03 16:47:48 +00:00
2017-09-11 17:10:04 +00:00
!!! note
We also set a [circuit breaker expression ](/basics/#backends ) for one of the backends by setting the `traefik.backend.circuitbreaker` annotation on the service.
2017-02-03 16:47:48 +00:00
2016-05-06 10:34:30 +00:00
[examples/k8s/cheese-services.yaml ](https://github.com/containous/traefik/tree/master/examples/k8s/cheese-services.yaml )
2017-04-30 18:17:57 +00:00
```shell
2017-05-15 17:59:57 +00:00
kubectl apply -f https://raw.githubusercontent.com/containous/traefik/master/examples/k8s/cheese-services.yaml
2016-05-06 10:34:30 +00:00
```
Now we can submit an ingress for the cheese websites.
```yaml
apiVersion: extensions/v1beta1
kind: Ingress
metadata:
name: cheese
2017-03-03 19:30:22 +00:00
annotations:
kubernetes.io/ingress.class: traefik
2016-05-06 10:34:30 +00:00
spec:
rules:
2017-05-04 12:31:38 +00:00
- host: stilton.minikube
2016-05-06 10:34:30 +00:00
http:
paths:
- path: /
backend:
serviceName: stilton
servicePort: http
2017-05-04 12:31:38 +00:00
- host: cheddar.minikube
2016-05-06 10:34:30 +00:00
http:
paths:
- path: /
backend:
serviceName: cheddar
servicePort: http
2017-05-04 12:31:38 +00:00
- host: wensleydale.minikube
2016-05-06 10:34:30 +00:00
http:
paths:
- path: /
backend:
serviceName: wensleydale
servicePort: http
```
2018-01-15 08:40:05 +00:00
2016-05-06 10:34:30 +00:00
[examples/k8s/cheese-ingress.yaml ](https://github.com/containous/traefik/tree/master/examples/k8s/cheese-ingress.yaml )
2017-09-11 17:10:04 +00:00
!!! note
we list each hostname, and add a backend service.
2016-05-06 10:34:30 +00:00
2017-04-30 18:17:57 +00:00
```shell
2017-05-15 17:59:57 +00:00
kubectl apply -f https://raw.githubusercontent.com/containous/traefik/master/examples/k8s/cheese-ingress.yaml
2016-05-06 10:34:30 +00:00
```
2017-09-11 17:10:04 +00:00
Now visit the [Træfik dashboard ](http://traefik-ui.minikube/ ) and you should see a frontend for each host.
2018-01-15 08:40:05 +00:00
Along with a backend listing for each service with a server set up for each pod.
2016-05-06 10:34:30 +00:00
2017-08-26 10:12:44 +00:00
If you edit your `/etc/hosts` again you should be able to access the cheese websites in your browser.
2016-05-06 10:34:30 +00:00
2017-04-30 18:17:57 +00:00
```shell
2017-05-04 12:31:38 +00:00
echo "$(minikube ip) stilton.minikube cheddar.minikube wensleydale.minikube" | sudo tee -a /etc/hosts
2016-05-06 10:34:30 +00:00
```
2018-01-15 08:40:05 +00:00
- [Stilton ](http://stilton.minikube/ )
- [Cheddar ](http://cheddar.minikube/ )
- [Wensleydale ](http://wensleydale.minikube/ )
2016-05-06 10:34:30 +00:00
2018-01-15 08:40:05 +00:00
## Path-based Routing
2016-05-06 10:34:30 +00:00
2017-08-26 10:12:44 +00:00
Now lets suppose that our fictional client has decided that while they are super happy about our cheesy web design, when they asked for 3 websites they had not really bargained on having to buy 3 domain names.
2016-05-06 10:34:30 +00:00
No problem, we say, why don't we reconfigure the sites to host all 3 under one domain.
```yaml
apiVersion: extensions/v1beta1
kind: Ingress
metadata:
name: cheeses
annotations:
2017-03-03 19:30:22 +00:00
kubernetes.io/ingress.class: traefik
2017-06-01 12:04:24 +00:00
traefik.frontend.rule.type: PathPrefixStrip
2016-05-06 10:34:30 +00:00
spec:
rules:
2017-05-04 12:31:38 +00:00
- host: cheeses.minikube
2016-05-06 10:34:30 +00:00
http:
paths:
- path: /stilton
backend:
serviceName: stilton
servicePort: http
- path: /cheddar
backend:
serviceName: cheddar
servicePort: http
- path: /wensleydale
backend:
serviceName: wensleydale
servicePort: http
```
2018-01-15 08:40:05 +00:00
2016-05-06 10:34:30 +00:00
[examples/k8s/cheeses-ingress.yaml ](https://github.com/containous/traefik/tree/master/examples/k8s/cheeses-ingress.yaml )
2017-09-11 17:10:04 +00:00
!!! note
2018-01-15 08:40:05 +00:00
We are configuring Træfik to strip the prefix from the url path with the `traefik.frontend.rule.type` annotation so that we can use the containers from the previous example without modification.
2016-05-06 10:34:30 +00:00
2017-04-30 18:17:57 +00:00
```shell
2017-05-15 17:59:57 +00:00
kubectl apply -f https://raw.githubusercontent.com/containous/traefik/master/examples/k8s/cheeses-ingress.yaml
2016-05-06 10:34:30 +00:00
```
2017-04-30 18:17:57 +00:00
```shell
2017-05-04 12:31:38 +00:00
echo "$(minikube ip) cheeses.minikube" | sudo tee -a /etc/hosts
2016-05-06 10:34:30 +00:00
```
You should now be able to visit the websites in your browser.
2018-01-15 08:40:05 +00:00
- [cheeses.minikube/stilton ](http://cheeses.minikube/stilton/ )
- [cheeses.minikube/cheddar ](http://cheeses.minikube/cheddar/ )
- [cheeses.minikube/wensleydale ](http://cheeses.minikube/wensleydale/ )
2017-02-15 21:37:47 +00:00
2018-01-15 08:40:05 +00:00
## Specifying Routing Priorities
2017-07-29 16:35:23 +00:00
2018-01-15 08:40:05 +00:00
Sometimes you need to specify priority for ingress routes, especially when handling wildcard routes.
This can be done by adding the `traefik.frontend.priority` annotation, i.e.:
2017-07-29 16:35:23 +00:00
```yaml
apiVersion: extensions/v1beta1
kind: Ingress
metadata:
name: wildcard-cheeses
annotations:
2017-10-09 12:16:03 +00:00
traefik.frontend.priority: "1"
2017-07-29 16:35:23 +00:00
spec:
rules:
- host: *.minikube
http:
paths:
- path: /
backend:
serviceName: stilton
servicePort: http
kind: Ingress
metadata:
name: specific-cheeses
annotations:
2017-10-09 12:16:03 +00:00
traefik.frontend.priority: "2"
2017-07-29 16:35:23 +00:00
spec:
rules:
- host: specific.minikube
http:
paths:
- path: /
backend:
serviceName: stilton
servicePort: http
```
2018-01-15 08:40:05 +00:00
Note that priority values must be quoted to avoid numeric interpretation (which are illegal for annotations).
2017-07-29 16:35:23 +00:00
2017-07-05 21:50:15 +00:00
## Forwarding to ExternalNames
When specifying an [ExternalName ](https://kubernetes.io/docs/concepts/services-networking/service/#services-without-selectors ),
2018-01-15 08:40:05 +00:00
Træfik will forward requests to the given host accordingly and use HTTPS when the Service port matches 443.
2017-08-26 10:12:44 +00:00
This still requires setting up a proper port mapping on the Service from the Ingress port to the (external) Service port.
2017-07-05 21:50:15 +00:00
2018-01-15 08:40:05 +00:00
## Disable passing the Host Header
2017-05-02 13:08:18 +00:00
2018-01-15 08:40:05 +00:00
By default Træfik will pass the incoming Host header to the upstream resource.
2017-09-11 17:10:04 +00:00
2018-01-15 08:40:05 +00:00
However, there are times when you may not want this to be the case. For example, if your service is of the ExternalName type.
2017-02-15 21:37:47 +00:00
2018-01-15 08:40:05 +00:00
### Disable globally
2017-05-02 13:08:18 +00:00
2018-01-15 08:40:05 +00:00
Add the following to your TOML configuration file:
2017-09-11 17:10:04 +00:00
2017-02-16 00:11:31 +00:00
```toml
disablePassHostHeaders = true
```
2018-01-15 08:40:05 +00:00
### Disable per Ingress
2017-05-02 13:08:18 +00:00
2018-01-15 08:40:05 +00:00
To disable passing the Host header per ingress resource set the `traefik.frontend.passHostHeader` annotation on your ingress to `"false"` .
2017-02-15 21:37:47 +00:00
2018-01-15 08:40:05 +00:00
Here is an example definition:
2017-09-11 17:10:04 +00:00
2017-02-15 21:37:47 +00:00
```yaml
apiVersion: extensions/v1beta1
kind: Ingress
metadata:
name: example
annotations:
2017-03-03 19:30:22 +00:00
kubernetes.io/ingress.class: traefik
2017-02-15 21:37:47 +00:00
traefik.frontend.passHostHeader: "false"
spec:
rules:
- host: example.com
http:
paths:
- path: /static
backend:
serviceName: static
servicePort: https
```
And an example service definition:
2017-09-11 17:10:04 +00:00
2017-02-15 21:37:47 +00:00
```yaml
apiVersion: v1
kind: Service
metadata:
name: static
spec:
ports:
- name: https
port: 443
type: ExternalName
externalName: static.otherdomain.com
```
2018-01-15 08:40:05 +00:00
If you were to visit `example.com/static` the request would then be passed on to `static.otherdomain.com/static` , and `static.otherdomain.com` would receive the request with the Host header being `static.otherdomain.com` .
2017-02-16 00:11:31 +00:00
2017-08-28 12:33:07 +00:00
!!! note
2018-01-15 08:40:05 +00:00
The per-ingress annotation overrides whatever the global value is set to.
So you could set `disablePassHostHeaders` to `true` in your TOML configuration file and then enable passing the host header per ingress if you wanted.
2017-03-03 19:30:22 +00:00
2017-10-09 10:36:03 +00:00
## Partitioning the Ingress object space
2017-04-30 18:17:57 +00:00
2017-10-09 10:36:03 +00:00
By default, Træfik processes every Ingress objects it observes. At times, however, it may be desirable to ignore certain objects. The following sub-sections describe common use cases and how they can be handled with Træfik.
2017-09-11 17:10:04 +00:00
2017-10-09 10:36:03 +00:00
### Between Træfik and other Ingress controller implementations
2017-04-30 18:17:57 +00:00
2017-10-09 10:36:03 +00:00
Sometimes Træfik runs along other Ingress controller implementations. One such example is when both Træfik and a cloud provider Ingress controller are active.
The `kubernetes.io/ingress.class` annotation can be attached to any Ingress object in order to control whether Træfik should handle it.
2018-04-30 18:28:03 +00:00
If the annotation is missing, contains an empty value, or the value `traefik` , then the Træfik controller will take responsibility and process the associated Ingress object.
2018-06-22 14:54:03 +00:00
It is also possible to set the `ingressClass` option in Træfik to a particular value. Træfik will only process matching Ingress objects.
2018-04-30 18:28:03 +00:00
For instance, setting the option to `traefik-internal` causes Træfik to process Ingress objects with the same `kubernetes.io/ingress.class` annotation value, ignoring all other objects (including those with a `traefik` value, empty value, and missing annotation).
2017-10-09 10:36:03 +00:00
2018-06-22 14:54:03 +00:00
!!! note
Letting multiple ingress controllers handle the same ingress objects can lead to unintended behavior.
It is recommended to prefix all ingressClass values with `traefik` to avoid unintended collisions with other ingress implementations.
2017-10-09 10:36:03 +00:00
### Between multiple Træfik Deployments
2018-04-30 18:28:03 +00:00
Sometimes multiple Træfik Deployments are supposed to run concurrently.
For instance, it is conceivable to have one Deployment deal with internal and another one with external traffic.
2017-10-09 10:36:03 +00:00
2018-04-30 18:28:03 +00:00
For such cases, it is advisable to classify Ingress objects through a label and configure the `labelSelector` option per each Træfik Deployment accordingly.
To stick with the internal/external example above, all Ingress objects meant for internal traffic could receive a `traffic-type: internal` label while objects designated for external traffic receive a `traffic-type: external` label.
The label selectors on the Træfik Deployments would then be `traffic-type=internal` and `traffic-type=external` , respectively.
2017-04-30 18:17:57 +00:00
2018-07-03 14:50:06 +00:00
## Traffic Splitting
It is possible to split Ingress traffic in a fine-grained manner between multiple deployments using _service weights_ .
One canonical use case is canary releases where a deployment representing a newer release is to receive an initially small but ever-increasing fraction of the requests over time.
The way this can be done in Træfik is to specify a percentage of requests that should go into each deployment.
For instance, say that an application `my-app` runs in version 1.
A newer version 2 is about to be released, but confidence in the robustness and reliability of new version running in production can only be gained gradually.
Thus, a new deployment `my-app-canary` is created and scaled to a replica count that suffices for a 1% traffic share.
Along with it, a Service object is created as usual.
The Ingress specification would look like this:
```yaml
apiVersion: extensions/v1beta1
kind: Ingress
metadata:
annotations:
traefik.ingress.kubernetes.io/service-weights: |
my-app: 99%
my-app-canary: 1%
name: my-app
spec:
rules:
- http:
paths:
- backend:
serviceName: my-app
servicePort: 80
path: /
- backend:
serviceName: my-app-canary
servicePort: 80
path: /
```
Take note of the `traefik.ingress.kubernetes.io/service-weights` annotation: It specifies the distribution of requests among the referenced backend services, `my-app` and `my-app-canary` .
With this definition, Træfik will route 99% of the requests to the pods backed by the `my-app` deployment, and 1% to those backed by `my-app-canary` .
Over time, the ratio may slowly shift towards the canary deployment until it is deemed to replace the previous main application, in steps such as 5%/95%, 10%/90%, 50%/50%, and finally 100%/0%.
A few conditions must hold for service weights to be applied correctly:
- The associated service backends must share the same path and host.
- The total percentage shared across all service backends must yield 100% (see the section on [omitting the final service ](#omitting-the-final-service ), however).
- The percentage values are interpreted as floating point numbers to a supported precision as defined in the [annotation documentation ](/configuration/backends/kubernetes#general-annotations ).
### Omitting the Final Service
When specifying service weights, it is possible to omit exactly one service for convenience reasons.
For instance, the following definition shows how to split requests in a scenario where a canary release is accompanied by a baseline deployment for easier metrics comparison or automated canary analysis:
```yaml
apiVersion: extensions/v1beta1
kind: Ingress
metadata:
annotations:
traefik.ingress.kubernetes.io/service-weights: |
my-app-canary: 10%
my-app-baseline: 10%
name: app
spec:
rules:
- http:
paths:
- backend:
serviceName: my-app-canary
servicePort: 80
path: /
- backend:
serviceName: my-app-baseline
servicePort: 80
path: /
- backend:
serviceName: my-app-main
servicePort: 80
path: /
```
This configuration assigns 80% of traffic to `my-app-main` automatically, thus freeing the user from having to complete percentage values manually.
This becomes handy when increasing shares for canary releases continuously.
2017-09-12 17:56:05 +00:00
## Production advice
### Resource limitations
The examples shown deliberately do not specify any [resource limitations ](https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/ ) as there is no one size fits all.
In a production environment, however, it is important to set proper bounds, especially with regards to CPU:
2018-02-21 16:34:04 +00:00
- too strict and Træfik will be throttled while serving requests (as Kubernetes imposes hard quotas)
- too loose and Træfik may waste resources not available for other containers
2017-09-12 17:56:05 +00:00
When in doubt, you should measure your resource needs, and adjust requests and limits accordingly.