Ansible playbooks to building the hard way Kubernetes cluster, This playbook is a fully automated command to bring up a Kubernetes cluster on VM or Baremetal.
Feature list:
- Support build virtual cluster using vagrant.
- Kubernetes v1.7.0+.
- Kubernetes common addons.
- Support CNI(calico, flannel, ..., etc) and CRI(docker, containerd).
- Build HA using Keepalived and HAProxy.
- Ingress controller.
- Ceph on Kubernetes(v10.2.0+).
In this section you will deploy a cluster using vagrant.
Prerequisites:
- Ansible version: v2.4 (or newer).
- Vagrant: >= 1.7.0.
- VirtualBox: >= 5.0.0.
- Mac OS X need to install
sshpass
tool.
$ brew install http://git.io/sshpass.rb
The getting started guide will use Vagrant with VirtualBox. It can deploy your Kubernetes cluster with a single command:
$ ./tools/setup -m 2048 -n calico -i eth1
Cluster Size: 1 master, 2 worker.
VM Size: 1 vCPU, 2048 MB
VM Info: ubuntu16, virtualbox
CNI: calico, Binding iface: eth1
Start deploying?(y):
- Use
sudo ./tools/setup -p libvirt -i eth1
command to setup libvirt provider .
If you want to access API you need to create RBAC object define the permission of role. For example using cluster-admin
role:
$ kubectl create clusterrolebinding open-api --clusterrole=cluster-admin --user=system:anonymous
Login the addon's dashboard:
- Dashboard: https://API_SERVER:6443/api/v1/namespaces/kube-system/services/https:kubernetes-dashboard:/proxy/
- Logging: https://API_SERVER:6443/api/v1/namespaces/kube-system/services/kibana-logging/proxy/
- Monitor: https://API_SERVER:6443/api/v1/namespaces/kube-system/services/monitoring-grafana/proxy/
As of release 1.7 Dashboard no longer has full admin privileges granted by default, so you need to create a token to access the resources:
$ kubectl -n kube-system create sa dashboard
$ kubectl create clusterrolebinding dashboard --clusterrole cluster-admin --serviceaccount=kube-system:dashboard
$ kubectl -n kube-system get sa dashboard -o yaml
apiVersion: v1
kind: ServiceAccount
metadata:
creationTimestamp: 2017-11-27T17:06:41Z
name: dashboard
namespace: kube-system
resourceVersion: "69076"
selfLink: /api/v1/namespaces/kube-system/serviceaccounts/dashboard
uid: 56b880bf-d395-11e7-9528-448a5ba4bd34
secrets:
- name: dashboard-token-vg52j
$ kubectl -n kube-system describe secrets dashboard-token-vg52j
...
token: eyJhbGciOiJSUzI1NiIsInR5cCI6IkpXVCJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJkYXNoYm9hcmQtdG9rZW4tdmc1MmoiLCJrdWJlcm5ldGVzLmlvL3NlcnZpY2VhY2NvdW50L3NlcnZpY2UtYWNjb3VudC5uYW1lIjoiZGFzaGJvYXJkIiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9zZXJ2aWNlLWFjY291bnQudWlkIjoiNTZiODgwYmYtZDM5NS0xMWU3LTk1MjgtNDQ4YTViYTRiZDM0Iiwic3ViIjoic3lzdGVtOnNlcnZpY2VhY2NvdW50Omt1YmUtc3lzdGVtOmRhc2hib2FyZCJ9.bVRECfNS4NDmWAFWxGbAi1n9SfQ-TMNafPtF70pbp9Kun9RbC3BNR5NjTEuKjwt8nqZ6k3r09UKJ4dpo2lHtr2RTNAfEsoEGtoMlW8X9lg70ccPB0M1KJiz3c7-gpDUaQRIMNwz42db7Q1dN7HLieD6I4lFsHgk9NPUIVKqJ0p6PNTp99pBwvpvnKX72NIiIvgRwC2cnFr3R6WdUEsuVfuWGdF-jXyc6lS7_kOiXp2yh6Ym_YYIr3SsjYK7XUIPHrBqWjF-KXO_AL3J8J_UebtWSGomYvuXXbbAUefbOK4qopqQ6FzRXQs00KrKa8sfqrKMm_x71Kyqq6RbFECsHPA
Copy and paste the
token
to dashboard.
In this section you will manually deploy a cluster on your machines.
Prerequisites:
- Ansible version: v2.4.0 (or newer).
- Linux distributions: Ubuntu 16+/CentOS 7.x.(CoreOS and SUSE coming soon)
- All Master/Node should have password-less access from
Deploy
node.
For machine example:
IP Address | Role | CPU | Memory |
---|---|---|---|
172.16.35.9 | vip | - | - |
172.16.35.13 | master1 | 4 | 8G |
172.16.35.10 | node1 | 4 | 8G |
172.16.35.11 | node2 | 4 | 8G |
172.16.35.12 | node3 | 4 | 8G |
Add the machine info gathered above into a file called inventory
. For inventory example:
[etcds]
172.16.35.13
[masters]
172.16.35.13
[nodes]
172.16.35.[10:12]
[kube-cluster:children]
masters
nodes
[kube-addon:children]
masters
Set the variables in group_vars/all.yml
to reflect you need options. For example:
# Kubenrtes version, only support 1.7.0+.
kube_version: 1.8.4
# CRI plugin,
# Supported runtime: docker, containerd.
cri_plugin: docker
# CNI plugin,
# Supported network: flannel, calico, canal, weave or router.
network: calico
pod_network_cidr: 10.244.0.0/16
# CNI opts: flannel(--iface=enp0s8), calico(interface=enp0s8), canal(enp0s8).
cni_iface: interface=eth1
lb_vip_address: 172.16.35.9
# Extra addons
kube_dashboard: true
kube_monitoring: true
If everything is ready, just run cluster.yml
to deploy cluster:
$ ansible-playbook cluster.yml
And then run addons.yml
to create addons:
$ ansible-playbook addons.yml
You can reset cluster with the reset.yml
playbook:
$ ansible-playbook reset.yml
Now, check the service as follow:
$ kubectl get po,svc --namespace=kube-system
NAME READY STATUS RESTARTS AGE IP NODE
po/haproxy-master1 1/1 Running 0 2h 172.16.35.13 master1
...