How to Install Kubernetes (k8s) 1.7 on CentOS 7 / RHEL 7

Kubernetes is a cluster and orchestration engine for docker containers. In other words Kubernetes is  an open source software or tool which is used to orchestrate and manage docker containers in cluster environment. Kubernetes is also known as k8s and it was developed by Google and donated to “Cloud Native Computing foundation”

In Kubernetes setup we have one master node and multiple nodes. Cluster nodes is known as worker node or Minion. From the master node we manage the cluster and its nodes using ‘kubeadm‘ and ‘kubectl‘  command.

Kubernetes can be installed and deployed using following methods:

  • Minikube ( It is a single node kubernetes cluster)
  • Kops ( Multi node kubernetes setup into AWS )
  • Kubeadm ( Multi Node Cluster in our own premises)

In this article we will install latest version of Kubernetes 1.7 on CentOS 7 / RHEL 7 with kubeadm utility. In my setup I am taking three CentOS 7 servers with minimal installation. One server will acts master node and rest two servers will be minion or worker nodes.

Kubernetes-settup-Diagram

On the Master Node following components will be installed

  • API Server  – It provides kubernetes API using Jason / Yaml over http, states of API objects are stored in etcd
  • Scheduler  – It is a program on master node which performs the scheduling tasks like launching containers in worker nodes based on resource availability
  • Controller Manager – Main Job of Controller manager is to monitor replication controllers and create pods to maintain desired state.
  • etcd – It is a Key value pair data base. It stores configuration data of cluster and cluster state.
  • Kubectl utility – It is a command line utility which connects to API Server on port 6443. It is used by administrators to create pods, services etc.

On Worker Nodes following components will be installed

  • Kubelet – It is an agent which runs on every worker node, it connects to docker  and takes care of creating, starting, deleting containers.
  • Kube-Proxy – It routes the traffic to appropriate containers based on ip address and port number of the incoming request. In other words we can say it is used for port translation.
  • Pod – Pod can be defined as a multi-tier or group of containers that are deployed on a single worker node or docker host.

Installations Steps of Kubernetes 1.7 on CentOS 7 / RHEL 7

Perform the following steps on Master Node

Step 1: Disable SELinux & setup firewall rules

Login to your kubernetes master node and set the hostname and disable selinux using following commands

~]# hostnamectl set-hostname 'k8s-master'
~]# exec bash
~]# setenforce 0
~]# sed -i --follow-symlinks 's/SELINUX=enforcing/SELINUX=disabled/g' /etc/sysconfig/selinux

Set the following firewall rules.

[root@k8s-master ~]# firewall-cmd --permanent --add-port=6443/tcp
[root@k8s-master ~]# firewall-cmd --permanent --add-port=2379-2380/tcp
[root@k8s-master ~]# firewall-cmd --permanent --add-port=10250/tcp
[root@k8s-master ~]# firewall-cmd --permanent --add-port=10251/tcp
[root@k8s-master ~]# firewall-cmd --permanent --add-port=10252/tcp
[root@k8s-master ~]# firewall-cmd --permanent --add-port=10255/tcp
[root@k8s-master ~]# firewall-cmd --reload
[root@k8s-master ~]# modprobe br_netfilter
[root@k8s-master ~]# echo '1' > /proc/sys/net/bridge/bridge-nf-call-iptables

Note: In case you don’t have your own dns server then update /etc/hosts file on master and worker nodes

192.168.1.30 k8s-master
192.168.1.40 worker-node1
192.168.1.50 worker-node2

Disable Swap in all nodes using “swapoff -a” command and remove or comment out swap partitions or swap file from fstab file

Step 2: Configure Kubernetes Repository

Kubernetes packages are not available in the default CentOS 7 & RHEL 7 repositories, Use below command to configure its package repositories.

[root@k8s-master ~]# cat <<EOF > /etc/yum.repos.d/kubernetes.repo
> [kubernetes]
> name=Kubernetes
> baseurl=https://packages.cloud.google.com/yum/repos/kubernetes-el7-x86_64
> enabled=1
> gpgcheck=1
> repo_gpgcheck=1
> gpgkey=https://packages.cloud.google.com/yum/doc/yum-key.gpg
>         https://packages.cloud.google.com/yum/doc/rpm-package-key.gpg
> EOF [root@k8s-master ~]#

Step 3: Install Kubeadm and Docker

Once the package repositories are configured, run the beneath command to install kubeadm and docker packages.

[root@k8s-master ~]# yum install kubeadm docker -y

Start and enable kubectl and docker service

[root@k8s-master ~]# systemctl restart docker && systemctl enable docker
[root@k8s-master ~]# systemctl  restart kubelet && systemctl enable kubelet

Step 4: Initialize Kubernetes Master with ‘kubeadm init’

Run the beneath command to  initialize and setup kubernetes master.

[root@k8s-master ~]# kubeadm init

Output of above command would be something like below

kubeadm-init-output

As we can see in the output that kubernetes master has been initialized successfully. Execute the beneath commands to use the cluster as root user.

[root@k8s-master ~]# mkdir -p $HOME/.kube
[root@k8s-master ~]# cp -i /etc/kubernetes/admin.conf $HOME/.kube/config
[root@k8s-master ~]# chown $(id -u):$(id -g) $HOME/.kube/config

Step 5: Deploy pod network to the cluster

Try to run below commands to get status of cluster and pods.

kubectl-get-nodes

To make the cluster status ready and kube-dns status running, deploy the pod network so that containers of different host communicated each other.  POD network is the overlay network between the worker nodes.

Run the beneath command to deploy network.

[root@k8s-master ~]# export kubever=$(kubectl version | base64 | tr -d '\n')
[root@k8s-master ~]# kubectl apply -f "https://cloud.weave.works/k8s/net?k8s-version=$kubever"
serviceaccount "weave-net" created
clusterrole "weave-net" created
clusterrolebinding "weave-net" created
daemonset "weave-net" created
[root@k8s-master ~]#

Now run the following commands to verify the status

[root@k8s-master ~]# kubectl get nodes
NAME         STATUS    AGE       VERSION
k8s-master   Ready     1h        v1.7.5
[root@k8s-master ~]# kubectl  get pods  --all-namespaces
NAMESPACE     NAME                                 READY     STATUS    RESTARTS   AGE
kube-system   etcd-k8s-master                      1/1       Running   0          57m
kube-system   kube-apiserver-k8s-master            1/1       Running   0          57m
kube-system   kube-controller-manager-k8s-master   1/1       Running   0          57m
kube-system   kube-dns-2425271678-044ww            3/3       Running   0          1h
kube-system   kube-proxy-9h259                     1/1       Running   0          1h
kube-system   kube-scheduler-k8s-master            1/1       Running   0          57m
kube-system   weave-net-hdjzd                      2/2       Running   0          7m
[root@k8s-master ~]#

Now let’s add worker nodes to the Kubernetes master nodes.

Perform the following steps on each worker node

Step 1: Disable SELinux & configure firewall rules on both the nodes

Before disabling SELinux set the hostname on the both nodes as ‘worker-node1’ and ‘worker-node2’ respectively

~]# setenforce 0
~]# sed -i --follow-symlinks 's/SELINUX=enforcing/SELINUX=disabled/g' /etc/sysconfig/selinux
~]# firewall-cmd --permanent --add-port=10250/tcp
~]# firewall-cmd --permanent --add-port=10255/tcp
~]# firewall-cmd --permanent --add-port=30000-32767/tcp
~]# firewall-cmd --permanent --add-port=6783/tcp
~]# firewall-cmd  --reload
~]# echo '1' > /proc/sys/net/bridge/bridge-nf-call-iptables

Step 2: Configure Kubernetes Repositories on both worker nodes

~]# cat <<EOF > /etc/yum.repos.d/kubernetes.repo
> [kubernetes]
> name=Kubernetes
> baseurl=https://packages.cloud.google.com/yum/repos/kubernetes-el7-x86_64
> enabled=1
> gpgcheck=1
> repo_gpgcheck=1
> gpgkey=https://packages.cloud.google.com/yum/doc/yum-key.gpg
>         https://packages.cloud.google.com/yum/doc/rpm-package-key.gpg
> EOF

Step 3: Install kubeadm and docker package on both nodes

[root@worker-node1 ~]# yum  install kubeadm docker -y
[root@worker-node2 ~]# yum  install kubeadm docker -y

Start and enable docker service

[root@worker-node1 ~]# systemctl restart docker && systemctl enable docker
[root@worker-node2 ~]# systemctl restart docker && systemctl enable docker

Step 4: Now Join worker nodes to master node

To join worker nodes to Master node, a token is required. Whenever kubernetes master initialized , then in the output we get command and token.  Copy that command and run on both nodes.

[root@worker-node1 ~]# kubeadm join --token a3bd48.1bc42347c3b35851 192.168.1.30:6443

Output of above command would be something like below

kubeadm-node1

[root@worker-node2 ~]# kubeadm join --token a3bd48.1bc42347c3b35851 192.168.1.30:6443

Output would be something like below

kubeadm-join-node2

Now verify Nodes status from master node using kubectl command

[root@k8s-master ~]# kubectl get nodes
NAME           STATUS    AGE       VERSION
k8s-master     Ready     2h        v1.7.5
worker-node1   Ready     20m       v1.7.5
worker-node2   Ready     18m       v1.7.5
[root@k8s-master ~]#

As we can see master and worker nodes are in ready status. This concludes that kubernetes 1.7 has been installed successfully and also we have successfully joined two worker nodes.  Now we can create pods and services.

Please share your feedback and comments in case this article helps you to install latest version of kubernetes 1.7

109 thoughts on “How to Install Kubernetes (k8s) 1.7 on CentOS 7 / RHEL 7”

  1. Can someone tell me how to determine the cluster CIDR of the master node which I have initialized using `kubeadm init` as detailed in this article? My master node is on CentOS 7 and I am trying to join a Windows node to the cluster and it requires me to pass the Cluster CIDR to the script.

  2. Hi

    I am getting Node status is “Not Ready” when it connected to master node for the first time.

    # kubectl get nodes
    NAME STATUS ROLES AGE VERSION
    labserver NotReady 1h v1.9.0
    rhel74 Ready master 3h v1.9.0

  3. thank so much for your blog.
    i am getting this error on ‘https://10.0.1.166:6443/’

    {
    “kind”: “Status”,
    “apiVersion”: “v1”,
    “metadata”: {

    },
    “status”: “Failure”,
    “message”: “forbidden: User \”system:anonymous\” cannot get path \”/\””,
    “reason”: “Forbidden”,
    “details”: {

    },
    “code”: 403
    }

    please help me

  4. Hi,

    After “kubeadm join” is executed on the node, weave-net fails to start.

    On the node, “journalctl -xe” shows:
    reflector.go:205] github.com/weaveworks/weave/prog/weave-npc/main.go:229: Failed to list *v1.Pod: Get ‘https://10.96.0.1:443/api/v1/pods?resourceVersion=0’: dial tcp 10.96.0.1:443: i/o timeout

    But the kube-apiserver is ‘https://192.168.56.109:6443’.

    And I tried with curl -k ‘https://192.168.56.109:6443//api/v1/pods’ on the node:
    {
    “kind”: “Status”,
    “apiVersion”: “v1”,
    “metadata”: {

    },
    “status”: “Failure”,
    “message”: “pods is forbidden: User \”system:anonymous\” cannot list pods at the cluster scope”,
    “reason”: “Forbidden”,
    “details”: {
    “kind”: “pods”
    },
    “code”: 403

    but no response with curl -k ‘https://10.96.0.1:443/api/v1/pods’.

    Any help is appreciated.
    Thanks.

  5. Could you share steps if the master server is restarted will the services come itself or what is the procedure you recommend to start and ensure node(minions) sync together.
    Also where does the etcd,kube-apiserver,scheduler,controller-manager,flannel/weave, configuration available.
    How to start individually

  6. Hey Pradeep

    Thanks for very useful tutorial
    I just successfully completed a cluster build, using k8s 1.9.3 and docker 1.12.6 (centos 7)

    Only two comments:
    1. When setting up the MASTER:
    no need to do `systemctl restart kubelet` before running the `kubeadm init`.
    Just do the `systemctl enable kublet` and then run `kubeadm init` which will set everything up and start the kublet service.

    If you try to start it before the init part – it will error out, complaining about being unable to load some CA certs.

    2. When adding the firewall rules for the worker nodes:
    the bridge config `echo ‘1’ > /proc/sys/net/bridge/bridge-nf-call-iptables` will fail, as the file does not exist yet.

    do the first step would be to:
    – install kubeadm and docker
    – enable and start docker service
    – enable kubelet service
    – join the node to cluster, which will automatically start kublet service

    ALSO:
    if you are using a minimal install from ISO (like I was – on virtual machines, with just default install settings). Make sure you disable swap !
    None of the kubeadm stuff will work if your machines have active swap (it will error out, complaining about it, asking you to disable it).

  7. I have made a cluseter , in which master is not in ready state , how master can be bring up to ready state and how to assign role of worker nodes?

    [root@k8s-master tmp]# kubectl get nodes
    NAME STATUS ROLES AGE VERSION
    k8s-master NotReady master 38m v1.9.3
    worker-node1 Ready 35m v1.9.3
    worker-node2 Ready 37m v1.9.3

    1. Even I have the similar issues; don’t know how to make it work yet…there should be very specific reason for it to not to come READY state..

      [root@etcd2 ~]# kubectl get nodes
      NAME STATUS ROLES AGE VERSION
      etcd1.hdp3.cisco.com Ready 48m v1.18.3
      etcd2.hdp3.cisco.com NotReady master 136m v1.18.3
      etcd3.hdp3.cisco.com Ready 48m v1.18.3

      [root@etcd2 ~]# kubectl get pods –all-namespaces
      NAMESPACE NAME READY STATUS RESTARTS AGE
      kube-system coredns-66bff467f8-25pg4 1/1 Running 0 133m
      kube-system coredns-66bff467f8-r64mg 1/1 Running 0 133m
      kube-system etcd-etcd2.hdp3.cisco.com 1/1 Running 1 133m
      kube-system kube-apiserver-etcd2.hdp3.cisco.com 1/1 Running 1 133m
      kube-system kube-controller-manager-etcd2.hdp3.cisco.com 1/1 Running 1 133m
      kube-system kube-proxy-9vm77 1/1 Running 0 45m
      kube-system kube-proxy-pk2bh 1/1 Running 1 133m
      kube-system kube-proxy-z2tv7 1/1 Running 0 46m
      kube-system kube-scheduler-etcd2.hdp3.cisco.com 1/1 Running 1 133m
      kube-system weave-net-6cnmh 2/2 Running 0 25m
      kube-system weave-net-h4gr5 2/2 Running 0 25m
      kube-system weave-net-tb4td 2/2 Running 0 25m

  8. Hi Pradeep,

    Thank you very much for sharing this. I was struggling to set kubernetes for a period of time. Your article helped a lot and I configured kubernetes cluster successfully. Thanks again.

  9. I successfully installed kubernetes cluster but getting error once i tried to access Web UI :

    {
    “kind”: “Status”,
    “apiVersion”: “v1”,
    “metadata”: {

    },
    “status”: “Failure”,
    “message”: “services \”kube-dns:dns\” is forbidden: User \”system:anonymous\” cannot get services/proxy in the namespace \”kube-system\””,
    “reason”: “Forbidden”,
    “details”: {
    “name”: “kube-dns:dns”,
    “kind”: “services”
    },
    “code”: 403
    }

Leave a Comment

Your email address will not be published. Required fields are marked *