create a kubernetes cluster

How To Create a Kubernetes Cluster Using Kubeadm on Ubuntu 18.04

This tutorial will guide you on setting up a Kubernetes cluster from scratch using Ansible and Kubeadm and further deploying a containerized Nginx application with it.

Introduction

Kubernetes (also known as k8s or “kube”) is an open-source container orchestration platform that automates many of the manual processes involved in deploying, managing, and scaling containerized applications. Kubernetes has a rapidly growing open-source community, actively contributing to the project. Take a look at our blog post that will introduce you to everything you need to know about the basics of the Kubernetes platform.

Kubeadm is a tool, which configures several integrated elements, parts, and pieces such as the API server, Controller Manager, and Kube DNS. It also helps in automating the installation. However, it doesn’t create users or handle the installation of operating-system-level dependencies and their configuration and cannot provision your infrastructure.

Ansible is an open-source tool for software provisioning and application deployment. Saltstack is open-source software for information technology automation driven by events. These are the two tools that make creating additional clusters or recreating existing clusters less vulnerable to errors and can be used for these preliminary tasks.

Goals:

Your cluster will include the following physical resources:

1. One master node:

A master node is a node that controls and manages a set of worker nodes (workloads runtime) and resembles a cluster in Kubernetes. It also holds the node’s resources plan to determine the proper action for the triggered event. It runs etcd, an open-source distributed key-value store used to hold and manage cluster data among components that schedule workloads to worker nodes.

For example, the scheduler would figure out which worker node will host a newly scheduled POD.

2. Two worker nodes:

Worker nodes are the nodes that carry on with their assigned work even if the master node goes down once scheduling is complete. Worker nodes are the servers where your workloads (i.e. containerized applications and services) will run. You can also increase the cluster’s capacity by adding workers.

Once you complete this tutorial, you will have a fully functional cluster ready to run workloads (i.e. containerized applications and services) assuming that the servers in the cluster have sufficient CPU and RAM resources for your applications to run. After you have successfully set up the cluster, you can run almost any traditional UNIX application. It could be containerized on your cluster, including web applications, databases, daemons, and command-line tools.

The cluster itself will consume around 300-500MB of memory and 10% of CPU on each node.

Prerequisites:

  1. You must have an SSH key pair on your local Linux machine and know how to use SSH keys. However, if you haven’t used SSH keys before, you can see this tutorial to help you set up SSH keys on your local machine.
  2. Three servers running Ubuntu 18.04 with at least 4GB RAM and 4 vCPUs each. You should be able to SSH into each server as the root user with your SSH key pair. Follow this tutorial to install your Ubuntu server.
  3. Ansible installed on your local machine.
  4. You must also be familiar with Ansible playbooks.
  5. You will also need to know how to launch a container from a Docker image. Look at “Step 5 — Working with Docker Images in Ubuntu” in How To Install and Use Docker on Ubuntu 18.04 if you need a refresher.

Step 1 — Setting up the Workspace Directory and Ansible Inventory File

You first need to set up Ansible on your local machine. It will help you execute commands on your remote server. It also eases the manual deployment effort by automating it. For this, you will need to create a directory on your local machine that will serve as your temporary digital storage area (Workspace).

Once you create a directory, you will create a hosts file to store all the information about each server’s IP addresses and group. It will help you store the inventory information within. As stated earlier, there would be three servers, one master, and two workers. The master server will be the master with an IP displayed as master_ip. The other two servers will be workers and will have the IPs worker_1_ip and worker_2_ip.

You need to create a directory named ~/kube-cluster in the home directory of your local machine and enter the directory using the cd command into it:

The ~/kube-cluster directory will now act as the temporary digital storage area (workspace) inside which you will run all local commands for creating a Kubernetes cluster using kubeadm. The directory will contain all of your Ansible playbooks and be used for the rest of the tutorial.

Creating Hosts File

Create a file named ~/kube-cluster/hosts using nano or your favorite text editor:

Now you will need to add the following text, which will specify information about the logical structure of your cluster:

As mentioned, that inventory file will help you store all the information about the IP addresses of your servers and the groups that each server belongs to. ~/kube-cluster/hosts will be your inventory file and (masters and workers) will be the two Ansible groups that you’ve added to it specifying the logical structure of your cluster.

The Master group is the group that specifies that Ansible should run remote commands as the root user. It also lists the master node’s IP (master_ip) that can be listed by the server entry named “master”. Likewise, the Workers group has two entries for the worker servers (worker_1_ip and worker_2_ip) that also specify the ansible_user as root.

The last line of the file tells Ansible to use the remote servers’ Python 3 interpreters for its management operations. At last, you need to save and close the file after you’ve added the text. After setting up the workspace directory and Ansible inventory file, let’s move on to the next step of installing operating system level dependencies and creating configuration settings.

Step 2 — Creating a Non-Root User on All Remote Servers

In this step, you will learn how to create a non-root user with sudo privileges on all servers so that you can SSH into them manually as an unprivileged user.

This can be useful for frequently performed operations for the preservation of a cluster. Moreover, this step will help you perform the task more accurately and less error-prone, decreasing the chances of altering or deleting important files unintentionally. If you want to change the setup of files owned by root or see system information with commands such as top/htop and view a list of running containers, the following step will help you perform all the tasks.

Creating the Playbook

Create a file named ~/kube-cluster/initial.yml in the workspace:

Next, you need to add the following play. A play in Ansible is a collection of steps to be performed that target specific servers and groups. There can be one or many plays in a playbook.

The following play will create a non-root sudo user:

Following is a breakdown of what our playbook does:

  1. This playbook will create the non-root user ubuntu.
  2. As you need to run sudo commands without a password prompt, this play will configure the sudoers file to allow the ubuntu user to do so.
  3. The main purpose of the above task was to allow you to SSH into each server as a ubuntu user. This playbook adds the public key of your local machine (usually ~/.ssh/id_rsa.pub) to the remote ubuntu user’s authorized key list.

Now, after adding the text, you need to save and close the file.

Running the playbook

After that, we need to execute our playbook that will create the non-root user ubuntu by simply running on local machines:

The execution of this command will take some time after which you will see the following output:

creating non root user create a kubernetes cluster

After this step is complete, you can move on to installing Kubernetes-specific dependencies in the next step.

Step 3 — Installing Kubernetes’ Dependencies

In this step, you will learn how to install the operating-system-level packages required by Kubernetes with Ubuntu’s package manager.

These packages are:

  1. Docker: Docker is a platform and tool for building, distributing, and running Docker containers. You can easily set up Docker by following our tutorial on how to install & operate Docker on Ubuntu in the public cloud. However, support for other runtimes such as rkt is under active development in Kubernetes.
  2. Kubeadm: kubeadm is a CLI tool that performs the actions necessary to get a minimum viable cluster up and running. That will help you install and build various components of the cluster in a standard way.
  3. kubelet: The kubelet is the primary “node agent” that runs on each node and handles node-level operations.
  4. kubectl: kubectl is also a CLI tool that talks to your cluster and issues commands through its API Server.
Creating the playbook

Create a file named ~/kube-cluster/kube-dependencies.yml in the workspace:

Now you need to add the following plays to the file to install these following packages to your servers:

The first play in the playbook does the following:

  1. This play will help you install operating-system-level packages, Docker –  the container runtime.
  2. It Installs apt-transport-https, which allows you to add external HTTPS sources to your APT sources list.
  3. Adds the Kubernetes APT repository’s apt-key for key verification.
  4. Adds the Kubernetes APT repository to your remote servers’ APT sources list.
  5. Installs kubelet and kubeadm.

The second play performs an important and solo task that includes installing kubectl on your master node. Now, after adding the text you need to save and close the file.

Running the playbook

After that, we need to execute our playbook  by simply running on local machines:

The execution of this command will take some time after which you will see the following output:

installing k8s dependencies create a kubernetes cluster

After execution, Docker, kubeadm and kubelet will be installed on all of the remote servers. Kubectl is not a required component and is only needed for executing cluster commands. Installing it only on the master node makes sense in this context since you will run kubectl commands only from the master. Note, however, that kubectl commands can be run from any of the worker nodes or from any machine where it can be installed and configured to point to a cluster.

All system dependencies are now installed. Let’s set up the master node and initialize the cluster.

Step 4 — Setting up the Master Node

In this step, you will learn a few concepts such as Pods and Pod Network Plugins since your cluster will include both once you set up your master node.

Pods are the smallest, most basic deployable objects in Kubernetes. Pods contain one or more containers, such as Docker containers. When a Pod runs multiple containers, the containers are managed as a single entity and share the Pod’s resources.

Each pod has its own IP address, and a pod on one node should be able to access a pod on another node using the pod’s IP. However, the communication between pods is more complex. It needs a separate component that can transparently route traffic from a pod on one node to a pod on another. Pod network plugins are used for this functionality. Many pod network plugins are available, but we will use a Flannel as it is a stable and efficient option.

Creating the Playbook

Create an Ansible playbook named master.yml on your local machine:

Further, you need to add the following play to the file to initialize the cluster and install Flannel:

Here’s a breakdown of this play:

  1. The first task in this play will set up the cluster by running kubeadm init. For specifying the private subnet that the pod IPs will be assigned we pass the argument --pod-network-cidr=10.244.0.0/16. Flannel uses the above subnet by default. We are using this to tell kubeadm to use the same subnet.
  2. The second task is used for creating a .kube directory at /home/ubuntu. Configuration information such as the admin key files, which are required to connect to the cluster and the cluster’s API address, will be held by this directory.
  3. The third task is used for copying the /etc/kubernetes/admin.conf file that was generated from kubeadm init to your non-root user’s home directory. This will allow you to use kubectl to access the newly-created cluster.
  4. The last task runs kubectl apply to install Flannel. kubectl apply -f descriptor.[yml|json] is the syntax for telling kubectl to create the objects described in the descriptor.[yml|json] file. The kube-flannel.yml file contains the descriptions of objects required for setting up Flannel in the cluster.

Now, after adding the text you need to save and close the file.

Running the Playbook

After that, you need to execute our playbook by simply running on local machines:

The execution of this command will take some time after which you will see the following output:

setting up master node create a kubernetes cluster

Now SSH into it with the following command to check the status of the master node :

Once inside the master node, execute:

You will now see the following output:

get nodes

On getting the above output you can proclaim that all the setup tasks have been accomplished by the master node and can start accepting worker nodes and executing tasks as it enters in a Ready state. You can now add the workers from your local machine.

Step 5 — Setting up the Worker Nodes

After setting the master node now we can move to our next step of setting up the worker nodes. Adding worker nodes to the cluster can be simply done by executing a single command on each worker server. The important information such as the IP address, port of the master’s API server, and a secure token are included in this command. Although you should note that all the nodes will not be able to join the cluster, only those nodes will be able to join the cluster that pass in the secure token.

Creating the playbook

This command will help you navigate back to your workspace and create a playbook named workers.yml:

Add the following text to the file to add the workers to the cluster:

Here is what the playbook does. There are two plays in the above code:

  1. The first play is used for getting the joining command that needs to be run on the worker nodes. The format of the command will be: kubeadm join --token sha256:<hash><token><master-ip>:<master-port> --discovery-token-ca-cert-hash sha256:<hash>;. The task needs to get the correct token and hash values. Once it gets the correct input the task sets it as a fact so that the second play will be able to access that info.
  2. The second play is written only to perform a solo task – to make the two worker nodes be a part of the cluster by simply running the joint command on all worker nodes.

After adding the text, you need to save and close the file.

Running the playbook

After that, we need to execute our playbook by running the following command on the worker machines:

The execution of this command will take some time after which you will see the following output:

setting up worker nodes

Now, your Kubernetes cluster is fully set up and functional with the workers ready to run workloads. Before moving to the next step, let’s verify that the cluster is working as planned.

Step 6 — Verifying the Cluster

There could be instances where a cluster fails during the setup. It could be because of a network error between the master and worker, or a node issue. So we need to verify the cluster before scheduling applications and ensure that no malfunctioning occurs. For this, you will need to check the current state of the cluster from the master node to ensure that the nodes are ready. You can get back the connection with the following command if the nodes are not ready or you get disconnected:

Use the following commands to get the status of the cluster:

The execution of this command will take some time after which you will see the following output:

get nodes

 

 

 

 

You need to check whether all nodes that are a part of the cluster are in the ready state. If a few nodes have Not Ready as the STATUS, it shows that the worker nodes haven’t finished their setup yet. However, before re-running kubectl get nodes and checking the updated output you should wait for another five to ten minutes. If some of the nodes still show Not Ready as their status, you should go and verify the previous steps and re-run the commands. Only if the nodes have the value Ready for STATUS, they are part of the cluster and ready to run workloads. After successfully executing the 6th step, your cluster is now verified. Now let’s schedule an example Nginx application on the cluster.

Step 7 — Running an Application on the Cluster

Creating Deployment

After creating the cluster successfully, you can deploy any containerized application to your cluster. You can use the following commands below for other containerized applications if you are within the master node. Next, execute the following command to create a deployment named nginx :

You need to change the Docker image name and any relevant flags (such as ports and volumes). To keep things familiar, you can deploy Nginx using deployments and services to see how applications can be deployed to the cluster.

A Kubernetes deployment is a resource object in Kubernetes that provides declarative updates to applications. A deployment allows you to describe an application’s life cycle, such as container image, replicas, and the update strategy. A deployment ensures the desired number of pods are running and available at all times. If a pod crashes during the cluster’s lifetime, it spawns it again. The update process is also wholly recorded, and versioned with options to pause, continue, and roll back to previous versions. The above command to create a deployment named Nginx will help you deploy a pod with one container from the Docker registry’s Nginx Docker Image.

Setting up Node Port

Next, we need to create a NodePort. NodePort is an open port on every node of your cluster. Kubernetes transparently routes incoming traffic on the NodePort to your service, even if your application is running on another node. For this we can use this command to create a NodePort resource named Nginx that will expose the app publicly:

A service is another Kubernetes object responsible for exposing an interface to those pods, which enables network access from either within the cluster or between external processes and the service. It can be defined as an abstraction on the top of the pod which provides a single IP address and DNS name by which pods can be accessed. With service, it is very easy to manage load balancing configuration.

Run the following command:

This will output text similar to the following:

get services

After getting the output, Kubernetes will assign a random port that is greater than 30000 automatically, while also making sure that the port assigned is not already bound by another service. The third line of the above output will help you retrieve the port that Nginx is running on.

To verify it is working visit http://worker_1_ip:nginx_port or http://worker_2_ip:nginx_port through a browser on your local machine. You will see Nginx’s familiar welcome page.

Removing Deployment

If you would like to remove the Nginx application, you need to first delete the nginx service from the master node:

For verifying that the application is finally deleted, you need to run this command:

You will get the following output:

check services

Post that, you need to delete the deployment using the following command:

You can use this command to verify if the deployment is finally deleted:

get deployments

Conclusion:

This tutorial will help you properly set up a cluster on Ubuntu 18.04 using Kubeadm and Ansible. Now that your cluster is set up you can easily start deploying your own applications and services.

Here is a list of links with additional details that will guide you in the process:

  1. Dockerizing applications – This link contains examples that guide you on how to load applications using Docker. Such as Dockerizing PostgreSQL, a CouchDB service, etc.
  2. Pod Overview – This link exhibits details on how to use a pod, the functioning of pods, and how pods are related to other Kubernetes objects. Pods are an important part of Kubernetes so understanding them will help you to succeed in your task.
  3. Deployments Overview – It will help you learn about deployments. A deployment provides declarative updates for Pods and ReplicaSets. You will learn how to update, roll over, and roll back a deployment.
  4. Services Overview -This link will guide you about cover services that are another frequently used object in Kubernetes clusters. A service in Kubernetes is an abstraction that defines a logical set of Pods and a policy by which you can access them. Understanding the types of services and the options they have are essential for running both stateless and stateful applications.

Furthermore, take a look at our other tutorials focusing on Docker and Kubernetes that you can find on our blog:

There are also many other important concepts such as Volumes, Ingresses, and Secrets that you can use while deploying production applications.

Happy Computing!