What is Kubernetes?
In this article, we review what Kubernetes and Kubeadm are, how to install, create a cluster, and set up worker nodes using Kubeadm. If you are not yet familiar with Kubernetes, we recommend reading our article on the fundamental basics of Kubernetes.Continue reading “How to Install Kubernetes Using Kubeadm on Ubuntu 18”
This article will demonstrate how to install and configure Minikube to set up a small Kubernetes cluster. We will then examine Kubernetes in more detail to apply that knowledge to a real-world example.Continue reading “How to Install and Configure Minikube on Ubuntu”
What is Kubernetes?
Kubernetes (or K8s) is an open-source container orchestration system for automating computer application deployment, scaling, and management. Kubernetes manages and runs Docker containers on numerous hosts. The project was started by Google and is supported by many companies, including Microsoft, RedHat, IBM.Continue reading “How to Install Kubernetes on Ubuntu, macOS, and Windows”
What is K3s?
K3s is a lightweight version of Kubernetes. It is a highly available Kubernetes certified distribution designed for production workloads in unattended, limited resource, remote locations, or inside an IoT appliance. The developers of K3s declare that K3s is capable of almost everything that K8s can do.Continue reading “How to Install and Configure K3s on Ubuntu 18.04”
In this tutorial, we will be reviewing what the fundamental similarities and distinctions are between Kubernetes and Docker Swarm. Kubernetes and Docker are two of the major players in container orchestration. Both Kubernetes and Docker Swarm continue to grow in popularity as they are increasingly used by those working with container deployment, orchestration, and management. Across all vertical markets, businesses continue to find new methods of utilization and practice with more uses constantly being discovered.Continue reading “Kubernetes vs. Docker Swarm: A Comparison”
What is Containerization?
Containerization is a form of virtualized operating system developed as a response to the many problems of hardware-level virtualization. Because the latter runs a full-blown guest operating system, it is very resource-intensive and incurs a significant amount of overhead, but containerization is much lighter. Since the containers share the host machine’s kernel, the resources are not wasted on running separate operating system tasks. This allows for a much quicker and lightweight deployment of applications.Continue reading “How to Install and Use Containerization”
What is a Service Mesh?
A service mesh is a layer of communication and control between applications or microservices and the network stack they typically communicate over. This layer controls communication and helps microservices share data. This service-to-service interaction is governed by logic built into the service mesh layer. Simply put, microservices are small or partial applications, or individual functions, and the network stack is the physical networking layer.
How is a Service Mesh Achieved?
Typically, a service mesh is implemented via sidecars. These sidecars are usually attached to every microservice on your network. The service mesh abstracts communication between microservices into these sidecars and the sidecars communicate with one another in a mesh topology.
Because most modern apps are now smaller than their predecessors, they have shrunken the communication logic written into each app, microservice, function, etc. With one or two microservices, or maybe even a dozen, the communication logic generally isn’t difficult to deal with. At scale, or when you have hundreds of such apps, a service mesh can be thought of as ‘required’.
So, a service mesh consists of the following components:
- A communication layer of control between apps or microservices
- The network stack they typically communicate over
- Abstracts communication and implements ‘sidecars’ with every microservice
If you’re used to using services such as GitHub or Bitbucket for continuous integration, chances are you’re wondering how you can setup continuous deployment for your website on Cloud Sites. Since Cloud Sites doesn’t have git or SSH access, you might think it’s impossible. Luckily, with a service like DeployBot and their SFTP deployment tools, it is actually very simple to deploy your code from your repository to Cloud Sites with a simple click. Here’s a quick tour on how to get up and running with DeployBot and Cloud Sites.