A Kubernetes cluster consists out of a control plane (master nodes) and worker nodes (running the workloads). The Kubernetes master is responsible for maintaining the desired state for your cluster. In case of using a managed Kubernetes service, the control plane is managed by the Cloud provider.
The result of deploying a managed Kubernetes service is a Kubernetes API and some hosts to run your workload on. From this point on you will use Kubernetes API objects to describe your cluster’s desired state. This requires a good understanding of Kubernetes. Kubernetes is complex and even after a couple of years working with it, you’ll still wonder if you got it all under control.
Let’s say you’re an IT Operations guy and your company has a couple of development teams who would like to use Kubernetes. They have asked you to set this up.
Maybe the first question you need to ask yourself is: Are we going to setup a cluster that will be used as a shared platform and can be used by multiple teams/projects, or are we going to provide teams/projects with their own cluster(s) and they need to figure it out themselves. Both will have their own challenges. Let’s zoom in:
In a shared setup you would offer Kubernetes as a Service to the development teams. The service offering would be to onboard teams easily and provide them with a private space where they can deploy their workloads and offer them a set of tools for logging, metrics and alerting. This sounds easy, but how would you do this? And:
We see multiple organizations delivering Kubernetes clusters to teams and projects as if they were donuts. Sometimes teams even have their own cloud subscription and can deploy whatever services they like. In this case the team itself is responsible for their own cluster. Some would say this enables better isolation, but:
What choice to make?
Yes, setting up a shared cluster for multiple teams/projects sounds like the most challenging, but is has some advantages:
Using multiple clusters probably allows for greater levels of isolation between tenants and customizing maintenance lifecycles, but how to manage all these clusters? Managing the cluster itself isn’t probably the hardest part, but setting up and managing all the API objects is!
Setting up a shared container platform based on Kubernetes is hard and will take a huge amount of time. There are however a couple of solutions that claim to offer an enterprise container management platform that provides solutions for all these challenges. But do they really? Creating a container platform on top of a managed Kubernetes service requires a tight integration with other cloud services like L7 load balancers, DNS and certificate management. Do you know any solutions that are capable of doing this and also offer multi tenancy and a unified platform experience for all teams, no matter if you use AKS, EKS or GKE?
We do! Take a look at Otomi Container Platform.