Skip to content

Creating a Kubernetes cluster

If you want to create a Kubernetes cluster, you can do so via the Cleura Cloud Management Panel using Gardener. This guide shows you how to do that.

Prerequisites

  • If this is your first time using Gardener in Cleura Cloud, you need to activate the service from the Cleura Cloud Management Panel.
  • To access the Kubernetes cluster from your computer, you must install kubectl on your machine.

Creating a Kubernetes cluster in Cleura Cloud Management Panel

To get started, navigate to https://cleura.cloud and, in the side panel, choose Containers → Gardener. You will land on a page named Gardener / Shoot Clusters. There, you can create and manage your Gardener-based clusters. In Gardener terminology, a Kubernetes cluster is referred as a Shoot (as in, new plant growth).

At the top right-hand side of the central pane, click on Create Kubernetes cluster.

Gardener page in Cleura Cloud Management Panel

A new pane named Create Gardener Shoot Cluster slides over from the right-hand side of the browser. Type in a name for the new shoot cluster, and select a region. Also, select the version of Kubernetes the new cluster will be running. In the example below, we have chosen version 1.29.3, which, at the time of writing, was the latest supported in Cleura Cloud.

"Create Gardener Shoot Cluster" panel, showing options to type in a cluster name, select region and Kubernetes version

Next, you may accept the proposed network address (in CIDR notation) for the worker nodes, or type in the one you prefer.

Accept the proposed CIDR network address for the worker nodes or type in the CIDR network address you prefer

Alternatively, click on the drop-down menu at the right-hand side of “Network for worker nodes”. From the available menu items, choose a pre-existing network. Be sure to select one that has an assigned subnet, and is connected to a router.

Select a pre-existing network

Then, define a new subnet for the worker nodes by typing in a CIDR network address. If the new subnet overlaps with one of the subnets of the network you just selected, you will see an error message in red.

Select a pre-existing network

Instead, you should define a subnet that does not overlap with any of the subnets of your selected network.

New, valid subnet for the worker nodes

Now, scroll down a bit until you bring the Worker Groups section into full view. Make sure there is at least one worker group defined. (By default, you start with one.) Pay attention to the values you set for the following parameters:

  • Flavor: The flavor your worker nodes will use; this determines the number of CPU cores and the amount of RAM allocated to them.
  • Volume Size: The amount of local storage allocated to each worker node.
  • Autoscaler Min: The minimum number of worker nodes to run in the cluster at any time.
  • Autoscaler Max: The maximum number of worker nodes the cluster automatically scales to, in the event that the current number of nodes cannot handle the deployed workload.
  • Max Surge: The maximum number of additional nodes to deploy in an autoscaling event.

"Create Gardener Shoot Cluster" panel

For a test cluster, you can leave each parameter at its default value. Scroll to the bottom of the pane and click the green Create button.

"Create Gardener Shoot Cluster" panel

In the list of clusters, you will see your new Gardener shoot bootstrapping. The animated icon on the left-hand side of the cluster row marks the progress. Creating the cluster may take several minutes.

Shoot cluster bootstrapping

A note on quotas

Your Gardener worker nodes are subject to quotas applicable to your Cleura Cloud project. It would be best if you kept that in mind when selecting the worker node flavor, setting the volume size value, and also the Autoscaler Max value, so you will not be at risk of violating any quota.

For example, if your project is configured with the default quotas, and you select the b.4c16gb flavor for your worker nodes, your cluster would be able to run with a maximum of 3 worker nodes (since their total memory footprint would be 3×16=48 GiB, just short of the default 50 GiB limit). A 4th node would push your total memory allocation to 64 GiB, violating your quota.

If necessary, be sure to request a quota increase via our Service Center.

Viewing details and monitoring

After the new shoot cluster finishes bootstrapping, you may click on its row to bring all relevant details into view. For instance, click the Details tab to get networking information about the worker nodes.

Networking information about the worker nodes

Next, go to the Monitoring tab. There, you will notice two orange buttons: one for launching Prometheus, and one for launching Plutono.

Monitoring tab for launching Prometheus or Plutono

As an example, we have clicked the button for Plutono. Prior to launching the dashboard, a pop-up window appears. This shows the default username (admin), and the masked generated password. You get this pop-up every time you click on either of the orange buttons. In any case, to copy the password into the clipboard, click the related button shown below. Then, click the orange button labeled Open Dashboard.

Copy the Plutono password

After typing in the default username and pasting the password, you get access to the Prometheus or Plutono dashboard for your shoot cluster. In the example below, we have the Plutono dashboard displaying graphical information regarding the cluster nodes.

Plutono dashboard displaying information about the worker nodes

Interacting with your cluster

Once your new shoot cluster is operational, you can start interacting with it.