DEV Community

Cover image for 5-Step Approach: Projectsveltos for Kubernetes Cluster Autoscaler Deployment on Hybrid-Cloud
Eleni Grosdouli
Eleni Grosdouli

Posted on • Edited on

5-Step Approach: Projectsveltos for Kubernetes Cluster Autoscaler Deployment on Hybrid-Cloud

Introduction

In the previous post, we talked about the Projectsveltos, and how it can be installed and used as a way to deploy Kubernetes add-on deployments on different environments, independent if they are on-prem or in the Cloud. The example provided was based on the Rancher Kubernetes Engine Government (RKE2) infrastructure in an on-prem environment.

Today, we will take a slightly different approach. We will work with a hybrid cloud setup. The cluster04 will act as our Sveltos management cluster, and the Kubernetes Cluster Autoscaler will get deployed to a Linode Kubernetes Engine (LKE) cluster with the assistance of Sveltos.

Diagram

ArgoCD - Sveltos - Kubernetes Cluster Autoscaler

Prerequisites

For this demonstration, I have already installed ArgoCD, deployed Sveltos to cluster04 and created an LKE cluster. For the first two, follow step 1 and step 2 from my previous post. If you want to learn how to create an LKE cluster, check out the link.

Note: While creating the LKE cluster, the below Node pools created.

LKE Node Pools

Lab Setup

- - - - - -+ - - - - - - - - - - + - - - - - - - - - - -+
| Cluster Name | Type | Version |
+ - - - - - - -+ - - - - - - - - - - + - - - - - - - - -+
| cluster04 | Management Cluster | RKE2 v1.26.11+rke2r1 |
| linode-autoscaler | Managed k8s Cluster | k8s v1.27.8 |
+ - - - - - -+ - - - - - - - - - - + - - - - - - - - - -+
Enter fullscreen mode Exit fullscreen mode

Step 1: Register the LKE Cluster with Sveltos

Once the LKE cluster is in a "Running" state, we will use the sveltosctl to register it. For the registration, we need three things: a service account, a kubeconfig associated with that account and a namespace. If you are unsure how to create a Service Account and an associated kubeconfig, there is a script publicly available to help you out.

Registration

$ sveltosctl register cluster --namespace=projectsveltos --cluster=linode-autoscaler --kubeconfig=linode-autoscaler.yaml
Enter fullscreen mode Exit fullscreen mode

Verification

$ kubectl get sveltoscluster -n projectsveltos
NAME                READY   VERSION
cluster01           true    v1.26.6+rke2r1
linode-autoscaler   true    v1.27.8
Enter fullscreen mode Exit fullscreen mode

Step 2: Cluster Labelling

To deploy and manage Kubernetes add-ons with the help of Sveltos, the concept of ClusterProfile and cluster labelling comes into play. ClusterProfile is the CustomerResourceDefinition used to instruct Sveltos which add-ons to deploy on a set of clusters.

For this demonstration, we will set the unique label 'env=autoscaler' as we want to differentiate this cluster from the other already existing cluster01.

$ kubectl get sveltoscluster -n projectsveltos --show-labels
NAME                READY   VERSION          LABELS
cluster01           true    v1.26.6+rke2r1   env=test,sveltos-agent=present
linode-autoscaler   true    v1.27.8          sveltos-agent=present
Enter fullscreen mode Exit fullscreen mode

Add labels

$ kubectl label sveltosclusters linode-autoscaler env=autoscaler -n projectsveltos
Enter fullscreen mode Exit fullscreen mode

Verification

$ kubectl get sveltoscluster -n projectsveltos --show-labels
NAME                READY   VERSION          LABELS
cluster01           true    v1.26.6+rke2r1   env=test,sveltos-agent=present
linode-autoscaler   true    v1.27.8          env=autoscaler,sveltos-agent=present
Enter fullscreen mode Exit fullscreen mode

Step 3: Kubernetes Cluster Autoscaler ClusterProfile

The Cluster Autoscaler will get deployed with the Helm chart. Apart from that, we need a Kubernetes secret to store the cloud-config for the Linode environment.

The whole deployment process is orchestrated and managed by Sveltos. The secret which contains the needed information alongside the Helm chart, will get deployed with the Sveltos ClusterProfile.

Note: The configuration is done on cluster04. Cluster04 is our Sveltos management cluster.

Secret

---
apiVersion: v1
kind: Secret
metadata:
  name: cluster-autoscaler-cloud-config
  namespace: autoscaler
type: Opaque
stringData:
  cloud-config: |-
    [global]
    linode-token={Your own Linode Token}
    lke-cluster-id=147978
    defaut-min-size-per-linode-type=1
    defaut-max-size-per-linode-type=2

    [nodegroup "g6-standard-1"]
    min-size=1
    max-size=2

    [nodegroup "g6-standard-2"]
    min-size=1
    max-size=2
Enter fullscreen mode Exit fullscreen mode
  • "linode-token": This will get replaced with your own token. To generate a Linode token check the link here.
  • "lke-cluster-id": You can get the ID of the LKE cluster with the following CURL request, curl -H "Authorization: Bearer $TOKEN" https://api.linode.com/v4/lke/clusters
  • "nodegroup": As mentioned above, the existing LKE cluster has two node pools. The name of the node pools is defined as the nodegroups in the secret above.

Note: The min and max numbers of the nodegroups can be customised based on your needs. As this is a demo environment, I wanted to keep the cost of the deployment as low as possible.

Create Secret (cluster04)

Before we can create the secret in a Sveltos ClusterProfile, first we need to create the secret locally and add the type 'addons.projectsveltos.io/cluster-profile'.

$ kubectl create secret generic cluster-autoscaler-cloud-config --from-file=secret.yaml --type=addons.projectsveltos.io/cluster-profile

$ kubectl get secret
NAME                              TYPE                                       DATA   AGE
cluster-autoscaler-cloud-config   addons.projectsveltos.io/cluster-profile   1      2m25s
Enter fullscreen mode Exit fullscreen mode

Create ClusterProfile

---
apiVersion: config.projectsveltos.io/v1alpha1
  kind: ClusterProfile
  metadata:
    name: cluster-autoscaler
  spec:
    clusterSelector: env=autoscaler
    syncMode: Continuous
    policyRefs:
    - deploymentType: Remote
      kind: Secret
      name: cluster-autoscaler-cloud-config
      namespace: default
    helmCharts:
    - chartName: autoscaler/cluster-autoscaler
      chartVersion: v9.34.1
      helmChartAction: Install
      releaseName: autoscaler-latest
      releaseNamespace: autoscaler
      repositoryName: autoscaler
      repositoryURL: https://kubernetes.github.io/autoscaler
      helmChartAction:  Install
      values: |
        autoDiscovery:
          clusterName: linode-autoscaler
        cloudProvider: linode
        extraVolumeSecrets:
          cluster-autoscaler-cloud-config:
            mountPath: /config
            name: cluster-autoscaler-cloud-config
        extraArgs:
          logtostderr: true
          stderrthreshold: info
          v: 2
          cloud-config: /config/cloud-config
        image:
          pullPolicy: IfNotPresent
          pullSecrets: []
          repository: registry.k8s.io/autoscaling/cluster-autoscaler
          tag: v1.28.2
Enter fullscreen mode Exit fullscreen mode
  • "policyRefs": Will create the secret with the name "cluster-autoscaler-cloud-config" in the cluster with the Sveltos tag set to"env=autoscaler" and in the namespace "autoscaler".

Note: the 'namespace: default' refers to the management cluster. Where cluter04 should look for the secret. In this case, it is the default namespace.

  • "helmCharts": Install the latest Cluster Autoscaler Helm chart

  • "values": Overwrite the default Helm values with the ones required from the Linode cloud provider. More details can be found here.

Apply ClusterProfile (cluster04)

$ kubectl apply -f "cluster-autoscaler.yaml"
Enter fullscreen mode Exit fullscreen mode

Verification

$ sveltosctl show addons
+----------------------------------+---------------+------------+---------------------------------+---------+-------------------------------+--------------------+
|             CLUSTER              | RESOURCE TYPE | NAMESPACE  |              NAME               | VERSION |             TIME              |  CLUSTER PROFILES  |
+----------------------------------+---------------+------------+---------------------------------+---------+-------------------------------+--------------------+
| projectsveltos/linode-autoscaler | helm chart    | autoscaler | autoscaler-latest               | 9.34.1  | 2024-01-05 12:15:54 +0000 UTC | cluster-autoscaler |
| projectsveltos/linode-autoscaler | :Secret       | autoscaler | cluster-autoscaler-cloud-config | N/A     | 2024-01-05 12:16:03 +0000 UTC | cluster-autoscaler |
+----------------------------------+---------------+------------+---------------------------------+---------+-------------------------------+--------------------+
Enter fullscreen mode Exit fullscreen mode

Verification - linode-autoscaler

$ kubectl get pods -n autoscaler
NAME                                                           READY   STATUS    RESTARTS   AGE
autoscaler-latest-linode-cluster-autoscaler-5fd8bfbb6f-r5zf6   1/1     Running   0          5m25s
Enter fullscreen mode Exit fullscreen mode

Step 4: Cluster Multi Replica Deployment

We will use Sveltos to create a busybox deployment on the LKE cluster with 500 replicas.

Create Deployment and Configmap Resource (cluster04)

---
apiVersion: apps/v1
kind: Deployment
metadata:
  name: busybox-workload
  namespace: default
  labels:
    app: busybox
spec:
  replicas: 500
  strategy:
    type: RollingUpdate
  selector:
    matchLabels:
      app: busybox
  template:
    metadata:
      labels:
        app: busybox
    spec:
      containers:
      - name: busybox
        image: busybox
        imagePullPolicy: IfNotPresent
        command: ['sh', '-c', 'echo Demo Workload ; sleep 600']
Enter fullscreen mode Exit fullscreen mode
$ kubectl create configmap busybox --from-file=busybox.yaml
Enter fullscreen mode Exit fullscreen mode

Create ClusterProfile for busybox Deployment (cluster04)

---
apiVersion: config.projectsveltos.io/v1alpha1
kind: ClusterProfile
metadata:
  name: busybox
spec:
  clusterSelector: env=autoscaler
  policyRefs:
  - kind: ConfigMap
    name: busybox
    namespace: default
Enter fullscreen mode Exit fullscreen mode
$ kubectl apply -f "clusterprofile-busybox-autoscaler.yaml"
Enter fullscreen mode Exit fullscreen mode
$ sveltosctl show addons --clusterprofile=busybox
+--------------------------+-----------------+-----------+------------------+---------+-------------------------------+------------------+
|         CLUSTER          |  RESOURCE TYPE  | NAMESPACE |       NAME       | VERSION |             TIME              | CLUSTER PROFILES |
+--------------------------+-----------------+-----------+------------------+---------+-------------------------------+------------------+
| projectsveltos/linode-autoscaler | apps:Deployment | default   | busybox-workload | N/A     | 2024-01-05 12:25:10 +0000 UTC | busybox          |
+--------------------------+-----------------+-----------+------------------+---------+-------------------------------+------------------+
Enter fullscreen mode Exit fullscreen mode

As the cluster has only two small nodes, it cannot keep up with the load of the 500 replicas. The autoscaler will kick in and add the required size of the nodes available from the pool.

Verification

$ kubernetes logs autoscaler-latest-linode-cluster-autoscaler-5fd8bfbb6f-r5zf6 -n autoscaler -f

I0103 17:39:49.638775       1 linode_manager.go:84] LKE node group after refresh:
I0103 17:39:49.638804       1 linode_manager.go:86] node group ID g6-standard-1 := min: 1, max: 2, LKEClusterID: 148988, poolOpts: {Count:1 Type:g6-standard-1 Disks:[]}, associated LKE pools: { poolID: 218632, count: 1, type: g6-standard-1, associated linodes: [ID: "218632-48499cbb0000", instanceID: 53677574] }
I0103 17:39:49.638826       1 linode_manager.go:86] node group ID g6-standard-2 := min: 1, max: 2, LKEClusterID: 148988, poolOpts: {Count:1 Type:g6-standard-2 Disks:[]}, associated LKE pools: { poolID: 218633, count: 1, type: g6-standard-2, associated linodes: [ID: "218633-4d4d521d0000", instanceID: 53677576] }
I0103 17:39:49.639930       1 klogx.go:87] Pod default/busybox-workload-5d94965f98-f558s is unschedulable
I0103 17:39:49.639939       1 klogx.go:87] Pod default/busybox-workload-5d94965f98-rfhqp is unschedulable
I0103 17:39:49.639943       1 klogx.go:87] Pod default/busybox-workload-5d94965f98-wsfcw is unschedulable
I0103 17:39:49.639946       1 klogx.go:87] Pod default/busybox-workload-5d94965f98-25s5c is unschedulable
I0103 17:39:49.639948       1 klogx.go:87] Pod default/busybox-workload-5d94965f98-r7fv4 is unschedulable
I0103 17:39:49.639951       1 klogx.go:87] Pod default/busybox-workload-5d94965f98-nmbxc is unschedulable
I0103 17:39:49.639954       1 klogx.go:87] Pod default/busybox-workload-5d94965f98-dbrh4 is unschedulable
E0103 17:39:49.640116       1 orchestrator.go:450] Couldn't get autoscaling options for ng: g6-standard-1
E0103 17:39:49.640137       1 orchestrator.go:450] Couldn't get autoscaling options for ng: g6-standard-2
I0103 17:39:49.640587       1 orchestrator.go:185] Best option to resize: g6-standard-1
I0103 17:39:49.641940       1 orchestrator.go:189] Estimated 1 nodes needed in g6-standard-1
I0103 17:39:49.641961       1 orchestrator.go:295] Final scale-up plan: [{g6-standard-1 1->2 (max: 2)}]
I0103 17:39:49.641972       1 executor.go:147] Scale-up: setting group g6-standard-1 size to 2
Enter fullscreen mode Exit fullscreen mode

Linode UI - Provisioning Nodes

Step 5: Undeploy Busybox

Finally, once the tests are complete, it is time to remove the busybox deployment from the cluster and allow 10 good minutes for the autoscaler to scale down the unneeded nodes. For the undeploy process, we only need to delete the busybox ClusterProfile.

Undeploy (LKE)

$ kubectl delete -f "clusterprofile-busybox-autoscaler.yaml"
Enter fullscreen mode Exit fullscreen mode

Verification

$ kubernetes logs autoscaler-latest-linode-cluster-autoscaler-5fd8bfbb6f-r5zf6 -n autoscaler -f

I0103 18:02:27.179597       1 linode_manager.go:84] LKE node group after refresh:
I0103 18:02:27.179813       1 linode_manager.go:86] node group ID g6-standard-1 := min: 1, max: 2, LKEClusterID: 148988, poolOpts: {Count:1 Type:g6-standard-1 Disks:[]}, associated LKE pools: { poolID: 218632, count: 1, type: g6-standard-1, associated linodes: [ID: "218632-48499cbb0000", instanceID: 53677574] }
I0103 18:02:27.179840       1 linode_manager.go:86] node group ID g6-standard-2 := min: 1, max: 2, LKEClusterID: 148988, poolOpts: {Count:1 Type:g6-standard-2 Disks:[]}, associated LKE pools: { poolID: 218633, count: 1, type: g6-standard-2, associated linodes: [ID: "218633-4d4d521d0000", instanceID: 53677576] }
I0103 18:02:27.181433       1 static_autoscaler.go:547] No unschedulable pods
I0103 18:02:27.181449       1 pre_filtering_processor.go:67] Skipping lke148988-218632-48499cbb0000 - node group min size reached (current: 1, min: 1)
I0103 18:02:27.181623       1 pre_filtering_processor.go:67] Skipping lke148988-218633-4d4d521d0000 - node group min size reached (current: 1, min: 1)
I0103 18:02:36.040964       1 node_instances_cache.go:156] Start refreshing cloud provider node instances cache
I0103 18:02:36.040995       1 node_instances_cache.go:168] Refresh cloud provider node instances cache finished, refresh took 12.52µs
I0103 18:02:38.019810       1 linode_manager.go:84] LKE node group after refresh:
I0103 18:02:38.019983       1 linode_manager.go:86] node group ID g6-standard-2 := min: 1, max: 2, LKEClusterID: 148988, poolOpts: {Count:1 Type:g6-standard-2 Disks:[]}, associated LKE pools: { poolID: 218633, count: 1, type: g6-standard-2, associated linodes: [ID: "218633-4d4d521d0000", instanceID: 53677576] }
I0103 18:02:38.020015       1 linode_manager.go:86] node group ID g6-standard-1 := min: 1, max: 2, LKEClusterID: 148988, poolOpts: {Count:1 Type:g6-standard-1 Disks:[]}, associated LKE pools: { poolID: 218632, count: 1, type: g6-standard-1, associated linodes: [ID: "218632-48499cbb0000", instanceID: 53677574] }
I0103 18:02:38.021154       1 static_autoscaler.go:547] No unschedulable pods
I0103 18:02:38.021181       1 pre_filtering_processor.go:67] Skipping lke148988-218632-48499cbb0000 - node group min size reached (current: 1, min: 1)
I0103 18:02:38.021187       1 pre_filtering_processor.go:67] Skipping lke148988-218633-4d4d521d0000 - node group min size reached (current: 1, min: 1)
Enter fullscreen mode Exit fullscreen mode

Linode UI - Undeploy

As expected, Sveltos took care of the complete lifecycle of the different Kubernetes deployments in a simple and straightforward manner!

👏 Support this project
Every contribution counts! If you enjoyed this article, check out the Projectsveltos GitHub repo. You can star 🌟 the project if you find it helpful.

The GitHub repo is a great resource for getting started with the project. It contains the code, documentation, and many more examples.

Thanks for reading!

Top comments (0)