DEV Community

Cover image for Getting Started with Longhorn Distributed Block Storage and Cloud Native Distributed SQL
Jimmy Guerrero for YugabyteDB Distributed PostgreSQL Database

Posted on • Updated on

Getting Started with Longhorn Distributed Block Storage and Cloud Native Distributed SQL

Longhorn is cloud native distributed block storage for Kubernetes that is easy to deploy and upgrade, 100 percent open source, and persistent. Longhorn’s built-in incremental snapshot and backup features keep volume data safe, while its intuitive UI makes scheduling backups of persistent volumes easy to manage. Using Longhorn, you get maximum granularity and control, and can easily create a disaster recovery volume in another Kubernetes cluster and fail over to it in the event of an emergency.

YugabyteDB is a cloud native, distributed SQL database that runs in Kubernetes environments, so it can interoperate with Longhorn and many other CNCF projects. If you’re not familiar with YugabyteDB, it is an open source, high-performance distributed SQL database built on a scalable and fault-tolerant design inspired by Google Spanner. Yugabyte’s SQL API (YSQL) is PostgreSQL wire compatible.

If you are an engineer looking for a way to easily start your application development on top of a 100 percent cloud native infrastructure, this article is for you. In this article, we’ll give you step-by-step instructions on how to deploy a completely cloud native infrastructure stack. The stack will consist of Google Kubernetes Engine, Rancher enterprise Kubernetes management tooling, Longhorn distributed block storage, and a YugabyteDB distributed SQL database.

Why Longhorn and YugabyteDB?

YugabyteDB is deployed as a StatefulSet on Kubernetes and requires persistent storage. Longhorn can be used for backing YugabyteDB local disks, allowing the provisioning of large-scale persistent volumes. Here are a few benefits to using Longhorn and YugabyteDB together:

  • There’s no need to manage the local disks — they are managed by Longhorn.
  • Longhorn and YugabyteDB can provision large-sized persistent volumes.
  • Both Longhorn and YugabyteDB support multi-cloud deployments, helping organizations avoid cloud lock-in.

Additionally, Longhorn can do synchronous replication inside a geographic region. In a scenario where YugabyteDB is deployed across regions, and a node in any one region fails, YugabyteDB would have to rebuild this node with data from another region, which would incur cross-region traffic. This could prove to be more expensive and yield lower recovery performance. With Longhorn and YugabyteDB working together, you can rebuild the node seamlessly because Longhorn replicates locally inside the region. This means YugabyteDB does not end up having to copy data from another region, which ends up being less expensive and higher in performance. In this deployment setup, YugabyteDB would only need to do a cross-region node rebuild if the entire region failed.

Prerequisites

Below is the environment that we’ll use to run a YugabyteDB cluster on top of a Google Kubernetes cluster with Longhorn.

  • YugabyteDB (Using Helm Charts) – version 2.1.8
  • Rancher (Using Docker Run) – version 2.4.5
  • Longhorn (Using Rancher UI) – version 1.0
  • A Google Cloud Platform account

Setting Up a Kubernetes Cluster and Rancher on Google Cloud Platform

Rancher is an open source project that provides a container management platform built for Kubernetes. Rancher makes it easy to run Kubernetes everywhere, meet IT requirements, and empower DevOps teams.

Rancher requires a Linux host with 64-bit Ubuntu 16.04 or 18.04 and at least 4GB of memory. In this example, we will set up a Google Kubernetes Engine (GKE) cluster using the Rancher UI that we will first install on a GCP VM instance.

A summary of the steps required to set up a Kubernetes Cluster on GCP with Rancher includes:

  • Create a Service Account with the required IAM roles in GCP
  • Create a VM instance running Ubuntu 18.04
  • Install Rancher on the VM instance
  • Generate a Service Account private key
  • Set up the GKE cluster through the Rancher UI

Creating a Service Account with the Required IAM Roles, and a VM Instance Running Ubuntu

The first thing we need to do is create a Service Account attached to a GCP project. To do this, go to IAM & admin > Service accounts.

create a Service Account attached to a google cloud platform account distributed sql longhorn howto

Select Create New Service Account, give it a name and click create.

Next, we then need to add the required roles to the Service Account to be able to set up the Kubernetes cluster using Rancher. Add the roles shown below and create the Service Account.

add the required roles to the Service Account distributed sql longhorn tutorial

Once the roles have been added, click Continue and Done.

Now we now have to create an Ubuntu VM instance that is hosted on the GCP. To do this, go to Compute Engine > VM Instances > Create New Instance.

create an Ubuntu VM instance that is hosted on the GCP distributed sql longhorn how to

For the purposes of this demo I’ve selected an n1-standard-2 machine type. To select the Ubuntu image, click Boot Disk > Change and choose Ubuntu under Operating System and Ubuntu 18.04 LTS under Version.

choose Ubuntu under Operating System and Ubuntu 18.04 LTS under Version distributed sql longhorn how to

Make sure to check Firewall > Allow HTTPS traffic.

check Firewall and Allow HTTPS traffic distributed sql yugabytedb longhorn tutorial

Creating the VM instance with the above settings may take a few minutes. Once created connect to the VM using SSH. With a terminal connected to the VM, the next step is to install Rancher by executing the following command.

$ sudo docker run -d --restart=unless-stopped -p 80:80 -p 443:443 rancher/rancher
Enter fullscreen mode Exit fullscreen mode

Note: If Docker is not found, follow the instructions here to get it installed on your Ubuntu VM.

Installing Longhorn

To access the Rancher server UI and create a login, open a browser and go to the IP address where it was installed.

For example: https:///login

To access the Rancher server UI and create a login, open a browser and go to the IP address where it was installed distributed sql longhorn how to

Note: If you run into any issues attempting to access the Rancher UI, try to load the page using Chrome Incognito mode or disable the browser cache, more info here.

Follow the prompts to create a new account.

Follow the prompts to create a new account Rancher server UI

Once you’ve created the account, go to https:///g/clusters and click on Add Cluster to create a GKE cluster.

click on Add Cluster to create a GKE cluster distributed sql yugabytedb longhorn how to

Select GKE and give the cluster a name.

We now have to add the private key from the GCP service account that we created earlier. This can be found under IAM & admin > Service Accounts > Create Key.

add the private key from the GCP service account distributed sql yugabytedb longhorn tutorial

This will generate a JSON file that contains the details of the private key.

generate a JSON file that contains the details of the private key distributed sql longhorn tutorial

Copy the contents of the JSON file into the Service Account section in the Rancher UI and click Next.

Copy the contents of the JSON file into the Service Account section in the Rancher UI distributed sql longhorn

For the purposes of this tutorial I have selected an n1-standard-4 machine type, turned on Node Pool Autoscaling, and set the max node count to 24. Click Create.

select an n1-standard-4 machine type, turned on Node Pool Autoscaling, and set the max node count to 24

Verify that the cluster has been created by ensuring that the status of the cluster is set to Active. Be patient this will take several minutes.

Verify that the cluster has been created by ensuring that the status of the cluster is set to Active

The cluster should also be accessible from the GCP project by going to Kubernetes Engine > Clusters.

cluster should also be accessible from the GCP project distributed sql longhorn tutorial

Installing Longhorn on GKE

With Rancher now installed, we can use its UI to install and set up Longhorn on the GKE Cluster.

Click on the cluster, in case longhorn-demo, and then select System.

Click on the cluster, in case longhorn-demo, and then select System

Next click on Apps > Launch, search for Longhorn and click on the tile.

click on Apps then Launch, search for Longhorn and click on tile distributed sql yugabytedb longhorn tutorial

Stick with the defaults and click on Launch. Once installed, you can access the Longhorn UI by clicking on the /index.html link.

access the Longhorn UI by clicking on the index dot html link distributed sql longhorn blog

Verify that Longhorn is installed and that the GKE cluster nodes are visible.

Verify Longhorn is installed and the GKE cluster nodes are visible distributed sql yugabytedb longhorn

Installing YugabyteDB on the GKE Cluster using Helm

The next step is to install YugabyteDB on the GKE cluster. This can be done by completing the steps here, which are summarized below:

Verify and upgrade Helm

First, check to see if Helm is installed by using the Helm version command:

$ helm version
Enter fullscreen mode Exit fullscreen mode

From the output below, you can see that we are running Helm 3.

version.BuildInfo{Version:"v3.2.1", GitCommit:"fe51cd1e31e6a202cba7dead9552a6d418ded79a", GitTreeState:"clean", GoVersion:"g
o1.13.10"}
Enter fullscreen mode Exit fullscreen mode

Add the yugabytedb chart

$ helm repo add yugabytedb https://charts.yugabyte.com
Enter fullscreen mode Exit fullscreen mode

Fetch repository updates

$ helm repo update
Enter fullscreen mode Exit fullscreen mode

Validate the chart version

$ helm search repo yugabytedb/yugabyte
Enter fullscreen mode Exit fullscreen mode

To output you should be able to verify that you will be installing the latest version of YugabyteDB. In this case, version 2.1.8.

NAME                  CHART VERSION     APP VERSION
yugabytedb/yugabyte   2.1.8             2.1.8.2-b1
Enter fullscreen mode Exit fullscreen mode

Create the yb-demo namespace

$ kubectl create namespace yb-demo
Enter fullscreen mode Exit fullscreen mode

Install YugabyteDB

Next, let’s install YugabyteDB by passing options for low resource environments and specifying the longhornstorage class. This will take awhile, so again, please be patient! You can find detailed Helm instructions in the Docs.

$ helm install yb-demo yugabytedb/yugabyte --set resource.master.requests.cpu=0.5,resource.master.requests.memory=0.5Gi,resource.tserver.requests.cpu=0.5,resource.tserver.requests.memory=0.5Gi,enableLoadBalancer=True,storage.master.storageClass=longhorn,storage.tserver.storageClass=longhorn --namespace yb-demo --wait
Enter fullscreen mode Exit fullscreen mode

Check the status of the YugabyteDB cluster

To check the status of the YugabyteDB cluster, execute the command below:

$ helm status yb-demo -n yb-demo
Enter fullscreen mode Exit fullscreen mode

You can also verify that all the components are installed and communicating by visiting GKE’s Services & Ingress and Workloads pages.

verify that all the components are installed and communicating by visiting GKEs Services & Ingress and Workloads pages

You can view the YugabyteDB install in the administrative UI by visiting the endpoint for the yb-master-ui service on port 7000.

YugabyteDB Admin UI

You can also log into the PostgreSQL compatible shell by executing:

$ kubectl exec -n yb-demo -it yb-tserver-0 -- ysqlsh -h yb-tserver-0.yb-tservers.yb-demo

ysqlsh (11.2-YB-2.1.8.2-b0)
Type "help" for help.
yugabyte=#
Enter fullscreen mode Exit fullscreen mode

You are now ready to start creating database objects and manipulating data!

Managing YugabyteDB Volumes Using Longhorn

Next, reload the Longhorn dashboard page to verify that the YugabyteDB volumes are correctly set up. The number of volumes should now be visible.

number of volumes should now be visible yugabytedb in longhorn dashboard

To manage the volumes, click on Volume. The individual volumes should be visible.

manage the volumes, individual volumes are visible yugabytedb longhorn blog post how to

The volumes can now be managed by selecting them and choosing the required operations.

volumes can now be managed by selecting them and choosing the required operations

That’s it! You now have a three node YugabyteDB cluster running on GKE with Longhorn as its distributed block store.

Top comments (0)