DEV Community

Kenichiro Nakamura
Kenichiro Nakamura

Posted on

Ingest data into Databricks from Azure Data Explorer with KeyVault

In previous article, I explained how to read data from ADX in Azure Databricks.

Some of you may worry about pasting secret directly in notebook, which is valid concern. I explain how we can use Azure KeyVault to securely store and use secret instead.


Create Azure KeyVault resource.

Alt Text

Store Secret in KeyVault

Move secrets into KeyVault. I store client id into keyvault too.

1. Open KeyVault resource from Azure Portal and select Secrets.

2. Add "clientid" and "clientsecret" secrets and paste values from notebook which you created in the previous article.

Alt Text

Create Secret Store in Databricks

Next, setup secret store for databricks.

1. Go to https://your_databricks_cluster_address#secrets/createScope. You can find your URI from Databricks overview pane in Azure Portal.

2. Enter following information.

  • Name: This becomes scope name
  • Manage Principal: All Users
  • DNS Name: You can get "Vault URI" from Azure KeyVault resource in Azure portal
  • Resource ID: You can get this from URI in browser when you access KeyVault resource. Copy the address after /subscriptions

Alt Text

If you want to change Manage Principal to only creator, then you need to use Premium Tier.

3. Click "Create" to complete.

Once secret store is created, use Databricks CLI to confirm it.

1. Install Python 3.6+ or 2.7.9+ in your computer.

2. Install CLI via pip install.

pip install databricks-cli

3. Run configuration command.

databricks configure --token

4. It asks you the URL of Azure Databricks. You can find it in Azure Portal.

Alt Text

5. Then it asks you access token. You can get it from Azure Databricks portal. Select user icon on right top corner.

Alt Text

6. Click "Generate New Token" to obtain token.

Alt Text

7. List the secret store with scope. You should see two secrets.

databricks secrets list --scope key-vault-secrets

Alt Text

Use secrets in notebook

Finally replace current notebook to use the secret.

1. Go to notebook and open the one you created.

2. Replace the code where you define client_id and client_secret. Use dbutils to get value from secret store.

client_id = dbutils.secrets.get(scope = "key-vault-secrets", key = "clientid")
client_secret = dbutils.secrets.get(scope = "key-vault-secrets", key = "clientsecret")

3. Run the notebook to confirm it works as expected.


Azure KeyVault is great place to store secrets and you don't have to worry about hardcoding values in your code. It also works as central repository for keys, so that you only need to update the keyvault whenever you regenerated secret!


Setup Databricks CLI
Secret scopes
dbutils secret Utilities

Top comments (0)