Skip to main content

Databricks Integration

Connect Fleak to Databricks using OAuth M2M credentials to access Unity Catalog tables in your workflows.

Databricks is a unified analytics platform that provides collaborative data engineering and data science capabilities. The Databricks integration in Fleak allows users to connect to their Databricks workspace and write processed data from workflows into Unity Catalog tables using the Databricks Sink node.

Setting Up Databricks Integration

  1. Click on your username in the top left corner of the Fleak dashboard
  2. Select Settings from the dropdown menu
  3. Click on Integrations in the settings sidebar

Integrations Settings Menu

Add Databricks Integration

On the Integrations page, click Add New Integration to see available integration types. Select Add a Databricks account to create a new Databricks integration.

Add New Integration

note

Snowflake integration is coming soon and is not yet available for configuration.

Configure Connection Details

Fill in the following connection details:

  • Connection Name: A descriptive name for your connection (e.g., "My Databricks Connection")
  • Host URL: Your Databricks workspace URL (e.g., https://xxx.cloud.databricks.com)
  • Client ID: The OAuth Client ID from your Databricks Service Principal
  • Client Secret: The OAuth Client Secret from your Databricks Service Principal

These are OAuth M2M (Machine-to-Machine) credentials from your Databricks Service Principal, ensuring secure, auto-refreshing authentication for 24/7 pipeline stability.

Create Databricks Integration

Click Create to save the integration.

info

Connection scope (catalog, schema, volume, warehouse) is configured when creating data assets, not during integration setup.

Using Databricks in Workflows

Once connected, Fleak will be able to see all the tables in Unity Catalog. You can use the Databricks Sink node in your workflows to write data into those tables.

note

The service principal credentials must have appropriate permissions to read catalog metadata and write to the target tables.

For more information about configuring the Databricks Sink Node, please visit the Databricks Sink Node reference page.