Databricks Plugin
Note: Ask DataHub Plugins is currently in Private Beta. To enable this feature, please reach out to your DataHub customer support representative.
The Databricks Plugin connects Ask DataHub to your Databricks workspace via the Databricks MCP server, enabling conversational analytics directly from the chat. Like the Snowflake plugin, DataHub acts as the semantic layer — finding the right data — while Databricks serves as the data layer — executing queries and returning results.
Why Connect Databricks?
With the Databricks plugin enabled, Ask DataHub can:
- Conversational analytics — ask questions in plain language and get answers backed by real data. DataHub finds the right tables using metadata, then Databricks executes the SQL query.
- Data exploration — understand dataset structure, preview sample data, and inspect column values across your Unity Catalog.
- Data debugging — investigate data issues detected by DataHub assertions by querying the actual data at the source.
Because the plugin uses OAuth, each user authenticates with their own Databricks account. Users only see data they are authorized to access.
Example prompts:
- "Query the top 10 customers by lifetime value"
- "How many rows were added to the events table today?"
- "Show me 5 sample rows from the orders table"
- "Check for null values in the email column of the users table"
Prerequisites
- A Databricks workspace with Managed MCP Servers enabled
- Admin access to create OAuth app connections in Databricks (Settings > App connections)
- DataHub Cloud with Ask DataHub Plugins enabled
- Platform admin access in DataHub to configure the plugin
Admin Setup
The Databricks plugin uses User OAuth authentication — each user authenticates with their own Databricks account. Setup involves creating an OAuth app connection in Databricks, then configuring the plugin in DataHub.
Step 1: Create an OAuth App Connection in Databricks
- In Databricks, navigate to Settings > App connections
- Click Add connection and fill in the details:
| Field | Value |
|---|---|
| Application Name | DataHub (or any name you'll recognize) |
| Redirect URLs | https://<your-datahub-url>/integrations/oauth/callback |
| Access scopes | Check SQL |
| Generate a client secret | Checked |
The Redirect URL must match the OAuth Callback URL shown in the DataHub plugin creation form exactly (e.g. https://your-org.acryl.io/integrations/oauth/callback). You can copy this URL from the DataHub form when creating the plugin in Step 3.
- Click Add

Step 2: Collect Credentials
After creating the connection, a dialog will show your Client ID and Client Secret. Copy both immediately — the client secret won't be shown again.

Step 3: Create Plugin in DataHub
- Navigate to Settings > AI > Plugins in DataHub
- Click + Create and select Custom MCP
- Fill in the plugin details:
| Field | Value |
|---|---|
| Name | Databricks |
| Description | A description for the plugin |
| MCP Server URL | https://<your-workspace-hostname>/api/2.0/mcp/sql |
| Authentication Type | User OAuth (Each user authenticates) |
- DataHub will automatically discover the OAuth configuration from Databricks, including the authorization URL, token URL, and available scopes. Provide the following:
| Field | Value |
|---|---|
| Client ID | From Step 2 |
| Client Secret | From Step 2 |
| Default Scopes | sql (minimum required) |

The sql scope is the minimum required for Ask DataHub to execute queries. If you selected All APIs when creating the Databricks app connection, additional scopes will be available — select only what you need.
- Optionally add Instructions for the AI Assistant, ensure Enable for Ask DataHub is on, and click Create
User Setup
Navigate to Settings > My AI Settings, find the Databricks plugin, and click Connect. You'll be redirected to Databricks to authenticate, then back to DataHub. See the overview for more details.
Troubleshooting
OAuth Fails
- Verify the Redirect URL in Databricks matches the OAuth Callback URL shown in DataHub exactly
- Ensure the Client ID and Client Secret are entered correctly
- If your workspace has IP access restrictions, ensure DataHub's outbound IPs are allowlisted
Query Failures
- Verify the user has access to the tables they're querying in Unity Catalog
- Ensure the
sqlscope was selected when creating the app connection in Databricks - Check that a SQL warehouse is available and running in the workspace