Execute SQL queries on your Databricks SQL warehouse
List all Databricks jobs in your workspace
Get the status of a specific Databricks job by ID
Get detailed information about a specific Databricks job
The Databricks Connector provides a seamless interface between language models and your Databricks environment. It enables natural language interaction with your data warehouses and job management systems, allowing you to execute SQL queries, list jobs, and monitor job statuses directly through conversational prompts. This connector bridges the gap between AI assistants and your Databricks infrastructure, making data analysis and workflow management more accessible through natural language. It's particularly useful for data scientists, analysts, and engineers who want to quickly access Databricks resources without switching contexts.
The Databricks Connector allows language models to interact directly with your Databricks environment, enabling SQL query execution and job management through natural language requests.
Before setting up the connector, ensure you have:
git clone https://github.com/JordiNeil/mcp-databricks-server.git
cd mcp-databricks-server
python -m venv .venv
source .venv/bin/activate # On Windows: .venv\Scripts\activate
pip install -r requirements.txt
.env
file in the root directory with your Databricks credentials:DATABRICKS_HOST=your-databricks-instance.cloud.databricks.com
DATABRICKS_TOKEN=your-personal-access-token
DATABRICKS_HTTP_PATH=/sql/1.0/warehouses/your-warehouse-id
To set up the connector, you'll need to gather the following credentials:
Host: Your Databricks instance URL without the https://
prefix (e.g., your-instance.cloud.databricks.com
)
Personal Access Token:
HTTP Path for SQL Warehouse:
Before running the server, you can verify your connection settings:
python test_connection.py
This will attempt to connect to your Databricks environment and report any issues.
Start the MCP server with:
python main.py
The server will start and listen for requests from language models.
Once the server is running, language models can interact with your Databricks environment through natural language. Examples include:
If you encounter connection issues:
https://
prefix.env
file is properly formatted and in the correct location.env
file to version control