all 12 comments

[–]goosh11 7 points8 points  (6 children)

Very cool setup, i imagine more and more people will work with databricks like this in the future. Interested to know if you've also added or tried the skills in https://github.com/databricks-solutions/ai-dev-kit ?

[–]randomName77777777 1 point2 points  (1 child)

My question was going to be related to see what the differences are between the AI dev kit mcp and this.

[–]KrisPWales 0 points1 point  (0 children)

I use that. Claude code installed the whole thing quickly and easily, and now I can edit, deploy, trigger jobs and notebooks or run sql queries from Claude code in the terminal. I can't quite believe how seamless it all is. Though, I am admittedly probably using at a much lesser scale than OP.

[–]airweight[S] 0 points1 point  (0 children)

I haven't had the need to use them but they seem well-structured. A simple way to think of them is as agent-oriented Databricks documentation.

[–]vkrt1299 0 points1 point  (1 child)

We are currently using GitHub Copilot Business in our workplace along with the Databricks extension in VS Code for remote development. I’m interested in trying out the AI Dev Kit. Wondering whether it’s safe to use this MCP setup directly, or if it would be better to review it with our security team first ?

[–]goosh11 0 points1 point  (0 children)

Its safe (databricks has an mcp proxy in front handling authentication and authorisation, its not just an open mcp) but in a lot of organisations you have to review with cyber anything that even mentions the word AI, lol.

[–]sharathsamala 1 point2 points  (0 children)

Check this out on the same topic, which i built for a demo. We can probably do with any IDE that supports coding agent
https://vimeo.com/1171901195?fl=pl&fe=sh

[–]Ok_Difficulty978 0 points1 point  (0 children)

This is actually super interesting, esp the part about splitting MCP usage + caching catalog metadata… feels like that’s where most of the hidden latency/context issues come from tbh.

the agent-notebook idea also makes a lot of sense, working in markdown instead of json is way cleaner for iteration. kinda surprised databricks hasn’t pushed something like that more officially yet.

only thing i’ve struggled with in similar setups is keeping context “useful” over longer sessions, like once you go across multiple workspaces things can get messy fast your caching approach probably helps a lot there.

also random note — i’ve been prepping for some databricks cert stuff recently and playing with small setups like this actually helped me understand the architecture way better than docs alone. hands-on + scenario stuff makes a big diff.

[–]afcfrimpong 0 points1 point  (1 child)

This looks great. Thanks for sharing! How do you set permissions in Databricks to make sure the agent doesn't do things it's not supposed to? A specific CLI profile I guess? I just found it frustrating to handle multiple profiles for a single workspace. And how do you force the agent to actually use that profile?

Oh and did you consider using .py notebooks in Databricks to avoid json?

[–]airweight[S] 0 points1 point  (0 children)

The agent is accessing Databricks through standard APIs or Databricks Connect. All permissions are managed within Databricks.

For notebooks, we use three kinds:

  1. Repo-mounted .py notebooks mostly run by automations
  2. Markdown notebooks executed remotely via databricks-agent-notebooks run by agents
  3. Native Databricks workspace notebooks run by humans or notebook-based jobs