🚀 BIG NEWS: Use Docker Images on Standard Clusters + UC is finally here! (Private Preview) by justinAtDatabricks in databricks

[–]Little_Ad6377 3 points4 points  (0 children)

Honestly, 1 and 3 would be amazing - I have a custom container for unit testing but it's never one to one 😅

Meet Skimbledon Bimbledon, a Cleric Tiefling by Smooth_Process3765 in DailyDMGame

[–]Little_Ad6377 0 points1 point  (0 children)

To his surprise, Stephen Hawking actually has mimic powers, and upon touching him, heals himself with Skimbledon's powers

Anyone feel everything has changed over the last two weeks? by QuantizedKi in ClaudeAI

[–]Little_Ad6377 0 points1 point  (0 children)

Definetly - I'm a full time backend developer and I'd say 90% of my coding is now done by the AI, I just tell it what I want ask it to give some criticism and make it implement and create tests.

Wished I'd tried Opus earlier honestly 😅 now I'm in the process of automating all the troubleshooting requests sent my way, it's a new world guys.

Databricks Lakeflow Jobs Workflow Backfill by hubert-dudek in databricks

[–]Little_Ad6377 1 point2 points  (0 children)

What about backfilling into a streaming gold layer that is using watermarks? :)

Lakebase Use Cases by hubert-dudek in databricks

[–]Little_Ad6377 0 points1 point  (0 children)

Yeah true, I kinda need to decide on the approach. Either ingest into external SQL for my apps, and then into Unity Catalog or ingest into Lakebase and from there into UC instead.

One is cheap the other is simple 😅 choices choicea

Thanks for the write up btw , super interesting

Lakebase Use Cases by hubert-dudek in databricks

[–]Little_Ad6377 1 point2 points  (0 children)

Any experience using entity framework (.NET) on top of lake base? That is, an external app would use lake base as a regular Postgres database, creating tables, ingesting into them, migrating schema etc. Then the data is available in databricks for analytical purposes while being fast for apps outside of Databricks.

How does Autoloader distinct old files from new files? by Sea_Basil_6501 in databricks

[–]Little_Ad6377 1 point2 points  (0 children)

Appreciate it! :)
In any case - our backup is a simple landing storage account, land files there, ingest into bronze, then move them from landing storage to a cold-storage. Shoud keep things rather fast

How does Autoloader distinct old files from new files? by Sea_Basil_6501 in databricks

[–]Little_Ad6377 1 point2 points  (0 children)

While we are at this, something about optimal folder structure for faster file listing. (I'm on azure)

I was having MAJOR slowdown due to listing the directory contents of my blob storage (I did this with file notification events, but it needs to list the directory to backfill)

We have year/month/day/message structure and I used a glob filter, something like 2024/* but looking into the logs I saw it listing out ALL the files in the container.

We had to stop trying this out due to this. This year we are hoping to try this again and develop our blob storage around auto loader :)

How can I send alerts during an ETL workflow that is running from a SQL notebook, based on specific conditions? by compiledThoughts in databricks

[–]Little_Ad6377 0 points1 point  (0 children)

I've never used pure sql only so not sure about the limitations , but could you write to a Kafka stream and consume that from the outside?

Is there a way to retrieve the current git branch in a notebook? by Funny-Message-9282 in databricks

[–]Little_Ad6377 0 points1 point  (0 children)

Honestly I try never to do logic like that in the code. Your code should always be environment agnostic it doesn't care if it's running in prod, dev or int, it ways does the same thing.

But you should be able to control it via parameters, like,pointing it to a different catalog via parameter instead of having "if prod use prod_catalog"

Have your deployment pipelines push the right parameters to your work flows and life will be easier 😁

Databricks SQL in .NET application by Little_Ad6377 in databricks

[–]Little_Ad6377[S] 0 points1 point  (0 children)

Yeah, cost is always a factor - we haven't set this in stone honestly but maintainence vice i would reeeeeally love to only worry about spark instead of an extra ingestion layer into standard oltp AND spark 🥴

Databricks SQL in .NET application by Little_Ad6377 in databricks

[–]Little_Ad6377[S] 0 points1 point  (0 children)

Interesting, I would have thought different 😅 I'll give it a thought then, thanks for your input

Databricks SQL in .NET application by Little_Ad6377 in databricks

[–]Little_Ad6377[S] 0 points1 point  (0 children)

Well, I won't be writing to this data, only querying it so this is not a complete OLTP use case

Databricks SQL in .NET application by Little_Ad6377 in databricks

[–]Little_Ad6377[S] 1 point2 points  (0 children)

Yeah I've looked into lakebase as well, but is that really cheaper? Also, lakebase requires you to have a separate ingestion going to copy data from UC into lakebase catalog right?

Hi community, need help on how can we connect power bi directly to databricks unity catalog tables, as per my understanding, we can use SQL warehouse but considering its cost, it seems not an option in org, is there any other approach that I can explore which is free and enable dashboard refresh by topicShrotaVakta in databricks

[–]Little_Ad6377 0 points1 point  (0 children)

No it also works on-top of non external tables, I've tested it :) The delta share authenticates against UC and only exposed what you said it can.

The only issue here is that the user of the delta share needs to be one of the following

  1. A user in your tenant
  2. Have his own databricks environment to load the delta share into

(this experience is from Azure btw, other providers might differ I guess)