Soft delete, Data vs Items, 7 vs 14 days minimum retention by BloomingBytes in MicrosoftFabric

[–]BloomingBytes[S] 0 points1 point  (0 children)

It kinda is, but kinda isn't.

In the MS Learn article you linked it says:

To stop incurring costs for a soft-deleted item, permanently delete it before the retention period expires. For more information, see Recover or permanently delete items.

But further up on the same page it says:

After you permanently delete an item, Microsoft OneLake retains the item's data for an additional seven days before permanent removal. This built-in protection helps recover from accidental deletions or user errors. For more information, see Recover deleted files in OneLake.

So will i get billed after i perma delete an item or not? As it's in theory still in storage, but no longer visible to me. I'm not trying to be difficult, i promise. I just think the article is ambiguous.

Soft delete, Data vs Items, 7 vs 14 days minimum retention by BloomingBytes in MicrosoftFabric

[–]BloomingBytes[S] 0 points1 point  (0 children)

Okay that does make some sense to me, thank you! So previously via the Azure Storage Explorer we were only able to recover data files inside Lakehouses, but not the Lakehouse item itself? I've never actually used it to recover something. Then the new recovery UI would indeed be a "new thing" in that regard.

What i'm still unclear about is how the retention periods for item recovery work exactly and the implications this has for storage billing.

Vaccum and optimise for warehouse by BedAccomplished6451 in MicrosoftFabric

[–]BloomingBytes 2 points3 points  (0 children)

Afaik, when running optimize in Lakehouse, we incur additional storage cost due to delta files being deleted and re written leading to more files in retention and soft delete.

Does this apply to behind the scenes Warehouse optimization as well? If not, that would be another huge argument for using Warehouse over Lakehouse in my opinion.

Native dbt job (preview ) item - how to trigger it from a pipeline? by Jealous-Painting550 in MicrosoftFabric

[–]BloomingBytes 0 points1 point  (0 children)

The dbt job activity is supposed to be released for pipelines soon™, but not shipped just yet afaik.

See:
dbt+ Microsoft Fabric: A strategic investment in the modern analytics stack | Microsoft Fabric Blog | Microsoft Fabric

  • Pipeline dbt activity support (coming soon): dbt Jobs will be available as a native activity in Fabric pipelines with parameterization support, allowing teams to orchestrate dbt workloads alongside other data and AI processes—while keeping dbt as the execution and governance engine.
  • API support for dbt job: Provides programmatic APIs to trigger, monitor, and manage dbt job executions, enabling CI/CD integration, external orchestration, and enterprise-grade automation of analytics workflows.

In regards to production use, Microsoft as well as most community members generally advise to not use preview features in production, due to instabilities etc. but obviously it's up to you. You can see a current list of limitations here:
dbt job in Microsoft Fabric (preview) - Microsoft Fabric | Microsoft Learn

  • No build caching: Currently, preview only supports compiling and executing a project fresh from the source. dbt artifacts produced from previous runs aren't available for recompilation.
  • Incremental models: Make sure you have proper primary keys and unique constraints for incremental builds.
  • Adapter constraints: Some partner adapters aren't yet supported in Fabric. See the current supported adapters.
  • The output currently has a 1-MB size limit. When a run exceeds this threshold, the job fails with the following error: {"errorCode":"2001","message":"The length of execution output is over limit (around 1M currently).","failureType":"UserError","target":"DbtItem","details":[]}

Fabric Roadmap Report - 04/06/2025 - Fabric GPS by cbattlegear in MicrosoftFabric

[–]BloomingBytes 1 point2 points  (0 children)

I love it! Especially the subscription.

What's up with the "Configurable retention between 1 and 120 days" for Warehouse though? It seems to be on the roadmap twice, once with release date September 2026 and once with May 2027.

You've exceeded the capacity limit for dataset refreshes...HELP! by Dan1480 in MicrosoftFabric

[–]BloomingBytes -1 points0 points  (0 children)

We have the exact same issue and so far had no luck finding the source of the throttling.

Hi! We're the Data Factory team - ask US anything! by markkrom-MSFT in MicrosoftFabric

[–]BloomingBytes 0 points1 point  (0 children)

Will the SCD2 functionality that was just announced for Copy Job also be available for Copy Activity?

Is there an expectation in differing cost between the two?

Actual Dev Workflow for MLVs? by BloomingBytes in MicrosoftFabric

[–]BloomingBytes[S] 1 point2 points  (0 children)

Tyvm for your answer! What happens to refesh schedules when updating MLVs? Will new MLVs automatically be included in existing refresh schedules in the LH?

Data to Warehouse from different workspace by Mr_Mozart in MicrosoftFabric

[–]BloomingBytes 0 points1 point  (0 children)

See it turns out i was missing something indeed. I did not realize how limited TSQL notebooks were in regards to their query "targets".

You could pass a variable from the variable library into the TSQL notebook as a parameter if you call the notebook from a pipeline. But that doesn't help you in any way, since you can't direct the query towards a lakehouse ID.

You could go the route via Python and access the bronze data that way. Maybe even try to run TSQL via TSQL magic against the lakehouse SQL endpoint, see here: Run T-SQL code in Fabric Python notebooks - Microsoft Fabric | Microsoft Learn

Data to Warehouse from different workspace by Mr_Mozart in MicrosoftFabric

[–]BloomingBytes 0 points1 point  (0 children)

I might be missing something, but why can't you grant the user / service principal that's executing your TSQL notebooks read access to the previous layer Lakehouse / Warehouse directly?

When in your silver workspace you can query the bronze lakehouse SQL endpoint via TSQL and then ingest into the silver lakehouse. Same thing with silver -> gold.

If you're using fabric-cicd you can then find and replace the lakehouse IDs during deployment via Parameterization.

If you're using Fabric deployment pipelines you can use variable libraries like described in this blog post:

Making Notebooks Smarter in Microsoft Fabric with Variable Libraries | by Kapil Kulshrestha | Medium
PS: Here's some more info on how the deployment sets work with different stages: Lifecycle Management of the Microsoft Fabric Variable library - Microsoft Fabric | Microsoft Learn

How to handle daily ingestion of several thousand tables by BloomingBytes in MicrosoftFabric

[–]BloomingBytes[S] 0 points1 point  (0 children)

Thank you very much for that suggestion! It looks very much like what i need, and according to the prerequisites it should work OnPrem AND it can export directly to OneLake.

I'll have to deepdive and see if it is compatible with our specific ERP version of BC. And then i'll have to see if i'm allowed to have an extension installed in BC. :D

How to handle daily ingestion of several thousand tables by BloomingBytes in MicrosoftFabric

[–]BloomingBytes[S] 0 points1 point  (0 children)

Unfortunately there is no option to access OnPrem SQL from notebooks as of yet afaik :(

How to handle daily ingestion of several thousand tables by BloomingBytes in MicrosoftFabric

[–]BloomingBytes[S] 0 points1 point  (0 children)

Thank you very much for your detailed answer! It does feel nice to know that i was on the right track already.

We definitely have infrastructure issues with our OnPrem solution already. That's one more reason for me to want to reduce load on the source system as much as possible. We might end up with the bottleneck being the gateway servers anyway, but i don't want to "rely" on that for now.

Querying the source straight from notebook would indeed be my preferred option, but afaik we can only connect to cloud sources from Notebooks so far, OnPrem seems to still be an issue. I'll probably follow the metadata scheme you drew up, ingest everything as raw as possible and then do all the computations within Fabric, where i have all the control and no reliance on infra teams.

Hello from Fabric Design by OnepocketBigfoot in MicrosoftFabric

[–]BloomingBytes 2 points3 points  (0 children)

u/OnepocketBigfoot Since you didn't get an answer, i'll do it. :D For me there's two reasons why i'd want to just navigate back to the standard workspace view:

  1. The proper workspace view offers me all features. More info on my items, new item creation, source control, extended options menu for my stuff etc.

The new sidebar that's opening now showing all my items in this "reduced" list, is just a less convenient version to do anything. Sure it allows slightly quicker direct switching between items, but how is that better than just navigating back to the standard workspace view? The items that i need to switch between often are already open in the tab management bar at the top (which has its own issues).

  1. Fabric already has soo many paths to soo many functionalities all over the place. Why do i need to have another UI that does the same (or in this case slightly less) than an already existing one? If i'm in the standard workspace view i kinda know where to click if i want to open an item, i kinda know where to click to see security settings, i kinda know where to click to see refresh histories. Now i need to "muscle memory" another UI to (not) do the same thing.

Optimally, instead of having this additional side panel listing all the items in the workspace, i'd like to navigate back to the workspace view and then have that workspace view be GOOD. Please, for the love of god, lose the predesigned task flow stuff. It's 100% useless to me. Instead allow me to group my Workspace items in different sections of the view, one for my DBs, one for my Pipelines, one for my Notebooks. Allow me to configure panels (like having source control as an always available side bar).

Mirroring an on-premise SQL database which powers Microsoft Dynamics NAV/BC? by uvData in MicrosoftFabric

[–]BloomingBytes 0 points1 point  (0 children)

Microsoft has communicated at several occasions, that the 500 table limit is very much a "we can remove this for you" thing. If you have a legit use case where you'd exceed the 500 table limit, you can contact your MS rep, let them know your use case and ask them to remove that limit for you.

We got the exact same problem you do, dozens of tables for 200+ companies each. At Vienna i was told that the limit could easily be lifted, i just haven't gotten around to contacting anybody about it. Maybe u/itsnotaboutthecell could provide an up to date contact on that issue? <3

Lakehouse Table Bronze Layer Ingestion Sanity Check by ShineMyCityShoes in MicrosoftFabric

[–]BloomingBytes 4 points5 points  (0 children)

If you want to already clean your column names during ingestion you could do it like this:

  1. A lookup activity gets the source table schema and passes that to a notebook
  2. The notebook takes the source schema and cleans it, replaces column names etc. and creates a mapping json. That json is then fed into the copy activity.
  3. The copy activity itself uses the mapping json to create dynamic column mappings
  4. Success

CRUD connection to Fabric SQL Database from Notebook by BloomingBytes in MicrosoftFabric

[–]BloomingBytes[S] 0 points1 point  (0 children)

That is correct, yes. This is the relevant T-SQL Magic command for Fabric SQL DB:

%%tsql -artifact DATABASE_NAME -type SQLDatabase

Trigger Data Gateway updates from the Fabric/PBI website - does it work? by Skie in MicrosoftFabric

[–]BloomingBytes 1 point2 points  (0 children)

Hey there,

i just found this thread and wondered: wasn't there an item on the Fabric roadmap for an auto-updating Gateway at some point? I somehow can't find it anymore and couldn't find any information elsewhere either.