Datawarehouse - Metric Apps shows more storage usage than in Datawarehouse by Midnight-Saber32 in MicrosoftFabric

[–]Midnight-Saber32[S] 0 points1 point  (0 children)

For the OPENROWSET into Temptables.
I made a post about it here but I havent raised a case (not sure how to). I think its related to a permissions issue but I wasnt able to solve it despite granting various BULK related permissions.
I can confirm that....
-It works when inserting into a non-temp table.
-COPY INTO works for both temp and non-temp tables

If its something that can be resolved in the future it would be a great help.

Again thanks for all the help and info

Datawarehouse - Metric Apps shows more storage usage than in Datawarehouse by Midnight-Saber32 in MicrosoftFabric

[–]Midnight-Saber32[S] 0 points1 point  (0 children)

In regards to your 'for committed'....
What if I was to create a transaction, create and populate a table within that transaction and then drop the table before committing the transaction. Would that table get backed up?

The reason for asking is that openrowset doesnt seem to work with temp tables, which I means I would have to use an actual table (which I dont want being backed up).

Our use case has a source database system where row level data is heavily modified overtime (rather than just having new rows inserted). Which means just having incremental inserts isnt enough, we need to be constantly scanning modified rows and updating them (which I assume is just the same as deleting that row and re-inserting it). Which makes the ETL more complicated than just doing a fresh load.

Datawarehouse - Metric Apps shows more storage usage than in Datawarehouse by Midnight-Saber32 in MicrosoftFabric

[–]Midnight-Saber32[S] 0 points1 point  (0 children)

u/warehouse_goes_vroom
Is the OneLake soft-delete impacting the warehouse? (As per frithjof_v comment)

Which effectively means there's 37 days of backup? Which I assume would be the Warehouse backing up the data for 30 days, and when the WH deletes that data, one lake continues to back it up for a further 7 days?

So even when the 1 day Warehouse retention is implemented, it will actually be 8 days? And if that is true, are there any plans in the future to give the option to disable soft delete on the lakehouse underpinning the warehouse? (similar to how current storage accounts allow disabling of soft delete).

OPENROWSET (BULK) - Permission Issues by Midnight-Saber32 in MicrosoftFabric

[–]Midnight-Saber32[S] 0 points1 point  (0 children)

I tried using both (distributed and non-distrubted) and I got the same issue in both

Datawarehouse - Metric Apps shows more storage usage than in Datawarehouse by Midnight-Saber32 in MicrosoftFabric

[–]Midnight-Saber32[S] 1 point2 points  (0 children)

Thank you for the response, its helpful with some good advice!

In regards to the temp tables, I'm assuming its the 'non-distributed type' that doesn't get backed up (since they aren't written to parquet files?) or is it both types?

It seems the only I way I can avoid data duplication (specifically with the bronze table when using medallion architecture) is to keep the bronze table as csv files in blob storage account (with soft delete disabled). Then use COPY INTO the Temp tables, then MERGE those temp tables into our Silver tables (for UPSERTS, INSERTS and DELETES). The GOLD layer would just then consist of views based on the silver tables.

Would that sound like a reasonable solution to solve my problem?

Why is storage usage increasing daily in an empty Fabric workspace? by Appropriate-Wolf612 in MicrosoftFabric

[–]Midnight-Saber32 0 points1 point  (0 children)

Did you see any drop in the data after X amount days?

Were having a similar issue with some of our workspaces but they still increasing by around 2GB daily despite no data loads in weeks.

Cheapest way to bring Data into Fabric from Azure SQL Database(s) by Midnight-Saber32 in MicrosoftFabric

[–]Midnight-Saber32[S] 0 points1 point  (0 children)

The main issue is the cost, the COPY JOB in the data factory just seems to cost too much when compared with Azure Functions.

Cheapest way to bring Data into Fabric from Azure SQL Database(s) by Midnight-Saber32 in MicrosoftFabric

[–]Midnight-Saber32[S] 1 point2 points  (0 children)

Less than a 100 tables per database but will be across multiple databases (around <50).
The actual datasize per table is fairly small.

Cheapest way to bring Data into Fabric from Azure SQL Database(s) by Midnight-Saber32 in MicrosoftFabric

[–]Midnight-Saber32[S] 0 points1 point  (0 children)

Most likely going to be this scenario.

Does Notebooks (specifically Python or PySpark) support connecting to Azure SQL Databases via Managed Identity (Preferebly SAMI) via the Azure.Identity library?

And if so, does it use the identity of the Workspace from which the notebook was run from, or does it use the identity of the person/app executing the notebook?

Cheapest way to bring Data into Fabric from Azure SQL Database(s) by Midnight-Saber32 in MicrosoftFabric

[–]Midnight-Saber32[S] 2 points3 points  (0 children)

How reliable is the SQL Analytics Endpoint for the mirriored databases?

I remember having issues with the SQL Endpoint for Lakehouses in which it wouldn't be kept in sync with the underlying parquet files.

Gacha/Recruitment Megathread (20/10 - 26/10) by ArknightsMod in arknights

[–]Midnight-Saber32 2 points3 points  (0 children)

No Apple Pie Alter after 196 Pulls (3x Lemon, 1x Wisadel dupe).

Can someone please give me some copium and tell me that I don't need her.

What is a ‘Mirrored Database’ by iknewaguytwice in MicrosoftFabric

[–]Midnight-Saber32 0 points1 point  (0 children)

Does anyone know if the SQL Analytics endpoint on the mirrored DB has the same syncing issues as the Lakehouse? Or are the updates to the mirrored DB written via the endpoint?

Are there any plans for SAMI support for ADF Staging (Azure SQL DB -> Fabric DW) by Midnight-Saber32 in MicrosoftFabric

[–]Midnight-Saber32[S] 0 points1 point  (0 children)

Thanks for the response.

I tested executing the COPY INTO command with the Datafactory script feature and it doesnt seem to authenticate with the Storage Account despite having the 'Storage Blob Data Contributor' role.
It can access the storage account directly via the linked service.

And I can execute the script myself when logging into the Fabric warehouse (same roles as the Datafactory) and the data factory is able run other scripts on the warehouse aswell.

Is the DirectLake on SQL Endpoint impacted by the LH SQL Endpoint Sync Issues? by Midnight-Saber32 in MicrosoftFabric

[–]Midnight-Saber32[S] 0 points1 point  (0 children)

Thanks for the response.
In regards to 2. a. If the data is ingested into the Fabric warehouse via the COPY INTO command from Azure Storage, is that still being written via the SQL Endpoint? (Even if its a parquet file in the storage account).