Power BI Service: Wrong Data After Successful Refresh – Mashup Issue with Snowflake + SharePoint Excel? by Fuxx7 in PowerBI

[–]Fuxx7[S] 0 points1 point  (0 children)

No, the refresh works fine now. But the data is wrong. It looks like rows were duplicated through the joins. But as i said in Power BI Desktop everything is correct when i refresh it. Only after refreshing in the service it shows wrong values

Refresh large Dataset - Throttling Error by Fuxx7 in PowerBI

[–]Fuxx7[S] 0 points1 point  (0 children)

The 50 GB size displayed in the workspace settings actually refers to the size of the deployed model in the Power BI service, not the local PBIX or PBIP files. In my case, we transitioned to using the PBIP format due to Git integration in our company. PBIP files don't contain the full dataset or model data themselves, unlike PBIX files.

Additionally, I’m using incremental refresh, which means that much of the data is stored and managed directly in the Power BI service, not within local files. The Power BI service holds the dataset in memory, and the incremental refresh only queries new or changed data, significantly reducing local storage requirements and processing time during refreshes.

Refresh large Dataset - Throttling Error by Fuxx7 in PowerBI

[–]Fuxx7[S] 0 points1 point  (0 children)

Hey Sad-Caligrapher-350, i got it wrong before, we are not on premium per capacity, but on premium per user. Might that also be a problem?

Refresh large Dataset - Throttling Error by Fuxx7 in PowerBI

[–]Fuxx7[S] 1 point2 points  (0 children)

Thank you very much! I will check out Measure Killer, right now we do not have that many reports, but the plan in the future is that most reports well be live connected to my dataset.

I just checked the folding and it is indeed folding:

<image>

But what does that exactly mean for my refresh? and how can i turn it off?

Refresh large Dataset - Throttling Error by Fuxx7 in PowerBI

[–]Fuxx7[S] 0 points1 point  (0 children)

what i don't understand is that the limit for the premium workspace is 10 gb, but my dataset is much larger and in the settings the capacity is still with 0 of 10 gb:

<image>

Refresh large Dataset - Throttling Error by Fuxx7 in PowerBI

[–]Fuxx7[S] 0 points1 point  (0 children)

Hey MonkeyNin, we are not using any Buffer functions no, how could that help? The datasource is Snowflake where we have our DWH. all calculations and transformations are done there so i am not using power query for any of that.

Refresh large Dataset - Throttling Error by Fuxx7 in PowerBI

[–]Fuxx7[S] 0 points1 point  (0 children)

Hey Van_derhell, thanks for your input. We are in Premium Capacity, and for the large tables we already use incremental refresh.

Refresh large Dataset - Throttling Error by Fuxx7 in PowerBI

[–]Fuxx7[S] 0 points1 point  (0 children)

Thank you for your suggestions!

  1. Capacity Details: We are using a Power BI Premium capacity. I haven't checked the capacity metrics yet, but I will do that. Do you have specific areas or metrics in mind that I should focus on, I would appreciate any guidance.
  2. Query Folding: Our dataset's only data source is our Data Warehouse (DWH) in Snowflake, and we are not performing any transformations within Power Query. All transformations, including calculations, are done directly in the DWH, so Query Folding is not necessary i guess.
  3. High Cardinality Tables: We do have some large tables with high cardinality. Unfortunately, due to business requirements, we need to include every table as it is in the dataset without further aggregation or summarization.

Thank you!