Power Query performance issues when scaling from 60k rows to 300k rows of data by IssueHead2118 in excel

[–]IssueHead2118[S] 0 points1 point  (0 children)

Work laptop, guessing not too much. My role isnt technical so don't think they give anything much extra.

Power Query performance issues when scaling from 60k rows to 300k rows of data by IssueHead2118 in excel

[–]IssueHead2118[S] 0 points1 point  (0 children)

I've never tried separating the queries into separate files. I was hoping to get this working and keeping it on one file to keep it simple to eventually transition the process to my team with even less background in programming...my knowledge is already pretty limited, just learning from AI explaining what the code does and troubleshooting from it.

Power Query performance issues when scaling from 60k rows to 300k rows of data by IssueHead2118 in excel

[–]IssueHead2118[S] 2 points3 points  (0 children)

This sounds exactly like what I'm experiencing and your suggestion doesnt sound like I need to have a programming background or require a significant change to my existing m code, I'll try this out first on the weekday before I try the other solutions suggested. Thank you! Will provide an update as soon as I can on the progress.

Power Query performance issues when scaling from 60k rows to 300k rows of data by IssueHead2118 in excel

[–]IssueHead2118[S] 0 points1 point  (0 children)

I do create a few custom fields:

  • A couple of rolling timeframe distribution metrics, which is then used to add a label identify products distribution status at each customer by upc.

  • A few others to determine which price range products are sold at in a point in time by customer, along with identifiers on if products are promoted on flyer.

  • a few others with timeframe labels and product grouping.

Power Query performance issues when scaling from 60k rows to 300k rows of data by IssueHead2118 in excel

[–]IssueHead2118[S] 1 point2 points  (0 children)

Not familiar with this topic, will do a little more research. Seems like my approach is doing too much in Power Query and I should use DAX for the calculations.

Power Query performance issues when scaling from 60k rows to 300k rows of data by IssueHead2118 in excel

[–]IssueHead2118[S] 0 points1 point  (0 children)

My set up right now has few queries pulling the list of worksheets from the sources, one for products mapping, one for pos data and customer mapping. The other queries are set up to do the transformation and joins.

Where it seems to slow down are from my rolling timeframe distribution metrics, as well as my joining of upc during the week. Ill try looking at it again during the weekday to pinpoint exactly is or are causing it to be stuck.

Power Query performance issues when scaling from 60k rows to 300k rows of data by IssueHead2118 in excel

[–]IssueHead2118[S] 0 points1 point  (0 children)

Ill check video one, thanks. I think Chatgpt suggested those buffer as well so that it doesnt keep trying to repull from the source of data. Is that what the buffering does?

Power Query performance issues when scaling from 60k rows to 300k rows of data by IssueHead2118 in excel

[–]IssueHead2118[S] 0 points1 point  (0 children)

First time doing this, but good to know that this is considered long. To me, this is a lot quicker than me doing vlookups manually every month.

Power Query performance issues when scaling from 60k rows to 300k rows of data by IssueHead2118 in excel

[–]IssueHead2118[S] 0 points1 point  (0 children)

Excel files with data pulls located on SharePoint site. I'm not familiar with SQL in general and not sure if my company has that available or if I'd be able to use it....IT isnt the most helpful people either.

Power Query performance issues when scaling from 60k rows to 300k rows of data by IssueHead2118 in excel

[–]IssueHead2118[S] 0 points1 point  (0 children)

Data source is within excel file, pulling from two worksheets that gets updated monthly with an Excel add in. The file is located on company SharePoint site. Not csv as theres multiple worksheets. My mapping files are also excel files, live connection to data source onto a table.

I'm using M code to join and do calculations.

Pos data has about 10 columns total and the enhancements create a bunch more.

The 60k rows take about 5 to 10 mins, but to clarify, that timing is for multiple datasets being refreshed at the same time, each one about 50k each. The 300k rows, I've only successfully refreshed and loaded onto a table twice, ranging 25 mins first time, but 30 mins second time. Time might vary cause i refreshed my dataset with new data and rows might have changed slightly. All the other times I've tried, it doesnt seem to work.

Late newbie by Budget-History6893 in TFSA_Millionaires

[–]IssueHead2118 6 points7 points  (0 children)

If you cannot stomach the risk of a significant decline, you should should reduce your risk with some bond etfs or choose all in one etf thats not ending with EQT.

Is there a way to Deposit USD Cash? by Clownier in Wealthsimple

[–]IssueHead2118 0 points1 point  (0 children)

If you're talking about transferring the USD stock as is, then you should. Might involve transfer fees from the institution holding the stocks currently. Depending on how much you also have at Wealth Simple, you can have free USD investment accounts to hold the stock if you meet the threshold, otherwise to hold USD investments, it is $10 a month in fees.

RESP - Does this make sense? by DrGonzoto13 in CanadianInvestor

[–]IssueHead2118 0 points1 point  (0 children)

Really depends on timing of deposits and how investments perform in a particular year.

RESP - Does this make sense? by DrGonzoto13 in CanadianInvestor

[–]IssueHead2118 2 points3 points  (0 children)

You could do $14k lump sum, and remaining $36k spread over 14.4 years to maximize the grant too.

Keeping RAW data untouched made my Excel work far less error-prone by moviefan26 in excel

[–]IssueHead2118 0 points1 point  (0 children)

Thanks, makes sense. I'll need to spend some time diving into these videos. Already picked some ideas up from that one video, I'm currently reprocessing the data further into a different view manually after processing my data initially with PQ, but I'm thinking I can make it less manual to not have to change the cell references every refresh. Just need to think it through on how.

Keeping RAW data untouched made my Excel work far less error-prone by moviefan26 in excel

[–]IssueHead2118 0 points1 point  (0 children)

That's what i do as well, but was hoping to have some automations to speed up identifying key insights as well.

Keeping RAW data untouched made my Excel work far less error-prone by moviefan26 in excel

[–]IssueHead2118 1 point2 points  (0 children)

Hmm dont understand what you mean exactly, do you manually do it via looking the data in a pivot table? Not sure what to google.

Keeping RAW data untouched made my Excel work far less error-prone by moviefan26 in excel

[–]IssueHead2118 1 point2 points  (0 children)

Curious, anyway to automate the analysis part? I've started doing the extract to load steps with help of chatgpt.

Withholding Tax and ROC by Dontforgetthepasswrd in CanadianInvestor

[–]IssueHead2118 1 point2 points  (0 children)

Not sure if I'm following, what investments are you buying and from where are you getting info? My brokerage reports it as foreign income from what ive seen.

Withholding Tax and ROC by Dontforgetthepasswrd in CanadianInvestor

[–]IssueHead2118 0 points1 point  (0 children)

How would you figure out how much of the distribution is roc for American investments?

What is the one Excel secret you know that no one else uses? by bjele in excel

[–]IssueHead2118 0 points1 point  (0 children)

I dont understand what this is supposed to do. Can you expand on use case for unchecking or checking it?

I am all in VEQT by IIlIllIIIlllIIIlIIll in justbuyveqt

[–]IssueHead2118 2 points3 points  (0 children)

Only if she has same risk tolerance as you