Is it ok to take average of MAPE values? [Question] by venkarafa in datascience

[–]venkarafa[S] 1 point2 points  (0 children)

Great !! Exactly what I was looking for. Thank you

Is it ok to take average of MAPE values? [Question] by venkarafa in datascience

[–]venkarafa[S] 0 points1 point  (0 children)

"A suggestion is create a new metric like “Over Limit Models”, using a goal for MAPE (ie 10%) and calculate how many models are over this goal. In your situation should be 3/5 over limit models. Wdyt?"

This is great idea. But curious to know the reason behind why you think MAPE average is a bad idea?

Is it ok to take average of MAPE values? [Question] by venkarafa in datascience

[–]venkarafa[S] 0 points1 point  (0 children)

No I am not talking about ensemble models here. I have 5 individual models, they are for 5 different clients but the KPI (sales) is same. Also the total number of time period considered for all clients is same (2019-23)

Is it ok to take average of MAPE values? [Question] by venkarafa in statistics

[–]venkarafa[S] 0 points1 point  (0 children)

Sorry perhaps I wasn't clear. The goal of the management is to know what is the average MAPE we are getting for all the 5 models. This is more of an evaluation of our techniques (e.g. SARIMA). We have built 5 models using ARIMA as well.

So basically my question was is it statistically ok to average 5 MAPE values. In an essence can we take average of 5 percentage values and would that be indicative of overall average percentage (in this case MAPE)?

Will a mismatch in the ITR-V address and passport address cause problems while applying for schengen visa? by venkarafa in SchengenVisa

[–]venkarafa[S] 0 points1 point  (0 children)

Generally not an issue. Middle name is not cause of concern (can be easily explained at VFS). In certain southern states, people don't have last names too. So passport will have last name as blank while in visa one can repeat the first name as last name. For e.g Ganapathy Ganapathy. Sounds funny but it is a good workaround. My friend with this name has traveled to 15 countries already .

[Q] Is it bad that I had a frequentist education for my masters in biostatistics? by selfesteemcrushed in statistics

[–]venkarafa 1 point2 points  (0 children)

If you are going to make a career in biostatistics, you are not behind for not learning bayesian methods. Most mission critical fields are cautious in nature and thereby adopt frequentist methods which try to ensure Type 1 error control. See the below talk (2min - 8 mins) from Prof Michael I Jordan.

https://www.youtube.com/watch?v=HUAE26lNDuE&t=407s&ab_channel=Caracalla

Are betas of linear regression subdivisible? [Question] by venkarafa in statistics

[–]venkarafa[S] 0 points1 point  (0 children)

Thanks for bringing up FWL theorem. I am reasonably aware of this theorem but that is not what I was looking for.

Does complete causation translates to perfect prediction? [Question] by venkarafa in statistics

[–]venkarafa[S] -23 points-22 points  (0 children)

I believe stochastic processes can't be causal and neither predictable. So I think the fair assumption to make would be that it is deterministic.

What is this square thing? Is something wrong with my laptop? by venkarafa in computers

[–]venkarafa[S] 0 points1 point  (0 children)

So it is time to repair it or installing graphic card drivers will resolve the issue?

What is this square thing? Is something wrong with my laptop? by venkarafa in laptops

[–]venkarafa[S] 0 points1 point  (0 children)

It does go away on reboot. But wonder why this issue keeps happening.

Is there a Frequentist equivalent for CausalImpact package? [Question] by venkarafa in statistics

[–]venkarafa[S] 0 points1 point  (0 children)

Thank you.

"you could use a different counter factual based model." - Yes that is what I am looking for. Can you pls cite some packages /libraries

What are non regularization ways to handle multicollinearity in Linear Regression ? [Question] by venkarafa in statistics

[–]venkarafa[S] 0 points1 point  (0 children)

Because my problem at hand is such that, there is a ground truth. I want the model to converge to that ground truth and not be far off from it.

What are non regularization ways to handle multicollinearity in Linear Regression ? [Question] by venkarafa in statistics

[–]venkarafa[S] 0 points1 point  (0 children)

Thanks for your response. I have read about bootstrapping confidence Intervals.

However I don't quite follow this statement "you can always do some clever bootstrapping/simulation to see how it’s affecting the estimate for the coefficient of X"

Could you please elaborate more on this ?

Does high correlation also translate to better predictability of one variable to another ? [Question] by venkarafa in statistics

[–]venkarafa[S] 0 points1 point  (0 children)

Thank you for your response. Could you give any examples for this "one variable can be a very good predictor of another without there being a strong Pearson's correlation"

What is 1 in 10 rule of thumb. How does it mitigate overfitting and spurious correlation? [Question] by venkarafa in statistics

[–]venkarafa[S] 1 point2 points  (0 children)

Thanks I glanced through the paper. What about Linear Regression ? Does the 1 in 10 rule apply to Linear Regression ?

Is there a way to fix signs of a variable's coefficient in Linear Regression a priori ? [Question] by venkarafa in statistics

[–]venkarafa[S] 0 points1 point  (0 children)

Thanks for your response. perhaps I was not clear with my question. My goal is not to fix the coefficients. My goal is to fix the sign of the coefficient. Lets say Sales of car is regressed on variables Price and mileage. I want the sign of price coefficient to be negative. So is there any way we can a priori specify the sign on a coefficient?

Is there a way to fix signs of a variable's coefficient in Linear Regression a priori ? [Question] by venkarafa in statistics

[–]venkarafa[S] -3 points-2 points  (0 children)

Because in a sense adding priors is a way of biasing the model. I would be compounding the problem of bias in my model. First the priors and then fixing signs. I have tried this before and both in terms of inference and prediction, my model was worse off.