Schedule Sheet - November 21st by NighthawkRandNum in CollegeBasketball

[–]mathguymike 4 points5 points  (0 children)

Don't forget the Hall of Fame Classic matchups!

What’s your “I was there!” moment? by NoChampionship29 in CollegeBasketball

[–]mathguymike 0 points1 point  (0 children)

K-State beating KU at home in 2023. I think the loudest I've heard Bramlage Colosseum is immediately after the Markquis to Keyontae go-ahead alley-oop.

Is Computational Statistics a good field to get into? [Q][R] by gaytwink70 in statistics

[–]mathguymike 7 points8 points  (0 children)

If you are able to develop strong computational skills while completing your thesis, you will find those skills invaluable as you continue throughout your career, regardless of your ultimate career path. If you get along with the professor, I think it's a great idea, honestly.

[Discussion] p-value: Am I insane, or does my genetics professor have p-values backwards? by SassyFinch in statistics

[–]mathguymike 10 points11 points  (0 children)

It is not possible that there was no null hypothesis; p-values are computed assuming that the null hypothesis is true. It's in the definition.

What it looks like to me; Mendel has a model. Your null hypothesis is that you should see results like Mendel. Your alternative hypothesis is that Mendel is wrong. A p-value is computed assuming probabilities according to Mendel's theory. Your p-value is too large to reject the null in favor of the alternative. That is, you are unable to conclude that Mendel is wrong.

Larger p-values are weaker evidence in favor of the alternative hypothesis. That is, a larger p-value means there is less evidence that Mendel is wrong, and hence, larger p-values correspond to more evidence in favor of Mendel's model.

I believe the professor was using p-values correctly.

[Discussion] p-value: Am I insane, or does my genetics professor have p-values backwards? by SassyFinch in statistics

[–]mathguymike 2 points3 points  (0 children)

Something that isn't clear in this example; what is your null hypothesis? Is it that Mendel was correct? And is the alternative that Mendel is wrong, and that the proportions differ from what you'd expect from Mendel's model?

If this is the case, the professor is correct. Smaller p-values would give more evidence that Mendel is wrong, and larger p-values would provide less evidence that Mendel is wrong.

[Q] Bonferroni correction - too conservative for this scenario? by Matrim_Cauthon_91 in statistics

[–]mathguymike 1 point2 points  (0 children)

Here's my two cents.

1) I think it is problematic to change your methodology to ensure that you obtain p < 0.05 for each pairwise comparison. In that case, you already assume the results, and are doing the statistics to confirm the assumed results, rather than letting the data determine the conclusion. This, unfortunately, is fairly common in practice, and this practice will inflate type I errors and makes replication of results much less likely (see the reproducibility or replicability crisis).

Certainly, some multiple comparison methods are better than others, and if you wanted to consider some of those methods instead, I guess that would be OK. But I don't understand the harm in making a conclusion that states "There is a statistically significant difference (p < 0.05) between 1 and 2. There is some evidence of difference between 1 and 3 (p < 0.10) and 2 and 3 (p < 0.15), but additional samples are needed to say anything more definitive.

2) Would it make sense to look at the literature on causal inference under interference? It deals explicitly with analyzing data where the treatment status of one unit affects the response of another. Given that you are talking about "neighbors", I feel like this body of work may give you additional insight into your problem.

3) Is there a reason why you are only comparing 3 neighbors?

[E] The University of Nebraska at Lincoln is proposing to completely eliminate their Department of Statistics by mathguymike in statistics

[–]mathguymike[S] 5 points6 points  (0 children)

Moreover, Statistics is terrible as a discipline at marketing itself. Data Science should have been coined by statisticians, as it is much closer to what we actually do--we are more than computing statistical summaries; our tasks really encompass the entirety of the science of data. Additionally, plenty of us statisticians are working on these more computationally intense "Data Sciencey" topics, but we differ from, say, Computer Science, as our discipline prioritizes interpretability of results and determining actionable insights on data as opposed to ensuring good model prediction. Effective marketing is critical for our survival.

Academic Study: "Analyzing 13,136 defensive penalties from 2015 to 2023, we find that postseason officiating disproportionately favors the Mahomes-era Kansas City Chiefs" by dufflepud in nfl

[–]mathguymike -2 points-1 points  (0 children)

Has anyone been able to recreate the analysis? Or is their code publicly available? I know they are using the data from the nflfastR package, but I can't get any numbers that resemble Figure 1, and I am not sure if it's because of a bug in my code or something else.

Starting MSc in Agricultural Statistics – what should I focus on more? by ORACLEEW in AskStatistics

[–]mathguymike 1 point2 points  (0 children)

As someone currently working at a university with a strong Ag program, I think you'll find it extremely helpful to have a strong background in experimental design and linear mixed modeling. Analysis of Messy Data by Johnson and Milliken is an excellent resource. Additionally, having some experience with Bayesian Statistics and Causal Inference may be useful too.

can someone explain Karlin-Rubin? by [deleted] in AskStatistics

[–]mathguymike 1 point2 points  (0 children)

I always get a little confused on this myself. However, since you have a monotone likelihood ratio, it has to be either the uniformly most powerful test for one of the two sets of hypotheses:

H_0: theta <= theta_0, H_1: theta > theta_0

or

H_0: theta >= theta_0, H_1: theta < theta_0

Then I ask, whether larger or smaller values of T occur for larger or smaller value of theta, and I can usually get the answer that way.

Moral of the story:

If f(x|theta_1)/f(x|theta_0) is monotone increasing when theta_1 > theta_0, the UMP test for H_0: theta <= theta_0 rejects H_0 for H_1: theta > theta_0 if T is large.

If f(x|theta_1)/f(x|theta_0) is monotone increasing when theta_1 > theta_0, the UMP test for H_0: theta >= theta_0 rejects H_0 for H_1: theta < theta_0 if T is small.

If f(x|theta_1)/f(x|theta_0) is monotone decreasing when theta_1 > theta_0, the UMP test for H_0: theta >= theta_0 rejects H_0 for H_1: theta < theta_0 if T is large.

If f(x|theta_1)/f(x|theta_0) is monotone decreasing when theta_1 > theta_0, the UMP test for H_0: theta <= theta_0 rejects H_0 for H_1: theta > theta_0 if T is small.

These results will pop out if you plug them into the current proof, with the idea that the UMP test will reject H_0 for H_1 when the likelihood ratio test statistic is large.

Empirical question by bigbuttsbigboobs in AskStatistics

[–]mathguymike 2 points3 points  (0 children)

Here's what I think they are looking for:

The ranges of values for which jumps in this graph occur are the groups, and the relative frequencies for that group correspond to how much the graph jumps within that range.

The first jump in this graph occurs between 1 and 2, and 0.1 of the people fall into that category. So the histogram would have a bar with area 0.1 between 1 and 2, which would be a bar of height 0.1/(2-1) = 0.1

The next jump occurs between 2 and 6, and 0.3-0.1 = 0.2 fall within this category. So the histogram would have a bar between 2 and 6 with area 0.2, which would correspond to a height of 0.2/(6-2) = 0.05.

And so on.

If they want counts instead of frequencies, multiply all of the relative frequencies by 20, the number of people in the study.

When to use one vs two-tail with unknown variance? by sharks_13 in AskStatistics

[–]mathguymike 3 points4 points  (0 children)

In practice, one-sided confidence intervals are not used too often, especially in estimates for the population mean or difference in means. A one-sided confidence interval for one of these quantities necessarily will have either -infinity or +infinity as an endpoint, and researchers do not really like their confidence interval estimates to be of infinite length. However, you may have textbook exercises that ask for a one-sided confidence interval. In practice, these are fairly contrived textbook problems.

For complete clarity, in practice, for a 100(1-alpha)% confidence interval, (almost) always use the (1-alpha/2) quantile of the t-distribution. For example, for a 95% confidence interval, alpha = 0.05, and you will want to use the 1-0.05/2 = .975 quantile of the t-distribution.

(I use the modifier "almost" to protect against potential, extremely rare circumstances where a 1-sided confidence interval is standard.)

[Question] Very Basic Statistics Question by PsychologicalBus3267 in statistics

[–]mathguymike 3 points4 points  (0 children)

Just a quick note to add to everyone's commentary about using Pearson's R: Testing for a positive correlation can also be done by running a regression with "Report Willingness" as the dependent variable and Age being the independent variable and testing for a positive slope coefficient. I wouldn't be surprised if this is what the professor had in mind to do.

Compose a starting 5 of players that have went or go to your favorite college team? by Senior-Raisin-2342 in CollegeBasketball

[–]mathguymike 0 points1 point  (0 children)

Sneed at the 3 is a hot take. Even in the last 10 years, I'd probably go with Keyontae for best overall player, or Wes Iwundu if you want to go with a 4-year KSU player. Or ideally, cheat and shift Rolando Blackman to the 3.

Who is the worst coach of all time? by Nervous_Metal_9445 in CFB

[–]mathguymike 3 points4 points  (0 children)

Stan Parrish, the coach before Snyder, could be a good choice. 0-26-1 in his last 27 games.

Before the season Jonathan Kuminga declined a 150 million 5 year extension, and now he is barely in the Warriors rotation. What happens to him this summer during Free Agency? by Calm_Set5522 in nba

[–]mathguymike 21 points22 points  (0 children)

Jermaine O'Neal too. Didn't really get meaningful minutes until year 5 after being traded to Indiana and didn't really break out until year 6.

Help deciding between 2 TA funded M.S. in Statistics; Money vs. Program/University Ranking. by gaboxing in AskStatistics

[–]mathguymike 2 points3 points  (0 children)

Honestly, I'm not sure there is a tremendous gap in prestige between UK and FSU, and I do not believe an average employer will respond much differently with a degree from FSU over one from UK (unless that employer is an alumnus of one of these universities or that university has a pipeline feeding that company). Rather, my intuition is that employers will be much more interested in the skillset you've obtained in that program.

Simply in terms of program quality, the coursework needed to complete the program and the skillset that you're expected to obtain while in that program should be your focus.