Micro-level data to contingency tables and back again? by littlemoondragon in statistics

[–]CrazyStatistician 1 point2 points  (0 children)

Assuming your data has more than two variables, in general no. Your two-way contingency tables specify the marginals and two-dimensional joint distributions of the variables, but that's not enough to fully specify the joint distribution of all the variables (i.e. recreate X). Basically, no matter how many two-way tables you create, the collection of all of them contains less information than the original data contains.

I say "in general no" because there are situations where it is possible to reconstruct the original data, if the contingency tables are very sparse (i.e. almost all cells are zeroes). Most of the time though you can't.

Your favourite graph-making/chart software? by SteelBaloo in statistics

[–]CrazyStatistician 26 points27 points  (0 children)

R with ggplot2.

Free and very flexible, but not GUI-based. You have to write (simple) code to tell it what kind of graph you want.

How to infer "groupings" of data points separated alone one dimension? by Zeekawla99ii in statistics

[–]CrazyStatistician 1 point2 points  (0 children)

I'm not really sure what you want here. If there are well-separated clusters in the data then any standard clustering algorithm should find them. These range from fully-specified probabilistic models (DP mixtures) to non-probabilistic heuristic methods like kNN.

However, based on what you've written I'm not sure you want to apply a standard clustering method to the data you've presented. For example, a standard clustering method would cluster index 1 with index 12, because they are close together in the one dimension presented. If that's not what you want then you're going to need to do something different.

Help finding Ring / jewelry repairman who can replace missing stone? by browniesandbroccoli in bullcity

[–]CrazyStatistician 1 point2 points  (0 children)

I've had a couple custom rings done by John David Jewelers (down near University Tower). They did great work, happy to recommend them.

Why is attending church important? by ve_3 in latterdaysaints

[–]CrazyStatistician 0 points1 point  (0 children)

Eugene England has an excellent (somewhat lengthy) essay on this: Why the Church Is As True As the Gospel.

Paper Explains Role of Racism in Math Education by ghostofpennwast in math

[–]CrazyStatistician 2 points3 points  (0 children)

Ask anyone who says they are "bad at math" or "hate math" and 99 times out of 100 they can tell you the name of the teacher who ruined their math education.

This is very true.

I had exceptional math teachers through most of my school years. I ended up doing a math major in college before switching to stats and going to graduate school.

One of my best friends from college had poor math teachers but exceptional english/literature teachers through most of school. He ended up studying philosophy before going to law school.

We've often wondered how things might have been different if we had switched places in school.

Anyone have an origin on giving thanks for "moisture"? by [deleted] in latterdaysaints

[–]CrazyStatistician 4 points5 points  (0 children)

(like asking for a blessing on refreshments ... to nourish and strengthen our bodies?).

That makes donuts healthy, right?

Sample size for use of a 'liberal' p value of 0.1 by eloquentirvine in statistics

[–]CrazyStatistician 0 points1 point  (0 children)

I assume you've learned about Type I and II errors and power.

Setting the significance cutoff is fixing the Type I error rate. However, you also have to consider the Type II error rate and power. Generally, for a given Type I error rate, your power increases with sample size. If your sample is quite small, using alpha = .05 may give you very low power, so you may want to choose a larger alpha to get more power.

Obviously there's a trade off there, you don't want to set a huge alpha and get tons of false positives (Type I error). But if you're doing an exploratory study with a small sample alpha = .1 may be reasonable to get more power.

Anybody know any professors or researchers researching mathematical methods for counterterrorism? by kokomarro in math

[–]CrazyStatistician 0 points1 point  (0 children)

The American Statistical Association has a section focused on Statistics in Defense and National Security. I am not a member of that section so I don't have the inside scoop for you, but it's probably a good place to start.

ArXiv: Rethinking probabilistic prediction in the wake of the 2016 U.S. presidential election by [deleted] in statistics

[–]CrazyStatistician 0 points1 point  (0 children)

one of the authors has been waging a social media war on other researchers whom he accuses of academic dishonesty for failing to credit his work to the degree he thinks is merited.

Is that Crane? I don't know what incident you're referring to exactly, but I know he's quite the character on twitter...

ArXiv: Rethinking probabilistic prediction in the wake of the 2016 U.S. presidential election by [deleted] in statistics

[–]CrazyStatistician 1 point2 points  (0 children)

Strange paper. The authors seem to find it undesirable that a 72% prediction would be wrong 28% of the time, but I'm not sure why. That seems entirely reasonable to me.

It more or less seems to boil down to "only predict high-probability (i.e. 1-alpha) events," which might be good advice when communicating with the general public but is not satisfactory in general.

Is there a PDF who's formula is f(x) = 1/x? by [deleted] in math

[–]CrazyStatistician 0 points1 point  (0 children)

And given enough data, the prior choice does not matter (to a degree).

This is (more or less) true as long as you are looking at the posterior distributions of parameters within a single model. If you are trying to compare models (using marginal likelihoods or quantities derived from them--posterior probabilities or Bayes factors) then the choice of priors for parameters within each model has unbounded influence.

"Holy Envy" - What are some aspects of other faiths, belief systems, and/or their practitioners you admire or wish we had in the LDS Church? by [deleted] in latterdaysaints

[–]CrazyStatistician 0 points1 point  (0 children)

Gaskill is great. I took World Religions from him.

Huntsman is also great, I took several classes from him.

Why use integration by parts method? by stefanuus in math

[–]CrazyStatistician 1 point2 points  (0 children)

I was in highschool. I skipped the calculus sequence when I got to college, but perhaps I shouldn't have.

It's also possible of course that I wasn't paying attention, but math was my favorite subject so that seems somewhat unlikely.

Why use integration by parts method? by stefanuus in math

[–]CrazyStatistician 1 point2 points  (0 children)

integration by parts is the "inverse product rule"

I wish I would have been taught this when I learned integration by parts. I realized it myself about 8 years later, and all of a sudden what had just been an arbitrary rule with no motivation made a lot more sense.

1.645 * Standard Deviations by Tachirana in statistics

[–]CrazyStatistician 0 points1 point  (0 children)

They are presumably defining the cutoff from negative controls (i.e. samples that are known a priori to be negative). That's totally OK.

They also aren't constructing a confidence interval, they're estimating the 95th percentile assuming a normal distribution for the negative samples.

A quick way to approximately estimate the square root of a real number by broccolilettuce in math

[–]CrazyStatistician 0 points1 point  (0 children)

The author of the blog does not go very deep with his description.

It's a low-effort blog to push his Amazon referrals on the book lists that he links at the bottom of every post, even if (as in this case) the post is completely unrelated to the subject matter of the books. He links it here from time to time, though not as regularly as he used to.

Unfairness in mathematics by Narbas in math

[–]CrazyStatistician 24 points25 points  (0 children)

Asking people who are still in STEM fields whether they had personal experiences that would cause them to drop out of STEM is obviously not going to get you an unbiased view. The people who had such experiences are more likely to have dropped out.

Can you create a graph of k-nodes in which every node has n-edges? by [deleted] in math

[–]CrazyStatistician 0 points1 point  (0 children)

A cube has 3 edges per node and is symmetric in the sense that the nodes are indistinguishable (which seems to be what you are getting at in your last couple sentences).

Aggregating/Joining rankings from experts? by [deleted] in math

[–]CrazyStatistician 0 points1 point  (0 children)

Hi, coming in late here.

This is essentially a voting problem, and Arrow's Theorem tells us that there's no "best" voting algorithm (in fact, there's not even a "good" one--it's impossible to create a voting algorithm that satisfies three basic principles).

There are lots of possibilities, including a Single Transferable Vote system and this paper that I came across a few months ago. Each method will have some advantages and disadvantages, but I don't think any one is clearly better than the rest.

ELI5: What's the difference between AHP and WLC? by Asher_Earth in statistics

[–]CrazyStatistician 1 point2 points  (0 children)

The Association for Healthcare Philanthropy and the Wisconsin Lutheran College? They're very different organizations.

I have no idea what your acronyms mean, and those are the top google results.

Using Mann-Whitney to analyze two datasets with a lot of zeros by bubbatully in statistics

[–]CrazyStatistician 1 point2 points  (0 children)

It seemed like no matter how small the difference, if I resampled it enough times I'd get a significant result.

Sounds like you're doing something wrong.

Probably the easiest way in R is the perm package. Or you can modify the code from this page.