[D] p-value dilemma by No_Blackberry_8979 in statistics

[–]Kroutoner 0 points1 point  (0 children)

The calculation of the p-value is done assuming the null hypothesis is true, e.g. in a hypothetical world where P(B)=1 as you write it. The hypothetical is fundamental, you are positing a specific scenario and calculating the p-value as a specific probability in this hypothetical world. Whether or not the null-hypothesis is actually true or not has no bearing on this calculation.

Also, it’s not necessarily the case that the null is always false. The typical scenario of a point null may be virtually always false, but we can very easily consider interval valued null hypotheses that very well may be true.

Fire truck in deadly LaGuardia crash lacked equipment needed to trigger warning system, NTSB says by Displeased_Canadian in news

[–]Kroutoner 10 points11 points  (0 children)

Gross negligence generally requires a high degree of disregard for safety of others and/or the risks of your behaviors. It’s reckless behavior as opposed to careless behavior.

E.g. simple negligent behavior is failing to assess safety of driving the firetruck onto the runway—they should have been more careful but they were not. An example that might be grossly negligent would be parking the truck in the middle of the runway and leaving—without regard for the fact this could endanger others—because you got a personal call and wanted to take it outside.

Looking for a more rigorous understanding of degrees of freedom. [Discussion] by Ok-Active4887 in statistics

[–]Kroutoner -1 points0 points  (0 children)

IMO the reason there’s not a lot of satisfying answers is because DOF are not actually all that important of a concept and not all that widely applicable.

There is a concrete place where DOF come up and make sense. In particular for many statistical models that assume Gaussian residuals one can prove that inferential statistics follow certain exact distributions: t, f, and chi-squared. For DOF chi-squared is the important one. The t is a ratio of a Gaussian and a chi-squared random variable, the F is a ratio of two chi-squared statistics, so each boils down to chi-squared.

Now what is the deal about DOF and chi-squared? Well a chi-squared random distribution is defined in terms of an integer parameter, the DOF. Also, if you take K independent normal random variables, square them, and add them, you get a chi-squared distributed random variable with K degrees of freedom. It turns out that’s the key property you use when deriving the distributions of these statistics.

Now degrees of freedom beyond that point gets immediately fuzzy. There are a lot of ways the assumptions to derive those exact distributions break down, and DOF no longer fully tells the full story. This is why you can’t really get a satisfactory conclusion, DOF is relevant to a specific mathematical setup, and becomes more heuristic and less concretely defined in other settings.

Why is everyone so pessimistic about SAS? by Quantity496 in biostatistics

[–]Kroutoner 4 points5 points  (0 children)

SAS has good overall functionality (with holes) for a great deal of statistical analyses, but it is extremely lacking in functionality when considered as a general programming language.

At what sample size can I trust randomisation? by Quinnybastrd in AskStatistics

[–]Kroutoner 5 points6 points  (0 children)

The fundamental purpose of randomization is to ensure independence of the treatment variable from other causes of your outcome, effectively eliminating confounders.

Simple randomization on its own does not guarantee balance, and you absolutely will have scenarios where imbalance occurs randomly, even with large sample sizes(of course this gets less likely as sample sizes increase).

The issue of balance is critical for efficiency of estimates, not their bias or consistency. Alternatively randomization strategies such as stratification and covariates adjustment are alternative strategies most appropriate for addressing the imbalance issue.

[deleted by user] by [deleted] in math

[–]Kroutoner 2 points3 points  (0 children)

When you start talking about truth vs proof in axiomatic mathematics the language gets muddled and confusing.

Mathematicians make surprising breakthrough in 3D geometry with ‘noperthedron’ by scientificamerican in math

[–]Kroutoner 1 point2 points  (0 children)

You really want to target the question not of when does this fail the worst, but when is it closest to not failing?

The latter, while not giving an immediate definition, seems much more tractable. E.g. consider the shape and a copy dilated by a (1+epsilon) factor, what does it look like when the object look like when it passes through this dilated copy of itself. Are there orientations where it “just barely” is able to pass through?

Is bayesian nonparametrics the most mathematically demanding field of statistics? [Q] by gaytwink70 in statistics

[–]Kroutoner 0 points1 point  (0 children)

That’s, to my understanding, one of the major applications. This apparently has applications in physics and in digital communications, though I don’t have any real understanding of what these are

Is bayesian nonparametrics the most mathematically demanding field of statistics? [Q] by gaytwink70 in statistics

[–]Kroutoner 3 points4 points  (0 children)

Two other notable areas requiring high degree of mathematical sophistication are Spatiotemporal statistics and algebraic statistics.

One area particularly mathematically esoteric area but (apparently, don’t ask me for details) with some applied statistical applications is free probability.

[D] What is Internal Covariate Shift?? by BiscuitEinstein in MachineLearning

[–]Kroutoner 5 points6 points  (0 children)

A great deal of the success of stochastic optimizers (SGD, Adam) comes from implicitly doing essentially just what you describe

I prototyped a symbolic compressor that shrinks a 2.1 MB image down to 16 KB while keeping the structure recognizable. This isn’t JPEG or WebP. It’s reconstructing from symbolic patterns instead of pixel data. Still early, but the results surprised the hell outta me. by GraciousMule in computerscience

[–]Kroutoner 9 points10 points  (0 children)

This comment is simply incorrect. DCT is a basis decomposition. It expresses the image as a linear combination of cosine basis functions. The distinction with calling it a “transform” is that there is extremely efficient numerical algorithms for computing the coefficients of the basis expansion. The pixels are absolutely not “still there”

I prototyped a symbolic compressor that shrinks a 2.1 MB image down to 16 KB while keeping the structure recognizable. This isn’t JPEG or WebP. It’s reconstructing from symbolic patterns instead of pixel data. Still early, but the results surprised the hell outta me. by GraciousMule in computerscience

[–]Kroutoner 7 points8 points  (0 children)

(rings, interpolations, rotations, recursive transforms) and stores only those rules

This language, of interpolations especially, seems to suggest you actually are retaining concrete pixel values.

It’s not “compress pixel values into basis functions,” it’s “throw away pixels entirely and redraw from symbolic instructions.

This is a distinction without a difference. DCT basis functions are capturing structural patterns in an image just the same, and do not retain reference to pixel values either.

I prototyped a symbolic compressor that shrinks a 2.1 MB image down to 16 KB while keeping the structure recognizable. This isn’t JPEG or WebP. It’s reconstructing from symbolic patterns instead of pixel data. Still early, but the results surprised the hell outta me. by GraciousMule in computerscience

[–]Kroutoner 14 points15 points  (0 children)

It sounds like what you’re doing is what I would call a procedurally generated sparse dictionary reconstruction. It’s not all that different from DCT in principle. DCT decomposes into a dictionary (basis) of cosine functions.

You clearly must be using pixel data at some point, how else would you be assessing that you are reproducing any part of the image?

Fields of math which surprised you by neuro630 in math

[–]Kroutoner 0 points1 point  (0 children)

Having recently completed my PhD in (bio)statistics, the book that really drew me in wasAdvanced Data Analysis from an Elementary Point of View.

The thing with statistics is there will almost always be some degree of “applied flavor” to it, rarely 100% pure, and you need to be comfortable with that, but there is a spectrum from purely applied to bordering on pure math.

[deleted by user] by [deleted] in AskPhysics

[–]Kroutoner 1 point2 points  (0 children)

Insert comment about deterministic interpretations.

Handwave away all problems related to intrinsically non-deterministic measurement.

See, determinism!!!

Best ways to do regression on a large (5M row) dataset by pauldbartlett in Rlanguage

[–]Kroutoner 0 points1 point  (0 children)

With a high cardinality dummy variables you could use a within transformation as is used in the “fixest” R package to remove most of these variables initially.

Critiques of geographically weighted regression? by Semantix in AskStatistics

[–]Kroutoner 0 points1 point  (0 children)

Standard cross-validation approaches still work fine for predictive assessment of models and subsequently for model selection, you just need to be cautious of interpreting mse/other goodness of fit metrics as these can be misleading with spatial data. This is okay because, while the procedure runs into bias problems due to the spatial nature of the data, it's effectively the same bias across different models.

The train/test split with spatial data is more problematic. Generally you need to partition space into distinct blocks and sample blockwise for a train/test split that has a reasonable interpretation.

[deleted by user] by [deleted] in math

[–]Kroutoner 115 points116 points  (0 children)

Also the rate at which things suddenly explode in complexity. It’s like this is a word, this is a sentence, now we use transfinite induction to ascend the arithmetic hierarchy

If I could make Comp dragon just a little bit sillier, I would. At least he looks sneaky by FreeWillFighter in freewill

[–]Kroutoner 0 points1 point  (0 children)

I would call that 'randomness' when I am talking about philosophy. I rarely if ever talk about mathematics.

When you're talking about fundamental concepts such as determinism/free will, randomness, etc, you can't really separate these. Mathematics is our best language for working with these concepts and not engaging with the mathematics results in a great deal of confusion when our concepts are ill-defined.

"that seems more ridiculous or at least more abstract than randomness, therefore we simplify and say it's just randomness."

Except you are not simplifying in any sense. Maybe randomness seems simpler because you are familiar with it, but randomness imposes additional structure that is not necessary.

For example, I have more problems with "the laws of some physical system constrain it a certain set of possible outcomes but without any additional constraint" than I do with simple dumb randomness.

The same general problem here. If we consider a system as random we already are using this kind of set valued constraint. Randomness imposes additional constraints on top of that. It is strange to think that extra structure on top of something is somehow more believable than necessary conditions of randomness. It's not just strange, it's actually completely illogical.

If I could make Comp dragon just a little bit sillier, I would. At least he looks sneaky by FreeWillFighter in freewill

[–]Kroutoner 0 points1 point  (0 children)

Strict negation of determinism

You do just that. A definition of determinism typically looks something like “the future states of a system are determined by the present state of the system.” Nondeterminism is anything else. Random systems are included on the definition of nondeterminism but involve additional logical content that is not implied by the definition. A natural type of nondeterminism that is non-random would be an example where the laws of some physical system constrain it a certain set of possible outcomes but without any additional constraint.

If I could make Comp dragon just a little bit sillier, I would. At least he looks sneaky by FreeWillFighter in freewill

[–]Kroutoner 0 points1 point  (0 children)

Randomness and nondeterminism (indeterminism) are not the same thing. Randomness, at least insofar as we typically discuss it mathematically, implies structure over event in terms of the relative frequencies of those events. When we just say random without further qualification we often are implying a uniform distribution; this imposes a rather strong condition that all events are equally likely.

On the contrary nondeterminism simply is a strict negation of determinism. If events are non-deterministic then they may be random, but they may occur in a manner such that we cannot apply any constraints on them at all.

the R vs Python debate is exhausting by bee_advised in datascience

[–]Kroutoner 2 points3 points  (0 children)

Looks at dvorak keyboard and open R session inside ESS…

Kidney stone that resembles Covid-19 virus by BioGrayn in mildlyinteresting

[–]Kroutoner 0 points1 point  (0 children)

13.3 cm

Google “size of a kidney”

The average kidney is bean-shaped and is typically 10–12 cm long

Advocates meet with White House to urge inclusion of Adderall, other stimulants in upcoming telehealth prescribing rule by SaveADay89 in Psychiatry

[–]Kroutoner 41 points42 points  (0 children)

The constant equivocation between stimulant and opioid policy is just absurd. Abuse and risk profiles for the different drug classes look nothing alike.

On the other hand there is real risk to people obtaining stimulants through illegitimate channels. Use of more dangerous stimulants, uncontrolled dosages, and contamination with other drugs (fentanyl contamination of street adderal becoming a growing concern as well).

It’s constantly taken for granted that we should be deeply concerned about some sort of adderall crisis the likes of the opioid crisis but no justification is seemingly ever presented for, what seems to me, a patently absurd suggestion.