Posted by ibobev 2 days ago
No, the central limit theorem specifically doesn't address that. It says that the sum of iid random variables is well approximated by a normal distribution near the mean; it doesn't tell you how well that approximation works in the tails. The rarer the event you're modeling is, the less relevant the normal approximation is.
What are "most cases"?
a vast amount of fluff for less than a college statistics professor would (hopefully) be able to impart with a chalkboard in 10 minutes, when Quanta has the ability to prepare animated diagrams like 3Blue1Brown but chooses not to use it
they could go down myriad paths, like how it provides that random walks on square lattices are asymptotically isotropic, or give any other simple easy-to-understand applications (like getting an asymptotic on the expected # of rolls of an n-sided die before the first reoccurring face) or explain what a normal distribution is, but they only want to tell a story to convey a feeling
they are a blight upon this world for not using their opportunity to further public engagement in a meaningful way
Perhaps you're just not in their intended audience?
https://news.ycombinator.com/item?id=45800657
3b1b doesn't have the same goal as Quanta, or as introductory guides. It's actually not that great a teaching tool (it's truly great at what it is for, which is (a) appreciation and motivation, and (b) allowing people to signal how smart they are on message board threads by talking about how much people would get out of watching 3b1b).
This is prose writing about math. It's something you're meant to read for enjoyment. If you don't enjoy it, fine; I don't enjoy cowboy fiction. So I don't read it. I don't so much look for opportunities to yell at how much I hate "The Ballad of Easy Breezy".
My compliant is only that there should be a dozen more just like them, each competing with each other for the best, most engaging math and science content. This would allow for more a broader audience skillevel to be reached.
As it stands, we’re lucky even to have Quanta and 3b1b.
I think there is hope though, quite a few new-ish creators on YouTube are following in Grant’s footsteps and producing very technically detailed and informative content at similar quality levels.
by the metric of "if this expository piece were to be taken to a time before its subject had been considered and presented to researchers, how useful would its outline be towards reproducing the theory in its totality," Quanta's writings (on both classical and research math) mostly score 0
Seems a bit like Ted Talks. Lightweight popcorn for the simple minded.
> suppose that a large sample of observations is obtained, each observation being randomly produced in a way that does not depend on the values of the other observations, and the average (arithmetic mean) of the observed values is computed. If this procedure is performed many times, resulting in a collection of observed averages, the central limit theorem says that if the sample size is large enough, the probability distribution of these averages will closely approximate a normal distribution.
> Laplace distilled this structure into a simple formula, the one that would later be known as the central limit theorem. No matter how irregular a random process is, even if it’s impossible to model, the average of many outcomes has the distribution that it describes. “It’s really powerful, because it means we don’t need to actually care what is the distribution of the things that got averaged,” Witten said. “All that matters is that the average itself is going to follow a normal distribution.”
This is not really true, because the central limit theorem requires a huge assumption: that the random process has finite variance. I believe that distributions that don't satisfy that assumption, which we can call heavy-tailed distributions, are much more common in the real world than this discussion suggests. Pointing out that infinities don't exist in the real world is also missing the point, since a distribution that just has a huge but finite variance will require a correspondingly huge number of samples to start behaving like a normal distribution.
Apart from the universality, the normal distribution has a pretty big advantage over others in practice, which is that it leads to mathematical models that are tractable in practice. To go into a slightly more detail, in mathematical modeling, often you define some mathematical model that approximates a real-world phenomenon, but which has some unknown parameters, and you want to determine those parameters in order to complete the model. To do that, you take measurements of the real phenomenon, and you find values for the parameters that best fit the measurements. Crucially, the measurements don't need to be exact, but the distribution of the measurement errors is important. If you assume the errors are independent and normally distributed, then you get a relatively nice optimization problem compared to most other things. This is, in my opinion, about as much responsible for the ubiquity of normal distributions in mathematical modeling as the universality from the central limit theorem.
However, as most people who solve such problems realize, sometimes we have to contend with these things called "outliers," which by another name are really samples from a heavy-tailed distribution. If you don't account for them somehow, then Bad Things(TM) are likely to happen. So either we try to detect and exclude them, or we replace the normal distribution with something that matches the real data a bit better.
Anyway, to connect this all back to the central limit theorem, it's probably fair to say measurement errors tend to be the combined result of many tiny unrelated effects, but the existence of outliers is pretty strong evidence that some of those effects are heavy-tailed and thus we can't rely on the central limit theorem giving us a normal distribution.
The point on convergence rates re: the central limit theorem is also a major point otherwise clever people tend to miss, and which comes up in a lot of modeling contexts. Many things which make sense "in the limit" likely make no sense in real world practical contexts, because the divergence from the infinite limit in real-world sizes is often huge.
EDIT: Also from a modeling standpoint, say e.g. Bayesian, I often care about finding out something like the "range" of possible results for (1) a near-uniform prior, (2), a couple skewed distributions, with the tail in either direction (e.g. some beta distributions), and (3) a symmetric heavy-tailed distribution (e.g. Cauchy). If you have these, anything assuming normality is usually going to be "within" the range of these assumptions, and so is generally not anything I would care about.
Basically, in practical contexts, you care about tails, so assuming they don't meaningfully exist is a non-starter. Looking at non-robust stats of any kind today, without also checking some robust models or stats, just strikes me as crazy.
The sum of independent identically distributed random variables, if they converge at all, converge to a Levy stable distribution (aka fat-tailed, heavy tailed, power law). In this sense, Levy stable distributions are more "normal" than the normal distribution. They also show up with regular frequency all over nature.
As you point out, infinite variance might be dismissed but, in practice, this just ends up getting larger and larger "outliers" as one keeps drawing from the distribution. Infinities are, in effect, a "verb" and so an infinite variance, in this context, just means the distributions spits out larger and larger numbers the more you sample from it.
This is a tautology to the extreme.
If sums of independent identically distributed random variables converge to a distribution, they converge to a Levy stable distribution [0]. Tails of the Levy stable distribution are power law, which makes them not Gaussian.
Second, your "aka" is incorrect --- there is all sorts of clumping that is not a normal distribution.
> your "aka" is incorrect --- there is all sorts of clumping that is not a normal distribution.
That it's "incredibly common for people to label "bell curves" by eyeball, regardless of whether they are normal curves" is not just not relevant, it's anti-relevant ... the central limit theorem says that the distribution of the means is always a bell curve--a normal distribution--not merely a "bell curve".
Anyway, this is covered in far more detail in other comments and material elsewhere, so this is my last contribution.