CRITICAL THINKING - Fundamentals: Bayes' Theorem [HD]
Video Statistics and Information
Channel: Wireless Philosophy
Views: 338,971
Rating: 4.9009356 out of 5
Keywords: Khan Academy, Philosophy, Wireless Philosophy, Wiphi, video, lecture, course, critical thinking, logic, mathematics, probability, Thomas Bayes, Bayes Theorem, epistemology, bayesian, base rate fallacy, Ian Olasov, CUNY, City University of New York, confirmation theory, statistics
Id: OqmJhPQYRc8
Channel Id: undefined
Length: 6min 20sec (380 seconds)
Published: Fri Apr 22 2016
Please note that this website is currently a work in progress! Lots of interesting data and statistics to come.
I'm in medicine and this reminds me of the saying "Uncommon presentations of common diseases is more common than common presentations of uncommon diseases".
If it helps, the way that I was (conceptually) taught Bayes' Theorem was thus:
1) You hear strange noises in the room next door
2) You hypothesize that the noises come from dinosaurs mating
3) You note that "that's exactly what I would hear if there were dinosaurs mating in the next room"
4) BUT then you note that that likelihood of there being dinosaurs in the next room is very LOW...
5)...and the likelihood of hearing (any) strange noises coming from an adjacent room is probably pretty high...
6) ...so in the end, the actual probability of your original hypothesis is very very low.
While it's a non empirical mnemonic, it was a weird image that it stuck with me and I can always reconstruct the formulaic version of Bays theorem from just that example.
I'm not a philosophy type, but I figured you all would be a little upset at his definition of probability being "The likelihood or chance of something happening". Otherwise cool video!
Could you use Bayes theorem to explain the Monty hall problem? First the probability is 1/3, then some new information is given. But which probabilities should go where in the theorem?
If someone wants to expertly make and evaluate likelihood claims, then that person needs to master (not just have a vague acquaintance with) statistics.
IMO, when someone makes a likelihood claim, one should have no qualms asking to be shown a formal likelihood calculation - it's how to cut through the crap.
Edit: aaand as expected this is
hugelycontentious in this subreddit.Edit2: ummm.
Arguably the single most important lesson omitted from common American education.
I like that they put in the Base Rate Fallacy as an important concept to check. I find it is incredibly common even in popular beliefs, understandings, and observations.
For example, one of the most popular business management books of the past decade is Jim Collins' Good to Great. Most people think of the book, and the research it is based on, as providing the qualities of a company that will help it go from good to great. (It defines what these mean in quantitative terms, of course.) That is, you are interested in the probability that your company will go from good to great if it has these properties. People believe that having these properties gives a high probability of going good to great. That is, it is presented that P(good to great | specific properties) is high.
The problem is that the research actually investigates the probability of the evidence given the hypothesis. It finds the probability of the properties given that a company has gone from good to great, P(specific properties | good to great) and presents only those properties that have high values -- in their case 100%. (To get this they also create vague and subjective properties, which is another problem.) It gets more complicated in that they have a "control" group of companies that didn't go from good to great and made sure those one also didn't have the properties, meaning the result book describes properties for which P(property X | good to great) = 1 AND P(property X | NOT good to great) = 0, based on their cherry picked groups (not randomly selected).
They never investigated P(good to great) which is very small, P(NOT good to great) which is very large, or P(property X) which is completely unknown. So the book is useless to me.
The same base rate error is used in Dan Buettner's "Blue Zones" of unusually long-lived communities of the world, described in a TED Talk entitled "How to live to be 100+". The title claims P(100+ | lifestyle choice X) but, unfortunately, investigates P(lifestyle choice X | 100+), with no mention of P(100+) or P(lifestyle choice X).
I think the same error is true of much of the "social justice" movements and their very bad statistics, particularly with respect to the concept of privilege, and particularly in the use of the Progressive Stack. (Let's ignore that they treat all people by the identity group even if we have direct evidence of the privilege level of the individual.)
What they are actually arguing is that, based on identity properties like gender, skin color, or sexual orientation, they can designate a person's privilege and provide a counteracting policy to exclude members of the higher privileged identity groups and help the members of the lower privileged identity groups.
So, for example, a policy of excluding whites or males will come from the belief that whites and males have power and wealth, and this is determined by noting that those with power and wealth are statistically likely to be white and male. That is, they use the probability that somebody is white given they have wealth and power -- P(white | wealth & power) -- and the probability that they are male -- P(male | wealth & power) -- are higher than for others with wealth and power.
But this is a base rate fallacy, as it says the people with wealth & power tend to be white and male, not the whites and males tend to have wealth & power. That is, their policies are based on the believe that P(wealth & power | white) and P(wealth & power | male) are quite high, compared to others, but they use the inverse probabilities in their reasoning. (Oddly, the conditional information of the actual wealth & power of a given individual seems irrelevant in such policies for some reason.)
I'm finding more and more examples of where people make these serious mistakes and make serious policies based on them. Business practices, lifestyle choices, and social policies are being built on them based on erroneous thinking.
It's still not clear to me why this base rate fallacy is so common a psychological error. Most people understand that all crows are birds, but only a few birds are crows, but when you switch from absolute essentialist statements to probabilities, the concept gets mixed up so easily.
How does a person determine prior probabilities of H and E?
Anyone who wants a more in-depth explanation should check this out:
http://www.yudkowsky.net/rational/bayes
It's the first explanation that actually made it click in my head