Random variables and probability distributions
Law of large numbers Introduction to the law of large numbers
Law of large numbers
- Let's learn a little bit about the law of large numbers
- which is on many levels, one of the most intuitive laws
- in mathematics and in probability theory.
- But because it's so applicable to so many things, it's often
- a misused law or sometimes, slightly misunderstood.
- So just to be a little bit formal in our mathematics,
- let me just define it for you first and
- then we'll talk a little bit about the intuition.
- So let's say I have a random variable, X.
- And we know its expected value or its population mean.
- The law of large numbers just says that
- if we take a sample of n observations of our random variable,
- and if we were to average all of those observations--
- and let me define another variable.
- Let's call that x sub n with a line on top of it.
- This is the mean of n observations of
- our random variable.
- So it's literally this is my first observation.
- So you can kind of say I run the experiment once and
- I get this observation and I run it again, I get that observation.
- And I keep running it n times and
- then I divide by my number of observations.
- So this is my sample mean.
- This is the mean of all the observations I've made.
- The law of large numbers just tells us that my sample mean
- will approach my expected value of the random variable.
- Or I could also write it as my sample mean will approach
- my population mean for n approaching infinity.
- And I'll be a little informal with what does approach or
- what does convergence mean?
- But I think you have the general intuitive sense that
- if I take a large enough sample here that I'm going to end up
- getting the expected value of the population as a whole.
- And I think to a lot of us that's kind of intuitive.
- That if I do enough trials that over large samples, the trials
- would kind of give me the numbers that I would expect
- given the expected value and the probability and all that.
- But I think it's often a little bit misunderstood
- in terms of why that happens.
- And before I go into that let me give you
- a particular example.
- The law of large numbers will just tell us that-- let's say
- I have a random variable-- X is equal to the number of heads
- after 100 tosses of a fair coin-- tosses or flips
- of a fair coin.
- First of all, we know what the expected value of
- this random variable is.
- It's the number of tosses, the number of trials times
- the probabilities of success of any trial.
- So that's equal to 50.
- So the law of large numbers just says if I were to take a sample
- or if I were to average the sample of a bunch of these trials,
- so you know, I get-- my first time I run this trial
- I flip 100 coins or have 100 coins in a shoe box and
- I shake the shoe box and I count the number of heads, and I get 55.
- So that Would be X1.
- Then I shake the box again and I get 65.
- Then I shake the box again and I get 45.
- And I do this n times and then I divide it by
- the number of times I did it.
- The law of large numbers just tells us that this the average
- the average of all of my observations,
- is going to converge to 50 as n approaches infinity.
- Or for n approaching 50.
- I'm sorry, n approaching infinity.
- And I want to talk a little bit about why this happens
- or intuitively why this is.
- A lot of people kind of feel that oh, this means that
- if after 100 trials that if I'm above the average that somehow
- the laws of probability are going to give me more heads
- or fewer heads to kind of make up the difference.
- That's not quite what's going to happen.
- That's often called the gambler's fallacy.
- Let me differentiate.
- And I'll use this example.
- So let's say-- let me make a graph.
- And I'll switch colors.
- This is n, my x-axis is n.
- This is the number of trials I take.
- And my y-axis, let me make that the sample mean.
- And we know what the expected value is, we know
- the expected value of this random variable is 50.
- Let me draw that here.
- This is 50.
- So just going to the example I did.
- So when n is equal to-- let me just [INAUDIBLE]
- So my first trial I got 55 and so that was my average.
- I only had one data point.
- Then after two trials, let's see, then I have 65.
- And so my average is going to be 65 plus 55 divided by 2.
- which is 60.
- So then my average went up a little bit.
- Then I had a 45, which will bring my average
- down a little bit.
- I won't plot a 45 here.
- Now I have to average all of these out.
- What's 45 plus 65?
- Let me actually just get the number just
- so you get the point.
- So it's 55 plus 65.
- It's 120 plus 45 is 165.
- Divided by 3.
- 3 goes into 165 5-- 5 times 3 is 15.
- It's 53.
- No, no, no.
- So the average goes down back down to 55.
- And we could keep doing these trials.
- So you might say that the law of large numbers tells this,
- OK, after we've done 3 trials and our average is there.
- So a lot of people think that somehow the gods of probability
- are going to make it more likely that we get fewer
- heads in the future.
- That somehow the next couple of trials are going to have to
- be down here in order to bring our average down.
- And that's not necessarily the case.
- Going forward the probabilities are always the same.
- The probabilities are always 50% that
- I'm going to get heads.
- It's not like if I had a bunch of heads to start off with or
- more than I would have expected to start off with, that
- all of a sudden things would be made up and I would get more tails.
- That would the gambler's fallacy.
- That if you have a long streak of heads or you have
- a disproportionate number of heads, that at some point
- you're going to have-- you have a higher likelihood of having
- a disproportionate number of tails.
- And that's not quite true.
- What the law of large numbers tells us is that it doesn't care
- Let's say after some finite number of trials
- Your average actually-- it's a low probability of this happening,
- but let's say your average is actually up here.
- Is actually at 70.
- You're like, wow, we really diverged a good bit from
- the expected value.
- But what the law of large numbers says, well,
- I don't care how many trials this is.
- We have an infinite number of trials left.
- And the expected value for that infinite number of trials,
- especially in this type of situation is going to be this.
- So when you average a finite number that averages out to
- some high number, and then an infinite number that's going to
- converge to this, you're going to over time, converge back
- to the expected value.
- And that was a very informal way of describing it,
- but that's what the law or large numbers tells you.
- And it's an important thing.
- It's not telling you that if you get a bunch of heads that
- somehow the probability of getting tails is going
- to increase to kind of make up for the heads.
- What it's telling you is, is that no matter what happened
- over a finite number of trials, no matter what the average is
- over a finite number of trials, you have
- an infinite number of trials left.
- And if you do enough of them it's going to converge back
- to your expected value.
- And this is an important thing to think about.
- But this isn't used in practice every day with the lottery and with casinos
- because they know that if you do large enough samples
- and we could even calculate
- if you do large enough samples,
- what's the probability that things deviate significantly?
- But casinos and the lottery every day operate on this principle
- that if you take enough people-- sure,
- in the short-term or with a few samples,
- a couple people might beat the house.
- But over the long-term the house is always going to win
- because of the parameters of the games that
- they're making you play.
- Anyway, this is an important thing in probability and
- I think it's fairly intuitive.
- Although, sometimes when you see it formally explained
- like this with the random variables and that
- it's a little bit confusing.
- All it's saying is that as you take more and more samples,
- the average of that sample is going to
- approximate the true average.
- Or I should be a little bit more particular.
- The mean of your sample is going to converge to
- the true mean of the population or
- to the expected value of the random variable.
- Anyway, see you in the next video.
Be specific, and indicate a time in the video:
At 5:31, how is the moon large enough to block the sun? Isn't the sun way larger?
Have something that's not a question about this content?
This discussion area is not meant for answering homework questions.
Share a tip
When naming a variable, it is okay to use most letters, but some are reserved, like 'e', which represents the value 2.7831...
Thank the author
This is great, I finally understand quadratic functions!
Have something that's not a tip or thanks about this content?
This discussion area is not meant for answering homework questions.
At 2:33, Sal said "single bonds" but meant "covalent bonds."
For general discussions about Khan Academy, visit our Reddit discussion page.
Here are posts to avoid making. If you do encounter them, flag them for attention from our Guardians.
- disrespectful or offensive
- an advertisement
- low quality
- not about the video topic
- soliciting votes or seeking badges
- a homework question
- a duplicate answer
- repeatedly making the same post
- a tip or thanks in Questions
- a question in Tips & Thanks
- an answer that should be its own question