My next math class doesn't start until next week, so I've been keeping myself busy by teaching myself a little statistics and probability. (I'm not sure why, since said math class is linear algebra.) Anywho, I ran into a bit of a pickle and I'm not sure what the problem is. So I figured the math nerds at Poly might be able to help.
I know that, for example, if a particular area usually gets rain once every ten days, then the odds of it not raining at all for a week are .9^7 = 48%, and thus the odds it rains at least once are 52%. I also know that if the odds aren't constant per day, I can just multiply repeatedly. So if Monday-Friday has a 10% chance of rain, and Saturday-Sunday has a 20% chance, I can do .9^5 * .8^2.
Then I was thinking, what if the length of time between days of rain follows a normal distribution? So if the mean length is 5 days and the standard deviation is 1 day, then I can plug those numbers into the normal probability density function and come up with the odds that the length of time between rains is x days. The odds it rains on any particular day are small, but I know that you can add probabilities of the normal function. So the odds that it rains after 5 days are going to sum to 50%.
But if I then take the individual probabilities for a given day (based on the normal function, mean, and standard deviation), subtract them from 1 and multiply them together as in the first section, I get a totally different answer. It comes out to a 43% chance that it doesn't rain after 5 days.
So, obviously, I'm approaching this problem from two different places and I wouldn't expect the math to come out the same, except that I feel like I'm asking the same question with both techniques. So why are the answers different?
As an aside, this isn't the actual problem I was doing. The problem I was doing is significantly more depressing but also involves a lot more terms. I kind of thought that having a lot of terms would smooth it out because the density function is continuous, but it actually diverges significantly more than my sample problem here does. So, what's the deal?
I know that, for example, if a particular area usually gets rain once every ten days, then the odds of it not raining at all for a week are .9^7 = 48%, and thus the odds it rains at least once are 52%. I also know that if the odds aren't constant per day, I can just multiply repeatedly. So if Monday-Friday has a 10% chance of rain, and Saturday-Sunday has a 20% chance, I can do .9^5 * .8^2.
Then I was thinking, what if the length of time between days of rain follows a normal distribution? So if the mean length is 5 days and the standard deviation is 1 day, then I can plug those numbers into the normal probability density function and come up with the odds that the length of time between rains is x days. The odds it rains on any particular day are small, but I know that you can add probabilities of the normal function. So the odds that it rains after 5 days are going to sum to 50%.
But if I then take the individual probabilities for a given day (based on the normal function, mean, and standard deviation), subtract them from 1 and multiply them together as in the first section, I get a totally different answer. It comes out to a 43% chance that it doesn't rain after 5 days.
So, obviously, I'm approaching this problem from two different places and I wouldn't expect the math to come out the same, except that I feel like I'm asking the same question with both techniques. So why are the answers different?
As an aside, this isn't the actual problem I was doing. The problem I was doing is significantly more depressing but also involves a lot more terms. I kind of thought that having a lot of terms would smooth it out because the density function is continuous, but it actually diverges significantly more than my sample problem here does. So, what's the deal?
Comment