Fitting of Probability Distributions
This is a modal window.
The media could not be loaded, either because the server or network failed or because the format is not supported.
Formal Metadata
Title |
| |
Title of Series | ||
Part Number | 12 | |
Number of Parts | 16 | |
Author | ||
License | CC Attribution - ShareAlike 3.0 Unported: You are free to use, adapt and copy, distribute and transmit the work or content in adapted or unchanged form for any legal and non-commercial purpose as long as the work is attributed to the author in the manner specified by the author or licensor and the work or content is shared also in adapted form only under the conditions of this | |
Identifiers | 10.5446/12882 (DOI) | |
Publisher | ||
Release Date | ||
Language |
Content Metadata
Subject Area | |
Genre |
1
2
3
4
5
6
7
9
10
11
12
13
14
15
16
00:00
MathematicsOpen setLimit of a functionLimit (category theory)AreaSign (mathematics)Term (mathematics)Lecture/Conference
03:20
Random variableLimit (category theory)Limit of a functionMereologyFunctional (mathematics)Right anglePower (physics)Term (mathematics)InfinityDot productMaxima and minimaProduct (business)Numerical analysisDistribution (mathematics)Natural numberIdentical particlesOrdnungsstatistikMultiplication signCumulative distribution functionSequenceSeries (mathematics)Rule of inferenceFactory (trading post)Uniformer RaumQuantum stateGroup actionLecture/Conference
10:49
Derivation (linguistics)Fraction (mathematics)Population densityVarianceMaß <Mathematik>Multiplication signTerm (mathematics)RootArithmetic meanSquare numberDistribution (mathematics)Maxima and minimaTotal S.A.InfinityOperator (mathematics)FrequencyCumulative distribution functionMereologyMany-sorted logicStandard deviationFunctional (mathematics)ExplosionMortality rateTransformation (genetics)Normal distributionLecture/Conference
18:19
RootSquare numberAreaNumerical analysisForestDivisorShift operatorSigma-algebraPower (physics)Lecture/Conference
24:30
Numerical analysisMaxima and minimaRandomizationFraction (mathematics)RootSquare numberLimit (category theory)Random variableMultiplication signExpressionMoment (mathematics)SequenceAreaFunction (mathematics)Standard deviationOrientation (vector space)RoutingState of matterRight angleLecture/Conference
30:30
Numerical analysisDistribution (mathematics)FrequencyMaxima and minimaRandom variableMortality rateArithmetic meanTrailState of matterDifferent (Kate Ryan album)Limit (category theory)Set theoryRootAreaStatisticsProbability theoryVarianceGeneralized extreme value distributionTheoremHeuristicDecision theoryGreatest elementReliefFree groupUniformer RaumMultiplication signLecture/Conference
33:35
StatisticsDegree (graph theory)Chi-squared distributionDegrees of freedom (physics and chemistry)Sign (mathematics)VarianceNormal (geometry)Multiplication signDistribution (mathematics)MetreIndependence (probability theory)Negative numberSquare numberDecision theoryLecture/Conference
37:23
Different (Kate Ryan album)SummierbarkeitMathematicsIndependence (probability theory)Differential (mechanical device)Numerical analysisChi-Quadrat-TestGamma functionMultiplication signComputabilitySign (mathematics)Square numberRandom variableSet theoryDegrees of freedom (physics and chemistry)Direction (geometry)Normal (geometry)Lecture/Conference
42:41
Set theoryLie groupObservational studyGrothendieck topologyPopulation densityPoint (geometry)Mortality rateChainRandom variableMultiplication signFunctional (mathematics)WahrscheinlichkeitsfunktionLine (geometry)Projective planeRootFrequencyLimit (category theory)Price indexSpacetimeInfinityProduct (business)Logical constantLinear regressionDifferential (mechanical device)Distribution (mathematics)Variable (mathematics)2 (number)RandomizationProbability density functionINTEGRALDerivation (linguistics)Lecture/Conference
49:16
Multiplication signComputer programmingInfinityLogical constantEnergy levelDifferent (Kate Ryan album)Spherical capChainPopulation densityAreaPrime idealPower (physics)Probability density functionGamma functionAnalytic continuationDerivation (linguistics)Right angleAlpha (investment)Square numberLecture/Conference
55:52
MereologyExponentiationMultiplication signAlpha (investment)Power (physics)DivisorFraction (mathematics)Population densityLine (geometry)Right angleGamma functionTerm (mathematics)Potenz <Mathematik>Equaliser (mathematics)Probability distributionDirection (geometry)Inequality (mathematics)Functional (mathematics)Distribution (mathematics)Independence (probability theory)Open set1 (number)State of matterArithmetic progressionLecture/Conference
01:02:03
Gamma functionDecision theoryDistribution (mathematics)Population densityGrothendieck topologyStudent's t-testSampling (statistics)Moving averageRoutingGraph coloringSpacetimeProduct (business)RootFunctional (mathematics)INTEGRALStudent's t-distributionObservational studyStatisticsLecture/Conference
01:05:10
Maß <Mathematik>Moment (mathematics)Vector spaceDivisorFunctional (mathematics)EstimatorParameter (computer programming)Sampling (statistics)Square numberCartesian coordinate systemProduct (business)Variable (mathematics)VarianceRankingMortality rateIndependence (probability theory)Category of beingUniqueness quantificationMetreMultilaterationPopulation densityRule of inferenceExpected valueMereologyMultiplication signRandom variableSigma-algebraWater vaporCentral limit theoremStatisticsNormal distributionSummierbarkeitBinomial distributionRandomizationAverageDistribution (mathematics)Arithmetic meanProof theoryLecture/Conference
01:13:14
Independence (probability theory)Multiplication signCoefficientWater vaporRight angleDiagonalTerm (mathematics)FrequencyExpression1 (number)Moment (mathematics)SummierbarkeitExponentiationLecture/Conference
01:21:18
Military baseModulformSquare numberAxiom of choiceNormal (geometry)Vector spaceIdentical particlesMultiplication signMortality rateFunctional (mathematics)Selectivity (electronic)Order (biology)Arithmetic meanRule of inferenceSummierbarkeitVarianceRandom variableTerm (mathematics)Independence (probability theory)Normal distributionRandomizationExponentiationMereologyMoment (mathematics)Lecture/Conference
01:29:23
Axiom of choiceVector spaceResultantFunctional (mathematics)Square numberMereologyDifferent (Kate Ryan album)Musical ensembleSampling (statistics)ModulformNormal (geometry)Independence (probability theory)Sparse matrixDivisorRandom variableArithmetic meanSummierbarkeitMoment (mathematics)VariancePotenz <Mathematik>Right angleProduct (business)Term (mathematics)Lecture/Conference
Transcript: English(auto-generated)
00:05
Okay, there's one exercise that was a little more difficult than the rest. Maybe I'll go through that. I don't think I assigned this one.
00:46
Did I assign that? Yes? Okay, let's do that one.
01:23
Okay, that the limit of a n is a, says this. Given epsilon bigger than zero, there is a capital N, so that a n is between a minus epsilon and a plus epsilon for all little n bigger than or equal to capital N. Do you agree with that statement?
01:40
That's what it means to have limit a. Alright? So because of that, this is the limit we want to treat. There's an a n in there, so let's replace a n on one side by a plus epsilon, the other side by a plus epsilon.
02:24
Do you agree with that? Okay, and what does this side do as n goes to infinity?
02:42
This converges to something, this one, e to the a minus epsilon. And this converges to e to the a plus epsilon.
03:09
Okay?
03:24
So, maybe you remember, do you know the terms lim inf and lim sup? No? Okay. Well, if this limit exists, what does it have to be between?
03:48
Right. It has to be between this and that, right? So the limit, let's assume the limit exists.
04:10
And this is true for every epsilon bigger than zero. No matter what epsilon I pick, this holds in the limit.
04:29
And what can I do with epsilon now? I can let epsilon go to zero, because the two n's squeeze down to e to the a.
05:01
Right? And any question on this? Okay, technically speaking, I should say, for those who do know this language, this holds.
05:52
And then you let epsilon go to zero, and then you conclude the lim inf is equal to the lim sup, and that means the limit exists and is their common value, which must be e to the a.
06:01
So for more rigorous treatment, if you've had these terms, I think you might have had them in calculus, when you're talking about limits of sequences. But anyway, I think the idea is clear from here. And then this was supposed to be used in another exercise, on the extreme order statistic for uniformly distributed random variables.
06:30
And I forget which number that was. 29? Okay, number 29.
06:58
So we take u1 through un to be iid uniformly distributed on the interval from zero to one.
07:05
And then, after you take your sample, you call the largest one that you see,
07:24
u sub parenthesis n. Okay, that's the max of the, that's the largest observation. And the exercise was to find the CDF, first to find the CDF for this. So let's find the CDF for that.
07:46
That was the easy part of the exercise. Does everybody know by now what this stands for? Okay, cumulative distribution function.
08:01
Well, the CDF for any random variable is this, if you put the random variable here, it's the probability that the random variable is less than or equal to x.
08:22
It's a function of x. It's this function of x we're interested in. And this says what? Un less than or equal to x. Un is the largest of all the observations. So if the largest is less than or equal to x, well, the largest is less than or equal to x if and only if all of them are less than or equal to x.
08:54
And these are independent, so this becomes, since these commas mean n or intersection,
09:01
this becomes a product of the probabilities of u1 less than or equal to x, of u2 less than or equal to x, et cetera. And I could put probably u2 less than or equal to x here, and then a bunch of dots and probably un is less than or equal to x there. But what does this I mean in IID?
09:25
Identically distributed. They have the same distribution function. So that you'd repeat the same thing over and over again, but how many times do you repeat it? N times. So we get that to the nth power. And what is this?
09:40
Well, depends on x. If x is bigger than 1, it's 1. Why is that?
10:01
What's the CDF for a uniformly distributed random variable on an interval from 0 to 1 look like? It's that, right? 0 until you get to 0, and it's the function x, f of x equal x, and then it's 1.
10:24
Why is it 1? Because if it's a uniformly distributed random variable on the interval from 0 to 1, the random variable has to have the value less than or equal to 1. So the probability that it's less than or equal to a number bigger than 1 is 1. If x is less than 0, then this probability is 0.
10:44
All the numbers are between 0 and 1, so they can't be less than 0. The maximum can't be less than 0. Otherwise, if x is between, then this thing is x, and we raise it to nth power.
11:25
Okay, let's find the expected value of un.
11:52
I'll put 0 in there because I know the density will vanish off the interval from 0 to 1.
12:02
Why? What's the density? Here's the CDF. The density is the derivative of this. What's the derivative of the constant function? 0. What's the derivative of this constant function? 0. So the PDF is 0, but for values of x less than 0 and for values of x bigger than 1.
12:24
You should have here the integral from minus infinity to infinity x times the density of un, but the density vanishes off the unit interval, so I don't have to go from 0 to 1. And what's the density between 0 and 1? It's the derivative of this. That would be n times x to the n minus 1.
12:51
And so inside here we get x to the n. The integral of x to the n is x to the n plus 1 over n plus 1, and we evaluate between 0 and 1.
13:04
The value at 1 is 1 over n plus 1 times n, and the value at 0 is 0. So there's the expected value of the max. Now let's find the variance of the max.
13:28
Well, that would be the expected value of the maximum squared minus the square of the expected value. We just computed this. It's here.
13:43
How do we compute this? Well, here we're computing the expected value if we put x there. What will we put if we're doing squared? We put x squared there.
14:08
OK, and this becomes integral from 0 to 1 n times x to what power? From here we get x squared. From here we get x to the n minus 1.
14:21
So that would be an x to the n plus 1 altogether? OK, so here's the CDF, and then the PDF is the derivative of that.
14:58
So we get x to the n plus 1 when we multiply these two together.
15:09
And then minus n divided by n plus 1 squared. And this becomes n over n plus 2 minus n over n plus 1 squared.
15:32
And let's combine this. We have to find common denominator. Common denominator would be n plus 2, n plus 1 squared.
15:45
We have to multiply this by this denominator. And we have to multiply n squared by this denominator.
16:13
And what does this become?
16:22
When I multiply the first term out here, I get n squared plus 2n plus 1. I multiply that by n, so I get n cubed plus 2n squared plus n.
16:41
And here I get minus n cubed minus 2n squared. For a grand total of what? Just n is left.
17:07
So there's the variance of un.
17:33
OK, so far so good? Now, the exercise says something about the distribution of the standardized max.
17:48
What's the standard normal distribution? Have we used that term? Has the book used that term? What do you think the standard normal would be? Mean 0, variance 1. So a standardized random variable means you somehow transform it so it has mean 0 and variance 1.
18:04
How can you transform a random variable so it has mean 0? Subtract the mean. So we'd subtract n over n plus 1. Now, how can you make a random variable have variance 1? Yeah, exactly. Divide by square root of the variance.
18:20
OK, so we'll take un, u parenthesis n, subtract that and divide by the square root of this.
18:41
Try to compute this. Less than or equal to x.
19:25
For simplicity, let me call this sigma n squared.
20:25
Well, if this number is between 0 and 1, we know what this is. It's what? It's that number to the power n. So if, and if it's bigger than 0, say.
21:15
Now let's put back in this for sigma n squared.
21:24
This should not be sigma n squared here, sorry. It should be square root of that, right? It's the square root of sigma n squared. So I better put square roots here. I do that because I can erase the 2 on the board, but I don't think you can erase the 2 in your notes.
22:47
Is that right now? OK, I'll start again. OK, so I should have square root of n over n plus 2,
23:06
and then x over n plus 1, plus n over n plus 1.
24:08
Can you see that? So the way to get from here to here is add a 1 here in the numerator and subtract a 1.
24:22
If I add a 1, n plus 1 over n plus 1 is 1. If I subtract a 1, that gives me minus 1 over n plus 1, which I put over here. OK? Now what was the exercise we did just a moment ago?
24:45
It was if a n goes to a, then 1 plus a n over n to the n goes to e to the a.
25:11
We almost have that here. This would be the a n. What does that converge to? Does that numerator converge to anything? What does this ratio n over n plus 2 do as n goes to infinity?
25:24
It goes to 1. So the numerator here goes to x minus 1. This numerator goes to x minus 1. And the denominator, is that n?
25:41
Not quite. It's a quibble that it's not, but we can fix that. Let's replace this with n. But we want to be honest. We want to keep the same quantity we had before. So I'll put an n here, and then in the numerator I'll put n over n plus 1.
26:03
So I'm going to rewrite this as, first I'm going to put the 1 because over there I put 1 first. Over n to the n. And then I'll put n over n plus 1 here times square root n over n plus 2 x minus 1.
26:29
And that's equal to this. See this n plus 1? That's in the denominator here, right? So that means it's the same as being down here. And when I have an n up here and an n down there, what do they do?
26:42
They cancel. So that gives that thing. Now, this is the a n. And what does that do as n goes to infinity? n over n plus 1 goes to 1.
27:01
n over n plus 2 goes to 1. So this whole numerator goes to x minus 1. So what does this expression go to then? What does this probability converge to? This converges to e to the x minus 1. So long as this is true, this should hold in the limit.
27:25
Let's see what this says. Mu n is n over n plus 1, so this should say that x is less than or equal to 1 minus mu n over sigma n.
27:41
And what does this do as n goes to infinity? 1 minus mu n is, what's 1 minus mu n? That's converging to 1, right? So 1 minus mu n is converging to 0. In fact, 1 minus mu n would be n plus 1 over n plus 1 minus n over n plus 1 is 1 over n plus 1, right?
28:37
This n plus 1 and this one down here will cancel.
28:42
It's n plus 1 squared, leaving n over n plus 2 square root. And what does this do as n goes to infinity? It goes to 1. So this is good as long as x is less than or equal to 1.
29:00
And the CDF for x bigger than that would be 1. So limiting CDF for the standardized max, what does this do when x is equal to 1?
29:54
e to the 0 or 1. So this CDF looks like what?
30:02
When x goes to minus infinity, this goes to 0. It dies off exponentially. It goes up to 1 at 1 and then it's flat. Turns out, if you ever go into insurance, this kind of random variable comes up a lot, the maximum.
30:28
I think I mentioned earlier why would an insurance company be interested in the maximum of a sequence of random variables. Well, the claims they receive on their policies are random variables. And the maximum would be the largest payout they have to make.
30:44
And so they want to charge enough money so that they can cover the largest payout they have to make. So they want to know the distribution of the largest payoff. n is the number of claims they've received. So they probably are thinking about a large number of claims over a period of years.
31:04
So they want to know how this behaves. So what they do is they have some distribution for the claims. It may not be uniform. It could be something else. And they subtract the mean and they divide by the square root of the variance. And it turns out that this thing always has a limit and there are three kinds of limits it can have.
31:26
There are three different limiting CDFs you can get depending on what distribution you start with. These are called the extreme value distributions. This is just one example of one of the extreme value distributions. You could start with another underlying distribution instead of uniform.
31:42
You would get a different one. But in the limit, there are only three possible limits that you can get. So they're very interested in the distribution of the largest claim that they can get. And it turns out that in the limit, there are only a couple different kinds of limits you can get. This is a very famous theorem in something called extreme value statistics.
32:06
And this is all in the area of risk management. People are risk averse. Some people are risk averse. How many of you think you're risk averse? How many of you go on the roller coaster ride at the Grey Ghost at Knott's Berry Farm?
32:27
Anybody go on that? So you're not risk averse. Or perhaps you invest your money in unwise companies. Or perhaps you go to Las Vegas and you're going to win.
32:43
But then there are some people who are very risk averse and they want to get rid of their risk. They sell it to somebody else and there's a whole industry. And this whole industry of risk management, or helping people who are risk averse, uses a lot of probability theory.
33:03
Okay, so that's it for those exercises. This is probably harder than anybody. Probably nobody was able to get this, but that's okay. So there will be a new set of exercises posted today.
33:22
They'll be due Friday. It'll be on chapters 6 and 7. And I think that'll be the end of the course. I don't remember what the syllabus said we're going to do. But I think we'll finish 6 and 7. Okay, so last time we discussed chi-squared distributions.
33:58
If z is normal with mean 0 and variance 1, then I'll define it to be u.
34:07
u equals z-squared is called a chi-squared with one degree of freedom.
34:30
If z1, z2, zn are independent, normal 0, 1 random variables,
34:43
then u sub n, I'll call it, equals z1-squared plus z2-squared plus zn-squared is chi-squared with n degrees of freedom.
35:04
These come up in a lot of statistical analyses.
35:21
Some other common distributions arising from the normal are, if m and n are non-negative integers,
36:26
and u is chi-squared with m degrees of freedom, v is chi-squared with n degrees of freedom, and u and v are independent, that's called the fmn distribution.
37:16
Ratios are useful statistics to analyze data.
37:39
Now last time we did mention that the chi-squared is actually a gamma.
38:08
We believe it's gamma of n over 2 and 1 half. That means this is the alpha, that's a lambda.
39:36
I haven't done a lot of computations where we compute PDFs or CDFs for ratios,
39:42
so let's do that now with this one.
40:09
So that would be the CDF differentiated, so we want to differentiate this thing.
40:26
So that means we should try to figure out what this is. Rather than write the derivative every time, I'm just going to compute this first and then I'll differentiate afterwards.
41:07
Let's multiply through by v over n. Does that change this sign at all? What were u and v? Chi-squares with m and n degrees of freedom respectively. What were chi-squares?
41:21
Sums of squares of normals. So chi-squared random variables are always non-negative or positive. So when I multiply through by v over n, I don't change the direction here at all.
41:44
And now I'm going to multiply through by little m. I'm going to write this m over n v x.
42:14
Using the independence of u and v, I could get to this.
42:26
Maybe I should take a couple steps here.
42:40
This is saying that the pair u, v is in a certain set. The pair has to lie below the line in u, v space that has slope m over n x.
43:10
That's this line, right? u less than or equal to that means... Well, maybe I have the wrong one. u less than or equal to that means I guess we're above the line.
43:21
The first variable has to be smaller than this multiple and the second. So the probability that u is less than or equal to m over n x, v would be the probability that the pair u, v is in that set A.
43:44
Maybe I'll index it by m over n x. So this is the set A, m over n x. Is that too quick? At this set, well, if the pair u, v is up here,
44:06
is it true that u is less than or equal to m over n v x?
44:21
Here's the fixed value of v. If I go along this line here, horizontal line with constant y coordinate v, when I start out here, is u bigger than m over n x, v or smaller? When u is 0, it's smaller and it's smaller all the way up until I get to that line there.
44:44
So this is the right region. And to compute a probability like this, you integrate the joint density over that region.
45:06
But they're independent.
45:28
So the joint density is a product of the individual densities. And now what we should do is set limits of regression to describe where we're integrating.
45:55
So our v can go from 0 to infinity.
46:04
And for a fixed value of v, u goes from here to m over n x, v. So this is the integral, v going from 0 to infinity, u going from 0 to m over n x, v,
46:23
f sub capital U of little u, f sub capital V of little v, du is the integral, dv is the outer one. Now let me kick this guy out.
46:49
And what is this integral? It's the integral of the density for capital U from 0 to something. What's that called? Three letters.
47:00
Last letter is f. CDF. That's the CDF for U evaluated at the upper limit. So I'm going to write that first. This is the CDF for U, so that's the probability that U is less than or equal to that value.
47:27
Okay? So that's what I wrote here. So always remember this.
47:40
If you want to compute the probability of something about a pair of random variables, you need their joint density or joint distribution or joint probability mass function. And you express whatever this restriction is as a set and you integrate the CDF over that set
48:00
or you sum the probability mass function over that set. So here that set was this. So this should be automatic every time. Alright, so we got to this point. And we wanted to differentiate.
48:21
So let's differentiate that expression. Well, here's the only place X appears, so we differentiate this.
48:52
This is the CDF for U, so when we differentiate, we should get the PDF. But we have to use the chain rule.
49:01
We're differentiating with respect to X, so we'll get this thing outside. So this will become integral from 0 to infinity. This PDF at M over NVX, because that's the derivative of this function of that variable. But then we have to multiply by the derivative of this with respect to X, which would be M over NV,
49:25
and then we have the PDF for V. Okay? So that's the chain rule. Right? This is capital F sub U at M over NVX.
49:43
So if we want to differentiate that with respect to X, we get capital F U prime at M over NVX times the derivative of what appeared inside with respect to X, that's M over NV. Okay? And this is the PDF now. This is little f.
50:03
Okay? And then we have to integrate this from 0 to infinity. But you won't have to do any integration. Let's substitute what the PDFs are now. I wrote them on the board.
50:23
They're going to be constants out in front. First one, I guess, will be M over N from here. And then we take the, evaluate the PDF for U at M over NVX.
50:47
Now U has these constants out in front, so let's put those out in front.
51:01
And then V will have one half to the N over gamma of a half. And then we take the value of this at, we raise that to the M over 2 minus 1.
51:29
And then we raise V to the N over 2 minus 1.
51:49
I promise you, it's not as bad as it looks. Okay. Now what in here doesn't get, doesn't have a V?
52:02
Well, this doesn't have a V. Let's pull that out, and the X doesn't have a V. Let's pull that out. Okay?
52:20
Now out in front, we have an M over N. Well, let's see. Let's take care of the halves first. We have a half to the, whoops, I forgot, N over 2 here. We have a half to the M plus N over 2. We have a gamma of one half squared.
52:46
Here's an M over N. But we also have an M over N to the M over 2 minus 1. Yeah?
53:03
Oh, you're right. Let's see. No? Oh, it's here. You're right. I didn't put it, I left out the V. Thanks. I left out the V. Yeah. I forgot to put this V down here.
53:27
Okay. So, M over N to the M over 2 minus 1. And here's an M over N to the first power, so we'll get altogether M over N to the M over 2.
53:52
And what's the, so I think that's all of the constants. Oh, well, I forgot this one. X to the M over 2 minus 1.
54:13
N times integral, zero to infinity. V to some power. What power?
54:20
Here's the first power. Here's to the M over 2 minus 1. Here's to the N over 2 minus 1. I add all that up. 1 plus M over 2 minus 1 plus M over 2 minus 1. That is V to the M plus N over 2 minus 1.
54:48
And then E to the minus V.
55:15
Okay. So, it's late. Let's take a break and come back in 10 minutes and finish it.
55:24
Okay. So, let's continue. But amend your notes. I've put gamma of a half here. It should be gamma of M over 2. Remember the gamma density is lambda to the alpha over gamma of alpha. X to the alpha minus 1.
55:40
E to the minus lambda X. So, that should be gamma of M over 2 here. Gamma of N over 2 there. So, that shows up here. It should have gamma of M over 2. Gamma of N over 2. And when I was evaluating the density for U at this,
56:02
I did put that in M over NVX to the M over 2 minus 1 for the power. But I forgot to do that in the exponential part. So, this should not be. This should be M over NVX here.
56:28
I forgot to put the, I put V in for the exponential part. So, I should be evaluating this at M over NVX.
56:40
So, that should go in here. It should also go in the exponent. I forgot to put it in the exponent. This is the density for U. Not to evaluate it here. I did it correctly there, but I forgot to put it in the exponent. So, that'll show up when I do the exponents.
57:10
I have an E to the minus V over 2 and then this. So, if I add the exponents, I'll add minus V over 2 M over NVX and minus V over 2.
57:23
I can factor out the minus V over 2. I get 1 plus M over NVX times V over 2. Or, maybe I'll write it this way, minus 1 plus M over NVX over 2 times V.
57:44
Okay? That's what happens when I combine these two exponentials. So, once again, I forgot to put M over NVX in the exponent part of the density for U. I put it in the power part, but you also have to put it in the exponent part.
58:02
And correct the densities for U and V to have a gamma of M over 2, not a gamma of a half. It's M over 2 here. And for V, N over 2. Gamma of N over 2. Okay? And that brings us to this here. And the last part I have to carry over from there is E to this power.
58:29
DV. Okay? So, how do we do this integral? We use the fact that this is a density, right?
58:51
And in that, what is the alpha? Alpha is M plus N over 2. And lambda is one half 1 plus M over NVX.
59:12
You compare. Alpha is M plus N over 2. Lambda is half times quantity 1 plus M over NVX.
59:32
And so, this becomes, let me put this down here.
01:00:00
and here I'd get lambda to the alpha in the denominator that'd be 1 half 1 plus m over n x raised to the power alpha right which is m plus n over 2
01:00:22
this is the lambda to the alpha another way of saying this here is that integral from 0 to infinity x to the alpha minus 1 e to the minus lambda x dx is gamma of alpha
01:00:40
over lambda to the alpha so I've just put the lambda to the alpha here with lambda equal half times the quantity 1 plus m over n x alpha equal m plus n over 2 and up here should go gamma of alpha well you can
01:01:08
combine the x terms but that's the density so there are a couple of steps, key steps one is if you're computing the probability
01:01:20
distribution function for a ratio you write that if the denominator is always a non-negative random variable you can multiply through by that and not change the direction of inequality and then you have to express this as the probability that the pair uv lies in some region
01:01:41
in this case would be above a line and then you integrate the joint density for the pair over the region above the line and if they're independent the joint density factors and the rest was just collecting terms busy work and then using the fact that you know
01:02:03
that the gamma density is a density and then you don't have to do any integration it's kind of ugly I have to admit it doesn't look very nice okay so another
01:02:28
distribution that comes up is called the t sometimes referred to as the student t distribution this is called the student t distribution
01:03:40
sometimes a student doesn't refer to people like yourselves it refers to
01:03:47
pen name of a statistician who worked for Guinness and he had to work with small samples and to analyze data from small samples he came up with this
01:04:02
statistic here we'll study this later I'll give it to you as a challenge to compute the density for this that would follow the same steps we've just done how do you compute the density you can write down the distribution function
01:04:22
what would that look like you'd have a what instead of this ratio you'd have a z here and a root u over n here you could multiply through by the root u over n so you'd have z less than or equal x root u over n this is the pair then that means the pair
01:04:42
z u must lie in a certain region in space write down the integral over that region of the joint density what's the joint density going to be for z and u it's going to be a product of the individual densities and then hope for some miracle
01:05:00
and then there will be one but i won't do it the density is written down in the book but see if you can derive it yourselves
01:05:34
okay so let's consider the following suppose x1
01:06:00
through xn are independent random variables well for now i don't need independent the sample mean is just this
01:06:29
they're average and the sample variance i always forget yeah right i want to make sure it's one over n minus one
01:06:54
here n minus one goes here the reason for
01:07:20
that is if the if the x's are
01:07:44
independent or let's say iid with variance common variance equal to sigma squared then the expected value of s squared is sigma squared this is supposed to be
01:08:10
an estimate for sigma squared if i have a one over n here i wouldn't get sigma squared there
01:08:21
i'd get n minus one over n times sigma squared if s squared is supposed to estimate some parameter namely sigma squared if the expected value of the estimator is equal to the true value then we call that an unbiased statistic so we normalize by one over n minus one here
01:08:43
to get an unbiased statistic up here um what would the expected value of x bar equal if we're in this situation again if these are identically distributed then the expected value of x bar is mu
01:09:18
why is that well what's the expected value of a sum of random variables
01:09:22
it's the sum of the expected values so what do we get here in the sum we get n times mu but then divide by n you get mu okay so this is supposed to be an estimator for mu and its expected value when you take the expected expectation of the randomness you get mu so that's called an unbiased estimator
01:09:42
its expected value is equal to the parameter it's supposed to be estimating so let's examine the case where we have normal random variables suppose these are
01:10:26
normal mean zero variance one random variables which is often the case because of the central limit theorem many samples and applications can be assumed to be
01:10:41
normal random variables also we have the fact that Poisson distribution when properly centered and scaled is close to a normal random variable and binomial random variables when we subtract the mean divide by the square to the variance that's going to be close to
01:11:00
normal so a lot of the samples you get are assumed to be normal random variables in that case x bar is independent of the random vector x1
01:11:20
minus x bar x2 minus x bar x and minus x bar it's kind of strange x bar appears in both places
01:11:44
yet they're independent so one way to prove independence and independence is look at the joint density in short factors another way is to look at the moment generating function
01:12:04
and show that it factors into a product if it does remember then use the uniqueness property of moment generating functions a moment generating function can only come from one distribution so if it comes from a distribution where the x bar part factors away from this part then you
01:12:24
know they're independent so let's look at the moment generating function for this for the vector value case we have n plus one random variables here so we need n plus one real variables and here's what we do we look at e to the t naught x bar plus t1 x1 minus x bar
01:12:45
plus t2 x2 minus x bar plus etc t n x n minus x bar and what we're going to do is we'll show that this factors into and that will be enough to show
01:13:34
independence okay okay let's recall this fact and
01:14:22
let's start with the left hand side okay and what's x bar what do we get
01:15:22
here what's the coefficient of um there's
01:15:42
there's there are a bunch of x ones in here right what's the coefficient of x1 in this expression in the exponent what do we get for the coefficient of x1 from here well x1 is the first term in the sum it's multiplied by
01:16:00
one over n and then we multiply also by t naught over n right so maybe i'll just work with the exponent for a moment or two here i got t naught over n
01:16:42
um here i won't get a what do i get here for the coefficient of x1 is that right and what do i get all along here the next term would be
01:17:04
t2 x2 minus x bar so i'll get a minus t2 over n times x1 from here then minus t3 over n etc and then here i get a minus tn over n all times x1
01:17:26
and what do i get for the coefficient of x2 from here i get a t naught over n from here i get a minus t1 over n from here i'll get a t2 times 1 minus 1 over n
01:17:44
then i get a minus t3 over n etc all the way up to minus t sub n over n times x2 and i keep going like that until i get down to the nth one okay
01:18:18
on the diagonal i get t1 times 1 minus 1 over n then t2 times 1 minus
01:18:25
1 over n and all the way down to here right let's recombine
01:18:48
there are terms with just like t1 t2 etc right so i get t sum i equal 1 to n ti times xi
01:19:07
well let's see what do i get times everyone has a t naught over n right
01:19:22
every one of them has a t0 over n every one of them has a that for x1 i get a t1 for x2 i get a t2 for x3 i get a t3 for x and i get a tn so i get a plus ti and then i get a minus
01:19:49
t sub i over n minus uh what some of you some j equal 1 to n tj over n i'm going to write that this way 1
01:20:06
over n sum j equal 1 to n tj
01:20:28
so let's do this one first if i have t naught over n times xi and i add up i going from 1 to n what do i get i get it i can follow up the t naught right
01:20:41
and then i get a 1 over n times the sum of the xi's that would be x plus the sum i equal 1 to n ti xi and then minus
01:21:03
what um this thing i could call t bar so we could rewrite it this way this is
01:22:04
not x bar is it but it's n times x bar so i combine that with this but uh actually that's not a that's not as
01:22:41
useful as i i would like i think i'll just stick with this let's just stop here okay we'll use this one okay so you don't need these what i wrote is right but we'll stop here okay now let's decide what kind of random variable is this we have a sum of independent normal
01:23:06
random variables what kind of random variable is t naught over n plus ti minus t bar xi that's a normal random variable what's its mean zero and its variance is
01:23:31
t naught over n plus ti minus t bar squared
01:23:46
if i have a normal random variable x is normal if x is normal means zero and variance one what kind of random variable is a times x what's the expected value of this zero still what's the expected value what's the variance the variance would just be the expected value of this squared
01:24:02
that would be a squared okay so we just get this squared so what kind of random variable would this whole sum be here they're independent normals the sum of independent normals is normal this is a normal random variable what's the expected value of
01:24:31
this each sum and as i take the value zero so the whole sum would have expected value zero and what about the variance of this
01:24:41
the variance of the sum is the sum of the variances for independent random variables so the variance would be the sum of these things now let's examine what's what this
01:25:22
variance is if you have a a plus b squared it's a squared plus
01:26:08
2 ab plus b squared so i just did that now what's the first term here all of these are the same so i just multiply this by n
01:26:23
that would be t naught squared over n what about this when i add up this what do i get when i add up the ti's i get
01:26:46
n times t bar when i add up these what do i get well they're all the same i get n times this term
01:27:03
but i'm subtracting this from this so what does this what does this become zero this part is zero okay so this random variable up there is
01:27:23
normal with this variance and mean zero okay so and that's what goes in this exponent this random variable is normal with mean zero and variance that so this is equal to
01:27:47
e to the mu is zero sigma squared is that thing over there and so
01:28:00
so this is so i have one here one times this random variable this is this is my this is now my x i'm taking t equal one so i get just the half the variance t naught squared over n or 2n and then times e to the
01:28:26
one half sum i equal one to n ti minus t bar squared so i find that the expected value of this thing the moment
01:28:48
generating function for this complicated vector form is this what is the expected value of e to the t naught x bar can i get that from this
01:29:13
is t naught x bar equal to what i see there for some choice of t1 through tn
01:29:24
yeah what choice take all the t i's to be zero except t zero then this will be zero that'll be zero that'll be zero what do you get here then zero so this would be
01:29:41
okay and what is this part that's if i take if i take t naught to be zero i that's what i get here and what if i take t naught to be zero i
01:30:02
just get this so what's the moral of the story the moment generating function for this is the product of this moment generating function in that one so what does that mean it means this random variable x bar and the vector x1 minus x bar through
01:30:25
x n minus x bar are independent okay and the miracle was right here
01:30:54
the moment generating function for normal looks like this it has this particular form so when we're
01:31:01
computing the variance what happened the t naught part separated from the ti minus t bar part and we're taking exponential of this so that means it factors that's that's why the factoring took place because this term cancelled
01:31:21
okay so x bar is independent of this when you have independent normal samples all right so this implies a very important result x bar and s squared are independent
01:32:20
s squared remember was one over n minus one
01:32:23
sum i equal one to n x i minus x bar squared this is a function of you know use g of x1 minus x bar x2 minus x bar x n minus x bar right
01:32:45
it's a it depends on what these differences s squared is a function of these differences x bar is independent of this vector this is a function of the vector that's independent of x bar so a function of
01:33:04
if you have two random variables y and x are independent that implies y and g of x are independent okay so we're just taking a function of these things and they're independent of x bar
01:33:21
so so when you're dealing with sample means and sample variances they are independent random variables if your sample is normal it's not true if it's not normal okay so we can stop there today have a good weekend
Recommendations
Series of 1 medium