probability theory presentation 14
TRANSCRIPT
-
8/8/2019 Probability Theory Presentation 14
1/26
BST 401 Probability Theory
Xing Qiu Ha Youn Lee
Department of Biostatistics and Computational BiologyUniversity of Rochester
October 15, 2009
Qiu, Lee BST 401
http://find/http://goback/ -
8/8/2019 Probability Theory Presentation 14
2/26
Outline
Qiu, Lee BST 401
http://find/http://goback/ -
8/8/2019 Probability Theory Presentation 14
3/26
Expectation (I)
If Y = (X), then EY =(x)dP.
Special case: (X) = Xk. EXk is called the kth moment ofX.
E|X|k: the kth absolute momentof X (remember the Lp
norm?).
E(X EX)k: the kth central momentof X.
E|X EX|k: the kth absolute central momentof X.
The second central moment (which is also the second
absolute central moment) of X is called the varianceof X.
Denote m = EX.var(X) = E(X m)2 = E(X2 2mX + m2) = E(X2) m2.
Qiu, Lee BST 401
http://find/http://goback/ -
8/8/2019 Probability Theory Presentation 14
4/26
Expectation (I)
If Y = (X), then EY =(x)dP.
Special case: (X) = Xk. EXk is called the kth moment ofX.
E|X|k: the kth absolute momentof X (remember the Lp
norm?).
E(X EX)k: the kth central momentof X.
E|X EX|k: the kth absolute central momentof X.
The second central moment (which is also the second
absolute central moment) of X is called the varianceof X.
Denote m = EX.var(X) = E(X m)2 = E(X2 2mX + m2) = E(X2) m2.
Qiu, Lee BST 401
http://find/http://goback/ -
8/8/2019 Probability Theory Presentation 14
5/26
Expectation (I)
If Y = (X), then EY =(x)dP.
Special case: (X) = Xk. EXk is called the kth moment ofX.
E|X|k: the kth absolute momentof X (remember the Lp
norm?).
E(X EX)k: the kth central momentof X.
E|X EX|k: the kth absolute central momentof X.
The second central moment (which is also the second
absolute central moment) of X is called the varianceof X.
Denote m = EX.var(X) = E(X m)2 = E(X2 2mX + m2) = E(X2) m2.
Qiu, Lee BST 401
http://find/http://goback/ -
8/8/2019 Probability Theory Presentation 14
6/26
Expectation (I)
If Y = (X), then EY =(x)dP.
Special case: (X) = Xk. EXk is called the kth moment ofX.
E|X|k: the kth absolute momentof X (remember the Lp
norm?).
E(X EX)k: the kth central momentof X.
E|X EX|k: the kth absolute central momentof X.
The second central moment (which is also the second
absolute central moment) of X is called the varianceof X.
Denote m = EX.var(X) = E(X m)2 = E(X2 2mX + m2) = E(X2) m2.
Qiu, Lee BST 401
http://find/http://goback/ -
8/8/2019 Probability Theory Presentation 14
7/26
Expectation (I)
If Y = (X), then EY =(x)dP.
Special case: (X) = Xk. EXk is called the kth moment ofX.
E|X|k: the kth absolute momentof X (remember the Lp
norm?).
E(X EX)k: the kth central momentof X.
E|X EX|k: the kth absolute central momentof X.
The second central moment (which is also the second
absolute central moment) of X is called the varianceof X.
Denote m = EX.var(X) = E(X m)2 = E(X2 2mX + m2) = E(X2) m2.
Qiu, Lee BST 401
http://find/http://goback/ -
8/8/2019 Probability Theory Presentation 14
8/26
Expectation (I)
If Y = (X), then EY =(x)dP.
Special case: (X) = Xk. EXk is called the kth moment ofX.
E|X|k: the kth absolute momentof X (remember the Lp
norm?).
E(X EX)k: the kth central momentof X.
E|X EX|k: the kth absolute central momentof X.
The second central moment (which is also the second
absolute central moment) of X is called the varianceof X.
Denote m = EX.var(X) = E(X m)2 = E(X2 2mX + m2) = E(X2) m2.
Qiu, Lee BST 401
http://find/http://goback/ -
8/8/2019 Probability Theory Presentation 14
9/26
Expectation (I)
If Y = (X), then EY =(x)dP.
Special case: (X) = Xk. EXk is called the kth moment ofX.
E|X|k: the kth absolute momentof X (remember the Lp
norm?).
E(X EX)k: the kth central momentof X.
E|X EX|k: the kth absolute central momentof X.
The second central moment (which is also the second
absolute central moment) of X is called the varianceof X.
Denote m = EX.var(X) = E(X m)2 = E(X2 2mX + m2) = E(X2) m2.
Qiu, Lee BST 401
http://find/http://goback/ -
8/8/2019 Probability Theory Presentation 14
10/26
Expectation (II)
The existing of a higher order moment implies the existing
of a lower order moment.
A special case of Chebyshevs inequality: (m = EX,=STD)
P(|X m| > k) 1
k2
Qiu, Lee BST 401
http://find/http://goback/ -
8/8/2019 Probability Theory Presentation 14
11/26
Expectation (II)
The existing of a higher order moment implies the existing
of a lower order moment.
A special case of Chebyshevs inequality: (m = EX,=STD)
P(|X m| > k) 1
k2
Qiu, Lee BST 401
http://find/http://goback/ -
8/8/2019 Probability Theory Presentation 14
12/26
Covariance/correlation
cov(X, Y) = E(X EX)(Y EY).corr(X, Y) = cov(X, Y)/
X
Y.
Why correlation is always in [1, 1]? Cauchy-Schwarz.
Variance of summation.
var(X1 + X2 + . . .) =
varXi + 2
i,j=1;i
-
8/8/2019 Probability Theory Presentation 14
13/26
Covariance/correlation
cov(X, Y) = E(X EX)(Y EY).corr(X, Y) = cov(X, Y)/
X
Y.
Why correlation is always in [1, 1]? Cauchy-Schwarz.
Variance of summation.
var(X1 + X2 + . . .) =
varXi + 2
i,j=1;i
-
8/8/2019 Probability Theory Presentation 14
14/26
Covariance/correlation
cov(X, Y) = E(X EX)(Y EY).corr(X, Y) = cov(X, Y)/XY.
Why correlation is always in [1, 1]? Cauchy-Schwarz.
Variance of summation.
var(X1 + X2 + . . .) =
varXi + 2
i,j=1;i
-
8/8/2019 Probability Theory Presentation 14
15/26
Independence (I))
Two events A, B are independent if P(A B) = P(A)P(B).
Two random variables X, Y are independence if for any
A,B B, P(X A,Y B) = P(X A)P(Y B).
Two -algebras F and G are independent if all events ineach algebra are independent.
The second definition is a special case of 3 because X
induces a -algebra FX = X1(B) on . Note that
FX F (coarser).
For more than one variables: X1,X2, . . .. We have to makesure for all B1,B2, . . . B, X
1(B1),X1(B2), . . . are
independent.
Qiu, Lee BST 401
http://find/http://goback/ -
8/8/2019 Probability Theory Presentation 14
16/26
Independence (I))
Two events A, B are independent if P(A B) = P(A)P(B).
Two random variables X, Y are independence if for any
A,B B, P(X A,Y B) = P(X A)P(Y B).
Two -algebras F and G are independent if all events ineach algebra are independent.
The second definition is a special case of 3 because X
induces a -algebra FX = X1(B) on . Note that
FX F (coarser).
For more than one variables: X1,X2, . . .. We have to makesure for all B1,B2, . . . B, X
1(B1),X1(B2), . . . are
independent.
Qiu, Lee BST 401
http://find/http://goback/ -
8/8/2019 Probability Theory Presentation 14
17/26
Independence (I))
Two events A, B are independent if P(A B) = P(A)P(B).
Two random variables X, Y are independence if for any
A,B B, P(X A,Y B) = P(X A)P(Y B).
Two -algebras F and G are independent if all events ineach algebra are independent.
The second definition is a special case of 3 because X
induces a -algebra FX = X1(B) on . Note that
FX F (coarser).
For more than one variables: X1,X2, . . .. We have to makesure for all B1,B2, . . . B, X
1(B1),X1(B2), . . . are
independent.
Qiu, Lee BST 401
http://find/http://goback/ -
8/8/2019 Probability Theory Presentation 14
18/26
Independence (I))
Two events A, B are independent if P(A B) = P(A)P(B).
Two random variables X, Y are independence if for any
A,B B, P(X A,Y B) = P(X A)P(Y B).
Two -algebras F and G are independent if all events ineach algebra are independent.
The second definition is a special case of 3 because X
induces a -algebra FX = X1(B) on . Note that
FX F (coarser).
For more than one variables: X1,X2, . . .. We have to makesure for all B1,B2, . . . B, X
1(B1),X1(B2), . . . are
independent.
Qiu, Lee BST 401
http://find/http://goback/ -
8/8/2019 Probability Theory Presentation 14
19/26
Independence (I))
Two events A, B are independent if P(A B) = P(A)P(B).
Two random variables X, Y are independence if for any
A,B B, P(X A,Y B) = P(X A)P(Y B).
Two -algebras F and G are independent if all events ineach algebra are independent.
The second definition is a special case of 3 because X
induces a -algebra FX = X1(B) on . Note that
FX F (coarser).
For more than one variables: X1,X2, . . .. We have to makesure for all B1,B2, . . . B, X
1(B1),X1(B2), . . . are
independent.
Qiu, Lee BST 401
http://find/http://goback/ -
8/8/2019 Probability Theory Presentation 14
20/26
(Independence (II))
Distribution function and independence: X1,X2, . . . areindependent iff F(x1, x2, . . . ) = F(x1)F(x2) . . ..
The part is just a special case of the definition. The
part relies on the Carathodory extension theorem.In terms of density function: the same multiplication rule.
If X1,X2, . . . are independent, then f1(X1), f2(X2), . . . areindependent. This is because fi(X) can only induce a
-algebra that is coarser than FX. Corollary (4.5) is aslightly stronger version of this proposition.
Qiu, Lee BST 401
http://find/http://goback/ -
8/8/2019 Probability Theory Presentation 14
21/26
(Independence (II))
Distribution function and independence: X1,X2, . . . areindependent iff F(x1, x2, . . . ) = F(x1)F(x2) . . ..
The part is just a special case of the definition. The
part relies on the Carathodory extension theorem.In terms of density function: the same multiplication rule.
If X1,X2, . . . are independent, then f1(X1), f2(X2), . . . areindependent. This is because fi(X) can only induce a
-algebra that is coarser thanF
X. Corollary (4.5) is aslightly stronger version of this proposition.
Qiu, Lee BST 401
http://find/http://goback/ -
8/8/2019 Probability Theory Presentation 14
22/26
(Independence (II))
Distribution function and independence: X1,X2, . . . areindependent iff F(x1, x2, . . . ) = F(x1)F(x2) . . ..
The part is just a special case of the definition. The
part relies on the Carathodory extension theorem.In terms of density function: the same multiplication rule.
If X1,X2, . . . are independent, then f1(X1), f2(X2), . . . areindependent. This is because fi(X) can only induce a
-algebra that is coarser thanF
X. Corollary (4.5) is aslightly stronger version of this proposition.
Qiu, Lee BST 401
http://find/http://goback/ -
8/8/2019 Probability Theory Presentation 14
23/26
(Independence (II))
Distribution function and independence: X1,X2, . . . areindependent iff F(x1, x2, . . . ) = F(x1)F(x2) . . ..
The part is just a special case of the definition. The
part relies on the Carathodory extension theorem.In terms of density function: the same multiplication rule.
If X1,X2, . . . are independent, then f1(X1), f2(X2), . . . areindependent. This is because fi(X) can only induce a
-algebra that is coarser thanF
X. Corollary (4.5) is aslightly stronger version of this proposition.
Qiu, Lee BST 401
http://find/http://goback/ -
8/8/2019 Probability Theory Presentation 14
24/26
Independence and Expectation
Being independent implies being uncorrelated
EXY = EX EY but not vice versa.
This is because covariance is a summary statistic which is
sort of the average dependence and being dependent
means independence almost everywhere. Example: X,Y are two discrete random variables. X Y meansP(X = t,Y = s) = P(X = t)P(Y = s) for every t, s, sothere are many, many constraint equations. On the other
hand, EXY = EXEY is just one constraint equation
t X(t)Y(t) =
t
s XtYs.
The convolution formula for the sum of two independent
r.v.s.
Qiu, Lee BST 401
http://find/http://goback/ -
8/8/2019 Probability Theory Presentation 14
25/26
Independence and Expectation
Being independent implies being uncorrelated
EXY = EX EY but not vice versa.
This is because covariance is a summary statistic which is
sort of the average dependence and being dependent
means independence almost everywhere. Example: X,Y are two discrete random variables. X Y meansP(X = t,Y = s) = P(X = t)P(Y = s) for every t, s, sothere are many, many constraint equations. On the other
hand, EXY = EXEY is just one constraint equation
t X(t)Y(t) =
t
s XtYs.
The convolution formula for the sum of two independent
r.v.s.
Qiu, Lee BST 401
http://find/http://goback/ -
8/8/2019 Probability Theory Presentation 14
26/26
Independence and Expectation
Being independent implies being uncorrelated
EXY = EX EY but not vice versa.
This is because covariance is a summary statistic which is
sort of the average dependence and being dependent
means independence almost everywhere. Example: X,Y are two discrete random variables. X Y meansP(X = t,Y = s) = P(X = t)P(Y = s) for every t, s, sothere are many, many constraint equations. On the other
hand, EXY = EXEY is just one constraint equation
t X(t)Y(t) =
t
s XtYs.
The convolution formula for the sum of two independent
r.v.s.
Qiu, Lee BST 401
http://find/http://goback/