discrete random variablesdiscrete random variables chapter 3 { lecture 12 yiren ding shanghai qibao...

14
Discrete Random Variables Chapter 3 – Lecture 12 Yiren Ding Shanghai Qibao Dwight High School April 4, 2016 Yiren Ding Discrete Random Variables 1 / 14

Upload: others

Post on 09-Mar-2021

33 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Discrete Random VariablesDiscrete Random Variables Chapter 3 { Lecture 12 Yiren Ding Shanghai Qibao Dwight High School April 4, 2016 Yiren Ding Discrete Random Variables 1 / 14 Outline

Discrete Random VariablesChapter 3 – Lecture 12

Yiren Ding

Shanghai Qibao Dwight High School

April 4, 2016

Yiren Ding Discrete Random Variables 1 / 14

Page 2: Discrete Random VariablesDiscrete Random Variables Chapter 3 { Lecture 12 Yiren Ding Shanghai Qibao Dwight High School April 4, 2016 Yiren Ding Discrete Random Variables 1 / 14 Outline

Outline

1 Independence of Random VariablesCumulative Distribution FunctionDefinition of Independence

2 Alternative DefinitionCorollaries

3 Expected Value and VarianceIndependence and Expected ValueIndependence and VarianceSquare-root Law

4 Convolution RuleExample

Yiren Ding Discrete Random Variables 2 / 14

Page 3: Discrete Random VariablesDiscrete Random Variables Chapter 3 { Lecture 12 Yiren Ding Shanghai Qibao Dwight High School April 4, 2016 Yiren Ding Discrete Random Variables 1 / 14 Outline

Independence of Random Variables Cumulative Distribution Function

Definition 1 (Cumulative Distribution Function).

The cumulative distribution function (CDF) of a discrete randomvariable X is defined by

FX (x) = P(X ≤ x) =∑t≤x

P(X = t).

The functino FX (x) tells you the probability that the random variableX takes on values at most as extreme as x .

For example, the median of a random variable X can be definedmore precisely as the value m = x1+x2

2 , where

x1 = max{x : FX (x) ≤ 0.5}x2 = min{x : FX (x) ≥ 0.5}.

Yiren Ding Discrete Random Variables 3 / 14

Page 4: Discrete Random VariablesDiscrete Random Variables Chapter 3 { Lecture 12 Yiren Ding Shanghai Qibao Dwight High School April 4, 2016 Yiren Ding Discrete Random Variables 1 / 14 Outline

Independence of Random Variables Definition of Independence

Definition 2 (Independence of Random Variables).

Let X and Y be two random variables (discrete or continuous) that aredefined on the same sample space with probability measure P. And let FXand FY denote their CDF’s, respectively. The random variables X and Yare said to be independent if

P(X ≤ x ,Y ≤ y) = FX (x)FY (x),

for any x and y , where P(X ≤ x ,Y ≤ y) represents the probability ofoccurrence of both events {X ≤ x} and {Y ≤ y}.

In words, the random variable X and Y are independent if the eventof X taking on a value less than or equal to x and the event that Ytakes on value less than or equal to y are independent for all x and y .

It is worth noting that this definition works for any random variable,not just discrete.

Yiren Ding Discrete Random Variables 4 / 14

Page 5: Discrete Random VariablesDiscrete Random Variables Chapter 3 { Lecture 12 Yiren Ding Shanghai Qibao Dwight High School April 4, 2016 Yiren Ding Discrete Random Variables 1 / 14 Outline

Alternative Definition

Theorem 1 (Alternative Definition).

For any two sets A,B ⊆ R, Definition 2 is equivalent to

P(X ∈ A,Y ∈ B) = P(X ∈ A)P(Y ∈ B).

Proof. We first prove this for the special case where

A = {r : am ≤ r ≤ aM} and B = {r : bm ≤ r ≤ bM}.

For convenience, let aXb denote a ≤ X ≤ b, and P = P(X ∈ A,Y ∈ B):

P = P(amXaM , bmYbM) = P(XaM , bmYbM)− P(Xam, bmYbM)

= P(XaM ,YbM)− P(XaM ,Ybm)− P(Xam,YbM) + P(XaM ,Ybm)

= FX (aM)FY (bM)− FX (aM)FY (bm)− FX (am)FY (bM) + FX (aM)FY (bm)

= (FX (aM)− FX (am))(FY (bM)− FY (bm)) = P(amXaM)P(bmYbM)

= P(X ∈ A)P(Y ∈ B).

Yiren Ding Discrete Random Variables 5 / 14

Page 6: Discrete Random VariablesDiscrete Random Variables Chapter 3 { Lecture 12 Yiren Ding Shanghai Qibao Dwight High School April 4, 2016 Yiren Ding Discrete Random Variables 1 / 14 Outline

Alternative Definition

Proof of Theorem 1

Now, for arbitrary sets A and B, we can always write

A =⋃i

Ai and B =⋃j

Bj ,

where Ai and Bi are subsets of the form {r : α ≤ r ≤ β}, for some α, β ∈ R.

Therefore, by the axioms of probability theory, we have,

P = P(x ∈⋃i

Ai ,Y ∈⋃j

Bj) =∑i

P(x ∈ Ai ,Y ∈⋃j

Bj)

=∑i

∑j

P(x ∈ Ai ,Y ∈ Bj) =∑i

∑j

P(x ∈ Ai )P(Y ∈ Bj)

=∑j

P(X ∈ A)P(Y ∈ Bj) = P(X ∈ A)P(Y ∈ B).

Yiren Ding Discrete Random Variables 6 / 14

Page 7: Discrete Random VariablesDiscrete Random Variables Chapter 3 { Lecture 12 Yiren Ding Shanghai Qibao Dwight High School April 4, 2016 Yiren Ding Discrete Random Variables 1 / 14 Outline

Alternative Definition Corollaries

Corollary 1.

If X and Y are independent random variables, then the random variablesf (X ) and g(Y ) are independent for any two functions f and g .

Proof. Suppose that X and Y are independent. Let A′ and B ′ denotethe images f (A) and g(B), respectively.

Since X and Y are independent, by Theorem 1:

P(f (X ) ∈ A′, g(Y ) ∈ B ′) = P(X ∈ A,Y ∈ B)

= P(X ∈ A)P(Y ∈ B)

= P(f (X ) ∈ A′)P(g(Y ) ∈ B ′).

Hence f (X ) and g(X ) are also independent by Theorem 1.

Yiren Ding Discrete Random Variables 7 / 14

Page 8: Discrete Random VariablesDiscrete Random Variables Chapter 3 { Lecture 12 Yiren Ding Shanghai Qibao Dwight High School April 4, 2016 Yiren Ding Discrete Random Variables 1 / 14 Outline

Alternative Definition Corollaries

Corollary 2 (Independence of Discrete Random Variables).

Discrete random variables X and Y are independent if and only if

P(X = x ,Y = y) = P(X = x)P(Y = y) for all x , y .

Proof. This is the easiest proof in the world.

Simply let A = {x} and B = {y} in Theorem 1, and we’re done.

Yiren Ding Discrete Random Variables 8 / 14

Page 9: Discrete Random VariablesDiscrete Random Variables Chapter 3 { Lecture 12 Yiren Ding Shanghai Qibao Dwight High School April 4, 2016 Yiren Ding Discrete Random Variables 1 / 14 Outline

Expected Value and Variance Independence and Expected Value

Theorem 2.

If the random variable X and Y are independent, then

E (XY ) = E (X )E (Y ),

assuming that E (X ) and E (Y ) exist and are finite.

Proof. Let I and J denote the range of X and Y , respectively, anddefine the random variable Z by Z = XY .

Then we have

E (Z ) =∑z

P(Z = z) =∑z

∑xy=z

P(X = x ,Y = y)

=∑z

∑xy=z

xyP(X = x ,Y = y)

Yiren Ding Discrete Random Variables 9 / 14

Page 10: Discrete Random VariablesDiscrete Random Variables Chapter 3 { Lecture 12 Yiren Ding Shanghai Qibao Dwight High School April 4, 2016 Yiren Ding Discrete Random Variables 1 / 14 Outline

Expected Value and Variance Independence and Expected Value

Theorem 2 Proof

Since X and Y are independent, by Corollary 2,

E (Z ) =∑x ,y

xyP(X = x ,Y = y) =∑x ,y

xyP(X = x)P(Y = y)

=∑x∈I

xP(X = x)∑y∈J

yP(Y = y) = E (X )E (Y )

Note that the converse of this theorem is not true!

For example, suppose two fair dice are tossed. Denote by the randomvariable V1 the number appearing on the first die and by V2 thenumber appearing on the second die.

Let X = V1 + V2 and Y = V1 − V2. It is obvious that X and Y arenot independent. (Why?) Verify that E (XY ) = E (X )E (Y ).

Yiren Ding Discrete Random Variables 10 / 14

Page 11: Discrete Random VariablesDiscrete Random Variables Chapter 3 { Lecture 12 Yiren Ding Shanghai Qibao Dwight High School April 4, 2016 Yiren Ding Discrete Random Variables 1 / 14 Outline

Expected Value and Variance Independence and Variance

Theorem 3.

If the random variables X and Y are independent, then

var(X + Y ) = var(X ) + var(Y ).

Proof. Let µX = E (X ) and µY = E (Y ). We have

var(X + Y ) = E ((X + Y )2)− (µX + µY )2

= E (X 2 + 2XY + Y 2)− µ2X − 2µXµY − µ2Y= E (X 2) + ����2µXµY + E (Y 2)− µ2X −����2µXµY − µ2Y= var(X ) + var(Y )

Note that in general, if X1,X2, ...,Xn are independent,

var(X1 + X2 + · · ·+ Xn) = var(X1) + var(X2) + · · · var(Xn).

Yiren Ding Discrete Random Variables 11 / 14

Page 12: Discrete Random VariablesDiscrete Random Variables Chapter 3 { Lecture 12 Yiren Ding Shanghai Qibao Dwight High School April 4, 2016 Yiren Ding Discrete Random Variables 1 / 14 Outline

Expected Value and Variance Square-root Law

Corollary 3.

If the random variables X1,X2, ...,Xn are i.i.d. (independently identicallydistributed) with standard deviation σ, then the standard deviation of thesum X1 + X2 + · · ·+ Xn is given by

σ(X1 + X2 + · · ·+ Xn) = σ√n.

This is one of the most important results used in statistics. It isgenerally stated as the famous square-root law:

σ

(X1 + X2 + · · ·+ Xn

n

)=

σ√n.

The term (X1 + X2 + · · ·Xn)/n represents the sample mean of nsamples, and is itself a random variable. The Central Limit Theoremis closely associated with this random variable.

Yiren Ding Discrete Random Variables 12 / 14

Page 13: Discrete Random VariablesDiscrete Random Variables Chapter 3 { Lecture 12 Yiren Ding Shanghai Qibao Dwight High School April 4, 2016 Yiren Ding Discrete Random Variables 1 / 14 Outline

Convolution Rule

Theorem 4 (Convolution Rule).

If discrete random variables X and Y have the set of nonnegative integersas the range of possible values, and are independent, then

P(X + Y = k) =k∑

j=0

P(X = j)P(Y = k − j) for k = 0, 1, ....

Proof. By definition of X + Y and independence,

P(X + Y = k) =∑

x+y=k

P(X = x ,Y = y)

=k∑

j=0

P(X = j ,Y = k − j)

=k∑

j=0

P(X = j)P(Y = k − j)

Yiren Ding Discrete Random Variables 13 / 14

Page 14: Discrete Random VariablesDiscrete Random Variables Chapter 3 { Lecture 12 Yiren Ding Shanghai Qibao Dwight High School April 4, 2016 Yiren Ding Discrete Random Variables 1 / 14 Outline

Convolution Rule Example

Example 1.

Suppose the random variables X and Y are independent and have Poissondistribution with respective means λ and µ. What is the probabilitydistribution of X + Y ?

By the convolution rule and the binomial theorem,

P(X + Y = k) =k∑

j=0

e−λλj

j!e−µ

µk−j

(k − j)!

=e−(λ+µ)

k!

k∑j=0

(k

j

)λjµk−j

= e−(λ+µ)(λ+ µ)k

k!, for k = 0, 1, ....

Hence, X + Y is also Poisson distributed with mean λ+ µ.

Yiren Ding Discrete Random Variables 14 / 14