math 256 probability and random processes
DESCRIPTION
OKAN UNIVERSIT Y FACULTY OF ENGINEERING AND ARCHITECTURE. MATH 256 Probability and Random Processes. 04 Random Variables. Yrd . Doç . Dr. Didem Kivanc Tureli [email protected] [email protected]. Fall 2011. Probability Mass Function. Is defined for a discrete variable X. - PowerPoint PPT PresentationTRANSCRIPT
MATH 256 Probability and Random Processes
Yrd. Doç. Dr. Didem Kivanc [email protected]
14/10/2011 Lecture 3
OKAN UNIVERSITYFACULTY OF ENGINEERING AND ARCHITECTURE
04 Random Variables
Fall 2011
Probability Mass Function
• Is defined for a discrete variable X.
4/10/2011 Lecture 3 2
p a P X a
0 1,2,...( ) 0
ip x ip a
p x x
for for all other values of
1
1ii
p x
• Suppose that
• Then since x must be one of the values xi,
Example of probability mass function
4/10/2011 Lecture 3 3
0 0 1 4
1 1 1 2
2 2 1 4
p P X
p P X
p P X
Expectation of a random variable• If X is a discrete random variable having a probability mass
function p(x) then the expectation or the expected value of X denoted by E[X] is defined by
4/10/2011 Lecture 3 4
: ( ) 0
( )x p x
E X xp x
• In other words, • Take every possible value for X• Multiply it by the probability of getting that value• Add the result.
Examples of expectation• For example, suppose you have a fair coin. You flip the coin,
and define a random variable X such that – If the coin lands heads, X = 1
– If the coin lands tails, X = 2
• Then the probability mass function of X is given by
4/10/2011 Lecture 3 5
11 2
2p p 1 2 if 1 or 2,
0 otherwise.x x
p x Or we can write
1 11 2 1.5
2 2E X
Expectation of a function of a random variable
• To find E[g(x)], that is, the expectation of g(X)• Two step process: – find the pmf of g(x)– find E[g(x)]
4/10/2011 Lecture 3 6
4/10/2011 Lecture 3 7
Let X denote a random variable that takes on any of the values –1, 0, and 1 with respective probabilities
{ 1} 0.2 { 0} 0.5 { 1} 0.3P X P X P X Compute 2E X
SolutionLet Y = X 2.
{ 1} { 1} { 1} 0.5{ 0} { 0} 0.5
P Y P X P XP Y P X
2 1(0.5) 0(0.5) 0.5E X E Y
Then the probability mass function of Y is given by
0.5 if 0 or 10 otherwise.
y yp y
Proposition 4.1
4/10/2011 Lecture 3 8
If X is a discrete random variable that takes on one of the values xi, i ≥ 1 with respective probabilities p(xi), then any real valued function g.
i ii
E g X g x p x Check if this holds for the previous example:
22 2 21 0.2 0 0.5 1 0.3
1 0.2 0.3 0 0.50.5
E X
Proof of Proposition 4.1
4/10/2011 Lecture 3 9
:
:
:
i j
i j
i j
i i i ii j i g x y
j ij i g x y
j ij i g x y
j jj
g x p x g x p x
y p x
y p x
y P g X y
E g X
4/10/2011 Lecture 3 10
4/10/2011 Lecture 3 11
4/10/2011 Lecture 3 12
4/10/2011 Lecture 3 13
Variance
4/10/2011 Lecture 3 14
Variance
4/10/2011 Lecture 3 15
4/10/2011 Lecture 3 16
Some more comments about variance
4/10/2011 Lecture 3 17
Bernoulli Random Variables
4/10/2011 Lecture 3 18
Binomial Random Variables
4/10/2011 Lecture 3 19
Poisson Random Variable
4/10/2011 Lecture 3 20
Some examples of random variables that generally obey the Poisson probabilitylaw [that is, they obey Equation (7.1)] are as follows:1. The number of misprints on a page (or a group of pages) of a book2. The number of people in a community who survive to age 1003. The number of wrong telephone numbers that are dialed in a day4. The number of packages of dog biscuits sold in a particular store each day5. The number of customers entering a post office on a given day6. The number of vacancies occurring during a year in the federal judicial system7. The number of α-particles discharged in a fixed period of time from some radioactivematerial
Expectation of the sum of r.v.s• Start with this:
• It’s really quite easy to show that:
• And from this we show that
4/10/2011 Lecture 3 21