Purpose At the end of the course the student should be able to handle problems involving
probability distributions of a discrete or a continuous random variable.
Objectives
By the end of this course the student should be able to;
(1) Define the probability mass, density and distribution functions, and to use these to
determine expectation, variance, percentiles and mode for a given distribution
(2) Appreciate the form of the probability mass functions for the binomial, geometric,
hypergeometic and Poisson distributions, and the probability density functions for the
uniform, exponential gamma , beta and normal, functions, and their applications
(3) Apply the moment generating function and transformation of variable techniques
(4) Apply the principles of statistical inference for one sample problems.
DESCRIPTION
Random variables: discrete and continuous, probability mass, density and distribution
functions, expectation, variance, percentiles and mode. Moments and moment generating
function. Moment generating function and transformation Change of variable technique for
univariate distribution. Probability distributions: hypergeometric, binomial, Poisson, uniform,
normal, beta and gamma. Statistical inference including one sample normal and t tests.
Pre-Requisites: STA 2100 Probability and Statistics I, SMA 2104 Mathematics for Science
Course Text Books
1) RV Hogg, JW McKean & AT Craig Introduction to Mathematical Statistics, 6th ed.,
Prentice Hall, 2003 ISBN 0-13-177698-3
2) J Crawshaw & J Chambers A Concise Course in A-Level statistics, with worked examples,
3rd ed. Stanley Thornes, 1994 ISBN 0-534- 42362-0
Course Journals:
1) Journal of Applied Statistics (J. Appl. Stat.) [0266-4763; 1360-0532]
2) Statistics (Statistics) [0233-1888]
Further Reference Text Books And Journals:
a) HJ Larson Introduction to Probability Theory and Statistical Inference(Probability and
Mathematical Statistics) 3rd ed., Wiley, 1982
b) Uppal, S. M. , Odhiambo, R. O. & Humphreys, H. M. Introduction to Probability and
Statistics. JKUAT Press, 2005
c) I Miller & M Miller John E Freund’s Mathematical Statistics with Applications, 7th ed.,
Pearsons Education, Prentice Hall, New Jersey, 2003 ISBN: 0131246461
d) Statistical Science (Stat. Sci.) [0883-4237]
e) Journal of Mathematical Sciences
f) The Annals of Applied Probability
1
, 1. RANDOM VARIABLES
1.1 Introduction
In application of probability, we are often interested in a number associated with the
outcome of a random experiment. Such a quantity whose value is determined by the outcome
of a random experiment is called a random variable. It can also be defined as any quantity
or attribute whose value varies from one unit of the population to another.
A discrete random variable is function whose range is finite and/or countable, Ie it can only
assume values in a finite or countably infinite set of values. A continuous random variable is
one that can take any value in an interval of real numbers. (There are uncountably many real
numbers in an interval of positive length.)
1.2 Discrete Random Variables and Probability Mass Function
Consider the experiment of flipping a fair coin three times. The number of tails that appear is
noted as a discrete random variable. X= number of tails that appear in 3 flips of a fair coin.
There are 8 possible outcomes of the experiment: namely the sample space consists of
S HHH , HHT , HTH , HTT , THH , THT , TTH , TTT
X 0 1, 1, 2 , 1, 2, 2, 3
are the corresponding values taken by the random variable X.
Now, what are the possible values that X takes on and what are the probabilities of X taking a
particular value?
From the above we see that the possible values of X are the 4 values
X = 0, 1, 2, 3
Ie the sample space is a disjoint union of the 4 events {X = j } for j=0,1,2,3
Specifically in our example:
X 0 = HHH X 1 = HHT, HTH, THH
X 2 = TTH, HTT, THT X 3 = TTT
Since for a fair coin we assume that each element of the sample space is equally likely (with
probability 18 , we find that the probabilities for the various values of X, called the probability
distribution of X or the probability mass function (pmf). can be summarized in the following
table listing the possible values beside the probability of that value
x 0 1 2 3
1 3 3 1
P(X=x) 8 8 8 8
Note: The probability that X takes on the value x, ie p(X x) , is defined as the sum of the
probabilities of all points in S that are assigned the value x.
We can say that this pmf places mass 83 on the value X = 2 .
The “masses” (or probabilities) for a pmf should be between 0 and 1.
The total mass (i.e. total probability) must add up to 1.
Definition: The probability mass function of a discrete variable is a graph, table, or formula
that specifies the proportion (or probabilities) associated with each possible value the random
variable can take. The mass function P(X x) (or just p(x) has the following properties:
0 p(x) 1 and p(x) 1
all x
More generally, let X have the following properties
i) It is a discrete variable that can only assume values x1 , x2 , .... xn
2
, ii) The probabilities associated with these values are
P( X x1 ) p1 , P( X x2 ) p2 ……. P( X xn ) pn
n
Then X is a discrete random variable if 0 pi 1 and p
i 1
i 1
Remark: We denote random variables with capital letters while realized or particular values
are denoted by lower case letters.
Example 1
Two tetrahedral dice are rolled together once and the sum of the scores facing down was
noted. Find the pmf of the random variable ‘the sum of the scores facing down.’
Solution
+ 1 2 3 4 Therefore t is given the pmf by the table
1 2 3 4 5 below
2 3 4 5 6 x 2 3 4 5 6 7 8
1 1 3 1 3 1 1
3 4 5 6 7 P(X=x) 16 8 16 4 16 8 16
4 5 6 7 8 This can also be written as a function
x161 for x 2 , 3, 4 , 5
X = 1, 2, 3, 4 , 5 , 6 , 7 , 8 P( X x ) 9 x
16 for x 6 , 7, 8
Example 2
The pmf of a discrete random variable W is given by the table below
w -3 -2 -1 0 1
P(W=w) 0.1 0.25 0.3 0.15 d
Find the value of the constant d, P 3 w 0 , Pw -1 and P 1 w 1
Solution
p(W w) 1 0.1 0.25 0.3 0.15 d 1 d 0.2
all w
P 3 w 0 P(W 3) P(W 2) P(W 1) 0.65
Pw -1 Pw 0 Pw 1 0.15 0.2 0.35
P 1 w 1 P(W 0) 0.15
Example 3
A discrete random variable Y has a pmf given by the table below
y 0 1 2 3 4
P(Y=y) c 2c 5c 10c 17c
Find the value of the constant c hence computes P1 Y 3
Solution
p(Y y) 1 c(1 2 5 10 17) 1 c 351
ally
P1 Y 3 P(Y 1) P(Y 2) 352 355 15
Exercise
1. A die is loaded such that the probability of a face showing up is proportional to the face
number. Determine the probability of each sample point.
2. Roll a fair die and let X be the square of the score that show up. Wtrie down the
probability distribution of X hence compute P X 15 and P3 X 30
3
, 3. Let X be the random variable the number of fours observed when two dice are rolled
together once. Show that X is a discrete random variable.
4. The pmf of a discrete random variable X is given by P( X x ) kx for x 1, 2 , 3, 4 , 5 , 6
Find the value of the constant k, P X 4 and P3 X 6
5. A fair coin is flip until a head appears. Let N represent the number of tosses required to
realize a head. Find the pmf of N
6. A discrete random variable Y has a pmf given by P(Y y ) c 34 for y 0 ,1, 2 , .....
x
Find the value of the constant c and PX 3
2x
7. Verify that f(x) for x 0 ,1, 2 , .....k can serve as a pmf of a random variable X.
k (k 1)
8. For each of the following determine c so that the function can serve as a pmf of a random
variable X.
a) f(x) cx for x 1, 2 , 3, 4 , 5 c) f(x) c 16 for x 0 ,1, 2 , 3.....
x
b) f(x) cx 2 for x 0 ,1, 2 , .....k d) f(x) c2 x for x for x 0 ,1, 2 , ....
9. A coin is loaded so that heads is three times as likely as the tails. For 3 independent
tosses of the coin find the pmf of the total number of heads realized and the probability of
realizing at most 2 heads.
1.3 Continuous Random Variables and Probability Density Function
A continuous random variable can assume any value in an interval on the real line or in a
collection of intervals. The sample space is uncountable. For instance, suppose an experiment
involves observing the arrival of cars at a certain period of time along a highway on a
particular day. Let T denote the time that lapses before the 1st arrival, the T is a continuous
random variable that assumes values in the interval [0 , )
Definition: A random variable X is continuous if there exists a nonnegative function f so that, for
every interval B, P X B f(x) dx. The function f = f(x) is called the probability density
B
function of X.
Definition: Let X be a continuous random variable that assumes values in the interval
( , ) , The f(x) is said to be a probability density function (pdf) of X if it satisfies the
following conditions
b
f(x) 0 for all x , pa x b f(x) dx 1 and f(x) dx 1
a -
The support of a continuous random variable is the smallest interval containing all values of x
where f(x) >= 0.
Remark A crucial property is that, for any real number x, we have P(X x) 0 (implying
there is no difference between P(X x) and P(X x) ); that is it is not possible to talk about
the probability of the random variable assuming a particular value. Instead, we talk about the
probability of the random variable assuming a value within a given interval. The probability
of the random variable assuming a value within some given interval from x a to x b is
defined to be the area under the graph of the probability density function between
x a and x b .
Example 1
Let X be a continuous random variable. Show that the function
4