Introduction To Probability Theory
Introduction To Probability Theory
K. Suresh Kumar
Department of Mathematics
Indian Institute of Technology Bombay
LECTURES 22-23
MX (t) = E[etX ], t ∈ I,
MX (t) = (1 − p) + pet , t ∈ R.
Now we will state and indicate the proofs of various properties of moment
generating functions.
Here M (k) (t) denote the kth derivative of MX (t) at t. Proof follows from
dk E[etX ] h dk etX i
= E , t ∈ (−h, h).
dtk dtk
So proof is all about justifying differentation under the ’integral’ sign.
The above theorem is bad for mgfs. What I meant is the following
”existence of mgf arround a neighourhood of 0, makes the random variable
very nice. This implies an unpleasent property of mgfs, i.e. they won’t
exists arround zero unless the random variable is nice”. We will see some
examples to illustrate this point.
We know that X doesn’t have finite mean (exercise). Hence Theorem 0.2
points to the non existence of mgf. In fact, we show that mgf of X exists
only at t = 0. For t > 0, consider
1 ∞ tx 1
Z
tX
E[e ] = e dx
π −∞ 1 + x2
1 ∞ tx 1
Z
≥ e dx
π 0 1 + x2
t ∞ x
Z
≥ dx
π 0 1 + x2
Z ∞
t 1
= dy.
2π 1 y
The RHS integral diverges to ∞. Hence by comparison of integrals, it follows
that E[etX ] diverges to ∞. i.e. MX (t) doesn’t exists for t > 0. Using a
simliar argument one can show that MX (t) doesn’t exists for t < 0 (exercise).
So MX (t) exists only at 0.
Example 0.4 (Log normal distribution) Let X ∼ eY , Y ∼ N (µ, σ 2 ). Then
X is said to be ’log normally’ distributed.
Note
n2
EX n = E[enY ] = MY (n) = e 2 , n ≥ 1.
Now since X ≥ 0, clearly MX (t) exists for t ≥ 0. Now for t > 0,
Y
E[etX ] = E[ete ]
Z ∞
1 y y2
= √ ete e− 2 dy
2π −∞
Z ∞
1
≥ √ edy,
2π K
for some K > 0 large enough. In the first inequality we used the fact that
there exists a K > 0
y y2
tee − ≥ 1 for all y ≥ K.
2
Now since the last integral diverges to ∞, it follows that E[etX diverges to
∞. Hence MX (t) doesn’t exists for t > 0, though EX n exists for all n ≥ 1.
Theorem 0.3 Let X1 , X2 , · · · , Xn be indedependent random variables with
mgfs MXi (t) exists on a common interval I, then the mgf of the sum Sn =
X1 + · · · + Xn exists on I and is given by
MSn (t) = Πnk=1 MXk (t), t ∈ I.
0.1. MOMENT GENERATING FUNCTION 5
n n
MX (t) = (1 − p)n + (1 − p)n−1 pet + (1 − p)n−2 p2 e2t + · · · + pn ent .
1 2
(1) n n−1 t n
MX (t) = (1 − p) pe + 2 (1 − p)n−2 p2 e2t + · · · + npn ent ,
1 2
(2) n n−1 t 2 n
MX (t) = (1 − p) pe + 2 (1 − p)n−2 p2 e2t + · · · + n2 pn ent ,
1 2
(k) n n−1 t k n
MX (t) = (1 − p) pe + 2 (1 − p)n−2 p2 e2t + · · · + nk pn ent ,
1 2
k = 3, 4, · · · .
Therefore
k n n−1 k n
EX = (1 − p) p+2 (1 − p)n−2 p2 + · · · + nk pn , k = 1, 2, · · ·
1 2
Theorem 0.4 Let X and Y be two random variables such that MX (t) =
MY (t) on some interval I, then X and Y has the same distribution function.
This proof is beyond the scope of this course. Any way we will see a similar
result for the characteristic functions soon.
6
The proof follows from the Riesz criterion (sufficient condition) for moment
determinancy given by
µ 1
2k 2k
lim inf < ∞.
k→∞ (2k)!
n!
lim √ 1 = 1.
n→∞ 2πnn+ 2 e−n
Remark 0.1 Theorem 0.5 gives a partial converse for Theorem 0.2. i.e., ”
if µk = EX k exists for all k, then MX (t) is uniquely determined by µk ’s” is
not true in general. But with an extra condition that µk doesn’t grow rapidly
1
1 2k
(for example if limk→∞ 2k µ2k = 0, then mgf is uniquely determined. This
gives a partial answer to the question, when given all moments determine a
distribution uniquely, because mgfs determinds distributions uniquely, a fact
we will not state or prove in this course.
with
n k
P {X = k} = p (1 − p)n−k := pk , k = 0, 1, · · · , n.
k
Hence
n
X
MX (t) = ketk pk , t ∈ R.
k=0
0.2. CHARACTERISTIC FUNCTIONS 7
Therefore
n
(m)
X
MX (t) = k m etk pk , t ∈ R.
k=0
ΦX (t) = EeitX , t ∈ R.
A digression: Before going further, I will give some very brief working
knowledge for complex valued functions defined on R.
Theorem 0.6 For any random variable X, its characteristic function φX (·)
is uniformly continuous on R and satisfies
(i) ΦX (0) = 1
(ii) |ΦX (t)| ≤ 1
(iii) ΦX (−t) = ΦX (t) , where for z a complex number, z denote the
conjugate.
Proof:
We prove (iii), (i) and (ii) are exercises.
Theorem 0.7 If the random variable X has finite moments upto order n.
Then Φ has continuous derivatives upto order n.More over
(k)
ik EX k = ΦX (0), k = 1, 2, . . . , n .
Proof.
Consider
ΦX (t + h) − ΦX (t) (eihX − 1)
= E[eitX ]
h h
(eihX − 1)
|eitX | ≤ |X|
h
10
(eihX − 1)
lim E[eitX ] = E[iXeitX ] .
h→0 h
Therefore
Φ0X (t) = E[iXeitX ] .
Put t = 0, we get
(1)
ΦX (0) = i EX .
For higher order derivatives, repeat the above arguments.
Proof. Before proceeding towards the sketch of proof, a word about the
integral on the rhs. The integral on the rhs is interpreted as an improper
Riemann integral which may not be in general absolutely integrable. At this
stage, student may not worry about this.
Consider
Z ∞ −ita Z ∞ −ita
1 e − e−itb 1 e − e−itb itX
ΦX (t)dt = Ee dt
2π −∞ it 2π −∞ Z ∞ −ita it
1 e − e−itb itX
= E e dt (0.1)
2π it
Z ∞ −∞
eit(X−a) − eit(X−b)
= E dt .
−∞ 2πit
The second equality follows from the change of order of integration (This
in fact, requires to consider the integrals on finite intervels say for example
[−T, T ](i.e. proper integrals) and use change of variable formula there and
then let T → ∞). Now
∞
0
eit(X−a) − eit(X−b) e−it(X−a) − e−it(X−b)
Z Z
dt = dt (0.2)
−∞ 2πit 0 2πit
0.2. CHARACTERISTIC FUNCTIONS 11
Similarly, the other integral. Combining (0.1), (0.3) and (0.4), we com-
plete the proof.
Remark
R ∞ sin αx 0.2 In this remark, I will give the computation R ∞ of the integral
sin x
0 x dx. It is enough to compute the Dirichlet integral 0 x dx because
for other values of α, the integral follows from the Dirichlet integral easily.
For example, Z ∞ Z ∞
sin(−x) sin x
dx = − dx.
0 x 0 x
R∞
Before even proceeding to compute this, I will show that 0 sinx x dx is not
absolutely integrable but is integrable. First recall that
Z ∞ Z T
sin x sin x
dx := lim dx.
0 x ε→0,T →∞ ε x
Consider
Z T Z T
sin x 1
dx = (1 − cos x)0 dx
1 x 1 x
h 1 − cos x iT Z T 1 − cos x
integration by parts = + dx
x 1 1 x2
Now since
sin x
lim = 1
x↓0 x
12
sin x
the function f (x) = x ,x > 0, = 1, x = 0 is Riemann integrable and hence
Z 1
sin x
lim dx exists.
ε→0 ε x
The last integral has two parts in which first one is known to diverge to ∞
and the second one converges using the same arguement as for the Dirichlet
integral. Hence the integral on the lhs also diverges.
Now we compute the Dirichlet integral. To this end, first we need the
value of the following. Consider for any given u > 0
Z ∞ Z T
eix e−xu dx = lim e(i−u)x dx
0 T →∞ 0
h e(i−u)x iT
= lim
T →∞i−u 0
−1 u+i
= = .
i−u 1 + u2
Hence equating the real and imaginary parts, we get the values of the fol-
lowing improper Riemann integrals
Z ∞ Z ∞
−xu u 1
cos xe dx = 2
, sin xe−xu dx = .
0 1+u 0 1 + u2
0.2. CHARACTERISTIC FUNCTIONS 13
Now consider
Z ∞ Z ∞ Z ∞
sin x
dx = sin x e−xu dudx
0 x
Z0 ∞ Z ∞ 0
Proof:
Using Inversion theorem, we have
F1 (b) = F2 (b)