EE5111 – Estimation Theory Mini Project 3 solved

$40.00

Category: You will receive a download link of the .ZIP file upon Payment

Description

5/5 - (1 vote)

Study of Expectation Maximization (EM) algorithm
The objective of this exercise is to understand the Expectation Maximization (EM) algorithm.
Assume that there are two biased coins, labelled as A and B. Coin A lands up as head with
probability p and coin B lands up as head with probability q. An experiment is performed with
n independent trials and in each trial, a coin is chosen at random (coin A with probability π
and coin B with probability 1 − π) and tossed m times (independently).
Let z
(i) ∈ {A, B} be the label of the coin selected and x
(i) ∈ {H, T}
m be the observed sequence
in the i
th trial of the experiment, i ∈ 1, 2, . . . , n. The labels of the coins {z
(i)
, i ∈ 1, 2, . . . , n}
remain hidden and the observer could only record the faces {x
(i)
, i ∈ 1, 2, . . . , n}. Hence, the
observations x
(i)
can be considered as i.i.d samples from a Mixture of two Binomial models:
PX(x) = π × PX(x|z
(i) = A) + (1 − π) × PX(x|z
(i) = B)
Our aim is to obtain maximum likelihood estimate for parameter vector θ using Expectation
Maximization (EM) algorithm. Generate the observations (x
(i)
, z(i)
) for following two cases
(i) π = 0.50, p = 0.35 and q = 0.60.
(ii) π = 0.25, p = 0.35 and q = 0.60.
Choose m = 1, 10 and n = 10, 1000, 10000. Run the EM algorithm for following experiments
Experiment: 1 Assume that we know π. Thus, the vector of parameters is given by θ =
[p q]
T
.
(a) Use observations from case (i) and assume that π = 0.50 and initial estimates are
[0.45 0.50]T
(b) Use observations from case (ii) and assume that π = 0.50 and initial estimates are
[0.45 0.50]T
(c) Use observations from case (ii) and assume that π = 0.25 and initial estimates are
[0.45 0.50]T
Experiment: 2 Assume that we do not know π. Thus, the vector of parameters is given by
θ = [π p q]
T
.
(a) Use observations from case (ii) and initial estimates are [0.50 0.45 0.50]T
(b) Repeat above experiment with initial estimate [0.50 0.50 0.45]T
Experiment: 3 Now use a Beta prior over π. Derive the update equations for MAP estimate.
Use the following priors Beta (1, 1), Beta (1, 3), Beta (2, 6), Beta (3, 9).
Make the following inferences from the algorithm for the aforementioned choices of n, m and θˆ0:
1
(1) Plot the learning curve1 and show the convergence of the algorithm2
for each experiment.
(2) Observe how the final estimate of θ from the algorithm (call it θˆEM ), and number of
iterations needed for convergence, change when we increase m and n.
(3) Report your observation about case (b) and case (c) of experiment 1, where in former case
we are assuming a wrong value of π and in later case we are assuming the true value of π
(4) Report your observation about experiment 2 when we swap the prior for p and q.
(5) Compare the result of experiment 3 with the result of experiment 2.
1Plot of the estimate at iteration k, θˆk vs. iteration index, k.
2You shall consider that the algorithm has converged at iteration k when the update to any of the parameter
is not more than  = 10−6
(i.e., ||θˆk − θˆk−1||∞ = max(|θˆk − θˆk−1|) ≤ ).
2