0% found this document useful (0 votes)
10 views2 pages

EM Algorithm for Coin Bias Estimation

The mini project focuses on understanding the Expectation Maximization (EM) algorithm through experiments involving two biased coins. It requires generating observations under different parameters and running the EM algorithm to estimate the parameters. The project also includes tasks such as plotting learning curves, analyzing convergence, and comparing results across different experiments and assumptions about the parameters.
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
10 views2 pages

EM Algorithm for Coin Bias Estimation

The mini project focuses on understanding the Expectation Maximization (EM) algorithm through experiments involving two biased coins. It requires generating observations under different parameters and running the EM algorithm to estimate the parameters. The project also includes tasks such as plotting learning curves, analyzing convergence, and comparing results across different experiments and assumptions about the parameters.
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd

EE5111 - Estimation Theory

(Spring 2019)

Mini Project 3
Released: March 14, 2020 Deadline: March 31, 2020

Study of Expectation Maximization (EM) algorithm


The objective of this exercise is to understand the Expectation Maximization (EM) algorithm.
Assume that there are two biased coins, labelled as A and B. Coin A lands up as head with
probability p and coin B lands up as head with probability q. An experiment is performed with
n independent trials and in each trial, a coin is chosen at random (coin A with probability π
and coin B with probability 1 − π) and tossed m times (independently).

Let z (i) ∈ {A, B} be the label of the coin selected and x(i) ∈ {H, T }m be the observed sequence
in the ith trial of the experiment, i ∈ 1, 2, . . . , n. The labels of the coins {z (i) , i ∈ 1, 2, . . . , n}
remain hidden and the observer could only record the faces {x(i) , i ∈ 1, 2, . . . , n}. Hence, the
observations x(i) can be considered as i.i.d samples from a Mixture of two Binomial models:

PX (x) = π × PX (x|z (i) = A) + (1 − π) × PX (x|z (i) = B)

Our aim is to obtain maximum likelihood estimate for parameter vector θ using Expectation
Maximization (EM) algorithm. Generate the observations (x(i) , z (i) ) for following two cases

(i) π = 0.50, p = 0.35 and q = 0.60.

(ii) π = 0.25, p = 0.35 and q = 0.60.

Choose m = 1, 10 and n = 10, 1000, 10000. Run the EM algorithm for following experiments

Experiment: 1 Assume that we know π. Thus, the vector of parameters is given by θ =


[p q]T .

(a) Use observations from case (i) and assume that π = 0.50 and initial estimates are
[0.45 0.50]T
(b) Use observations from case (ii) and assume that π = 0.50 and initial estimates are
[0.45 0.50]T
(c) Use observations from case (ii) and assume that π = 0.25 and initial estimates are
[0.45 0.50]T

Experiment: 2 Assume that we do not know π. Thus, the vector of parameters is given by
θ = [π p q]T .

(a) Use observations from case (ii) and initial estimates are [0.50 0.45 0.50]T
(b) Repeat above experiment with initial estimate [0.50 0.50 0.45]T

Experiment: 3 Now use a Beta prior over π. Derive the update equations for MAP estimate.
Use the following priors Beta (1, 1) , Beta (1, 3) , Beta (2, 6) , Beta (3, 9).

Make the following inferences from the algorithm for the aforementioned choices of n, m and θ̂ 0 :

1
(1) Plot the learning curve 1 and show the convergence of the algorithm2 for each experiment.

(2) Observe how the final estimate of θ from the algorithm (call it θ̂ EM ), and number of
iterations needed for convergence, change when we increase m and n.

(3) Report your observation about case (b) and case (c) of experiment 1, where in former case
we are assuming a wrong value of π and in later case we are assuming the true value of π

(4) Report your observation about experiment 2 when we swap the prior for p and q.

(5) Compare the result of experiment 3 with the result of experiment 2.

1
Plot of the estimate at iteration k, θ̂ k vs. iteration index, k.
2
You shall consider that the algorithm has converged at iteration k when the update to any of the parameter
is not more than  = 10−6 (i.e., ||θ̂ k − θ̂ k−1 ||∞ = max(|θ̂ k − θ̂ k−1 |) ≤ ).

You might also like