Problem 1: Otherwise, 0 X 0 1), 0 ( ) (
Problem 1: Otherwise, 0 X 0 1), 0 ( ) (
داﻧﺸﮕﺎه ﺗﻬﺮان
داﻧﺸﻜﺪه ﻣﻬﻨﺪﺳﻲ ﺑﺮق و ﻛﺎﻣﭙﻴﻮﺗﺮ
درس ﺑﺎزﺷﻨﺎﺧﺖ اﻟﮕﻮ
آﺑﺎنﺁﺑﺎﻧﻤﺎﻩ
1389 2310ﺩﻭﺷﻨﺒﻪ
دوﺷﻨﺒﻪ:ﻣﻮﻋﺪ ﺗﺤﻮﻳﻞ ﺗﻤﺮﻳﻦ ﺳﺮي دوم
Problem 1
Let {x k }, k 1,2,..., N denote independent training samples from one of the following densities.
Obtain the Maximum Likelihood estimate of θ in each case.
a) f ( x k ; ) exp(x k ) xk 0 0 Exponential Density
2
xk x
b) f ( x k ; ) 2 exp(- k 2 ) xk 0 0 Rayleigh Density
2
1
c) f ( x k ; ) x k 0 xk 1 0 Beta Density
Problem 2
Let x have a uniform density
1 0 x
f X ( x ) ~ U (0, )
0 otherwise,
a) Suppose that n samples D {x1 , x 2 ,..., x n } are drawn independently according to
f X ( x ) . Show that the maximum likelihood estimate for is max[D], i.e., the value of
the maximum element in D.
b) Suppose that n=5 points are drawn from the distribution and the maximum value of
which happens to be max k xk = 0.6. Plot the likelihood f X ( D ) in the range 0 1.
Explain in words why you do not need to know the values of the other four points.
Problem 3
Let us review the Expectation Maximization (EM) algorithm . We want to estimate a
parametric pdf / y(y;O) using observations on X. The random vector X is re lated to Y
through a many to one mappin g g( . ) , where g : R n ~ R I and /I > ! . This is in deed
estimation of di stribution of Y using incomplete data X = {x ... } i_I'
where x ... = g(y ... ) ,
k ~ 1 , 2, ... , N
Now, let use the idea of Maximum Likelihood (ML) estimation, i.e. fin ding the
parameters 0 in such a way that the observed outcome become the most likely outcome of
the random experiment. Therefore, we should maximize the fo llowing joint distribution
which is ca lled the likelihood function
N
1(0) ~ f( {y ,}: _,; OI X) ~ IT f , (y ,; 0 I x, )
,-,
or equivalently we may maximize the lOll-likelihood function
N
L(O) ~ ln f( {y , }:_,;OI X) ~ L:ln f , (y , ;0 I x, )
,-,
If observation x ... corresponds to a unique y ... (complete data) then L(O) is a deterministic
,
functio n of 0 and the ML estimate of 0 would be 0 ML = arg maxL (O) . However, if
2
-
j\il-step - Compute eCr + 1) by maximizing Q(9;9(t»
-
- -
,
9(1 + 1) ~ argmllxQ(9;9(1))
Now let use EM algorithm in mixhu"e density model
, ,
", (x:9) ~ L f( x:9 1j)p j Pj ~ O L Pj = l
) _1 ) _1
3
Problem 4
Problem 5 – optional
Nonparametric Estimate and Classification
The figure below illustrates two dimensional training samples from two classes. and
represent classes 1 and 2, respectively. Classify the test sample (0.5,0) t using:
a) Parzen window approximation using the uniform window function
1
1 u i , i 1,2
(u1 , u 2 ) 2
0 otherwise
i- v N is a square with a side of hN 2
ii- v N is a circle with a radius of rN 1
b) Unbiased K Nearest neighbor approximation with k=3, and
i- Euclidean distance: d ( x, y ) (x
i
i yi ) 2