Escolar Documentos
Profissional Documentos
Cultura Documentos
Prior
10
5
5
0 0
0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1
Normalized frequency f0 Normalized frequency f0
Prior and Posterior estimates
• Thus, the third curve is the product of the first two (with
normalization), and one Bayesian alternative is to use the
maximum as the estimate.
Prior and Posterior estimates
p( x|µ) 0.5
0
0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1
µ
Prior density before observing any data
0.05
p(µ)
0.025
0
0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1
µ
Posterior density after observing 3 heads
0.1
p(µ| x)
0.05
0
0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1
µ
Cost Functions
1. Quadratic: C(x) = x2
• This results in
ZZ Z Z
C(θ−θ̂)p(θ | x)p(x) dx dθ = C(θ − θ̂)p(θ | x) dθ p(x) dx
| {z }
(∗)
1
Note, that there’s a slight shift in the paradigm. The double integral results in the theoretical estimate that
requires the knowledge of p(x). When minimizing only the inner integral, we get the optimum for a particular
realization, not all possible realizations.
1. Quadratic Cost Solution (or the MMSE
estimator)
2
Prior PDF, p(θ), refers to the parameter distribution before any observations are made. Posterior PDF, p(θ | x),
refers to the parameter distribution after observing the data.
2. Absolute Cost Solution
with
0, |x| < δ
C(x) =
1, |x| > δ
3. Hit-or-miss Cost Solution (or the MAP
estimator)
Z Z θ̂−δ Z∞
C(θ− θ̂)p(θ | x) dθ = 1·p(θ | x) dθ+ 1·p(θ | x) dθ
−∞ θ̂+δ
or in a simplified form
Z Z θ̂+δ
C(θ − θ̂)p(θ | x) dθ = 1 − 1 · p(θ | x) dθ
θ̂−δ
3. Hit-or-miss Cost Solution (or the MAP
estimator)
Z θ̂+δ
p(θ | x) dθ
θ̂−δ
p(x | θ)p(θ)
θ̂MAP = arg max
θ p(x)
3. Hit-or-miss Cost Solution (or the MAP
estimator)
• Now,
∂ 3 (µ − 0.5)
ln [p(x|µ)p(µ)] = − = 0,
∂µ µ σ2
when
µ2 − 0.5µ − 3σ2 = 0.
This happens when
p √
0.5 ± 0.25 − 4 · 1 · (−3σ2 ) 0.25 + 12σ2
µ= = 0.25 ± .
2 2
Example
θ̂MMSE = E(θ | x)
or more explicitly
R
R θ1 p(θ | x) dθ
θ2 p(θ | x) dθ
θ̂MMSE =
..
R .
θp p(θ | x) dθ
Vector Parameter Case for MMSE
x = Hθ + w,
E(θ | x) = µθ + (C−1 T −1 −1 T −1
θ + H Cw H) H Cw (x − Hµθ ).
or in vector form
x = Hθ + w,
where
1 0
cos 2πf0 sin 2πf0
a
H=
cos 4πf0 sin 4πf0
and θ=
b
..
.
cos(2(N − 1)πf0 ) sin(2(N − 1)πf0 )
Example: MMSE Estimation of Sinusoidal
Parameters
E(θ | x) = µθ + (C−1 T −1 −1 T −1
θ + H Cw H) H Cw (x − Hµθ )
Example: MMSE Estimation of Sinusoidal
Parameters
3
Note the correspondence with Ridge regression. It holds that Ridge regression is equivalent to the Bayesian
estimator with Gaussian prior for the coefficients. It also holds that the LASSO is equivalent to the Bayesian
estimator with Laplacian prior.
Example: MMSE Estimation of Sinusoidal
Parameters
N
HT H ≈ I
2
• Thus,
−1
1 N 1
E(θ | x) ≈ 2
I+ 2 I HT x
σθ 2σw σ2w
1
σ2w
= 1 N
HT x.
σ2θ
+ 2σ2w
Example: MMSE Estimation of Sinusoidal
Parameters
2 X
N−1
" #
1
âMMSE = 2σ2 /N
x[n] cos 2πf0 n
1+ N
σ2θ n=0
2 X
" N−1 #
1
b̂MMSE = 2σ2 /N
x[n] sin 2πf0 n
1+ N
σ2θ n=0
Example: MMSE Estimation of Sinusoidal
Parameters
2 X
N−1
âMVU = x[n] cos 2πf0 n
N
n=0
2 X
N−1
b̂MVU = x[n] sin 2πf0 n
N
n=0
Example: MMSE Estimation of Sinusoidal
Parameters
50 50
40 40
30 30
20 20
10 10
0 0
−1 −0.5 0 0.5 1 −1 −0.5 0 0.5 1
50 50
40 40
30 30
20 20
10 10
0 0
−1 −0.5 0 0.5 1 −1 −0.5 0 0.5 1
Example: MMSE Estimation of Sinusoidal
Parameters
50 50
40 40
30 30
20 20
10 10
0 0
−1 −0.5 0 0.5 1 −1 −0.5 0 0.5 1
50 50
40 40
30 30
20 20
10 10
0 0
−1 −0.5 0 0.5 1 −1 −0.5 0 0.5 1
Example: MMSE Estimation of Sinusoidal
Parameters
• Assume that
θ exp(−θx[n]) if x[n] > 0
p(x[n] | θ) =
0, if x[n] < 0
• Now,
Y
"N−1 #
ln p(x | θ) + ln p(θ) = ln θ exp(−θx[n]) + ln[λ exp(−λθ)]
n=0
X
" N−1
!#
N
= ln θ exp −θ x[n] + ln[λ exp(−λθ)]
n=0
= N ln θ − Nθx̄ + ln λ − λθ
• Differentiation produces
d N
ln p(x | θ) + ln p(θ) = − Nx̄ − λ
dθ θ
Example: MAP Estimator
1
θ̂ = λ
x̄ + N
Example: Deconvolution
nX
s −1
x = Hs + w
ŝ = Cs (Cs + σ2 I)−1 x
rss [0]
ŝ[0] = x[0]
rss [0] + σ2