Escolar Documentos
Profissional Documentos
Cultura Documentos
Bayesian Decision
Theory
(Sections 2.3-2.5)
Error Bounds
Bayesian Decision Theory
1 ( "
1 x−µ %
2+
2
N(x; µ, σ ) = exp *− $ '-
2π σ *) 2 # σ & -,
2 1 # 1 t −1 &
N(x; µ, σ ) = 1/2
exp %− (x − µ ) Σ (x − µ )(
(2π )d/2 Σ $ 2 '
x = (x1, x2, …, xd)t
µ = (µ1, µ2, …, µd)t mean vector
S = d*d covariance matrix
|S| and S-1 are determinant and inverse respectively
Pattern Classification, Chapter 2 (Part 2) 5
Discriminant Functions
for the Normal Density
1 t −1 d 1
gi (x) = − (x − µi ) Σ i (x − µi ) − ln2π − ln Σ i + ln P(ω i )
2 2 2
gi (x) = w it x + w i0
where :
µi 1 t
w i = 2 ; w i0 = − 2 µi µi + ln P(ω i )
σ 2σ
€
(ω i0 is called the threshold for the ith category!)
€
Discriminant Functions
for the Normal Density…
• A classifier that uses linear discriminant functions is called
“a linear machine”
1 s2 P( w i )
x0 = ( µ i + µ j ) - ln ( µi - µ j )
2 µi - µ j
2
P( w j )
1 t −1 d 1
gi (x) = − (x − µi ) Σ i (x − µi ) − ln2π − ln Σ i + ln P(ω i )
2 2 2
• Expand the term and disregard the quadratic expression
€ t
where :
gi (x) = w x + w i0
i 1 t −1
−1
w i = ∑ µ; w i0 = − µi ∑ µi + ln P(ω i )
2
1
x 0 = ( µi + µ j ) −
[
ln P(ω i ) /P(ω j ) ] .( µi − µ j )
t −1
2 ( µi − µ j ) Σ ( µi − µ j )
é0.3 0.2ù
S=ê ú
C ë 0.2 0.3û
B
B
A: (0.5, 0.5)
AA B: (0, 1)
C: (1.5, 1.5)
Mahalanobis
Distance
Covariance Matrix:
é0.3 0.2ù
S=ê ú
ë 0.2 0.3û
C
A: (0.5, 0.5)
B B: (0, 1)
B
C: (1.5, 1.5)
A
A
Euclid(A,B) = 0.5
Euclid(A,B) = 2
Mahal(A,B) = 5
Mahal(A,C) = 4
19 Pattern Classification, Chapter 2 (Part 3)
20 Pattern Classification, Chapter 2 (Part 3)
Discriminant Functions for
the Normal Density…
• Case Si = arbitrary
• The covariance matrices are different for each category
g i ( x ) = x tWi x + w it x = w i 0
where :
1 -1
Wi = - S i
2
w i = S i- 1 µ i
1 t -1 1
w i 0 = - µ i S i µ i - ln S i + ln P ( w i )
2 2
P(ω1)=P(ω2)
Disconnected
decision
regions
Pattern Classification, Chapter 2 (Part
23
3)
Pattern Classification, Chapter 2 (Part
24
3)
Discriminant Functions for
the Normal Density
Decision Regions for Two-
Dimensional Gaussian Data
Decision Regions for Two-
Dimensional Gaussian Data
Error Probabilities and Integrals
Error Probabilities and Integrals
• 2-class problem: There are two types of errors
• Multi-class problem
– Simpler to computer the prob. of being correct (more ways to be
wrong than to be right)
Error Bounds for Normal Densities
• Inequality:
Chernoff Bound
• Inequality:
Chernoff bound
• To derive a bound for the error,
Chernoff bound
Assuming conditional prob. are normal,
where
When the two covariance matrices are equal, k(1/2) is the same
as the Mahalanobis distance between the two means
Bhattacharya bound
Error bounds for
gaussian distribution