CLASSIFICATION CHAPTER 15 Supervised Classification A Dermanis Supervised

  • Slides: 20
Download presentation
CLASSIFICATION CHAPTER 15 Supervised Classification A. Dermanis

CLASSIFICATION CHAPTER 15 Supervised Classification A. Dermanis

Supervised Classification The known pixels in each one of the predecided classes ω1, ω2,

Supervised Classification The known pixels in each one of the predecided classes ω1, ω2, . . . , ωK, form corresponding “sample sets” S 1, S 2, . . . , SK with n 1, n 2, . . . , n. K number of pixels respectively. Estimates from each sample set Si, (i = 1, 2, …, K ) : Class mean vectors: 1 mi = n i x x Si Class covariance matrices: 1 Ci = n i (x – m ) x Si i i T Supervised classification methods: Parallelepiped Euclidean distance (minimization) Mahalanobis distance (minimization) Maximum likelihood Bayesian (maximum a posteriori probability density) A. Dermanis

Classification with Euclidean distance d. E(x, x ) = || x – x ||

Classification with Euclidean distance d. E(x, x ) = || x – x || = (a) Simple (x 1 – x 1)2 + (x 2 – x 2)2 + … + (x. B – x B)2 || x – mi || = min || x – mk || x i k Assign each pixel to the class of the closest center (class mean) Boundaries between class regions = = perpendicular at middle of segment joining the class centers A. Dermanis

Classification with Euclidean distance d. E(x, x ) = || x – x ||

Classification with Euclidean distance d. E(x, x ) = || x – x || = (b) with threshold T (x 1 – x 1)2 + (x 2 – x 2)2 + … + (x. B – x B)2 || x – mi || = min || x – mk || x – mi || T x i Assign each pixel to the class of the closest center (class mean) if distance < threshold || x – mi || > T, i x 0 Leave pixel unclassified (class ω0) if all class centers are at distances larger than threshold A. Dermanis

Classification with Euclidean distance d. E(x, x ) = || x – x ||

Classification with Euclidean distance d. E(x, x ) = || x – x || = RIGHT (x 1 – x 1)2 + (x 2 – x 2)2 + … + (x. B – x B)2 WRONG The role of statistics (dispersion) in classification A. Dermanis

Classification with the parallelepiped method standard deviations for each band ij = (Ci)jj j

Classification with the parallelepiped method standard deviations for each band ij = (Ci)jj j = 1, 2, …, B parallelepipeds Pi x = [x 1 … xj … x. B]T Pj mij – k ij xj mij + k ij j = 1, 2, …, B Classification: x P j x i x P i x 0 i A. Dermanis

Classification with the Mahalanobis distance: C= 1 N i d. M(x, x ) =

Classification with the Mahalanobis distance: C= 1 N i d. M(x, x ) = (x – mi)T = x Si Classification (simple): Classification with threshold: (x – x )T C– 1 (x – x ) 1 N n. C i i (total covariance matrix) i d. M(x, mi) < d. M(x, mk), k i d. M(x, mi) T, x i d. M(x, mi) > T, i x 0 A. Dermanis

Classification with the maximum likelihood method Probability distribution density function or likelihood function of

Classification with the maximum likelihood method Probability distribution density function or likelihood function of class ωi: li(x) = 1 (2 )B/2 | Ci |1/2 Classification: exp [ – 1 (x – mi)T Ci– 1 (x – mi) ] 2 li(x) > lk(x) k i x i Equivalent use of decision function: di(x) = 2 ln[li(x)] + B ln(2 ) = – ln | Ci | – (x – mi)T Ci– 1 (x – mi) di(x) > dk(x) k i x i A. Dermanis

Classification using the Bayesian approach N: total number of pixels in the image (i.

Classification using the Bayesian approach N: total number of pixels in the image (i. e. in each band) B: number of bands, ω1, ω2, …, ωK : the K classes present in the image Ni : number of image pixels belonging to the class ωi (i = 1, 2, …, K) nx : number of pixels with value x (= vector of values in all bands) nxi : number of pixels with value x which also belong to the class ωi A. Dermanis

Classification using the Bayesian approach N: total number of pixels in the image (i.

Classification using the Bayesian approach N: total number of pixels in the image (i. e. in each band) B: number of bands, ω1, ω2, …, ωK : the K classes present in the image Ni : number of image pixels belonging to the class ωi (i = 1, 2, …, K) nx : number of pixels with value x (= vector of values in all bands) nxi : number of pixels with value x which also belong to the class ωi A. Dermanis

Classification using the Bayesian approach N: total number of pixels in the image (i.

Classification using the Bayesian approach N: total number of pixels in the image (i. e. in each band) B: number of bands, ω1, ω2, …, ωK : the K classes present in the image Ni : number of image pixels belonging to the class ωi (i = 1, 2, …, K) nx : number of pixels with value x (= vector of values in all bands) nxi : number of pixels with value x which also belong to the class ωi Basic identity: A. Dermanis

Classification using the Bayesian approach N: total number of pixels in the image (i.

Classification using the Bayesian approach N: total number of pixels in the image (i. e. in each band) B: number of bands, ω1, ω2, …, ωK : the K classes present in the image Ni : number of image pixels belonging to the class ωi (i = 1, 2, …, K) nx : number of pixels with value x (= vector of values in all bands) nxi : number of pixels with value x which also belong to the class ωi Basic identity: A. Dermanis

Classification using the Bayesian approach N: total number of pixels in the image (i.

Classification using the Bayesian approach N: total number of pixels in the image (i. e. in each band) B: number of bands, ω1, ω2, …, ωK : the K classes present in the image Ni : number of image pixels belonging to the class ωi (i = 1, 2, …, K) nx : number of pixels with value x (= vector of values in all bands) nxi : number of pixels with value x which also belong to the class ωi Basic identity: A. Dermanis

Ni p( i) = N probability of a pixel to belong to the class

Ni p( i) = N probability of a pixel to belong to the class ωi nx p(x) = N probability of a pixel to have the value x nxi p(x | i) = Ni probability of a pixel belonging to the class ωi to have value x (conditional probability) nxi p( i | x) = nx probability of a pixel having value x to belong to the class ωi (conditional probability) nxi p(x, i) = N probability of a pixel to have the value x and to simultaneously belong to ωi (joint probability) A. Dermanis

Ni p( i) = N probability of a pixel to belong to the class

Ni p( i) = N probability of a pixel to belong to the class ωi nx p(x) = N probability of a pixel to have the value x nxi p(x | i) = Ni probability of a pixel belonging to the class ωi to have value x (conditional probability) nxi p( i | x) = nx probability of a pixel having value x to belong to the class ωi (conditional probability) nxi p(x, i) = N probability of a pixel to have the value x and to simultaneously belong to ωi (joint probability) A. Dermanis

Ni p( i) = N probability of a pixel to belong to the class

Ni p( i) = N probability of a pixel to belong to the class ωi nx p(x) = N probability of a pixel to have the value x nxi p(x | i) = Ni probability of a pixel belonging to the class ωi to have value x (conditional probability) nxi p( i | x) = nx probability of a pixel having value x to belong to the class ωi (conditional probability) nxi p(x, i) = N probability of a pixel to have the value x and to simultaneously belong to ωi (joint probability) formula of Bayes A. Dermanis

The Bayes theorem: Pr(A B) Pr(A | B) = Pr(B) Pr(A | B) Pr(B)

The Bayes theorem: Pr(A B) Pr(A | B) = Pr(B) Pr(A | B) Pr(B) = Pr(A B) = Pr(B | A) Pr(A | B) Pr(B | A) = Pr(A) event A = occurrence of the value x event B = occurence of the class ωi p(x| i) p( i|x) = p(x) Classification: p( i |x) > p( k |x) k i x i p(x) = not necessary (common constant factor) Classification: p(x | i) p( i) > p(x | k) p( k) k i A. Dermanis

p(x| i) p( i) = max [p(x| k) p( k) Classification: for Gaussian distribution:

p(x| i) p( i) = max [p(x| k) p( k) Classification: for Gaussian distribution: k p(x | i) = li(x) = x i 1 1 T C – 1 (x – m ) } exp { – – (x – m ) i i i (2 )B/2 | Ci |1/2 2 Instead of p(x | i) p( i) = max equivalent ln[p(x | i) p( i)] = ln[p(x | i) + ln[p( i) = max – – 1 (x – mi)T Ci– 1 (x – mi) – 1– ln[ | Ci | + ln[p( i)] = max 2 2 or finally: (x – mi)T Ci– 1 (x – mi) + ln[ | Ci | + ln[p( i)] = min A. Dermanis

Bayesian Classification for Gaussian distribution : (x – mi)T Ci– 1 (x – mi)

Bayesian Classification for Gaussian distribution : (x – mi)T Ci– 1 (x – mi) + ln[ | Ci | + ln[p( i)] = min SPECIAL CASES: p( 1) = p( 2) = … = p( K) (x – mi)T Ci– 1 (x – mi) + ln[ | Ci | = min Maximum Likelihood ! p( 1) = p( 2) = … = p( K) C 1 = C 2 = … = CK = C (x – mi)T Ci– 1 (x – mi) = min Mahalanobis distance ! p( 1) = p( 2) = … = p( K) C 1 = C 2 = … = CK = I (x – mi)T (x – mi) = min Euclidean distance ! A. Dermanis

Want to learn more ? A. Dermanis L. Biagi: Telerilevamento Casa Editrice Ambrosiana

Want to learn more ? A. Dermanis L. Biagi: Telerilevamento Casa Editrice Ambrosiana