1 / 32

Pattern Recognition: Statistical and Neural

Nanjing University of Science & Technology. Pattern Recognition: Statistical and Neural. Lonnie C. Ludeman Lecture 14 Oct 14, 2005. Lecture 14 Topics. 1. Review structures of Optimal Classifier 2. Define Linear functions, hyperplanes, boundaries, unit normals,various distances

bmontague
Download Presentation

Pattern Recognition: Statistical and Neural

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Nanjing University of Science & Technology Pattern Recognition:Statistical and Neural Lonnie C. Ludeman Lecture 14 Oct 14, 2005

  2. Lecture 14 Topics 1. Review structures of Optimal Classifier 2. Define Linear functions, hyperplanes, boundaries, unit normals,various distances 3. Use of Linear Discriminant functions for defining classifiers- Examples

  3. Motivation! Motivation! Motivation!

  4. Optimum Decision Rules: 2-class Gaussian Case 1: K1 = K2 Quadratic Processing C1 if > T1 - (x – M1)TK1-1(x – M1) + (x – M2)TK2-1(x – M2) < C2 Case 2: K1 = K2 = K Linear Processing C1 > T2 if ( M1 – M2)T K-1 x < C2 Review 1

  5. Optimum Decision Rules: 2-class Gaussian (cont) Case 3: K1 = K2 = K = s2 I Linear Processing C1 > T3 if ( M1 – M2)T x < C2 Review 2

  6. M-Class General Gaussian MPE and MAP Case 1: K1 = K2 Select ClassCj if Qj(x) is MINIMUM Qi(x) = (x – Mj)TKj-1(x – Mj) } – 2 ln P(Cj) + ln | Ki | Case 2: K1 = K2 = … = KM = K Select ClassCj if Lj(x) is MAXIMUM Lj(x) = MjTK-1x – ½ MjTK-1Mj+lnP(Cj) Review 3

  7. M-Class General Gaussian: Bayes exp(- ½ (x – Mk)TKk-1(x – Mk) ) 1 p(x|Ck) = ½ N/2 (2 ) Kk Ck :X ~ N( Mk, Kk ) , P(Ck) Bayes decision rule is determined form a set of yi(x) defined by where Review 4

  8. ½ N/2 Kj (2 ) M Cij exp(- ½ (x – Mj)TKj-1(x – Mj)) P(Cj) yi(x) = j=1 Taking the ln of the yi(x) for this case does not simplify to a linear or quadratic processor The structure of the optimum classifier uses a sum of exp( quadratic forms) and thus is a special form of nonlinear processing using quadratic forms. Review 5

  9. Linear and Gaussian assumptions Quadratic processing Reasons for studying linear, quadratic and other special forms of non linear processing If Gaussian we can find or learn a usable decision rule and the rule is optimum If non-Gaussian case we can find or learn a usable decision rule; however the rule is NOT necessarily optimum

  10. Linear functions One Variable f(x1) = w1x1+ w2 Two Variables f(x1, x2 ) = w1x1+ w2x2 + w3 Three Variables f(x1, x2 , x3) = w1x1+ w2x2 + w2x2 + w3

  11. w1x1+ w2 = 0 Constant w1x1+ w2x2 + w3 = 0 Line w1x1+ w2x2 + w3x3 + w4 = 0 Plane ? w1x1+ w2x2 + w3x3 + w4x4 + w5 = 0 Answer = Hyperplane

  12. Hyperplanes n-dimensional Hyperplane w1x1+ w2x2 + … + wnxn + wn+1 = 0 Define x= [ x1,x2 ,… , xn ]T w0=[ w1, w2 ,… , wn]T An alternative representation of a Hyperplane is w0 x + wn+1 = 0 T

  13. Hyperplanes as boundaries for Regions Positive side of Hyperplane boundary R+ = {x: } Negative side of Hyperplane boundary } R- = {x: Hyperplane boundary w0 x + wn+1 = 0

  14. Definitions (1) Unit Normal u

  15. (2) Distance from a point y to the hyperplane (3) Distance from the origin to the hyperplane

  16. (4) Linear Discriminate Functions where Augmented Pattern Vector Weight vector

  17. Linear Decision Rule: 2-Class Case using single linear discriminant function given: d(x)=w1x1+ w2x2 + … + wnxn + wn+1 for a vector xif No claim of optimality !!!

  18. Linear Decision Rule: 2-Class Case using two linear discriminant function given two discriminant functions define decision rule by except on boundaries d1(x) = 0 and d2(x) = 0 where we decide randomly between C1 and C2

  19. Decision regions (2-class case) using two linear discriminant functions and AND logic

  20. Decision regions (2-class case) using two linear discriminant functions(continued)

  21. Decision regions (2-class case) alternative formulation using two linear discriminant functions

  22. Decision regions (2-class case) using alternative form of two linear discriminant functions equivalent to

  23. Decision regions (3-class case) using two linear discriminant functions

  24. Decision regions (4-class case) using two linear discriminant functions

  25. Decision region R1 (M-class case) using K linear discriminant functions

  26. Example: Piecewise linear boundaries Given the following discriminant functions

  27. Define the following decision rule Example Continued If d1(x) > 0 AND d2(x) > 0 OR d3(x) > 0 AND d4(x) > 0 AND d5(x) > 0 AND d6(x) > 0 then decide x comes from class C1, on the boundaries decide randomly, otherwise decide C2 Show the decision regions in the two dimensional pattern space

  28. Solution:

  29. Lecture 14 Summary 1. Reviewed structures of Optimal Classifier 2. Defined Linear functions, hyperplanes, boundaries, unit normals,various distances 3. Used Linear Discriminant functions for defining classifiers- Examples

  30. End of Lecture 14

More Related