220 likes | 368 Views
Part 3 Vector Quantization and Mixture Density Model. CSE717, SPRING 2008 CUBS, Univ at Buffalo. Vector Quantization. Quantization Represents continuous range of values by a set of discrete values Example: floating-point representation of real numbers in computer Vector Quantization
E N D
Part 3 Vector Quantization and Mixture Density Model CSE717, SPRING 2008 CUBS, Univ at Buffalo
Vector Quantization • Quantization • Represents continuous range of values by a set of discrete values • Example: floating-point representation of real numbers in computer • Vector Quantization • Represent a data space (vector space) by discrete set of vectors
Vector Quantizer A mapping from vector space onto a finite subset of the vector space Y = y1,y2,…,yN finite subset of IRk, referred to as the codebook of Q Q is usually determined by training data
Partition of Vector Quantizer The vector space is partitioned into N cells by the vector quantizer
Properties of Partition Vector quantizer Q defines a complete and disjoint partition of IRk into R1,R2,…,RN
Quantization Error • Quantization Error for single vector x is a suitable distance measure • Overall Quantization Error
Nearest-Neighbor Condition The minimum quantization error of a given codebook Y is given by partition y3 x y1 y2
Centroid Condition • Centroid of a cell Ri • is minimized by choosing as the codebook • For Euclidean distance d Centroid
Vector Quantizer Design – General Steps • Determine initial codebook Y0 • Adjust partition of sample data for the current codebook Ym using nearest-neighbor condition • Update the codebook Ym→Ym+1 using centroid condition • Check a certain condition of convergence. If it converges, return the current codebook Ym+1; otherwise go to step 2 N Converge? Y
Mixture Density Model A mixture model of N random variables X1,…,XN is defined as follows: is a random variables defined on N labels
Mixture Density Model Suppose the p.d.f.’s of X1,…,XN are and then
Example: Gaussian Mixture Model of Two Components Histogram of samples Mixture Density
Estimation of Gaussian Mixture Model ML Estimation (Value X and label are given) Samples in the format of (-0.39, 0), (0.12, 0), (0.94, 1), (1.67, 0), (1.76, 1), … S1 (Subset of ): (0.94, 1), (1.76, 1), … S2 (Subset of ): (-0.39, 0), (0.12, 0), (1.67, 0), …
Estimation of Gaussian Mixture Model EM Algorithm (Value X is given, label is unknown) 1. Choose initial values of 2. E-Step: For each sample xk, label is missing. But we can estimate using its expected value Samples in the format of ; is missing (-0.39), (0.12), (0.94), (1.67), (1.76), … x1 x2
Estimation of Gaussian Mixture Model EM Algorithm (Value X is given, label is unknown) 3. M-Step: We can estimate again using the labels estimated in the E-Step:
Estimation of Gaussian Mixture Model EM Algorithm (Value X is given, label is unknown) 4. Termination: The log likelihood of n samples At the end of m-th iteration, if terminate; otherwise go to step 2 (the E-Step).