360 likes | 522 Views
A Study on the 3D Facial Feature Extraction and Motion Recovery using Multi-modal Information. Sang Hoon Kim, Hankyong National Univ., Korea, south wind1104@hanmail.net http://auto.hankyong.ac.kr/kimsh. Experience.
E N D
A Study on the 3D Facial Feature Extraction and Motion Recovery using Multi-modal Information Sang Hoon Kim, Hankyong National Univ., Korea, south wind1104@hanmail.net http://auto.hankyong.ac.kr/kimsh
Experience 1989-1994 Researcher, Institute of LG Semiconductor co.,Ltd 1994-1999 Research Assistant, Dept. of Electronics, Korea University 1999-2001 Senior Research Scientist, Imaging and Media Research Center, KIST(Korea Institute of Science and Technology) 1999- present Assistant professor in Dept. of Information and Control, Hankyong National University, Kyonggi-do, Korea Research Interests Face detection and tracking in real time Pose estimation Multi-modal fusion for object detection and recognition
Problems are face detection • How to detect facial area exactly from complex background • in variable conditions. • How to define and extract facial features location(and how many • features are proper to do next step) • How to recover 3D facial features and global motion from • the 2D information. • How to combine the whole results to synchronize a synthetic • face model with a real face. global motion recovery and application Evaluation method is important !! to prove the accuracy of 3D recovered information
Categorization of methods for face detection • Knowledge-based top-down methods • Bottom-up feature based methods (this work classified here from the survey) • Template matching • Appearance-based methods “ Detecting Faces in Images: A Survey ” IEEE Trans. on PAMI, vol.24, no.1, Jan 2002 Ming-Hsuan Yang, David J. Kriegman, Narendra Ahuja
Transform 3D information into FAP of MPEG4 and synchronize a generic model with a real face facial features and motion extraction using paraperspective camera model Facial features extraction using moving color information Overall Blockdiagram
Generalized Skin Color Distribution(GSCD) The skin color distribution in the normalized (r,g) domain can be approximated by the 2D Gaussian distribution GSCD in (r,g) domain
GSCD color transform 1. Color Transform defined as Z(x,y) = GSCD ( r(x,y) , g(x,y) ) 2. Result intensity value of enhanced facial regions suggests the possibility of being facial color GSCD input color image Ic(x,y) gray level image Z(x,y)
LoG and MPC Disparity Map Left camera Image grab Enhanced image Disparity map Right camera LoG MPC stereo matching Disparity D(x,y) is Vth : predefined threshold value
MPC similarity measure disparity map results with Random-dot stereo image (a) left image (b) right image (c) SAD (d) NCC (e) MPC similarity measures matching ratio comparison when noises are increased ( 0% ---> 20% )
Object 1 frequency Object 2 Background disparity threshold 2Disparity Histogram(DH) 1. Definition : occurrence frequency of each disparity value obtained from MPC disparity map 2. Meaning : the locations and number of objects
Range Segmentation 1. The outline of the DH curve is smoothed with average filtering 2. Region having the smallest disparity value is defined as background (Although it has skin-colored components, they can be regarded as noises or useless data) 3. Regions having frequency values less than a specified threshold are assumed to have no object and merged to background region 4. Regions having continuous disparity values greater than a specified threshold are defined as objects with same label
Moving Color Enhancement 1 This idea based on the fact that regions detected as faces only using skin color are unstable and have variable skin color space due to the lighting condition, so adding the motion information to the skin color enhanced region would be effective so as to increase the probability of faces So given an GSCD transformed image, region that hasa low gray value, which means low probability of face, should be detected as a face only when it haslarge motion, while region that has high gray level, which means high probability of face, can be detected as a face even when it has small motion.
Moving Color Enhancement 2 The SWUPC(Sigmoid function Weighted Unmatched Pixel Count) operation emphasizes only the region with a motion in the skin-color enhanced region. where
(e) (d) (a) (c) (b) • (a) present frame (b) previous frame • (c) UPC result (d) WUPC result (e) SWUPC result Moving color enhanced results
Test image 1 Step 2 Step 3 Step 4 Step 5 Final detection
Test image 2 Step 2 Step 3 Step 4 Step 5 Final detection
Test image 3 Step 2 Step 3 Step 4 Step 5 Final detection
Test image 4 Step 2 Step 3 Step 4 Step 5 Final detection
Test image 5 Step 2 Step 3 Step 4 Step 5 Final detection
Test image 6 face region detection result using range,skin color and motion information (a)input color image(t=1)(b)input color image(t=2) (c)MPC disparity map(d)range segmented image(g)skin color transformed image (h)swupc transform image (i)final face area
Present image GSCD Previous image GSCD BWCD Final eyes, eyebrows SWUPC Morphological Result Erosion Dilation Extracted facial features Automatic Facial Features Extraction - eyes, eyebrows
Hypothetical Image Plane Image Plane Focal length, l World Origin paraperspective camera model Paraperspective Camera model why use? This model can describe the change of object’s depth information Refer to “ A Paraperspective factorization method for shape and motion recovery” Conrad J.Poelman and Takeo Kanade
Basic matrix Final M,S recovered (1) A recovered (2) (3) (8) SVD- M,S separation (4) (7) (6) (5) Sequential SVD Process (Matrix time series)
-100,0,-50 0,0,0 0,100,-50 0,-100,-50 100,0,-50 Pyramid type 3D test object
20th frame with motion viewed from two cameras(left and right camera) 40th frame with motion viewed from two cameras(left and right camera) Multi-view Synthetic Pyramid test images
Motion and Shape Recovery Results using synthetic face image synchronized with real face
(front view) (left view) (right view) (upward face) (downward face) Real facial information --> Face Animation Parameter
Conclusions 1. This technique works well when detecting faces using GSCD(Generalized Skin Color Distribution) combined with the Range Segmentation using MPC(Matching Pixel Count) DH(Disparity Histogram) and Object’s moving color information(SWUPC) 2. Automatic facial features extraction(95%) using moving color information(SWUPC, BWCD) and morphological process. 3. Motion and 3D facial features recovery(80%, 208 of 210 frames) using multi-view paraperspective model and factorization(SVD) method 4. The results can be used for various applications