750 likes | 822 Views
Novel view synthesis from still pictures. by Frédéric ABAD under the supervision of Olivier FAUGERAS 1 Luc ROBERT 2 Imad ZOGHLAMI 2 1 ROBOTVIS Team INRIA Sophia Antipolis 2 REALVIZ SA. Novel view synthesis. Given data: Few reference photographs Reference camera calibration
E N D
Novel view synthesis from still pictures by Frédéric ABAD under the supervision of Olivier FAUGERAS1 Luc ROBERT2 Imad ZOGHLAMI2 1 ROBOTVIS Team INRIA Sophia Antipolis 2 REALVIZ SA
Novel view synthesis • Given data: • Few reference photographs • Reference camera calibration • Objective: • Photo-realistic image generation • Free virtual camera motion • In particular: correct handling of parallax and image resolution
Novel view synthesis • Usual approaches: • Model-based rendering (light simulation with mathematical models) • Image-based rendering (image interpolation) • Our approach: • Hybrid image-model based rendering (texture mapping)
Our approach • Based on a hybrid scene representation • Rough 3D model + few images (reference images and masks) • Layer factorization • Rendering engine (main processing step) • View-dependent texture mapping • Double layered-structure • Refinement step (post-processing step) • Rendering errors occur when the 3D model is too rough • Mask extraction (pre-processing step) • Segmentation of the layers in the reference images
Our approach • Hybrid scene representation • Rendering engine (main processing step) • Refinement step (post-processing step) • Mask extraction (pre-processing step)
Our approach • Hybrid scene representation • Rendering engine (main processing step) • Refinement step (post-processing step) • Mask extraction (pre-processing step)
Scene representation • Hybrid representation: • Few reference images • Rough 3D model (built by image-based modeling) • 3D structure decomposed into layers • Binary layer masks extracted from the reference images
Scene representation (example) • Reference images
Scene representation (example) • 3D model
Scene representation (example) • Layer map
Scene representation (example) • Masks extracted from reference image #1
Our approach • Hybrid scene representation • Rendering engine (main processing step) • Refinement step (post-processing step) • Mask extraction (pre-processing step)
Our approach • Hybrid scene representation • Rendering engine (main processing step) • Refinement step (post-processing step) • Mask extraction (pre-processing step)
Rendering engine • View-dependent texture mapping [Debevec:96] • Efficient combination of the different reference images with respect to the virtual viewpoint. • Optimal image resolution • Double layered-structure, three steps: • Independant rendering of each geometric layer with the best 3 reference textures • Intra-layer compositing (for VDTM) • Inter-layer compositing (for occlusion processing)
View-dependent texture mapping Basic texture mapping Reference image weighting
Double layered-structure Intra-layer compositing Inter-layer compositing Intra-layer compositing
Rendering engine (example) • Results: hole-filling by VDTM
Rendering engine (example) • Results: generated movie
Our approach • Hybrid scene representation • Rendering engine (main processing step) • Refinement step (post-processing step) • Mask extraction (pre-processing step)
Our approach • Hybrid scene representation • Rendering engine (main processing step) • Refinement step (post-processing step) • Mask extraction (pre-processing step)
Refinement step • Rendering errors occur with basic texture mapping if the 3D model is too rough (‘Geometric Rendering Errors’ or GRE) • GRE’s are responsible for ‘ghosting artefacts’ with view-dependent texture mapping
Refinement step • Origin of the Geometric Rendering Errors
Refinement step • Origin of the ghosting artefacts
Refinement step • Our correcting approach: 1) Detect GRE’s in auxiliary reference images 2) Propagate them in new generated images 3) Correct them by image morphing
Our correcting approach • Step 1: detect GRE’s in an auxiliary image Model-based stereo [Debevec:96]
Our correcting approach • Step 2: GRE propagation by point prediction Known point Searched point Known point Known point Known point
Point prediction methods • Example 1: epipolar transfer
Point prediction methods • Example 2: Shashua’s cross-ratio method [Shashua:93]
Point prediction methods • Other point prediction methods: • 3D point reconstruction and projection • Trifocal transfer • Compact cross-ratio method • Irani’s parallax-based multi-frame rigidity constraint method
Our correcting approach • Step 3: correct the GRE’s by image morphing
Our correcting approach • Experimental comparison of point prediction methods: • Epipolar transfer: simplest implementation but imprecise and instable close to the trifocal plane • Irani’s approach: complex, imprecise and instable method • Cross-ratio approaches: simple, precise and stable methods
Our correcting approach • Experimental application to deghosting
Experimental application to deghosting • Before deghosting +
Experimental application to deghosting • After deghosting +
Experimental application to deghosting • Comparison before/after Before deghosting After deghosting
Our approach • Hybrid scene representation • Rendering engine (main processing step) • Refinement step (post-processing step) • Mask extraction (pre-processing step)
Our approach • Hybrid scene representation • Rendering engine (main processing step) • Refinement step (post-processing step) • Mask extraction (pre-processing step)
Mask extraction • Extract layer masks from reference images Reference image Ii Layer Cj Mask Mij
Mask extraction • Region-based image segmentation: pixel labelling by energy minimization • Energies • Optimization techniques
Mask extraction • Region-based image segmentation: pixel labelling by energy minimization • Energies • Optimization techniques
Mask extraction • Energies: Data attachment term + Regularization term
Energies • Data attachment term • Ensures the adaptation of the labelling to the observed data in the image • Inverse of the labelling likelihood
Data attachment term • Usual segmentation criteria: • Luminance: luma-key • Color: chroma-key • Texture: texture-key • Emphasis on a new geometric criterion: planar-key
Regularization term • Ensures one stable and unique solution • ‘Markovian Random Field’ a priori • 4-connexity neighborhood • Second order cliques • Generalized Potts Model potential function
Planar-key • Exploits geometric a priori knowledge: Scene made with planar patches (3D model = triangular mesh) • 1 label = 1 plane = 1 homography (between the image to segment and an auxiliary image) • Data attachment energy: dissimilarity between the labelled pixel and its image by the homography associated with the label
Dissimilarity(p,HC.p) < Dissimilarity(p,HA.p) D(C,p) < D(A,p) Dissimilarity(p,HC.p) < Dissimilarity(p,HB.p) D(C,p) < D(B,p) Planar-key
Planar-key • Example: Auxiliary image Main image Structure of the scene Segmented image
Planar-key • Technique more complex than it seems: • Dissimilarity measures • Photometric discrepancy robustness • Geometric inaccuracy robustness • Occlusion shadow error management