750 likes | 833 Views
Explore a novel approach for generating photo-realistic images with free virtual camera motion using hybrid scene representation. Learn about rendering techniques, refinement steps, and mask extraction methods. Correct rendering errors and ghosting artifacts for optimal results. Experiment with different point prediction methods for deghosting.
E N D
Novel view synthesis from still pictures by Frédéric ABAD under the supervision of Olivier FAUGERAS1 Luc ROBERT2 Imad ZOGHLAMI2 1 ROBOTVIS Team INRIA Sophia Antipolis 2 REALVIZ SA
Novel view synthesis • Given data: • Few reference photographs • Reference camera calibration • Objective: • Photo-realistic image generation • Free virtual camera motion • In particular: correct handling of parallax and image resolution
Novel view synthesis • Usual approaches: • Model-based rendering (light simulation with mathematical models) • Image-based rendering (image interpolation) • Our approach: • Hybrid image-model based rendering (texture mapping)
Our approach • Based on a hybrid scene representation • Rough 3D model + few images (reference images and masks) • Layer factorization • Rendering engine (main processing step) • View-dependent texture mapping • Double layered-structure • Refinement step (post-processing step) • Rendering errors occur when the 3D model is too rough • Mask extraction (pre-processing step) • Segmentation of the layers in the reference images
Our approach • Hybrid scene representation • Rendering engine (main processing step) • Refinement step (post-processing step) • Mask extraction (pre-processing step)
Our approach • Hybrid scene representation • Rendering engine (main processing step) • Refinement step (post-processing step) • Mask extraction (pre-processing step)
Scene representation • Hybrid representation: • Few reference images • Rough 3D model (built by image-based modeling) • 3D structure decomposed into layers • Binary layer masks extracted from the reference images
Scene representation (example) • Reference images
Scene representation (example) • 3D model
Scene representation (example) • Layer map
Scene representation (example) • Masks extracted from reference image #1
Our approach • Hybrid scene representation • Rendering engine (main processing step) • Refinement step (post-processing step) • Mask extraction (pre-processing step)
Our approach • Hybrid scene representation • Rendering engine (main processing step) • Refinement step (post-processing step) • Mask extraction (pre-processing step)
Rendering engine • View-dependent texture mapping [Debevec:96] • Efficient combination of the different reference images with respect to the virtual viewpoint. • Optimal image resolution • Double layered-structure, three steps: • Independant rendering of each geometric layer with the best 3 reference textures • Intra-layer compositing (for VDTM) • Inter-layer compositing (for occlusion processing)
View-dependent texture mapping Basic texture mapping Reference image weighting
Double layered-structure Intra-layer compositing Inter-layer compositing Intra-layer compositing
Rendering engine (example) • Results: hole-filling by VDTM
Rendering engine (example) • Results: generated movie
Our approach • Hybrid scene representation • Rendering engine (main processing step) • Refinement step (post-processing step) • Mask extraction (pre-processing step)
Our approach • Hybrid scene representation • Rendering engine (main processing step) • Refinement step (post-processing step) • Mask extraction (pre-processing step)
Refinement step • Rendering errors occur with basic texture mapping if the 3D model is too rough (‘Geometric Rendering Errors’ or GRE) • GRE’s are responsible for ‘ghosting artefacts’ with view-dependent texture mapping
Refinement step • Origin of the Geometric Rendering Errors
Refinement step • Origin of the ghosting artefacts
Refinement step • Our correcting approach: 1) Detect GRE’s in auxiliary reference images 2) Propagate them in new generated images 3) Correct them by image morphing
Our correcting approach • Step 1: detect GRE’s in an auxiliary image Model-based stereo [Debevec:96]
Our correcting approach • Step 2: GRE propagation by point prediction Known point Searched point Known point Known point Known point
Point prediction methods • Example 1: epipolar transfer
Point prediction methods • Example 2: Shashua’s cross-ratio method [Shashua:93]
Point prediction methods • Other point prediction methods: • 3D point reconstruction and projection • Trifocal transfer • Compact cross-ratio method • Irani’s parallax-based multi-frame rigidity constraint method
Our correcting approach • Step 3: correct the GRE’s by image morphing
Our correcting approach • Experimental comparison of point prediction methods: • Epipolar transfer: simplest implementation but imprecise and instable close to the trifocal plane • Irani’s approach: complex, imprecise and instable method • Cross-ratio approaches: simple, precise and stable methods
Our correcting approach • Experimental application to deghosting
Experimental application to deghosting • Before deghosting +
Experimental application to deghosting • After deghosting +
Experimental application to deghosting • Comparison before/after Before deghosting After deghosting
Our approach • Hybrid scene representation • Rendering engine (main processing step) • Refinement step (post-processing step) • Mask extraction (pre-processing step)
Our approach • Hybrid scene representation • Rendering engine (main processing step) • Refinement step (post-processing step) • Mask extraction (pre-processing step)
Mask extraction • Extract layer masks from reference images Reference image Ii Layer Cj Mask Mij
Mask extraction • Region-based image segmentation: pixel labelling by energy minimization • Energies • Optimization techniques
Mask extraction • Region-based image segmentation: pixel labelling by energy minimization • Energies • Optimization techniques
Mask extraction • Energies: Data attachment term + Regularization term
Energies • Data attachment term • Ensures the adaptation of the labelling to the observed data in the image • Inverse of the labelling likelihood
Data attachment term • Usual segmentation criteria: • Luminance: luma-key • Color: chroma-key • Texture: texture-key • Emphasis on a new geometric criterion: planar-key
Regularization term • Ensures one stable and unique solution • ‘Markovian Random Field’ a priori • 4-connexity neighborhood • Second order cliques • Generalized Potts Model potential function
Planar-key • Exploits geometric a priori knowledge: Scene made with planar patches (3D model = triangular mesh) • 1 label = 1 plane = 1 homography (between the image to segment and an auxiliary image) • Data attachment energy: dissimilarity between the labelled pixel and its image by the homography associated with the label
Dissimilarity(p,HC.p) < Dissimilarity(p,HA.p) D(C,p) < D(A,p) Dissimilarity(p,HC.p) < Dissimilarity(p,HB.p) D(C,p) < D(B,p) Planar-key
Planar-key • Example: Auxiliary image Main image Structure of the scene Segmented image
Planar-key • Technique more complex than it seems: • Dissimilarity measures • Photometric discrepancy robustness • Geometric inaccuracy robustness • Occlusion shadow error management