230 likes | 393 Views
Lagadic. Visual Servoing in Robotics, Computer Vision, and Augmented Reality François Chaumette IRISA / INRIA Rennes. http://www.irisa.fr/lagadic. The Lagadic group. Spin-off of the Vista project in January 2004 Created as an Inria project in December 2004 Currently 13 people:
E N D
Lagadic Visual Servoing in Robotics, Computer Vision, and Augmented Reality François Chaumette IRISA / INRIA Rennes http://www.irisa.fr/lagadic
The Lagadic group Spin-off of the Vista project in January 2004 Created as an Inria project in December 2004 Currently 13 people: • François Chaumette, DR 2 • Éric Marchand, CR 1, HDR 2004 • Alexandre Krupa, CR 2, recruited in Sep. 2004(LSIIT, Strasbourg) • Fabien Spindler, IR 2 • 1 temporary research scientist: C. Collewet from Cemagref • 1 temporary assistant prof.: A. Remazeilles, INSA Rennes • 5 Ph. D. students: Master in Rennes (2), Strasbourg (2) and Grenoble (1) • 1 post-doc: S. Segvic from Croatia • 1 temporary engineer: F. Dionnet from LRP Paris
Research field Visual servoing : vision-based control of a dynamic system • Modeling: • Control law: Usually, highly nonlinear and coupled potential problems Objective: cook so that is as linear as possible
Objectives • Modeling visual features • for usual cameras (perspective projection) • for omni-directional cameras • for 2D ultrasound images • Considering high level tasks in complex environments • Robot navigation • Additional constraints (occlusions, joint limits avoidance, etc.) • Visual tracking • real-time • accurate for 6 dof • robust • mono-object • geometrical structure
Application fields • Robotics • Manipulating/grasping objects, target tracking • Nuclear/submarine/space/medical, etc. • Eye-in-hand/eye-to-hand systems • Robot arms, mobile robots, UAV • Augmented reality • Insert virtual objects in real images • Virtual reality • Viewpoint generation • Virtual cinematography • Control of virtual humanoid • Cogniscience
Experimental platforms • Eye-in-hand, eye-to hand systems, mobile robot, medical robot • Experimental validation, tests before transfer, demonstrations Experimental results very time consuming (same image never acquired, and useless after 40 ms)
Modeling image moments • Determination of the analytical form of the interaction matrix for any moment • Determination of combinations of moments (from invariants) for decoupling and linearizing properties usual choice with moments
Visual servoing from ultrasound images • Modeling features • No observation outside B-scan corresponding to the current 2D ultrasound image • Automation of spatial calibration procedure • Adaptive visual servoing to position B-scan on a cross-wire phantom • Robotized 3D «free-hand» ultrasound imaging • Conventional 2D ultrasound probe moved by a medical robot • Thanks to calibration step, B-Scans positioned in a 3D reference frame (collaboration with Visages) • Application field: remote examination
Navigation from an image database to • Appearance-based representation • Topological description of the environment with key images (no 3D reconstruction) • Image path retrieval from indexing techniques (collaboration with Texmex) • Qualitative visual servoing • Navigation expressed as visual features to be seen (and not successive poses to be reached) • Confident interval for features • Automatic update of features used for navigation (by imposing a progress within the visibility corridor) from
Tasks sequencing • Idea : to give as much freedom as possible to take constraints (joint limits, occlusions, obstacles) into account • Scheme more reactive than reactive path planning • Scheme more versatile than classical visual servoing • Redundancy framework revisited: directional redundancy • non linear projection operator to increase the free space where secondary tasks are applied • Visual elementary task managed by a stack • Remove the good task for ensuring the constraints • Put the task back when possible
3D model-based tracking • Virtual visual servoing scheme for pose computation • Virtually moves a camera so that the projection of the 3D model of the object corresponds to the observed image • Statistically robust pose estimation to deal with outliers and occlusions (M-estimation) • Real-time capabilities • Application to visual servoing and augmented reality • Extension to articulated object tracking
Texture and contours-based tracking • 2D model-based tracking • Estimation of an homography • Consider both edges and image intensities • 3D model-based tracking • Introducing spatio-temporal constraints in model-based tracking • Joint estimation of pose and displacement
Collaborations • Inside Inria : Visages (medical imaging) Icare (Predit Mobivip, Robea Bodega) • In France : 5 Robea projects • Omni-directional vision: Lasmea, Crea, Lirmm • Small helicopters: I3S, CEA • Mobile robot navigation (Lasmea, UTC) • Outside France : • ANU Canberra: modeling, helicopters • ISR Lisbon: jacobian learning • KTH Stockholm, CSIRO Melbourne, Urbana-Champaign
Publications • Main journals : IEEE TRA(O): 6, IJRR: 5 • Main conferences: ICRA:18, IROS:14 • Best paper award : IEEE TRA 2002, RFIA’2004 • Finalist papers : IROS’2004, AMDO’2004, ICRA’2004, IROS’2005
Transfert • Marker-less: 3D model-based tracker transferred to Total-Immersion for augmented reality (RIAM SORA) • France Télécom R&D: Augmented reality in urban environment • ESA: vision-based manipulation on the ISS with Eurobot
Software • ViSP: Open source software environment for visual servoing • Currently available for Linux and Mac OS with QPL license • Written in C++ (~ 100 000 lines of code) • Library of canonical vision-based tasks through many visual features • Suitable for 2D, 2½ D, 3D control laws • Eye-in-hand / eye-to-hand • Redundancy framework • Visual tracking algorithms • Independence wrt. the robotics platform, frame grabber • Simulator included (interface with OpenGL)
Positioning wrt. INRIA & French labs • INRIA scientific and technological challenges: • (4): Coupling models and data to simulate and control complex systems • (5): Combining simulation, visualization and interaction (real-time, augmented reality) • (7): Fully integrating ICST into medical technology (medical imaging, medical robotics) • Inside INRIA: • Icare (Num A: Control and complex systems): visual servoing and control • Vista, Movi, Isa: visual tracking • Other French labs: • LASMEA: visual tracking, position-based visual servoing • LSIIT: visual servoing for medical robotics • LRP, I3S
Worldwide positioning • Pioneering lab: CMU (1984 – 1994, no more active) • Main labs: • USA: (S. Hutchinson, G. Hager) • Australia (P. Corke), Japan (K. Hashimoto) • Europe: KTH (more recently) • Other labs : almost everywhere (Italy, Spain, Portugal, Germany, Canada, Mexico, Brazil, South Korea, China, etc.) • Visual tracking: Cambridge, EPFL • Lagadic: High visibility in the robotics community • AE IEEE TRA(O) • Look for “visual servoing” ∪ “visual servo” in Google Scholar
Evolution wrt. past objectives • From the 2001 Vista evaluation experts report: “Vista is planning to split off its activities in visual servoing and active vision as a separate project. This is an excellent decision” • Evolution wrt. scientific objectives: 80 % well done • Complex objects of unknown shape: image moments • Outliers: M-estimator integrated in the control loop • Applications in robotics: underwater, space, flying robots • Applications outside robotics: virtual reality, augmented reality • Visual servoing directly on image intensity: future objective
Objectives: modeling visual features • Spherical projection: • same model for perspective projection and omni-directional cameras • nice geometrical properties • Modeling directly the image intensity (no image processing, many unknown parameters, cooking very challenging) • Enclosing volume for 3D objects (global and sufficient information) • Mobile/flying robots: non holonomic or underactuated systems (modeling and control)
site B visual data site A control flow force multi-sensor control 2D trajectory 2D ultrasound images Objectives: medical robotics • Modeling adequate ultrasound features and their interaction • Automatic control of the probe motion to assist medical examination • Automatically follow an organ of interest along the patient skin • Hybrid force/vision control schemes • Remote examination without using haptic device • Robot control combining ultrasound images, force measurement and visual data of the patient provided by a remote camera • Autonomous exploration of a given area (organ, tumor)
Objectives: real-time visual tracking • New camera models • Omnidirectional cameras (3D model-based tracking) • Model-based vs model-free approaches • Structure estimation • Joint estimation of pose and structure “a la Debevec” • Model with some degrees of freedom following the work with articulated object • On line structure estimation during visual servoing • Joint estimation of depth and displacement (controlled SLAM) • Initialization • Object detection, recognition and localization • Image-based model of the considered object (collaboration with Vista and EPFL through FP6 Pegase proposal)