680 likes | 953 Views
Real-Time Tracking. Axel Pinz Image Based Measurement Group EMT – Institute of Electrical Measurement and Measurement Signal Processing TU Graz – Graz University of Technology http://www.emt.tugraz.at/~tracking http://www.emt.tugraz.at/~pinz axel.pinz@tugraz.at. Defining the Terms.
E N D
Real-Time Tracking Axel Pinz Image Based Measurement Group EMT – Institute of Electrical Measurement and Measurement Signal Processing TU Graz – Graz University of Technology http://www.emt.tugraz.at/~tracking http://www.emt.tugraz.at/~pinz axel.pinz@tugraz.at
Defining the Terms • Real-Time • Task dependent, “in-the-loop” • Navigation: “on-time” • Video rate: 30Hz • High-speed tracking: several kHz • Tracking • DoF: Degrees of Freedom • 2D: images, videos 2 / 3 DoF • 3D: scenes, object pose 6 DoF
Applications • Surveillance • Augmented reality • Surgical navigation • Motion capture (MoCap) • Autonomous navigation • Telecommunication • Many industrial applications
Example: Augmented Reality [ARToolkit, Billinghurst, Kato, Demo at ISAR2000, Munich] http://www.hitl.washington.edu/research/shared_space/download/
AgendaStructure of the SSIP Lecture • Intro, terminology, applications • 2D motion analysis • Geometry • 3D motion analysis • Practical considerations • Existing systems • Summary, conclusions
2D Motion Analysis • Change detection • Can be anything (not necessarily motion) • Optical flow computation • What is moving in which direction ? • Hard in real time • Data reduction required ! • Interest operators • Points, lines, regions, contours • Modeling required • Motion models, object models • Probabilistic modeling, prediction
Change Detection [Pinz, Bildverstehen, 1994]
Optical Flow (1) [Brox, Bruhn, Papenberg, Weickert] ECCV04 best paper award • Estimating the displacement field • Assumptions: • Gray value constancy • Gradient constancy • Smoothness • ... • Error Minimization
Optical Flow (2) [Brox, Bruhn, Papenberg, Weickert] ECCV04 best paper award !! Not in real-time !!
Interest Operators • Reduce the amount of data • Track only salient features • Support region – ROI (region of interest) Feature in ROI: Edge / Line Corner Blob Contour
2D Point Tracking[Univ. Erlangen, VAMPIRE, EU-IST-2001-34401] • Corner detection Initialization • Calculate “cornerness” c • Threshold sensitivity, # of corners • E.g.: “Harris” / “Plessey” corners in ROI • Cross-correlation in ROI
2D Point Tracking[Univ. Erlangen, VAMPIRE, EU-IST-2001-34401]
Edge Tracking[Rapid 95, Harris, RoRapid 95, Armstrong, Zisserman]
CONDENSATION (2) • CONditional DENSity propagATION • Requires a good initialization • Works with active contours • Maintains / adapts a contour model • Can keep more than one hypothesis
AgendaStructure of the SSIP Lecture • Intro, terminology, applications • 2D motion analysis • Geometry • 3D motion analysis • Practical considerations • Existing systems • Summary, conclusions
Geometry • Having motion in images: • What does it mean? • What can be measured? • Projective camera • Algebraic projective geometry • Camera calibration • Computer Vision • Reconstruction from uncalibrated views • There are excellent textbooks [Faugeras 1994, Hartley+Zisserman 2001, Ma et al. 2003]
Projective Camera (1) • Pinhole camera model: • p = (x,y)Tis the image ofP = (X,Y,Z )T • (x,y) ... image-, (X,Y,Z) ... scene-coordinates • o ... center of projection • (x,y,z) ... camera coordinate system • (x,y,-f) ... image plane x x P(X,Y,Z) f o z y y Z X p(x,y) Y
Projective Camera (2) • Pinhole camera model: • If scene- = camera-coordinate system X f o Z x
Projective Camera (3) • Frontal pinhole camera model: • (x,y,+f) ... image plane • Normalized camera: f=+1 P(X,Y,Z) f x x o p z y y Z X Y
Projective Camera (4) • “real” camera: • 5 Intrinsic parameters (K) • Lens distortion • 6 Extrinsic parameters (M: R, t) • … arbitrary scale
Algebraic Projective Geometry [Semple&Kneebone 52] • Homogeneous coordinates • Duality points lines • Homography H describes any transformation • E.g.: image image transform: x’ = Hx • All transforms can be described by 3x3 matrices • Combination of transformations: Matrix product Translation Rotation
Camera Calibration (1) • Recover the 11 camera parameters: • 5 Intrinsic parameters (K: fsx, fsy, fs, u0, v0) • 6 Extrinsic parameters (M: R, t) • Calibration target: • At least 6 point correspondences • System of linear equations • Direct (initial) solution for K and M
Camera Calibration (2) • Iterative optimization • K, M, lens distortion • E.g. Levenberg-Marquart • Practical solutions require more points • Many algorithms [Tsai 87, Zhang 98, Heikkilä 00] • Overdetermined systems • Robustness against outliers • E.g. RANSAC • Refer to [Hartley, Zisserman, 2001]
What can be measured ... • … with a calibrated camera • Viewing directions • Angles between viewing directions • 3D reconstruction: more than 1 view required • … with uncalibrated camera(s) • Computer Vision research of the past decade • Hierarchy of geometries: Projective – oriented projective – affine – similarity – Euclidean
AgendaStructure of the SSIP Lecture • Intro, terminology, applications • 2D motion analysis • Geometry • 3D motion analysis • Practical considerations • Existing systems • Summary, conclusions
3D Motion Analysis:Location and Orientation head coord. system R t scene coord. system 6 DoF pose in real-time Extrinsic parameters in real-time
3D Motion Analysis: • Tracking technologies, terminology • Camera pose (PnP) • Stereo, E, F, epipolar geometry • Model-based tracking • Confluence of 2D and 3D • Fusion • Kalman Filter
Tracking Technologies (1) • Mechanical tracking • “Magnetic tracking” • Acoustic – time of flight • “Optical” vision-based • Compass • GPS, … External effort required ! No „self-contained“ system [Allen, Bishop, Welch. Tracking: Beyond 15 minutes of thought. SIGGRAPH’01]
Tracking Technologies (2)Examples [Allen, Bishop, Welch. Tracking: Beyond 15 minutes of thought. SIGGRAPH’01]
Research at EMT:Hybrid Tracking – HT Combine 2 technologies: • Vision-based • Good results for slow motion • Motion blur, occlusion, wrong matches • Inertial • Good results for fast motion • Drift, noise, long term stability • Fusion of complementary sensors ! Mimicks human cognition !
Vision-Based Tracking MoreTerminology • Measure position and orientation in real-time • Obtain trajectories of object(s) • Moving observer, egomotion – “inside-out” • Stationary observer – “outside-in Tracking” • Combinations of the above • Degrees of Freedom – DoF • 3 DoF (mobile robot) • 6 DoF (head tracking in AR)
Inside-out Tracking • monocular • exterior parameters • 6 DoF from 4 points • wearable, fully mobile corners blobs natural landmarks
Outside-in Tracking stereo-rig IR-illumination • no cables • 1 marker/device: • 3 DoF • 2 markers: 5 DoF • 3 markers: 6 DoF devices
Camera Pose Estimation • Pose estimation: Estimate extrinsic parameters from known / unknown scene find R, t • Linear algorithms [Quan, Zan, 1999] • Iterative algorithms [Lu et al., 2000] • Point-based methods • No geometry, just 3D points • Model-based methods • Object-model, e.g. CAD
PnP(1)Perspective n-Point Problem • Calibrated camera K, C = (KKT)-1 • n point correspondences scene image • Known scene coordinates of pi, and known distances dij= ||pi – pj || • Each pair (pi,pj) defines an angle • can be measured (2 lines of sight, calibrated camera) constraint for the distance ||c –pi|| pi pj dij xi xj c
PnP (2) pi pj dij xi xj c uj ui
PnP (3) • P3P, 3 points: underdetermined, 4 solutions • P4P, 4 points: overdetermined, 6 equations, 4 unknowns 4 x P3P, then find a common solution • General problem: PnP, n points
PnP (4) Once the xihave been solved: • project image points scene p’i= xi K-1ui • find a common R, t for p’i pi (point-correspondences solve a simple system of linear equations)
Stereo Reconstruction z Elementary stereo geometry in “canonical configuration” 2 h … “baseline” b Pr - Pl … “disparity” d There is just column disparity Depth computation: P(x,y,z) Pz y Cl Cr x f Pl Pr h h xl=0 x=0 xr=0
Stereo (2) • 2 cameras, general configuration: Epipolar geometry X Y ur ul v ll lr Cl el er Cr
Stereo (3) • Uncalibrated cameras: Fundamental matrix F • Calibrated cameras: Essential matrix E • 3x3, Rank 2 • Many Algorithms • Normalized 8-point [Hartley 97] • 5-point (structure and motion) [Nister 03]
Model-Based Tracking Confluence of 2D and 3D [Deutscher, Davison, Reid 01]
3D Motion Analysis: • Tracking technologies, terminology • Camera pose (PnP) • Stereo, E, F, epipolar geometry • Model-based tracking • Confluence of 2D and 3D • Fusion • Kalman Filter • General considerations • Kalman Filter • EMT HT project
General Considerations • We have: • Several sensors (vision, inertial, ...) • Algorithms to deliver pose streams for each sensor (at discrete times; rates may vary depending on sensor, processor, load, ...) • Thus, we need: • Algorithms for sensor fusion (weighting the confidence in a sensor, sensor accuracy, ...) • Pose estimation including a temporal model [Allen, Bishop, Welch. Tracking: Beyond 15 minutes of thought. SIGGRAPH’01]
Sensor Fusion • Dealing with ignorance: • Imprecision, ambiguity, contradiction, ... • Mathematical models • Fuzzy sets • Dempster-Shafer evidence theory • Probability theory • Probabilistic modeling in Computer Vision • The topic of this decade ! • Examples: CONDENSATION, mean shift
Kalman Filter (1) [Welch, Bishop. An Introduction to the Kalman Filter. SIGGRAPH’01] http://www.cs.unc.edu/~welch/kalman
Kalman Filter (2) • Estimate a process • with a measurement xn ... State of the process zm ... Measurement p, v ... Process and measurement noise (zero mean) A ... n x n Matrix relates the previous with the current time step B ... n x l Matrix relates optional control input u to x H ... n x m Matrix relates state x to measurement z