520 likes | 551 Views
EE 4780: Introduction to Computer Vision. Introduction. EE 4780. Instructor: Bahadir K. Gunturk Office: EE 225 Email: bahadir@ece.lsu.edu Tel: 8-5621 Office Hours: MW 10:00 – 12:00. EE 4780. We will learn the fundamentals of digital image processing and computer vision.
E N D
EE 4780: Introduction to Computer Vision Introduction
EE 4780 • Instructor: Bahadir K. Gunturk • Office: EE 225 • Email: bahadir@ece.lsu.edu • Tel: 8-5621 • Office Hours: MW 10:00 – 12:00
EE 4780 • We will learn the fundamentals of digital image processing and computer vision. • Lecture slides, problems sets, solutions, study materials, etc. will be posted on the class website. [www.ece.lsu.edu/gunturk/EE4780] • Textbook is not required. • References: • Gonzalez/Woods, Digital Image Processing, Prentice-Hall, 2/e. • Forsyth/Ponce, Computer Vision: A Modern Approach, Prentice-Hall. • Duda, Hart, and Stork, “Pattern Classification,” John Wiley&Sons, 2001. • Shapiro/Stockman, Computer Vision, Prentice-Hall. • Horn, “Robot Vision,” MIT Press, 1986.
Grading Policy • Your grade will be based on • Problem Sets: 30% • Midterm: 30% • Final: 40% • Problem Sets • Mini projects: Theoretical problems and MATLAB assignments • 4-5 Problem Sets • Individually or in two-person teams
Digital Image Acquisition Sensor array • When photons strike, electron-hole pairs are generated on sensor sites. • Electrons generated are collected over a certain period of time. • The number of electrons are converted to pixel values. (Pixel is short for picture element.)
Digital Image Acquisition • Two types of quantization: • There are finite number of pixels. (Spatial resolution) • The amplitude of pixel is represented by a finite number of bits. (Gray-scale resolution)
Matrix Representation of Images • A digital image can be written as a matrix
Bit Depth – Grayscale Resolution 8 bits 7 bits 6 bits 5 bits
Bit Depth – Grayscale Resolution 4 bits 3 bits 2 bits 1 bit
Video = vertical position = horizontal position = frame number
Why do we process images? • To facilitate their storage and transmission • To prepare them for display or printing • To enhance or restore them • To extract information from them • To hide information in them
Image Processing Example • Image Restoration Original image Blurred Restored by Wiener filter
Image Processing Example • Noise Removal Noisy image Denoised by Median filter
Image Processing Example • Image Enhancement Histogram equalization
Image Processing Example • Artifact Reduction in Digital Cameras Original scene Captured by a digital camera Processed to reduce artifacts
Image Processing Example • Image Compression Original image 64 KB JPEG compressed 15 KB JPEG compressed 9 KB
Image Processing Example • Object Segmentation “Rice” image Edges detected using Canny filter
Image Processing Example • Resolution Enhancement
Image Processing Example • Watermarking Original image Watermarked image Generate watermark Hidden message Secret key
Image Processing Example • Face Recognition Search in the database Surveillance video
Image Processing Example • Fingerprint Matching
Image Processing Example • Segmentation
Image Processing Example • Texture Analysis and Synthesis Photo Computer generated Pattern repeated
Image Processing Example • Face detection and tracking http://www-2.cs.cmu.edu/~har/faces.html
Image Processing Example • Face Tracking
Image Processing Example • Object Tracking
Image Processing Example • Virtual Controls
Image Processing Example • Visually Guided Surgery
Cameras • First camera was invented in 16th century. • It used a pinhole to focus light rays onto a wall or translucent plate. • Take a box, prick a small hole in one of its sides with a pin, and then replace the opposite side with a translucent plate. • Place a candle on the pinhole side, you will see an inverted image of the candle on the translucent plate.
Pinhole Camera Model • If the pinhole were really reduced to a point, exactly one light ray would pass through each point in the image plane. • In reality, each point in the image place collects light from a cone of rays. • In addition, real cameras are equipped with lenses. • Still, pinhole model is an acceptable approximation of the imaging process.
Pinhole Cameras Pinhole too big - many directions are averaged, blurring the image Pinhole too small - diffraction effects blur the image
Perspective Projection • Far objects appear smaller than the close ones. Focal point
Perspective Projection • Perspective projection equations
Cameras With Lenses • Most cameras are equipped with lenses. • There are two main reasons for this: • To gather light. For an ideal pinhole, a single light ray would reach each point the image plane. Real pinholes have a finite size, so each point in the image plane is illuminated by a cone of light rays. The larger the hole, the wider the cone and the brighter the image => blurry pictures. Shrinking the pinhole produces sharper images, but reduces the amount of light and may introduce diffraction effects. • To keep the picture in sharp focus while gathering light from a large area.
Geometric Optics • Ignoring diffraction, interferences, etc., the behavior of lenses is dictated by the laws of geometric optics. • Light travels in straight lines in homogeneous media. • When a light ray is reflected from a surface, this ray, its reflection, and the surface normal are coplanar, and the incident and the reflection angles are identical. • When a ray passes from one medium to another, it is refracted. Snell’s Law Refraction indices
Paraxial Geometric Optics • In paraxial (or first-order) geometric optics, the angles between all light rays going through a lens and the normal to lens surfaces are small. • If the angles are small, their sines and tangents are equal to the angles, to the first order.
Thin Lenses • For a thin lens surrounded by a vacuum (refraction index = 1), the formula is where Refraction index of lens
Real Lenses • Better approximation of thick lenses can be obtained:
Real Lenses • Thin lens assumption is not correct. Small angle approximations are not valid. • Rays do not focus at a single point. Spherical aberration Spherical aberration can be eliminated completely by designing aspherical lenses.
Real Lenses • The index of refraction is a function of wavelength. • Light at different wavelengths follow different paths. Chromatic aberration
Real Lenses Chromatic Aberration
Real Lenses • Special lens systems (achromatic doublets) using two or more pieces of glass with different refractive indexes can reduce or eliminate this problem. However, not even these lens systems are completely perfect and still can lead to visible chromatic aberrations, especially at wide angles. • Third-order lens model might be helpful to quantify the aberration.
Real Lenses Stop, put for reducing spherical aberrations • Barrel Distortion & Pincushion Distortion
Real Lenses • Barrel Distortion & Pincushion Distortion Corrected Distorted http://www.vanwalree.com/optics/distortion.html http://www.dpreview.com/learn/?/Image_Techniques/Barrel_Distortion_Correction_01.htm
Real Lenses Vignetting effect in a two-lens system. The shaded part of the beam never reaches the second lens. The brightness drop in the image perimeter.
Real Lenses Optical vignetting example. Left: f/1.4. Right: f/5.6. f-number
Real Lenses Flare Hood may prevent flares