430 likes | 483 Views
Welcome to 18-491 Fundamentals of Signal Processing! This course covers mechanics, content review, and material preview. Grades are based on homework and exams. Textbook: "Discrete-Time Signal Processing." Course includes time and frequency domain processing, digital filter design, and various application areas. Learn about signal representation and filtering techniques. Take care of yourself and seek help if needed. Course topics encompass speech processing, image processing, and more. Join us for an informative journey in signal processing!
E N D
INTRODUCTION TO 18-491FUNDAMENTALS OF SIGNAL PROCESSING Richard M. Stern 18-491 lecture January 14, 2019 Department of Electrical and Computer Engineering Carnegie Mellon University Pittsburgh, Pennsylvania 15213
Welcome to 18-491 Fundamentals of Signal Processing (DSP)! • Today will • Review mechanics of course • Review course content • Preview material in 18-491 (DSP)
Important people (for this course at least) • Instructor: Richard Stern • PH B24, 8-2535, rms@cs.cmu.edu • Course management assistant: Michelle Mahouski • HH 1112, 8-4951, mmahousk@andrew.cmu.edu
More important people • Teaching interns: • Tyler Supradeep VuongRangarajan
Some course details • Meeting time and place: • Lectures here and now • Recitations Friday 10:30 – 12:20, 12:30 – 2:20, SH 214 • Pre-requisites (you really need these!): • Signals and Systems 18-290 • Some MATLAB or background (presumably from 18-290)
Does our work get graded? • Yes! • Grades based on: • Homework (including MATLAB problems) (33%) • Three exams (67%) • Two midterms (March 6 and April 3), and final exam • Plan on attending the exams!
Textbooks • Major text: • Oppenheim, Schafer, Yoder, and Padgett: Discrete-Time Signal Processing • Plan on purchasing a hard copy new or used • Material to be supplemented by class notes at end of course • Some other texts listed in syllabus
Other support sources • Office hours: • Two hours per week for instructor and each TA, times TBA • You can schedule additional times with me as needed • Course home page: • http://www.ece.cmu.edu/~ece491 • Canvas to be used for grades (but probably not much else) • Piazza to be used for class discussions
Academic stress and sources of help • This is a hard course • Take good care of yourself • If you are having trouble, seek help • Teaching staff • CMU Counseling and Psychological Services (CaPS) • We are here to help!
Academic integrity (i.e. cheating and plagiarism) • CMU’s take on academic integrity: • http://www.cmu.edu/policies/documents/Cheating.html • ECE’s take on academic integrity: • http://www.ece.cmu.edu/programs-admissions/masters/academic-integrity.html • Most important rule: Don’t cheat! • But what do we mean by that? • Discussing general strategies on homework with other students is OK • Solving homework together is NOT OK • Accessing material from previous years isNOT OK • “Collaborating” on exams is REALLY REALLYNOT OK!
18-491: major topic areas • Signal processing in the time domain: convolution • Frequency-domain processing: • The DTFT and the Z-transform • Complementary signal representations • Sampling and change of sampling rate • The DFT and the FFT • Digital filter implementation • Digital filter design • Selected applications
Complementary signal representations • Unit sample response • Discrete-time Fourier transforms • Z-transforms • Difference equations • Poles and zeros of an LSI system
Some application areas (we may not get to all of these) • Linear prediction and lattice filters • Adaptive filtering • Optimal Wiener filtering • Two-dimensional DSP (image processing) • Short-time Fourier analysis • Speech processing
Signal representation: why perform signal processing? • A speech waveform in time: “Welcome to DSP I”
A time-frequency representation of “welcome” is much more informative
Downsampling the waveform Downsampling the waveform by factor of 2:
Consequences of downsampling by 2 Original: Downsampled:
Upsampling the waveform Upsampling by a factor of 2:
Consequences of upsampling by 2 Original: Upsampled:
Linear filtering the waveform y[n] x[n] Filter 1: y[n] = 3.6y[n–1]+5.0y[n–2]–3.2y[n–3]+.82y[n–4] +.013x[n]–.032x[n–1]+.044x[n–2]–.033x[n–3]+.013x[n–4] Filter 2: y[n] = 2.7y[n–1]–3.3y[n–2]+2.0y[n–3]–.57y[n–4] +.35x[n]–1.3x[n–1]+2.0x[n–2]–1.3x[n–3]+.35x[n–4]
Output of Filter 1 in the frequency domain Original: Lowpass:
Output of Filter 2 in the frequency domain Original: Highpass:
Let’s look at the lowpass filter from different points of view … y[n] x[n] Difference equation for Lowpass Filter 1: y[n] = 3.6y[n–1]+5.0y[n–2]–3.2y[n–3]+.82y[n–4] +.013x[n]–.032x[n–1]+.044x[n–2]–.033x[n–3]+.013x[n–4]
Lowpass filtering in the time domain: the unit sample response
Lowpass filtering in the frequency domain: magnitude and phase of the DTFT
Pitch Pulse train source Vocal tract model Noise source Another type of modeling: the source-filter model of speech A useful model for representing the generation of speech sounds: Amplitude p[n]
Signal modeling: let’s consider the “uh” in “welcome:”
An application of LPC modeling: separating the vocal tract excitation and and filter Original speech: Speech with 75-Hz excitation: Speech with 150 Hz excitation: Speech with noise excitation: Comment: this is a major techniques used in speech coding
Classical signal enhancement: compensation of speech for noise and filtering • Approach of Acero, Liu, Moreno, et al. (1990-1997)… • Compensation achieved by estimating parameters of noise and filter and applying inverse operations “Clean” speech Degraded speech x[m] h[m] z[m] Linear filtering n[m] Additive noise
“Classical” combined compensation improves accuracy in stationary environments • Threshold shifts by ~7 dB • Accuracy still poor for low SNRs Complete retraining –7 dB 13 dB Clean VTS (1997) Original CDCN (1990) “Recovered” CMN (baseline)
Another type of signal enhancement: adaptive noise cancellation • Speech + noise enters primary channel, correlated noise enters reference channel • Adaptive filter attempts to convert noise in secondary channel to best resemble noise in primary channel and subtracts • Performance degrades when speech leaks into reference channel and in reverberation
Simulation of noise cancellation for a PDA using two mics in “endfire” configuration • Speech in cafeteria noise, no noise cancellation • Speech with noise cancellation • But…. simulation assumed no reverb
Signal separation: speech is quite intelligible, even when presented only in fragments • Procedure: • Determine which time-frequency time-frequency components appear to be dominated by the desired signal • Reconstruct signal based on “good” components • A Monaural example: • Mixed signals - • Separated signals -
Practical signal separation: Audio samples using selective reconstruction based on ITD RT60 (ms) 0 300 No Proc Delay-sum ZCAE-bin ZCAE-cont
Phase vocoding: changing time scale and pitch • Changing the time scale: • Original speech • Faster by 4:3 • Slower by 1:2 • Transposing pitch: • Original music • After phase vocoding • Transposing up by a major third • Transposing down by a major third Comment: this is one of several techniques used to perform autotuning
Summary • Lots of interesting topics that teach us how to understand signals and design filters • An emphasis on developing a solid understanding of fundamentals • Will introduce selected applications to demonstrate utility of techniques • I hope that you have as much fun in signal processing as I have had!