1 / 39

1. Introduction:

How Does auditory perception organization works ? by Elvira Perez and Georg Meyer Dept. Psychology, Liverpool University, UK Hoarse Meeting, Chrysler Ulm, Germany 28 th -30 th October, 2004. 1. Introduction:. Ears receive mixtures of sounds.

Download Presentation

1. Introduction:

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. How Does auditory perception organization works?by Elvira Perez and Georg MeyerDept. Psychology, Liverpool University, UKHoarse Meeting, Chrysler Ulm, Germany 28th-30th October, 2004

  2. 1. Introduction: Ears receive mixtures of sounds. We can tolerate surprisingly high levels of noise and still orientate our attention to whatever we want to attend. But... how the auditory system can do this so accurately?

  3. 1. Introduction: • Auditory scene analysis (Bregman, 1990) is a theoretical framework that aims to explain auditory perceptual organisation. • Basics: • Environment contains multiples objects • Decomposition into its constituent elements. • Grouping. • It proposes two grouping mechanisms: • 1. ‘Bottom-up’: Primitive cues (F0, intensity, location) Grouping mechanism based on Gestalt principles. • 2. ‘Top-down’: Schema-based (speech pattern matching)

  4. 1. Introduction: • Primitive process (Gestalt. Koffka, 1935): • Similarity • Good continuation • Common fate • Disjoint locations • Closure

  5. 1. Introduction • Criticisms: Too simplistic. Whatever cannot be explained through the primitive processes, it is explained by the schema-based processes. • Primitive processes only work in the lab. • Sine-wave replicas of utterances (Remez et al., 1992) • Phonetic principles of organization find a single speech stream, whereas auditory principles find several simultaneous whistles. • Grouping by phonetic rather than by simple auditory coherence.

  6. 3. Experiments (baseline): • The purpose of these studies is to explore how noise (a chirp) affects speech perception. • The stimulus used is a vowel-nasal syllable which is perceived as /en/ if presented in isolation but as /em/ if it is presented with a frequency modulated sine wave in the position where the second formant transition would be expected. • In the three experiments participants categorised the synthetic syllable heard as /em/ or /en/. • Direction, duration, and position of the chirp were the values manipulated.

  7. vowel nasal 2700 2000 Formant frequ. (Hz) 800 375 100200ms 3. Experiments The perception of a nasal /n/ change to /m/ when adding a chirp between the vowel and nasal F2

  8. In 80% of the trials the participants heard the difference between up and down chirp. vowel nasal 100 200ms chirp down chirp up Experiment 1 Baseline/Direction

  9. vowel nasal 100 200ms Experiment 2 Duration

  10. vowel nasal Experiment 3 Position 100 200ms

  11. 5. Conclusions: Chirps from 4 ms to 20 ms duration and in a range of 1kHz-2kHz, independently of their direction, are apparently integrated into the speech signal and change the percept from /en/ to /em/. Subjects very clearly hear two objects, so that some scene analysis is taking place since the chirp is not integrated completely into the speech. Duplex perception with one ear. It seems that listeners can also discriminate the direction motion of the chirp when they focus their attention in the chirp and a more high level of auditory processing takes places (80% accuracy).

  12. Mr. Background Noise • Do human listeners actively generate representation of background noise to improve speech recognition? • Hypothesis: Recognition performance should be highest if the spectral and temporal structure of interfering noise is regular so that a good noise model can be generated  unpredictable noise.

  13. Experiment 4 & 5 • Stimuli: chirps + /en/ • Ten subjects • The amplitude of the chirp vary (5 conditions: 0dB -8dB -14dB -20dB no-chirp) • Background noise (down chirps): • Quantity: Lots (170/20ms) vs Few (19/20ms) • Time of appearance: Regular vs Irregular • Categorization task 2FC. • Threshold shifts.

  14. Regular condition en en en en en en Irregular condition en en en en en en

  15. Each point in the scatter is the mean threshold over all subjects for a give session. The solid lines show the Boltzmann fit (Eq.(1) for each individual subject in the fifth different conditions. All the fits have the same upper and lower asymptotes.

  16. Exp. 4 rand/reg lots vs. few (t = -3.34, df = 38, p = 0.001). control vs. lots (t = -3.34, df = 38, p = 0.001). No effect between irregular and regular.

  17. Two aspects change from exp. 4 to 5: • Amplitude scale of the chirps (0dB -4dB -8dB -16dB no-chirp). • The conditions lots now includes 100/20’’ and before 170/20’’.

  18. Exp.5 rand/reg lots vs few (t = 2.27, df = 38, p = 0.05). control vs. lots (t = 3.12, df = 38, p < 0.05). No effect between irregular and regular.

  19. 5. Conclusions Only the amount of background noise seems to affect the performance of the recognition task. The regularity of the background noise seems an irrelevant cue to improve auditory stream segregation and therefore, speech perception. Counterintuitive phenomenon.

  20. Irrelevant sound effect (ISE) (Colle & Welsh, 1976) disrupts in serial recall. • The level of meaning (reverse vs forward speech), predictability of the sequence (random vs regular), and similarity (semantic or physical) of the IS to the target material, seems to have little impact in the focal task. (Jones et al., 1990). • Changing state: The degree of variability orphysical change within an auditory stream is the primary determinant of the degree of distrupion in the focal task.

  21. Frequency Frequency Time Time Smooth change Abrupt change

  22. Zoom in Top Down Up Bottom

  23. Experiment 6 • Stimuli: Synthesised vowel-nasal + background FM tone + Chirps • Three blocks (200trial each block): First Control, second Smooth or Abrupt(counterbalanced order) • Chirps: Four different frequencies: Up/Down/Top/Bottom • Five amplitudes: 0dB, -4dB, -8dB, -16dB, no-chirp

  24. Experiment 6 • Subjects: 42 • Musicians vs Non musicians • Female vs Male • Nationality (27) • Age(27.7), languages spoken (3) • Hearing issues (AP, RP, Tinnitus)

  25. down vs up YES control vs. smooth YES control vs abrupt YES No effect between smooth and abrupt. Results Tukey test p < 0.001

  26. Results Musicians vs. Non Musicians …no differences.

  27. More analysis… • Take away the intermediate conditions • Results remain the same • Habituation • Differences between musicians and non musicians; overall, 10, 5, 3 first blocks, but not the first block. • Again control vs smooth/abrupt

  28. 6. Conclusions • It seems that listeners do not use pattern prediction as a cue for auditory perceptual organisation. • … or they do it extremely fast (3ms), or is due to STM (pre-perceptual auditory storage) • Attention must be focused on an object (background noise) for a change in that object to detected (Rensink, et al, 1997) • … or we just ignore the information contained in the transitions for not being reliable.

  29. Conflict area vowel nasal 2700 2000 Formant frequ. (Hz) 800 375

  30. Ignore is not reliable vowel nasal 2700 2000 Formant frequ. (Hz) 800 375

  31. vowel nasal 2700 2000 Formant frequ. (Hz) 800 375 vowel nasal 2700 2000 Formant frequ. (Hz) 800 375 = /em/ = /em/

  32. Which information are we taking in account? • Do we combine cues? • Do we measure the variances associated which each cue to test its reliability? • Maximum likelihood integrator • Antecedents: The nervous system seems to combine visual and haptic info similar to a MLE (Ernst, et al. 2002)

  33. nasal nasal transition transition noise noise

  34. Conditions VowelTransitionNasal e n Full syllable ( ) e n No transition ( ) e n No nasal ( ) e n No transition/nasal

  35. Preliminary results

  36. New Methodology • Until now Method of Constants: Several stimulus levels are chosen beforehand, and groups of observations are placed at each of these stimulus levels. The order of observations is randomized. A conventional method of estimation is used in fitting the psychometric function to the resulting data. • Adaptative procedures: The stimulus level on any one trial is determined by the preceding stimuli and responses. • Sequential experiment: The course of the experiment is dependent on the experimental data.

  37. Up-Down procedures or staircase method: The stimuli level (amplitude of the speech signal) is decreased after a positive response (or increase after a negative). • On each trial, the participant is required to give both a binary judgment (em/en/e) and a confident rating. The binary judgment are used to decide on the direction of change in the stimulus level, and the confident ratings are used to decide on the steps size (dB). • Advantages: Most of the observations are placed at or near the 50% level

  38. Conclusions • …In the next meeting.

  39. Thank you

More Related