1 / 14

Combining Multiple Models for Speech Information Retrieval

Muath Alzghool and Diana Inkpen. Combining Multiple Models for Speech Information Retrieval. University of Ottawa Canada. Presentation Outline. Task: Speech Information Retrieval. Data: Mallach collection (Oard et al, 2004) . System description. Model fusion.

snow
Download Presentation

Combining Multiple Models for Speech Information Retrieval

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Muath Alzghool and Diana Inkpen Combining Multiple Models for Speech Information Retrieval University of Ottawa Canada LREC 2008

  2. Presentation Outline • Task: Speech Information Retrieval. • Data: Mallach collection (Oard et al, 2004). • System description. • Model fusion. • Experiments using model fusion. • Results of the cross-language experiments. • Results of manual keywords and summaries. • Conclusion and future work. LREC 2008

  3. The Mallach collection • Used in the Cross-Language Speech Retrieval (CLSR) task at Cross-Language Evaluation Forum (CLEF) 2007. • 8104 “documents” (segments) from 272 interviews with Holocaust survivors, totaling 589 hours of speech: ASR transcripts with a word error rate of 25-38%. • Additional metadata: automatically-assigned keywords, manually-assigned keywords, and a manual 3-sentence summary. • A set of 63 training topics and 33 test topics, created in English from actual user requests and translated into Czech, German, French, and Spanish by native speakers. • Relevance judgments were generated standard pooling. LREC 2008

  4. Segments LREC 2008

  5. Example topic (English) LREC 2008

  6. Example topic (French) LREC 2008

  7. System Description • SMART: Vector Space Model (VSM). • Terrier: Divergence from Randomness models (DFR). • Two Query Expansion Methods: • Based on thesaurus (novel technique). • Blind relevance feedback (12 terms from the top 15 documents): based on Bose-Einstein 1 model (Bo1 from Terrier). • Model Fusion: sum of normalized weighted similarity scores (novel way to compute weights). • Combined output of 7 machine translation tools. LREC 2008

  8. Model Fusion • Combine the results of different retrieval strategies from SMART (14 runs) and Terrier (1 run). • Each technique will retrieve different sets of relevant documents; therefore combining the results could produce a better result than any of the individual techniques. LREC 2008

  9. Experiments using Model Fusion • Applied the data fusion methods to 14 runs produced by SMART and one run produced by Terrier. • % change is given with respect to the run providing better performance in each combination on the training data. • Model fusion helps to improve the performance (MAP and Recall score) on the test data. • Monolingual (English): 6.5% improvement (notstatistically significant). • Cross-language experiments (French) : 21.7% improvement (significant). LREC 2008

  10. Experiments using Model Fusion (MAP) LREC 2008

  11. Experiments using Model Fusion (Recall) LREC 2008

  12. Results of the cross-language experiments • The cross-language results for French are very close to Monolingual (English) on training data (the difference is not significant), but not on test data (the difference is significant). • The difference is significant between cross-language results for Spanish and Monolingual (English) on training data but not on test data (the difference is not significant). LREC 2008

  13. Results of manual keywords and summaries Experiments on manual keywords and manual summaries showed high improvements comparing to Auto-English. Our results (for manual and automatic runs) are the highest to date on this data collection in CLEF/CLSR. LREC 2008

  14. Conclusion and future work • Model fusion helps to improve the retrieval significantly for some experiments (Auto-French) and for other not significantly (Auto-English). • The idea of using multiple translations proved to be good (based on previous experiments). • Future work • we plan to investigate more methods of model fusion. • Removing or correcting some of the speech recognition errors in the ASR content words. LREC 2008

More Related