260 likes | 362 Views
Effective Time Ratio: A measure for Web search engine with document snippet. Outline. Introduction Effective Time Ratio Experiment and Results Conclusion. Existing IR Measures. search engine quality. doc relevance. Snippets. Snippets : False Positive.
E N D
Effective Time Ratio:A measure for Web searchengine with document snippet
Outline • Introduction • Effective Time Ratio • Experiment and Results • Conclusion
Existing IR Measures search engine quality doc relevance
Snippets:False Positive Waste time examining irrelevant document!
Snippets:False Negative Miss really relevant document!
Outline • Introduction • Effective Time Ratio • Experiment and Results • Conclusion
IntuitionReduce time wasted on examining irrelevant information
Effective Time Ratio (ETR) of getting relevant information of total search
Precision:a special case of ETR • Assumption: constant time(T) on each doc P@5 = 0.4 of getting rel information = 2T of total search = 5T ETR@5 = 0.4
User Behavior Assumptions • Examination Assumption[Dupret+08,Craswell+08, Guo+09a,Guo+09b] • Cascade Assumption[Craswell+08, Guo+09a,Guo+09b, Chapelle+09] • Examination Time Assumption • Snippet Examination Time: T1 • Doc Examination Time: T2 Position i Position (i+1)
ETR for Search Engine with Snippets of getting rel information T1 + T2 of total search 5T1 + 3T2
Factors affecting ETR • Retrieval System Performance(P@N) • Snippet Quality • First Type Error(False Positive) • Second Type Error(False Negative) P1 = Pr( | ) P2 = Pr( | )
Expected ETR • C = T1/T2
Properties • It helps (increases ETS) by using perfect snippets (with no first/second type error)
Properties • ETR can reflect snippet generation/retrieval system quality
Outline • Introduction • Effective Time Ratio • Experiment and Results • Conclusion
User Study Data • 10 Users • 50 Questions from commercial query log • 25 open + 25 close • Various topics/difficulty level • For each question, 100 top results(snippet + doc) from a search engine
User Study • Each user examines the results to answer the question • Write the answer • Report the satisfaction degree • Three assessors judges each snippet/doc as relevant/irrelevant
Measure Validation • Correlation[Huffman+07] between • User reported satisfaction • Measure value based on judgments • Higher correlation indicates the measure can reflect user satisfaction better
Measures • Traditional IR measures • P@N, RR, DCG, CP • Version of measure including snippet quality [Turpin+09] • P’@N, RR’, DCG’, CP’ • ETR and its cumulated version • ETR and CETR
Results • Up-bound: 0.706 (avg correlation between users) • Traditional Measures • Measures including snippet quality[Turpin+09] • ETR and its cumulated version
Results • Open Questions • Close Questions
Conclusion • Search engine quality is also affected by snippets • ETR can reflect both retrieval system and snippet generation algorithm quality • ETR can reflect user satisfaction better than existing measures