480 likes | 634 Views
Introduction to Collaborative Web . Sergej Zerr zerr@L3S.de. 23. April 2013. 1. Outline. Collaborative Advantages Wisdom of crowds Conditions for a successful collaboration Using Collaborative Data Gathering Data from Social Web / Mechanical Turk Inter Rater Agreement
E N D
Introduction to Collaborative Web . Sergej Zerr zerr@L3S.de 23. April 2013 1
Outline • Collaborative Advantages • Wisdom of crowds • Conditions for a successful collaboration • Using Collaborative Data • Gathering Data from Social Web / Mechanical Turk • Inter Rater Agreement • Search and Sensemaking processes, the overview • Collaboration opportunities in (Web) search • Overview Papers • Collaborative Search, Search result relevance judgments,.. • Small experiment • Can we collaborate? • Discussion Sergej Zerr 2
Collaboration Oftenweneedmorethanonehand Sometimesmorethanonebrain “Why the Many Are Smarter Than the Few and How Collective Wisdom Shapes Business, Economies, Societies and Nations” James Suroewicki Sergej Zerr 3
Direct Collaboration in WWW can be used: Collaborative tagging, Favorite assignments, Click logs, Data gathering, Reccomendations, ect., ect.,ect…. Tags: Rainbow, Sea, Island, Green, Palmtree, Maui Sergej Zerr 4
Social Computer Expert Masses Equivalent, or greater, utility under the curve Utility # ofcontributors 10 10,000+ Sergej Zerr 5
Social Computer Expert $$$$ Masses $ Utility 100,000 amateurs 1.6 Million articles 5 years to develop Real-Time Updates 4,000 experts 80,000 articles 200 years to develop Annual Updates # ofcontributors 10 10,000+ Sergej Zerr 6
Collaboration: Paradox Collaborative workneedstobemanaged efficiently http://spectrum.ieee.org/computing/hardware/multicore-is-bad-news-for-supercomputers Kasparovwonagainsttheworld Sergej Zerr 7
Collaboration Sergej Zerr 8
http://en.wikipedia.org/wiki/Groupthink Groupthink Symptoms: Irving Lester Janis (26 May 1918 - 15 November 1990) • Direct pressure on dissenters • Self-censorship • Illusion of unanimity • Self-appointed ‘mindguards’ • Illusion of invulnerability • Collective rationalization • Belief in inherent morality • Stereotyped views of out-groups Sergej Zerr 9
Collaboration “The best collective decisions are the product of disagreement and contest, not consensus or compromise.” “The best way for a group to be smart is for each person in it to think and act as independently as possible.” Sergej Zerr 10
Many Complementary Games • ESP Game: label images Image retrieval by text • Squigl: match the labels to areas Object recognition • Matchin: find the better image Image ranking • FlipIt: memory with similar images • Near duplicate detection • Other areas covered as well: label songs, find synonyms, describe videos • See: www.gwap.com by Luis von Ahn Sergej Zerr 11
Re-Using Human Power Sergej Zerr 12
Outline • Collaborative Advantages • Wisdom of crowds • Conditions for a successful collaboration • Using Collaborative Data • Gathering Data from Social Web / Mechanical Turk • Inter Rater Agreement • Search and Sensemaking processes, the overview • Collaboration opportunities in (Web) search • Overview Papers • Collaborative Search, Search result relevance judgments,.. • Small experiment • Can we collaborate? • Discussion Two different images that share the same ESP labels: man and woman Sergej Zerr 13
Social Computer • Human can (yet) solve some tasks • more efficient and/or accurate • as a machine would do. • Captcha • Classification (OCR) • Image tagging • Speech recognition Sergej Zerr 14
Social Computer – Good Question „Mr. Burns saw Homer withthebinoculars“ Matthew Lease and Omar Alonso: http://de.slideshare.net/mattlease/crowdsourcing-for-search-evaluation-and-socialalgorithmic-search Sergej Zerr 15
Social Computer – Good Interface • Throw the coin and tell us the result • Head? • Tail? • Results • Head 61 • Tail 39 • People often tend just to select the first option • Better: Some preliminary textual answer • Coin type? • Head or tail. Matthew Lease and Omar Alonso: http://de.slideshare.net/mattlease/crowdsourcing-for-search-evaluation-and-socialalgorithmic-search Sergej Zerr 16
L3S Research – “What about image privacy?” Sergej Zerr , Stefan Siersdorfer , Jonathon Hare , Elena Demidova Privacy-Aware Image Classificationand Search , SIGIR‘12 Sergej Zerr 17
L3S Research • Gathering average community notion of privacy • We crawled “most recently uploaded” Flickr photos (2 Months) • Started a social annotation game (over the course of 2 weeks) • 81 users (colleagues, social networks , forum users) , 6 teams Sergej Zerr , Stefan Siersdorfer , Jonathon Hare , Elena Demidova Privacy-Aware Image Classificationand Search , SIGIR‘12 Sergej Zerr 18
Sampling Data from Social Web: Tags 1-20 tags per photo >75 tags per photo Sergej Zerr 19
Sampling Data from Social Web: User activity Simple random sample canresult in a setdominatedbyfew power user Sergej Zerr 20
Inter–Rater Reliability: What about ambiguity? Nederland, netherlands, holland, dutch Rotterdam, wielrennen, cycling, duck le grand depart, tour de france, Reklame, caravan, Funny Fotos Sergej Zerr 21
Inter–Rater Reliability “Where is the cat?” V V X V V X X X X X X X X X V V X V V V X X X X Results Sergej Zerr 22
Inter–Rater Reliability “Where is the cat?” • Ideas • Introduce „Honeypot“ – a setofgroundtruthobjectsandselectonlygoodraters • Select onlyraterswho rate closetoaveragerating. • Other startegies? V V X V V X X X X X X X X X V V X V V V X X X X Results Sergej Zerr 23
Inter–Rater Reliability: A B Naive approach: 3 cases out of 6 = 0.5 agreement Kilem L. Gwet, Handbook of inter-raterreliability 2010 Sergej Zerr 24
Inter–Rater Reliability: Cohen’s Kappa(1960) Idea: Weneedtoremoveagreementachived just bychance Kilem L. Gwet, Handbook of inter-raterreliability 2010 Sergej Zerr 25
Inter–Rater Reliability: Missing Values Idea: Use partial ratingstoestimate marginal probabilityonly Kilem L. Gwet, Handbook of inter-raterreliability 2010 Sergej Zerr 26
Inter–Rater Reliability: Extentions • Multiple Raters/Categories: • Fleiss 1971 – Averageoverrandompairsofratersforrandomobjects • AdjustmentforOrdinalandInterval Data, Weighting: • weightjudgementsusingdistancesbetweencategories. • ReduceKappa Paradox: • Split judgementsintomorecertain / marginal. • Measures: AC1, AC2(ordinalandintervaldata) • Check forstatisticalsignificance: • The numberofcategoriesand/orratersmatters. Kilem L. Gwet, Handbook of inter-raterreliability 2010 Sergej Zerr 27
Inter–Rater Reliability: Kappa Interpretations Pleasenote: These interpretationswereproventobeusefullmostly in medicaldomain(diagnosis) Kilem L. Gwet, Handbook of inter-raterreliability 2010 Sergej Zerr 28
Useful human power for annotating the Web • 5000 people playing simultaneously can label all images on Google in 30 days! • Individual games in Yahoo! and MSN average over 5,000 players at a time • Possible contributions: attach labels to images in other languages, categorize web pages into topics Sergej Zerr 29
Outline • Collaborative Advantages • Wisdom of crowds • Conditions for a successful collaboration • Using Collaborative Data • Gathering Data from Social Web / Mechanical Turk • Inter Rater Agreement • Search and Sensemaking processes, the overview • Collaboration opportunities in (Web) search • Overview Papers • Collaborative Search, Search result relevance judgments,.. • Small experiment • Can we collaborate? • Discussion Sergej Zerr 30
What is relevant? Sergej Zerr 31
Search and Sensemaking Process Query Annotation/Organization Resultlist Can collaborationimprovethesensemakingprocessatanystep? Do youusecollaborativesearch? Sergej Zerr 32
What are the typical collaborative search tasks? • Watched over someone’s shoulder as he/she searched the Web, and suggested alternate query terms. • E-mailed someone links to share the results of a Web search. • E-mailed someone a textual summary to share the results of a Web search. • Called someone on the phone to tell them about the results of a Web search. Morris, M.R. A Survey of Collaborative Web Search Practices. In Proceedings of 26th CHI Conference 2008 Around 90% of Microsoft employees are engaged in collaborative search activities. Morris, M.R. A Survey of Collaborative Web Search Practices. In Proceedings of 26th CHI Conference 2008 Sergej Zerr 33
Collaborative Search: • Query formulation • Search process/browsing • Save/Bookmark • Annotate/Organize • Howtosupportusers in collaborativesearching? • Ideas • Tools(Web 2.0) Sergej Zerr 34
Indirect Collaboration on Web: Google uses search/click logs. For PageRank algorithm each link to a page serves as a vote for that page. Amazon uses search/click logs. For item recommendation similar users are the indirect voters for the product. ect., ect. Sergej Zerr 35
Outline • Collaborative Advantages • Wisdom of crowds • Conditions for a successful collaboration • Using Collaborative Data • Gathering Data from Social Web / Mechanical Turk • Inter Rater Agreement • Search and Sensemaking processes, the overview • Collaboration opportunities in (Web) search • Overview Papers • Collaborative Search, Search result relevance judgments,.. • Small experiment • Can we collaborate? • Discussion Sergej Zerr 36
Software support for Co-located search (CoSearch). Amershi, S., Morris, M. CoSearch: System for colocated collaborative Web search. In Proceedings of 26th CHI Conference 2008 Sergej Zerr 37
Software support for Co-located search (CoSearch). Amershi, S., Morris, M. CoSearch: System for colocated collaborative Web search. In Proceedings of 26th CHI Conference 2008 Sergej Zerr 38
Spatial distributed Search (SearchTogether) • (a) integrating messaging • (b) query awareness, • (c) current results • (d) recommendation queue • (e)(f)(g) search buttons • (h) page-specific metadata • (i) toolbar • (j) browser Morris, M.R. & Horvitz, E. SearchTogether: An Interface for Collaborative Web Search. In Proceedings of the UIST 2007 Sergej Zerr 39
Spatial distributed Search (SearchTogether) • 38% af all result lists were the consequence of using history • 70 positive and 9 negative ratings • 22 of 36 recommendation were viewed by the recipients Morris, M.R. & Horvitz, E. SearchTogether: An Interface for Collaborative Web Search. In Proceedings of the UIST 2007 Sergej Zerr 40
Improvement Through Collaborative Ranking Agrahri, A., Manickam, D., Riedl, J. Can people collaborate to improve the relevance of search results? In Proceedings of the ACM International Conference on Recommendation Systems 2008 Sergej Zerr 41
WeSearch: Collaborative Sensemaking (Hardware support) Meredith Ringel Morris, Jarrod Lombardo, and Daniel Wigdor. 2010. WeSearch: supporting collaborative search and sensemaking on a tabletop display. In Proceedings of the 2010 ACM conference on Computer supported cooperative work (CSCW '10). ACM, New York, NY, USA, 401-410. Sergej Zerr 42
WeSearch: Collaborative Sensemaking Meredith Ringel Morris, Jarrod Lombardo, and Daniel Wigdor. 2010. WeSearch: supporting collaborative search and sensemaking on a tabletop display. In Proceedings of the 2010 ACM conference on Computer supported cooperative work (CSCW '10). ACM, New York, NY, USA, 401-410. Sergej Zerr 43
Hardware support for Co-located(TeamSearch). • Circles are categories: people, location, year, event Morris, M.R., Paepcke, A., and Winograd, T. Team-Search: Comparing Techniques for Co-Present Collaborative Search of Digital Media. Sergej Zerr 44
Hardware support for Co-located(TeamSearch). Morris, M.R., Paepcke, A., and Winograd, T. Team-Search: Comparing Techniques for Co-Present Collaborative Search of Digital Media. Sergej Zerr 45
Outline • Collaborative Advantages • Wisdom of crowds • Conditions for a successful collaboration • Using Collaborative Data • Gathering Data from Social Web / Mechanical Turk • Inter Rater Agreement • Search and Sensemaking processes, the overview • Collaboration opportunities in (Web) search • Overview Papers • Collaborative Search, Search result relevance judgments,.. • Small experiment • Can we collaborate? • Discussion Sergej Zerr 46
Smalltalk, what should be the properties of a perfect collaborative system? Sergej Zerr 48