SlideShare ist ein Scribd-Unternehmen logo
1 von 48
Downloaden Sie, um offline zu lesen
23. April 2013 1
Introduction to Collaborative Web .
Sergej Zerr
zerr@L3S.de
Sergej Zerr 2
Outline
• Collaborative Advantages
• Wisdom of crowds
• Conditions for a successful collaboration
•Using Collaborative Data
• Gathering Data from Social Web / Mechanical Turk
• Inter Rater Agreement
• Search and Sensemaking processes, the overview
• Collaboration opportunities in (Web) search
• Overview Papers
• Collaborative Search, Search result relevance judgments,..
•Small experiment
• Can we collaborate?
• Discussion
Sergej Zerr 3
Collaboration
Often we need more than one hand
Sometimes more than one brain
“Why the Many Are Smarter Than the Few and
How Collective Wisdom Shapes Business,
Economies, Societies and Nations”
James Suroewicki
Sergej Zerr 4
Direct Collaboration in WWW can be used:
Collaborative tagging,
Favorite assignments,
Click logs,
Data gathering,
Reccomendations,
ect., ect.,ect….
Tags: Rainbow, Sea, Island, Green, Palmtree, Maui
Sergej Zerr 5
Social Computer
Expert
Utility
Masses
# of contributors10 10,000+
Equivalent, or greater, utility
under the curve
Sergej Zerr 6
Social Computer
Expert $$$$
Utility
Masses $
# of contributors10 10,000+
4,000 experts
80,000 articles
200 years to develop
Annual Updates
100,000 amateurs
1.6 Million articles
5 years to develop
Real-Time Updates
Sergej Zerr 7
Collaboration: Paradox
http://spectrum.ieee.org/computing/hardware/multicore-is-bad-news-for-supercomputers
Collaborative work needs to be managed
efficiently
Kasparov won against the world
Sergej Zerr 8
Collaboration
Criteria Description
Diversity of opinion Each person should have private information.
Independence People's opinions aren't determined by the opinions of
those around them.
Decentralization People are able to specialize and draw on local
knowledge.
Aggregation Some mechanism exists for turning private judgments
into a collective
Sergej Zerr 9
Groupthink Symptoms:
Irving Lester Janis (26 May 1918 - 15 November 1990)
• Illusion of invulnerability
• Collective rationalization
• Belief in inherent morality
• Stereotyped views of out-groups
http://en.wikipedia.org/wiki/Groupthink
• Direct pressure on dissenters
• Self-censorship
• Illusion of unanimity
• Self-appointed „mindguards‟
Sergej Zerr 10
Collaboration
“The best collective decisions are the product of
disagreement and contest, not consensus or compromise.”
“The best way for a group to be smart is for each person in it
to think and act as independently as possible.”
Sergej Zerr 11
Many Complementary Games
• ESP Game: label images
 Image retrieval by text
• Squigl: match the labels to areas
 Object recognition
• Matchin: find the better image
 Image ranking
• FlipIt: memory with similar images
Near duplicate detection
• Other areas covered as well: label songs, find synonyms, describe videos
• See: www.gwap.com by Luis von Ahn
Sergej Zerr 12
Re-Using Human Power
Sergej Zerr 13
Outline
• Collaborative Advantages
• Wisdom of crowds
• Conditions for a successful collaboration
•Using Collaborative Data
• Gathering Data from Social Web / Mechanical Turk
• Inter Rater Agreement
• Search and Sensemaking processes, the overview
• Collaboration opportunities in (Web) search
• Overview Papers
• Collaborative Search, Search result relevance judgments,..
•Small experiment
• Can we collaborate?
• Discussion
Two different images that share the same ESP labels:
man and woman
Sergej Zerr 14
Social Computer
Human can (yet) solve some tasks
more efficient and/or accurate
as a machine would do.
• Captcha
• Classification (OCR)
• Image tagging
• Speech recognition
Sergej Zerr 15
Social Computer – Good Question
Matthew Lease and Omar Alonso: http://de.slideshare.net/mattlease/crowdsourcing-for-search-evaluation-and-socialalgorithmic-search
„Mr. Burns saw Homer with the binoculars“
Sergej Zerr 16
Social Computer – Good Interface
Matthew Lease and Omar Alonso: http://de.slideshare.net/mattlease/crowdsourcing-for-search-evaluation-and-socialalgorithmic-search
Throw the coin and tell us the result
• Head?
• Tail?
Results
• Head 61
• Tail 39
People often tend
just to select the
first option 
Better: Some preliminary textual answer
• Coin type?
• Head or tail.
Sergej Zerr 17
L3S Research – “What about image privacy?”
Sergej Zerr , Stefan Siersdorfer , Jonathon Hare , Elena Demidova Privacy-Aware Image Classification and Search , SIGIR„12
Sergej Zerr 18
L3S Research
Gathering average community notion of privacy
• We crawled “most recently uploaded” Flickr photos (2 Months)
• Started a social annotation game (over the course of 2 weeks)
• 81 users (colleagues, social networks , forum users) , 6 teams
Sergej Zerr , Stefan Siersdorfer , Jonathon Hare , Elena Demidova Privacy-Aware Image Classification and Search , SIGIR„12
Sergej Zerr 19
Sampling Data from Social Web: Tags
0
50000
100000
150000
200000
250000
1 10 100 1000
#photos
#Tags per Photo
Tags Distribution of 2Mio
Flickr "Streetart" Images
>75 tags per photo
1-20 tags per photo
Sergej Zerr 20
Sampling Data from Social Web: User activity
Simple random sample can result in a set dominated by few power user
Sergej Zerr 21
Inter–Rater Reliability: What about ambiguity?
Nederland, netherlands, holland, dutch
Rotterdam, wielrennen, cycling, duck
le grand depart, tour de france,
Reklame, caravan, Funny Fotos
Sergej Zerr 22
Inter–Rater Reliability “Where is the cat?”
X XV V V V
XX XVV V
X X X X X X
V VX XX X Results
Sergej Zerr 23
Inter–Rater Reliability “Where is the cat?”
X X
XX X
V V V V
VV V
X X X X X X
V VX XX X Results
• Ideas
• Introduce „Honeypot“ – a set of ground truth objects and select only good raters
• Select only raters who rate close to average rating.
• Other startegies?
Attractive?
1 1
0 1
0 0
1 0
1 0
1 1
Sergej Zerr 24
Inter–Rater Reliability:
A
total
Yes No
B Yes 2 1 3
No 2 1 3
total 4 2 6
Naive approach: 3 cases out of 6 = 0.5 agreement
Kilem L. Gwet, Handbook of inter-rater reliability 2010
A B
Sergej Zerr 25
Inter–Rater Reliability: Cohen‟s Kappa(1960)
Idea: We need to remove agreement achived just by chance
e
ea
p
pp



1
ˆ
49.0
100
60
*
100
45
100
40
*
100
55
ep
75.0
100
4035
22211211
2211





nnnn
nn
pa
A
total
Yes No
B Yes 35 20 55
No 5 40 45
total 40 60 100
51.
49.1
49.75.0
ˆ 



Kilem L. Gwet, Handbook of inter-rater reliability 2010
A
total
Yes No
B Yes n11 n12
No n21 n22
total
Sergej Zerr 26
Inter–Rater Reliability: Missing Values
Idea: Use partial ratings to estimate marginal probability only
A
total
Yes No X
B Yes 30 18 2 50
No 5 34 3 42
X 5 3 0 8
total 40 55 5 100
e
ea
p
pp



1
ˆ
431.0
100
55
*
100
42
100
40
*
100
50
ep
74.
)85(100
3430
)( 2121
2211







xxxx
a
nnnnn
nn
p
54.
431.1
431.74.0
ˆ 



Kilem L. Gwet, Handbook of inter-rater reliability 2010
A
total
Yes No X
B Yes n11 n12 nx1
No n21 n22 nx2
X n1x n2x 0
total
Sergej Zerr 27
Inter–Rater Reliability: Extentions
• Multiple Raters/Categories:
• Fleiss 1971 – Average over random pairs of raters for random objects
• Adjustment for Ordinal and Interval Data, Weighting:
• weight judgements using distances between categories.
• Reduce Kappa Paradox:
• Split judgements into more certain / marginal.
• Measures: AC1, AC2(ordinal and interval data)
• Check for statistical significance:
• The number of categories and/or raters matters.
Kilem L. Gwet, Handbook of inter-rater reliability 2010
Sergej Zerr 28
Inter–Rater Reliability: Kappa Interpretations
Koch
Kappa Strenght of Agreement
<0.0 Poor
0.0 – 0.20 Slight
0.21 - 0.40 Fair
0.41 - 0.60 Moderate
0.61 - 0.80 Substantual
0.81 - 100 Almost Perfect
Fleiss
Kappa Strenght of Agreement
0.0-0.40 Poor
0.41 – 0.75 Intermediate to Good
>0.75 Excellent
Altman
Kappa Strenght of Agreement
<0.20 Poor
0.21 - 0.40 Fair
0.41 - 0.60 Moderate
0.61 - 0.80 Good
0.81 - 100 Very Good
Please note: These interpretations were proven to be usefull mostly in medical domain (diagnosis)
Kilem L. Gwet, Handbook of inter-rater reliability 2010
Sergej Zerr 29
Useful human power for annotating the Web
• 5000 people playing simultaneously can label all images on Google
in 30 days!
• Individual games in Yahoo! and MSN average over 5,000 players at
a time
Possible contributions: attach labels to images in other languages,
categorize web pages into topics
Sergej Zerr 30
Outline
• Collaborative Advantages
• Wisdom of crowds
• Conditions for a successful collaboration
• Using Collaborative Data
• Gathering Data from Social Web / Mechanical Turk
• Inter Rater Agreement
• Search and Sensemaking processes, the overview
• Collaboration opportunities in (Web) search
• Overview Papers
• Collaborative Search, Search result relevance judgments,..
•Small experiment
• Can we collaborate?
• Discussion
Sergej Zerr 31
What is relevant?
Sergej Zerr 32
Search and Sensemaking Process
Query
Result list
Can collaboration improve the sensemaking process at any step?
Do you use collaborative search?
Annotation/Organization
Sergej Zerr 33
What are the typical collaborative search tasks?
0.00
5.00
10.00
15.00
20.00
25.00
30.00
Travel
planning
General
shopping
tasks
Literature
search
Technical
information
Fact finding Social
planning
Medical
information
Real estate
Morris, M.R. A Survey of Collaborative Web Search Practices. In Proceedings of 26th CHI Conference 2008
Around 90% of Microsoft employees are engaged in collaborative search
activities.
• Watched over someone‟s shoulder as
he/she searched the Web, and suggested
alternate query terms.
• E-mailed someone links to share the results
of a Web search.
• E-mailed someone a textual summary to
share the results of a Web search.
• Called someone on the phone to tell them
about the results of a Web search.
Morris, M.R. A Survey of Collaborative Web Search Practices. In Proceedings of 26th CHI Conference 2008
Sergej Zerr 34
Collaborative Search:
• Query formulation
• Search process/browsing
• Save/Bookmark
• Annotate/Organize
How to support users in collaborative searching?
- Ideas
- Tools(Web 2.0)
Sergej Zerr 35
Indirect Collaboration on Web:
Google uses search/click logs. For PageRank algorithm each link to a page
serves as a vote for that page.
Amazon uses search/click logs. For item recommendation similar users are the
indirect voters for the product.
ect., ect.
Sergej Zerr 36
Outline
• Collaborative Advantages
• Wisdom of crowds
• Conditions for a successful collaboration
• Using Collaborative Data
• Gathering Data from Social Web / Mechanical Turk
• Inter Rater Agreement
• Search and Sensemaking processes, the overview
• Collaboration opportunities in (Web) search
• Overview Papers
• Collaborative Search, Search result relevance judgments,..
•Small experiment
• Can we collaborate?
• Discussion
Sergej Zerr 37
Software support for Co-located search (CoSearch).
Amershi, S., Morris, M. CoSearch: System for colocated collaborative Web search. In Proceedings of 26th CHI Conference 2008
Sergej Zerr 38
Software support for Co-located search (CoSearch).
Amershi, S., Morris, M. CoSearch: System for colocated collaborative Web search. In Proceedings of 26th CHI Conference 2008
Qualitative criteria 5 point Likert scale (5=strongly agree)
Was CoSearch easy to use? 3
Were the colors useful? 4.5
Was the query queue useful? 4
Were color tabs useful? 4
Mobile view useful? 4
I would use CoSearch at work and in the school 4
I would use CoSearch at home 3.5
Sergej Zerr 39
Spatial distributed Search (SearchTogether)
Morris, M.R. & Horvitz, E. SearchTogether: An Interface for Collaborative Web Search. In Proceedings of the UIST 2007
• (a) integrating messaging
• (b) query awareness,
• (c) current results
• (d) recommendation
queue
• (e)(f)(g) search buttons
• (h) page-specific
metadata
• (i) toolbar
• (j) browser
Sergej Zerr 40
Spatial distributed Search (SearchTogether)
Morris, M.R. & Horvitz, E. SearchTogether: An Interface for Collaborative Web Search. In Proceedings of the UIST 2007
Qualitative criteria (5 points Likert scale) Collective tokens
SearchTogether helped to complete the joint task 3.9
SearchTogether is more effective than other tools 4.1
• 38% af all result lists were the
consequence of using history
• 70 positive and 9 negative ratings
• 22 of 36 recommendation were
viewed by the recipients
Sergej Zerr 41
Improvement Through Collaborative Ranking
Agrahri, A., Manickam, D., Riedl, J. Can people collaborate to improve the relevance of search results? In Proceedings of the ACM International Conference on Recommendation Systems 2008
Sergej Zerr 42
WeSearch: Collaborative Sensemaking (Hardware support)
Meredith Ringel Morris, Jarrod Lombardo, and Daniel Wigdor. 2010. WeSearch: supporting collaborative search and sensemaking on a tabletop display. In Proceedings of the 2010 ACM conference on
Computer supported cooperative work (CSCW '10). ACM, New York, NY, USA, 401-410.
Sergej Zerr 43
WeSearch: Collaborative Sensemaking
Meredith Ringel Morris, Jarrod Lombardo, and Daniel Wigdor. 2010. WeSearch: supporting collaborative search and sensemaking on a tabletop display. In Proceedings of the 2010 ACM conference on
Computer supported cooperative work (CSCW '10). ACM, New York, NY, USA, 401-410.
Qualitative criteria (7 point Likert scale)
The system was easy to use 6
High awareness af others„ activity 5
Export record for later view was useful 6
Prefer clip cutting instead of clip search 6
Sergej Zerr 44
Hardware support for Co-located(TeamSearch).
Morris, M.R., Paepcke, A., and Winograd, T. Team-Search: Comparing Techniques for Co-Present Collaborative Search of Digital Media.
• Circles are categories:
people, location, year, event
Sergej Zerr 45
Hardware support for Co-located(TeamSearch).
Morris, M.R., Paepcke, A., and Winograd, T. Team-Search: Comparing Techniques for Co-Present Collaborative Search of Digital Media.
Qualitative criteria (7 points Likert scale) Collective tokens Parallel tokens
I worked closely with the other members of my
group to accomplish this task.
5.75 4.88
Members of the group communicated with each
other effectively.
5.75 5
The group worked effectively as a team on this
task.
5.75 4.81
Sergej Zerr 46
Outline
• Collaborative Advantages
• Wisdom of crowds
• Conditions for a successful collaboration
• Using Collaborative Data
• Gathering Data from Social Web / Mechanical Turk
• Inter Rater Agreement
• Search and Sensemaking processes, the overview
• Collaboration opportunities in (Web) search
• Overview Papers
• Collaborative Search, Search result relevance judgments,..
•Small experiment
• Can we collaborate?
• Discussion
23. Nov. 2010Sergej Zerr
Sergej Zerr 48
Smalltalk, what should be the properties of a perfect
collaborative system?

Weitere ähnliche Inhalte

Ähnlich wie Collaborative search intro 2013

UX Mind Games - Jared Schwartz, Beaconfire RedEngine
UX Mind Games - Jared Schwartz, Beaconfire RedEngineUX Mind Games - Jared Schwartz, Beaconfire RedEngine
UX Mind Games - Jared Schwartz, Beaconfire RedEngineRoshani Kothari
 
Making Sense of (Big) Data with Visual Analytics
Making Sense of (Big) Data with Visual AnalyticsMaking Sense of (Big) Data with Visual Analytics
Making Sense of (Big) Data with Visual AnalyticsKai Xu
 
Workshop - Neo4j Graph Data Science
Workshop - Neo4j Graph Data ScienceWorkshop - Neo4j Graph Data Science
Workshop - Neo4j Graph Data ScienceNeo4j
 
Understanding Queries through Entities
Understanding Queries through EntitiesUnderstanding Queries through Entities
Understanding Queries through EntitiesPeter Mika
 
How Flickr Helps us Make Sense of the World
How Flickr Helps us Make Sense of the WorldHow Flickr Helps us Make Sense of the World
How Flickr Helps us Make Sense of the Worldmor
 
An HCI-design research approach to search
An HCI-design research approach to searchAn HCI-design research approach to search
An HCI-design research approach to searchguestd2853ae
 
Searching the Web is like Searching Your Own Mind
Searching the Web is like Searching Your Own MindSearching the Web is like Searching Your Own Mind
Searching the Web is like Searching Your Own MindSean Connolly
 
Large scale social recommender systems at LinkedIn
Large scale social recommender systems at LinkedInLarge scale social recommender systems at LinkedIn
Large scale social recommender systems at LinkedInMitul Tiwari
 
Getting started in ds (july 17) atlanta
Getting started in ds (july 17)   atlantaGetting started in ds (july 17)   atlanta
Getting started in ds (july 17) atlantaThinkful
 
Using Data Science for Behavioural Game Design
Using Data Science for Behavioural Game DesignUsing Data Science for Behavioural Game Design
Using Data Science for Behavioural Game DesignGameCamp
 
[CS570] Machine Learning Team Project (I know what items really are)
[CS570] Machine Learning Team Project (I know what items really are)[CS570] Machine Learning Team Project (I know what items really are)
[CS570] Machine Learning Team Project (I know what items really are)Kunwoo Park
 
What Can We Learn from Games
What Can We Learn from GamesWhat Can We Learn from Games
What Can We Learn from GamesDouglas Kiang
 
Presentationforwebdesign 170214213118
Presentationforwebdesign 170214213118Presentationforwebdesign 170214213118
Presentationforwebdesign 170214213118kaikikazuki
 
Social Search in a Professional Context
Social Search in a Professional ContextSocial Search in a Professional Context
Social Search in a Professional ContextDaniel Tunkelang
 
MPI Presentation on Gameification DeMarle
MPI Presentation on Gameification DeMarleMPI Presentation on Gameification DeMarle
MPI Presentation on Gameification DeMarleAnn DeMarle
 
Graph Algorithms for Developers
Graph Algorithms for DevelopersGraph Algorithms for Developers
Graph Algorithms for DevelopersNeo4j
 
"Play to Change" Workshop - WebVisions Portland
"Play to Change" Workshop - WebVisions Portland"Play to Change" Workshop - WebVisions Portland
"Play to Change" Workshop - WebVisions PortlandCarolyn Chandler
 
Making Sense of (Big) Bata with Visual Analytics
Making Sense of (Big) Bata with Visual AnalyticsMaking Sense of (Big) Bata with Visual Analytics
Making Sense of (Big) Bata with Visual AnalyticsKai Xu
 

Ähnlich wie Collaborative search intro 2013 (20)

UX Mind Games - Jared Schwartz, Beaconfire RedEngine
UX Mind Games - Jared Schwartz, Beaconfire RedEngineUX Mind Games - Jared Schwartz, Beaconfire RedEngine
UX Mind Games - Jared Schwartz, Beaconfire RedEngine
 
Making Sense of (Big) Data with Visual Analytics
Making Sense of (Big) Data with Visual AnalyticsMaking Sense of (Big) Data with Visual Analytics
Making Sense of (Big) Data with Visual Analytics
 
Workshop - Neo4j Graph Data Science
Workshop - Neo4j Graph Data ScienceWorkshop - Neo4j Graph Data Science
Workshop - Neo4j Graph Data Science
 
Science Game Lab
Science Game LabScience Game Lab
Science Game Lab
 
Understanding Queries through Entities
Understanding Queries through EntitiesUnderstanding Queries through Entities
Understanding Queries through Entities
 
How Flickr Helps us Make Sense of the World
How Flickr Helps us Make Sense of the WorldHow Flickr Helps us Make Sense of the World
How Flickr Helps us Make Sense of the World
 
An HCI-design research approach to search
An HCI-design research approach to searchAn HCI-design research approach to search
An HCI-design research approach to search
 
Searching the Web is like Searching Your Own Mind
Searching the Web is like Searching Your Own MindSearching the Web is like Searching Your Own Mind
Searching the Web is like Searching Your Own Mind
 
Large scale social recommender systems at LinkedIn
Large scale social recommender systems at LinkedInLarge scale social recommender systems at LinkedIn
Large scale social recommender systems at LinkedIn
 
Getting started in ds (july 17) atlanta
Getting started in ds (july 17)   atlantaGetting started in ds (july 17)   atlanta
Getting started in ds (july 17) atlanta
 
Using Data Science for Behavioural Game Design
Using Data Science for Behavioural Game DesignUsing Data Science for Behavioural Game Design
Using Data Science for Behavioural Game Design
 
Psychology In UX
Psychology In UXPsychology In UX
Psychology In UX
 
[CS570] Machine Learning Team Project (I know what items really are)
[CS570] Machine Learning Team Project (I know what items really are)[CS570] Machine Learning Team Project (I know what items really are)
[CS570] Machine Learning Team Project (I know what items really are)
 
What Can We Learn from Games
What Can We Learn from GamesWhat Can We Learn from Games
What Can We Learn from Games
 
Presentationforwebdesign 170214213118
Presentationforwebdesign 170214213118Presentationforwebdesign 170214213118
Presentationforwebdesign 170214213118
 
Social Search in a Professional Context
Social Search in a Professional ContextSocial Search in a Professional Context
Social Search in a Professional Context
 
MPI Presentation on Gameification DeMarle
MPI Presentation on Gameification DeMarleMPI Presentation on Gameification DeMarle
MPI Presentation on Gameification DeMarle
 
Graph Algorithms for Developers
Graph Algorithms for DevelopersGraph Algorithms for Developers
Graph Algorithms for Developers
 
"Play to Change" Workshop - WebVisions Portland
"Play to Change" Workshop - WebVisions Portland"Play to Change" Workshop - WebVisions Portland
"Play to Change" Workshop - WebVisions Portland
 
Making Sense of (Big) Bata with Visual Analytics
Making Sense of (Big) Bata with Visual AnalyticsMaking Sense of (Big) Bata with Visual Analytics
Making Sense of (Big) Bata with Visual Analytics
 

Mehr von Vorname Nachname (10)

Truth Inference in Crowdsourcing: Is the Problem Solved?
Truth Inference in Crowdsourcing: Is the Problem Solved?Truth Inference in Crowdsourcing: Is the Problem Solved?
Truth Inference in Crowdsourcing: Is the Problem Solved?
 
AMA - Applied Machine Learning Academy (L3S)
AMA - Applied Machine Learning Academy (L3S)AMA - Applied Machine Learning Academy (L3S)
AMA - Applied Machine Learning Academy (L3S)
 
Ndcg
NdcgNdcg
Ndcg
 
Plane
PlanePlane
Plane
 
Mobile
MobileMobile
Mobile
 
Man
ManMan
Man
 
Computer Defect
Computer DefectComputer Defect
Computer Defect
 
Computer
ComputerComputer
Computer
 
Airport tower
Airport towerAirport tower
Airport tower
 
Cloud
CloudCloud
Cloud
 

Collaborative search intro 2013

  • 1. 23. April 2013 1 Introduction to Collaborative Web . Sergej Zerr zerr@L3S.de
  • 2. Sergej Zerr 2 Outline • Collaborative Advantages • Wisdom of crowds • Conditions for a successful collaboration •Using Collaborative Data • Gathering Data from Social Web / Mechanical Turk • Inter Rater Agreement • Search and Sensemaking processes, the overview • Collaboration opportunities in (Web) search • Overview Papers • Collaborative Search, Search result relevance judgments,.. •Small experiment • Can we collaborate? • Discussion
  • 3. Sergej Zerr 3 Collaboration Often we need more than one hand Sometimes more than one brain “Why the Many Are Smarter Than the Few and How Collective Wisdom Shapes Business, Economies, Societies and Nations” James Suroewicki
  • 4. Sergej Zerr 4 Direct Collaboration in WWW can be used: Collaborative tagging, Favorite assignments, Click logs, Data gathering, Reccomendations, ect., ect.,ect…. Tags: Rainbow, Sea, Island, Green, Palmtree, Maui
  • 5. Sergej Zerr 5 Social Computer Expert Utility Masses # of contributors10 10,000+ Equivalent, or greater, utility under the curve
  • 6. Sergej Zerr 6 Social Computer Expert $$$$ Utility Masses $ # of contributors10 10,000+ 4,000 experts 80,000 articles 200 years to develop Annual Updates 100,000 amateurs 1.6 Million articles 5 years to develop Real-Time Updates
  • 7. Sergej Zerr 7 Collaboration: Paradox http://spectrum.ieee.org/computing/hardware/multicore-is-bad-news-for-supercomputers Collaborative work needs to be managed efficiently Kasparov won against the world
  • 8. Sergej Zerr 8 Collaboration Criteria Description Diversity of opinion Each person should have private information. Independence People's opinions aren't determined by the opinions of those around them. Decentralization People are able to specialize and draw on local knowledge. Aggregation Some mechanism exists for turning private judgments into a collective
  • 9. Sergej Zerr 9 Groupthink Symptoms: Irving Lester Janis (26 May 1918 - 15 November 1990) • Illusion of invulnerability • Collective rationalization • Belief in inherent morality • Stereotyped views of out-groups http://en.wikipedia.org/wiki/Groupthink • Direct pressure on dissenters • Self-censorship • Illusion of unanimity • Self-appointed „mindguards‟
  • 10. Sergej Zerr 10 Collaboration “The best collective decisions are the product of disagreement and contest, not consensus or compromise.” “The best way for a group to be smart is for each person in it to think and act as independently as possible.”
  • 11. Sergej Zerr 11 Many Complementary Games • ESP Game: label images  Image retrieval by text • Squigl: match the labels to areas  Object recognition • Matchin: find the better image  Image ranking • FlipIt: memory with similar images Near duplicate detection • Other areas covered as well: label songs, find synonyms, describe videos • See: www.gwap.com by Luis von Ahn
  • 12. Sergej Zerr 12 Re-Using Human Power
  • 13. Sergej Zerr 13 Outline • Collaborative Advantages • Wisdom of crowds • Conditions for a successful collaboration •Using Collaborative Data • Gathering Data from Social Web / Mechanical Turk • Inter Rater Agreement • Search and Sensemaking processes, the overview • Collaboration opportunities in (Web) search • Overview Papers • Collaborative Search, Search result relevance judgments,.. •Small experiment • Can we collaborate? • Discussion Two different images that share the same ESP labels: man and woman
  • 14. Sergej Zerr 14 Social Computer Human can (yet) solve some tasks more efficient and/or accurate as a machine would do. • Captcha • Classification (OCR) • Image tagging • Speech recognition
  • 15. Sergej Zerr 15 Social Computer – Good Question Matthew Lease and Omar Alonso: http://de.slideshare.net/mattlease/crowdsourcing-for-search-evaluation-and-socialalgorithmic-search „Mr. Burns saw Homer with the binoculars“
  • 16. Sergej Zerr 16 Social Computer – Good Interface Matthew Lease and Omar Alonso: http://de.slideshare.net/mattlease/crowdsourcing-for-search-evaluation-and-socialalgorithmic-search Throw the coin and tell us the result • Head? • Tail? Results • Head 61 • Tail 39 People often tend just to select the first option  Better: Some preliminary textual answer • Coin type? • Head or tail.
  • 17. Sergej Zerr 17 L3S Research – “What about image privacy?” Sergej Zerr , Stefan Siersdorfer , Jonathon Hare , Elena Demidova Privacy-Aware Image Classification and Search , SIGIR„12
  • 18. Sergej Zerr 18 L3S Research Gathering average community notion of privacy • We crawled “most recently uploaded” Flickr photos (2 Months) • Started a social annotation game (over the course of 2 weeks) • 81 users (colleagues, social networks , forum users) , 6 teams Sergej Zerr , Stefan Siersdorfer , Jonathon Hare , Elena Demidova Privacy-Aware Image Classification and Search , SIGIR„12
  • 19. Sergej Zerr 19 Sampling Data from Social Web: Tags 0 50000 100000 150000 200000 250000 1 10 100 1000 #photos #Tags per Photo Tags Distribution of 2Mio Flickr "Streetart" Images >75 tags per photo 1-20 tags per photo
  • 20. Sergej Zerr 20 Sampling Data from Social Web: User activity Simple random sample can result in a set dominated by few power user
  • 21. Sergej Zerr 21 Inter–Rater Reliability: What about ambiguity? Nederland, netherlands, holland, dutch Rotterdam, wielrennen, cycling, duck le grand depart, tour de france, Reklame, caravan, Funny Fotos
  • 22. Sergej Zerr 22 Inter–Rater Reliability “Where is the cat?” X XV V V V XX XVV V X X X X X X V VX XX X Results
  • 23. Sergej Zerr 23 Inter–Rater Reliability “Where is the cat?” X X XX X V V V V VV V X X X X X X V VX XX X Results • Ideas • Introduce „Honeypot“ – a set of ground truth objects and select only good raters • Select only raters who rate close to average rating. • Other startegies?
  • 24. Attractive? 1 1 0 1 0 0 1 0 1 0 1 1 Sergej Zerr 24 Inter–Rater Reliability: A total Yes No B Yes 2 1 3 No 2 1 3 total 4 2 6 Naive approach: 3 cases out of 6 = 0.5 agreement Kilem L. Gwet, Handbook of inter-rater reliability 2010 A B
  • 25. Sergej Zerr 25 Inter–Rater Reliability: Cohen‟s Kappa(1960) Idea: We need to remove agreement achived just by chance e ea p pp    1 ˆ 49.0 100 60 * 100 45 100 40 * 100 55 ep 75.0 100 4035 22211211 2211      nnnn nn pa A total Yes No B Yes 35 20 55 No 5 40 45 total 40 60 100 51. 49.1 49.75.0 ˆ     Kilem L. Gwet, Handbook of inter-rater reliability 2010 A total Yes No B Yes n11 n12 No n21 n22 total
  • 26. Sergej Zerr 26 Inter–Rater Reliability: Missing Values Idea: Use partial ratings to estimate marginal probability only A total Yes No X B Yes 30 18 2 50 No 5 34 3 42 X 5 3 0 8 total 40 55 5 100 e ea p pp    1 ˆ 431.0 100 55 * 100 42 100 40 * 100 50 ep 74. )85(100 3430 )( 2121 2211        xxxx a nnnnn nn p 54. 431.1 431.74.0 ˆ     Kilem L. Gwet, Handbook of inter-rater reliability 2010 A total Yes No X B Yes n11 n12 nx1 No n21 n22 nx2 X n1x n2x 0 total
  • 27. Sergej Zerr 27 Inter–Rater Reliability: Extentions • Multiple Raters/Categories: • Fleiss 1971 – Average over random pairs of raters for random objects • Adjustment for Ordinal and Interval Data, Weighting: • weight judgements using distances between categories. • Reduce Kappa Paradox: • Split judgements into more certain / marginal. • Measures: AC1, AC2(ordinal and interval data) • Check for statistical significance: • The number of categories and/or raters matters. Kilem L. Gwet, Handbook of inter-rater reliability 2010
  • 28. Sergej Zerr 28 Inter–Rater Reliability: Kappa Interpretations Koch Kappa Strenght of Agreement <0.0 Poor 0.0 – 0.20 Slight 0.21 - 0.40 Fair 0.41 - 0.60 Moderate 0.61 - 0.80 Substantual 0.81 - 100 Almost Perfect Fleiss Kappa Strenght of Agreement 0.0-0.40 Poor 0.41 – 0.75 Intermediate to Good >0.75 Excellent Altman Kappa Strenght of Agreement <0.20 Poor 0.21 - 0.40 Fair 0.41 - 0.60 Moderate 0.61 - 0.80 Good 0.81 - 100 Very Good Please note: These interpretations were proven to be usefull mostly in medical domain (diagnosis) Kilem L. Gwet, Handbook of inter-rater reliability 2010
  • 29. Sergej Zerr 29 Useful human power for annotating the Web • 5000 people playing simultaneously can label all images on Google in 30 days! • Individual games in Yahoo! and MSN average over 5,000 players at a time Possible contributions: attach labels to images in other languages, categorize web pages into topics
  • 30. Sergej Zerr 30 Outline • Collaborative Advantages • Wisdom of crowds • Conditions for a successful collaboration • Using Collaborative Data • Gathering Data from Social Web / Mechanical Turk • Inter Rater Agreement • Search and Sensemaking processes, the overview • Collaboration opportunities in (Web) search • Overview Papers • Collaborative Search, Search result relevance judgments,.. •Small experiment • Can we collaborate? • Discussion
  • 31. Sergej Zerr 31 What is relevant?
  • 32. Sergej Zerr 32 Search and Sensemaking Process Query Result list Can collaboration improve the sensemaking process at any step? Do you use collaborative search? Annotation/Organization
  • 33. Sergej Zerr 33 What are the typical collaborative search tasks? 0.00 5.00 10.00 15.00 20.00 25.00 30.00 Travel planning General shopping tasks Literature search Technical information Fact finding Social planning Medical information Real estate Morris, M.R. A Survey of Collaborative Web Search Practices. In Proceedings of 26th CHI Conference 2008 Around 90% of Microsoft employees are engaged in collaborative search activities. • Watched over someone‟s shoulder as he/she searched the Web, and suggested alternate query terms. • E-mailed someone links to share the results of a Web search. • E-mailed someone a textual summary to share the results of a Web search. • Called someone on the phone to tell them about the results of a Web search. Morris, M.R. A Survey of Collaborative Web Search Practices. In Proceedings of 26th CHI Conference 2008
  • 34. Sergej Zerr 34 Collaborative Search: • Query formulation • Search process/browsing • Save/Bookmark • Annotate/Organize How to support users in collaborative searching? - Ideas - Tools(Web 2.0)
  • 35. Sergej Zerr 35 Indirect Collaboration on Web: Google uses search/click logs. For PageRank algorithm each link to a page serves as a vote for that page. Amazon uses search/click logs. For item recommendation similar users are the indirect voters for the product. ect., ect.
  • 36. Sergej Zerr 36 Outline • Collaborative Advantages • Wisdom of crowds • Conditions for a successful collaboration • Using Collaborative Data • Gathering Data from Social Web / Mechanical Turk • Inter Rater Agreement • Search and Sensemaking processes, the overview • Collaboration opportunities in (Web) search • Overview Papers • Collaborative Search, Search result relevance judgments,.. •Small experiment • Can we collaborate? • Discussion
  • 37. Sergej Zerr 37 Software support for Co-located search (CoSearch). Amershi, S., Morris, M. CoSearch: System for colocated collaborative Web search. In Proceedings of 26th CHI Conference 2008
  • 38. Sergej Zerr 38 Software support for Co-located search (CoSearch). Amershi, S., Morris, M. CoSearch: System for colocated collaborative Web search. In Proceedings of 26th CHI Conference 2008 Qualitative criteria 5 point Likert scale (5=strongly agree) Was CoSearch easy to use? 3 Were the colors useful? 4.5 Was the query queue useful? 4 Were color tabs useful? 4 Mobile view useful? 4 I would use CoSearch at work and in the school 4 I would use CoSearch at home 3.5
  • 39. Sergej Zerr 39 Spatial distributed Search (SearchTogether) Morris, M.R. & Horvitz, E. SearchTogether: An Interface for Collaborative Web Search. In Proceedings of the UIST 2007 • (a) integrating messaging • (b) query awareness, • (c) current results • (d) recommendation queue • (e)(f)(g) search buttons • (h) page-specific metadata • (i) toolbar • (j) browser
  • 40. Sergej Zerr 40 Spatial distributed Search (SearchTogether) Morris, M.R. & Horvitz, E. SearchTogether: An Interface for Collaborative Web Search. In Proceedings of the UIST 2007 Qualitative criteria (5 points Likert scale) Collective tokens SearchTogether helped to complete the joint task 3.9 SearchTogether is more effective than other tools 4.1 • 38% af all result lists were the consequence of using history • 70 positive and 9 negative ratings • 22 of 36 recommendation were viewed by the recipients
  • 41. Sergej Zerr 41 Improvement Through Collaborative Ranking Agrahri, A., Manickam, D., Riedl, J. Can people collaborate to improve the relevance of search results? In Proceedings of the ACM International Conference on Recommendation Systems 2008
  • 42. Sergej Zerr 42 WeSearch: Collaborative Sensemaking (Hardware support) Meredith Ringel Morris, Jarrod Lombardo, and Daniel Wigdor. 2010. WeSearch: supporting collaborative search and sensemaking on a tabletop display. In Proceedings of the 2010 ACM conference on Computer supported cooperative work (CSCW '10). ACM, New York, NY, USA, 401-410.
  • 43. Sergej Zerr 43 WeSearch: Collaborative Sensemaking Meredith Ringel Morris, Jarrod Lombardo, and Daniel Wigdor. 2010. WeSearch: supporting collaborative search and sensemaking on a tabletop display. In Proceedings of the 2010 ACM conference on Computer supported cooperative work (CSCW '10). ACM, New York, NY, USA, 401-410. Qualitative criteria (7 point Likert scale) The system was easy to use 6 High awareness af others„ activity 5 Export record for later view was useful 6 Prefer clip cutting instead of clip search 6
  • 44. Sergej Zerr 44 Hardware support for Co-located(TeamSearch). Morris, M.R., Paepcke, A., and Winograd, T. Team-Search: Comparing Techniques for Co-Present Collaborative Search of Digital Media. • Circles are categories: people, location, year, event
  • 45. Sergej Zerr 45 Hardware support for Co-located(TeamSearch). Morris, M.R., Paepcke, A., and Winograd, T. Team-Search: Comparing Techniques for Co-Present Collaborative Search of Digital Media. Qualitative criteria (7 points Likert scale) Collective tokens Parallel tokens I worked closely with the other members of my group to accomplish this task. 5.75 4.88 Members of the group communicated with each other effectively. 5.75 5 The group worked effectively as a team on this task. 5.75 4.81
  • 46. Sergej Zerr 46 Outline • Collaborative Advantages • Wisdom of crowds • Conditions for a successful collaboration • Using Collaborative Data • Gathering Data from Social Web / Mechanical Turk • Inter Rater Agreement • Search and Sensemaking processes, the overview • Collaboration opportunities in (Web) search • Overview Papers • Collaborative Search, Search result relevance judgments,.. •Small experiment • Can we collaborate? • Discussion
  • 48. Sergej Zerr 48 Smalltalk, what should be the properties of a perfect collaborative system?