This document summarizes a workshop paper that explored user perceptions of semantic similarity between near-duplicate multimedia files through a crowdsourcing task. Workers were shown triads of multimedia search results and asked to identify the odd one out and justify their choice. The responses revealed many dimensions of user-perceived similarity beyond technical metadata. Fourteen qualified workers provided judgments for over 300 triads in under 36 hours, which were then categorized into similarity dimensions through a card sorting process. The findings could guide the design of clustering algorithms for multimedia search results.
The Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdf
Discovering User Perceptions of Semantic Similarity
1. Discovering User Perceptions of
Semantic Similarity in
Near-duplicate Multimedia Files
Raynor Vliegendhart (speaker)
Martha Larson
Johan Pouwelse
WWW 2012 Workshop on Crowdsourcing Web Search (CrowdSearch 2012),
Lyon, France, April 17, 2012.
3. Question:
Are these the same? Why (not)?
Chrono Cross -
'Dream of the Shore Near Another World'
Violin/Piano Cover
Chrono Cross
Dream of the Shore Near Another World
Violin and Piano
sources: YouTube, IQYNEj51EUI (left), Iuh3YrJtK3M (right)
3
4. Question:
Are these the same? Why (not)?
Chrono Cross -
'Dream of the Shore Near Another World'
Violin/Piano Cover
Yes, it’s the
same song
Chrono Cross
Dream of the Shore Near Another World
Violin and Piano
sources: YouTube, IQYNEj51EUI (left), Iuh3YrJtK3M (right)
4
5. Question:
Are these the same? Why (not)?
Chrono Cross -
'Dream of the Shore Near Another World'
Violin/Piano Cover
No, these are
different performances
by different performers
Chrono Cross
Dream of the Shore Near Another World
Violin and Piano
sources: YouTube, IQYNEj51EUI (left), Iuh3YrJtK3M (right)
5
6. Problem:
What constitutes a near duplicate?
Functional near-duplicate multimedia items are items
that fulfill the same purpose for the user.
Once the user has one of these items, there is no
additional need for another.
6
7. Problem:
What constitutes a near duplicate?
Our work:
• Discovering new notions of user-perceived
similarity between multimedia files
• in a file-sharing setting
• through a crowdsourcing task.
7
11. Crowdsourcing Task:
Point the odd one out
• Three multimedia files displayed as search results
• Worker points the odd one out and justifies why
• Challenge: eliciting serious judgments
11
12. Crowdsourcing Task:
Eliciting serious judgments (1)
“Imagine that you downloaded
the three items in the list
and that you view them.”
Harry Potter and the Sorcerers Stone Audio
Book (478 MB)
Harry Potter and the Sorcerer s Stone
(2001)(ENG GER NL) 2Lions- (4.36 GB)
Harry Potter.And.The.Sorcerer.Stone.DVDR.
NTSC.SKJACK.Universal.S (4.46 GB)
12
13. Crowdsourcing Task:
Eliciting serious judgments (2)
• Don’t force workers to make a contrast
• Explain the definition of functional similarity
o The items are comparable. They are for all practical purposes the
same. Someone would never really need all three of these.
o Each item can be considered unique. I can imagine that someone
might really want to download all three of these items.
o One item is not like the other two. (Please mark that item in the list.)
The other two items are comparable.
13
17. Results
1000 test triads
3 validation triads + 28 validation triads mixed in
Recruitment Main HIT
HIT
(3 workers per test triad)
two HITs run concurrently
17
18. Results
1000 test triads
3 validation triads + 28 validation triads mixed in
Recruitment 8
Main HIT
HIT
14 qualified
workers
free-text judgments
< 36h for 308 test triads
18
19. Card Sort
• Print judgments on small pieces of paper
• Group similar judgments into piles
• Merge piles iteratively
• Label each pile
19
20. Card Sort
Example: “different language”
• “The third item is a Hindi language version of the movie”
• “This is a Spanish version of the movie represented by the other
two”
•…
20
21. User-perceived
Similarity Dimensions
Different movie vs. TV show Different movie
Normal cut vs. extended cut Movie vs. trailer
Cartoon vs. movie Comic vs. movie
Movie vs. book Audiobook vs. movie
Game vs. corresponding movie Sequels (movies)
Commentary document vs. movie Soundtrack vs. corresponding movie
Movie/TV show vs. unrelated audio album Movie vs. wallpaper
Different episode Complete season vs. individual episodes
Episodes from different season Graphic novel vs. TV episode
Multiple episodes vs. full season Different realization of same legend/story
Different songs Different albums
Song vs. album Collection vs. album
Album vs. remix Event capture vs. song
Explicit version Bonus track included
Song vs. collection of songs+videos Event capture vs. unrelated movie
Language of subtitles Different language
Mobile vs. normal version Quality and/or source
Different codec/container (MP4 audio vs. MP3) Different game
Crack vs. game Software versions
Different game, same series Different application
Addon vs. main application Documentation (pdf) vs. software
List (text document) vs. unrelated item Safe vs. X-Rated
21
23. Conclusions
• A wealth of user-perceived dimensions of similarity discovered,
some we could not have thought of
• Quick results due to interesting crowdsourcing task,
with the focus on engagement and encouraging serious workers
23
24. Future Work
• Expand experiments, larger worker volume
• Other multimedia search settings
• Crowdsourcing the card sorting process
• Use findings to guide design of clustering algorithms
Done: first version is deployed in Tribler
24