Reading the Correct History? Modeling Temporal Intention in Resource Sharing
1. Reading the Correct History?
Modeling Temporal Intention in
Resource Sharing
Hany SalahEldeen & Michael Nelson Reading the Correct History?
Hany M. SalahEldeen & Michael L. Nelson
Old Dominion University
Department of Computer Science
Web Science and Digital Libraries Lab.
2. Hany SalahEldeen & Michael Nelson 1 Reading the Correct History?
⢠We share web pages
What I share might not be what my readers read
Possible Scenario:
⢠Web pages change
⢠Readers explore shared pages
3. Motivation
A temporal inconsistency can arise in
the intention of the author regarding
the state of the resource between the
tweet time and the read timeâŚ
Hany SalahEldeen & Michael Nelson 2 Reading the Correct History?
Can we detect and model this
difference in intention?
4. The game plan
Hany SalahEldeen & Michael Nelson 3 Reading the Correct History?
Problem Illustration
Training data collection attempts
The TIRM model
Ground truth validation
Data collection
Feature extraction and modeling
Model evaluation
6. Clicking on the link in the tweet âŚ
Hany SalahEldeen & Michael Nelson 5 Reading the Correct History?
7. Using the Twitter expanded interface
Hany SalahEldeen & Michael Nelson 6 Reading the Correct History?
The attack on the embassy was in February
2013
8. Problem: There is an inconsistency
between what the tweetâs author intended
to share at time ttweet
and what the reader might actually read
upon clicking on the link at time tclick .
Hany SalahEldeen & Michael Nelson 7 Reading the Correct History?
9. Hany SalahEldeen & Michael Nelson 8 Reading the Correct History?
Implication: Since tweets are considered
the first draft of history⌠the historical
integrity of the tweets could be
compromised.
10. Solution: Detect the correct intention
Hany SalahEldeen & Michael Nelson 9 Reading the Correct History?
Option 1 Option 2 Option 3
11. The game plan
Hany SalahEldeen & Michael Nelson Reading the Correct History?
Problem Illustration
Training data collection attempts
The TIRM model
Ground truth validation
Data collection
Feature extraction and modeling
Model evaluation
12. Amazonâs Mechanical Turk (MT)
⢠Crowdsourcing Internet marketplace
⢠Co-ordinates the use of human intelligence to
perform tasks that computers are currently unable to
do.*
Hany SalahEldeen & Michael Nelson 10 Reading the Correct History?
* http://en.wikipedia.org/wiki/Amazon_Mechanical_Turk
13. Goal: Collect user intention data via MT
Hany SalahEldeen & Michael Nelson 11
Reading the Correct History?
Tweets dataset Intention Classification Tasks User Intention Data
Classifier
Train
⢠Problem:
â It is not as easy as it seems!
14. How not to classify temporal
intention 101
⢠Given a tweet, is the intended state of the link is
in:
Hany SalahEldeen & Michael Nelson 12 Reading the Correct History?
past state? current state? No information?
15. Ground truth collection
⢠A dataset of 100 tweets classified by:
â Our Web Science and Digital Libraries (WS-DL)
research group members
â MT workers
Hany SalahEldeen & Michael Nelson 13 Reading the Correct History?
16. The agreement was very lowâŚ
⢠Reliability of agreement between:
â WS-DL members = Fleissâ Ď° = 0.14
â MT workers = Fleissâ Ď° = 0.07
⢠Inter-rater agreement between the collective WS-DL
members and MT workers = Cohenâs Ď° = 0.04
ď Slight agreement
Hany SalahEldeen & Michael Nelson 14 Reading the Correct History?
17. So we removed the guessing part:
⢠The tweet is presented along with the two snapshots:
Hany SalahEldeen & Michael Nelson 15 Reading the Correct History?
at ttweet at tclick
18. ⌠and classified the 100 tweets
again
⢠Via a face to face meeting with WS-DL members.
⢠Resubmitted the new experiment to MT.
Hany SalahEldeen & Michael Nelson 16 Reading the Correct History?
19. The tweet, current and past
snapshots
Hany SalahEldeen & Michael Nelson 17 Reading the Correct History?
Past Version Current Version
20. The results remained very low
⢠For 9 MT assignments per tweet:
â If we allowed 4-5 splits we have 58% match with WS-DL.
â If we allowed 3-6 splits or better we got 31% match
ď Which is worse that flipping a coin!
Hany SalahEldeen & Michael Nelson 18 Reading the Correct History?
21. Observations
⢠Assigning a temporal intention is not
a trivial task.
⢠MT workers are accustomed to more
straightforward tasks.
⢠The concept of âtime on the webâ is
foreign to MT workers.
Hany SalahEldeen & Michael Nelson 19 Reading the Correct History?
22. The game plan
Hany SalahEldeen & Michael Nelson Reading the Correct History?
Problem Illustration
Training data collection attempts
The TIRM model
Ground truth validation
Data collection
Feature extraction and modeling
Model evaluation
23. Idea: We need to transform the
problem from intention to
relevance.
Hany SalahEldeen & Michael Nelson 20 Reading the Correct History?
24. Relevance tasks are simpler
⢠MT workers are more accustomed to classification tasks
and it requires minimum amount of explanation
Is that a cat?
- Yes
- No
Hany SalahEldeen & Michael Nelson 21 Reading the Correct History?
25. Hany SalahEldeen & Michael Nelson 22 Reading the Correct History?
Temporal Intention Relevancy Model
( TIRM)
Between ttweet and tclick:
The linked resource could have:
⢠Changed
⢠Not changed
The tweet and the linked resource could be:
⢠Still relevant
⢠No longer relevant
26. Hany SalahEldeen & Michael Nelson 23 Reading the Correct History?
Resource is changed but relevant
⢠The resource changed
⢠But it is still relevant
ď Intention: need the current version of the resource at any time
27. Hany SalahEldeen & Michael Nelson 24 Reading the Correct History?
Relevancy and Intention Mapping
Current
28. Hany SalahEldeen & Michael Nelson 25 Reading the Correct History?
Resource is changed and not relevant
ď Intention: need the past version of the resource at any time
⢠The resource changed
⢠But it is no longer relevant
29. Past
Hany SalahEldeen & Michael Nelson 26 Reading the Correct History?
Relevancy and Intention Mapping
Current
30. Hany SalahEldeen & Michael Nelson 27 Reading the Correct History?
Resource is not changed and relevant
ď Intention: need the past version of the resource at any time
⢠The resource is not changed
⢠And it is relevant
31. Past
Hany SalahEldeen & Michael Nelson 28 Reading the Correct History?
Relevancy and Intention Mapping
Current
Past
32. Hany SalahEldeen & Michael Nelson 29 Reading the Correct History?
Resource is not changed and not relevant
ď Intention: I am not sure which version of the resource I need
⢠The resource is not changed
⢠But it is not relevant
33. Past
Hany SalahEldeen & Michael Nelson 30 Reading the Correct History?
Relevancy and Intention Mapping
Current
Past Not Sure
34. The game plan
Hany SalahEldeen & Michael Nelson Reading the Correct History?
Problem Illustration
Training data collection attempts
The TIRM model
Ground truth validation
Data collection
Feature extraction and modeling
Model evaluation
35. Next step: validation
⢠MT workers ⥠judgments of the experts (WS-DL members)
Hany SalahEldeen & Michael Nelson 31 Reading the Correct History?
â
Is the content still relevant to the tweet?
36. Filtering the results
⢠We accepted raters with:
â At least 1000 accepted HITs
â 95% acceptance rate
⢠Average completion time = 61 seconds
⢠We removed:
â Any assignments that took <10 seconds ď hasty decision
â Low quality repetitive assignments and banned the raters
Hany SalahEldeen & Michael Nelson 32 Reading the Correct History?
37. Mechanical Turk Workers Vs. Experts
⢠For 100 tweets, WS-DL members % of agreement :
⢠Cohenâs Ď° = 0.854 ď almost perfect agreement
Hany SalahEldeen & Michael Nelson 33 Reading the Correct History?
Agreement in three or more votes 93%
Agreement in four or more votes 80%
Agreement with all five votes 60%
38. The game plan
Hany SalahEldeen & Michael Nelson 34 Reading the Correct History?
Problem Illustration
Training data collection attempts
The TIRM model
Ground truth validation
Data collection
Feature extraction and modeling
Model evaluation
39. Data collection
⢠From SNAP dataset we extracted:
â Tweets in English
â Each has an embedded URI pointing to an external resource.
â The embedded URI is shortened via Bit.ly
â The external resource:
⢠Still persists.
⢠Has at least 10 mementos.
⢠Is unique.
ď We extracted 5,937 unique instances
Hany SalahEldeen & Michael Nelson 35 Reading the Correct History?
40. Get the closest memento
Hany SalahEldeen & Michael Nelson 35 Reading the Correct History?
âŚ
t1 t2
tn
t4t3
Î1 Î2< ď Pick Memento @ t1
41. Sorted Time Delta between tweet and closest memento
Hany SalahEldeen & Michael Nelson 36 Reading the Correct History?
Randomly selected 1,124 instances
Time delta range: 3.07 minutes to 56.04 hours Average: 25.79 hours ~ 1 day
Tweet time
After Tweet time
Before Tweet time
42. Training dataset
⢠Rcurrent: The state of the resource at current time.
⢠Rclick: The state of the resource at click time.
Hany SalahEldeen & Michael Nelson 37 Reading the Correct History?
Relevant Assignments 929 82.65%
Non-Relevant Assignments 195 17.35%
5 MT workers agreeing (5-0 split) 589 52.40%
4 MT workers agreeing (4-1 split) 309 27.49%
3 MT workers agreeing (3-2 close call split) 226 20.11%
43. The game plan
Hany SalahEldeen & Michael Nelson 38 Reading the Correct History?
Problem Illustration
Training data collection attempts
The TIRM model
Ground truth validation
Data collection
Feature extraction and modeling
Model evaluation
44. Feature extraction
⢠For each tweet we perform:
â Link analysis
â Social Media Mining
â Archival Existence
â Sentiment Analysis
â Content Similarity
â Entity Identification
Hany SalahEldeen & Michael Nelson 39 Reading the Correct History?
45. Link analysis
⢠Since the tweets have embedded resources shortened by
Bit.ly we can extract:
â Total number of clicks
â Hourly click logs
â Creation dates
â Referring websites
â Referring countries.
⢠We calculate the depth of the resource in relation to its domain
(either it is a leaf node or a root page)
â We calculated the number of backslashes in the resourceâs URI
Hany SalahEldeen & Michael Nelson 40 Reading the Correct History?
46. Social Media Mining
⢠Twitter:
â Using Topsy.comâs API to
extract:
⢠Total number of tweets.
⢠The most recent 500.
⢠Number of tweets by
influential users.
Hany SalahEldeen & Michael Nelson 41 Reading the Correct History?
ď The collection of tweets extracted provided an extended context of the
resource authored by users in the twittersphere.
47. Social Media Mining
⢠Facebook:
â Mined too for likes, shares, posts, and clicks related to each
resource.
Hany SalahEldeen & Michael Nelson 42 Reading the Correct History?
48. Archival Existence
⢠Using Memento Time
Maps we get:
â Total mementos
available
â Different archives count.
â The closest archived
version to the tweet
time.
Hany SalahEldeen & Michael Nelson 43 Reading the Correct History?
49. Sentiment Analysis
⢠Using NLTK libraries of natural language text processing
⢠Extract the most prominent sentiment in the text
Hany SalahEldeen & Michael Nelson 44 Reading the Correct History?
50. Content Similarity
⢠Steps:
â We download the content HTML using Lynx browser.
â We apply boilerplate removal algorithm and full text extraction.
â Calculate the cosine similarity between the two pages.
Hany SalahEldeen & Michael Nelson 45 Reading the Correct History?
ď 70% similarity ď
51. Entity Identification
⢠By visual inspection we observed that the majority of tweets about
celebrities are related to current events.
⢠We harvested Wikipedia for lists of actors, politicians, and athletes.
⢠Checked the existence of a celebrity mention in the tweets.
Hany SalahEldeen & Michael Nelson 46 Reading the Correct History?
Actor: Johnny Depp
52. ⢠To remove confusion we removed the close calls
ď 898 instances remaining
Relevant Assignments 929 82.65%
Non-Relevant Assignments 195 17.35%
5 MT workers agreeing (5-0 split) 589 52.40%
4 MT workers agreeing (4-1 split) 309 27.49%
3 MT workers agreeing (3-2 close call split) 226 20.11%
Modeling and Classification
Hany SalahEldeen & Michael Nelson 47 Reading the Correct History?
53. The trained classifier
⢠From the feature extraction phase we extracted 39
different features to train the classifier.
⢠Using 10-fold cross validation, the Cost Sensitive Classifier
Based on Random Forests gave the highest success rate =
90.32%
Hany SalahEldeen & Michael Nelson 48 Reading the Correct History?
54. Testing the model
Hany SalahEldeen & Michael Nelson 49 Reading the Correct History?
10-Fold Cross-Validation Testing
Classifier
Mean Absolute
Error
Root Mean
Squared Error
Kappa
Statistic
Incorrectly
Classified %
Correctly
Classified %
Cost sensitive
classifier based on
Random Forest
0.15 0.27 0.39 9.68% 90.32%
Classifier Precision Recall F-measure Class
Cost sensitive classifier based on
Random Forest
0.93
0.53
0.96
0.37
0.95
0.44
Relevant
Non-Relevant
Weighted Average 0.89 0.90 0.90
55. Feature significance
⢠Since we have 39 features, we needed to understand the
effect of each feature and which are the strongest ones
affecting the classification
⢠We applied an attribute evaluator supervised algorithm
based on Ranker search to find the strongest features
Hany SalahEldeen & Michael Nelson 50 Reading the Correct History?
56. Most significant features sorted by
information gain
Hany SalahEldeen & Michael Nelson 51 Reading the Correct History?
Rank Feature Gain Ratio
1 Existence of celebrities in tweets 0.149
2 Number of mementos 0.090
3 Tweet similarity with current page 0.071
4 Similarity: Current & past page 0.0527
5 Similarity: Tweet & past page 0.04401
6 Original URIâs depth 0.0324
57. The game plan
Hany SalahEldeen & Michael Nelson Reading the Correct History?
Problem Illustration
Training data collection attempts
The TIRM model
Ground truth validation
Data collection
Feature extraction and modeling
Model evaluation
58. Model Evaluation
⢠Next step was to test the trained model against other
datasets and examine the results.
⢠We tested against:
â The remaining 4,813 from the original 5,937 instances after extracting the
1,124 used in training.
â The Tweet Collections based on historic events. (MJ, Obama, Iran, Syria, &
H1N1)
Hany SalahEldeen & Michael Nelson 52 Reading the Correct History?
59. Results of testing the model
against multiple datasets
Hany SalahEldeen & Michael Nelson 53 Reading the Correct History?
Dataset Status 200 Status 404 of other Relevant % Non-Relevant %
Extended 4,813 instances 96.77% 3.23% 96.74% 3.26%
MJâs Death 57.54% 42.46% 93.24% 6.76%
H1N1 Outbreak 8.96% 91.04% 97.48% 2.52%
Iran Elections 68.21% 31.79% 94.69% 5.31%
Obamaâs Nobel Prize 62.86% 37.14% 93.89% 6.11%
Syrian Uprising 80.80% 19.20% 70.26% 29.75%
60. Hany SalahEldeen & Michael Nelson 54 Reading the Correct History?
Idea: We need to transform the
problem from intention to
relevance.
RecapâŚ
Now we need to transform it back!
61. Mapping TIRM
⢠We used 70% similarity as a threshold of relevancy.
Hany SalahEldeen & Michael Nelson 55 Reading the Correct History?
62. Conclusions
⢠TIRM successfully transfers the temporal intention
problem to a temporal relevancy problem.
⢠Temporal relevancy is easier to solve and MT workers
provide almost perfect agreement with expertsâ opinions.
⢠We successfully collected a gold standard dataset of
temporal user intention.
⢠We found a temporal inconsistency in the shared
resource ranging from <1% to 25% according to the
dataset.
Hany SalahEldeen & Michael Nelson 56 Reading the Correct History?