SlideShare ist ein Scribd-Unternehmen logo
1 von 62
Downloaden Sie, um offline zu lesen
WHAT WIKIDATA
TEACHES US
ABOUT
KNOWLEDGE
ENGINEERING
Elena Simperl
King’s College London, UK
@esimperl
ISWS2022
July 2022
OVERVIEW
Knowledge graphs have
become a critical AI resource
We study them associo-
technical constructs
Our research
 Explores knowledge-graph engineering
empirically to understand practices and
improve processes and tools
 Uncovers links between social and technical
qualities of knowledge graphs to make them
better
Picture from https://medium.com/@sderymail/challenges-of-knowledge-graph-part-1-d9ffe9e35214
KNOWLEDGE ENGINEERING AS HUMAN-
MACHINE PROCESS
Human-
machine
interactions
Knowledge
engineering
communities
WIKIDATA
Collaborative knowledge graph,
Wikimedia project (2012),
23k active users, 99m items, 1.6b edits
Open license
RDF support, links to LOD cloud
THE KNOWLEDGE GRAPH
STATEMENTS, ITEMS, PROPERTIES
Item identifiers start with a Q, property identifiers
start with a P
8
Q84
London
Q334155
Sadiq Khan
P6
head of government
THE KNOWLEDGE GRAPH
ITEMS CAN BE CLASSES, ENTITIES, VALUES
9
Q7259
Ada Lovelace
Q84
London
Q334155
Sadiq Khan
P6
head of government
Q727
Amsterdam
Q515
city
Q6581097
male
Q59360
Labour party
Q145
United Kingdom
THE KNOWLEDGE GRAPH
STATEMENTS IN CONTEXT
Statements may include context
 Qualifiers (optional)
 References (required)
Two types of references
 Internal, linking to another item
 External, linking to webpage
10
Q84
London
Q334155
Sadiq Khan
P6
head
of government
9 May 2016
london.gov.uk
THE KNOWLEDGE GRAPH
CO-EDITED BY BOTS AND HUMANS
Human editors can register or work anonymously
Community creates bots to automate routine tasks
23k active human users, 340+ bots
IN THIS TALK
Assembling your KG team
KG quality as a function of
data provenance
Supporting editors and
improving retention
‘ONTOLOGIES ARE US’
Piscopo, A., Phethean, C., & Simperl, E. (2017). What Makes a
Good Collaborative Knowledge Graph: Group Composition and
Quality in Wikidata. International Conference on Social
Informatics, 305-322, Springer.
Piscopo, A., & Simperl, E. (2018). Who Models the World?:
Collaborative Ontology Creation and User Roles in
Wikidata. Proceedings of the ACM on Human-Computer
Interaction, 2(CSCW), 141.
BACKGROUND
Editors have varied tenure and interests
Their profile and editing behaviour impact
outcomes
 Group composition has varied effects
 Tenure and interest diversity can increase outcome
quality and group productivity
 Different editors groups focus on different types of
activities
Chen, J., Ren, Y., Riedl, J.: The effects of diversity on group productivity and member withdrawal in online volunteer groups. In: Proceedings of the 28th international
conference on human factors in computing systems - CHI ’10. p. 821. ACM Press, New York, USA (2010)
STUDY 1: ITEM QUALITY
Analysed edit history of items
Corpus of 5k items, whose quality has been
manually assessed (5 levels)*
Edit history focused on community make-up
Community is defined as set of editors of item
Considered features from group diversity
literature and Wikidata-specific aspects
*https://www.wikidata.org/wiki/Wikidata:Item_quality
RESEARCH HYPOTHESES
Activity Outcome
H1 Bots edits Item quality
H2 Bot-human interaction Item quality
H3 Anonymous edits Item quality
H4 Tenure diversity Item quality
H5 Interest diversity Item quality
DATA AND METHODS
Ordinal regression analysis, trained four models
Dependent variable: 5k labelled Wikidata items
Independent variables
 Proportion of bot edits
 Bot human edit proportion
 Proportion of anonymous edits
 Tenure diversity: Coefficient of variation
 Interest diversity: User editing matrix
Control variables: group size, item age
RESULTS
ALL HYPOTHESES SUPPORTED
H1
H2
H3 H4
H5
SUMMARY AND IMPLICATIONS
The more is
not always
the merrier
01
Bot edits are
key for quality,
but bots and
humans are
better
02
Registered
editors have
a positive
impact
Diversity
matters
04
Encourage
registration
01
Identify further
areas for bot
editing
02
Design effective
human-bot
workflows
03
Suggest items
to edit based
on tenure and
interests
04
03
ONGOING WORK
Analysing how
the community
awards quality
labels to items
Developing
robust machine
learning models
Spotting biases
STUDY 2: ONTOLOGY QUALITY
Analysed Wikidata ontology and its edit
context
Defined as the graph of all items linked through
P31 (instance of) & P279 (subclass of)
Calculated evolution of quality indicators and
editing activity in time and the links between them
Based on features from literature on ontology
evaluation and collaborative ontology engineering
DATA AND METHODS
Wikidata dumps from March 2013 (creation of P279)
to September 2017
 Analysed data in 55 monthly time frames
Literature survey to define Wikidata ontology quality
framework
Clustering to identify ontology editor roles
Lagged multiple regression to link roles and ontology
features
 Dependent variable: Changes in ontology quality across time
 Independent variables: Number of edits by different roles
 Control variables: Bot and anonymous edits
ONTOLOGY QUALITY: INDICATORS
Based on 7 ontology evaluation frameworks
Compiled structural indicators that can be computed from the data
23
Indicator Description Indicator Description
noi Number of instances ap; mp Average and median population
noc Number of classes rr Relationship richness
norc Number of root classes ir, mr Inheritance and median richness
nolc Number of leaf classes cr Class richness
nop Number of properties ad, md, maxd Average, median, max explicit depth
Sicilia, M. A., Rodríguez, D., García-Barriocanal, E., & Sánchez-Alonso, S. (2012). Empirical findings on ontology metrics. Expert Systems with
Applications, 39(8), 6706-6711.
ONTOLOGY QUALITY: RESULTS
LARGE ONTOLOGY, UNEVEN QUALITY
>1.5M classes, ~4000 properties
No of classes increases at same rate as overall
no of items, likely due to users incorrectly using
P31 & P279
ap and cr decrease over time, several classes
are either without instances, sub-classes or both
ir & maxd increase over time, part of the
Wikidata ontology is distributed vertically
24
EDITOR ROLES: METHODS
K-means, features based on previous studies in ontology engineering
Analysis by yearly cohort
25
Feature Description Feature Description
# edits Total number of edits per month. # property edits Total number of edits on
Properties in a month.
# ontology edits Number of edits on classes. # taxonomy edits Number of edits on P31 and P279
statements.
# discussion edits Number of edits on talk pages. p batch edits Number of edits done through
automated tools.
# modifying editsNumber of revisions on previously
existing statements.
item diversity Proportion between number of edits
and number of items edited.
admin True if user in an admin user
group, false otherwise.
lower admin True if user in a user group
with enhanced user rights,
false otherwise.
EDITOR ROLES: RESULTS
190,765 unique editors over 55 months (783k
total)
18k editors active for 10+ months
2 clusters, obtained using gap statistic (tested
2≥k≥8)
Leaders: more active minority (~1%), higher
number of contributions to ontology, engaged
within the community
Contributors: less active, lower number of
contributions to ontology and lower proportion of
batch edits
26
EDITOR ROLES: RESULTS
People who joined the project early tend to be
more active & are more likely to become leaders
Levels of activity of leaders decrease over time
(alternatively, people move on to different tasks)
27
RESEARCH HYPOTHESES
H1 Higher levels of leader activity are negatively correlated to
number of classes (noc), number of root classes (norc), and
number of leaf classes (nolc)
H2 Higher levels of leader activity are positively correlated to
inheritance richness (ir), average population (ap), and average
depth (ad)
28
ROLES & ONTOLOGY: RESULTS
H1 not supported
H2 partially supported
Only inheritance richness (ir) and average depth (ad)
related significantly with leader edits (p<0.01)
Bot edits significantly and positively affect the number of
subclasses and instances per class (ir & ap) (p<0.05)
29
CONCLUSIONS
Creating ontologies still a challenging task
Size of ontology renders existing automatic quality
assessment methods unfeasible
Broader curation efforts needed: large number of empty
classes
Editor roles less well articulated than in other ontology
engineering projects
Possible decline in motivation after several months
ONGOING WORK
The role of discussions in
Wikidata
Studying participation
pathways to observe
learning effects
Analysis of ShEx
schemas
THE CONTENT IS AS
GOOD AS ITS
REFERENCES
Amaral, G., Piscopo, A., Kaffee, L. A., Odinaldo, R., &
Simperl, E. (2021). Assessing the quality of sources in
Wikidata across languages: a hybrid approach. ACM
JDIQ 13(4), 1-35.
32
PROVENANCE IN WIKIDATA
Statements may include context
 Qualifiers (optional)
 References (required)
Two types of references
 Internal, linking to another item
 External, linking to webpage
33
Q84
London
Q334155
Sadiq Khan
P6
head
of government
9 May 2016
london.gov.uk
THE ROLE OF PROVENANCE
Wikidata aims to become a hub of references
Provenance increases trust in Wikidata
Lack of provenance hinders downstream reuse
Quality of references underexplored
Hartig, O. (2009). Provenance Information in the Web of Data. LDOW, 538.
STUDY 3
Approach to evaluate quality of external and
internal references in Wikidata
Quality defined by the Wikidata verifiability policy
 Relevant: support the statement they are attached to
 Authoritative: trustworthy, up-to-date, and free of bias for supporting a
particular statement
 Easy to access: users should be able to access the information in references with
low perceived effort
Large-scale (the whole of Wikidata)
Multiple languages (EN, ES, PT, SV, NL, JA)
RESEARCH QUESTIONS
RQ1 How easy it is to access sources?
RQ2 How authoritative are sources?
 According to existing guidance, do they match the
author and publisher types from the Wikidata policy?
RQ3 Are sources relevant to claims?
RESEARCH QUESTIONS
RQ4 Are external and internal references of
varying quality?
RQ5 Are references in different languages of
varying quality?
RQ6 Can we predict relevance, authoritativeness
and ease of use of a reference without relying on
the content of the references themselves? Can we
do this for references in different languages?
METHODS
THREE STAGE MIXED APPROACH
1. Descriptive statistics
Understand differences in external/internal references
Determine data distributions
Language
Predicates
Domains
Prepare data to crowdsource
METHODS
THREE STAGE MIXED APPROACH
2. Microtask crowdsourcing
Evaluate relevance, authoritativeness & ease
of access of a reference sample
Create training set for ML
3. Machine learning
Large-scale reference quality prediction
RQs 1-5
RQ6
MICROTASK CROWDSOURCING
2 tasks on Amazon’s Mechanical Turk
5 workers/task, majority voting
Quality assurance through golden data & attention checks
40
Feature Microtask Description
Relevance / Ease
of access
T1.1 Does the reference support the statement?
T1.2 How easy was the navigation to verify the reference?
T1.3 What was the biggest barrier to access, if any?
Authoritativeness
T2.1 Choose author type from list
T2.2 Choose publisher type from list
T2.3 Choose publisher sub-type from list
RQs 1-5
MACHINE LEARNING
Compared 5 algorithms
 Naïve Bayes, Random Forest, SVM, XGBoost, Neural Network
Features taken from URLs and Wikidata ontology
3 tasks: relevance, ease of navigation (data from T1.2), and
authoritativeness
RQ6
Features
Features of the representative URL extracted
Features of the reference node’s coding and structure
Features of the website available through the representative URL
Features of the claim node associated to this reference node
DATA
Snapshot up to April 16th, 2020
15m individual items used in Stage 1 (20% of
total)
385 random references from each language
(95% conf., 5% m. of error)
2,310 references in total
 Six languages, both internal and external
 Around 940 could be automatically checked (40% of total)
RESULTS: CROWDSOURCING
CROWDSOURCING WORKS
Agreement calculated per question and language
Agreement overall moderate (over 0.4) and often substantial (over
0.6)
 Multiple trusted metrics used for robustness
* Kappa metrics can’t be used for interval data
Task No. of microtasks Total workers Fleiss’ k Randolph’s k Kripp’s α
T1.1 460 microtasks 87 0.507702 0.745626 0.537647
T1.2 N/A* N/A* 0.614661
T1.3 0.405330 0.592810 0.415045
T2.1 370 microtasks 74 0.496904 0.781481 0.645795
T2.2 0.603499 0.695833 0.657171
T2.3 0.515551 0.604419 0.572782
RESULTS: CROWDSOURCING
MAJORITY OF REFERENCES ARE HIGH QUALITY
2310 references evaluated
Found 91.73% are relevant
Found 66.97% are authoritative
Most (95%) unauthoritative references link to Wikipedia
On ease of access:
5-level Likert Scale was used
78.43% fall on levels 3 (easy) or 4 (very easy)
6.66% fall on levels 0 (very difficult) and 1 (difficult)
76.44% of irrelevant references were due to working links with
wrong information (different from Wikidata)
19.37% were due to bad links
RQ1
RQ2
RQ3
RESULTS: CROWDSOURCING
VARYING REFERENCE QUALITY
RQ4
 External references had better agreement than internal
 10% better on average (Krippendorff’s Alpha)
 No noticeable quality difference between external and
internal references
 Quality differs a lot based on language
 Examples:
 English references are much harder to access (19.37% voted as
very difficult)
 Non-authoritative references: 4.94% for Dutch , but 43.12% for
Japanese
 Likely due to specific languages used for specific topics
RQ5
RESULTS: CROWDSOURCING
DATA FROM GOVERNMENT AND ACADEMIC SOURCES
Most common author type (T2.1)
 Organization (67%)
Most common publisher types (T2.2)
 Companies/Organizations (37.75%)
 Self-published (28.79%)
 Academic/Scientific (22.16%)
This also varies based on language
 Academic/Scientific for English is 83.12%
 Self-published for Swedish is 57.14%
RQ2
RQ5
RESULTS: MACHINE LEARNING
RANDOM FORESTS PERFORM BEST
F1 Acc
Relevance
Random Forest 0.78 0.93
XGBoost 0.76 0.91
Naïve Bayes 0.66 0.86
SVM 0.68 0.85
Neural Network 0.79 0.94
Authoritativeness
Random Forest 0.981 0.983
XGBoost 0.977 0.980
Naïve Bayes 0.974 0.977
SVM 0.975 0.979
Neural Network 0.982 0.984
RQ6
RESULTS: MACHINE LEARNING
RANDOM FORESTS PERFORM BEST
F1 Acc
Ease of Navigation
Random Forest 0.815 0.834
XGBoost 0.183 0.06
Naïve Bayes 0.811 0.806
SVM 0.780 0.824
Neural network 0.831 0.936
RQ6
KG embeddings trained on Wikidata did not improve the task overall
when compared to simple one-hot encoding
CONCLUSIONS
Crowdsourcing + ML works
Many external sources are high-quality
Bad references are mainly working links with bad
information
External references were much clearer to workers
There are topical biases in different languages,
leading quality to vary by language
ONGOING WORK
Editor study to
understand
how references
are chosen
Automatic
claim
verification
ENCOURAGING EDITOR
DIVERSITY, IMPROVING
RETENTION
Alghamdi, K., Shi, M. & Simperl, E. (2021). Learning to
recommend items to Wikidata editors. International
Semantic Web Conference, 163-181.
OVERVIEW
Content growth is
outpacing community
growth
Tenure and interest
diversity help with item
quality
Recommenders could
help retain new editors
Source: Lydia Pintcher @Wikidata
Pages per active editor
DATASETS
LONG TAIL DISTRIBUTION OF PARTICIPATION
All editors Editors w/
200+ items
WIKIDATAREC
MODEL
Recommendation
modelled as classification
Sparse data, editor is
interested in item if they
edit it, no explicit
feedback of lack of
interest
Features of items, items
relations and editors
RESULTS: WIKIDATAREC VS BASELINES
OUTPERFORMS COLLABORATIVE FILTERING AND CONTENT-
BASED BASELINES
Baselines: collaborative filtering (CF) and content-based
 GMF = neural network implementation of collaborative filtering (CF)
 BPR-MF = matrix factorization (MF), for recommenders with implicit feedback
 eALS = MF, for implicit feedback scenarios, optimises a different function
 YouTube DNN = leans item and relational embeddings in a neural network with
random initialisation
RESULTS: ABLATION STUDY
GRAPH FEATURES MAKE A DIFFERENCE
Performance increases when items and relations are considered
Item content contributed more than relations
WikidataRec w NMoR (neural mixture of representations) performs
significantly better than WikidataRec w/o NMoR
RESULTS: ACTIVE EDITORS
FURTHER PERFORMANCE INCREASE ON LESS SPARSE DATA
Wikidata-14M
Wikidata-Active-Editor
CONCLUSIONS
WikidataRec shows recommendations are
technically feasible
Alternative means to collect editor
preferences and learn about their
motivations needed because of data
sparsity
ONGOING WORK
Study with editors suggested editors work
within the same topics for some time, then
change to new topics
Extension to WikidataRec for topic
recommendations yields better results than
item recommendations
WikidataRecSeq taking for edit sequences
under development
HUMAN-MACHINE
INTERACTIONS IN
KNOWLEDGE ENGINEERING
Use of AI tools e.g. language
models in KG engineering:
explainability, trust
Use of KGs in downstream AI
applications e.g. through
embeddings
WHAT WIKIDATA
TEACHES US ABOUT
KNOWLEDGE
ENGINEERING
Wikidata logs are a
huge source of
observational data in
knowledge engineering
Sociotechnical methods
improve on purely
technical ones to address
community growth and
diversity, content
reliability
Many cross-disciplinary
methods e.g. psychology,
sociology, neurosciences
underexplored
WHAT
WIKIDATA
TEACHES US
ABOUT
KNOWLEDGE
ENGINEERING
• Piscopo, A., Phethean, C., & Simperl, E. (2017). What
Makes a Good Collaborative Knowledge Graph:
Group Composition and Quality in Wikidata.
International Conference on Social Informatics, 305-
322, Springer.
• Piscopo, A., & Simperl, E. (2018). Who Models the
World?: Collaborative Ontology Creation and User
Roles in Wikidata. Proceedings of the ACM on
Human-Computer Interaction, 2(CSCW), 141.
• Amaral, G., Piscopo, A., Kaffee, L. A., Odinaldo, R., &
Simperl, E. (2021). Assessing the quality of sources in
Wikidata across languages: a hybrid approach. ACM
JDIQ 13(4), 1-35.
• Alghamdi, K., Shi, M. & Simperl, E. (2021). Learning
to recommend items to Wikidata editors.
International Semantic Web Conference, 163-181.
• Kaffee, L. A., Vougiouklis, P., & Simperl, E. (2022).
Using natural language generation to bootstrap
missing Wikipedia articles: A human-centric
perspective. Semantic Web, 13(2), 163-194.

Weitere ähnliche Inhalte

Ähnlich wie What Wikidata teaches us about knowledge engineering

One does not simply crowdsource the Semantic Web: 10 years with people, URIs,...
One does not simply crowdsource the Semantic Web: 10 years with people, URIs,...One does not simply crowdsource the Semantic Web: 10 years with people, URIs,...
One does not simply crowdsource the Semantic Web: 10 years with people, URIs,...Elena Simperl
 
Describing Scholarly Contributions semantically with the Open Research Knowle...
Describing Scholarly Contributions semantically with the Open Research Knowle...Describing Scholarly Contributions semantically with the Open Research Knowle...
Describing Scholarly Contributions semantically with the Open Research Knowle...Sören Auer
 
2011 06-14 cristhian-parra_u_count
2011 06-14 cristhian-parra_u_count2011 06-14 cristhian-parra_u_count
2011 06-14 cristhian-parra_u_countCristhian Parra
 
Credit where credit is due: acknowledging all types of contributions
Credit where credit is due: acknowledging all types of contributionsCredit where credit is due: acknowledging all types of contributions
Credit where credit is due: acknowledging all types of contributionsmhaendel
 
How academic research on GitHub has evolved in the last several years
How academic research on GitHub has evolved in the last several yearsHow academic research on GitHub has evolved in the last several years
How academic research on GitHub has evolved in the last several yearsKeiko Ono
 
Open Access Statistics: An Examination how to Generate Interoperable Usage In...
Open Access Statistics: An Examination how to Generate Interoperable Usage In...Open Access Statistics: An Examination how to Generate Interoperable Usage In...
Open Access Statistics: An Examination how to Generate Interoperable Usage In...Daniel Beucke
 
Project E: Citation
Project E: CitationProject E: Citation
Project E: CitationLizLyon
 
Conf 2012-empirikom3
Conf 2012-empirikom3Conf 2012-empirikom3
Conf 2012-empirikom3Clay Spinuzzi
 
Research Objects: more than the sum of the parts
Research Objects: more than the sum of the partsResearch Objects: more than the sum of the parts
Research Objects: more than the sum of the partsCarole Goble
 
AI, Knowledge Representation and Graph Databases -
 Key Trends in Data Science
AI, Knowledge Representation and Graph Databases -
 Key Trends in Data ScienceAI, Knowledge Representation and Graph Databases -
 Key Trends in Data Science
AI, Knowledge Representation and Graph Databases -
 Key Trends in Data ScienceOptum
 
Sources of Change in Modern Knowledge Organization Systems
Sources of Change in Modern Knowledge Organization SystemsSources of Change in Modern Knowledge Organization Systems
Sources of Change in Modern Knowledge Organization SystemsPaul Groth
 
Crediting informatics and data folks in life science teams
Crediting informatics and data folks in life science teamsCrediting informatics and data folks in life science teams
Crediting informatics and data folks in life science teamsCarole Goble
 
Sci Know Mine 2013: What can we learn from topic modeling on 350M academic do...
Sci Know Mine 2013: What can we learn from topic modeling on 350M academic do...Sci Know Mine 2013: What can we learn from topic modeling on 350M academic do...
Sci Know Mine 2013: What can we learn from topic modeling on 350M academic do...William Gunn
 
Leveraging the Crowd: Supporting Newcomers to Build an OSS Community
Leveraging the Crowd: Supporting Newcomers to Build an OSS CommunityLeveraging the Crowd: Supporting Newcomers to Build an OSS Community
Leveraging the Crowd: Supporting Newcomers to Build an OSS CommunityMarco Aurelio Gerosa
 
Building Effective Visualization Shiny WVF
Building Effective Visualization Shiny WVFBuilding Effective Visualization Shiny WVF
Building Effective Visualization Shiny WVFOlga Scrivner
 
Shared Rationales in Group Activities
Shared Rationales in Group ActivitiesShared Rationales in Group Activities
Shared Rationales in Group ActivitiesLu Xiao
 

Ähnlich wie What Wikidata teaches us about knowledge engineering (20)

One does not simply crowdsource the Semantic Web: 10 years with people, URIs,...
One does not simply crowdsource the Semantic Web: 10 years with people, URIs,...One does not simply crowdsource the Semantic Web: 10 years with people, URIs,...
One does not simply crowdsource the Semantic Web: 10 years with people, URIs,...
 
VIVO 2011 Paper
VIVO 2011 PaperVIVO 2011 Paper
VIVO 2011 Paper
 
Describing Scholarly Contributions semantically with the Open Research Knowle...
Describing Scholarly Contributions semantically with the Open Research Knowle...Describing Scholarly Contributions semantically with the Open Research Knowle...
Describing Scholarly Contributions semantically with the Open Research Knowle...
 
2011 06-14 cristhian-parra_u_count
2011 06-14 cristhian-parra_u_count2011 06-14 cristhian-parra_u_count
2011 06-14 cristhian-parra_u_count
 
Credit where credit is due: acknowledging all types of contributions
Credit where credit is due: acknowledging all types of contributionsCredit where credit is due: acknowledging all types of contributions
Credit where credit is due: acknowledging all types of contributions
 
How academic research on GitHub has evolved in the last several years
How academic research on GitHub has evolved in the last several yearsHow academic research on GitHub has evolved in the last several years
How academic research on GitHub has evolved in the last several years
 
Digital repertoires of poetry metrics: towards a Linked Open Data ecosystem
Digital repertoires of poetry metrics: towards a Linked Open Data ecosystemDigital repertoires of poetry metrics: towards a Linked Open Data ecosystem
Digital repertoires of poetry metrics: towards a Linked Open Data ecosystem
 
Open Access Statistics: An Examination how to Generate Interoperable Usage In...
Open Access Statistics: An Examination how to Generate Interoperable Usage In...Open Access Statistics: An Examination how to Generate Interoperable Usage In...
Open Access Statistics: An Examination how to Generate Interoperable Usage In...
 
Project E: Citation
Project E: CitationProject E: Citation
Project E: Citation
 
Conf 2012-empirikom3
Conf 2012-empirikom3Conf 2012-empirikom3
Conf 2012-empirikom3
 
Research Objects: more than the sum of the parts
Research Objects: more than the sum of the partsResearch Objects: more than the sum of the parts
Research Objects: more than the sum of the parts
 
AI, Knowledge Representation and Graph Databases -
 Key Trends in Data Science
AI, Knowledge Representation and Graph Databases -
 Key Trends in Data ScienceAI, Knowledge Representation and Graph Databases -
 Key Trends in Data Science
AI, Knowledge Representation and Graph Databases -
 Key Trends in Data Science
 
Sources of Change in Modern Knowledge Organization Systems
Sources of Change in Modern Knowledge Organization SystemsSources of Change in Modern Knowledge Organization Systems
Sources of Change in Modern Knowledge Organization Systems
 
Crediting informatics and data folks in life science teams
Crediting informatics and data folks in life science teamsCrediting informatics and data folks in life science teams
Crediting informatics and data folks in life science teams
 
Sci Know Mine 2013: What can we learn from topic modeling on 350M academic do...
Sci Know Mine 2013: What can we learn from topic modeling on 350M academic do...Sci Know Mine 2013: What can we learn from topic modeling on 350M academic do...
Sci Know Mine 2013: What can we learn from topic modeling on 350M academic do...
 
BioNLPSADI
BioNLPSADIBioNLPSADI
BioNLPSADI
 
Leveraging the Crowd: Supporting Newcomers to Build an OSS Community
Leveraging the Crowd: Supporting Newcomers to Build an OSS CommunityLeveraging the Crowd: Supporting Newcomers to Build an OSS Community
Leveraging the Crowd: Supporting Newcomers to Build an OSS Community
 
Building Effective Visualization Shiny WVF
Building Effective Visualization Shiny WVFBuilding Effective Visualization Shiny WVF
Building Effective Visualization Shiny WVF
 
Shared Rationales in Group Activities
Shared Rationales in Group ActivitiesShared Rationales in Group Activities
Shared Rationales in Group Activities
 
Shamane-PhD-Defence-Final.pptx
Shamane-PhD-Defence-Final.pptxShamane-PhD-Defence-Final.pptx
Shamane-PhD-Defence-Final.pptx
 

Mehr von Elena Simperl

This talk was not generated with ChatGPT: how AI is changing science
This talk was not generated with ChatGPT: how AI is changing scienceThis talk was not generated with ChatGPT: how AI is changing science
This talk was not generated with ChatGPT: how AI is changing scienceElena Simperl
 
Knowledge graph use cases in natural language generation
Knowledge graph use cases in natural language generationKnowledge graph use cases in natural language generation
Knowledge graph use cases in natural language generationElena Simperl
 
Knowledge engineering: from people to machines and back
Knowledge engineering: from people to machines and backKnowledge engineering: from people to machines and back
Knowledge engineering: from people to machines and backElena Simperl
 
The web of data: how are we doing so far
The web of data: how are we doing so farThe web of data: how are we doing so far
The web of data: how are we doing so farElena Simperl
 
Open government data portals: from publishing to use and impact
Open government data portals: from publishing to use and impactOpen government data portals: from publishing to use and impact
Open government data portals: from publishing to use and impactElena Simperl
 
Ten myths about knowledge graphs.pdf
Ten myths about knowledge graphs.pdfTen myths about knowledge graphs.pdf
Ten myths about knowledge graphs.pdfElena Simperl
 
Data commons and their role in fighting misinformation.pdf
Data commons and their role in fighting misinformation.pdfData commons and their role in fighting misinformation.pdf
Data commons and their role in fighting misinformation.pdfElena Simperl
 
Are our knowledge graphs trustworthy?
Are our knowledge graphs trustworthy?Are our knowledge graphs trustworthy?
Are our knowledge graphs trustworthy?Elena Simperl
 
The web of data: how are we doing so far?
The web of data: how are we doing so far?The web of data: how are we doing so far?
The web of data: how are we doing so far?Elena Simperl
 
Crowdsourcing and citizen engagement for people-centric smart cities
Crowdsourcing and citizen engagement for people-centric smart citiesCrowdsourcing and citizen engagement for people-centric smart cities
Crowdsourcing and citizen engagement for people-centric smart citiesElena Simperl
 
Pie chart or pizza: identifying chart types and their virality on Twitter
Pie chart or pizza: identifying chart types and their virality on TwitterPie chart or pizza: identifying chart types and their virality on Twitter
Pie chart or pizza: identifying chart types and their virality on TwitterElena Simperl
 
High-value datasets: from publication to impact
High-value datasets: from publication to impactHigh-value datasets: from publication to impact
High-value datasets: from publication to impactElena Simperl
 
The story of Data Stories
The story of Data StoriesThe story of Data Stories
The story of Data StoriesElena Simperl
 
The human face of AI: how collective and augmented intelligence can help sol...
The human face of AI:  how collective and augmented intelligence can help sol...The human face of AI:  how collective and augmented intelligence can help sol...
The human face of AI: how collective and augmented intelligence can help sol...Elena Simperl
 
Qrowd and the city: designing people-centric smart cities
Qrowd and the city: designing people-centric smart citiesQrowd and the city: designing people-centric smart cities
Qrowd and the city: designing people-centric smart citiesElena Simperl
 
Inclusive cities: a crowdsourcing approach
Inclusive cities: a crowdsourcing approachInclusive cities: a crowdsourcing approach
Inclusive cities: a crowdsourcing approachElena Simperl
 
Making transport smarter, leveraging the human factor
Making transport smarter, leveraging the human factorMaking transport smarter, leveraging the human factor
Making transport smarter, leveraging the human factorElena Simperl
 

Mehr von Elena Simperl (20)

This talk was not generated with ChatGPT: how AI is changing science
This talk was not generated with ChatGPT: how AI is changing scienceThis talk was not generated with ChatGPT: how AI is changing science
This talk was not generated with ChatGPT: how AI is changing science
 
Knowledge graph use cases in natural language generation
Knowledge graph use cases in natural language generationKnowledge graph use cases in natural language generation
Knowledge graph use cases in natural language generation
 
Knowledge engineering: from people to machines and back
Knowledge engineering: from people to machines and backKnowledge engineering: from people to machines and back
Knowledge engineering: from people to machines and back
 
The web of data: how are we doing so far
The web of data: how are we doing so farThe web of data: how are we doing so far
The web of data: how are we doing so far
 
Open government data portals: from publishing to use and impact
Open government data portals: from publishing to use and impactOpen government data portals: from publishing to use and impact
Open government data portals: from publishing to use and impact
 
Ten myths about knowledge graphs.pdf
Ten myths about knowledge graphs.pdfTen myths about knowledge graphs.pdf
Ten myths about knowledge graphs.pdf
 
Data commons and their role in fighting misinformation.pdf
Data commons and their role in fighting misinformation.pdfData commons and their role in fighting misinformation.pdf
Data commons and their role in fighting misinformation.pdf
 
Are our knowledge graphs trustworthy?
Are our knowledge graphs trustworthy?Are our knowledge graphs trustworthy?
Are our knowledge graphs trustworthy?
 
The web of data: how are we doing so far?
The web of data: how are we doing so far?The web of data: how are we doing so far?
The web of data: how are we doing so far?
 
Crowdsourcing and citizen engagement for people-centric smart cities
Crowdsourcing and citizen engagement for people-centric smart citiesCrowdsourcing and citizen engagement for people-centric smart cities
Crowdsourcing and citizen engagement for people-centric smart cities
 
Pie chart or pizza: identifying chart types and their virality on Twitter
Pie chart or pizza: identifying chart types and their virality on TwitterPie chart or pizza: identifying chart types and their virality on Twitter
Pie chart or pizza: identifying chart types and their virality on Twitter
 
High-value datasets: from publication to impact
High-value datasets: from publication to impactHigh-value datasets: from publication to impact
High-value datasets: from publication to impact
 
The story of Data Stories
The story of Data StoriesThe story of Data Stories
The story of Data Stories
 
The human face of AI: how collective and augmented intelligence can help sol...
The human face of AI:  how collective and augmented intelligence can help sol...The human face of AI:  how collective and augmented intelligence can help sol...
The human face of AI: how collective and augmented intelligence can help sol...
 
Qrowd and the city: designing people-centric smart cities
Qrowd and the city: designing people-centric smart citiesQrowd and the city: designing people-centric smart cities
Qrowd and the city: designing people-centric smart cities
 
Qrowd and the city
Qrowd and the cityQrowd and the city
Qrowd and the city
 
Inclusive cities: a crowdsourcing approach
Inclusive cities: a crowdsourcing approachInclusive cities: a crowdsourcing approach
Inclusive cities: a crowdsourcing approach
 
The data we want
The data we wantThe data we want
The data we want
 
Data stories
Data storiesData stories
Data stories
 
Making transport smarter, leveraging the human factor
Making transport smarter, leveraging the human factorMaking transport smarter, leveraging the human factor
Making transport smarter, leveraging the human factor
 

Kürzlich hochgeladen

Dev Dives: Streamline document processing with UiPath Studio Web
Dev Dives: Streamline document processing with UiPath Studio WebDev Dives: Streamline document processing with UiPath Studio Web
Dev Dives: Streamline document processing with UiPath Studio WebUiPathCommunity
 
Artificial intelligence in cctv survelliance.pptx
Artificial intelligence in cctv survelliance.pptxArtificial intelligence in cctv survelliance.pptx
Artificial intelligence in cctv survelliance.pptxhariprasad279825
 
Bun (KitWorks Team Study 노별마루 발표 2024.4.22)
Bun (KitWorks Team Study 노별마루 발표 2024.4.22)Bun (KitWorks Team Study 노별마루 발표 2024.4.22)
Bun (KitWorks Team Study 노별마루 발표 2024.4.22)Wonjun Hwang
 
SAP Build Work Zone - Overview L2-L3.pptx
SAP Build Work Zone - Overview L2-L3.pptxSAP Build Work Zone - Overview L2-L3.pptx
SAP Build Work Zone - Overview L2-L3.pptxNavinnSomaal
 
SIP trunking in Janus @ Kamailio World 2024
SIP trunking in Janus @ Kamailio World 2024SIP trunking in Janus @ Kamailio World 2024
SIP trunking in Janus @ Kamailio World 2024Lorenzo Miniero
 
Pigging Solutions in Pet Food Manufacturing
Pigging Solutions in Pet Food ManufacturingPigging Solutions in Pet Food Manufacturing
Pigging Solutions in Pet Food ManufacturingPigging Solutions
 
Gen AI in Business - Global Trends Report 2024.pdf
Gen AI in Business - Global Trends Report 2024.pdfGen AI in Business - Global Trends Report 2024.pdf
Gen AI in Business - Global Trends Report 2024.pdfAddepto
 
Commit 2024 - Secret Management made easy
Commit 2024 - Secret Management made easyCommit 2024 - Secret Management made easy
Commit 2024 - Secret Management made easyAlfredo García Lavilla
 
"Federated learning: out of reach no matter how close",Oleksandr Lapshyn
"Federated learning: out of reach no matter how close",Oleksandr Lapshyn"Federated learning: out of reach no matter how close",Oleksandr Lapshyn
"Federated learning: out of reach no matter how close",Oleksandr LapshynFwdays
 
AI as an Interface for Commercial Buildings
AI as an Interface for Commercial BuildingsAI as an Interface for Commercial Buildings
AI as an Interface for Commercial BuildingsMemoori
 
Developer Data Modeling Mistakes: From Postgres to NoSQL
Developer Data Modeling Mistakes: From Postgres to NoSQLDeveloper Data Modeling Mistakes: From Postgres to NoSQL
Developer Data Modeling Mistakes: From Postgres to NoSQLScyllaDB
 
Streamlining Python Development: A Guide to a Modern Project Setup
Streamlining Python Development: A Guide to a Modern Project SetupStreamlining Python Development: A Guide to a Modern Project Setup
Streamlining Python Development: A Guide to a Modern Project SetupFlorian Wilhelm
 
Integration and Automation in Practice: CI/CD in Mule Integration and Automat...
Integration and Automation in Practice: CI/CD in Mule Integration and Automat...Integration and Automation in Practice: CI/CD in Mule Integration and Automat...
Integration and Automation in Practice: CI/CD in Mule Integration and Automat...Patryk Bandurski
 
Designing IA for AI - Information Architecture Conference 2024
Designing IA for AI - Information Architecture Conference 2024Designing IA for AI - Information Architecture Conference 2024
Designing IA for AI - Information Architecture Conference 2024Enterprise Knowledge
 
Kotlin Multiplatform & Compose Multiplatform - Starter kit for pragmatics
Kotlin Multiplatform & Compose Multiplatform - Starter kit for pragmaticsKotlin Multiplatform & Compose Multiplatform - Starter kit for pragmatics
Kotlin Multiplatform & Compose Multiplatform - Starter kit for pragmaticscarlostorres15106
 
Beyond Boundaries: Leveraging No-Code Solutions for Industry Innovation
Beyond Boundaries: Leveraging No-Code Solutions for Industry InnovationBeyond Boundaries: Leveraging No-Code Solutions for Industry Innovation
Beyond Boundaries: Leveraging No-Code Solutions for Industry InnovationSafe Software
 
Unleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding ClubUnleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding ClubKalema Edgar
 
Connect Wave/ connectwave Pitch Deck Presentation
Connect Wave/ connectwave Pitch Deck PresentationConnect Wave/ connectwave Pitch Deck Presentation
Connect Wave/ connectwave Pitch Deck PresentationSlibray Presentation
 
Automating Business Process via MuleSoft Composer | Bangalore MuleSoft Meetup...
Automating Business Process via MuleSoft Composer | Bangalore MuleSoft Meetup...Automating Business Process via MuleSoft Composer | Bangalore MuleSoft Meetup...
Automating Business Process via MuleSoft Composer | Bangalore MuleSoft Meetup...shyamraj55
 

Kürzlich hochgeladen (20)

Dev Dives: Streamline document processing with UiPath Studio Web
Dev Dives: Streamline document processing with UiPath Studio WebDev Dives: Streamline document processing with UiPath Studio Web
Dev Dives: Streamline document processing with UiPath Studio Web
 
Artificial intelligence in cctv survelliance.pptx
Artificial intelligence in cctv survelliance.pptxArtificial intelligence in cctv survelliance.pptx
Artificial intelligence in cctv survelliance.pptx
 
Bun (KitWorks Team Study 노별마루 발표 2024.4.22)
Bun (KitWorks Team Study 노별마루 발표 2024.4.22)Bun (KitWorks Team Study 노별마루 발표 2024.4.22)
Bun (KitWorks Team Study 노별마루 발표 2024.4.22)
 
SAP Build Work Zone - Overview L2-L3.pptx
SAP Build Work Zone - Overview L2-L3.pptxSAP Build Work Zone - Overview L2-L3.pptx
SAP Build Work Zone - Overview L2-L3.pptx
 
SIP trunking in Janus @ Kamailio World 2024
SIP trunking in Janus @ Kamailio World 2024SIP trunking in Janus @ Kamailio World 2024
SIP trunking in Janus @ Kamailio World 2024
 
Pigging Solutions in Pet Food Manufacturing
Pigging Solutions in Pet Food ManufacturingPigging Solutions in Pet Food Manufacturing
Pigging Solutions in Pet Food Manufacturing
 
Gen AI in Business - Global Trends Report 2024.pdf
Gen AI in Business - Global Trends Report 2024.pdfGen AI in Business - Global Trends Report 2024.pdf
Gen AI in Business - Global Trends Report 2024.pdf
 
Commit 2024 - Secret Management made easy
Commit 2024 - Secret Management made easyCommit 2024 - Secret Management made easy
Commit 2024 - Secret Management made easy
 
"Federated learning: out of reach no matter how close",Oleksandr Lapshyn
"Federated learning: out of reach no matter how close",Oleksandr Lapshyn"Federated learning: out of reach no matter how close",Oleksandr Lapshyn
"Federated learning: out of reach no matter how close",Oleksandr Lapshyn
 
Hot Sexy call girls in Panjabi Bagh 🔝 9953056974 🔝 Delhi escort Service
Hot Sexy call girls in Panjabi Bagh 🔝 9953056974 🔝 Delhi escort ServiceHot Sexy call girls in Panjabi Bagh 🔝 9953056974 🔝 Delhi escort Service
Hot Sexy call girls in Panjabi Bagh 🔝 9953056974 🔝 Delhi escort Service
 
AI as an Interface for Commercial Buildings
AI as an Interface for Commercial BuildingsAI as an Interface for Commercial Buildings
AI as an Interface for Commercial Buildings
 
Developer Data Modeling Mistakes: From Postgres to NoSQL
Developer Data Modeling Mistakes: From Postgres to NoSQLDeveloper Data Modeling Mistakes: From Postgres to NoSQL
Developer Data Modeling Mistakes: From Postgres to NoSQL
 
Streamlining Python Development: A Guide to a Modern Project Setup
Streamlining Python Development: A Guide to a Modern Project SetupStreamlining Python Development: A Guide to a Modern Project Setup
Streamlining Python Development: A Guide to a Modern Project Setup
 
Integration and Automation in Practice: CI/CD in Mule Integration and Automat...
Integration and Automation in Practice: CI/CD in Mule Integration and Automat...Integration and Automation in Practice: CI/CD in Mule Integration and Automat...
Integration and Automation in Practice: CI/CD in Mule Integration and Automat...
 
Designing IA for AI - Information Architecture Conference 2024
Designing IA for AI - Information Architecture Conference 2024Designing IA for AI - Information Architecture Conference 2024
Designing IA for AI - Information Architecture Conference 2024
 
Kotlin Multiplatform & Compose Multiplatform - Starter kit for pragmatics
Kotlin Multiplatform & Compose Multiplatform - Starter kit for pragmaticsKotlin Multiplatform & Compose Multiplatform - Starter kit for pragmatics
Kotlin Multiplatform & Compose Multiplatform - Starter kit for pragmatics
 
Beyond Boundaries: Leveraging No-Code Solutions for Industry Innovation
Beyond Boundaries: Leveraging No-Code Solutions for Industry InnovationBeyond Boundaries: Leveraging No-Code Solutions for Industry Innovation
Beyond Boundaries: Leveraging No-Code Solutions for Industry Innovation
 
Unleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding ClubUnleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding Club
 
Connect Wave/ connectwave Pitch Deck Presentation
Connect Wave/ connectwave Pitch Deck PresentationConnect Wave/ connectwave Pitch Deck Presentation
Connect Wave/ connectwave Pitch Deck Presentation
 
Automating Business Process via MuleSoft Composer | Bangalore MuleSoft Meetup...
Automating Business Process via MuleSoft Composer | Bangalore MuleSoft Meetup...Automating Business Process via MuleSoft Composer | Bangalore MuleSoft Meetup...
Automating Business Process via MuleSoft Composer | Bangalore MuleSoft Meetup...
 

What Wikidata teaches us about knowledge engineering

  • 1. WHAT WIKIDATA TEACHES US ABOUT KNOWLEDGE ENGINEERING Elena Simperl King’s College London, UK @esimperl ISWS2022 July 2022
  • 2. OVERVIEW Knowledge graphs have become a critical AI resource We study them associo- technical constructs Our research  Explores knowledge-graph engineering empirically to understand practices and improve processes and tools  Uncovers links between social and technical qualities of knowledge graphs to make them better Picture from https://medium.com/@sderymail/challenges-of-knowledge-graph-part-1-d9ffe9e35214
  • 3.
  • 4.
  • 5.
  • 6. KNOWLEDGE ENGINEERING AS HUMAN- MACHINE PROCESS Human- machine interactions Knowledge engineering communities
  • 7. WIKIDATA Collaborative knowledge graph, Wikimedia project (2012), 23k active users, 99m items, 1.6b edits Open license RDF support, links to LOD cloud
  • 8. THE KNOWLEDGE GRAPH STATEMENTS, ITEMS, PROPERTIES Item identifiers start with a Q, property identifiers start with a P 8 Q84 London Q334155 Sadiq Khan P6 head of government
  • 9. THE KNOWLEDGE GRAPH ITEMS CAN BE CLASSES, ENTITIES, VALUES 9 Q7259 Ada Lovelace Q84 London Q334155 Sadiq Khan P6 head of government Q727 Amsterdam Q515 city Q6581097 male Q59360 Labour party Q145 United Kingdom
  • 10. THE KNOWLEDGE GRAPH STATEMENTS IN CONTEXT Statements may include context  Qualifiers (optional)  References (required) Two types of references  Internal, linking to another item  External, linking to webpage 10 Q84 London Q334155 Sadiq Khan P6 head of government 9 May 2016 london.gov.uk
  • 11. THE KNOWLEDGE GRAPH CO-EDITED BY BOTS AND HUMANS Human editors can register or work anonymously Community creates bots to automate routine tasks 23k active human users, 340+ bots
  • 12. IN THIS TALK Assembling your KG team KG quality as a function of data provenance Supporting editors and improving retention
  • 13. ‘ONTOLOGIES ARE US’ Piscopo, A., Phethean, C., & Simperl, E. (2017). What Makes a Good Collaborative Knowledge Graph: Group Composition and Quality in Wikidata. International Conference on Social Informatics, 305-322, Springer. Piscopo, A., & Simperl, E. (2018). Who Models the World?: Collaborative Ontology Creation and User Roles in Wikidata. Proceedings of the ACM on Human-Computer Interaction, 2(CSCW), 141.
  • 14. BACKGROUND Editors have varied tenure and interests Their profile and editing behaviour impact outcomes  Group composition has varied effects  Tenure and interest diversity can increase outcome quality and group productivity  Different editors groups focus on different types of activities Chen, J., Ren, Y., Riedl, J.: The effects of diversity on group productivity and member withdrawal in online volunteer groups. In: Proceedings of the 28th international conference on human factors in computing systems - CHI ’10. p. 821. ACM Press, New York, USA (2010)
  • 15. STUDY 1: ITEM QUALITY Analysed edit history of items Corpus of 5k items, whose quality has been manually assessed (5 levels)* Edit history focused on community make-up Community is defined as set of editors of item Considered features from group diversity literature and Wikidata-specific aspects *https://www.wikidata.org/wiki/Wikidata:Item_quality
  • 16. RESEARCH HYPOTHESES Activity Outcome H1 Bots edits Item quality H2 Bot-human interaction Item quality H3 Anonymous edits Item quality H4 Tenure diversity Item quality H5 Interest diversity Item quality
  • 17. DATA AND METHODS Ordinal regression analysis, trained four models Dependent variable: 5k labelled Wikidata items Independent variables  Proportion of bot edits  Bot human edit proportion  Proportion of anonymous edits  Tenure diversity: Coefficient of variation  Interest diversity: User editing matrix Control variables: group size, item age
  • 19. SUMMARY AND IMPLICATIONS The more is not always the merrier 01 Bot edits are key for quality, but bots and humans are better 02 Registered editors have a positive impact Diversity matters 04 Encourage registration 01 Identify further areas for bot editing 02 Design effective human-bot workflows 03 Suggest items to edit based on tenure and interests 04 03
  • 20. ONGOING WORK Analysing how the community awards quality labels to items Developing robust machine learning models Spotting biases
  • 21. STUDY 2: ONTOLOGY QUALITY Analysed Wikidata ontology and its edit context Defined as the graph of all items linked through P31 (instance of) & P279 (subclass of) Calculated evolution of quality indicators and editing activity in time and the links between them Based on features from literature on ontology evaluation and collaborative ontology engineering
  • 22. DATA AND METHODS Wikidata dumps from March 2013 (creation of P279) to September 2017  Analysed data in 55 monthly time frames Literature survey to define Wikidata ontology quality framework Clustering to identify ontology editor roles Lagged multiple regression to link roles and ontology features  Dependent variable: Changes in ontology quality across time  Independent variables: Number of edits by different roles  Control variables: Bot and anonymous edits
  • 23. ONTOLOGY QUALITY: INDICATORS Based on 7 ontology evaluation frameworks Compiled structural indicators that can be computed from the data 23 Indicator Description Indicator Description noi Number of instances ap; mp Average and median population noc Number of classes rr Relationship richness norc Number of root classes ir, mr Inheritance and median richness nolc Number of leaf classes cr Class richness nop Number of properties ad, md, maxd Average, median, max explicit depth Sicilia, M. A., Rodríguez, D., García-Barriocanal, E., & Sánchez-Alonso, S. (2012). Empirical findings on ontology metrics. Expert Systems with Applications, 39(8), 6706-6711.
  • 24. ONTOLOGY QUALITY: RESULTS LARGE ONTOLOGY, UNEVEN QUALITY >1.5M classes, ~4000 properties No of classes increases at same rate as overall no of items, likely due to users incorrectly using P31 & P279 ap and cr decrease over time, several classes are either without instances, sub-classes or both ir & maxd increase over time, part of the Wikidata ontology is distributed vertically 24
  • 25. EDITOR ROLES: METHODS K-means, features based on previous studies in ontology engineering Analysis by yearly cohort 25 Feature Description Feature Description # edits Total number of edits per month. # property edits Total number of edits on Properties in a month. # ontology edits Number of edits on classes. # taxonomy edits Number of edits on P31 and P279 statements. # discussion edits Number of edits on talk pages. p batch edits Number of edits done through automated tools. # modifying editsNumber of revisions on previously existing statements. item diversity Proportion between number of edits and number of items edited. admin True if user in an admin user group, false otherwise. lower admin True if user in a user group with enhanced user rights, false otherwise.
  • 26. EDITOR ROLES: RESULTS 190,765 unique editors over 55 months (783k total) 18k editors active for 10+ months 2 clusters, obtained using gap statistic (tested 2≥k≥8) Leaders: more active minority (~1%), higher number of contributions to ontology, engaged within the community Contributors: less active, lower number of contributions to ontology and lower proportion of batch edits 26
  • 27. EDITOR ROLES: RESULTS People who joined the project early tend to be more active & are more likely to become leaders Levels of activity of leaders decrease over time (alternatively, people move on to different tasks) 27
  • 28. RESEARCH HYPOTHESES H1 Higher levels of leader activity are negatively correlated to number of classes (noc), number of root classes (norc), and number of leaf classes (nolc) H2 Higher levels of leader activity are positively correlated to inheritance richness (ir), average population (ap), and average depth (ad) 28
  • 29. ROLES & ONTOLOGY: RESULTS H1 not supported H2 partially supported Only inheritance richness (ir) and average depth (ad) related significantly with leader edits (p<0.01) Bot edits significantly and positively affect the number of subclasses and instances per class (ir & ap) (p<0.05) 29
  • 30. CONCLUSIONS Creating ontologies still a challenging task Size of ontology renders existing automatic quality assessment methods unfeasible Broader curation efforts needed: large number of empty classes Editor roles less well articulated than in other ontology engineering projects Possible decline in motivation after several months
  • 31. ONGOING WORK The role of discussions in Wikidata Studying participation pathways to observe learning effects Analysis of ShEx schemas
  • 32. THE CONTENT IS AS GOOD AS ITS REFERENCES Amaral, G., Piscopo, A., Kaffee, L. A., Odinaldo, R., & Simperl, E. (2021). Assessing the quality of sources in Wikidata across languages: a hybrid approach. ACM JDIQ 13(4), 1-35. 32
  • 33. PROVENANCE IN WIKIDATA Statements may include context  Qualifiers (optional)  References (required) Two types of references  Internal, linking to another item  External, linking to webpage 33 Q84 London Q334155 Sadiq Khan P6 head of government 9 May 2016 london.gov.uk
  • 34. THE ROLE OF PROVENANCE Wikidata aims to become a hub of references Provenance increases trust in Wikidata Lack of provenance hinders downstream reuse Quality of references underexplored Hartig, O. (2009). Provenance Information in the Web of Data. LDOW, 538.
  • 35. STUDY 3 Approach to evaluate quality of external and internal references in Wikidata Quality defined by the Wikidata verifiability policy  Relevant: support the statement they are attached to  Authoritative: trustworthy, up-to-date, and free of bias for supporting a particular statement  Easy to access: users should be able to access the information in references with low perceived effort Large-scale (the whole of Wikidata) Multiple languages (EN, ES, PT, SV, NL, JA)
  • 36. RESEARCH QUESTIONS RQ1 How easy it is to access sources? RQ2 How authoritative are sources?  According to existing guidance, do they match the author and publisher types from the Wikidata policy? RQ3 Are sources relevant to claims?
  • 37. RESEARCH QUESTIONS RQ4 Are external and internal references of varying quality? RQ5 Are references in different languages of varying quality? RQ6 Can we predict relevance, authoritativeness and ease of use of a reference without relying on the content of the references themselves? Can we do this for references in different languages?
  • 38. METHODS THREE STAGE MIXED APPROACH 1. Descriptive statistics Understand differences in external/internal references Determine data distributions Language Predicates Domains Prepare data to crowdsource
  • 39. METHODS THREE STAGE MIXED APPROACH 2. Microtask crowdsourcing Evaluate relevance, authoritativeness & ease of access of a reference sample Create training set for ML 3. Machine learning Large-scale reference quality prediction RQs 1-5 RQ6
  • 40. MICROTASK CROWDSOURCING 2 tasks on Amazon’s Mechanical Turk 5 workers/task, majority voting Quality assurance through golden data & attention checks 40 Feature Microtask Description Relevance / Ease of access T1.1 Does the reference support the statement? T1.2 How easy was the navigation to verify the reference? T1.3 What was the biggest barrier to access, if any? Authoritativeness T2.1 Choose author type from list T2.2 Choose publisher type from list T2.3 Choose publisher sub-type from list RQs 1-5
  • 41. MACHINE LEARNING Compared 5 algorithms  Naïve Bayes, Random Forest, SVM, XGBoost, Neural Network Features taken from URLs and Wikidata ontology 3 tasks: relevance, ease of navigation (data from T1.2), and authoritativeness RQ6 Features Features of the representative URL extracted Features of the reference node’s coding and structure Features of the website available through the representative URL Features of the claim node associated to this reference node
  • 42. DATA Snapshot up to April 16th, 2020 15m individual items used in Stage 1 (20% of total) 385 random references from each language (95% conf., 5% m. of error) 2,310 references in total  Six languages, both internal and external  Around 940 could be automatically checked (40% of total)
  • 43. RESULTS: CROWDSOURCING CROWDSOURCING WORKS Agreement calculated per question and language Agreement overall moderate (over 0.4) and often substantial (over 0.6)  Multiple trusted metrics used for robustness * Kappa metrics can’t be used for interval data Task No. of microtasks Total workers Fleiss’ k Randolph’s k Kripp’s α T1.1 460 microtasks 87 0.507702 0.745626 0.537647 T1.2 N/A* N/A* 0.614661 T1.3 0.405330 0.592810 0.415045 T2.1 370 microtasks 74 0.496904 0.781481 0.645795 T2.2 0.603499 0.695833 0.657171 T2.3 0.515551 0.604419 0.572782
  • 44. RESULTS: CROWDSOURCING MAJORITY OF REFERENCES ARE HIGH QUALITY 2310 references evaluated Found 91.73% are relevant Found 66.97% are authoritative Most (95%) unauthoritative references link to Wikipedia On ease of access: 5-level Likert Scale was used 78.43% fall on levels 3 (easy) or 4 (very easy) 6.66% fall on levels 0 (very difficult) and 1 (difficult) 76.44% of irrelevant references were due to working links with wrong information (different from Wikidata) 19.37% were due to bad links RQ1 RQ2 RQ3
  • 45. RESULTS: CROWDSOURCING VARYING REFERENCE QUALITY RQ4  External references had better agreement than internal  10% better on average (Krippendorff’s Alpha)  No noticeable quality difference between external and internal references  Quality differs a lot based on language  Examples:  English references are much harder to access (19.37% voted as very difficult)  Non-authoritative references: 4.94% for Dutch , but 43.12% for Japanese  Likely due to specific languages used for specific topics RQ5
  • 46. RESULTS: CROWDSOURCING DATA FROM GOVERNMENT AND ACADEMIC SOURCES Most common author type (T2.1)  Organization (67%) Most common publisher types (T2.2)  Companies/Organizations (37.75%)  Self-published (28.79%)  Academic/Scientific (22.16%) This also varies based on language  Academic/Scientific for English is 83.12%  Self-published for Swedish is 57.14% RQ2 RQ5
  • 47. RESULTS: MACHINE LEARNING RANDOM FORESTS PERFORM BEST F1 Acc Relevance Random Forest 0.78 0.93 XGBoost 0.76 0.91 Naïve Bayes 0.66 0.86 SVM 0.68 0.85 Neural Network 0.79 0.94 Authoritativeness Random Forest 0.981 0.983 XGBoost 0.977 0.980 Naïve Bayes 0.974 0.977 SVM 0.975 0.979 Neural Network 0.982 0.984 RQ6
  • 48. RESULTS: MACHINE LEARNING RANDOM FORESTS PERFORM BEST F1 Acc Ease of Navigation Random Forest 0.815 0.834 XGBoost 0.183 0.06 Naïve Bayes 0.811 0.806 SVM 0.780 0.824 Neural network 0.831 0.936 RQ6 KG embeddings trained on Wikidata did not improve the task overall when compared to simple one-hot encoding
  • 49. CONCLUSIONS Crowdsourcing + ML works Many external sources are high-quality Bad references are mainly working links with bad information External references were much clearer to workers There are topical biases in different languages, leading quality to vary by language
  • 50. ONGOING WORK Editor study to understand how references are chosen Automatic claim verification
  • 51. ENCOURAGING EDITOR DIVERSITY, IMPROVING RETENTION Alghamdi, K., Shi, M. & Simperl, E. (2021). Learning to recommend items to Wikidata editors. International Semantic Web Conference, 163-181.
  • 52. OVERVIEW Content growth is outpacing community growth Tenure and interest diversity help with item quality Recommenders could help retain new editors Source: Lydia Pintcher @Wikidata Pages per active editor
  • 53. DATASETS LONG TAIL DISTRIBUTION OF PARTICIPATION All editors Editors w/ 200+ items
  • 54. WIKIDATAREC MODEL Recommendation modelled as classification Sparse data, editor is interested in item if they edit it, no explicit feedback of lack of interest Features of items, items relations and editors
  • 55. RESULTS: WIKIDATAREC VS BASELINES OUTPERFORMS COLLABORATIVE FILTERING AND CONTENT- BASED BASELINES Baselines: collaborative filtering (CF) and content-based  GMF = neural network implementation of collaborative filtering (CF)  BPR-MF = matrix factorization (MF), for recommenders with implicit feedback  eALS = MF, for implicit feedback scenarios, optimises a different function  YouTube DNN = leans item and relational embeddings in a neural network with random initialisation
  • 56. RESULTS: ABLATION STUDY GRAPH FEATURES MAKE A DIFFERENCE Performance increases when items and relations are considered Item content contributed more than relations WikidataRec w NMoR (neural mixture of representations) performs significantly better than WikidataRec w/o NMoR
  • 57. RESULTS: ACTIVE EDITORS FURTHER PERFORMANCE INCREASE ON LESS SPARSE DATA Wikidata-14M Wikidata-Active-Editor
  • 58. CONCLUSIONS WikidataRec shows recommendations are technically feasible Alternative means to collect editor preferences and learn about their motivations needed because of data sparsity
  • 59. ONGOING WORK Study with editors suggested editors work within the same topics for some time, then change to new topics Extension to WikidataRec for topic recommendations yields better results than item recommendations WikidataRecSeq taking for edit sequences under development
  • 60. HUMAN-MACHINE INTERACTIONS IN KNOWLEDGE ENGINEERING Use of AI tools e.g. language models in KG engineering: explainability, trust Use of KGs in downstream AI applications e.g. through embeddings
  • 61. WHAT WIKIDATA TEACHES US ABOUT KNOWLEDGE ENGINEERING Wikidata logs are a huge source of observational data in knowledge engineering Sociotechnical methods improve on purely technical ones to address community growth and diversity, content reliability Many cross-disciplinary methods e.g. psychology, sociology, neurosciences underexplored
  • 62. WHAT WIKIDATA TEACHES US ABOUT KNOWLEDGE ENGINEERING • Piscopo, A., Phethean, C., & Simperl, E. (2017). What Makes a Good Collaborative Knowledge Graph: Group Composition and Quality in Wikidata. International Conference on Social Informatics, 305- 322, Springer. • Piscopo, A., & Simperl, E. (2018). Who Models the World?: Collaborative Ontology Creation and User Roles in Wikidata. Proceedings of the ACM on Human-Computer Interaction, 2(CSCW), 141. • Amaral, G., Piscopo, A., Kaffee, L. A., Odinaldo, R., & Simperl, E. (2021). Assessing the quality of sources in Wikidata across languages: a hybrid approach. ACM JDIQ 13(4), 1-35. • Alghamdi, K., Shi, M. & Simperl, E. (2021). Learning to recommend items to Wikidata editors. International Semantic Web Conference, 163-181. • Kaffee, L. A., Vougiouklis, P., & Simperl, E. (2022). Using natural language generation to bootstrap missing Wikipedia articles: A human-centric perspective. Semantic Web, 13(2), 163-194.