SlideShare ist ein Scribd-Unternehmen logo
1 von 40
Challenges and opportunities in research
evaluation: Toward a better evaluation
environment
Sergio Benedetto
Consiglio Direttivo ANVUR
sergio.benedetto@anvur.org
Barcelona, May 18, 2015
Research evaluation
2
A taxonomy
3
Research evaluation
Ex ante Ex post
before research takes place, to assess its
potential relevance, the prospects of
success and the cost appropriateness
after research has been concluded, to
assess its results in terms of scientific
quality and impact
Comparative Individual
aimed at defining a ranking of
individuals, research groups or
HEIs, often within a
homogeneous area of research
comparing qualification against
a threshold to promote or not
individuals, research groups
National research assessments
• What? …..Evaluated objects…
• Why? …..Goals…
• How?....Evaluation methodologies…
• When?....Continuity, frequency…
• With what consequences?..... On institutions, on
researchers, on society…
4
Research evaluation
5
What ?
• The volume of the scientific outcomes
• Their quality
• Their scientific impact
• Their impacts on the economy, society and/or culture
• More in general, the so-called “third mission” of HEIs, i.e. their
involvement with society at large
Objects of research evaluation
6
Research evaluation
7
Why ?
To inform HEI government bodies and other stakeholders about
the status of national research
To help Ministry of Education and Research (or other national
bodies) distributing resources to HEIs
To help HEI government bodies taking strategic decisions to
improve the quality and effectiveness of research and in internal
resource (positions, funds) assignment
Goals of comparative research evaluation
8
Research evaluation
9
How ?
10
Sample of research
outcomes
Expert
panels
Reading Bibliometry
Informed peer review
External peer
review
Bibliometry
Final evaluation Final evaluation
Criteria
Research evaluation
11
When?
RAE-REF in UK: 1986, 1989, 1992, 1996, 2001, 2008, 2014
VTR-VQR in Italy: 2006, 2013, 2017 (?)
ERA in Australia: 2010, 2012, 2015
The period of research evaluation
12
3 43 5 7 6
7 4
2 3
Research evaluation
13
With what
consequences?
No measuring technique leaves the measured object
unaffected, so:
14
• What are the intended consequences?
• Evaluation leads to an improvement in the quality of research (UK: Adams
& Gurney, 2010, Australia: Butler, 2003)
• Evaluation modifies dissemination channels, e.g., it makes the journal
article published in highly ranked journals the main publication outlet (Rin,
2009)
• Resources distribution to HEI based on assessment outcomes (UK, Italy,…)
• Improved HEIs infrastructure and archival repositories
• Enhanced “quality” recruitment
• Strategic positioning and more consistent policies
• Trust from society
No measuring technique leaves the measured object
unaffected, so:
15
• What could be the risks and unintended consequences?
• Worse publication practices: excessive segmentation of the
research results, clinging to the mainstream, safe disciplinary
research, citation stacking , coercive citation, ...)
• Research freedom limitation: too much emphasis on
accountability
• Underestimating the teaching activity (J. Warner, 1998)
• Misuse of assessment outcomes: evaluating individuals, apply
national-level criteria automatically to local issues
• …
• Collecting data and objects:
– Local (often incompatible) repositories
– Copyright issues
• Cleaning data:
– Human errors in uploading
– Names ambiguity
– Duplicated records
• Connecting data to “owners”:
– Researchers, institutions
• Transferring data and objects to evaluators (panels, peer reviewers,…)
– IP protection, “big data” issues
Research evaluation: The challenges
16
• What to evaluate and how to evaluate strictly depend on
• Size: Individuals, research groups, departments, institutions
• Scientific field: Hard and life sciences, social sciences and humanities
• Goal:
– Performance-based HEIs funding (to enhance average or excellence
performance?)
– Improve HEI-industry collaboration
– Incentivize social impact of research
– …
Research evaluation: The challenges
17
• Input indicators measure resources, human, physical and financial, devoted to research
– Typical examples are the number of (academic) staff employed or revenues such as
competitive, project funding for research
• Process indicators measure how research is conducted, including its management and
evaluation
– A typical example is the total of human resources employed by university departments,
offices or affiliated agencies to support and fulfill technology transfer activities
• Output indicators measure the quantity of research products
– Typical examples are the number of papers published or the number of PhDs delivered
• Outcome indicators relate to a level of performance, or achievement, for instance the
contribution research makes to the advancement of scientific‐scholarly knowledge
• Impact and benefits refers to the contribution of research outcomes for society, culture, the
environment and/or the economy
Research evaluation: Indicators
18
The quality of a publication is an elusive attribute
Research evaluation: Outcome indicators
19
Measured through proxies
Quantitative: Bibliometric indicators Qualitative: Peers’ opinion
The bibliometric evaluation
20
Based on measurable indicators of publication impact:
• The “quality” (the peer review ex-ante process, the acceptance ratio,…) and the
number of citations of the journal (Impact Factor, Eigen Factor, Source normalized
impact per paper (SNIP) , …)
• The number of citations of the article
• The number of citations of the author (h index and related indicators)
• Alternative scholarly impact metrics (altmetrics), which cover other aspects of the
impact of a work, such as
– how many data and knowledge bases refer to it
– article views
– artcile downloads
– mentions in social media and news media
(a) Normative theory of citations
- Citation as recognition of scientific value (Smith, 1981, Merton, 1988)
(b) Constructivist social theory of citations
- Citation as act of academic deference
- Citation as attempt at persuading (Gilbert, 1977)
- Assertive citation (Moed and Garfield, 2004)
- Citation as simple discourse articulation (Crossick, 2007)
The citational behaviour
21
“The main point which emerges is that citations stand at the intersection between two systems:
a rhetorical (conceptual, cognitive) system, through which scientists try to persuade each other
of their knowledge claims; and a reward (recognition, reputation) system, through which credit
for achievements is allocated” (Cozzens, 1989)
• The reliability of bibliometric indicators tends to decrease with the size of samples
they are applied to (institutions, department, research groups, individuals)
• Never confuse the impact of journals with the impact of articles they publish
(skewness of citations distribution)
• Use of a plurality of indicators (e.g., at journal level: IF, Article influence,
Eigenfactor, SJR, SNIP, …) reduces risks of manipulation: self-citation, citation
stacking,…
• Always normalize within a coherent, uniform scientific area wrt traditions of
publishing and citing
The bibliometric evaluation
22
Pros
• Efficient
• Fast
• Economic
• Not intrusive
• Objective
• Helps in identifying the origin and impact of scientific theories
Cons
• Different citational behaviours among disciplines and publication type (books vs articles)
• Self citations
• Data bases transparency and pitfalls
• Differences between OA and non-OA journals
• Language of publication
The bibliometric evaluation
23
• The “quality” of a publication cannot be assessed through
quantitative measures, just like the beauty of human beings or
artworks
• Can we assess the beauty of Leonardo’s Gioconda from the number
of tickets sold at Louvre or from the average time spent by visitors in
front of the painting?
• These are the arguments of those affirming the supremacy of peer
review against bibliometrics
The bibliometric evaluation
24
• Is the peer review the solution?
• Citing Richard Horton, editor of Lancet
“The mistake, of course, is to have thought that peer review was any more than a crude
means of discovering the acceptability—not the validity—of a new finding. Editors
and scientists alike insist on the pivotal importance of peer review. We portray peer
review to the public as a quasi-sacred process that helps to make science our most
objective truth teller. But we know that the system of peer review is biased,
unjust, unaccountable, incomplete, easily fixed, often insulting, usually
ignorant, occasionally foolish, and frequently wrong”
The peer review evaluation
25
• In several scientific areas a significant correlation has been found between
bibliometric indicators and peer review evaluations
• Italian VTE 2001-2003
- 9 areas of hard science and economics: High correlation (Franceschet, 2009)
• Italian VQR 2004-2010
– Hard and life science and economics: Higher correlation between bibliometrics and
peer review thank between the two peer reviews of the same article (Benedetto, 2013)
• Research Assessment Exercise (RAE) 1992 (UK)
- Genetics, anatomy, archeology: High correlation (Holmes & Oppenheim, 2001)
• Research Assessment Exercise (RAE) 2001 (UK)
- Psicology - Correlazione equal to 0.86 (Smith and Eysenck, 2002)
Bibliometrics and peer review
26
Bibliometric indicators, based on indexing international journals, mainly written in
English, and on extracting citational indicators, are not reliable in SSH
Challenges in research evaluation in SSH
27
How about peer review?
Publications characteristics make peer review even more difficult and less
reliable in SSH:
• Research outcomes difficult to be made objective and comparable
• Belonging bias (different schools of thought, …)
• Reduced number of potential peers (niche disciplines, marginal publication
language,…)
Journal classifications in SSH:
• Who does it?
• How is it done?
• How many classes?
• For what?
• Four examples:
– The ranking of Australian Research Council (2008)
– The ranking of European Science Foundation (ERIH project,
2007-2008)
– The ranking of AERES (2008)
– The ranking of ANVUR within the National Research Habilitation
Challenges in research evaluation in SSH
28
A tsunami of criticisms…
• UK historians: “crude and oversimplified”
• A set of journals classified in th best class A requested to be
cancelled from the ESF-ERIH lists
• Petition to AERES to withdraw the lists: “non transparent criteria”
• The ARC classification of ARC became an electoral issue in
Australia, and the new government declared it over
• Increased degree of acceptance after initial resistance in Italy
Journal classification in SSH
29
• Journal classification in SSH as a rough quantisation of the continuous ranking
induced by impact factor in hard and life sciences
• Journal IF is based on average number of citations received by published articles
in a period of time (2 or 5 years): it generates a continuous ranking of journals
within a homogenous scientific area
• Journal classification in SSH has a similar objective, but needs different
classification criteria, mainly of qualitative nature, and is limited to a small number
of classes (typically 2-3): it is a bridge between peer review and bibliometrics
• Since the number of citations is missing, journal classification cannot fully replace
peer review
Journal classification in SSH
30
• Different kind of research outputs beyond journal articles:
– Books
– Book chapters
– Translations
– Notes to court rulings (for law disciplines)
– Exhibitions and their catalogues
– Architectural designs
– Archeological excavations
– Artistic performance
– …
Challenges in research evaluation in SSH
31
Books evaluation
• Explore the feasibility of publishers classification (Spain)
• Use of indicators such as:
– Reviews on international journals
– Characteristics of the publishing series:
• Existence of an editorial board
• Transparent review procedures governing the decision to publish
• International diffusion of the publisher nooks
• …
Challenges in research evaluation in SSH
32
• Avoid global rankings of institutions based on a single score
aggregating many different indicators
• Use a multidimensional approach based on five steps1:
– Define the purpose and audience of the research assessment
– Involve institutions to be evaluated in step 1
– Identify the appropriate indicators
– Perform the assessment
– Identify the range of actions and decisions to be taken after assessment
1. Assessing Europe’s university-based research, Final Report of the Expert Group on Assessment of University-based
Research, 2010, http://ec.europa.eu/research/science-society/document_library/pdf_06/assessing-europe-university-
based-research_en.pdf
Toward an ideal evaluation environment
33
• A crucial first step is the availability of a national data base of researchers with
the list of their publications and other relevant information (research contracts,
awards, editorial responsibilities,…)
• An excellent example is the “Plataforma Lattes” in Brasil (http://lattes.cnpq.br)
• A bad example is the “Anagrafe nazionale della ricerca (ANPRePS) in Italy,
prescribed by a law in 2009 and never implemented
• The publications metadata records should be linked to the pdf of the
publications (taking into account copyright issues when relevant)
Toward an ideal evaluation environment
34
• All researchers should be uniquely identifiable through a single identifier,
linking the researcher to his/her publications and other information
• Use of ORCID identifier is one viable solution:
Non-profit organisation supported by members (majority of hem non-profit
organisations)
Free to individuals
Growing adoption: Sweden, Finland, Denmark, Norway, UK, Spain,
Portugal, Australia
• Italy launched recently the I.R.ID.E project, aiming at providing an ORCID
identifier to 80% of researchers by the end of 2016
Toward an ideal evaluation environment
35
• Local assessment in Institutions should be performed more frequently (yearly?) to
provide bridge between national assessments (4-5 years period)
• Local assessment must consider a wider range of context variables:
– The critical mass of research groups
– The strategic promotion of some areas
– The opening of new research frontiers, e.g., interdisciplinary
• National and local research assessments must be coordinated, to present
researchers with a coherent set of goals and incentives
Toward an ideal evaluation environment
36
• Research assessments must include an indicator of the performance variation with
time, so as to reward improvement even when the absolute performance is still
poor
• This implies a certain degree of persistence of indicators
• The evaluation of research outputs should use an informed peer review
methodology, where the panel in charge acquires information from:
 Biliometrics
 Expert peers
 …
to make the final decision
Toward an ideal evaluation environment
37
• To be effective, research assessment must be a shared experience in goals and
methodologies between evaluators and evaluated
• Evaluation criteria must be known a priori
• The evaluation results should not be applied to different contexts wrt the initial ones
 Outcomes evaluations addressed to institutional performance should never be used to
assess individuals
• Performance-based funding should not erode the institution survival quota
• Assessment methodology should not underestimate the inter (multi) disciplinary
research
A few final hints
38
If you cannot measure it, you cannot improve it. Lord Kelvin
39
Not everything that can be counted counts, and not everything
that counts can be counted. William B. Cameron, Informal
Sociology: “A Casual Introduction to Sociological Thinking” (1963)
40
Thank you

Weitere ähnliche Inhalte

Was ist angesagt?

Qualitative vs Quantitative
Qualitative vs QuantitativeQualitative vs Quantitative
Qualitative vs Quantitative
eilire91
 
Alternative Views Of Evaluation
Alternative Views Of EvaluationAlternative Views Of Evaluation
Alternative Views Of Evaluation
mandy fontenot
 
Methods for developing assessment instruments to generate useful data in t…
Methods for developing assessment instruments to generate useful data in t…Methods for developing assessment instruments to generate useful data in t…
Methods for developing assessment instruments to generate useful data in t…
Pat Barlow
 

Was ist angesagt? (20)

Assessing quality and bias in studies
Assessing quality and bias in studiesAssessing quality and bias in studies
Assessing quality and bias in studies
 
Qualitative vs Quantitative
Qualitative vs QuantitativeQualitative vs Quantitative
Qualitative vs Quantitative
 
Chapter10
Chapter10Chapter10
Chapter10
 
Impact evaluation methods: Qualitative Methods
Impact evaluation methods: Qualitative MethodsImpact evaluation methods: Qualitative Methods
Impact evaluation methods: Qualitative Methods
 
Dr. G. Sawarkar qualitative & quantitative evaluation methode
Dr. G. Sawarkar qualitative & quantitative evaluation methodeDr. G. Sawarkar qualitative & quantitative evaluation methode
Dr. G. Sawarkar qualitative & quantitative evaluation methode
 
Analytic Frameworks Quiz
Analytic Frameworks QuizAnalytic Frameworks Quiz
Analytic Frameworks Quiz
 
Chapter17
Chapter17Chapter17
Chapter17
 
Scale development
Scale developmentScale development
Scale development
 
Quality assessment in systematic literature review
Quality assessment in systematic literature reviewQuality assessment in systematic literature review
Quality assessment in systematic literature review
 
Grading Strength of Evidence
Grading Strength of EvidenceGrading Strength of Evidence
Grading Strength of Evidence
 
Systematic literature review: An introduction
Systematic literature review: An introductionSystematic literature review: An introduction
Systematic literature review: An introduction
 
Alternative Views Of Evaluation
Alternative Views Of EvaluationAlternative Views Of Evaluation
Alternative Views Of Evaluation
 
Assessing Quality of Individual Studies Quiz
Assessing Quality of Individual Studies QuizAssessing Quality of Individual Studies Quiz
Assessing Quality of Individual Studies Quiz
 
Mixed Methods Research Design
Mixed Methods Research DesignMixed Methods Research Design
Mixed Methods Research Design
 
Comparative and Non-Comparative
Comparative and Non-ComparativeComparative and Non-Comparative
Comparative and Non-Comparative
 
Notes from frankel and wallen
Notes from frankel  and wallenNotes from frankel  and wallen
Notes from frankel and wallen
 
Mixed Method Research
Mixed Method ResearchMixed Method Research
Mixed Method Research
 
Pseudo-evaluation and Quasi-evaluation Approach
Pseudo-evaluation and Quasi-evaluation ApproachPseudo-evaluation and Quasi-evaluation Approach
Pseudo-evaluation and Quasi-evaluation Approach
 
Methods for developing assessment instruments to generate useful data in t…
Methods for developing assessment instruments to generate useful data in t…Methods for developing assessment instruments to generate useful data in t…
Methods for developing assessment instruments to generate useful data in t…
 
Efficacy of Information interventions in reducing transfer anxiety from a cri...
Efficacy of Information interventions in reducing transfer anxiety from a cri...Efficacy of Information interventions in reducing transfer anxiety from a cri...
Efficacy of Information interventions in reducing transfer anxiety from a cri...
 

Ähnlich wie Challenges and opportunities in research evaluation: toward a better evaluation environment

Understanding impact through alternative metrics: developing library-based as...
Understanding impact through alternative metrics: developing library-based as...Understanding impact through alternative metrics: developing library-based as...
Understanding impact through alternative metrics: developing library-based as...
Kristi Holmes
 
CL8 Scientiometrics Module 6 RPE-Rijo TKMCE.pdf
CL8 Scientiometrics Module 6 RPE-Rijo TKMCE.pdfCL8 Scientiometrics Module 6 RPE-Rijo TKMCE.pdf
CL8 Scientiometrics Module 6 RPE-Rijo TKMCE.pdf
ssuserb76cdd
 

Ähnlich wie Challenges and opportunities in research evaluation: toward a better evaluation environment (20)

Beyond the Factor: Talking about Research Impact
Beyond the Factor: Talking about Research ImpactBeyond the Factor: Talking about Research Impact
Beyond the Factor: Talking about Research Impact
 
Evolving and emerging scholarly communication services in libraries: public a...
Evolving and emerging scholarly communication services in libraries: public a...Evolving and emerging scholarly communication services in libraries: public a...
Evolving and emerging scholarly communication services in libraries: public a...
 
Altmetricon 2015
Altmetricon 2015Altmetricon 2015
Altmetricon 2015
 
Introduction to Altmetrics for Medical and Special Librarians
Introduction to Altmetrics for Medical and Special LibrariansIntroduction to Altmetrics for Medical and Special Librarians
Introduction to Altmetrics for Medical and Special Librarians
 
Proving Your Value: The Librarians’ Contribution to the Promotion and Tenure ...
Proving Your Value: The Librarians’ Contribution to the Promotion and Tenure ...Proving Your Value: The Librarians’ Contribution to the Promotion and Tenure ...
Proving Your Value: The Librarians’ Contribution to the Promotion and Tenure ...
 
Quality in scientific research final
Quality in scientific research finalQuality in scientific research final
Quality in scientific research final
 
NISO Webinar: New Perspectives on Assessment How Altmetrics Measure Scholarly...
NISO Webinar: New Perspectives on Assessment How Altmetrics Measure Scholarly...NISO Webinar: New Perspectives on Assessment How Altmetrics Measure Scholarly...
NISO Webinar: New Perspectives on Assessment How Altmetrics Measure Scholarly...
 
Understanding impact through alternative metrics: developing library-based as...
Understanding impact through alternative metrics: developing library-based as...Understanding impact through alternative metrics: developing library-based as...
Understanding impact through alternative metrics: developing library-based as...
 
2016 AAUDE
2016 AAUDE2016 AAUDE
2016 AAUDE
 
Evaluating e reference
Evaluating e referenceEvaluating e reference
Evaluating e reference
 
In metrics we trust?
In metrics we trust?In metrics we trust?
In metrics we trust?
 
Assessing research impact mic 1 Sep 2015
Assessing research impact mic 1 Sep 2015Assessing research impact mic 1 Sep 2015
Assessing research impact mic 1 Sep 2015
 
Assessing Research Impact: Bibliometrics, Citations and the H-Index
Assessing Research Impact: Bibliometrics, Citations and the H-IndexAssessing Research Impact: Bibliometrics, Citations and the H-Index
Assessing Research Impact: Bibliometrics, Citations and the H-Index
 
Peer reviewing
 Peer reviewing Peer reviewing
Peer reviewing
 
Peer reviewing
Peer reviewingPeer reviewing
Peer reviewing
 
Peer reviewing
Peer reviewingPeer reviewing
Peer reviewing
 
CL8 Scientiometrics Module 6 RPE-Rijo TKMCE.pdf
CL8 Scientiometrics Module 6 RPE-Rijo TKMCE.pdfCL8 Scientiometrics Module 6 RPE-Rijo TKMCE.pdf
CL8 Scientiometrics Module 6 RPE-Rijo TKMCE.pdf
 
MedicReS Conference 2017 Istanbul - Journal Metrics: The Impact Factor and Al...
MedicReS Conference 2017 Istanbul - Journal Metrics: The Impact Factor and Al...MedicReS Conference 2017 Istanbul - Journal Metrics: The Impact Factor and Al...
MedicReS Conference 2017 Istanbul - Journal Metrics: The Impact Factor and Al...
 
Atmiya university. shree m n virani college of science 14 oct 2021. researc...
Atmiya university. shree m n virani college of science 14 oct 2021.   researc...Atmiya university. shree m n virani college of science 14 oct 2021.   researc...
Atmiya university. shree m n virani college of science 14 oct 2021. researc...
 
Russell Group PVCs 3 Jun 09
Russell Group PVCs 3 Jun 09Russell Group PVCs 3 Jun 09
Russell Group PVCs 3 Jun 09
 

Mehr von ORCID, Inc

Mehr von ORCID, Inc (20)

ORCID Overview: Why your Lifelong Identifier is Important in the Digital Age ...
ORCID Overview: Why your Lifelong Identifier is Important in the Digital Age ...ORCID Overview: Why your Lifelong Identifier is Important in the Digital Age ...
ORCID Overview: Why your Lifelong Identifier is Important in the Digital Age ...
 
Identifying Springer's Author (with ORCID iD) on SpringerLink (H. Aziz)
Identifying Springer's Author (with ORCID iD) on SpringerLink (H. Aziz)Identifying Springer's Author (with ORCID iD) on SpringerLink (H. Aziz)
Identifying Springer's Author (with ORCID iD) on SpringerLink (H. Aziz)
 
ORCID at UniKL (P. Hamzah)
ORCID at UniKL (P. Hamzah)ORCID at UniKL (P. Hamzah)
ORCID at UniKL (P. Hamzah)
 
ORCID Integration Videos
ORCID Integration VideosORCID Integration Videos
ORCID Integration Videos
 
ORCID as a Community Initiative (N. Miyairi)
ORCID as a Community Initiative (N. Miyairi)ORCID as a Community Initiative (N. Miyairi)
ORCID as a Community Initiative (N. Miyairi)
 
Introduction and Welcome to the 2017 ORCID Malaysia Workshop (N. Miyairi)
Introduction and Welcome to the 2017 ORCID Malaysia Workshop (N. Miyairi)Introduction and Welcome to the 2017 ORCID Malaysia Workshop (N. Miyairi)
Introduction and Welcome to the 2017 ORCID Malaysia Workshop (N. Miyairi)
 
MyRID: ORCID Integration in Malaysia (T. Dharmalingam)
MyRID: ORCID Integration in Malaysia (T. Dharmalingam)MyRID: ORCID Integration in Malaysia (T. Dharmalingam)
MyRID: ORCID Integration in Malaysia (T. Dharmalingam)
 
Spreading the ORCID Word: ORCID Communications Webinar (2016.12)
Spreading the ORCID Word: ORCID Communications Webinar (2016.12)Spreading the ORCID Word: ORCID Communications Webinar (2016.12)
Spreading the ORCID Word: ORCID Communications Webinar (2016.12)
 
ORCID @ Khalifa University
ORCID @ Khalifa UniversityORCID @ Khalifa University
ORCID @ Khalifa University
 
ORCID Integration with Institutional Repositories (D. Grenz)
ORCID Integration with Institutional Repositories (D. Grenz)ORCID Integration with Institutional Repositories (D. Grenz)
ORCID Integration with Institutional Repositories (D. Grenz)
 
Research in a world where machines read (M. Buys)
Research in a world where machines read (M. Buys)Research in a world where machines read (M. Buys)
Research in a world where machines read (M. Buys)
 
ORCID Collect & Connect: understanding integrations and the API (M. Buys)
ORCID Collect & Connect: understanding integrations and the API (M. Buys)ORCID Collect & Connect: understanding integrations and the API (M. Buys)
ORCID Collect & Connect: understanding integrations and the API (M. Buys)
 
Benefits to researchers who use ORCID (P. Purnell)
Benefits to researchers who use ORCID (P. Purnell)Benefits to researchers who use ORCID (P. Purnell)
Benefits to researchers who use ORCID (P. Purnell)
 
Research Management & Publishing (M. Jagerhorn)
Research Management & Publishing (M. Jagerhorn)Research Management & Publishing (M. Jagerhorn)
Research Management & Publishing (M. Jagerhorn)
 
ORCID overview: why your lifelong identifier is important in the digital age ...
ORCID overview: why your lifelong identifier is important in the digital age ...ORCID overview: why your lifelong identifier is important in the digital age ...
ORCID overview: why your lifelong identifier is important in the digital age ...
 
ORCID in the Publishing Workflow (Mochammad Tanzil Multazam)
ORCID in the Publishing Workflow (Mochammad Tanzil Multazam)ORCID in the Publishing Workflow (Mochammad Tanzil Multazam)
ORCID in the Publishing Workflow (Mochammad Tanzil Multazam)
 
What in the World is ORCID? (Haak)
What in the World is ORCID? (Haak)What in the World is ORCID? (Haak)
What in the World is ORCID? (Haak)
 
ORCID as a Community Initiative (Miyairi)
ORCID as a Community Initiative (Miyairi)ORCID as a Community Initiative (Miyairi)
ORCID as a Community Initiative (Miyairi)
 
ORCID Integration Videos
ORCID Integration VideosORCID Integration Videos
ORCID Integration Videos
 
The Latest on ORCID API v2
The Latest on ORCID API v2 The Latest on ORCID API v2
The Latest on ORCID API v2
 

Kürzlich hochgeladen

Chiulli_Aurora_Oman_Raffaele_Beowulf.pptx
Chiulli_Aurora_Oman_Raffaele_Beowulf.pptxChiulli_Aurora_Oman_Raffaele_Beowulf.pptx
Chiulli_Aurora_Oman_Raffaele_Beowulf.pptx
raffaeleoman
 
No Advance 8868886958 Chandigarh Call Girls , Indian Call Girls For Full Nigh...
No Advance 8868886958 Chandigarh Call Girls , Indian Call Girls For Full Nigh...No Advance 8868886958 Chandigarh Call Girls , Indian Call Girls For Full Nigh...
No Advance 8868886958 Chandigarh Call Girls , Indian Call Girls For Full Nigh...
Sheetaleventcompany
 

Kürzlich hochgeladen (20)

My Presentation "In Your Hands" by Halle Bailey
My Presentation "In Your Hands" by Halle BaileyMy Presentation "In Your Hands" by Halle Bailey
My Presentation "In Your Hands" by Halle Bailey
 
Chiulli_Aurora_Oman_Raffaele_Beowulf.pptx
Chiulli_Aurora_Oman_Raffaele_Beowulf.pptxChiulli_Aurora_Oman_Raffaele_Beowulf.pptx
Chiulli_Aurora_Oman_Raffaele_Beowulf.pptx
 
SaaStr Workshop Wednesday w/ Lucas Price, Yardstick
SaaStr Workshop Wednesday w/ Lucas Price, YardstickSaaStr Workshop Wednesday w/ Lucas Price, Yardstick
SaaStr Workshop Wednesday w/ Lucas Price, Yardstick
 
Dreaming Marissa Sánchez Music Video Treatment
Dreaming Marissa Sánchez Music Video TreatmentDreaming Marissa Sánchez Music Video Treatment
Dreaming Marissa Sánchez Music Video Treatment
 
The workplace ecosystem of the future 24.4.2024 Fabritius_share ii.pdf
The workplace ecosystem of the future 24.4.2024 Fabritius_share ii.pdfThe workplace ecosystem of the future 24.4.2024 Fabritius_share ii.pdf
The workplace ecosystem of the future 24.4.2024 Fabritius_share ii.pdf
 
Re-membering the Bard: Revisiting The Compleat Wrks of Wllm Shkspr (Abridged)...
Re-membering the Bard: Revisiting The Compleat Wrks of Wllm Shkspr (Abridged)...Re-membering the Bard: Revisiting The Compleat Wrks of Wllm Shkspr (Abridged)...
Re-membering the Bard: Revisiting The Compleat Wrks of Wllm Shkspr (Abridged)...
 
Presentation on Engagement in Book Clubs
Presentation on Engagement in Book ClubsPresentation on Engagement in Book Clubs
Presentation on Engagement in Book Clubs
 
Mohammad_Alnahdi_Oral_Presentation_Assignment.pptx
Mohammad_Alnahdi_Oral_Presentation_Assignment.pptxMohammad_Alnahdi_Oral_Presentation_Assignment.pptx
Mohammad_Alnahdi_Oral_Presentation_Assignment.pptx
 
Causes of poverty in France presentation.pptx
Causes of poverty in France presentation.pptxCauses of poverty in France presentation.pptx
Causes of poverty in France presentation.pptx
 
Introduction to Prompt Engineering (Focusing on ChatGPT)
Introduction to Prompt Engineering (Focusing on ChatGPT)Introduction to Prompt Engineering (Focusing on ChatGPT)
Introduction to Prompt Engineering (Focusing on ChatGPT)
 
Air breathing and respiratory adaptations in diver animals
Air breathing and respiratory adaptations in diver animalsAir breathing and respiratory adaptations in diver animals
Air breathing and respiratory adaptations in diver animals
 
Thirunelveli call girls Tamil escorts 7877702510
Thirunelveli call girls Tamil escorts 7877702510Thirunelveli call girls Tamil escorts 7877702510
Thirunelveli call girls Tamil escorts 7877702510
 
Busty Desi⚡Call Girls in Sector 51 Noida Escorts >༒8448380779 Escort Service-...
Busty Desi⚡Call Girls in Sector 51 Noida Escorts >༒8448380779 Escort Service-...Busty Desi⚡Call Girls in Sector 51 Noida Escorts >༒8448380779 Escort Service-...
Busty Desi⚡Call Girls in Sector 51 Noida Escorts >༒8448380779 Escort Service-...
 
ANCHORING SCRIPT FOR A CULTURAL EVENT.docx
ANCHORING SCRIPT FOR A CULTURAL EVENT.docxANCHORING SCRIPT FOR A CULTURAL EVENT.docx
ANCHORING SCRIPT FOR A CULTURAL EVENT.docx
 
lONG QUESTION ANSWER PAKISTAN STUDIES10.
lONG QUESTION ANSWER PAKISTAN STUDIES10.lONG QUESTION ANSWER PAKISTAN STUDIES10.
lONG QUESTION ANSWER PAKISTAN STUDIES10.
 
No Advance 8868886958 Chandigarh Call Girls , Indian Call Girls For Full Nigh...
No Advance 8868886958 Chandigarh Call Girls , Indian Call Girls For Full Nigh...No Advance 8868886958 Chandigarh Call Girls , Indian Call Girls For Full Nigh...
No Advance 8868886958 Chandigarh Call Girls , Indian Call Girls For Full Nigh...
 
VVIP Call Girls Nalasopara : 9892124323, Call Girls in Nalasopara Services
VVIP Call Girls Nalasopara : 9892124323, Call Girls in Nalasopara ServicesVVIP Call Girls Nalasopara : 9892124323, Call Girls in Nalasopara Services
VVIP Call Girls Nalasopara : 9892124323, Call Girls in Nalasopara Services
 
BDSM⚡Call Girls in Sector 97 Noida Escorts >༒8448380779 Escort Service
BDSM⚡Call Girls in Sector 97 Noida Escorts >༒8448380779 Escort ServiceBDSM⚡Call Girls in Sector 97 Noida Escorts >༒8448380779 Escort Service
BDSM⚡Call Girls in Sector 97 Noida Escorts >༒8448380779 Escort Service
 
AWS Data Engineer Associate (DEA-C01) Exam Dumps 2024.pdf
AWS Data Engineer Associate (DEA-C01) Exam Dumps 2024.pdfAWS Data Engineer Associate (DEA-C01) Exam Dumps 2024.pdf
AWS Data Engineer Associate (DEA-C01) Exam Dumps 2024.pdf
 
Call Girl Number in Khar Mumbai📲 9892124323 💞 Full Night Enjoy
Call Girl Number in Khar Mumbai📲 9892124323 💞 Full Night EnjoyCall Girl Number in Khar Mumbai📲 9892124323 💞 Full Night Enjoy
Call Girl Number in Khar Mumbai📲 9892124323 💞 Full Night Enjoy
 

Challenges and opportunities in research evaluation: toward a better evaluation environment

  • 1. Challenges and opportunities in research evaluation: Toward a better evaluation environment Sergio Benedetto Consiglio Direttivo ANVUR sergio.benedetto@anvur.org Barcelona, May 18, 2015
  • 3. 3 Research evaluation Ex ante Ex post before research takes place, to assess its potential relevance, the prospects of success and the cost appropriateness after research has been concluded, to assess its results in terms of scientific quality and impact Comparative Individual aimed at defining a ranking of individuals, research groups or HEIs, often within a homogeneous area of research comparing qualification against a threshold to promote or not individuals, research groups
  • 4. National research assessments • What? …..Evaluated objects… • Why? …..Goals… • How?....Evaluation methodologies… • When?....Continuity, frequency… • With what consequences?..... On institutions, on researchers, on society… 4
  • 6. • The volume of the scientific outcomes • Their quality • Their scientific impact • Their impacts on the economy, society and/or culture • More in general, the so-called “third mission” of HEIs, i.e. their involvement with society at large Objects of research evaluation 6
  • 8. To inform HEI government bodies and other stakeholders about the status of national research To help Ministry of Education and Research (or other national bodies) distributing resources to HEIs To help HEI government bodies taking strategic decisions to improve the quality and effectiveness of research and in internal resource (positions, funds) assignment Goals of comparative research evaluation 8
  • 10. 10 Sample of research outcomes Expert panels Reading Bibliometry Informed peer review External peer review Bibliometry Final evaluation Final evaluation Criteria
  • 12. RAE-REF in UK: 1986, 1989, 1992, 1996, 2001, 2008, 2014 VTR-VQR in Italy: 2006, 2013, 2017 (?) ERA in Australia: 2010, 2012, 2015 The period of research evaluation 12 3 43 5 7 6 7 4 2 3
  • 14. No measuring technique leaves the measured object unaffected, so: 14 • What are the intended consequences? • Evaluation leads to an improvement in the quality of research (UK: Adams & Gurney, 2010, Australia: Butler, 2003) • Evaluation modifies dissemination channels, e.g., it makes the journal article published in highly ranked journals the main publication outlet (Rin, 2009) • Resources distribution to HEI based on assessment outcomes (UK, Italy,…) • Improved HEIs infrastructure and archival repositories • Enhanced “quality” recruitment • Strategic positioning and more consistent policies • Trust from society
  • 15. No measuring technique leaves the measured object unaffected, so: 15 • What could be the risks and unintended consequences? • Worse publication practices: excessive segmentation of the research results, clinging to the mainstream, safe disciplinary research, citation stacking , coercive citation, ...) • Research freedom limitation: too much emphasis on accountability • Underestimating the teaching activity (J. Warner, 1998) • Misuse of assessment outcomes: evaluating individuals, apply national-level criteria automatically to local issues • …
  • 16. • Collecting data and objects: – Local (often incompatible) repositories – Copyright issues • Cleaning data: – Human errors in uploading – Names ambiguity – Duplicated records • Connecting data to “owners”: – Researchers, institutions • Transferring data and objects to evaluators (panels, peer reviewers,…) – IP protection, “big data” issues Research evaluation: The challenges 16
  • 17. • What to evaluate and how to evaluate strictly depend on • Size: Individuals, research groups, departments, institutions • Scientific field: Hard and life sciences, social sciences and humanities • Goal: – Performance-based HEIs funding (to enhance average or excellence performance?) – Improve HEI-industry collaboration – Incentivize social impact of research – … Research evaluation: The challenges 17
  • 18. • Input indicators measure resources, human, physical and financial, devoted to research – Typical examples are the number of (academic) staff employed or revenues such as competitive, project funding for research • Process indicators measure how research is conducted, including its management and evaluation – A typical example is the total of human resources employed by university departments, offices or affiliated agencies to support and fulfill technology transfer activities • Output indicators measure the quantity of research products – Typical examples are the number of papers published or the number of PhDs delivered • Outcome indicators relate to a level of performance, or achievement, for instance the contribution research makes to the advancement of scientific‐scholarly knowledge • Impact and benefits refers to the contribution of research outcomes for society, culture, the environment and/or the economy Research evaluation: Indicators 18
  • 19. The quality of a publication is an elusive attribute Research evaluation: Outcome indicators 19 Measured through proxies Quantitative: Bibliometric indicators Qualitative: Peers’ opinion
  • 20. The bibliometric evaluation 20 Based on measurable indicators of publication impact: • The “quality” (the peer review ex-ante process, the acceptance ratio,…) and the number of citations of the journal (Impact Factor, Eigen Factor, Source normalized impact per paper (SNIP) , …) • The number of citations of the article • The number of citations of the author (h index and related indicators) • Alternative scholarly impact metrics (altmetrics), which cover other aspects of the impact of a work, such as – how many data and knowledge bases refer to it – article views – artcile downloads – mentions in social media and news media
  • 21. (a) Normative theory of citations - Citation as recognition of scientific value (Smith, 1981, Merton, 1988) (b) Constructivist social theory of citations - Citation as act of academic deference - Citation as attempt at persuading (Gilbert, 1977) - Assertive citation (Moed and Garfield, 2004) - Citation as simple discourse articulation (Crossick, 2007) The citational behaviour 21 “The main point which emerges is that citations stand at the intersection between two systems: a rhetorical (conceptual, cognitive) system, through which scientists try to persuade each other of their knowledge claims; and a reward (recognition, reputation) system, through which credit for achievements is allocated” (Cozzens, 1989)
  • 22. • The reliability of bibliometric indicators tends to decrease with the size of samples they are applied to (institutions, department, research groups, individuals) • Never confuse the impact of journals with the impact of articles they publish (skewness of citations distribution) • Use of a plurality of indicators (e.g., at journal level: IF, Article influence, Eigenfactor, SJR, SNIP, …) reduces risks of manipulation: self-citation, citation stacking,… • Always normalize within a coherent, uniform scientific area wrt traditions of publishing and citing The bibliometric evaluation 22
  • 23. Pros • Efficient • Fast • Economic • Not intrusive • Objective • Helps in identifying the origin and impact of scientific theories Cons • Different citational behaviours among disciplines and publication type (books vs articles) • Self citations • Data bases transparency and pitfalls • Differences between OA and non-OA journals • Language of publication The bibliometric evaluation 23
  • 24. • The “quality” of a publication cannot be assessed through quantitative measures, just like the beauty of human beings or artworks • Can we assess the beauty of Leonardo’s Gioconda from the number of tickets sold at Louvre or from the average time spent by visitors in front of the painting? • These are the arguments of those affirming the supremacy of peer review against bibliometrics The bibliometric evaluation 24
  • 25. • Is the peer review the solution? • Citing Richard Horton, editor of Lancet “The mistake, of course, is to have thought that peer review was any more than a crude means of discovering the acceptability—not the validity—of a new finding. Editors and scientists alike insist on the pivotal importance of peer review. We portray peer review to the public as a quasi-sacred process that helps to make science our most objective truth teller. But we know that the system of peer review is biased, unjust, unaccountable, incomplete, easily fixed, often insulting, usually ignorant, occasionally foolish, and frequently wrong” The peer review evaluation 25
  • 26. • In several scientific areas a significant correlation has been found between bibliometric indicators and peer review evaluations • Italian VTE 2001-2003 - 9 areas of hard science and economics: High correlation (Franceschet, 2009) • Italian VQR 2004-2010 – Hard and life science and economics: Higher correlation between bibliometrics and peer review thank between the two peer reviews of the same article (Benedetto, 2013) • Research Assessment Exercise (RAE) 1992 (UK) - Genetics, anatomy, archeology: High correlation (Holmes & Oppenheim, 2001) • Research Assessment Exercise (RAE) 2001 (UK) - Psicology - Correlazione equal to 0.86 (Smith and Eysenck, 2002) Bibliometrics and peer review 26
  • 27. Bibliometric indicators, based on indexing international journals, mainly written in English, and on extracting citational indicators, are not reliable in SSH Challenges in research evaluation in SSH 27 How about peer review? Publications characteristics make peer review even more difficult and less reliable in SSH: • Research outcomes difficult to be made objective and comparable • Belonging bias (different schools of thought, …) • Reduced number of potential peers (niche disciplines, marginal publication language,…)
  • 28. Journal classifications in SSH: • Who does it? • How is it done? • How many classes? • For what? • Four examples: – The ranking of Australian Research Council (2008) – The ranking of European Science Foundation (ERIH project, 2007-2008) – The ranking of AERES (2008) – The ranking of ANVUR within the National Research Habilitation Challenges in research evaluation in SSH 28
  • 29. A tsunami of criticisms… • UK historians: “crude and oversimplified” • A set of journals classified in th best class A requested to be cancelled from the ESF-ERIH lists • Petition to AERES to withdraw the lists: “non transparent criteria” • The ARC classification of ARC became an electoral issue in Australia, and the new government declared it over • Increased degree of acceptance after initial resistance in Italy Journal classification in SSH 29
  • 30. • Journal classification in SSH as a rough quantisation of the continuous ranking induced by impact factor in hard and life sciences • Journal IF is based on average number of citations received by published articles in a period of time (2 or 5 years): it generates a continuous ranking of journals within a homogenous scientific area • Journal classification in SSH has a similar objective, but needs different classification criteria, mainly of qualitative nature, and is limited to a small number of classes (typically 2-3): it is a bridge between peer review and bibliometrics • Since the number of citations is missing, journal classification cannot fully replace peer review Journal classification in SSH 30
  • 31. • Different kind of research outputs beyond journal articles: – Books – Book chapters – Translations – Notes to court rulings (for law disciplines) – Exhibitions and their catalogues – Architectural designs – Archeological excavations – Artistic performance – … Challenges in research evaluation in SSH 31
  • 32. Books evaluation • Explore the feasibility of publishers classification (Spain) • Use of indicators such as: – Reviews on international journals – Characteristics of the publishing series: • Existence of an editorial board • Transparent review procedures governing the decision to publish • International diffusion of the publisher nooks • … Challenges in research evaluation in SSH 32
  • 33. • Avoid global rankings of institutions based on a single score aggregating many different indicators • Use a multidimensional approach based on five steps1: – Define the purpose and audience of the research assessment – Involve institutions to be evaluated in step 1 – Identify the appropriate indicators – Perform the assessment – Identify the range of actions and decisions to be taken after assessment 1. Assessing Europe’s university-based research, Final Report of the Expert Group on Assessment of University-based Research, 2010, http://ec.europa.eu/research/science-society/document_library/pdf_06/assessing-europe-university- based-research_en.pdf Toward an ideal evaluation environment 33
  • 34. • A crucial first step is the availability of a national data base of researchers with the list of their publications and other relevant information (research contracts, awards, editorial responsibilities,…) • An excellent example is the “Plataforma Lattes” in Brasil (http://lattes.cnpq.br) • A bad example is the “Anagrafe nazionale della ricerca (ANPRePS) in Italy, prescribed by a law in 2009 and never implemented • The publications metadata records should be linked to the pdf of the publications (taking into account copyright issues when relevant) Toward an ideal evaluation environment 34
  • 35. • All researchers should be uniquely identifiable through a single identifier, linking the researcher to his/her publications and other information • Use of ORCID identifier is one viable solution: Non-profit organisation supported by members (majority of hem non-profit organisations) Free to individuals Growing adoption: Sweden, Finland, Denmark, Norway, UK, Spain, Portugal, Australia • Italy launched recently the I.R.ID.E project, aiming at providing an ORCID identifier to 80% of researchers by the end of 2016 Toward an ideal evaluation environment 35
  • 36. • Local assessment in Institutions should be performed more frequently (yearly?) to provide bridge between national assessments (4-5 years period) • Local assessment must consider a wider range of context variables: – The critical mass of research groups – The strategic promotion of some areas – The opening of new research frontiers, e.g., interdisciplinary • National and local research assessments must be coordinated, to present researchers with a coherent set of goals and incentives Toward an ideal evaluation environment 36
  • 37. • Research assessments must include an indicator of the performance variation with time, so as to reward improvement even when the absolute performance is still poor • This implies a certain degree of persistence of indicators • The evaluation of research outputs should use an informed peer review methodology, where the panel in charge acquires information from:  Biliometrics  Expert peers  … to make the final decision Toward an ideal evaluation environment 37
  • 38. • To be effective, research assessment must be a shared experience in goals and methodologies between evaluators and evaluated • Evaluation criteria must be known a priori • The evaluation results should not be applied to different contexts wrt the initial ones  Outcomes evaluations addressed to institutional performance should never be used to assess individuals • Performance-based funding should not erode the institution survival quota • Assessment methodology should not underestimate the inter (multi) disciplinary research A few final hints 38
  • 39. If you cannot measure it, you cannot improve it. Lord Kelvin 39 Not everything that can be counted counts, and not everything that counts can be counted. William B. Cameron, Informal Sociology: “A Casual Introduction to Sociological Thinking” (1963)