uses and misuses of quantitative indicators of impact

37
Uses and Misuses of Quantitative Indicators of Impact Berenika Webster 7 October 2016 ULS/ISchool Digital Scholarship Workshop and Lecture Series 1

Upload: berenika-webster

Post on 14-Apr-2017

208 views

Category:

Science


2 download

TRANSCRIPT

Page 1: Uses and misuses of quantitative indicators of impact

1

Uses and Misuses of Quantitative Indicators of ImpactBerenika Webster

7 October 2016

ULS/ISchool Digital Scholarship Workshop and Lecture Series

Page 2: Uses and misuses of quantitative indicators of impact

2

Metrics are everywhere

Page 3: Uses and misuses of quantitative indicators of impact

3

Everyone talks about…

Productivity (publications counts) leads to “salami slicing”, maybe quantity vs quality

Impact (citation counts) but what impact?

Impact factor Speaks to prestige of outlet, not quality of individual paper

h-index simplistic it’s always highest in Google Scholar

Page 4: Uses and misuses of quantitative indicators of impact

4

What are we measuring exactly? San Francisco Declaration on Research Assessment,

2012 (DORA) Against using JIF to demonstrate impact of individual

publication

H-index problem Can we really show impact of a researcher through one

number?

Leiden Manifesto (2015) Bibliometrics practitioners stating some known truths

Page 5: Uses and misuses of quantitative indicators of impact

5

The backlash

Page 6: Uses and misuses of quantitative indicators of impact

6

• Over 12,000 individual signatories

• 800 institutional signatories (but not Pitt)

The steps that DORA recommends for universities and research institutions are measured and practical: be clear about the criteria used in researcher assessment; emphasise that a paper’s content is more important than where it is published; make sure to consider the value and impact of all types of research output; and use a broad range of measures when doing so. There is no blanket ban on metrics. Stephen Curry, Who is afraid of DORA, http://www.researchresearch.com/news/article/?articleId=1360100, 11 May 2016

Page 7: Uses and misuses of quantitative indicators of impact

7

I am not a number but…

Funders need to be responsible in the way that they use metrics, to resist the reduction of researchers’ careers to decimal points.

Researchers need to learn to use metrics to enhance the narratives that they develop to describe their ambitions and careers.

Providers need to understand that the data, analysis and visualizations they provide have a value over and beyond a simple service.

Mike Taylor, Metrics and The Social Contract: Using Numbers, Preserving Humanity https://www.digital-science.com/blog/perspectives, 26 July 2016

Page 8: Uses and misuses of quantitative indicators of impact

8

https://vimeo.com/133683418

Page 9: Uses and misuses of quantitative indicators of impact

9

Principle 1

Metrics-based evaluation can supplement and provide additional dimensions to qualitative assessment, but should never replace it.

Page 10: Uses and misuses of quantitative indicators of impact

10

Excellence in ResearchAustralia

ERA is a comprehensive quality evaluation of all research produced in Australian universities against national and international benchmarks. The ratings are determined and moderated by committees of distinguished researchers, drawn from Australia and overseas.

ERA is based on expert review informed by a range of indicators. The indicators used in ERA include a range of metrics such as citation profiles which are common to disciplines in the natural sciences, and peer review of a sample of research outputs which is more broadly common in the humanities and social sciences.

Page 11: Uses and misuses of quantitative indicators of impact

11

The REF team will provide the following information for each publication year in the period 2008 to 2012, and for each relevant ASJC code:

• The average (mean) number of times that journal articles and conference proceedings published worldwide in that year, in that ASJC code, were cited

• The number of times that journal articles and conference proceedings in that ASJC code would need to be cited to be in the top 1 per cent, 5 per cent, 10 per cent and 25 per cent of papers published worldwide in that year.

Page 12: Uses and misuses of quantitative indicators of impact

12

This work has shown that individual metrics give significantly different outcomes from the REF peer review process, showing that metrics cannot provide a like-for-like replacement for REF peer review. Publication year was a significant factor in the calculation of correlationwith REF scores, with all but two metrics showing significant decreases in correlation for more recent outputs. There is large variation in the availability of metrics data across the REF submission, with particular issues with coverage in units of assessment (UOAs) in REF Main Panel D. Finally, there is evidence to suggest issues for early career researchers (ECRs) and women in a small number of disciplines, as shown by statistically significant differences inthe REF scores for these groups at the UOA level.

Page 13: Uses and misuses of quantitative indicators of impact

13

Principle 2

Metrics used to evaluate research performance should reflect the research objectives of the institution, research groups or individual researchers.

No single metric or evaluation model can apply in all contexts.

Page 14: Uses and misuses of quantitative indicators of impact

14

Example: which impact?

https://www.insidehighered.com/news/2016/09/08/, 8 Sept. 2016

Which impact?• Academic• Social• Economic• Environmental

CitationsPatentsCommercialisation incomeChanged legislationCreated jobsImproved quality of lifeSaved lives

Page 15: Uses and misuses of quantitative indicators of impact

15

Example: quality and impact over volume

http://cra.org/resources/best-practice-memos/incentivizing-quality-and-impact-evaluating-scholarship-in-hiring-tenure-and-promotion/

Page 16: Uses and misuses of quantitative indicators of impact

16

Principle 3

Measure locally relevant research using appropriate metrics, including those that build on journal collections in local languages or that cover certain geographic locations. Big international citation databases (used most frequently to derive data used for constructing indicators) still mostly focus on English-language, western journals.  

Page 17: Uses and misuses of quantitative indicators of impact

17

Example: Scientific, popular & public debate publications at Norwegian HE institutions

http://www.scientometrics-school.eu/images/esss1_Sivertsen.pdf After Kyvik, 2005

Page 18: Uses and misuses of quantitative indicators of impact

18

Example: Polish Sociology and Spanish Law

Some research NEEDS to be published in local language (culture-creating role and intended audiences are local and/or practitioner)

Polish sociology in PSCI looks different than that in SSCI (Winclawska,1996/Webster,1998)

Current assessment regime in Spain (rewarding English-language publications) has a detrimental impact of Spanish law research (Hicks, 2015)

Page 19: Uses and misuses of quantitative indicators of impact

19

Principle 4

Metrics-based evaluation, to be trusted, should adhere to the standards of openness and transparency in data collection and analysis.

What data are collected? How is it collected? How are citations captured? What are the exact methods and calculations used to develop indicators? Is the process open to scrutiny by experts and by the assessed?

Page 20: Uses and misuses of quantitative indicators of impact

20

Example: how much of my output is captured?

Subject area Books and book chapters Conference papers Journal articlesHistory 45.6 3.8 50.6Politics and Policy 43.1 10.8 46.1Language 40.5 7.6 51.8Human Society 31.3 5.6 63Philosophy 29.8 5.4 64.8Economics 27.4 8 64.5Law 26.2 1.9 71.9The Arts 25.2 20.3 54.5Education 21.8 23.6 54.5Architecture 20.8 43.6 35.6Psychology 18.9 4.9 76.2Journalism, library 18.6 24.2 57.2Management 13 34 52.9Earth Sciences 8.6 9.2 82.2Medical & Health Sci 6.6 2.9 90.5Biological Sciences 6.6 2.7 90.7Agriculture 6.3 14.7 79Computing 5 62.3 32.8Mathematical Sciences 5 11.2 83.8Engineering 2.9 45.1 52Physical Sciences 2.7 7.3 90Chemical Sciences 2.3 1.9 95.7 L. Butler, 2006

Page 21: Uses and misuses of quantitative indicators of impact

21

Page 22: Uses and misuses of quantitative indicators of impact

22

Principle 5

Those who are evaluated should be able to verify data and the analyses used in the assessment process.

Are all relevant outputs identified, captured and analyzed?

Page 23: Uses and misuses of quantitative indicators of impact

23

Example

Page 24: Uses and misuses of quantitative indicators of impact

24

https://facultyinfo.pitt.edu/

Page 25: Uses and misuses of quantitative indicators of impact

25

Principle 6

Metrics cannot be applied equally across all disciplines

Page 26: Uses and misuses of quantitative indicators of impact

26

Bioc

hem

istry

, Gen

etics

an.

..

Chem

istry

Imm

unol

ogy

and

Micr

obio

...

Neur

oscie

nce

Chem

ical E

ngin

eerin

g

Phar

mac

olog

y, T

oxico

log.

..

Envi

ronm

enta

l Scie

nce

Mate

rials

Scie

nce

Agric

ultu

ral a

nd B

iolo

gica

...

Medi

cine

Phys

ics a

nd A

stro

nom

y

Ener

gy

Psyc

holo

gy

Earth

and

Pla

neta

ry S

cie...

Nurs

ing

Heal

th P

rofe

ssio

ns

Decis

ion

Scie

nces

Dent

istry

Engi

neer

ing

Vete

rinar

y

Busin

ess,

Mana

gem

ent a

n...

Math

emat

ics

Com

pute

r Scie

nce

Econ

omics

, Eco

nom

etric

s...

Socia

l Scie

nces

Arts

and

Hum

aniti

es

0

1

2

3

4

5

6

7

8

9

8.2 7.8 7.8 7.8 7.4

5.5 5.2 5.2 5 4.9 4.7 4.6 4.6 4.43.6 3.5 3.4 3.1 2.8 2.8 2.5 2.5 2.4 2.4 2.1 1.6

Cita

tions

per

Pub

licat

ion

Example: All disciplines are not equal (…bibliometrically)

Page 27: Uses and misuses of quantitative indicators of impact

27

Neurop

sycho

logy a

nd Ph

ysiolo

gical P

sycho

logy

Experi

mental

and C

ognit

ive Ps

ycholo

gy

Develo

pmen

tal an

d Educ

ation

al Psy

cholog

y

Clinic

al Psy

cholog

y

Appli

ed Ps

ycholo

gy

Social

Psycho

logy

0

1

2

3

4

5

6

7

76

4.9 4.8 4.74.1

Cita

tions

per

pub

licat

ion

Psychology – 4.6 CPP

Page 28: Uses and misuses of quantitative indicators of impact

28

Example: Not all disciplines are equal

Differences in citation curves at the category level

0%

2%

4%

6%

8%

10%

12%

2006 2005 2004 2003 2002 2001 2000 1999 1998 1997Cited year

% o

f tot

al c

itatio

ns to

the

cate

gory

Cell Biol (5.9)

Med, Gen Int (7.1)

Math (>10)

Multidisc (7.6)

Econ (>10)

Education(8.3)

Page 29: Uses and misuses of quantitative indicators of impact

29

Example: “novel research” (defined by concentration of distant references)

More likely to be in top 1% in its field and more likely be cited by other papers in top 1%

Delayed recognition Cited outside “home” field Less likely to be published in top IF journal

Wang J, Veugelers R, Stephan P (2016) Bias against novelty in science: A cautionary tale for users of bibliometric indicators. CEPR Discussion Paper No. DP11228; NBER Working Paper No. 22180.

Page 30: Uses and misuses of quantitative indicators of impact

30

Principle 7

Do not rely on a single quantitative indicator when evaluating individual researchers.

Page 31: Uses and misuses of quantitative indicators of impact

31

Author 1

Author 2

Author 3

15 150 1510 100 1010 50 105 25 55 5 54 1 04 0 03 0 03 0 01 0 0

Does not account for:

•insensitive to highly-cited publications

•citation characteristics of publication outlets

•citation characteristics of fields of science

•age of publications

•type of publications

•co-authorship

•self-citations

Page 32: Uses and misuses of quantitative indicators of impact

32

Source: SciVal

Page 33: Uses and misuses of quantitative indicators of impact

33

Principle 8

Sets of indicators can provide a more reliable and multi-dimensional view than a single indicator.

Page 34: Uses and misuses of quantitative indicators of impact

34

https://libraryconnect.elsevier.com/sites/default/files/ELS_LC_metrics_poster_V2.0_researcher_2016.pdf

Page 35: Uses and misuses of quantitative indicators of impact

35

Principles 9 and 10

Goodhart’s Law states that, “when a measure becomes a target, it ceases to be a good measure”.

Every evaluation system creates incentives (intended or unintended) and these, in turn, drive behaviors.

Use of a single indicator (like JIF) opens the evaluation system to undesirable behaviors like gaming. To mitigate against these behaviors multiple indicators should be used.

Furthermore, indicators should be reviewed and updated in line with changing goals of assessment, and new metrics should be considered as they become available.

Page 36: Uses and misuses of quantitative indicators of impact

36

Example: Assessment regime will modify behaviour

Explaining Australia's increased share of ISI publications: the effects of a funding formula based on publication counts (L. Butler, Res Eval, 2003)

“Significant funds are distributed to universities, and within universities, on the basis of aggregate publication counts, with little attention paid to the impact or quality of that output. In consequence, journal publication productivity has increased significantly in the last decade, but its impact has declined.”

Evidence for excellence: has the signal overtaken the substance? An analysis of journal articles submitted to RAE2008 (J. Adams, Digital Science Report, June 2014)

“What researchers actually do under assessment differs from what surveys say they believe about the signals of research excellence. When it comes to the RAE, with the exception of the humanities, academics prioritise journals over other publications, they accelerate publication rates at RAE time, they favour journals with high average citation impact and among those journals they are persuaded that a high Impact Factor beats a convincing individual article.” (p.8)

Page 37: Uses and misuses of quantitative indicators of impact

37

http://www.library.pitt.edu/bibliometric-services