The Leiden Manifesto for research metrics (LM) is a list of "ten principles to guide research evaluation",[1] published as a comment in Volume 520, Issue 7548 of Nature, on 22 April 2015. It was formulated by public policy professor Diana Hicks, scientometrics professor Paul Wouters, and their colleagues at the 19th International Conference on Science and Technology Indicators, held between 3–5 September 2014 in Leiden, The Netherlands.[2]
The LM was proposed as a guide to combat misuse of bibliometrics when evaluating scientific research literature. Examples of commonly used bibliometrics for science, or scientometrics, are the h-index, impact factor, and websites displaying indicators such as Altmetrics. According to Hicks et al., these metrics often pervasively misguide evaluations of scientific material.[3]
The Leiden Manifesto for research metrics | |
Authors: | Diana Hicks, Paul Wouters, Ludo Waltman, Sarah de Rijcke, Ismael Rafols |
Publisher: | Nature |
Pub Date: | 22 April 2015 |
Website: | leidenmanifesto.org |
Motivations for codification of the Leiden Manifesto arose from a growing worry that "impact-factor obsession" was leading to inadequate judgement of scientific material that should be worthy of fair evaluation. The lead author of the LM, Diana Hicks, hoped that publishing in Nature would expand the ideas already commonplace in the scientometrics sphere to the broader scientific community.[4] Although the principles of the LM are not new to scientometricians, Hicks et al. desired to create a unification of ideas to guide future editors and reviewers of scientific literature.
Interest in the reform of how research assessment is conducted is an ongoing debate in the scientific community. A well known declaration intended to quell the use of the impact factor, the San Francisco Declaration on Research Assessment was published around two years prior to the manifesto. The LM broadened the ideas presented in the DORA, which has now since been signed by over 2000 organizations and over 15000 individuals.[5]
One of the main concerns about overuse of citation-based performance indicators came from the observation that smaller research organizations and institutions may be negatively affected by their metric indices. In one public debate at the Centre for Science and Technology Studies at Leiden University, it was acknowledged that indicators which measure citations may give "more weight to publications from fields with a high expected number of citations than to publications from fields with a low expected number of citations".[6]
Although the main focus of the LM is based on the use of scientometrics for research evaluation, in its background, Hicks et al. also explain why overuse of metrics can adversely affect the wider scholarly community, such as the position of universities in global rankings.[7] [8] According to Hicks et al., scientific metrics such as citation rate are used far too much for ranking the quality of universities (and thus the quality of their research output).
The background for the LM describes why misusing metrics is becoming a larger problem in the scientific community. The journal impact factor, originally created by Eugene Garfield as a method for librarians to collect data to facilitate selecting journals to purchase, is now mainly used as a method of judging journal quality. This is seen by Hicks et al. as an abuse of data in order to examine research too hastily. For example, an impact factor, while a good metric to measure the size and experience of a journal, may or may not be sufficient enough to accurately describe the quality of its papers, and even less so for a single paper.
Consisting of ten concise principles, along with a description for each, the Leiden Manifesto aims to reconstruct the way that research evaluations by academic publishers and scientific institutions are done. Its emphasis lies in detailed and close evaluation of research, rather than the excessive use of quantitative data in evaluations. It aims to promote academic excellence and fairness with thorough scrutiny, as well as remove possible perverse incentives of using scientometrics, such as judgement of academic capability and university quality.
The ten principles of the Leiden Manifesto are as follows:
In 2016, the Leiden Manifesto was presented with the John Ziman Award by the European Association for the Study of Science and Technology for its effort to widen scientometrics knowledge to the scientific community as a whole. EASST president Fred Steward stated that the LM "emphasizes situatedness, in terms of different cognitive domains and research missions as well as the wider socioeconomic, national and regional context".[11] This award contributed to the solidification of the LM as a continuing public debate for the scholarly community. The award was received well by lead author Diana Hicks.
The Leiden Manifesto gained popularity following its publication, mainly by the scholarly publishing community which looked to reform their practices. The LM is often cited alongside other similar publications; those being the DORA and UK based academic review The Metric Tide.[12] In a public statement by the University of Leeds, support for committing to using the principles of the LM along with DORA and The Metric Tide was issued in order to further research excellence.[13]
LIBER, a collaboration of European research libraries issued a substantial review on the LM in 2017. It concluded as a viewpoint that the LM was a "solid foundation" on which academic libraries could base their assessment of metrics.[14]
Elsevier, a global leader in research publishing and information analytics, announced on 14 July 2020 that it was to endorse the LM to guide its development of improved research evaluation. Elsevier stated that the principles of the manifesto were already close in nature to their 2019 CiteScore metrics, which was in summary "improved calculation methodology" for "a more robust, fair and faster indicator of research impact".[15] This alignment further popularized the LM, and also illustrated the shift in practices of evaluation by prominent research publications.
The Loughborough University LIS-Bibliometrics committee chose to base their principles on those of the LM instead of the DORA, because according to their policy manager Elizabeth Gadd, the LM takes a "broader approach to the responsible use of all bibliometrics across a range of disciplines and settings".[16] Stephen Curry, the chair of the DORA steering committee, commented on this statement by emphasizing that DORA was aiming to extend its "disciplinary and geographical reach".[17] However, he still made it clear that a university should have the right to choose to follow either the DORA or the LM, or neither, as long as reasonable justification was provided.
David Moher et al. referenced the LM in a perspective for Issues in Science and Technology that the "right questions"(i.e. research planning, timeframe, reproducibility, and results) for assessing scientists were not being asked by academic institutions. Moher et al. criticize the obsession of journal impact factors and "gaming" by investigators of scientometrics.[18] Instead, Moher et al. advocate for usage of DORA and the LM when assessing individual scientists and research.
T. Kanchan and K. Krishan describe by a letter in Science and Engineering Ethics why the LM is "one of the best criteria" for assessing scientific research, especially considering the "rat race" for publications in the scholarly community. Kanchan and Krishan emphasize that use of the LM will lead to "progress of science and society at large".[19]