World Library  
Flag as Inappropriate
Email this Article


Article Id: WHEBN0036281866
Reproduction Date:

Title: Altmetrics  
Author: World Heritage Encyclopedia
Language: English
Subject: Bibliometrics, Open access, Article-level metrics, Journal ranking, Utopia Documents
Collection: Bibliometrics
Publisher: World Heritage Encyclopedia


In scholarly and scientific publishing, altmetrics are non-traditional metrics[1] proposed as an alternative[2] to more traditional citation impact metrics, such as impact factor and h-index. The term altmetrics was proposed in 2010,[3][4] as a generalization of article level metrics,[5] and has its roots in the #altmetrics hashtag. Although altmetrics are often thought of as metrics about articles, they can be applied to people, journals, books, data sets, presentations, videos, source code repositories, web pages, etc.[6] Altmetrics cover not just citation counts, but also other aspects of the impact of a work, such as how many data and knowledge bases refer to it, article views, downloads, or mentions in social media and news media.[7][8]


  • Adoption 1
  • Categories 2
    • Viewed 2.1
    • Discussed 2.2
    • Saved 2.3
    • Cited 2.4
  • Interpretation 3
  • Controversy 4
  • Ongoing research 5
  • See also 6
  • References 7
  • External links 8


Various websites and projects are calculating altmetrics, including ImpactStory,[9][10],[9][11] Plum Analytics,[9][12][13] and CitedIn.[14] Several publishers have started providing such information to readers, including BioMed Central, Public Library of Science,[15] Frontiers,[16] Nature Publishing Group,[17] and Elsevier. For example, Elsevier announced in a press release to be "increasingly looking at additional metrics, including so called Altmetrics, as a measure of influence of journals and authors".[18] Starting in March 2009, the Public Library of Science also introduced article-level metrics for all articles.[15] Funders have started showing interest in alternative metrics,[19] including the UK Medical Research Council.[20] Altmetrics have been used in applications for promotion review by researchers.[21] Furthermore, several universities, including the University of Pittsburgh are experimenting with altmetrics at an institute level.[21]

However, it is also observed that an article needs little attention to jump to the upper quartile of ranked papers,[22] suggesting that not enough sources of altmetrics are currently available to give a balanced picture of impact for the majority of papers.

Important in determining the relative impact of a paper, a service that calculates altmetrics statistics needs a considerably sized knowledge base. The following table shows the number of papers covered by services:

Website Number of papers 2.230.675[23]
ImpactStory 364.000[24]


Altmetrics are a very broad group of metrics, capturing various parts of impact a paper or work can have. A classification of altmetrics was proposed by ImpactStory in September 2012,[25] and a very similar classification is used by the Public Library of Science:[26]

  • Viewed - HTML views and PDF downloads
  • Discussed - journal comments, science blogs, WorldHeritage, Twitter, Facebook and other social media
  • Saved - Mendeley, CiteULike and other social bookmarks
  • Cited - citations in the scholarly literature, tracked by Web of Science, Scopus, CrossRef and others
  • Recommended - for example used by F1000Prime[27]


One of the first alternative metrics to be used was the number of views of a paper. Traditionally, an author would wish to publish in a journal with a high subscription rate, so many people would have access to the research. With the introduction of web technologies it became possible to actually count how often a single paper was looked at. Typically, publishers count the number of HTML views and PDF views. As early as 2004, the BMJ published the number of views for its articles, which was found to be somewhat correlated to citations.[28]


The discussion of a paper can be seen as a metric that captures the potential impact of a paper. Typical sources of data to calculate this metric include Facebook, Google+, Twitter, Science Blogs, and WorldHeritage pages. The correlation between the mentions and likes and citation by primary scientific literature has been studied, and a slight correlation at best was found, e.g. for articles in PubMed.[29] In 2008 the Journal of Medical Internet Research began publishing views and Tweets. These "tweetations" proved to be a good indicator of highly cited articles, leading the author to propose a "Twimpact factor", which is the number of Tweets it receives in the first seven days of publication, as well as a Twindex, which is the rank percentile of an article's Twimpact factor.[30]

Besides Twitter and other streams, blogging has shown to be a powerful platform to discuss literature. Various platforms exist that keep track of which papers are being blogged about. is one that uses this information for calculating metrics, while other tools just report where discussion is happening, such as ResearchBlogging and Chemical blogspace. Moreover, platforms may even provide a formal way of ranking papers or recommending papers otherwise, such as Faculty of 1000 does.


Even more informative is the number of people that bookmark a paper. The idea behind this metric is that someone would not bookmark a paper of little influence to their own work. Providers of such information include science specific social bookmarking services such as CiteULike and Mendeley.


Besides the traditional metrics based on citations in scientific literature, for example as obtained from Google Scholar, CrossRef, PubMed Central, and Scopus, altmetrics also adopts citations in secondary and other knowledge sources. For example, ImpactStory counts the number of times a paper has been referenced by WorldHeritage.[31]


While the concept of altmetrics is questioned,[32] the interpretation of altmetrics in particular is discussed. Proponents of altmetrics make clear that many of the metrics show influence or engagement, rather than impact on the progress of science.[26] It should be noted that even citation-based metrics do not indicate if a high score implies a positive impact on science; that is, papers are also cited in papers that disagree with the cited paper, an issue for example addressed by the Citation Typing Ontology project.[33]


The usefulness of metrics for estimating impact is controversial,[34][35] but the community shows a clear need: funders demand measurables on the impact of their spending. Like other metrics, altmetrics are prone to self-citation, gaming, and other mechanisms to boost one's apparent impact.[36] Additionally, it has been argued that the currently adopted metrics are suggestive of positive impact, while negative metrics are equally important.[37]

However, it should be kept in mind that the metrics are only one of the outcomes of tracking how research is used. Even more informative than knowing how often a paper is cited, is which papers are citing it. That information allows researchers to see how their work is impacting the field (or not). Providers of metrics also typically provide access to the information from which the metrics were calculated. For example, Web of Science shows which are the citing papers, ImpactStory shows which WorldHeritage pages are referencing the paper, and CitedIn shows which databases extracted data from the paper.[38]

Altmetrics can be gamed: for example, likes and mentions can be bought.[36] Altmetrics can be more difficult to standardize than citations. One example is the number of tweets linking to a paper where the number can vary widely depending on how the tweets are collected.[39]

Another source of objections against altmetrics, or any metrics, is how universities are using metrics to rank their employees.[40]

The score on each field does not directly tell you anything about the quality or impact of the paper. For example, a much discussed paper may merely be very controversial: papers discussed on Retraction Watch will typically get high altmetrics score, despite being retracted from the literature.

Altmetrics for more recent articles may be higher because of the increasing uptake of the social web and because articles may be mentioned mainly when they are published.[41] As a result it is not fair to compare the altmetric scores of articles unless they have been published in the same year and, especially for fast increasing social web sites, at similar times in the same year.

Ongoing research

The specific use cases and characteristics is an active research field in bibliometrics, providing the much needed data to measure the impact of altmetrics itself. Public Library of Science has an Altmetrics Collection[42] and the Information Standards Quarterly recently published a special issue on altmetrics.[43]

See also


  1. ^ "Altmetrics is the study and use of non-traditional scholarly impact measures that are based on activity in web-based environments" [1]
  2. ^ "The “alt” does indeed stand for “alternative”" Jason Priem, leading author in the Altimetrics Manifesto [2]
  3. ^ J. Priem (@jasonpriem), I like the term #articlelevelmetrics, but it fails to imply *diversity* of measures. Lately, I'm liking #altmetrics., 4:28 AM - 29 Sep 10, Tweet
  4. ^ J. Priem, D. Taraborelli, P. Groth, C. Neylon, altmetrics: a manifesto, 2010,
  5. ^ Peter Binfield Article-Level Metrics at PLoS - what are they, and why should you care on YouTube, 2009
  6. ^ E. Adie (@stew), @McDawg altmetrics could cover person, uni or journal level stats, non-trad contributions & items. ALMs a subset, fixes article as focus, 5:45 PM - 10 Aug 13, Tweet
  7. ^ McFedries, P. (2012). "Measuring the impact of altmetrics [Technically Speaking]". IEEE Spectrum 49 (8): 28.  
  8. ^ Galligan, F.; Dyas-Correia, S. (2013). "Altmetrics: Rethinking the Way We Measure". Serials Review 39: 56.  
  9. ^ a b c Liu, Jean; Euan Adie (8 July 2013). "New perspectives on article-level metrics: developing ways to assess research uptake and impact online". Insights 26 (2): 153.  
  10. ^ "FAQ". ImpactStory. Retrieved 13 August 2013. 
  11. ^ "About us". Altmetric. Retrieved 13 August 2013. 
  12. ^ "About Us". Plum Analytics. Retrieved 19 August 2013. 
  13. ^ "Plum Analytics Metrics Overview". Plum Analytics. Retrieved 19 August 2013. 
  14. ^ "Included resources". CitedIn. Retrieved 16 August 2013. 
  15. ^ a b "Article-Level Metrics Information". PLoS ONE. 2005-07-01. Retrieved 2012-05-29. 
  16. ^
  17. ^ Press release: Article level metrics on, 2012,
  18. ^ Press release: Elsevier Announces 2012 Journal Impact Factor Highlights, 2012,
  19. ^ Piwowar, Heather (2013). "Altmetrics: Value all research products". Nature.  
  20. ^ Viney, I. (2013). "Altmetrics: Research council responds". Nature 494 (7436): 176.  
  21. ^ a b Kwok, Roberta (2013). "Research impact: Altmetrics make their mark". Nature 500 (7463): 491.  
  22. ^ J. Kelly, Altmetric rankings, 22 August 2013,
  23. ^ Determined on 6 July 2014 using the altmetrics record for a specific paper, in the Score tab.
  24. ^ ImpactStory (@ImpactStory). "@egonwillighagen Sorry, we've got no automated reporting tool for the db size yet. But I just checked and it's just over 364k items." 2013-10-20, 8:10. Tweet.
  25. ^ "A new framework for altmetrics". ImpactStory Blog. 2012-09-14. Retrieved 2013-08-13. 
  26. ^ a b Lin, J.; Fenner, M. (2013). "Altmetrics in Evolution: Defining and Redefining the Ontology of Article-Level Metrics". Information Standards Quarterly 25 (2): 20.  
  27. ^ F1000Prime
  28. ^ Perneger, T. V (2004). "Relation between online "hit counts" and subsequent citations: Prospective study of research papers in the BMJ". BMJ 329 (7465): 546–7.  
  29. ^ Stefanie Haustein; Isabella Peters; Sugimoto, Cassidy R.; Mike Thelwall; Vincent Larivière (2013). "Tweeting biomedicine: an analysis of tweets and citations in the biomedical literature". arXiv:1308.1838 [cs.DL].
  30. ^ Eysenbach, Gunther (2011). "Can Tweets Predict Citations? Metrics of Social Impact Based on Twitter and Correlation with Traditional Metrics of Scientific Impact". Journal of Medical Internet Research 13 (4): e123.  
  31. ^ "FAQ: which metrics are measured?". ImpactStory. Retrieved 13 August 2013. 
  32. ^ Jump, Paul (23 August 2012). "Research Intelligence - Alt-metrics: fairer, faster impact data?". Times Higher Education. Retrieved 13 August 2013. 
  33. ^ Shotton, D. (2010). "CiTO, the Citation Typing Ontology". Journal of Biomedical Semantics 1 (Suppl 1): S6–S1.  
  34. ^ Mike Buschman; Andrea Michalek (April–May 2013). "Are Alternative Metrics Still Alternative?". asis&t Bulletin. Retrieved 21 August 2013. 
  35. ^ Cheung, M. K. (2013). "Altmetrics: Too soon for use in assessment". Nature 494 (7436): 176.  
  36. ^ a b J. Beall, Article-Level Metrics: An Ill-Conceived and Meretricious Idea, 2013,
  37. ^ Holbrook, J. B.; Barr, K. R.; Brown, K. W. (2013). "Research impact: We need negative metrics too". Nature 497 (7450): 439.  
  38. ^ Waagmeester, A.; Evelo, C. (2011). "Measuring impact in online resources with the CI­number (the CitedIn Number for online impact)". Nature Precedings.  
  39. ^ Chamberlain, S. (2013). "Consuming Article-Level Metrics: Observations and Lessons". Information Standards Quarterly 25 (2): 4–2.  
  40. ^ David Colquhoun, How should universities be run to get the best out of people?, 2007
  41. ^ Thelwall, M.; Haustein, S.; Larivière, V.; Sugimoto, C. R. (2013). "Do Altmetrics Work? Twitter and Ten Other Social Web Services". PLoS ONE 8 (5): e64841.  
  42. ^ Priem, J.; Groth, P.; Taraborelli, D. (2012). Ouzounis, Christos A, ed. "The Altmetrics Collection". PLoS ONE 7 (11): e48753.  
  43. ^ "Summer 2013 Volume 25, no. 2". Information Standards Quarterly. Retrieved 14 August 2013. 

External links

  • Altmetrics Manifesto
  • NISO Altmetrics Standards Project White Paper
  • Special issue or Research Trends: Issue 37 – June 2014
This article was sourced from Creative Commons Attribution-ShareAlike License; additional terms may apply. World Heritage Encyclopedia content is assembled from numerous content providers, Open Access Publishing, and in compliance with The Fair Access to Science and Technology Research Act (FASTR), Wikimedia Foundation, Inc., Public Library of Science, The Encyclopedia of Life, Open Book Publishers (OBP), PubMed, U.S. National Library of Medicine, National Center for Biotechnology Information, U.S. National Library of Medicine, National Institutes of Health (NIH), U.S. Department of Health & Human Services, and, which sources content from all federal, state, local, tribal, and territorial government publication portals (.gov, .mil, .edu). Funding for and content contributors is made possible from the U.S. Congress, E-Government Act of 2002.
Crowd sourced content that is contributed to World Heritage Encyclopedia is peer reviewed and edited by our editorial staff to ensure quality scholarly research articles.
By using this site, you agree to the Terms of Use and Privacy Policy. World Heritage Encyclopedia™ is a registered trademark of the World Public Library Association, a non-profit organization.

Copyright © World Library Foundation. All rights reserved. eBooks from Project Gutenberg are sponsored by the World Library Foundation,
a 501c(4) Member's Support Non-Profit Organization, and is NOT affiliated with any governmental agency or department.