With the citation benchmarking metrics, article performance is measured by comparing the article in question with the "average article" within the same field or journal. For some of these metrics, the calculation is made for you while other metrics require you to do the math and manually make the assessment.
The following metrics are provided for citation benchmarking:
Click on the tabs in the box below to see a description of each metric and how to get the scores.
The Field-Weighted Citation Impact (FWCI) score comes from the Scopus database and shows how the article's citation count compares to similar articles in the same field and timeframe.
In situations where article performance is being judged across a variety of subject areas each with different citation behaviors.
To get an articles's FWCI:
About the Relative Citation Ration (RCR)
iCite provides analysis for articles appearing in PubMed. Citation data are drawn from several data sources: PubMed Central, European PubMed Central, CrossRef, and Web of Science. iCite includes the Relative Citation Ratio (RCR) for an article that is a citation-based metric developed by NIH.
To Get the RCR:
For details about the information on the results page (and more), click on the Help link in the upper right corner of the screen.
CiteScore is a relatively new (December 2016) product from Elsevier, using citation data from the Scopus database to rank journals. In this metric you compare an article's citation count to what CiteScore says would be expected of the average article in this journal.
How to get the comparison between the article and the journal's CiteScore.
In this example, we want to compare an article's performance in 2015 to its journal's 2015 CiteScore; because the 2015 CiteScore is derived from the citation counts of articles from 2012 -2014, to keep the comparison as "apples to apples", the article we pick must have been published from 2012 through 2014. (If you want to compare an older article, use a CiteScore from a year that is no more than 3 years past the article's publication date. Because CiteScores go back only to 2011, articles published prior to 2008 cannot be assessed by this method.)
Lascaris E., Hemmati M., Buldyrev S.V., Stanley H.E., and Angell C.A. Search for a liquid-liquid critical point in model of silica. Journal of Chemical Physics, 140 (22), article number 224502, 2014.
The Journal Impact Factor (JIF) is the original journal ranking metric and it uses citation data from the Web of Science database. In this metric you compare an article's citation count to what the Web of Science (or Journal Citation Reports) says would be expected of the average article in this journal (aka, the Journal Impact Factor or JIF).
How to get the comparison between the article and the journal's JIF.
In this example, we want to compare an article's performance in 2015 to its journal's 2015 JIF; because the 2015 JIF is derived from the citation counts of articles from 2013 -2014, to keep the comparison as "apples to apples", the article we pick must have been published from 2013 or 2014. (If you want to compare an older article, use the Journal Citation Reports database to locate a JIF from a previous year that is no more than 2 years past the article's publication date. Because the Journal Citation Reports database ony keeps JIF scores for a decade, articles published more than 12 years prior to the current JIF cannot be assessed by this method.)
For this example we'll use ...
An article's citation count may be compared to where it falls on the journal's citation distribution curve (CDC). Unfortunately, most journals do not provide a citation distribution curve but a simple citation distribution "list" can be determined easily through either the Web of Science or the Scopus databases. When comparing an article's citation count to the citation distribution of a journal, we recommend using articles from the same year; so if you are assessing a 2014 article, find the citation distribution for articles published in that journal for 2014.
Because this metric is not using an "average" as the benchmark (as you get with CiteScore or the Journal Impact Factor) there is less chance of an outliers' effect artificially inflating the benchmark.
Determining a Journal's Citation Distribution via Scopus or Web of Science:
Although the numbers will be close, don't expect to get exactly the same numbers from both of these databases as they get their data from different sources and have different definitions of "Document Types". Do not mix scores from the different databases - if you are assessing more than one article, compare Scopus data to Scopus data and not Scopus data to Web of Science data.
If you are comfortable with creating graphs within an Excel file, you may also see the actual citation distribution curve and plot your article's citation count on the curve. See this article for how to do it using data from the Web of Science ...
Lariviere, Vincent; Kiermer, Veronique; MacCallum, Catriona J.; McNutt, Marcia; Patterson, Mark; Pulverer, Bernd; Swaminathan, Sowmya; Taylor, Stuart; and Curry, Stephen. A simple proposal for the publication of journal citation distributions. bioRxiv