H-index: Difference between revisions
m Various citation & identifier cleanup, plus AWB genfixes (arxiv version pointless when published) |
Citation bot (talk | contribs) m Alter: year, issue. Add: arxiv, pmc, pages, volume, hdl, citeseerx, year, issue, pmid, website. Removed parameters. Formatted dashes. You can use this bot yourself. Report bugs here. | Headbomb |
||
Line 1: | Line 1: | ||
{{About|the index of scientific research impact|the economic measure|Herfindahl index}} |
{{About|the index of scientific research impact|the economic measure|Herfindahl index}} |
||
{{DISPLAYTITLE:''h''-index}} |
{{DISPLAYTITLE:''h''-index}} |
||
The '''''h''-index''' is an [[author-level metric]] that attempts to measure both the [[productivity]] and [[citation impact]] of the [[scientific publication|publications]] of a [[scientist]] or scholar. The index is based on the set of the scientist's most cited papers and the number of citations that they have received in other publications. The index can also be applied to the productivity and impact of a [[scholarly journal]]<ref>{{cite web|url=http://googlescholar.blogspot.com.br/2012/04/google-scholar-metrics-for-publications.html|title=Google Scholar Metrics for Publications|last=Suzuki|first=Helder|date=2012| |
The '''''h''-index''' is an [[author-level metric]] that attempts to measure both the [[productivity]] and [[citation impact]] of the [[scientific publication|publications]] of a [[scientist]] or scholar. The index is based on the set of the scientist's most cited papers and the number of citations that they have received in other publications. The index can also be applied to the productivity and impact of a [[scholarly journal]]<ref>{{cite web|url=http://googlescholar.blogspot.com.br/2012/04/google-scholar-metrics-for-publications.html|title=Google Scholar Metrics for Publications|last=Suzuki|first=Helder|date=2012|website=googlescholar.blogspot.com.br|archive-url=|archive-date=|dead-url=|access-date=}}</ref> as well as a group of scientists, such as a department or university or country.<ref>{{Cite journal | doi = 10.1016/j.wneu.2011.01.015| pmid = 21839937| title = Finding a Way Through the Scientific Literature: Indexes and Measures| journal = World Neurosurgery| volume = 76| issue = 1–2| pages = 36–38| year = 2011| last1 = Jones | first1 = T. | last2 = Huggett | first2 = S. | last3 = Kamalski | first3 = J. }}</ref> The index was suggested in 2005 by [[Jorge E. Hirsch]], a physicist at [[University of California, San Diego|UCSD]], as a tool for determining [[theoretical physics|theoretical physicists]]' relative quality<ref name=Hirsch2005>{{Cite journal |last = Hirsch |first = J. E. |authorlink = Jorge E. Hirsch |date = 15 November 2005 |title = An index to quantify an individual's scientific research output |journal = [[Proceedings of the National Academy of Sciences|PNAS]] |volume = 102 |issue = 46 |pages = 16569–72 |doi = 10.1073/pnas.0507655102 |pmid = 16275915 |pmc = 1283832 |bibcode = 2005PNAS..10216569H |arxiv = physics/0508025 |year = 2005 }}</ref> and is sometimes called the ''Hirsch index'' or ''Hirsch number''. |
||
== Definition and purpose == |
== Definition and purpose == |
||
[[File:H-index-en.svg|300px|thumb|right|''h''-index from a plot of decreasing citations for numbered papers]] |
[[File:H-index-en.svg|300px|thumb|right|''h''-index from a plot of decreasing citations for numbered papers]] |
||
The definition of the index is that a scholar with an index of ''h'' has published ''h'' papers each of which has been cited in other papers at least ''h'' times.<ref name=McDonald2005>{{cite web |url = http://www.physorg.com/news7971.html |title = Physicist Proposes New Way to Rank Scientific Output |accessdate = 13 May 2010 |author = McDonald, Kim | |
The definition of the index is that a scholar with an index of ''h'' has published ''h'' papers each of which has been cited in other papers at least ''h'' times.<ref name=McDonald2005>{{cite web |url = http://www.physorg.com/news7971.html |title = Physicist Proposes New Way to Rank Scientific Output |accessdate = 13 May 2010 |author = McDonald, Kim |website = |publisher = [[PhysOrg]] |date = 8 November 2005 }}</ref> Thus, the ''h''-index reflects both the number of publications and the number of citations per publication. The index is designed to improve upon simpler measures such as the total number of citations or publications. The index works properly only for comparing scientists working in the same field; citation conventions differ widely among different fields. |
||
== Calculation == |
== Calculation == |
||
Line 35: | Line 35: | ||
The ''h''-index can be manually determined using citation databases or using automatic tools. Subscription-based databases such as [[Scopus]] and the [[Web of Science]] provide automated calculators. Harzing's ''Publish or Perish'' program calculates the ''h''-index based on [[Google Scholar]] entries. From July 2011 [[Google]] have provided an automatically-calculated ''h''-index and [[i10-index]] within their own [[Google Scholar]] profile.<ref>[https://scholar.google.com/intl/en/scholar/citations.html Google Scholar Citations Help], retrieved 2012-09-18.</ref> In addition, specific databases, such as the [[INSPIRE-HEP]] database can automatically calculate the ''h''-index for researchers working in [[high energy physics]]. |
The ''h''-index can be manually determined using citation databases or using automatic tools. Subscription-based databases such as [[Scopus]] and the [[Web of Science]] provide automated calculators. Harzing's ''Publish or Perish'' program calculates the ''h''-index based on [[Google Scholar]] entries. From July 2011 [[Google]] have provided an automatically-calculated ''h''-index and [[i10-index]] within their own [[Google Scholar]] profile.<ref>[https://scholar.google.com/intl/en/scholar/citations.html Google Scholar Citations Help], retrieved 2012-09-18.</ref> In addition, specific databases, such as the [[INSPIRE-HEP]] database can automatically calculate the ''h''-index for researchers working in [[high energy physics]]. |
||
Each database is likely to produce a different ''h'' for the same scholar, because of different coverage.<ref>{{Cite journal | doi = 10.1007/s11192-008-0216-y| title = Which h-index? – A comparison of WoS, Scopus and Google Scholar| journal = Scientometrics| volume = 74| issue = 2| pages = 257–71| year = 2007| last1 = Bar-Ilan | first1 = J. }}</ref> A detailed study showed that the Web of Science has strong coverage of journal publications, but poor coverage of high impact conferences. Scopus has better coverage of conferences, but poor coverage of publications prior to 1996; Google Scholar has the best coverage of conferences and most journals (though not all), but like Scopus has limited coverage of pre-1990 publications.<ref name="Meho2007">{{cite journal |last=Meho |first=L. I. |authorlink= |author2=Yang, K. |year=2007 |title=Impact of Data Sources on Citation Counts and Rankings of LIS Faculty: Web of Science vs. Scopus and Google Scholar |journal=[[Journal of the American Society for Information Science and Technology]] |volume=58 |issue=13 |pages=2105–25 |doi=10.1002/asi.20677 |url=}}</ref><ref name=Meho&Yang2006>{{Cite arXiv |date=23 December 2006 |author1=Meho, L. I. |author2=Yang, K |title=A New Era in Citation and Bibliometric Analyses: Web of Science, Scopus, and Google Scholar |eprint=cs/0612132 |postscript=}} (preprint of paper published as 'Impact of data sources on citation counts and rankings of LIS faculty: Web of Science versus Scopus and Google Scholar', in ''Journal of the American Society for Information Science and Technology'', Vol. '''58''', No. 13, 2007, 2105–25)</ref> The exclusion of conference proceedings papers is a particular problem for scholars in [[computer science]], where conference proceedings are considered an important part of the literature.<ref>{{Cite journal |title=Research Evaluation for Computer Science |journal=[[Communications of the ACM]] |volume=52 |issue=4 |year=2009 |pages=31–34 |last1=Meyer |first1=Bertrand |author1-link=Bertrand Meyer |last2=Choppy |first2=Christine |last3=Staunstrup |first3=Jørgen |last4=Van Leeuwen |first4=Jan |author4-link=Jan van Leeuwen |doi=10.1145/1498765.1498780 |url=http://www.informatics-europe.org/docs/research-eval.php |postscript={{inconsistent citations}}}}.</ref> Google Scholar has been criticized for producing "phantom citations," including [[gray literature]] in its citation counts, and failing to follow the rules of [[Boolean logic]] when combining search terms.<ref>{{cite journal |last=Jacsó |first=Péter |year=2006 |title=Dubious hit counts and cuckoo's eggs |journal=Online Information Review |volume=30 |issue=2 |pages=188–93 |doi=10.1108/14684520610659201 |url= |accessdate= }}</ref> For example, the Meho and Yang study found that Google Scholar identified 53% more citations than Web of Science and Scopus combined, but noted that because most of the additional citations reported by Google Scholar were from low-impact journals or conference proceedings, they did not significantly alter the relative ranking of the individuals. It has been suggested that in order to deal with the sometimes wide variation in ''h'' for a single academic measured across the possible citation databases, one should assume false negatives in the databases are more problematic than false positives and take the maximum ''h'' measured for an academic.<ref>{{cite journal |last=Sanderson |first=Mark |year=2008 |title=Revisiting ''h'' measured on UK LIS and IR academics |journal=Journal of the American Society for Information Science and Technology |volume=59 |issue=7 |pages=1184–90 |doi=10.1002/asi.20771 |url=}}</ref> |
Each database is likely to produce a different ''h'' for the same scholar, because of different coverage.<ref>{{Cite journal | doi = 10.1007/s11192-008-0216-y| title = Which h-index? – A comparison of WoS, Scopus and Google Scholar| journal = Scientometrics| volume = 74| issue = 2| pages = 257–71| year = 2007| last1 = Bar-Ilan | first1 = J. }}</ref> A detailed study showed that the Web of Science has strong coverage of journal publications, but poor coverage of high impact conferences. Scopus has better coverage of conferences, but poor coverage of publications prior to 1996; Google Scholar has the best coverage of conferences and most journals (though not all), but like Scopus has limited coverage of pre-1990 publications.<ref name="Meho2007">{{cite journal |last=Meho |first=L. I. |authorlink= |author2=Yang, K. |year=2007 |title=Impact of Data Sources on Citation Counts and Rankings of LIS Faculty: Web of Science vs. Scopus and Google Scholar |journal=[[Journal of the American Society for Information Science and Technology]] |volume=58 |issue=13 |pages=2105–25 |doi=10.1002/asi.20677 |url=}}</ref><ref name=Meho&Yang2006>{{Cite arXiv |date=23 December 2006 |author1=Meho, L. I. |author2=Yang, K |title=A New Era in Citation and Bibliometric Analyses: Web of Science, Scopus, and Google Scholar |eprint=cs/0612132 |postscript=}} (preprint of paper published as 'Impact of data sources on citation counts and rankings of LIS faculty: Web of Science versus Scopus and Google Scholar', in ''Journal of the American Society for Information Science and Technology'', Vol. '''58''', No. 13, 2007, 2105–25)</ref> The exclusion of conference proceedings papers is a particular problem for scholars in [[computer science]], where conference proceedings are considered an important part of the literature.<ref>{{Cite journal |title=Research Evaluation for Computer Science |journal=[[Communications of the ACM]] |volume=52 |issue=4 |year=2009 |pages=31–34 |last1=Meyer |first1=Bertrand |author1-link=Bertrand Meyer |last2=Choppy |first2=Christine |last3=Staunstrup |first3=Jørgen |last4=Van Leeuwen |first4=Jan |author4-link=Jan van Leeuwen |doi=10.1145/1498765.1498780 |url=http://www.informatics-europe.org/docs/research-eval.php |postscript={{inconsistent citations}}}}.</ref> Google Scholar has been criticized for producing "phantom citations," including [[gray literature]] in its citation counts, and failing to follow the rules of [[Boolean logic]] when combining search terms.<ref>{{cite journal |last=Jacsó |first=Péter |year=2006 |title=Dubious hit counts and cuckoo's eggs |journal=Online Information Review |volume=30 |issue=2 |pages=188–93 |doi=10.1108/14684520610659201 |url= |accessdate= }}</ref> For example, the Meho and Yang study found that Google Scholar identified 53% more citations than Web of Science and Scopus combined, but noted that because most of the additional citations reported by Google Scholar were from low-impact journals or conference proceedings, they did not significantly alter the relative ranking of the individuals. It has been suggested that in order to deal with the sometimes wide variation in ''h'' for a single academic measured across the possible citation databases, one should assume false negatives in the databases are more problematic than false positives and take the maximum ''h'' measured for an academic.<ref>{{cite journal |last=Sanderson |first=Mark |year=2008 |title=Revisiting ''h'' measured on UK LIS and IR academics |journal=Journal of the American Society for Information Science and Technology |volume=59 |issue=7 |pages=1184–90 |doi=10.1002/asi.20771 |url=|citeseerx=10.1.1.474.1990 }}</ref> |
||
== Comparing results across fields and career levels == |
== Comparing results across fields and career levels == |
||
Line 44: | Line 44: | ||
Among the 22 scientific disciplines listed in the [[Thomson Reuters]] Essential Science Indicators Citation Thresholds [thus excluding non-science academics], physics has the second most citations after space science.<ref name="ESI2010">{{cite web |title = Citation Thresholds (Essential Science Indicators) |url = http://www.sciencewatch.com/about/met/thresholds/ |website = Science Watch |publisher = Thomson Reuters |date = May 1, 2010 |accessdate = 13 May 2010|archiveurl=https://web.archive.org/web/20100505011410/http://sciencewatch.com/about/met/thresholds/ |archivedate=5 May 2010 }}</ref> During the period January 1, 2000 – February 28, 2010, a physicist had to receive 2073 citations to be among the most cited 1% of physicists in the world.<ref name="ESI2010" /> The threshold for space science is the highest (2236 citations), and physics is followed by clinical medicine (1390) and molecular biology & genetics (1229). Most disciplines, such as environment/ecology (390), have fewer scientists, fewer papers, and fewer citations.<ref name="ESI2010" /> Therefore, these disciplines have lower citation thresholds in the Essential Science Indicators, with the lowest citation thresholds observed in social sciences (154), computer science (149), and multidisciplinary sciences (147).<ref name="ESI2010" /> |
Among the 22 scientific disciplines listed in the [[Thomson Reuters]] Essential Science Indicators Citation Thresholds [thus excluding non-science academics], physics has the second most citations after space science.<ref name="ESI2010">{{cite web |title = Citation Thresholds (Essential Science Indicators) |url = http://www.sciencewatch.com/about/met/thresholds/ |website = Science Watch |publisher = Thomson Reuters |date = May 1, 2010 |accessdate = 13 May 2010|archiveurl=https://web.archive.org/web/20100505011410/http://sciencewatch.com/about/met/thresholds/ |archivedate=5 May 2010 }}</ref> During the period January 1, 2000 – February 28, 2010, a physicist had to receive 2073 citations to be among the most cited 1% of physicists in the world.<ref name="ESI2010" /> The threshold for space science is the highest (2236 citations), and physics is followed by clinical medicine (1390) and molecular biology & genetics (1229). Most disciplines, such as environment/ecology (390), have fewer scientists, fewer papers, and fewer citations.<ref name="ESI2010" /> Therefore, these disciplines have lower citation thresholds in the Essential Science Indicators, with the lowest citation thresholds observed in social sciences (154), computer science (149), and multidisciplinary sciences (147).<ref name="ESI2010" /> |
||
Numbers are very different in social science disciplines: The ''Impact of the Social Sciences'' team at [[London School of Economics]] found that social scientists in the United Kingdom had lower average h-indices. The h-indices for ("full") professors, based on [[Google Scholar]] data ranged from 2.8 (in law), through 3.4 (in political science), 3.7 (in sociology), 6.5 (in geography) and 7.6 (in economics). On average across the disciplines, a professor in the social sciences had an h-index about twice that of a lecturer or a senior lecturer, though the difference was the smallest in geography.<ref>{{cite web|url=http://blogs.lse.ac.uk/impactofsocialsciences/the-handbook/chapter-3-key-measures-of-academic-influence|title=Impact of Social Sciences – 3: Key Measures of Academic Influence| |
Numbers are very different in social science disciplines: The ''Impact of the Social Sciences'' team at [[London School of Economics]] found that social scientists in the United Kingdom had lower average h-indices. The h-indices for ("full") professors, based on [[Google Scholar]] data ranged from 2.8 (in law), through 3.4 (in political science), 3.7 (in sociology), 6.5 (in geography) and 7.6 (in economics). On average across the disciplines, a professor in the social sciences had an h-index about twice that of a lecturer or a senior lecturer, though the difference was the smallest in geography.<ref>{{cite web|url=http://blogs.lse.ac.uk/impactofsocialsciences/the-handbook/chapter-3-key-measures-of-academic-influence|title=Impact of Social Sciences – 3: Key Measures of Academic Influence|website=Impact of Social Sciences}}</ref> |
||
== Advantages == |
== Advantages == |
||
Line 59: | Line 59: | ||
<!-- * The ''h''-index does not account for confounding factors such as "gratuitous authorship", the so-called [[Matthew effect (sociology)|Matthew effect]], and the favorable citation bias associated with review articles. Again, this is a problem for all other metrics using publications or citations. --> |
<!-- * The ''h''-index does not account for confounding factors such as "gratuitous authorship", the so-called [[Matthew effect (sociology)|Matthew effect]], and the favorable citation bias associated with review articles. Again, this is a problem for all other metrics using publications or citations. --> |
||
* The ''h''-index has been found in one study to have slightly less predictive [[accuracy and precision]] than the simpler measure of mean citations per paper.<ref name="pmid17183295">{{cite journal |title=Measures for measures |journal=[[Nature (journal)|Nature]] |volume=444 |issue=7122 |pages=1003–04 |year=2006 |pmid=17183295 |doi=10.1038/4441003a |bibcode=2006Natur.444.1003L |last2=Jackson |first2=Andrew D. |first3=Benny E. |last3=Lautrup |author1 = Sune Lehmann}}</ref> However, this finding was contradicted by another study by Hirsch.<ref>{{cite journal |author=Hirsch J. E. |title=Does the h-index have predictive power? |journal=PNAS |volume=104 |pages=19193–98 |year=2007 |url=http://www.pnas.org/content/104/49/19193.full |doi=10.1073/pnas.0707962104 |pmid=18040045 |issue=49 |pmc=2148266 |bibcode=2007PNAS..10419193H |arxiv=0708.0646}}</ref> |
* The ''h''-index has been found in one study to have slightly less predictive [[accuracy and precision]] than the simpler measure of mean citations per paper.<ref name="pmid17183295">{{cite journal |title=Measures for measures |journal=[[Nature (journal)|Nature]] |volume=444 |issue=7122 |pages=1003–04 |year=2006 |pmid=17183295 |doi=10.1038/4441003a |bibcode=2006Natur.444.1003L |last2=Jackson |first2=Andrew D. |first3=Benny E. |last3=Lautrup |author1 = Sune Lehmann}}</ref> However, this finding was contradicted by another study by Hirsch.<ref>{{cite journal |author=Hirsch J. E. |title=Does the h-index have predictive power? |journal=PNAS |volume=104 |pages=19193–98 |year=2007 |url=http://www.pnas.org/content/104/49/19193.full |doi=10.1073/pnas.0707962104 |pmid=18040045 |issue=49 |pmc=2148266 |bibcode=2007PNAS..10419193H |arxiv=0708.0646}}</ref> |
||
* The ''h''-index is a [[natural number]] that reduces its discriminatory power. [[Frances P. Ruane|Ruane]] and [[Richard Tol|Tol]] therefore propose a [[rational number|rational]] ''h''-index that interpolates between ''h'' and ''h'' + 1.<ref>{{Cite journal |author=[[Frances P. Ruane|Frances Ruane]] & [[Richard Tol|Richard S. J. Tol]] |title=Rational (successive) h -indices: An application to economics in the Republic of Ireland |journal=[[Scientometrics (journal)|Scientometrics]] |volume=75 |issue=2 |year=2008 |pages=395–405 |doi=10.1007/s11192-007-1869-7|last2=Tol }}</ref> |
* The ''h''-index is a [[natural number]] that reduces its discriminatory power. [[Frances P. Ruane|Ruane]] and [[Richard Tol|Tol]] therefore propose a [[rational number|rational]] ''h''-index that interpolates between ''h'' and ''h'' + 1.<ref>{{Cite journal |author=[[Frances P. Ruane|Frances Ruane]] & [[Richard Tol|Richard S. J. Tol]] |title=Rational (successive) h -indices: An application to economics in the Republic of Ireland |journal=[[Scientometrics (journal)|Scientometrics]] |volume=75 |issue=2 |year=2008 |pages=395–405 |doi=10.1007/s11192-007-1869-7|last2=Tol |hdl=1871/31768 }}</ref> |
||
* The ''h''-index can be manipulated through self-citations,<ref name="10.1007/s11192-017-2330-1">{{cite journal |author=Gálvez RH |title=Assessing author self-citation as a mechanism of relevant knowledge diffusion |journal=Scientometrics |date=March 2017 |doi=10.1007/s11192-017-2330-1}}</ref><ref>{{Cite journal |author=Christoph Bartneck & Servaas Kokkelmans |title=Detecting h-index manipulation through self-citation analysis |journal=[[Scientometrics (journal)|Scientometrics]] |volume=87 |issue=1 |year=2011 |pages=85–98 |doi=10.1007/s11192-010-0306-5 |pmid=21472020 |pmc=3043246 |last2=Kokkelmans }}</ref><ref>{{Cite journal |author=Emilio Ferrara & Alfonso Romero |title=Scientific impact evaluation and the effect of self-citations: Mitigating the bias by discounting the h-index |journal=[[Journal of the American Society for Information Science and Technology]] |volume=64 |issue=11 |year=2013 |pages=2332–39 |doi=10.1002/asi.22976|last2=Romero |arxiv=1202.3119 }}</ref> and if based on [[Google Scholar]] output, then even computer-generated documents can be used for that purpose, e.g. using [[SCIgen]].<ref name=ike-antkare>{{cite report |author=Labbé, Cyril |title=Ike Antkare one of the great stars in the scientific firmament |date=2010 |work=Laboratoire d'Informatique de Grenoble RR-LIG-2008 (technical report) |publisher=[[Joseph Fourier University]] |url=http://rr.liglab.fr/research_report/RR-LIG-008.pdf |format=PDF}}</ref> |
* The ''h''-index can be manipulated through self-citations,<ref name="10.1007/s11192-017-2330-1">{{cite journal |author=Gálvez RH |title=Assessing author self-citation as a mechanism of relevant knowledge diffusion |journal=Scientometrics |volume=111 |issue=3 |pages=1801–1812 |date=March 2017 |doi=10.1007/s11192-017-2330-1}}</ref><ref>{{Cite journal |author=Christoph Bartneck & Servaas Kokkelmans |title=Detecting h-index manipulation through self-citation analysis |journal=[[Scientometrics (journal)|Scientometrics]] |volume=87 |issue=1 |year=2011 |pages=85–98 |doi=10.1007/s11192-010-0306-5 |pmid=21472020 |pmc=3043246 |last2=Kokkelmans }}</ref><ref>{{Cite journal |author=Emilio Ferrara & Alfonso Romero |title=Scientific impact evaluation and the effect of self-citations: Mitigating the bias by discounting the h-index |journal=[[Journal of the American Society for Information Science and Technology]] |volume=64 |issue=11 |year=2013 |pages=2332–39 |doi=10.1002/asi.22976|last2=Romero |arxiv=1202.3119 }}</ref> and if based on [[Google Scholar]] output, then even computer-generated documents can be used for that purpose, e.g. using [[SCIgen]].<ref name=ike-antkare>{{cite report |author=Labbé, Cyril |title=Ike Antkare one of the great stars in the scientific firmament |date=2010 |work=Laboratoire d'Informatique de Grenoble RR-LIG-2008 (technical report) |publisher=[[Joseph Fourier University]] |url=http://rr.liglab.fr/research_report/RR-LIG-008.pdf |format=PDF}}</ref> |
||
* The ''h''-index does not provide a significantly more accurate measure of impact than the total number of citations for a given scholar. In particular, by modeling the distribution of citations among papers as a random [[integer partition]] and the ''h''-index as the [[Durfee square]] of the partition, Yong<ref>Alexander Yong, [http://www.ams.org/notices/201409/rnoti-p1040.pdf Critique of Hirsch’s Citation Index: A Combinatorial Fermi Problem], [[Notices of the American Mathematical Society]], vol. 61 (2014), no. 11, pp. 1040–50</ref> arrived at the formula <math>h\approx 0.54\sqrt N</math>, where ''N'' is the total number of citations, which, for mathematics members of the National Academy of Sciences, turns out to provide an accurate (with errors typically within 10–20 percent) approximation of ''h''-index in most cases. |
* The ''h''-index does not provide a significantly more accurate measure of impact than the total number of citations for a given scholar. In particular, by modeling the distribution of citations among papers as a random [[integer partition]] and the ''h''-index as the [[Durfee square]] of the partition, Yong<ref>Alexander Yong, [http://www.ams.org/notices/201409/rnoti-p1040.pdf Critique of Hirsch’s Citation Index: A Combinatorial Fermi Problem], [[Notices of the American Mathematical Society]], vol. 61 (2014), no. 11, pp. 1040–50</ref> arrived at the formula <math>h\approx 0.54\sqrt N</math>, where ''N'' is the total number of citations, which, for mathematics members of the National Academy of Sciences, turns out to provide an accurate (with errors typically within 10–20 percent) approximation of ''h''-index in most cases. |
||
== Alternatives and modifications == |
== Alternatives and modifications == |
||
{{further|Author-level metrics}} |
{{further|Author-level metrics}} |
||
Various proposals to modify the ''h''-index in order to emphasize different features have been made.<ref name=BatistaEtal2006>{{Cite journal |author=Batista P. D. |title=Is it possible to compare researchers with different scientific interests? |journal=[[Scientometrics (journal)|Scientometrics]] |volume=68 |issue=1 |year=2006 |pages=179–89 |doi=10.1007/s11192-006-0090-4 |display-authors=1 |last2=Campiteli |first2=Mônica G. |last3=Kinouchi |first3=Osame }}</ref><ref>{{cite journal |last=Sidiropoulos |first=Antonis |authorlink= |author2=Katsaros, Dimitrios |author3=Manolopoulos, Yannis |year=2007 |title=Generalized Hirsch h-index for disclosing latent facts in citation networks |journal=Scientometrics |volume=72 |issue=2 |pages=253–80 |doi=10.1007/s11192-007-1722-z |url= |accessdate= |quote=}}</ref><ref>{{Cite journal |url=http://www.bmj.com/rapid-response/2011/10/31/v-index-fairer-index-quantify-individual-s-research-output-capacity |title=V-index: A fairer index to quantify an individual's research output capacity |author=Jayant S Vaidya |journal=BMJ |date=December 2005 |volume=331 |issue=7528 |pages= |
Various proposals to modify the ''h''-index in order to emphasize different features have been made.<ref name=BatistaEtal2006>{{Cite journal |author=Batista P. D. |title=Is it possible to compare researchers with different scientific interests? |journal=[[Scientometrics (journal)|Scientometrics]] |volume=68 |issue=1 |year=2006 |pages=179–89 |doi=10.1007/s11192-006-0090-4 |display-authors=1 |last2=Campiteli |first2=Mônica G. |last3=Kinouchi |first3=Osame }}</ref><ref>{{cite journal |last=Sidiropoulos |first=Antonis |authorlink= |author2=Katsaros, Dimitrios |author3=Manolopoulos, Yannis |year=2007 |title=Generalized Hirsch h-index for disclosing latent facts in citation networks |journal=Scientometrics |volume=72 |issue=2 |pages=253–80 |doi=10.1007/s11192-007-1722-z |url= |accessdate= |quote=|citeseerx=10.1.1.76.3617 }}</ref><ref>{{Cite journal |url=http://www.bmj.com/rapid-response/2011/10/31/v-index-fairer-index-quantify-individual-s-research-output-capacity |title=V-index: A fairer index to quantify an individual's research output capacity |author=Jayant S Vaidya |journal=BMJ |date=December 2005 |volume=331 |issue=7528 |pages=1339–c–40–c |doi=10.1136/bmj.331.7528.1339-c|pmid=16322034 |pmc=1298903 }}</ref><ref>Katsaros D., Sidiropoulos A., Manolopous Y., (2007), [http://sunsite.informatik.rwth-aachen.de/Publications/CEUR-WS//Vol-245/paper3.pdf Age Decaying H-Index for Social Network of Citations] in Proceedings of [http://ceur-ws.org/Vol-245/ Workshop on Social Aspects of the Web Poznan, Poland, April 27, 2007]</ref><ref>{{cite journal |last=Anderson |first=T.R. |author2=Hankin, R.K.S and Killworth, P.D. |year=2008 |title=Beyond the Durfee square: Enhancing the h-index to score total publication output |journal=Scientometrics |volume=76 |issue=3 |pages=577–88 |doi=10.1007/s11192-007-2071-2 }}</ref><ref>{{cite journal |last=Baldock |first=C. |author2=Ma, R.M.S and Orton, C.G. |year=2009 |title=The h index is the best measure of a scientist's research productivity |journal=Medical Physics |volume=36 |issue=4 |pages=1043–45 |doi=10.1118/1.3089421 |pmid=19472608 |bibcode=2009MedPh..36.1043B |last3=Orton |first3=Colin G. }}</ref> As the variants have proliferated, comparative studies have become possible showing that most proposals are highly correlated with the original h-index,<ref>{{cite journal |last1=Bornmann |first1=L. |display-authors=etal |year=2011 |title=A multilevel meta-analysis of studies reporting correlations between the h-index and 37 different h-index variants | url = |journal=Journal of Informetrics |volume=5 |issue=3 |pages=346–59 |doi=10.1016/j.joi.2011.01.006}}</ref> although alternative indexes may be important to decide between comparable CVs, as often the case in evaluation processes. |
||
* An individual ''h''-index normalized by the number of author has been proposed: <math>h_I = h^2/N_a^{(T)}</math>, with <math>N_a^{(T)}</math> being the number of authors considered in the <math>h</math> papers.<ref name=BatistaEtal2006/> It was found that the distribution of the ''h''-index, although it depends on the field, can be normalized by a simple rescaling factor. For example, assuming as standard the ''h''s for biology, the distribution of ''h'' for mathematics collapse with it if this ''h'' is multiplied by three, that is, a mathematician with ''h'' = 3 is equivalent to a biologist with ''h'' = 9. This method has not been readily adopted, perhaps because of its complexity. It might be simpler to divide citation counts by the number of authors before ordering the papers and obtaining the ''h''-index, as originally suggested by Hirsch. |
* An individual ''h''-index normalized by the number of author has been proposed: <math>h_I = h^2/N_a^{(T)}</math>, with <math>N_a^{(T)}</math> being the number of authors considered in the <math>h</math> papers.<ref name=BatistaEtal2006/> It was found that the distribution of the ''h''-index, although it depends on the field, can be normalized by a simple rescaling factor. For example, assuming as standard the ''h''s for biology, the distribution of ''h'' for mathematics collapse with it if this ''h'' is multiplied by three, that is, a mathematician with ''h'' = 3 is equivalent to a biologist with ''h'' = 9. This method has not been readily adopted, perhaps because of its complexity. It might be simpler to divide citation counts by the number of authors before ordering the papers and obtaining the ''h''-index, as originally suggested by Hirsch. |
||
* The ''m''-index is defined as ''h''/''n'', where ''n'' is the number of years since the first published paper of the scientist;<ref name=Hirsch2005/> also called ''m''-quotient.<ref>{{cite web |url=http://www.harzing.com/pop_hindex.htm |title=Reflections on the h-index |author=Anne-Wil Harzing |date=2008-04-23 | |
* The ''m''-index is defined as ''h''/''n'', where ''n'' is the number of years since the first published paper of the scientist;<ref name=Hirsch2005/> also called ''m''-quotient.<ref>{{cite web |url=http://www.harzing.com/pop_hindex.htm |title=Reflections on the h-index |author=Anne-Wil Harzing |date=2008-04-23 |website= |accessdate=2013-07-18}}</ref><ref>{{cite journal |pmid=21507617 |year=2011 |author1=von Bohlen und Halbach O |title=How to judge a book by its cover? How useful are bibliometric indices for the evaluation of "scientific quality" or "scientific productivity"? |volume=193 |issue=3 |pages=191–96 |doi=10.1016/j.aanat.2011.03.011 |journal=[[Annals of Anatomy]]}}</ref> |
||
* There are a number of models proposed to incorporate the relative contribution of each author to a paper, for instance by accounting for the rank in the sequence of authors.<ref>{{cite journal | last1 = Tscharntke | first1 = T. | last2 = Hochberg | first2 = M. E. | last3 = Rand | first3 = T. A. | last4 = Resh | first4 = V. H. | last5 = Krauss | first5 = J. | year = 2007 | title = Author Sequence and Credit for Contributions in Multiauthored Publications | url = | journal = PLoS Biology | volume = 5 | issue = 1| page = e18 | doi = 10.1371/journal.pbio.0050018 |pmc=1769438 | pmid=17227141}}</ref> |
* There are a number of models proposed to incorporate the relative contribution of each author to a paper, for instance by accounting for the rank in the sequence of authors.<ref>{{cite journal | last1 = Tscharntke | first1 = T. | last2 = Hochberg | first2 = M. E. | last3 = Rand | first3 = T. A. | last4 = Resh | first4 = V. H. | last5 = Krauss | first5 = J. | year = 2007 | title = Author Sequence and Credit for Contributions in Multiauthored Publications | url = | journal = PLoS Biology | volume = 5 | issue = 1| page = e18 | doi = 10.1371/journal.pbio.0050018 |pmc=1769438 | pmid=17227141}}</ref> |
||
* A generalization of the ''h''-index and some other indices that gives additional information about the shape of the author's citation function (heavy-tailed, flat/peaked, etc.) has been proposed.<ref>{{cite journal |last=Gągolewski |first=M. |author2=Grzegorzewski, P. |year=2009 |title=A geometric approach to the construction of scientific impact indices |journal=Scientometrics |volume=81 |issue=3 |pages=617–34 |doi=10.1007/s11192-008-2253-y }}</ref> |
* A generalization of the ''h''-index and some other indices that gives additional information about the shape of the author's citation function (heavy-tailed, flat/peaked, etc.) has been proposed.<ref>{{cite journal |last=Gągolewski |first=M. |author2=Grzegorzewski, P. |year=2009 |title=A geometric approach to the construction of scientific impact indices |journal=Scientometrics |volume=81 |issue=3 |pages=617–34 |doi=10.1007/s11192-008-2253-y }}</ref> |
||
* A successive Hirsch-type-index for institutions has also been devised.<ref>{{cite journal |last=Kosmulski |first=M. |year=2006 |title=I – a bibliometric index |journal=Forum Akademickie |volume=11 |page=31 }}</ref><ref>{{cite journal |last=Prathap |first=G. |year=2006 |title=Hirsch-type indices for ranking institutions' scientific research output |journal=Current Science |volume=91 | issue = 11 |page=1439 }}</ref> A scientific institution has a successive Hirsch-type-index of ''i'' when at least ''i'' researchers from that institution have an ''h''-index of at least ''i''. |
* A successive Hirsch-type-index for institutions has also been devised.<ref>{{cite journal |last=Kosmulski |first=M. |year=2006 |title=I – a bibliometric index |journal=Forum Akademickie |volume=11 |page=31 }}</ref><ref>{{cite journal |last=Prathap |first=G. |year=2006 |title=Hirsch-type indices for ranking institutions' scientific research output |journal=Current Science |volume=91 | issue = 11 |page=1439 }}</ref> A scientific institution has a successive Hirsch-type-index of ''i'' when at least ''i'' researchers from that institution have an ''h''-index of at least ''i''. |
||
* Three additional metrics have been proposed: ''h''<sup>2</sup> lower, ''h''<sup>2</sup> center, and ''h''<sup>2</sup> upper, to give a more accurate representation of the distribution shape. The three ''h''<sup>2</sup> metrics measure the relative area within a scientist's citation distribution in the low impact area, ''h''<sup>2</sup> lower, the area captured by the ''h''-index, ''h''<sup>2</sup> center, and the area from publications with the highest visibility, ''h''<sup>2</sup> upper. Scientists with high ''h''<sup>2</sup> upper percentages are perfectionists, whereas scientists with high ''h''<sup>2</sup> lower percentages are mass producers. As these metrics are percentages, they are intended to give a qualitative description to supplement the quantitative ''h''-index.<ref>{{cite journal |doi=10.1016/j.joi.2010.03.005 |title=The h index research output measurement: Two approaches to enhance its accuracy |year=2010 |last1=Bornmann |first1=Lutz |last2=Mutz |first2=Rüdiger |last3=Daniel |first3=Hans-Dieter |journal=Journal of Informetrics |volume=4 |issue=3 |pages=407–14 }}</ref> |
* Three additional metrics have been proposed: ''h''<sup>2</sup> lower, ''h''<sup>2</sup> center, and ''h''<sup>2</sup> upper, to give a more accurate representation of the distribution shape. The three ''h''<sup>2</sup> metrics measure the relative area within a scientist's citation distribution in the low impact area, ''h''<sup>2</sup> lower, the area captured by the ''h''-index, ''h''<sup>2</sup> center, and the area from publications with the highest visibility, ''h''<sup>2</sup> upper. Scientists with high ''h''<sup>2</sup> upper percentages are perfectionists, whereas scientists with high ''h''<sup>2</sup> lower percentages are mass producers. As these metrics are percentages, they are intended to give a qualitative description to supplement the quantitative ''h''-index.<ref>{{cite journal |doi=10.1016/j.joi.2010.03.005 |title=The h index research output measurement: Two approaches to enhance its accuracy |year=2010 |last1=Bornmann |first1=Lutz |last2=Mutz |first2=Rüdiger |last3=Daniel |first3=Hans-Dieter |journal=Journal of Informetrics |volume=4 |issue=3 |pages=407–14 }}</ref> |
||
* The [[g-index]] can be seen as the ''h''-index for an averaged citations count.<ref>{{cite journal |doi=10.1007/s11192-006-0144-7 |title=Theory and practise of the g-index |year=2013 |last1=Egghe |first1=Leo |journal=Scientometrics |volume=69 |pages=131–52}}</ref> |
* The [[g-index]] can be seen as the ''h''-index for an averaged citations count.<ref>{{cite journal |doi=10.1007/s11192-006-0144-7 |title=Theory and practise of the g-index |year=2013 |last1=Egghe |first1=Leo |journal=Scientometrics |volume=69 |pages=131–52|hdl=1942/981 }}</ref> |
||
* It has been argued that "For an individual researcher, a measure such as [[Erdős number]] captures the structural properties of network whereas the ''h''-index captures the citation impact of the publications. One can be easily convinced that ranking in coauthorship networks should take into account both measures to generate a realistic and acceptable ranking." Several author ranking systems such as [[eigenfactor]] (based on [[eigenvector centrality]]) have been proposed already, for instance the Phys Author Rank Algorithm.<ref>{{cite web |author1=Kashyap Dixit |author2=S Kameshwaran |author3=Sameep Mehta |author4=Vinayaka Pandit |author5=N Viswanadham |url=http://domino.research.ibm.com/library/cyberdig.nsf/papers/2B600A90C54E51B18525755800283D37/$File/RR_ranking.pdf |title=Towards simultaneously exploiting structure and outcomes in interaction networks for node ranking | |
* It has been argued that "For an individual researcher, a measure such as [[Erdős number]] captures the structural properties of network whereas the ''h''-index captures the citation impact of the publications. One can be easily convinced that ranking in coauthorship networks should take into account both measures to generate a realistic and acceptable ranking." Several author ranking systems such as [[eigenfactor]] (based on [[eigenvector centrality]]) have been proposed already, for instance the Phys Author Rank Algorithm.<ref>{{cite web |author1=Kashyap Dixit |author2=S Kameshwaran |author3=Sameep Mehta |author4=Vinayaka Pandit |author5=N Viswanadham |url=http://domino.research.ibm.com/library/cyberdig.nsf/papers/2B600A90C54E51B18525755800283D37/$File/RR_ranking.pdf |title=Towards simultaneously exploiting structure and outcomes in interaction networks for node ranking |website=IBM Research Report R109002 |date=February 2009}}; see also {{cite conference |doi=10.1145/1871437.1871470 |title=Outcome aware ranking in interaction networks |booktitle=Proceedings of the 19th ACM international conference on Information and knowledge management – CIKM '10 |year=2010 |last1=Kameshwaran |first1=Sampath |last2=Pandit |first2=Vinayaka |last3=Mehta |first3=Sameep |last4=Viswanadham |first4=Nukala |last5=Dixit |first5=Kashyap |isbn=9781450300995 |page=229}}</ref> |
||
* The ''c''-index accounts not only for the citations but for the quality of the citations in terms of the collaboration distance between citing and cited authors. A scientist has ''c''-index ''n'' if ''n'' of [his/her] ''N'' citations are from authors which are at collaboration distance at least ''n'', and the other (''N'' − ''n'') citations are from authors which are at collaboration distance at most ''n''.<ref>{{cite journal |author1=Bras-Amorós, M. |author2=Domingo-Ferrer, J. |author3=Torra, V | year = 2011 | title = A bibliometric index based on the collaboration distance between cited and citing authors | journal = Journal of Informetrics | volume = 5 | issue = 2 | pages = 248–64 | doi = 10.1016/j.joi.2010.11.001 | url = http://www.sciencedirect.com/science/article/pii/S1751157710000970 | format = | accessdate = }}</ref> |
* The ''c''-index accounts not only for the citations but for the quality of the citations in terms of the collaboration distance between citing and cited authors. A scientist has ''c''-index ''n'' if ''n'' of [his/her] ''N'' citations are from authors which are at collaboration distance at least ''n'', and the other (''N'' − ''n'') citations are from authors which are at collaboration distance at most ''n''.<ref>{{cite journal |author1=Bras-Amorós, M. |author2=Domingo-Ferrer, J. |author3=Torra, V | year = 2011 | title = A bibliometric index based on the collaboration distance between cited and citing authors | journal = Journal of Informetrics | volume = 5 | issue = 2 | pages = 248–64 | doi = 10.1016/j.joi.2010.11.001 | url = http://www.sciencedirect.com/science/article/pii/S1751157710000970 | format = | accessdate = |hdl=10261/138172 }}</ref> |
||
* An s-index, accounting for the non-entropic distribution of citations, has been proposed and it has been shown to be in a very good correlation with ''h''.<ref>{{cite journal |last1=Silagadze |first1=Z. K. |title=Citation entropy and research impact estimation |year= |
* An s-index, accounting for the non-entropic distribution of citations, has been proposed and it has been shown to be in a very good correlation with ''h''.<ref>{{cite journal |last1=Silagadze |first1=Z. K. |title=Citation entropy and research impact estimation |year=2325 |pages=2325–33 |volume=41 |issue=2010 |journal=Acta Phys. Pol. B |arxiv=0905.1039 |bibcode=2009arXiv0905.1039S}}</ref> |
||
* The ''e''-index, the square root of surplus citations for the ''h''-set beyond ''h''<sup>2</sup>, complements the ''h''-index for ignored citations, and therefore is especially useful for highly cited scientists and for comparing those with the same ''h''-index (iso-h-index group).<ref>{{cite journal |doi=10.1371/journal.pone.0005429 |title=The e-Index, Complementing the h-Index for Excess Citations |year=2009 |editor1-last=Joly |editor1-first=Etienne |last1=Zhang |first1=Chun-Ting |journal=PLoS ONE |volume=4 |issue=5 |page=e5429 |pmid=19415119 |pmc=2673580|bibcode=2009PLoSO...4.5429Z }}</ref><ref>{{cite journal |doi=10.1016/j.bbrc.2009.07.091 |title=Citation analysis: Maintenance of h-index and use of e-index |year=2009 |last1=Dodson |first1=M.V. |journal=Biochemical and Biophysical Research Communications |volume=387 |issue=4 |pages=625–26 |pmid=19632203}}</ref> |
* The ''e''-index, the square root of surplus citations for the ''h''-set beyond ''h''<sup>2</sup>, complements the ''h''-index for ignored citations, and therefore is especially useful for highly cited scientists and for comparing those with the same ''h''-index (iso-h-index group).<ref>{{cite journal |doi=10.1371/journal.pone.0005429 |title=The e-Index, Complementing the h-Index for Excess Citations |year=2009 |editor1-last=Joly |editor1-first=Etienne |last1=Zhang |first1=Chun-Ting |journal=PLoS ONE |volume=4 |issue=5 |page=e5429 |pmid=19415119 |pmc=2673580|bibcode=2009PLoSO...4.5429Z }}</ref><ref>{{cite journal |doi=10.1016/j.bbrc.2009.07.091 |title=Citation analysis: Maintenance of h-index and use of e-index |year=2009 |last1=Dodson |first1=M.V. |journal=Biochemical and Biophysical Research Communications |volume=387 |issue=4 |pages=625–26 |pmid=19632203}}</ref> |
||
* Because the ''h''-index was never meant to measure future publication success, recently, a group of researchers has investigated the features that are most predictive of future ''h''-index. It is possible to try the predictions using an online tool.<ref>{{cite journal |doi=10.1038/489201a |title=Future impact: Predicting scientific success |year=2012 |last1=Acuna |first1=Daniel E. |last2=Allesina |first2=Stefano |last3=Kording |first3=Konrad P. |authorlink3=Konrad Kording|journal=Nature |volume=489 |issue=7415 |pages=201–02 |pmid=22972278 |pmc=3770471|bibcode=2012Natur.489..201A }}</ref> However, later work has shown that since ''h''-index is a cumulative measure, it contains intrinsic auto-correlation that led to significant overestimation of its predictability. Thus, the true predictability of future ''h''-index is much lower compared to what has been claimed before.<ref>{{cite journal |doi=10.1038/srep03052 |title=On the Predictability of Future Impact in Science|year=2013 |last1=Penner |first1=Orion |last2=Pan |first2=Raj K. |last3=Petersen | first3=Alexander M. |last4=Kaski | first4=Kimmo | last5= Fortunato | first5=Santo |journal=Scientific Reports |volume=3 |page=3052 |pmid=24165898 |pmc=3810665|bibcode=2013NatSR...3E3052P|arxiv=1306.0114 }}</ref> |
* Because the ''h''-index was never meant to measure future publication success, recently, a group of researchers has investigated the features that are most predictive of future ''h''-index. It is possible to try the predictions using an online tool.<ref>{{cite journal |doi=10.1038/489201a |title=Future impact: Predicting scientific success |year=2012 |last1=Acuna |first1=Daniel E. |last2=Allesina |first2=Stefano |last3=Kording |first3=Konrad P. |authorlink3=Konrad Kording|journal=Nature |volume=489 |issue=7415 |pages=201–02 |pmid=22972278 |pmc=3770471|bibcode=2012Natur.489..201A }}</ref> However, later work has shown that since ''h''-index is a cumulative measure, it contains intrinsic auto-correlation that led to significant overestimation of its predictability. Thus, the true predictability of future ''h''-index is much lower compared to what has been claimed before.<ref>{{cite journal |doi=10.1038/srep03052 |title=On the Predictability of Future Impact in Science|year=2013 |last1=Penner |first1=Orion |last2=Pan |first2=Raj K. |last3=Petersen | first3=Alexander M. |last4=Kaski | first4=Kimmo | last5= Fortunato | first5=Santo |journal=Scientific Reports |volume=3 |issue=3052|page=3052 |pmid=24165898 |pmc=3810665|bibcode=2013NatSR...3E3052P|arxiv=1306.0114 }}</ref> |
||
* The h-index has been applied to Internet Media, such as [[YouTube]] channels. The h-index is defined as the number of videos with ≥ h × 10<sup>5</sup> views. When compared with a video creator's total view count, the h-index and g-index better capture both productivity and impact in a single metric.<ref>{{cite journal |doi=10.1002/asi.22936 |arxiv=1303.0766 |title=Bibliometrics for Internet media: Applying the h-index to YouTube |year=2013 |last1=Hovden |first1=R. |journal=Journal of the American Society for Information Science and Technology |volume= 64|issue= 11|pages=2326–31 }}</ref> |
* The h-index has been applied to Internet Media, such as [[YouTube]] channels. The h-index is defined as the number of videos with ≥ h × 10<sup>5</sup> views. When compared with a video creator's total view count, the h-index and g-index better capture both productivity and impact in a single metric.<ref>{{cite journal |doi=10.1002/asi.22936 |arxiv=1303.0766 |title=Bibliometrics for Internet media: Applying the h-index to YouTube |year=2013 |last1=Hovden |first1=R. |journal=Journal of the American Society for Information Science and Technology |volume= 64|issue= 11|pages=2326–31 }}</ref> |
||
* {{anchor|i10-index}}The i10-index indicates the number of academic publications an author has written that have at least ten citations from others. It was introduced in July 2011 by [[Google]] as part of their work on [[Google Scholar]].<ref>Connor, James; Google Scholar Blog. [http://googlescholar.blogspot.com/2011/11/google-scholar-citations-open-to-all.html "Google Scholar Citations Open To All"], Google, 16 November 2011, retrieved 24 November 2011</ref> |
* {{anchor|i10-index}}The i10-index indicates the number of academic publications an author has written that have at least ten citations from others. It was introduced in July 2011 by [[Google]] as part of their work on [[Google Scholar]].<ref>Connor, James; Google Scholar Blog. [http://googlescholar.blogspot.com/2011/11/google-scholar-citations-open-to-all.html "Google Scholar Citations Open To All"], Google, 16 November 2011, retrieved 24 November 2011</ref> |
||
* The h-index has been shown to have a strong discipline bias. However, a simple normalization <math>h/\langle h \rangle_d</math> by the average ''h'' of scholars in a discipline ''d'' is an effective way to mitigate this bias, obtaining a universal impact metric that allows comparison of scholars across different disciplines.<ref>{{cite journal |doi= 10.1016/j.joi.2013.09.002 |arxiv=1305.6339 |title= Universality of scholarly impact metrics |year=2013 |last1=Kaur |first1=Jasleen |last2=Radicchi |first2=Filippo |last3=Menczer |first3=Filippo |journal=Journal of Informetrics |volume= 7|issue= 4|pages= 924–32 }}</ref> Of course this method does not deal with academic age bias. |
* The h-index has been shown to have a strong discipline bias. However, a simple normalization <math>h/\langle h \rangle_d</math> by the average ''h'' of scholars in a discipline ''d'' is an effective way to mitigate this bias, obtaining a universal impact metric that allows comparison of scholars across different disciplines.<ref>{{cite journal |doi= 10.1016/j.joi.2013.09.002 |arxiv=1305.6339 |title= Universality of scholarly impact metrics |year=2013 |last1=Kaur |first1=Jasleen |last2=Radicchi |first2=Filippo |last3=Menczer |first3=Filippo |journal=Journal of Informetrics |volume= 7|issue= 4|pages= 924–32 }}</ref> Of course this method does not deal with academic age bias. |
||
* The h-index can be timed to analyze its evolution during one's career, employing different time windows.<ref>{{cite journal|last=Schreiber|first=Michael|date=2015|title=Restricting the h-index to a publication and citation time window: A case study of a timed Hirsch index|url=|journal=Journal of Informetrics|volume=9|pages=150–55|doi=10.1016/j.joi.2014.12.005|via=}}</ref> |
* The h-index can be timed to analyze its evolution during one's career, employing different time windows.<ref>{{cite journal|last=Schreiber|first=Michael|date=2015|title=Restricting the h-index to a publication and citation time window: A case study of a timed Hirsch index|url=|journal=Journal of Informetrics|volume=9|pages=150–55|doi=10.1016/j.joi.2014.12.005|via=|arxiv=1412.5050}}</ref> |
||
* The o-index corresponds to the [[geometric mean]] of the h-index and the most cited paper of a researcher.<ref name="DM2015">{{Cite journal |last = Dorogovtsev |first = S.N. |author2 = Mendes, J.F.F. |authorlink = Jose Fernando Ferreira Mendes |title = Ranking Scientists |journal = [[Nature Physics]] |volume = 11 |pages = 882–84 |year=2015 |doi = 10.1038/nphys3533 |arxiv = 1511.01545|bibcode = 2015NatPh..11..882D }}</ref> |
* The o-index corresponds to the [[geometric mean]] of the h-index and the most cited paper of a researcher.<ref name="DM2015">{{Cite journal |last = Dorogovtsev |first = S.N. |author2 = Mendes, J.F.F. |authorlink = Jose Fernando Ferreira Mendes |title = Ranking Scientists |journal = [[Nature Physics]] |volume = 11 |issue = 11 |pages = 882–84 |year=2015 |doi = 10.1038/nphys3533 |arxiv = 1511.01545|bibcode = 2015NatPh..11..882D }}</ref> |
||
== See also == |
== See also == |
||
Line 98: | Line 98: | ||
* {{Cite journal |author1=Lehmann, S. |author2=Jackson, A. D. |author3=Lautrup, B. E. |title=Measures for measures |journal=Nature |volume=444 |issue=7122 |pages=1003–04 |year=2006 |pmid=17183295 |doi=10.1038/4441003a |bibcode=2006Natur.444.1003L }} |
* {{Cite journal |author1=Lehmann, S. |author2=Jackson, A. D. |author3=Lautrup, B. E. |title=Measures for measures |journal=Nature |volume=444 |issue=7122 |pages=1003–04 |year=2006 |pmid=17183295 |doi=10.1038/4441003a |bibcode=2006Natur.444.1003L }} |
||
* {{Cite journal |author1=Panaretos, J. |author2=Malesios, C. |title=Assessing Scientific Research Performance and Impact with Single Indices |journal=Scientometrics |volume=81 |issue=3 |pages=635–70 |year=2009 |doi=10.1007/s11192-008-2174-9|arxiv=0812.4542 }} |
* {{Cite journal |author1=Panaretos, J. |author2=Malesios, C. |title=Assessing Scientific Research Performance and Impact with Single Indices |journal=Scientometrics |volume=81 |issue=3 |pages=635–70 |year=2009 |doi=10.1007/s11192-008-2174-9|arxiv=0812.4542 }} |
||
* {{Cite journal |title=Statistical Regularities in the Rank-Citation Profile of Scientists |journal=Scientific Reports |volume=181 |issue= |pages=1–7 |year=2011 |doi=10.1038/srep00181 |url=http://www.nature.com/srep/2011/111205/srep00181/full/srep00181.html |last2=Stanley |first2=H. Eugene |last3=Succi |first3=Sauro |first1=A. M. |last1=Petersen|bibcode=2011NatSR...1E.181P |arxiv=1103.2719 }} |
* {{Cite journal |title=Statistical Regularities in the Rank-Citation Profile of Scientists |journal=Scientific Reports |volume=181 |issue= 181|pages=1–7 |year=2011 |doi=10.1038/srep00181 |pmid=22355696 |pmc=3240955 |url=http://www.nature.com/srep/2011/111205/srep00181/full/srep00181.html |last2=Stanley |first2=H. Eugene |last3=Succi |first3=Sauro |first1=A. M. |last1=Petersen|bibcode=2011NatSR...1E.181P |arxiv=1103.2719 }} |
||
* {{Cite journal |last=Sidiropoulos |first=Antonis |authorlink= |author2=Katsaros, Dimitrios |author3=Manolopoulos, Yannis |year=2007 |title=Generalized Hirsch h-index for disclosing latent facts in citation networks |journal=Scientometrics |volume=72 |issue=2 |pages=253–80 |doi=10.1007/s11192-007-1722-z |url= |accessdate= |quote=}} |
* {{Cite journal |last=Sidiropoulos |first=Antonis |authorlink= |author2=Katsaros, Dimitrios |author3=Manolopoulos, Yannis |year=2007 |title=Generalized Hirsch h-index for disclosing latent facts in citation networks |journal=Scientometrics |volume=72 |issue=2 |pages=253–80 |doi=10.1007/s11192-007-1722-z |url= |accessdate= |quote=|citeseerx=10.1.1.76.3617 }} |
||
* {{Cite journal |last=Soler |first=José M. |year=2007 |title=A rational indicator of scientific creativity |journal=Journal of Informetrics |volume=1 |issue=2 |pages=123–30 |doi=10.1016/j.joi.2006.10.004 |url= |accessdate= |quote=|arxiv=physics/0608006 }} |
* {{Cite journal |last=Soler |first=José M. |year=2007 |title=A rational indicator of scientific creativity |journal=Journal of Informetrics |volume=1 |issue=2 |pages=123–30 |doi=10.1016/j.joi.2006.10.004 |url= |accessdate= |quote=|arxiv=physics/0608006 }} |
||
* {{Cite journal |last=Symonds |first=M. R. |title=Gender differences in publication output: towards an unbiased metric of research performance |journal=PLoS ONE |volume=1 |issue= 1|page=e127 |year=2006 |pmid=17205131 |doi=10.1371/journal.pone.0000127 |pmc=1762413 |bibcode=2006PLoSO...1..127S |display-authors=1 |editor1-last=Tregenza |editor1-first=Tom |last2=Gemmell |first2=Neil J. |last3=Braisher |first3=Tamsin L. |last4=Gorringe |first4=Kylie L. |last5=Elgar |first5=Mark A. }} |
* {{Cite journal |last=Symonds |first=M. R. |title=Gender differences in publication output: towards an unbiased metric of research performance |journal=PLoS ONE |volume=1 |issue= 1|page=e127 |year=2006 |pmid=17205131 |doi=10.1371/journal.pone.0000127 |pmc=1762413 |bibcode=2006PLoSO...1..127S |display-authors=1 |editor1-last=Tregenza |editor1-first=Tom |last2=Gemmell |first2=Neil J. |last3=Braisher |first3=Tamsin L. |last4=Gorringe |first4=Kylie L. |last5=Elgar |first5=Mark A. }} |
Revision as of 22:45, 29 August 2018
The h-index is an author-level metric that attempts to measure both the productivity and citation impact of the publications of a scientist or scholar. The index is based on the set of the scientist's most cited papers and the number of citations that they have received in other publications. The index can also be applied to the productivity and impact of a scholarly journal[1] as well as a group of scientists, such as a department or university or country.[2] The index was suggested in 2005 by Jorge E. Hirsch, a physicist at UCSD, as a tool for determining theoretical physicists' relative quality[3] and is sometimes called the Hirsch index or Hirsch number.
Definition and purpose
The definition of the index is that a scholar with an index of h has published h papers each of which has been cited in other papers at least h times.[4] Thus, the h-index reflects both the number of publications and the number of citations per publication. The index is designed to improve upon simpler measures such as the total number of citations or publications. The index works properly only for comparing scientists working in the same field; citation conventions differ widely among different fields.
Calculation
Formally, if f is the function that corresponds to the number of citations for each publication, we compute the h index as follows. First we order the values of f from the largest to the lowest value. Then, we look for the last position in which f is greater than or equal to the position (we call h this position). For example, if we have a researcher with 5 publications A, B, C, D, and E with 10, 8, 5, 4, and 3 citations, respectively, the h index is equal to 4 because the 4th publication has 4 citations and the 5th has only 3. In contrast, if the same publications have 25, 8, 5, 3, and 3, then the index is 3 because the fourth paper has only 3 citations.
- f(A)=10, f(B)=8, f(C)=5, f(D)=4, f(E)=3 → h-index=4
- f(A)=25, f(B)=8, f(C)=5, f(D)=3, f(E)=3 → h-index=3
If we have the function f ordered in decreasing order from the largest value to the lowest one, we can compute the h index as follows:
- h-index (f) =
The Hirsch index is equivalent to the Eddington number, an earlier metric used for evaluating cyclists. The h-index serves as an alternative to more traditional journal impact factor metrics in the evaluation of the impact of the work of a particular researcher. Because only the most highly cited articles contribute to the h-index, its determination is a simpler process. Hirsch has demonstrated that h has high predictive value for whether a scientist has won honors like National Academy membership or the Nobel Prize. The h-index grows as citations accumulate and thus it depends on the "academic age" of a researcher.
Input data
The h-index can be manually determined using citation databases or using automatic tools. Subscription-based databases such as Scopus and the Web of Science provide automated calculators. Harzing's Publish or Perish program calculates the h-index based on Google Scholar entries. From July 2011 Google have provided an automatically-calculated h-index and i10-index within their own Google Scholar profile.[5] In addition, specific databases, such as the INSPIRE-HEP database can automatically calculate the h-index for researchers working in high energy physics.
Each database is likely to produce a different h for the same scholar, because of different coverage.[6] A detailed study showed that the Web of Science has strong coverage of journal publications, but poor coverage of high impact conferences. Scopus has better coverage of conferences, but poor coverage of publications prior to 1996; Google Scholar has the best coverage of conferences and most journals (though not all), but like Scopus has limited coverage of pre-1990 publications.[7][8] The exclusion of conference proceedings papers is a particular problem for scholars in computer science, where conference proceedings are considered an important part of the literature.[9] Google Scholar has been criticized for producing "phantom citations," including gray literature in its citation counts, and failing to follow the rules of Boolean logic when combining search terms.[10] For example, the Meho and Yang study found that Google Scholar identified 53% more citations than Web of Science and Scopus combined, but noted that because most of the additional citations reported by Google Scholar were from low-impact journals or conference proceedings, they did not significantly alter the relative ranking of the individuals. It has been suggested that in order to deal with the sometimes wide variation in h for a single academic measured across the possible citation databases, one should assume false negatives in the databases are more problematic than false positives and take the maximum h measured for an academic.[11]
Comparing results across fields and career levels
Little systematic investigation has been done on how the h-index behaves over different institutions, nations, times and academic fields/disciplines [citation needed]. Hirsch suggested that, for physicists, a value for h of about 12 might be typical for advancement to tenure (associate professor) at major [US] research universities. A value of about 18 could mean a full professorship, 15–20 could mean a fellowship in the American Physical Society, and 45 or higher could mean membership in the United States National Academy of Sciences.[12]
For the most highly cited scientists in the period 1983–2002, Hirsch identified the top 10 in the life sciences (in order of decreasing h): Solomon H. Snyder, h = 191; David Baltimore, h = 160; Robert C. Gallo, h = 154; Pierre Chambon, h = 153; Bert Vogelstein, h = 151; Salvador Moncada, h = 143; Charles A. Dinarello, h = 138; Tadamitsu Kishimoto, h = 134; Ronald M. Evans, h = 127; and Axel Ullrich, h = 120. Among 36 new inductees in the National Academy of Sciences in biological and biomedical sciences in 2005, the median h-index was 57.[3] However, he points out that values of h will vary between different fields.[13]
Among the 22 scientific disciplines listed in the Thomson Reuters Essential Science Indicators Citation Thresholds [thus excluding non-science academics], physics has the second most citations after space science.[14] During the period January 1, 2000 – February 28, 2010, a physicist had to receive 2073 citations to be among the most cited 1% of physicists in the world.[14] The threshold for space science is the highest (2236 citations), and physics is followed by clinical medicine (1390) and molecular biology & genetics (1229). Most disciplines, such as environment/ecology (390), have fewer scientists, fewer papers, and fewer citations.[14] Therefore, these disciplines have lower citation thresholds in the Essential Science Indicators, with the lowest citation thresholds observed in social sciences (154), computer science (149), and multidisciplinary sciences (147).[14]
Numbers are very different in social science disciplines: The Impact of the Social Sciences team at London School of Economics found that social scientists in the United Kingdom had lower average h-indices. The h-indices for ("full") professors, based on Google Scholar data ranged from 2.8 (in law), through 3.4 (in political science), 3.7 (in sociology), 6.5 (in geography) and 7.6 (in economics). On average across the disciplines, a professor in the social sciences had an h-index about twice that of a lecturer or a senior lecturer, though the difference was the smallest in geography.[15]
Advantages
Hirsch intended the h-index to address the main disadvantages of other bibliometric indicators, such as total number of papers or total number of citations. Total number of papers does not account for the quality of scientific publications, while total number of citations can be disproportionately affected by participation in a single publication of major influence (for instance, methodological papers proposing successful new techniques, methods or approximations, which can generate a large number of citations), or having many publications with few citations each. The h-index is intended to measure simultaneously the quality and quantity of scientific output.
Criticism
There are a number of situations in which h may provide misleading information about a scientist's output:[16] Most of these however are not exclusive to the h-index.
- The h-index does not account for the typical number of citations in different fields. It has been stated that citation behavior in general is affected by field-dependent factors,[17] which may invalidate comparisons not only across disciplines but even within different fields of research of one discipline.[18]
- The h-index discards the information contained in author placement in the authors' list, which in some scientific fields is significant.[19][20]
- The h-index has been found in one study to have slightly less predictive accuracy and precision than the simpler measure of mean citations per paper.[21] However, this finding was contradicted by another study by Hirsch.[22]
- The h-index is a natural number that reduces its discriminatory power. Ruane and Tol therefore propose a rational h-index that interpolates between h and h + 1.[23]
- The h-index can be manipulated through self-citations,[24][25][26] and if based on Google Scholar output, then even computer-generated documents can be used for that purpose, e.g. using SCIgen.[27]
- The h-index does not provide a significantly more accurate measure of impact than the total number of citations for a given scholar. In particular, by modeling the distribution of citations among papers as a random integer partition and the h-index as the Durfee square of the partition, Yong[28] arrived at the formula , where N is the total number of citations, which, for mathematics members of the National Academy of Sciences, turns out to provide an accurate (with errors typically within 10–20 percent) approximation of h-index in most cases.
Alternatives and modifications
Various proposals to modify the h-index in order to emphasize different features have been made.[29][30][31][32][33][34] As the variants have proliferated, comparative studies have become possible showing that most proposals are highly correlated with the original h-index,[35] although alternative indexes may be important to decide between comparable CVs, as often the case in evaluation processes.
- An individual h-index normalized by the number of author has been proposed: , with being the number of authors considered in the papers.[29] It was found that the distribution of the h-index, although it depends on the field, can be normalized by a simple rescaling factor. For example, assuming as standard the hs for biology, the distribution of h for mathematics collapse with it if this h is multiplied by three, that is, a mathematician with h = 3 is equivalent to a biologist with h = 9. This method has not been readily adopted, perhaps because of its complexity. It might be simpler to divide citation counts by the number of authors before ordering the papers and obtaining the h-index, as originally suggested by Hirsch.
- The m-index is defined as h/n, where n is the number of years since the first published paper of the scientist;[3] also called m-quotient.[36][37]
- There are a number of models proposed to incorporate the relative contribution of each author to a paper, for instance by accounting for the rank in the sequence of authors.[38]
- A generalization of the h-index and some other indices that gives additional information about the shape of the author's citation function (heavy-tailed, flat/peaked, etc.) has been proposed.[39]
- A successive Hirsch-type-index for institutions has also been devised.[40][41] A scientific institution has a successive Hirsch-type-index of i when at least i researchers from that institution have an h-index of at least i.
- Three additional metrics have been proposed: h2 lower, h2 center, and h2 upper, to give a more accurate representation of the distribution shape. The three h2 metrics measure the relative area within a scientist's citation distribution in the low impact area, h2 lower, the area captured by the h-index, h2 center, and the area from publications with the highest visibility, h2 upper. Scientists with high h2 upper percentages are perfectionists, whereas scientists with high h2 lower percentages are mass producers. As these metrics are percentages, they are intended to give a qualitative description to supplement the quantitative h-index.[42]
- The g-index can be seen as the h-index for an averaged citations count.[43]
- It has been argued that "For an individual researcher, a measure such as Erdős number captures the structural properties of network whereas the h-index captures the citation impact of the publications. One can be easily convinced that ranking in coauthorship networks should take into account both measures to generate a realistic and acceptable ranking." Several author ranking systems such as eigenfactor (based on eigenvector centrality) have been proposed already, for instance the Phys Author Rank Algorithm.[44]
- The c-index accounts not only for the citations but for the quality of the citations in terms of the collaboration distance between citing and cited authors. A scientist has c-index n if n of [his/her] N citations are from authors which are at collaboration distance at least n, and the other (N − n) citations are from authors which are at collaboration distance at most n.[45]
- An s-index, accounting for the non-entropic distribution of citations, has been proposed and it has been shown to be in a very good correlation with h.[46]
- The e-index, the square root of surplus citations for the h-set beyond h2, complements the h-index for ignored citations, and therefore is especially useful for highly cited scientists and for comparing those with the same h-index (iso-h-index group).[47][48]
- Because the h-index was never meant to measure future publication success, recently, a group of researchers has investigated the features that are most predictive of future h-index. It is possible to try the predictions using an online tool.[49] However, later work has shown that since h-index is a cumulative measure, it contains intrinsic auto-correlation that led to significant overestimation of its predictability. Thus, the true predictability of future h-index is much lower compared to what has been claimed before.[50]
- The h-index has been applied to Internet Media, such as YouTube channels. The h-index is defined as the number of videos with ≥ h × 105 views. When compared with a video creator's total view count, the h-index and g-index better capture both productivity and impact in a single metric.[51]
- The i10-index indicates the number of academic publications an author has written that have at least ten citations from others. It was introduced in July 2011 by Google as part of their work on Google Scholar.[52]
- The h-index has been shown to have a strong discipline bias. However, a simple normalization by the average h of scholars in a discipline d is an effective way to mitigate this bias, obtaining a universal impact metric that allows comparison of scholars across different disciplines.[53] Of course this method does not deal with academic age bias.
- The h-index can be timed to analyze its evolution during one's career, employing different time windows.[54]
- The o-index corresponds to the geometric mean of the h-index and the most cited paper of a researcher.[55]
See also
References
- ^ Suzuki, Helder (2012). "Google Scholar Metrics for Publications". googlescholar.blogspot.com.br.
{{cite web}}
: Cite has empty unknown parameter:|dead-url=
(help) - ^ Jones, T.; Huggett, S.; Kamalski, J. (2011). "Finding a Way Through the Scientific Literature: Indexes and Measures". World Neurosurgery. 76 (1–2): 36–38. doi:10.1016/j.wneu.2011.01.015. PMID 21839937.
- ^ a b c Hirsch, J. E. (15 November 2005). "An index to quantify an individual's scientific research output". PNAS. 102 (46): 16569–72. arXiv:physics/0508025. Bibcode:2005PNAS..10216569H. doi:10.1073/pnas.0507655102. PMC 1283832. PMID 16275915.
{{cite journal}}
: CS1 maint: date and year (link) - ^ McDonald, Kim (8 November 2005). "Physicist Proposes New Way to Rank Scientific Output". PhysOrg. Retrieved 13 May 2010.
- ^ Google Scholar Citations Help, retrieved 2012-09-18.
- ^ Bar-Ilan, J. (2007). "Which h-index? – A comparison of WoS, Scopus and Google Scholar". Scientometrics. 74 (2): 257–71. doi:10.1007/s11192-008-0216-y.
- ^ Meho, L. I.; Yang, K. (2007). "Impact of Data Sources on Citation Counts and Rankings of LIS Faculty: Web of Science vs. Scopus and Google Scholar". Journal of the American Society for Information Science and Technology. 58 (13): 2105–25. doi:10.1002/asi.20677.
- ^ Meho, L. I.; Yang, K (23 December 2006). "A New Era in Citation and Bibliometric Analyses: Web of Science, Scopus, and Google Scholar". arXiv:cs/0612132. (preprint of paper published as 'Impact of data sources on citation counts and rankings of LIS faculty: Web of Science versus Scopus and Google Scholar', in Journal of the American Society for Information Science and Technology, Vol. 58, No. 13, 2007, 2105–25)
- ^ Meyer, Bertrand; Choppy, Christine; Staunstrup, Jørgen; Van Leeuwen, Jan (2009). "Research Evaluation for Computer Science". Communications of the ACM. 52 (4): 31–34. doi:10.1145/1498765.1498780Template:Inconsistent citations
{{cite journal}}
: CS1 maint: postscript (link). - ^ Jacsó, Péter (2006). "Dubious hit counts and cuckoo's eggs". Online Information Review. 30 (2): 188–93. doi:10.1108/14684520610659201.
- ^ Sanderson, Mark (2008). "Revisiting h measured on UK LIS and IR academics". Journal of the American Society for Information Science and Technology. 59 (7): 1184–90. CiteSeerX 10.1.1.474.1990. doi:10.1002/asi.20771.
- ^ Peterson, Ivars (December 2, 2005). "Rating Researchers". Science News. Retrieved 13 May 2010.
- ^ Meho, L.I. (2007) "The rise and rise of citation analysis". Physics World, January 2007, 32–36
- ^ a b c d "Citation Thresholds (Essential Science Indicators)". Science Watch. Thomson Reuters. May 1, 2010. Archived from the original on 5 May 2010. Retrieved 13 May 2010.
- ^ "Impact of Social Sciences – 3: Key Measures of Academic Influence". Impact of Social Sciences.
- ^ Wendl, Michael (2007). "H-index: however ranked, citations need context". Nature. 449 (7161): 403. Bibcode:2007Natur.449..403W. doi:10.1038/449403b. PMID 17898746.
- ^ Bornmann, L.; Daniel, H. D. (2008). "What do citation counts measure? A review of studies on citing behavior". Journal of Documentation. 64 (1): 45–80. doi:10.1108/00220410810844150.
- ^ Anauati, Maria Victoria and Galiani, Sebastian and Gálvez, Ramiro H., Quantifying the Life Cycle of Scholarly Articles Across Fields of Economic Research (November 11, 2014). Available at SSRN: http://ssrn.com/abstract=2523078
- ^ Sekercioglu, Cagan H. (2008). "Quantifying coauthor contributions" (PDF). Science. 322 (5900): 371. doi:10.1126/science.322.5900.371a. PMID 18927373.
- ^ Zhang, Chun-Ting (2009). "A proposal for calculating weighted citations based on author rank". EMBO Reports. 10 (5): 416–17. doi:10.1038/embor.2009.74. PMC 2680883. PMID 19415071.
- ^ Sune Lehmann; Jackson, Andrew D.; Lautrup, Benny E. (2006). "Measures for measures". Nature. 444 (7122): 1003–04. Bibcode:2006Natur.444.1003L. doi:10.1038/4441003a. PMID 17183295.
- ^ Hirsch J. E. (2007). "Does the h-index have predictive power?". PNAS. 104 (49): 19193–98. arXiv:0708.0646. Bibcode:2007PNAS..10419193H. doi:10.1073/pnas.0707962104. PMC 2148266. PMID 18040045.
- ^ Frances Ruane & Richard S. J. Tol; Tol (2008). "Rational (successive) h -indices: An application to economics in the Republic of Ireland". Scientometrics. 75 (2): 395–405. doi:10.1007/s11192-007-1869-7. hdl:1871/31768.
- ^ Gálvez RH (March 2017). "Assessing author self-citation as a mechanism of relevant knowledge diffusion". Scientometrics. 111 (3): 1801–1812. doi:10.1007/s11192-017-2330-1.
- ^ Christoph Bartneck & Servaas Kokkelmans; Kokkelmans (2011). "Detecting h-index manipulation through self-citation analysis". Scientometrics. 87 (1): 85–98. doi:10.1007/s11192-010-0306-5. PMC 3043246. PMID 21472020.
- ^ Emilio Ferrara & Alfonso Romero; Romero (2013). "Scientific impact evaluation and the effect of self-citations: Mitigating the bias by discounting the h-index". Journal of the American Society for Information Science and Technology. 64 (11): 2332–39. arXiv:1202.3119. doi:10.1002/asi.22976.
- ^ Labbé, Cyril (2010). Ike Antkare one of the great stars in the scientific firmament (PDF). Laboratoire d'Informatique de Grenoble RR-LIG-2008 (technical report) (Report). Joseph Fourier University.
- ^ Alexander Yong, Critique of Hirsch’s Citation Index: A Combinatorial Fermi Problem, Notices of the American Mathematical Society, vol. 61 (2014), no. 11, pp. 1040–50
- ^ a b Batista P. D.; et al. (2006). "Is it possible to compare researchers with different scientific interests?". Scientometrics. 68 (1): 179–89. doi:10.1007/s11192-006-0090-4.
- ^ Sidiropoulos, Antonis; Katsaros, Dimitrios; Manolopoulos, Yannis (2007). "Generalized Hirsch h-index for disclosing latent facts in citation networks". Scientometrics. 72 (2): 253–80. CiteSeerX 10.1.1.76.3617. doi:10.1007/s11192-007-1722-z.
- ^ Jayant S Vaidya (December 2005). "V-index: A fairer index to quantify an individual's research output capacity". BMJ. 331 (7528): 1339–c–40–c. doi:10.1136/bmj.331.7528.1339-c. PMC 1298903. PMID 16322034.
- ^ Katsaros D., Sidiropoulos A., Manolopous Y., (2007), Age Decaying H-Index for Social Network of Citations in Proceedings of Workshop on Social Aspects of the Web Poznan, Poland, April 27, 2007
- ^ Anderson, T.R.; Hankin, R.K.S and Killworth, P.D. (2008). "Beyond the Durfee square: Enhancing the h-index to score total publication output". Scientometrics. 76 (3): 577–88. doi:10.1007/s11192-007-2071-2.
{{cite journal}}
: CS1 maint: multiple names: authors list (link) - ^ Baldock, C.; Ma, R.M.S and Orton, C.G.; Orton, Colin G. (2009). "The h index is the best measure of a scientist's research productivity". Medical Physics. 36 (4): 1043–45. Bibcode:2009MedPh..36.1043B. doi:10.1118/1.3089421. PMID 19472608.
{{cite journal}}
: CS1 maint: multiple names: authors list (link) - ^ Bornmann, L.; et al. (2011). "A multilevel meta-analysis of studies reporting correlations between the h-index and 37 different h-index variants". Journal of Informetrics. 5 (3): 346–59. doi:10.1016/j.joi.2011.01.006.
- ^ Anne-Wil Harzing (2008-04-23). "Reflections on the h-index". Retrieved 2013-07-18.
- ^ von Bohlen und Halbach O (2011). "How to judge a book by its cover? How useful are bibliometric indices for the evaluation of "scientific quality" or "scientific productivity"?". Annals of Anatomy. 193 (3): 191–96. doi:10.1016/j.aanat.2011.03.011. PMID 21507617.
- ^ Tscharntke, T.; Hochberg, M. E.; Rand, T. A.; Resh, V. H.; Krauss, J. (2007). "Author Sequence and Credit for Contributions in Multiauthored Publications". PLoS Biology. 5 (1): e18. doi:10.1371/journal.pbio.0050018. PMC 1769438. PMID 17227141.
{{cite journal}}
: CS1 maint: unflagged free DOI (link) - ^ Gągolewski, M.; Grzegorzewski, P. (2009). "A geometric approach to the construction of scientific impact indices". Scientometrics. 81 (3): 617–34. doi:10.1007/s11192-008-2253-y.
- ^ Kosmulski, M. (2006). "I – a bibliometric index". Forum Akademickie. 11: 31.
- ^ Prathap, G. (2006). "Hirsch-type indices for ranking institutions' scientific research output". Current Science. 91 (11): 1439.
- ^ Bornmann, Lutz; Mutz, Rüdiger; Daniel, Hans-Dieter (2010). "The h index research output measurement: Two approaches to enhance its accuracy". Journal of Informetrics. 4 (3): 407–14. doi:10.1016/j.joi.2010.03.005.
- ^ Egghe, Leo (2013). "Theory and practise of the g-index". Scientometrics. 69: 131–52. doi:10.1007/s11192-006-0144-7. hdl:1942/981.
- ^ Kashyap Dixit; S Kameshwaran; Sameep Mehta; Vinayaka Pandit; N Viswanadham (February 2009). "Towards simultaneously exploiting structure and outcomes in interaction networks for node ranking" (PDF). IBM Research Report R109002.; see also Kameshwaran, Sampath; Pandit, Vinayaka; Mehta, Sameep; Viswanadham, Nukala; Dixit, Kashyap (2010). "Outcome aware ranking in interaction networks". Proceedings of the 19th ACM international conference on Information and knowledge management – CIKM '10. p. 229. doi:10.1145/1871437.1871470. ISBN 9781450300995.
{{cite conference}}
: Unknown parameter|booktitle=
ignored (|book-title=
suggested) (help) - ^ Bras-Amorós, M.; Domingo-Ferrer, J.; Torra, V (2011). "A bibliometric index based on the collaboration distance between cited and citing authors". Journal of Informetrics. 5 (2): 248–64. doi:10.1016/j.joi.2010.11.001. hdl:10261/138172.
- ^ Silagadze, Z. K. (2325). "Citation entropy and research impact estimation". Acta Phys. Pol. B. 41 (2010): 2325–33. arXiv:0905.1039. Bibcode:2009arXiv0905.1039S.
{{cite journal}}
: Check date values in:|year=
(help) - ^ Zhang, Chun-Ting (2009). Joly, Etienne (ed.). "The e-Index, Complementing the h-Index for Excess Citations". PLoS ONE. 4 (5): e5429. Bibcode:2009PLoSO...4.5429Z. doi:10.1371/journal.pone.0005429. PMC 2673580. PMID 19415119.
{{cite journal}}
: CS1 maint: unflagged free DOI (link) - ^ Dodson, M.V. (2009). "Citation analysis: Maintenance of h-index and use of e-index". Biochemical and Biophysical Research Communications. 387 (4): 625–26. doi:10.1016/j.bbrc.2009.07.091. PMID 19632203.
- ^ Acuna, Daniel E.; Allesina, Stefano; Kording, Konrad P. (2012). "Future impact: Predicting scientific success". Nature. 489 (7415): 201–02. Bibcode:2012Natur.489..201A. doi:10.1038/489201a. PMC 3770471. PMID 22972278.
- ^ Penner, Orion; Pan, Raj K.; Petersen, Alexander M.; Kaski, Kimmo; Fortunato, Santo (2013). "On the Predictability of Future Impact in Science". Scientific Reports. 3 (3052): 3052. arXiv:1306.0114. Bibcode:2013NatSR...3E3052P. doi:10.1038/srep03052. PMC 3810665. PMID 24165898.
- ^ Hovden, R. (2013). "Bibliometrics for Internet media: Applying the h-index to YouTube". Journal of the American Society for Information Science and Technology. 64 (11): 2326–31. arXiv:1303.0766. doi:10.1002/asi.22936.
- ^ Connor, James; Google Scholar Blog. "Google Scholar Citations Open To All", Google, 16 November 2011, retrieved 24 November 2011
- ^ Kaur, Jasleen; Radicchi, Filippo; Menczer, Filippo (2013). "Universality of scholarly impact metrics". Journal of Informetrics. 7 (4): 924–32. arXiv:1305.6339. doi:10.1016/j.joi.2013.09.002.
- ^ Schreiber, Michael (2015). "Restricting the h-index to a publication and citation time window: A case study of a timed Hirsch index". Journal of Informetrics. 9: 150–55. arXiv:1412.5050. doi:10.1016/j.joi.2014.12.005.
- ^ Dorogovtsev, S.N.; Mendes, J.F.F. (2015). "Ranking Scientists". Nature Physics. 11 (11): 882–84. arXiv:1511.01545. Bibcode:2015NatPh..11..882D. doi:10.1038/nphys3533.
Further reading
- Alonso, S.; Cabrerizo, F. J.; Herrera-Viedma, E.; Herrera, F. (2009). "h-index: A Review Focused in its Variants, Computation and Standardization for Different Scientific Fields". Journal of Informetrics. 3 (4): 273–89. doi:10.1016/j.joi.2009.04.001.
- Ball, Philip (2005). "Index aims for fair ranking of scientists". Nature. 436 (7053): 900. Bibcode:2005Natur.436..900B. doi:10.1038/436900a. PMID 16107806.
- Iglesias, Juan E.; Pecharromán, Carlos. "Scaling the h-index for different scientific ISI fields" (PDF).
{{cite journal}}
: Cite journal requires|journal=
(help) - Kelly, C. D.; Jennions, M. D. (2006). "The h index and career assessment by numbers". Trends Ecol. Evol. 21 (4): 167–70. doi:10.1016/j.tree.2006.01.005. PMID 16701079.
- Lehmann, S.; Jackson, A. D.; Lautrup, B. E. (2006). "Measures for measures". Nature. 444 (7122): 1003–04. Bibcode:2006Natur.444.1003L. doi:10.1038/4441003a. PMID 17183295.
- Panaretos, J.; Malesios, C. (2009). "Assessing Scientific Research Performance and Impact with Single Indices". Scientometrics. 81 (3): 635–70. arXiv:0812.4542. doi:10.1007/s11192-008-2174-9.
- Petersen, A. M.; Stanley, H. Eugene; Succi, Sauro (2011). "Statistical Regularities in the Rank-Citation Profile of Scientists". Scientific Reports. 181 (181): 1–7. arXiv:1103.2719. Bibcode:2011NatSR...1E.181P. doi:10.1038/srep00181. PMC 3240955. PMID 22355696.
- Sidiropoulos, Antonis; Katsaros, Dimitrios; Manolopoulos, Yannis (2007). "Generalized Hirsch h-index for disclosing latent facts in citation networks". Scientometrics. 72 (2): 253–80. CiteSeerX 10.1.1.76.3617. doi:10.1007/s11192-007-1722-z.
- Soler, José M. (2007). "A rational indicator of scientific creativity". Journal of Informetrics. 1 (2): 123–30. arXiv:physics/0608006. doi:10.1016/j.joi.2006.10.004.
- Symonds, M. R.; et al. (2006). Tregenza, Tom (ed.). "Gender differences in publication output: towards an unbiased metric of research performance". PLoS ONE. 1 (1): e127. Bibcode:2006PLoSO...1..127S. doi:10.1371/journal.pone.0000127. PMC 1762413. PMID 17205131.
{{cite journal}}
: CS1 maint: unflagged free DOI (link) - Taber, Douglass F. (2005). "Quantifying Publication Impact". Science. 309 (5744): 2166a. doi:10.1126/science.309.5744.2166a. PMID 16195445.
- Woeginger, Gerhard j. (2008). "An axiomatic characterization of the Hirsch-index". Mathematical Social Sciences. 56 (2): 224–32. doi:10.1016/j.mathsocsci.2008.03.001.