Movatterモバイル変換


[0]ホーム

URL:


Jump to content
WikipediaThe Free Encyclopedia
Search

h-index

From Wikipedia, the free encyclopedia
Measure of a scholar's citation impact
Not to be confused withHerfindahl–Hirschman index.
Part ofa series on
Citation metrics
Author-level
Citation
Journal-level

Theh-index is anauthor-level metric that measures both theproductivity andcitation impact of thepublications, initially used for an individualscientist or scholar. Theh-index correlates with success indicators such as winning theNobel Prize, being accepted for research fellowships and holding positions at top universities.[1] The index is based on the set of the scientist's most cited papers and the number of citations that they have received in other publications. The index has more recently been applied to the productivity and impact of ascholarly journal[2] as well as a group of scientists, such as a department or university or country.[3] The index was suggested in 2005 byJorge E. Hirsch, a physicist atUC San Diego, as a tool for determiningtheoretical physicists' relative quality[4] and is sometimes called theHirsch index orHirsch number.

Hirsch intended theh-index to address the main disadvantages of other bibliometric indicators. The total number of papers metric does not account for the quality of scientific publications. The total number of citations metric, on the other hand, can be heavily affected by participation in a single publication of major influence (for instance, methodological papers proposing successful new techniques, methods or approximations, which can generate a large number of citations). The index works best when comparing scholars working in the same field, since citation conventions differ widely among different fields.[5]

Theh-index is intended to measure simultaneously the quality and quantity of scientific output. TheKendall's correlation ofh-index with scientific awards in physics was found at 34 percent in 2010 and zero percent in 2019.[6]

Calculation

[edit]

Theh-index is defined as the maximum value ofh such that the given author/journal has published at leasth papers that have each been cited at leasth times.[4][7]

h-index from a plot of numbers of citations for an author's numbered papers (arranged in decreasing order)

Theh-index is the largesth such thath articles have at leasth citations each. For example, if an author has five publications, with 9, 7, 6, 2, and 1 citations (ordered from greatest to least), then the author'sh-index is 3, because the author has three publications with 3 or more citations. However, the author doesnot have four publications with 4 or more citations.

Clearly, an author'sh-index can only be as great as their number of publications. For example, an author with only one publication can have a maximumh-index of 1 (if their publication has 1 or more citations). On the other hand, an author with many publications, each with only 1 citation, would also have anh-index of 1.

Formally, iff is the function that corresponds to the number of citations for each publication, we compute theh-index as follows: First we order the values off from the largest to the lowest value. Then, we look for the last position in whichf is greater than or equal to theposition (we callh this position). For example, if we have a researcher with 5 publications A, B, C, D, and E with 10, 8, 5, 4, and 3 citations, respectively, theh-index is equal to 4 because the 4th publication has 4 citations and the 5th has only 3. In contrast, if the same publications have 25, 8, 5, 3, and 3 citations, then the index is 3 (i.e. the 3rd position) because the fourth paper has only 3 citations.

f(A)=10,f(B)=8,f(C)=5,f(D)=4,f(E)=3 →h-index=4
f(A)=25,f(B)=8,f(C)=5,f(D)=3,f(E)=3 →h-index=3

If we have the functionf ordered in decreasing order from the largestvalue to the lowest one, we can compute theh-index as follows:

h-index (f) =max{iN:f(i)i}{\displaystyle \max\{i\in \mathbb {N} :f(i)\geq i\}}

The Hirsch index is analogous to theEddington number, an earlier metric used for evaluating cyclists.[8]h-index is also related toSugeno integral andKy Fan metric.[9] Theh-index serves as an alternative to more traditional journalimpact factor metrics in the evaluation of the impact of the work of a particular researcher. Because only the most highly cited articles contribute to theh-index, its determination is a simpler process. Hirsch has demonstrated thath has high predictive value for whether a scientist has won honors likeNational Academy membership or theNobel Prize. Theh-index grows as citations accumulate and thus it depends on the "academic age" of a researcher.

Input data

[edit]

Theh-index can be manually determined by using citation databases or using automatic tools. Subscription-based databases such asScopus and theWeb of Science provide automated calculators. From July 2011Google has provided an automatically calculatedh-index andi10-index within its ownGoogle Scholar profile.[10] In addition, specific databases, such as theINSPIRE-HEP database can automatically calculate theh-index for researchers working inhigh energy physics.

Each database is likely to produce a differenth for the same scholar, because of different coverage.[11] A detailed study showed that the Web of Science has strong coverage of journal publications, but poor coverage of high impact conferences. Scopus has better coverage of conferences, but poor coverage of publications prior to 1996; Google Scholar has the best coverage of conferences and most journals (though not all), but like Scopus has limited coverage of pre-1990 publications.[12] The exclusion of conference proceedings papers is a particular problem for scholars incomputer science, where conference proceedings are considered an important part of the literature.[13] Google Scholar has been criticized for producing "phantom citations", includinggray literature in its citation counts, and failing to follow the rules ofBoolean logic when combining search terms.[14] For example, the Meho and Yang study found that Google Scholar identified 53% more citations than Web of Science and Scopus combined, but noted that because most of the additional citations reported by Google Scholar were from low-impact journals or conference proceedings, they did not significantly alter the relative ranking of the individuals. It has been suggested that in order to deal with the sometimes wide variation inh for a single academic measured across the possible citation databases, one should assume false negatives in the databases are more problematic than false positives and take the maximumh measured for an academic.[15]

Examples

[edit]

Little systematic investigation has been done on how theh-index behaves over different institutions, nations, times and academic fields.[16] Hirsch suggested that, for physicists, a value forh of about 12 might be typical for advancement to tenure (associate professor) at major [US] research universities. A value of about 18 could mean a full professorship, 15–20 could mean a fellowship in theAmerican Physical Society, and 45 or higher could mean membership in theUnited States National Academy of Sciences.[17] Hirsch estimated that after 20 years a "successful scientist" would have anh-index of 20, an "outstanding scientist" would have anh-index of 40, and a "truly unique" individual would have anh-index of 60.[4]

For the most highly cited scientists in the period 1983–2002, Hirsch identified the top 10 in the life sciences (in order of decreasingh):Solomon H. Snyder,h = 191;David Baltimore,h = 160;Robert C. Gallo,h = 154;Pierre Chambon,h = 153;Bert Vogelstein,h = 151;Salvador Moncada,h = 143;Charles A. Dinarello,h = 138;Tadamitsu Kishimoto,h = 134;Ronald M. Evans,h = 127; andRalph L. Brinster,h = 126. Among 36 new inductees in the National Academy of Sciences in biological and biomedical sciences in 2005, the medianh-index was 57.[4] However, Hirsch noted that values ofh will vary among disparate fields.[4]

Among the 22 scientific disciplines listed in theEssential Science Indicators citation thresholds (thus excludingnon-science academics), physics has the second most citations afterspace science.[18] During the period between January 1, 2000 and February 28, 2010, a physicist had to receive 2073 citations to be among the most cited 1% of physicists in the world.[18] The threshold for space science is the highest (2236 citations), and physics is followed byclinical medicine (1390) andmolecular biology &genetics (1229). Most disciplines, such as environment/ecology (390), have fewer scientists, fewer papers, and fewer citations.[18] Therefore, these disciplines have lower citation thresholds in the Essential Science Indicators, with the lowest citation thresholds observed in social sciences (154),computer science (149), andmultidisciplinary sciences (147).[18]

Numbers are very different in social science disciplines: TheImpact of the Social Sciences team atLondon School of Economics found that social scientists in the United Kingdom had lower averageh-indices. Theh-indices for ("full") professors, based onGoogle Scholar data ranged from 2.8 (in law), through 3.4 (inpolitical science), 3.7 (insociology), 6.5 (in geography) and 7.6 (in economics). On average across the disciplines, a professor in the social sciences had anh-index about twice that of a lecturer or a senior lecturer, though the difference was the smallest in geography.[19]

Criticism

[edit]

There are a number of situations in whichh may provide misleading information about a scientist's output.[20] The correlation betweenh-index and scientific awards dropped significantly since 2010 after the widespread usage ofh-index,[6] followingGoodhart's law. The decrease of correlation is partially attributed to the spread of hyperauthorship with more than 100 coauthors per paper.

Some of the following failures are not exclusive to theh-index but rather shared with otherauthor-level metrics:

  • Theh-index does not account for the number of authors of a paper. In the original paper, Hirsch suggested partitioning citations among co-authors. One such fractional index is known asfractional h-index, which accounts for multiple authors but is not widely available through the use of automatic tools.[6]
  • Theh-index does not account for the different typical number of citations in different fields, e.g. experimental over theoretical. Citation behavior in general is affected by field-dependent factors,[21] which may invalidate comparisons not only across disciplines but even within different fields of research of one discipline.[22]
  • Theh-index discards the information contained in author placement in the authors' list, which in some scientific fields is significant though in others it is not.[23][24]
  • Theh-index is aninteger, which reduces its discriminatory power.Ruane andTol therefore propose arationalh-index that interpolates betweenh andh + 1.[25]

Theh-index has been found in one study to have slightly less predictiveaccuracy and precision than the simpler measure of mean citations per paper.[26] However, this finding was contradicted by another study by Hirsch.[27] Theh-index does not provide a significantly more accurate measure of impact than the total number of citations for a given scholar. In particular, by modeling the distribution of citations among papers as a randominteger partition and theh-index as theDurfee square of the partition, Yong[28] arrived at the formulah0.54N{\displaystyle h\approx 0.54{\sqrt {N}}}, whereN is the total number of citations, which, for mathematics members of the National Academy of Sciences, turns out to provide an accurate (with errors typically within 10–20 percent) approximation ofh-index in most cases.

Susceptibility to manipulation

[edit]

Weaknesses apply to the purely quantitative calculation of scientific or academic output. Like other metrics that count citations, theh-index can be manipulated bycoercive citation, a practice in which an editor of a journal forces authors to add spurious citations to their own articles before the journal will agree to publish it.[29][30] Theh-index can be manipulated through self-citations,[31][32][33][34] and if based onGoogle Scholar output, then even computer-generated documents can be used for that purpose, e.g. usingSCIgen.[35] Theh-index can be also manipulated byhyperauthorship. Recent research shows clearly that the correlation of theh-index with awards that indicate recognition by the scientific community has substantially declined.[36]

Variants

[edit]

Various proposals to modify theh-index in order to emphasize different features have been made.[37][38][39][40][41][42] Many of these variants, such asg-index, are highly correlated with the originalh-index, which has led some researchers to consider them redundant.[43]

One metric which was found not to be highly correlated withh-index and is correlated with scientific awards as of 2019 is thefractional h-index.[6]

A Hirsch-type index for institutions has also been devised.[44][45] A scientific institution has a successive Hirsch-type index ofi when at leasti researchers from that institution have anh-index of at leasti.

Theh-index has been applied to internet media, such asYouTube channels. It is defined as the number of videosh a channel has with more thanh × 105 views. When compared with a video creator's total view count, theh-index andg-index better capture both productivity and impact in a single metric.[46]

More recently, a dual framework distinguishing betweendisruptive andconsolidating scientific impact has led to the development of new h-index variants.[47]

See also

[edit]

References

[edit]
  1. ^Bornmann, Lutz; Daniel, Hans-Dieter (July 2007). "What do we know about the h-index?".Journal of the American Society for Information Science and Technology.58 (9):1381–1385.doi:10.1002/asi.20609.S2CID 31323195.
  2. ^Suzuki, Helder (2012)."Google Scholar Metrics for Publications".googlescholar.blogspot.com.br.
  3. ^Jones, T.; Huggett, S.; Kamalski, J. (2011). "Finding a Way Through the Scientific Literature: Indexes and Measures".World Neurosurgery.76 (1–2):36–38.doi:10.1016/j.wneu.2011.01.015.PMID 21839937.
  4. ^abcdeHirsch, J. E. (15 November 2005)."An index to quantify an individual's scientific research output".PNAS.102 (46):16569–16572.arXiv:physics/0508025.Bibcode:2005PNAS..10216569H.doi:10.1073/pnas.0507655102.PMC 1283832.PMID 16275915.
  5. ^"Impact of Social Sciences – 3: Key Measures of Academic Influence".LSE Impact of Social Sciences Blog (Section 3.2). London School of Economics. 19 November 2010. Retrieved19 April 2020.
  6. ^abcdKoltun, V; Hafner, D. (2021)."The h-index is no longer an effective correlate of scientific reputation".PLOS ONE.16 (6) e0253397.arXiv:2102.03234.Bibcode:2021PLoSO..1653397K.doi:10.1371/journal.pone.0253397.PMC 8238192.PMID 34181681.Our results suggest that the use of the h-index in ranking scientists should be reconsidered, and that fractional allocation measures such as h-frac provide more robust alternativesCompanion webpage
  7. ^McDonald, Kim (8 November 2005)."Physicist Proposes New Way to Rank Scientific Output".PhysOrg. Retrieved13 May 2010.
  8. ^Jeffers, David; Swanson, John (November 2005)."How high is yourE?".Physics World.18 (10): 21.doi:10.1088/2058-7058/18/10/30. Retrieved2022-09-17.
  9. ^Mesiar, Radko; Gagolewski, Marek (December 2016). "H-Index and Other Sugeno Integrals: Some Defects and Their Compensation".IEEE Transactions on Fuzzy Systems.24 (6):1668–1672.Bibcode:2016ITFS...24.1668M.doi:10.1109/TFUZZ.2016.2516579.ISSN 1941-0034.S2CID 1651767.
  10. ^Google Scholar Citations Help, retrieved 2012-09-18.
  11. ^Bar-Ilan, J. (2007). "Whichh-index?: A comparison of WoS, Scopus and Google Scholar".Scientometrics.74 (2):257–271.doi:10.1007/s11192-008-0216-y.S2CID 29641074.
  12. ^
    • Meho, L. I.; Yang, K. (2007). "Impact of Data Sources on Citation Counts and Rankings of LIS Faculty: Web of Science vs. Scopus and Google Scholar".Journal of the American Society for Information Science and Technology.58 (13):2105–2125.doi:10.1002/asi.20677.
    • Yang, K. (23 December 2006). "A New Era in Citation and Bibliometric Analyses: Web of Science, Scopus, and Google Scholar".arXiv:cs/0612132. (preprint of paper published as 'Impact of data sources on citation counts and rankings of LIS faculty: Web of Science versus Scopus and Google Scholar', inJournal of the American Society for Information Science and Technology, Vol.58, No. 13, 2007, 2105–2125)
  13. ^Meyer, Bertrand; Choppy, Christine; Staunstrup, Jørgen;Van Leeuwen, Jan (2009)."Research Evaluation for Computer Science".Communications of the ACM.52 (4):31–34.doi:10.1145/1498765.1498780.S2CID 8625066..
  14. ^Jacsó, Péter (2006). "Dubious hit counts and cuckoo's eggs".Online Information Review.30 (2):188–193.doi:10.1108/14684520610659201.
  15. ^Sanderson, Mark (2008). "Revisitingh measured on UK LIS and IR academics".Journal of the American Society for Information Science and Technology.59 (7):1184–1190.CiteSeerX 10.1.1.474.1990.doi:10.1002/asi.20771.
  16. ^Turaga, Kiran K.; Gamblin, T. Clark (July 2012). "Measuring the Surgical Academic Output of an Institution: The "Institutional" H-Index".Journal of Surgical Education.69 (4):499–503.doi:10.1016/j.jsurg.2012.02.004.PMID 22677589.
  17. ^Peterson, Ivars (December 2, 2005)."Rating Researchers".Science News. Retrieved13 May 2010.
  18. ^abcd"Citation Thresholds (Essential Science Indicators)".Science Watch. Thomson Reuters. May 1, 2010. Archived fromthe original on 5 May 2010. Retrieved13 May 2010.
  19. ^"Impact of Social Sciences – 3: Key Measures of Academic Influence"(PDF).Impact of Social Sciences, LSE.ac.uk. 19 November 2010. Retrieved14 November 2020.
  20. ^Wendl, Michael (2007)."H-index: however ranked, citations need context".Nature.449 (7161): 403.Bibcode:2007Natur.449..403W.doi:10.1038/449403b.PMID 17898746.
  21. ^Bornmann, L.; Daniel, H. D. (2008). "What do citation counts measure? A review of studies on citing behavior".Journal of Documentation.64 (1):45–80.doi:10.1108/00220410810844150.hdl:11858/00-001M-0000-0013-7A94-3.S2CID 17260826.
  22. ^Anauati, Victoria; Galiani, Sebastian; Gálvez, Ramiro H. (2016). "Quantifying the Life Cycle of Scholarly Articles Across Fields of Economic Research".Economic Inquiry.54 (2):1339–1355.doi:10.1111/ecin.12292.hdl:10.1111/ecin.12292.ISSN 1465-7295.S2CID 154806179.
  23. ^Sekercioglu, Cagan H. (2008)."Quantifying coauthor contributions"(PDF).Science.322 (5900): 371.doi:10.1126/science.322.5900.371a.PMID 18927373.S2CID 47571516.
  24. ^Zhang, Chun-Ting (2009)."A proposal for calculating weighted citations based on author rank".EMBO Reports.10 (5):416–417.doi:10.1038/embor.2009.74.PMC 2680883.PMID 19415071.
  25. ^Ruane, F. P.;Tol, R. S. J. (2008)."Rational (successive) H-indices: An application to economics in the Republic of Ireland".Scientometrics.75 (2):395–405.doi:10.1007/s11192-007-1869-7.hdl:1871/31768.S2CID 6541932.
  26. ^Lehmann, Sune; Jackson, Andrew D.; Lautrup, Benny E. (2006). "Measures for measures".Nature.444 (7122):1003–1004.Bibcode:2006Natur.444.1003L.doi:10.1038/4441003a.PMID 17183295.S2CID 3099364.
  27. ^Hirsch, J. E. (2007)."Does theh-index have predictive power?".PNAS.104 (49):19193–19198.arXiv:0708.0646.Bibcode:2007PNAS..10419193H.doi:10.1073/pnas.0707962104.PMC 2148266.PMID 18040045.
  28. ^Yong, Alexander (2014)."Critique of Hirsch's Citation Index: A Combinatorial Fermi Problem"(PDF).Notices of the American Mathematical Society.61 (11):1040–1050.arXiv:1402.4357.doi:10.1090/noti1164.S2CID 119126314.
  29. ^Wilhite, A. W.; Fong, E. A. (2012). "Coercive Citation in Academic Publishing".Science.335 (6068):542–543.Bibcode:2012Sci...335..542W.doi:10.1126/science.1212540.PMID 22301307.S2CID 30073305.
  30. ^Noorden, Richard Van (February 6, 2020)."Highly cited researcher banned from journal board for citation abuse".Nature.578 (7794):200–201.Bibcode:2020Natur.578..200V.doi:10.1038/d41586-020-00335-7.PMID 32047304.
  31. ^Gálvez, Ramiro H. (March 2017). "Assessing author self-citation as a mechanism of relevant knowledge diffusion".Scientometrics.111 (3):1801–1812.doi:10.1007/s11192-017-2330-1.S2CID 6863843.
  32. ^Bartneck, Christoph; Kokkelmans, Servaas (2011)."Detectingh-index manipulation through self-citation analysis".Scientometrics.87 (1):85–98.doi:10.1007/s11192-010-0306-5.PMC 3043246.PMID 21472020.
  33. ^Ferrara, Emilio; Romero, Alfonso (2013). "Scientific impact evaluation and the effect of self-citations: Mitigating the bias by discounting theh-index".Journal of the American Society for Information Science and Technology.64 (11):2332–2339.arXiv:1202.3119.doi:10.1002/asi.22976.S2CID 12693511.
  34. ^
  35. ^Labbé, Cyril (2010).Ike Antkare one of the great stars in the scientific firmament(PDF).Laboratoire d'Informatique de Grenoble RR-LIG-2008 (Technical report).Joseph Fourier University.
  36. ^Koltun, Vladlen; Hafner, David (2021)."The h-index is no longer an effective correlate of scientific reputation".PLOS ONE.16 (6) e0253397.arXiv:2102.03234.Bibcode:2021PLoSO..1653397K.doi:10.1371/journal.pone.0253397.PMC 8238192.PMID 34181681.
  37. ^Batista, P. D.; et al. (2006). "Is it possible to compare researchers with different scientific interests?".Scientometrics.68 (1):179–189.arXiv:physics/0509048.doi:10.1007/s11192-006-0090-4.S2CID 119068816.
  38. ^Dimitrios, Katsaros; Manolopoulos, Yannis (2007). "Generalized Hirschh-index for disclosing latent facts in citation networks".Scientometrics.72 (2):253–280.CiteSeerX 10.1.1.76.3617.doi:10.1007/s11192-007-1722-z.S2CID 14919467.
  39. ^Vaidya, Jayant S. (December 2005)."V-index: A fairer index to quantify an individual's research output capacity".BMJ.331 (7528):1339–1340.doi:10.1136/bmj.331.7528.1339-c.PMC 1298903.PMID 16322034.
  40. ^Katsaros D., Sidiropoulos A., Manolopous Y., (2007),Age DecayingH-Index for Social Network of Citations in Proceedings ofWorkshop on Social Aspects of the Web Poznan, Poland, April 27, 2007
  41. ^Anderson, Thomas R.; Hankin, Robin K. S.; Killworth, Peter D. (12 July 2008)."Beyond the Durfee square: Enhancing the h-index to score total publication output".Scientometrics.76 (3). Springer:577–588.doi:10.1007/s11192-007-2071-2.ISSN 0138-9130.
  42. ^Baldock, Clive; Ma, Ruimin; Orton, Colin G. (5 March 2009). "The h index is the best measure of a scientist's research productivity".Medical Physics.36 (4). Wiley:1043–1045.Bibcode:2009MedPh..36.1043B.doi:10.1118/1.3089421.ISSN 0094-2405.PMID 19472608.
  43. ^Bornmann, L.; et al. (2011). "A multilevel meta-analysis of studies reporting correlations between theh-index and 37 differenth-index variants".Journal of Informetrics.5 (3):346–359.doi:10.1016/j.joi.2011.01.006.
  44. ^Kosmulski, M. (2006). "I – a bibliometric index".Forum Akademickie.11: 31.
  45. ^Prathap, G. (2006). "Hirsch-type indices for ranking institutions' scientific research output".Current Science.91 (11): 1439.
  46. ^Hovden, R. (2013). "Bibliometrics for Internet media: Applying theh-index to YouTube".Journal of the American Society for Information Science and Technology.64 (11):2326–2331.arXiv:1303.0766.doi:10.1002/asi.22936.S2CID 38708903.
  47. ^Yang, Alex J.; et al. (2023). "From consolidation to disruption: A novel way to measure the impact of scientists and identify laureates".Information Processing & Management.60 (5) 103420.doi:10.1016/j.ipm.2023.103420.

Further reading

[edit]

External links

[edit]
Journals
Papers
Grey literature
Other publication types
Impact and ranking
Reform and access
Versioning
Indexes and search engines
Related topics
Lists
Authority control databases: NationalEdit this at Wikidata
Retrieved from "https://en.wikipedia.org/w/index.php?title=H-index&oldid=1322602255"
Categories:
Hidden categories:

[8]ページ先頭

©2009-2025 Movatter.jp