See discussions, stats, and author profiles for this publication at: https://www.researchgate.net/publication/242915900 Bibliometric performance measures Article in Scientometrics · July 1996 DOI: 10.1007/BF02129596 CITATIONS READS 200 692 2 authors: Francis Narin Kimberly Hamilton Retired President, CHI Research, Inc The Patent Board 96 PUBLICATIONS 10,802 CITATIONS 9 PUBLICATIONS 1,736 CITATIONS SEE PROFILE SEE PROFILE Some of the authors of this publication are also working on these related projects: "UTILIZATION OF BASIC SCIENTIFIC RESEARCH AS EVIDENCED BY US PATENT CITATIONS" Coauthors Carpenter and Narin, Research Management Magazine of IRI, XXIII-2, March 1980. View project All content following this page was uploaded by Francis Narin on 18 February 2015. The user has requested enhancement of the downloaded file. Jointly published by Elsev&r Science Ltd, Oxford and Akad~miai Kiad6, Budapest Scientometrics, Vol. 36, No. 3 (1996) 293-310 BIBLIOMETRIC PERFORMANCE MEASURES F. NARIN, KIMBERLY S. HAMILTON CHI Computer Horizons Inc., 10 White Horse Pike, Haddon Heights, NJ 08035 (USA) r (Received May 24, 1996) Three different types of bibliometrics - literature bibliometrics, patent bibliometrics, and linkage bibliometric can all be used to address various government performance and results questions. Applications of these three bibliometric types will be described within the framework of Weinberg's internal and external criteria, whether the work being done is good science, efficiently and effectively done, and whether it is important science from a technological viewpoint. Within all bibliometrics the fundamental assumption is that the frequency with which a set of papers or patents is cited is a measure of the impact or influence of the set of papers. The literature bibliometric indicators are counts of publications and citations received in the scientific literature and various derived indicators including such phenomena as cross-sectoral citation, coauthorship and concentration within influential journals. One basic observation of literature bibliometrics, which carries over to patent bibliometrics, is that of highly skewed distributions - with a relatively small number of high-impact patents and papers, and large numbers of patents and papers of minimal impact. The key measure is whether an agency is producing or supporting highly cited papers and patents. The final set of data are in the area of linkage bibliometrics, looking at citations from patents to scientific papers. These are particularly relevant to the external criteria, in that it is quite obvious that institutions and supporting agencies whose papers are highly cited in patents are making measurable contributions to a nation's technological progress. Introduction It seems quite appropriate for b i b l i o m e t r i c techniques to play a m a j o r role in the application o f the U . S . G P R A ( G o v e r n m e n t P e r f o r m a n c e and Results Act) to scientific research. M a n y b i b l i o m e t r i c techniques w e r e d e v e l o p e d with the evaluation o f research as a primary goal, especially e v a l u a t i o n o f research departments, international aggregates, for which there are virtually no institutions, and other quantitative alternatives.1 In this paper we will discuss three different categories o f b i b l i o m e t r i c analysis: literature b i b l i o m e t r i c s , patent bibliometrics, and linkage bibliometrics, all o f which may be used in establishing quantitative indicators o f both the outputs and the outcomes o f g o v e r n m e n t p r o g r a m s . 0138- 9130/96/US $ 15.00 Copyright 9 1996 Akaddmiai Kiadr, Budapest All rights reserved F. NARIN, K. S. HAMILTON:BIBLIOMETRICPERFORMANCEMEASURES The drive to develop techniques for evaluating research programs arose in the first years of financial stress following the post World War II period of unlimited science growth. As a result of the development of radar, atomic energy and other technological wonders in the war years, the essential contribution of science to technology became apparent, and of technology to national policy, and the broad support of science became a national goal. Inevitably, of course, the growth in support had to level off, and the scientific community had to face the stressful prospect of limited funding. The challenge of assuring that the output of publicly supported science was worthy of the dollars being devoted to it was attacked in two elegant papers published by Alvin Weinberg in the early 1960's. 2,3 Weinberg set out to systematize the criteria for scientific choices. Weinberg categorized the criteria as internal and extemal. The internal criteria deal with the question "Is it good science?" whereas the external criteria deal more explicitly with what kind of impact that particular area of science was having. It is very important to note that he felt that these two sets of criteria, internal and external, should be viewed separately, a point that seems to be lost in much of the current discussion of program evaluation. Weinberg is an eloquent writer and we will basically summarize what he said in his own words, by excerpting sections from his 1963 paper. "As science grows, its demands on our society's resources grow. It seems inevitable that science's demands will eventually be limited by what society can allocate to it. We shall then have to make choices. These choices are of two kinds. We shall have to choose among different, often incommensurable, fields of science -between, for example, high-energy physics and oceanography or between molecular biology and science of metals. We shall also have to choose among the different institutions that receive support for science from the government - among university, governmental laboratories and industry. The first choice I call scientific choice; the second, institutional choice. My purpose is to suggest criteria for making scientific choices - to formulate a scale of values which might help establish priorities among scientific fields whose only common characteristic is that they all derive support from the government." "Internal criteria are generated within the scientific field itself and answer the question: How well is the science done? External criteria are generated outside the scientific field and answer the question: Why pursue this particular science? Though both are important, I think the external criteria are the more important." "Two internal criteria can be easily identified: (1) Is the field ready for exploitation? 294 Scientometrics36 (1996) F. NARIN, K. S. HAMILTON: BIBLIOMETRIC PERFORMANCE MEASURES (2) Are the scientists in the field really competent?" "Three external criteria can be recognized: (1) technological merit (2) scientific merit, and (3) social merit." The first is fairly obvious: once we have decided, one way or another, that a certain technological end is worthwhile, we must support the scientific research necessary to achieve that end." "The criteria of scientific merit and social merit are much more difficult: scientific merit because we have given little thought to defining scientific merit in the broadest sense, social merit because it is difficult to define the values of our society." "Relevance to neighboring fields of science is, therefore, a valid measure of the scientific merit of a field of basic science." "... that field has the most scientific merit which contributes most heavily to and illuminates most brightly its neighboring scientific disciplines." The imperative for justifying scientific investment also led to the first TRACES study (Technology in Retrospect and Critical Events in Science) which tried to link, in a relatively systematic way, the development of important innovations to the science base. 4 The initiation of the Science Indicators Report Series by the National Science Board in 19725 was another step in responding to the need for quantitative output and outcomes indicators. In the subsequent sections of this paper we will show how the three types of bibliometrics, literature, patent and linkage, can all be used to address two categories of GPR-related questions: is the work being done by an agency good science, efficiently and effectively done, and is it important from a technological viewpoint, in essence Weinberg's internal and external criteria. We will discuss both some of the basic indicators that can be developed, and some of their limitations. However, at the outset we want to make an extremely important point about the skewness of virtually all bibliometric distribution, and how that must be carefully taken into account in any analysis and interpretation of bibliometric data. 6 The fundamental point is that virtually every distribution of scientific productivity or impact is highly skewed, with a small number of highly productive scientists, hot areas, research institutions, etc., and a relatively large number of institutions and producers of much smaller impact. The distributions are more often logarithmic than linear, with difference factors of 10 to 100 between the most productive entities and the least productive entities, in any scientific or technological distribution. Scientometrics 36 (1996) 295 F. NARIN, K. S. HAMILTON: BIBLIOMETRIC PERFORMANCE MEASURES For example, the influence of individual scientific journals - where influences is defined as the weighted average number of citations per paper in a journal - is distributed logarithmically, over a range of more than 100 between the most influential, most cited journals in a subfield and the least influential, least cited journals in a subfield. 7 Similarly, recent data has shown that the number of patents being produced by individual inventors, even within a single company, is logarithmically distributed, just as the original Lotka's Law showed for scientific papers. 8,9 The reason this is very important GPRA related work is that for one individual, or even a small group, a very large amount of the ultimate impact on society may be centered around a few relatively rare, extremely high impact discoveries, papers, inventions, etc. Further, in highly skewed distributions the means and the medians can be quite different, and one may often want to do statistical testing in the logarithm of the distribution, rather than in the distribution itself, in order to have something even approaching a normal curve. Therefore, in all of the subsequent discussions of publication and citation rates, it must be kept in mind that comparisons should always take skewness into account, and that differences of five-to-ten percent in impact between two research institutions are almost certainly insignificant in a realm where the differences between two research institutions are often factors of five-to-ten or more. Within all of the bibliometric analyses, of course, there are a series of basic tenants that are widely and almost universally accepted. The tenants include: 1. Counts of the number of papers, and the numbers of patents, provide basic indicators of the amount of scientific and technological productivity. 2. Counts of citations to these papers and patents, and from patents to papers, provide indicators both of the quality (impact) of research, and of the linkage between basic and applied research, between one subfield and another, and between technology and science. 3. Counts of coauthorships, and especially international coauthorships, are an indicator of quality, and that scientists who cooperate with their colleagues in other institutions and overseas are more likely to be doing quality research than those that are relatively isolated. 4. There are natural differences in citation and publication pattems across research and technology, unrelated to quality of the science, and one must be extremely careful in doing comparisons of any bibliometric parameters without fully adjusting for these differences: for example, between the citation parameters in a relatively lightly citing subfield such as acoustics and a very heavily citing subfield such as biochemistry and molecular biology. 296 Scientometrics 36 (1996) F. NARIN, K. S. HAMILTON: BIBLIOMETRIC PERFORMANCE MEASURES Literature indicators for GPRA The bibliometric indicators which have been in use for the longest time are, of course, literature indicators: indicators of the scientific performance of organizations, agencies and countries based on counts of publications and citations in the scientific literature. These have been used most extensively in the United States in various studies done in the 1970's and 1980's for the National Institutes of Health 1~ and, of course, in the pioneering work at the National Science Foundation in the development of the Science Indicators series, starting with Science Indicators 1972, ongoing to Science and Engineering Indicators-1996, which has just been published. Amongst the literature indicators which are relevant to Weinberg's internal criteria are the number of publications, publications per dollar, the number of citations received by the papers, as well as various influence and impact surrogates for citation rate. In addition to those basic productivity and impact measures, another measure applicable to the internal criteria of whether an agency is producing good science is coauthorship, since it is widely accepted that having papers coauthored with one's colleagues at prestigious universities and research institutes, and with colleagues overseas, are indicators of acceptance within the scientific community. The number of papers produced by a given research institution is certainly a measure of its output, and a very basic one for GPRA. In most bibliometric studies of the physical, biological and mathematical science this is confined to articles, notes and reviews in refereed research journals, such as those covered in the Science Citation Index (SCI). This measure has particular significance when it is placed within the context of an organization's budget. Figure 1 shows such a placement, adapted from a paper published in 1976, graphing the number of biomedical papers published by the top 122 U.S. universities, compared to the amount of funds these schools received from NIH three years before. 11 The correlation, of course, is extremely high, approximately r =: 0.95, indicating that virtually all variation in the number of biomedical papers produced by those institutions is fully accounted for by the funding received: large and small universities produce essentially the same number of papers on a per dollar of funding basis. In general, we have never found that the size of an institution is of any significance in measurement of either the productivity or the quality of its research output. Productivity and quality vary widely, but are not primarily driven by organizational size. Small, highly specialized institutions can produce papers of just as high quality papers per funding increment as large, well known institutions. All bibliometric indicators of the quality of research produced are normally based on either direct citation counts, or various citation surrogates such as journal influence Scientometrics 36 (1996) 297 F. NARIN, K. S. HAMILTON:BIBLIOMETRICPERFORMANCEMEASURES or journal impact factor, and all are subject to one extremely important constraint: the data must be normalized for differences in field, subfield, and sometimes specialty parameters. Citation densities, that is the number of references per paper, the number of times a paper is cited, and time lags all vary widely from one field to another, and one subfield to another, and sometimes even within a subfield by specialty area. As a general rule, citation densities are highest in very hot and fast areas of science such as molecular biology and genetics, and much lower and slower in some of the more traditional areas of astronomy, descriptive biology, mathematics, and so forth. Citation patterns are also quite light in most of the engineering fields, and very different in the Social Sciences. 800 700 600 ll* S s 500 _~ 4oo g 3oo s 9 < 200 ~ 9 ..'". ," 100 0 99 9 D t, . "-eeq9 b s9 * ! I i i i I, i I 2000 4000 6000 8000 10000 12000 14000 16000 Average NIH Constant 1967 Dollars (thousands), 1965-1969 Fig. 1. Publication output v e r s u s NIH funds for 122 universities Perhaps the most fundamental challenge facing any evaluation of the impact of an institution's programs or publications is to properly normalize and adjust for field and subfield differences. This point is illustrated by Fig. 2, which shows the average citations per paper received in a five year window for papers in biomedical subfields, where subfield is, in this case, CHI's division of approximately 1,300 biomedical journals into 49 different subfields. Obviously, the factor of 4 or more between the citation frequency of papers in the most highly cited subfield biochemistry and molecular biology compared to the ~r;D,,tnmatrie~ ~6 (1996) F, NARIN, K. S. HAMILTON:BIBLIOMETRICPERFORMANCEMEASURES least highly cited specific biomedical subfield pharmacy indicates that citation studies of biomedical publication performance must be very carefully adjusted for differences in the subfield. For example, a highly prestigious paper in a lightly citing subfield may be less frequently cited than an average paper in a highly citing subfield, and it is simply not valid to directly compare citation counts in one of the lightly citing clinical subfields with citation counts in the much more highly citing basic subfields. 5 Year Citing Window) o o Fig. 2. Cumulativecites per paper to 1986 Clinical Medicine & Biomedical Research papers from 1986-91 papers by subfield One type of normalization sometimes done, with which we totally disagree, is to divide a papers' citation count by the average citation count for papers in the specific journal in which the paper appeared. This has the perverse effect of penalizing a research institution that encourages its scientists to publish in the most prestigious and best journals in its field. Scientometrics 36 (1996) 299 F. NARIN, K. S. HAMILTON: BIBLIOMETRIC PERFORMANCE MEASURES The correct way to do citation normalization is on a subfield basis, dividing citation counts by subfield analysis. In highly precise studies one should further adjust for the citation frequencies within specialized areas. Coauth0rship rates may be used to determine another internal indicator of performance, based on the generally accepted notion that scientists who coauthor with people outside their institutions, and especially with their colleagues overseas, are the scientists who are the most widely accepted in their community. A particularly interesting demonstration of this is contained in a study done for the European Community, which showed that Community papers that were coauthored with scientists outside the home country were more than twice as highly cited as those that were authored at a single institution within a single country; papers coauthored within one country were cited intermediate between the single institution papers and the internationally coauthored papers. 12 Coauthorship rates should always, of course, be adjusted for the dynamic changes that are occurring in coauthorship, and for the size of countries. International coauthorship has been increasing steadily in the United States and everywhere else, and is quite field dependent, in complicated ways. For example, although institutional coauthorship rates are generally higher in the field of Clinical Medicine than in the field of Biomedical Research, international coauthorship is higher in Biomedical Research than in Clinical Medicine. It is extremely important that one adjust coauthorship data for fields and subfield differences and, of course, year, since coauthorship rates have been increasing monotonically for more than 20 years. The final internal criteria for bibliometric performance to be discussed here relates to the concept of research level. CHI Research has classified each of some 3,000 journals covered in the Science Citation Index (SCI) on a research level scale ranging from 1 (most applied journals) to 4 (most basic journals). This scale allows one to characterize the degree of basicness of a research program, or the publication of a research institution. The four research levels and prototype journals are: Level 1 Applied Technology (Clinical Observation in Biomedicine) Engineering-Technological Science (Clinical Mix in Biomedicine) J Iron & Steel Inst J Am Med Assn J Nuc Sci & Tech Proc IEEE New Eng J Med Level 3 Applied Research (Clinical Investigation in Biomedicine) J Appl Phys Cancer Res Level 4 Basic Scientific Research Phys Rev J Am Ch Soc J Biol Chem Level 2 300 Scientometrics 36 (1996) F. NARIN, K. S. HAMILTON: BIBLIOMETRIC PERFORMANCE MEASURES A particularly telling application of this is illustrated in Fig. 3, redrawn from a paper published almost 20 years ago, which shows the average research level for the publications of the 95 largest U.S. hospitals, medical schools, clinics, and research universities. Also shown there is the average research level for papers published by the U.S. National Institutes of Health (NIH) Intramural program. The figure quite graphically illustrates the usefulness of this indicator: the average level of the papers from the hospitals, clinics and medical schools is, for every single one, more clinical than the average level of the papers from the major universities. Furthermore, the papers of the NIH Intramural program lie right in the middle of the distribution: the NIH Intramural program is more clinical than any university program, more basic than any hospital or medical school program, and, in fact, very beautifully spans the two constituencies which the intramural program at NIH serves so well. The construction of indicators relevant to Weinberg's external criteria within the scientific literature are normally based on cross-citing patterns, either citations to a basic subfield from an applied subfield, or citations to university and other public science from industrial, private science. An illustration of this kind of indicator is shown in Fig. 4 which shows the trend in citation from industry papers to university sector papers in all fields. It is quite clear that industrial science continues to heavily use the more basic papers produced at the universities, and the universities have, if anything, a growing role in producing the science base of industrial technology. There are, of course, various technical problems which impact any bibliometric evaluation. One of the most important problems, especially for program evaluation, is that of constructing the basic program bibliography. For a support program where there are many different grantee institutions, agency records are almost always quite incomplete, and usually in quite non-standard form. Just establishing the journal, volume, page and year for the bibliography of the papers supported in a typical U.S. government agency program is a massive job, and often takes many person months of unification, yet such a standard form is absolutely imperative for any kind of analysis, before you can even establish the field and subfield of the papers, or whether the publications are, in fact, papers in refereed journals, in journals covered by the SCI, and so forth and so on. Even the construction of an institutional bibliography is not trivial. Papers from Harvard University contain literally hundreds of different variations of the name Harvard University, and finding all of them in the literature is a substantial challenge. In recent years the Institute for Scientific Information (ISI) has done some standardization of these names in the SCI, and that is a help in using that data. However, many universities contain hospitals and research labs that have names other Scientometrics 36 (1996) 301 F. NARIN, K. S. HAMILTON: BIBLIOMETRIC PERFORMANCE MEASURES than that of the university, and those addresses often do not explicitly mention the university name, and so that a substantial amount of research is necessary to obtain a decent bibliography. In creating the U.S. Science Indicators over the years, CHI has created a thesaurus of over 600,000 different names which are associated with 3,500 major U.S. research institutions, and it. takes a substantial amount of time, effort and care to construct an institutions bibliography from the many names under which its scientists will publish. MEDICAL SCHOOLS, HOSPITALS, AND CLINICS ] UNIVERSITIES 15 E 0 vltn i 21o (1 = Most Clinical) 2's 37oI NIH INTRAMURAL 3'5 ,'o (4 = Mwst Basic) Average Level Fig. 3. Average level for biomedical papers of NIH and the 95 largest U.S. universities, medical schools, hospitals and clinics. Source: The Intramural Role of the NIH as a Biomedical Research Institute, F. NARIN, S. B. KEITH Federation Proceedings, Vol. 37, No. 8, June 1978 Another important factor that must be carefully normalized for are the time effects. Specifically, of course, citations accumulate over time, and as it was mentioned earlier, citation data must be over comparable time periods, and within the same subfield or area of science. However, in addition to that, the time patterns of citation are far from uniform across country, and any valid evaluative indicator must use a fixed window and be appropriately normalized. To illustrate this Fig. 5 shows a phenomenon we call the "Citation Time Anomaly", namely the anomaly in the percent of references from a country's papers to earlier papers for the same country. The 302 Scientometrics 36 (1996) F. NARIN, K. S. HAMILTON: BIBLIOMETRICPERFORMANCE MEASURES ordinate is the percent of the references from each country's papers which are to the same country's papers, published in the prior year indicated on the abscissa. 0.00 5.00 10.00 15.00 20.00 25.00 30.00 35.00 i i : = ', [ t Industry Sector University Sector 1985 Cixl~ 1981-83 Federal Gov Sector 1986 Ctt.'~ 1982-84 ~] 1997 Cili,~l 1983-88 [ ] 1989 CainQ 1984-86 tr---I C~ 1989 Crting 1985-87 [ [~ 1990 Cklng 1986-88 [ ] 1991 Citing 1997-89 FFRDC Sector FFRDC = Federarly Funded Research & Development Cenler Fig. 4. Percent of industry paper cites to various sectors' papers 1981 journal set Clearly the effects are extremely strong, with some 52 percent of the German physics references to German papers in the same year, only 40 percent in the first year down to 21 percent for papers published ten years earlier. Similarly in U.S. Clinical Medicine, 83 percent of the essentially immediate references are to U.S. papers, whereas after ten years only percent are to U.S. papers. This effect is quite general, and shows just as strongly for almost any country and for almost any field or subfield, and obviously could have major effects on an evaluation. Specifically, if one is evaluating the short time citation rate to non-U.S, papers in a database that is as heavily U.S. as the Science Citation Index is, then the short term Scientometrics 36 (1996) 303 F. NARIN, K. S. HAMILTON: BIBLIOMETRIC PERFORMANCE MEASURES indicators will give far lower citation ratios to the non-U.S, papers than long term data, since many of the quick U.S. references are to U.S. papers. This, of course, would effect any short term citation indicator substitute, such as the impact factor, and argues strongly for being very careful to adjust for the citation window in any kind of cross national Citation analysis. For comparison of institutions within one country this is not as important, but for cross national comparisons it is quite a striking, important phenomenon. Percent of Cites frc~m U . S . P a p e c s t o U . S . C l i n i c a l M e d i c i n e Papers 85 - - 1 9 9 3 8o -- -- -- 1991 ........ 70 __ 1989 - 65 line represen~ o n e d t i n o year . . . . 1987 ...... 1988 N g m b e r e~ Ymlrs P l k x t o C ~ l g Year Pmcent of C#es h'om W. Gmmm P~oem to W. Gmman 1993 - -........ 30 1991 1989 [ [ 20 10 0 Each line represents o n e chino yee,r ~ ~ J i i i t i i i t i N u m b e r o f Years Pdor t o CRIng Yelw Fig. 5. "Citation time anomaly" 304 Scientometrics 36 (1996) F. NARIN, K. S. HAMILTON: BIBLIOMETRIC PERFORMANCE MEASURES Linkage indicators for GPRA A series of very direct indicators of the impact that an agency's science is having on industrial technology may be developed from linkage statistics, by characterizing the citations from patented industrial technology to the agency's scientific research papers. The analysis of linkage from technology to science has been underway for a number of years, and is becoming a crucially important means of demonstrating the importance of science via Weinberg's external criteria. This type of analysis of industrial impact is now being used at a number of research agencies. 13-15 In a paper published in 1992 we summarized the status of linkage knowledge at that time. 16 When a typical U.S. patent is issued it contains on its front page seven or eight references to earlier U.S. patents, one reference to a foreign patent, and 1.3 non patent references (NPR), which are reference to material other than earlier U.S, or foreign patents. Of these non patent references approximately 43 percent or a total of 60,000 per year in recent years are to scientific research papers in the central core of 3500 or so journals covered by the Science Citation Index. There are a number of exceedingly important aspects of this general relationship from a GPRA viewpoint. First, technology-to-science linkage is increasing rapidly, with a two to three-fold increase in citations from patents to papers over the last six years, as shown in Fig. 6, which summarizes the citations from U.S. patents to U.S. authored papers in two 11 year citation windows. The number of papers cited has more than doubled, the number of citations to these papers has increased similarly, and the number of support sources on the cited papers has more than tripled. This rapidly increasing citation is to rather basic, mainstream science. The citation is heavily to Level 3 and 4 Applied and Basic research papers; in the Biomedical and Chemical areas predominately to the Level 4, very basic journals; in the electrical and computer industry more mixed, with a heavy citation to Level 3 Applied Physics journals. This citation is also quite rapid, with the age of the cited scientific papers almost as young as the age of the cited patents. 17 It is largely to public science, to scientific work done at universities or government laboratories, and supported by government agencies. The top 10 research support agencies cited in 1993 and 1994 U.S. patents include 6 NIH Institutes, the National Science Foundation, the American Cancer Society, the Department of Energy, and the Navy.. This data clearly demonstrates the crucial contribution that public science makes to patented industrial technology. Scientometrics 36 (1996) 305 F. NARIN, K. S. HAMILTON: BIBLIOMETRIC PERFORMANCE MEASURES # of Support Sourcas Cited * 80000 r ~ all cited papers 9 supported papers 70000 ; [~ unsupported papers i ] 60000 ~- 50000 40000 30000 i I i ! 20000 T~ 10000 1957/88 1993/94 citLng citing 1975-85 1981-91 1987/85 1993/94 citing citing 1975-85 1981-91 1997/85 1993/94 citing citing ~975-85 1981-91 9 Includes only those papers found in the librar'/ Fig. 6. Citations from U.S. patents to U.S.-authored SCI journal papers Patent indicators for GPRA The patents generated by various government laboratories, and by the grantees and contractors of government agencies, may also be used in performance assessments, just as papers are. As with papers, patents data may be used to construct indicators of internal and external performance for an agency, internal in that their existence, per se, is an indicator of contributions to technology, and external in that the impact of those patents on other technology, and ebpecially on U.S. private sector technology, is an indicator of the contribution of those technological inventions to society. 306 Scientometrics 36 (1996) F. NARIN, K. S. HAMILTON: BIBLIOMETRIC PERFORMANCE MEASURES The most basic of the internal indicators are, of course, the numbers and distributions of patents across different technologies. As a general rule government agencies have been relatively steady in their patenting over the last few decades, whereas university patenting has been increasing quite rapidly. Figure 7, for example, shows the rate of patenting by the U.S. government agencies and U.S. universities . Clearly, 9niversities have been increasing their patenting at a very much more rapid rate than~tlae government agencies. One general point about university patenting in the U.S. is that it is rather heavily concentrated in biotechnology, one of the very, very hot areas of U.S. technology. Other properties of those patents, particularly their relatively high science linkage, also indicate that university patents are in leading edge areas of technology. 2500 US Academic 21XX) ~ 1513t3 i ///1 m | z 1000 \ \ \\ 500 0 1997 j ~ F ! i I I ' 1988 1989 1990 1991 1992 1993 1994 1995 Fig. 7. U.S. patenting trends As a measure of the quality of patents one can compute various citation indicators. The one we tend to use most is Current Impact Index (CII), which is, in essence, the citation rate for the last five years of an agency's patents, as cited by the most current Scientometrics 36 (1996) 307 F. NARIN, K. S. HAMILTON: BIBLIOMETRIC PERFORMANCE MEASURES year. For example, in computing CII for 1995 we will take all the agency's patents in each of the five years 1990-1994, note how many times each patent is cited, divide it by the expected citation rates for all patents in the same product group and year, and average those ratios, so that a CII equal to 1.0 is the expected value. Science Linkage (SL), the average number of times each patent references scientific materials, is another indicator of quality patents. For example, for the 5 years, 1991-1995, U.S. Navy patents have CII = 0.69 and SL = 1.09, while U.S. Dept. of Health and Human Services (largely NIH patents) have CII = 0.75 but SL = 8.10. While neither agencies' patents are highly cited (expected CII = 1.00) they are quite science-linked, especially HHS. Overall for all U.S. invented U.S. patents in the time period the SL = 0.99, so that both NIH and NRL are in relatively highly science linked areas with NIH, of course, far more heavily science-linked. Evaluating the efficiency of a government laboratory or program in producing patents is very problematic, and perhaps impossible. The R&D dollars associated with each patent produced in a large industrial laboratory tends to be in the one to five million dollar range, and we have observed similar ranges for some government laboratories. However, most government laboratories have fundamental missions far removed from the production of patentable technology, and in many cases the patents are the product of only certain parts of the labs. Therefore, comparisons of patents per R&D dollar are far more problematic than comparisons of papers per R&D dollar in the university environment, where publication is a central objective of the scientific process. There are a number of indicators of the external contribution that agency and laboratory patents make, both direct indicators in terms of licensing, and indirect in their contribution to the advances of technology in general. The indirect contributions can be tracked by looking at the pattern of citation to laboratories' patents, as is illustrated by Fig. 8, which shows citations from the top assignees citing to NRL patents, Another way of looking at the patent impact from an agency viewpoint is to look at the licensing information, at how many licenses the laboratory has obtained, and how much revenue is derived from those licenses. While we don't have detailed licensing revenue information, licensing activity is a reflection of the growing technology transfer activities of most U.S. labs and universities, and there is an interesting interaction that we have found between the bibliometric characteristics of patents and their licenses. In particular, we looked at the patents which had been licensed for most of the Navy laboratories, and found that there are two important bibliometric characteristics of licensed patents: first, these 308 Scientometrics 36 (1996) F. NARIN, K. S. HAMILTON: BIBLIOMETRIC PERFORMANCE MEASURES patents are cited much more often than the other lab patents, and second, they tend to be much more science linked. This is illustrative of a rather interesting convergence of the internal and external criteria. Those patents which are most highly cited and most science linked, both internal indicators of the likely impact of that technology, are also the patents that tend to be most heavily licensed, and are therefore making the most ~lirect external contribution to the economy. I~ oro. .... '"J 16 SumJtomoElec t US Air .9 I,owa S~. Univ ]13_j AIcstel Alsthom CGE Philips i .... ..... ]/2"/ ;Si GE PLC Mitsubishi Elec Raytheon NASA Sumitorno Metals ,. [ ~1 ! ,~IL._ / .~ Northrup Grurnrnan 8P NRL 231 Patents ~ .. ./ / //' J/" /" "I'/ " ]j~/ 10 ToshibaNEC Fujitsu Fig. 8. Companies with 10 or more patents citing to 1985-89 NRL patents Which brings us to our final point. As was perceived at the very beginning of the modern age of science, the most important characteristic of research is the quality of the work. Highly cited, science dependent, quality patents, and the related basic scientific work, seem to have wide economic impact. This has also been Shown in the sample cases studied from the beginnings of modern bibliometric analysis, and now is Scientometrics 36 (1996) 309 F. NARIN, K. S. HAMILTON: BIBLIOMETRIC PERFORMANCE MEASURES shown statistically in technology and in its interface with science: high impact, basic scientific research is the type of research directly driving our technological advancement. References 1. 2. 3. 4. 5. 6. 7. 8. 9. 10. 11. 12. 13. 14. 15. 16. 17. F. NARIN, "Evaluative Bibliometrics: The Use of Publication and Citation Analysis in the Evaluation of Scientific Activity", Report prepared for the National Science Foundation, Contract NSF C-627, NTIS Accession #PB252339/AS, (1976), 456pp. A.M. WEINBERG,Criteria for scientific choice II: The two cultures, Minerva, 3 (1964) 3-14. A.M. WEINBERG,Criteria for scientific choice", Minerva, 1 (1963) 159-171. IIT Research Institute, "Technology in Retrospect and Critical Events in Science (TRACES)". Report prepared for the National Science Foundation, Contract NSF-C535, (1968), 72pp. National Science Foundation, Science Indicators 1972, U.S. Government Printing Office, Reports of the National Science Board. P.O. SEGLEN,The skewness of science, Journal of the American Society for Information Science, 43 (9), (1992) 628-638. G. P1NSKI, F. NARIN, Citation influence for journal aggregates of scientific publications: theory, with application to the literature of physics, Information Processing and Management, 12 (5), (1976) 297-312. F. NARIN, A. BREITZMAN,Inventive Productivity, Research Policy, 24 (1995) 507-519. A.J. LOTKA, The frequency distribution of scientific productivity, Journal of the Washington Academy of Science, 16 (1926) 317-323. F. NARIN, H.H. GEE, "An Analysis of Research Publications Supported by NIH, 1973-1980", NIH Program Evaluation Report, U.S. Department of Health and Human Services, Public Health Service, National Institutes of Health, (1986). J.D. FRAME, F. NARIN, NIH funding and biomedical publication output, Federation Proceedings, 35 (14) (1976) 2529-2532. F. NARIN, K. STEVENS, E. WHITLOW, Scientific cooperation in Europe and the citation of multinationally authored papers, Scientometrics, 21 (1991) 313-323. L.B. ELLWEIN, P. KROLL, F. NAR1N, Linkage between research sponsorship and patented eye-care technology, To be published (1996). J. ANDERSON, N. WILLIAMS, D. SEEMUNGAL, F. NARiN, D. OLIVASTRO, Human genetic technology: Exploring the links between science and innovation. To be published in Technology Analysis and Strategic Management, (1996). F. NARIN, "Linking Biomedical Research to Outcomes - The Role of Bibliometrics and Patent Analysis", Presented at The Economics Round Table, National Institutes of Health, (1995), 27pp. F. NARIN, D. OLIVASTRO, Status report - linkage between technology and science, Research Policy, 21 (3) (1992) 237-249. F. NARIN, E. NOMA, IS technology becoming science?, Scientometrics, 7 (1985) 369-381. 310 View publication stats Scientometrics 36 (1996)