[*Definition of h-index corrected (Tue, 4/2 1:45pm EST)]
The research I did on facts about this year's tenure-track hires (so far) got me thinking about something else: how our profession evaluates philosophical quality and impact. In many (I think most) other academic disciplines, individual articles and journals appear to be measured on the basis of measurable impact on the discipline. Journals appeared to ranked, for instance -- that is, their relative prestige in the discipline appears to be largely determined by -- their impact factor, which is simply the average number of citations to that journal's recent articles (another, possibly better, measure is the journal's eingenfactor). The impact of particular articles is typically also measured on these grounds. Finally, the impact of a given scholar is often understood in terms of what is known as the h-index (or Hirsch index), which understands a scholar's impact as "the largest number h such that h publications have at least h citations."
Here's what got me started thinking about this stuff. First, the more research I did on people's publications (both in TT-hiring and beyond), the more I noticed that many papers published in highly-ranked journals (e.g. Philosophical Quarterly, etc.) that have few, if any citations -- fewer citations that papers in lower-ranked journals. Second, very few philosophy journals seem to post their impact-factor. Third, a lot of evaluation in our discipline -- in particular, evaluation of how "good" someone is -- seems to be a matter of reputation. People say things like, "Awesome Journal is better than So-So Journal because Awesome Journal tends to have better papers in it" (I've heard this sort of thing a lot). Similarly, one often hears about how one must publish in top-25 journals, how search committees (particularly at R1 institutions) care about this, etc.
Here's the thing, though. Consider a paper published 10 years ago in, say, Phil Review. Suppose this paper has, I don't know, two citations. If the research I've done is any indication, you'd be surprised how much this happens. Lots of journal articles -- even articles in "top" places -- fall stillborn from the presses. Few, if any, people cite them. On the other hand, suppose a paper published two years ago in So-So Journal has 10 citations, two of which are standalone papers based on it (i.e. responding to it, etc.). Whereas the former paper (the one in the top journal) appears to have little to no impact on the profession, the latter paper has had more.
Now, here's the thing: in other disciplines, this sort of difference would presumably be recognized. People in, say, psychology and physics make a big deal out of how many times particular papers have or have not been cited. Yes, publishing in a top journal is an honor -- but if a paper published there makes no impact on future work, people notice that. Is this true in our discipline? My (admittedly anecodotal) impression is that it isn't really. One constantly hears about where people have published ("S-/he published an article in Phil Review"). But one rarely hears any mention of articles' impact (aside from "famous" articles).
Here's why I worry about this. There are many well-discussed issues with the peer-review process in our discipline. For one thing, reviewers are widely reported to engage in the abonimable process of "Google reviewing" (breaking blind-review by Googline paper titles). For another thing, in contrast to people in tenure-track positions -- particularly, at R1 institutions -- who don't come up for tenure for 5-6 years and thus can afford to have papers out at top journals for a few years, people in adjunct positions, post-docs and VAPs don't have this luxury: they need publications now, and so are often likely to send papers to lower-ranked journals where they have a better shot of getting accepted. One could go on.
Given these and many other issues of bias (e.g. "in-crowds" in the discipline, etc.), it seems to me that our discipline should take a good, hard(-er) look at how we evaluate philosophical impact and merit. Offhand, it seems to me that our discipline should attempt to be more objective in these matters than at present. We should look not look primarily at where articles have been published (even if a journal's stature is admittedly defeasible evidence of the general quality of work that appears within it); we should look, primarily, at the impact that articles have actually had in the profession -- which can be measured in many relatively objective ways (not just mere citation numbers, but eigenfactors, which measure, roughly, how "influential" a given article is).
What say you all?
Problem with citation data is that some areas are bigger than others, and some topics interest people outside philosophy. Bad work in bioethics routinely gets many more citations than better work in metaphysics: I know, having produced both.
Posted by: Neil | 04/01/2013 at 04:35 AM
I’m skeptical. It’s true that citation data are easier to track than other things. But that doesn’t give us good reason to care about them. We should, I think, be wary of embracing standards merely because they are more objective. This is, to beat your favorite drum, probably part of why we fetishize rigor: it is easier for readers to agree on what is rigorous than on what is insightful. If we ought to care about citation data it cannot be because citations are easy to track. So a defense of tracking citation numbers needs to say more about why we should care about citation data. This raises a few questions:
1: What do citations data track that matters? And for what purpose are they supposed to matter? Influence might matter when we’re trying to think about the the merit of papers that have been around for a decade or more, but it strikes me as a terrible way to evaluate the work of a paper published a year or two ago by someone on the job market. When we’re considering the merit of especially recent work, we want to think about what influence it could have, or perhaps should have, not what influence it has had.
2: (Neil’s point is a worry about this) Suppose citations track something that matters for some purpose. Does it do so consistently enough across the discipline to be helpful? You might suggest, for example, that influence is a good way to think about at least some part of the merit of the publication records of philosophers who are up for promotion to full professor. Even if that's so, we have to address the fact Neil points to: some subdisciplines are quite different from others, and this bound to make measures of influence favor some subdisciplines over others.
3: What effects, if any, would it have if we began to emphasize people to aim to have influence? What kind of papers should we expect to written in a discipline that encourages the production of "influential" work?
Posted by: Ben Bryan | 04/01/2013 at 11:33 AM
Wouldn't all the same biases that distort what papers get accepted also distort what papers get cited? Someone who is not well connected is less likely to have his/her published work read in the first place. And those who do read it will be more inclined to ignore it (presumably if you know someone and already think highly of him/her, you will be less likely to ignore his/her paper on X when you write a paper on X).
Posted by: B.M. | 04/01/2013 at 12:07 PM
Neil, Ben, and B.M.: Thanks, all, for your comments. I certainly understand the worries that each of you raise. But here's an initial thought in reply.
It seems to me that we face something of a Scylla/Charybdis (or "rock and a hard place") dilemma here.
On the one hand, we can go with seemingly "subjective" measures of impact and quality -- i.e. judgments by people in the discipline at large. This seems to me to be the emphasis in philosophy. Yet this approach has costs, as there are all kinds of ways that judgments about quality and impact can be biased (e.g. in favor of people from top departments, people in TT positions, who have certain luxuries of time in sending things out to top journals, etc.).
Some (though by no means all) of these problems would seem to be addressable in terms of more "objective" measures (i.e. citation rates), as citation rates aren't based on subjective judgments about what is "good" but rather actual *impact* in the discipline (i.e. discussion, etc.). Despite the problems you all raise -- and I don't deny they exist -- these things seem to me relevant. A paper in a top journal which, say, ten years later has one or two citations has, I think, clearly made less of a contribution to the field than a paper in a lesser journal that has been cited a lot (even if it is a bad paper, bad papers can contribute to *better* future work, whereas a "good" paper that is never cited never does much of anything).
But, as you all point out, there are problems here too. Citation rates can be misleading on several grounds you all have pointed out: different sizes of disciplines, ignoring differences in quality of work, etc.
Here, then, is my suggestion. If both approaches have benefits and drawbacks -- and I think this is right -- there should be some *balance* to how they are used in evaluating people.
This is, I think, how it's done in many other fields. My wife, for instance, works in psychology. In their field, they make both kinds of judgments -- judgments about paper-quality and journal placement, but also judgments based on citation-numbers.
I didn't mean to suggest that we should do away with judgments of quality (though I realize the tenor of my post might have indicated that). My worry, rather, was that philosophy perhaps focuses too much on reputational judgments and not *enough* on numbers (viz. I've never heard someone say something like, "Yeah, so-and-so's paper was in Phil Review, but nobody's ever written anything about it" -- which still seems to me a relevant kind of judgment to make).
Finally, B.M., a quick reply to your comment: that is of course possible. However, my (admittedly anecdotal) research suggests that bias affects citation rates less than more subjective judgments of "quality." In my research, I regularly came across articles by well-known people that have hardly been cited at all and articles by lesser-known people, in lower-ranked journals, that have a lot of citations. In other words, I get the (anecdotal) impression that bias affects what people write about and cite less than, say, judgments about how good an article is based on what journal it has appeared in. But again, this is just an impression. It could be wrong.
Anyway, thanks again for all your comments!
Posted by: Marcus Arvan | 04/01/2013 at 01:48 PM
Citations are an imperfect measure. I've been told that one paper of mine put an end to debate concerning a particular topic, and that's why it's not cited! Cited work tends to be first or come at a critical juncture, but need not be the best. Indeed, I suspect some papers are heavily cited *because* they're wrong.
Nevertheless, I agree that we should rely more on objective measures. There's definitely an echo chamber effect in philosophy: this is a good paper because it's published in a top journal, and we all "know" the top journals publish the best material (as though even the most selective journals don't make mistakes about what to publish!). So yes, less of the reputational bias.
Posted by: Michael Cholbi | 04/01/2013 at 10:42 PM
Incidentally, I don't think you've got the h-index definition right in your post. If you published 10 papers, each of which is cited 10 times, your h-index is 10. If, like Gettier, you published just one article, but it has 1,700 citations, your h-index is the same as someone who published one article with just one citation - namely, 1.
Your h-index can't be higher than the number of articles you've published, no matter how well cited they are.
So the idea behind the h-index is that it rewards people who write a lot of papers that are cited a lot of times. Those who write only a few papers, regardless of how well cited they are, cannot get a high h-index.
Posted by: Chris Stephens | 04/02/2013 at 12:20 AM
Your definition of the h-index is not accurate. The h-index is defined as "the largest number h such that h publications have at least h citations." So a person with 20 papers, 10 of which are cited at least 10 times has an h index of 10; a person with only 10 papers, all of them are cited 10 times or more has the same h-index. In other words, you aren't penalized for writing papers that aren't cited.
Also, the total number of citations plays only a modest role. To take an extreme example, suppose you're someone like Gettier with few papers, but one of them is very influential (amassing 1000s of citations), your h-index cannot become higher than say, 1 or 2, even if you have a total citation count of over 2000.
The h-index is a good measure of combined productivity and impact for mature researchers. For instance, in Google Scholar you can find that people like Dan Dennett and David Chalmers have high h-indices. Dennett's h-index (66) is higher than Chalmers' (44) probably because the former is cited to a greater extent outside of philosophy than the latter (although both have a lot of influence outside of our discipline).
I don't think the h-index is a good evaluation tool for young researchers, but on the other hand, I think should count for something if you are early career and have a high h-index (e.g., in tenure reviews). It is possible to bump your h-index up by self-citation, as these are included, but it is very hard to raise one's own h-index significantly only through self-citation.
A while ago I looked at the h-index of journals in philosophy. You can find these data on Google scholar. What I found particularly interesting in this list is that journals that tend to get more citations do not correlate particularly well with perceived prestige in our discipline.
Here is the top ten list of journals in philosophy according to their h-index over 5 years:
1. Synthese
2. Philosophical Studies
3. Mind & Language
4. Noûs
5. The Journal of Philosophy
6. Mind
7. Journal of Consciousness Studies
8. Philosophy Compass
9. Philosophy and Phenomenological Research
10. Phenomenology and the Cognitive Sciences
See here for my blogpost about this: http://www.newappsblog.com/2013/02/the-h-index-of-philosophy-journals-and-their-relative-prestige.html
Posted by: Helen De Cruz | 04/02/2013 at 08:32 AM
Chris and Helen: Thanks for the correction!
Posted by: Marcus Arvan | 04/02/2013 at 09:52 AM
This is an important question.
Bad papers may trigger a number of replies, pointing out the errors. So replies and citations are flawed. But then so are other measures...
I wonder whether it would be possible to institute a sort of Amazon type review and rate system for philosophy papers and books, where only bona fide academics could do the reviewing a rating? Academia.edu would be the obvious place to do this as it already has a large body of registered users…
Clearly there are potential problems with this, but it might be worth adding to the other flawed methods of assessment...
Posted by: Dan Dennis | 04/03/2013 at 12:47 AM