Future UK RAEs: Peer review or Metrics-Based

Jonathan Levitt jonathan at LEVITT.NET
Sun Apr 9 10:02:37 EDT 2006


Hi,

Steven wrote "1) peer review has already been done for published articles,
so the issue is not (i) peer review vs. metrics but (ii)  peer review plus
metrics vs. peer review plus metrics plus 'peer re-review' (by the RAE
panels)."  .  To me the issue is not so much peer review vs. metrics, but
finding a combination of peer review and citation/usage metrics that seems
particularly likely to be effective at measuring research quality.

In principle, the sequel to the RAE could take into account the accrual
reviews from the journal that published the article (rather than conduct
their own reviews), but in my view there would need to be stringent checks
on their authenticity.

Best regards,
Jonathan.


----- Original Message -----
From: "Stevan Harnad" <harnad at ECS.SOTON.AC.UK>
To: <SIGMETRICS at LISTSERV.UTK.EDU>
Sent: Thursday, April 06, 2006 10:56 PM
Subject: Re: [SIGMETRICS] Future UK RAEs to be Metrics-Based


> Adminstrative info for SIGMETRICS (for example unsubscribe):
> http://web.utk.edu/~gwhitney/sigmetrics.html
>
> The following article is excellent and accurate overall.
>
>    Cliffe, Rebeca (2006) Research Assessment Exercise: Bowing
>    out in Favour of Metrics. EPS Insights: 3 April 2006
>    http://technology.guardian.co.uk/weekly/story/0,,1747334,00.html
>
> One can hardly quarrel with the following face-valid summary from
> this article:
>
>>   "The move to a new metrics based system [for RAE] will no doubt please
>>   those who see a role for institutional repositories in monitoring
>>   research quality. The online environment has thrown up new metrics,
>>   which could be used alongside traditional measures such as citations.
>>   Usage can be measured at the point of consumption -- the number
>>   of "hits" on a particular article can indicate the uptake of the
>>   research. Web usage would be expected to be an early indicator of
>>   how often the article is later cited. Some believe that institutional
>>   repositories should be used as the basis for ongoing assessment of all
>>   UK peer-reviewed research output by mandating that researchers should
>>   place material in repositories. They argue that this would allow
>>   usage to be measured earlier, through downloads of both pre-prints
>>   and post-prints. Of course, this course of action would also advance
>>   the cause of open access by making this research available free."
>
> But there are a few points of detail on which this otherwise accurate
> report could be made even more useful:
>
>    (1) peer review has already been done for published articles, so
>    the issue is not (i) peer review vs. metrics but (ii)  peer review
>    plus metrics vs. peer review plus metrics plus "peer re-review"
>    (by the RAE panels). It is the re-review of already peer-reviewed
>    publications that is the wasteful practice that needs to be scrapped,
>    given that peer review has already been done, and that metrics are
>    already highly correlated with the RAE ranking outcome anyway.
>
>    (2) For the fields in which the current RAE outcome is not already
>    highly correlated with metrics, further work is needed; obviously
>    works other than peer-reviewed article or books (e.g., artwork,
>    multimedia) will have to be evaluated in other ways, but for
>    science, engineering, biomedicine, social science, and most fields
>    of humanities, books and articles are the form that research output
>    takes, and they will be amenable to the increasingly powerful and
>    diverse forms of metrics that are being devised and tested. (Many
>    will be tested in parallel with the 2008 RAE, which will still be
>    conducted the old, wasteful way; some of the metrics may also be
>    testable retrospectively against prior RAE outcomes.)
>
>>   "Proponents of a metrics based system point to studies that show
>>   how average citation frequencies of articles can closely predict
>>   the scores given by the RAE for departmental quality, even though
>>   the RAE does not currently count these."
>
> True, but the highest metric correlate of the present RAE outcome
> is reportedly prior research funding (0.98). Yet it would be a big
> mistake to scrap all other metrics and base the RAE rank on just
> prior funding. That would just generate a massive Matthew Effect and
> essentially make top-sliced RAE funding redundant with direct competitive
> research project funding (thereby essentially "bowing out" of the dual
> RAE/RCUK funding system altogether, reducing it to just research project
> funding). What is remarkable about the high correlation between citation
> counts and RAE ranks (0.7 - 0.9), even though the correlation is not quite
> as high as with prior funding (0.98), is that citations are not presently
> counted in the RAE (whereas prior funding is)! Not only are citations a
> more independent metric of research performance than prior funding, but
> counting them directly -- along with the many other candidate metrics --
> can enrich and diversify the RAE evaluation, rather than just make it
> into a self-fulfilling prophecy.
>
>>   "However, metrics tend to work better for the sciences than the
>>   humanities. Whereas the citation of science research is seen as an
>>   indicator of the quality and impact of the research, in the humanities
>>   this is not the case.  Humanities research is based around critical
>>   discourse and an author may be citing an article simply to disagree
>>   with its argument."
>
> I don't think this is quite accurate. It might be true that humanities
> research makes less explicit use of citation counts today than science
> research. It might even be true that the correlation between citation
> counts and research productivity and importance is lower in the
> humanities than in the sciences (though I am not aware of studies to that
> effect). And it may also be true that citation counts in the humanities
> are less correlated with RAE rankings than they are in the sciences. But
> the familiar canard about articles being cited, not because they are
> valid but important, but in order to disagree with them, has too much
> the flavour of the a-priori dismissiveness of citation analysis that we
> hear in *all* disciplines from those who have not really investigated
> it, or the evidence for/against it, but are simply expressing their own
> personal prejudices on the subject.
>
> Let's see the citation counts for humanities articles and books, and
> their correlation with other performance indicators as well as RAE
> rankings rather than dismissing them a-priori on the basis of anecdotes.
>
>>   "Also, an analysis of RAE 2001 submissions revealed that while some
>>   90% of research outputs listed by British researchers in the fields
>>   of Physics and Chemistry were mapped by ISI data, in Law the figure
>>   was below 10%, according to Ian Diamond of the Economic and Social
>>   Research Council (ESRC) (Oxford Workshop on the use of Metrics in
>>   Research Assessment)."
>
> I am not sure what "mapped by ISI data" means, but if it means that ISI
> does not
> cover enough of the pertinent journals in Law, then the empirical question
> is:
> what are the pertinent journals? can citation counts be derived from their
> online
> versions, using the publishers' websites and/or subscribing institutions'
> online
> versions? how well does this augmented citation count correlate with the
> ISI
> subsample (<10%)? and how well do both correlate with RAE ranking? (Surely
> ISI
> coverage should not be the determinant of whether or not a metric is
> valid.)
>
>>   "Ultimately, a combination of qualitative and quantitative indicators
>>   would seem to be the best approach."
>
> What is a "qualitative indicator"? A peer judgment of quality? But
> that quality judgment has already been made by the peer-reviewers of
> the journal in which the article was published -- and every field has a
> hierarchy of quality among journals that is known (and may even sometimes
> be correlated with the journal's impact factor, if one compares like
> with like in terms of subject matter). What is the point of repeating
> the peer review exercise? And especially if here too it turns out to be
> correlated with metrics? Is it?
>
>>   "While metrics are likely to be used to simplify the research
>>   assessment process, the merits of a qualitative element would be to
>>   ensure that over-reliance on quantitative factors does not unfairly
>>   discriminate against research which is of good quality but has not
>>   been cited as highly as other research due to factors such as its
>>   local impact."
>
> Why not ask the panels first to make quality judgments on the journals in
> which
> the papers were published, and then see whether those rankings correlate
> with the
> author/article citation metrics? and whether they correlate with the RAE
> rankings
> based on the present time-consuming qualitative re-evaluations? If the
> correlations prove lower than in the other fields (even when augmented by
> prior
> funding and other metrics) *then* there may be a case for special
> treatment of
> the humanities. Otherwise, the special pleading on behalf of uncited
> research
> sounds as anecdotal, arbitrary and ad hoc as the claim that high citations
> in humanities betoken disagreement rather than usage and importance,
> as in other fields.
>
>>   "Unless more appropriate metrics can be developed for the humanities,
>>   it would seem that an element of expert peer review must remain in
>>   whatever metrics based system emerges from the ashes of the RAE."
>
> It has not yet been shown whether the same metrics that correlate highly
> with RAE outcome in other fields (funding, citations) truly fail to
> do so in the humanities. If they do fail to correlate sufficiently,
> there are still many candidate metrics to try out (co-citations,
> downloads, book citations, CiteRank, latency/longevity, exogamy/endogamy,
> hubs/authorities, co-text, etc.) before having to fall back on repeating,
> badly, the peer evaluation that should have already have been done,
> properly, by the journals in which the research first appeared.
>
> Stevan Harnad
>
>> Research Assessment Exercise:  http://www.rae.co.uk
>> Economic and Social Research Council:  http://www.eserc.ac.uk
>>
>> Open access: practical matters become the key focus, EPS Insights, 10
>> March 2005
>> http://www.epsltd.com/accessArticles.asp?articleType=1&updateNoteID=1538
>>
>> Citation Analysis in the Open Access World, imi, September 2004
>> http://www.epsltd.com/accessArticles.asp?articleType=2&articleID=236&imiID=294



More information about the SIGMETRICS mailing list