CWTS Leiden Ranking 2015
Loet Leydesdorff
loet at LEYDESDORFF.NET
Fri May 22 02:53:18 EDT 2015
PS. Size-dependent is rather stable. This is also normalized at the micro-cluster level!
These results also accord with my intuition.
The problem must thus be that the denominators fluctuate. Is this correct, Nees Jan?
Would this be an argument for using P(10%) rather than PP(10%)? Or is this field-specific? (This is SSH).
Best,
Loet
size dependent
2006-2009
2007-2010
2008-2011
2009-2012
2010-2013
Delft Univ Technol
13
12
11
11
12
Eindhoven Univ Technol
11
11
13
13
13
Erasmus Univ Rotterdam
2
4
4
4
4
Leiden Univ
8
9
9
8
8
Maastricht Univ
7
7
8
9
9
Radboud Univ Nijmegen
6
6
6
6
6
Tilburg Univ
9
8
7
7
7
Univ Amsterdam
1
1
1
1
1
Univ Groningen
5
5
5
5
5
Univ Twente
12
13
12
12
10
Utrecht Univ
3
3
3
3
2
VU Univ Amsterdam
4
2
2
2
3
Wageningen Univ & Res Ctr
10
10
10
10
11
_____
Loet Leydesdorff
Emeritus University of Amsterdam
Amsterdam School of Communication Research (ASCoR)
<mailto:loet at leydesdorff.net> loet at leydesdorff.net ; <http://www.leydesdorff.net/> http://www.leydesdorff.net/
Honorary Professor, <http://www.sussex.ac.uk/spru/> SPRU, University of Sussex;
Guest Professor <http://www.zju.edu.cn/english/> Zhejiang Univ., Hangzhou; Visiting Professor, <http://www.istic.ac.cn/Eng/brief_en.html> ISTIC, Beijing;
Visiting Professor, <http://www.bbk.ac.uk/> Birkbeck, University of London;
<http://scholar.google.com/citations?user=ych9gNYAAAAJ&hl=en> http://scholar.google.com/citations?user=ych9gNYAAAAJ&hl=en
From: Loet Leydesdorff [mailto:loet at leydesdorff.net]
Sent: Friday, May 22, 2015 7:10 AM
To: 'ASIS&T Special Interest Group on Metrics'
Subject: RE: [SIGMETRICS] CWTS Leiden Ranking 2015
Dear David and colleagues,
The time series provides an impression of the stability.
Since I happen to be most familiar with this system, I chose “Social Science and Humanities” for the 13 Dutch universities,
PP(top-10%), fractional counting, etc. (default values). Size-independent ranking; that is, normalized over the 4,000 clusters or micro-fields.
SSH
2006-2009
2007-2010
2008-2011
2009-2012
2010-2013
Delft Univ Technol
13
12
13
11
12
Eindhoven Univ Technol
2
3
11
12
13
Erasmus Univ Rotterdam
1
2
5
6
8
Leiden Univ
7
10
7
1
3
Maastricht Univ
9
8
6
8
5
Radboud Univ Nijmegen
10
7
4
5
6
Tilburg Univ
11
11
8
7
7
Univ Amsterdam
3
4
2
4
1
Univ Groningen
8
6
9
10
9
Univ Twente
12
13
12
13
10
Utrecht Univ
5
5
3
2
4
VU Univ Amsterdam
4
1
1
3
2
Wageningen Univ & Res Ctr
6
9
10
9
11
cid:image002.png at 01D0945D.7631D0F0
Correlations
v06t09
v07t10
v08t11
v09t12
v10t13
Kendall's tau_b
v06t09
Correlation Coefficient
1.000
.718**
.359
.205
.282
Sig. (2-tailed)
.
.001
.088
.329
.180
N
13
13
13
13
13
v07t10
Correlation Coefficient
.718**
1.000
.538*
.282
.256
Sig. (2-tailed)
.001
.
.010
.180
.222
N
13
13
13
13
13
v08t11
Correlation Coefficient
.359
.538*
1.000
.692**
.718**
Sig. (2-tailed)
.088
.010
.
.001
.001
N
13
13
13
13
13
v09t12
Correlation Coefficient
.205
.282
.692**
1.000
.667**
Sig. (2-tailed)
.329
.180
.001
.
.002
N
13
13
13
13
13
v10t13
Correlation Coefficient
.282
.256
.718**
.667**
1.000
Sig. (2-tailed)
.180
.222
.001
.002
.
N
13
13
13
13
13
**. Correlation is significant at the 0.01 level (2-tailed).
*. Correlation is significant at the 0.05 level (2-tailed).
Thus, the system is very dynamic. J
As can be expected, the size-dependent ranking with the same parameters is much more stable. (Univ. Amsterdam 1, Utrecht and VU following at the second and third place, with the exception of EUR in the second place in 2006-2009).
Best,
Loet
_____
Loet Leydesdorff
Emeritus University of Amsterdam
Amsterdam School of Communication Research (ASCoR)
<mailto:loet at leydesdorff.net> loet at leydesdorff.net ; <http://www.leydesdorff.net/> http://www.leydesdorff.net/
Honorary Professor, <http://www.sussex.ac.uk/spru/> SPRU, University of Sussex;
Guest Professor <http://www.zju.edu.cn/english/> Zhejiang Univ., Hangzhou; Visiting Professor, <http://www.istic.ac.cn/Eng/brief_en.html> ISTIC, Beijing;
Visiting Professor, <http://www.bbk.ac.uk/> Birkbeck, University of London;
<http://scholar.google.com/citations?user=ych9gNYAAAAJ&hl=en> http://scholar.google.com/citations?user=ych9gNYAAAAJ&hl=en
From: ASIS&T Special Interest Group on Metrics [mailto:SIGMETRICS at LISTSERV.UTK.EDU] On Behalf Of David Wojick
Sent: Friday, May 22, 2015 4:02 AM
To: SIGMETRICS at LISTSERV.UTK.EDU
Subject: Re: [SIGMETRICS] CWTS Leiden Ranking 2015
Yes, Christina, there obviously are distant parts of the graph of science that that are not connected to other parts. But we are talking about drawing local lines. That is, saying that some local citation connections are part of a cluster while others are not.
My point is merely that there are lots of ways of drawing these local lines. All clusters are in that sense artificial constructs, based on algorithmic assumptions. My question is how sensitive are the university rankings to this particular construct? I have even suggested an alternative construct as a sensitivity test.
Mind you, I am assuming that the 4000 fields are a construct, not an empirical claim. If it is being claimed that it has been discovered that science actually consists of precisely 4000 micro fields then we have a very different discussion.
My best wishes,
David
Sent from my IPad
On May 21, 2015, at 7:08 PM, "Pikas, Christina K." <Christina.Pikas at JHUAPL.EDU> wrote:
“The citation network of science is seamless, or mostly so”
There are definitely components - parts of the graph that are not connected to other parts.
“Thus any segmentation of this seamless network into clusters must require algorithmic assumptions that are more or less arbitrary, in the sense that alternative assumptions are available. “
There are well-known and accepted community detection techniques and clustering techniques that work on networks where there is only one component. In community detection techniques, you typically maximize modularity, a measure of the extent nodes connect more to each other than to nodes not in the group. There are metrics for the other techniques, too. This is not new science at all. (fwiw, my paper using community detection for science blogs is OA archived here: http://terpconnect.umd.edu/~cpikas/ScienceBlogging/PikasEScience08.pdf - remainder of site is horribly out of date – don’t recommend reading it!)
The paper discusses how they used 12 (I think?) different levels of clustering before they decided on the one with 4000. I was at a meeting today (hi Chris, hi Nancy) so didn’t see these as they were sent so maybe someone has sent the paper already.
Christina
------
Christina K. Pikas
Librarian
The Johns Hopkins University Applied Physics Laboratory
Baltimore: 443.778.4812
D.C.: 240.228.4812
Christina.Pikas at jhuapl.edu
From: ASIS&T Special Interest Group on Metrics [mailto:SIGMETRICS at listserv.utk.edu] On Behalf Of David Wojick
Sent: Thursday, May 21, 2015 6:56 PM
To: SIGMETRICS at listserv.utk.edu
Subject: Re: [SIGMETRICS] CWTS Leiden Ranking 2015
Thank you, Nees Jan, but the paper is behind pay wall. Perhaps you could send me a copy, to dwojick at craigellachie.us.
In any case my thinking is this. The citation network of science is seamless, or mostly so. That is, there are few,if any, cases where there occurs a cluster such that all the authors therein cite each other and no one else. Thus any segmentation of this seamless network into clusters must require algorithmic assumptions that are more or less arbitrary, in the sense that alternative assumptions are available.
One wonders, therefore, to what extent the university rankings are sensitive to the specific assumptions made in order to cluster the citation network and create the fields? Have you tested this sensitivity?
I suggest a test. Change the algorithm such that it creates 2000 micro fields instead of 4000. Then rank the universities and see what difference it makes. Mind you this is just a crude first thought and better tests may be possible.
I do not question the value of the work, but as you know using metrics in this way is itself a sensitive issue.
My best regards,
David
On May 21, 2015, at 1:51 AM, "Eck, N.J.P. van" <ecknjpvan at CWTS.LEIDENUNIV.NL> wrote:
Dear David,
The 4000 fields are constructed using a clustering algorithm based on citation relations between publications. A detailed explanation is provided in the following paper: http://dx.doi.org/10.1002/asi.22748.
The clustering methodology for constructing the fields is fully transparent. The methodology is documented in the above-mentioned paper, and the computer software that is required to implement the methodology is freely available (open source) at www.ludowaltman.nl/slm/. It is true that the results produced by the clustering methodology are not transparent. The assignment of individual publications to the 4000 fields is not visible. As already mentioned, this is something that hopefully can be improved in the future. Please keep in mind that there is a growing consensus among bibliometricians that the use of the Web of Science subject categories for field normalization of bibliometric indicators is unsatisfactory and does not yield sufficiently accurate results. The normalization approach that is taken in the Leiden Ranking offers a more accurate alternative, but indeed the transparency of the Web of Science subject categories is lost.
Best regards,
Nees Jan
From: ASIS&T Special Interest Group on Metrics [mailto:SIGMETRICS at LISTSERV.UTK.EDU] On Behalf Of David Wojick
Sent: Wednesday, May 20, 2015 11:23 PM
To: SIGMETRICS at LISTSERV.UTK.EDU
Subject: Re: [SIGMETRICS] CWTS Leiden Ranking 2015
Dear Nees Jan,
How do you apply 4000 field categories to individual papers? A semantic algorithm? Is this explained on the website? It sounds very difficult.
Also if the categories are not visible how is the methodology transparent?
My best wishes,
David
http://insidepublicaccess.com/
At 04:06 PM 5/20/2015, you wrote:
Dear Loet,
Yes, your understanding is correct. MNCS, TNCS, PP(top 10%), P(top 10%), and the other field-normalized impact indicators all use the 4000 fields for the purpose of normalization. The Web of Science subject categories are not used.
Unfortunately, the 4000 fields are not visible. Because these fields are defined at the level of individual publications rather than at the journal level, there is no easy way to make the fields visible. This is something that hopefully can be improved in the future.
We have decided to move from 800 to 4000 fields because our analyses indicate that with 800 fields there still is too much heterogeneity in citation density within fields. A detailed analysis of the effect of performing field normalization at different levels of aggregation is reported in the following paper by Javier Ruiz-Castillo and Ludo Waltman: http://dx.doi.org/10.1016/j.joi.2014.11.010. In this paper, it is also shown that at the level of entire universities field-normalized impact indicators are quite insensitive to the choice of an aggregation level.
Best regards,
Nees Jan
From: ASIS&T Special Interest Group on Metrics [ <mailto:SIGMETRICS at LISTSERV.UTK.EDU> mailto:SIGMETRICS at LISTSERV.UTK.EDU] On Behalf Of Loet Leydesdorff
Sent: Wednesday, May 20, 2015 9:28 PM
To: SIGMETRICS at LISTSERV.UTK.EDU
Subject: Re: [SIGMETRICS] CWTS Leiden Ranking 2015
Dear Nees Jan,
As always impressive! Thank you.
Are the approximately 4,000 fields also visible in one way or another? Do I correctly understand that MNCS is defined in relation to these 4,000 fields and not to the 251 WCs? Is there a concordance table between the fields and WCs as there is between WCs and five broad fields in the Excel sheet?
I think that I understand from your and Ludo’s previous publications how the 4,000 fields are generated. Why are there 4,000 such fields in 2015, and 800+ in 2014? Isn’t it amazing that trends can despite the discontinuities be smooth? Or are indicators robust across these scales?
Best wishes,
Loet
Loet Leydesdorff
Emeritus University of Amsterdam
Amsterdam School of Communication Research (ASCoR)
loet at leydesdorff.net ; http://www.leydesdorff.net/
Honorary Professor, SPRU, <http://www.sussex.ac.uk/spru/> University of Sussex;
Guest Professor Zhejiang Univ. <http://www.zju.edu.cn/english/> , Hangzhou; Visiting Professor, ISTIC, <http://www.istic.ac.cn/Eng/brief_en.html> Beijing;
Visiting Professor, Birkbeck <http://www.bbk.ac.uk/> , University of London;
http://scholar.google.com/citations?user=ych9gNYAAAAJ <http://scholar.google.com/citations?user=ych9gNYAAAAJ&hl=en> &hl=en
From: ASIS&T Special Interest Group on Metrics [ <mailto:SIGMETRICS at LISTSERV.UTK.EDU> mailto:SIGMETRICS at LISTSERV.UTK.EDU] On Behalf Of Eck, N.J.P. van
Sent: Wednesday, May 20, 2015 8:27 PM
To: SIGMETRICS at LISTSERV.UTK.EDU
Subject: [SIGMETRICS] CWTS Leiden Ranking 2015
Release of the CWTS Leiden Ranking 2015
Today CWTS has released the 2015 edition of the Leiden Ranking. The CWTS Leiden Ranking 2015 offers key insights into the scientific performance of 750 major universities worldwide. A sophisticated set of bibliometric indicators provides statistics on the scientific impact of universities and on universities’ involvement in scientific collaboration. The CWTS Leiden Ranking 2015 is based on Web of Science indexed publications from the period 2010–2013.
Improvements and new features in the 2015 edition
Compared with the 2014 edition of the Leiden Ranking, the 2015 edition includes a number of enhancements. First of all, the 2015 edition offers the possibility to perform trend analyses. Bibliometric statistics are available not only for the period 2010–2013 but also for earlier periods. Second, the 2015 edition of the Leiden Ranking provides new impact indicators based on counting publications that belong to the top 1% or top 50% of their field. And third, improvements have been made to the presentation of the ranking. Size-dependent indicators are presented in a more prominent way, and it is possible to obtain a convenient one-page overview of all bibliometric statistics for a particular university.
Differences with other university rankings
Compared with other university rankings, the Leiden Ranking offers more advanced indicators of scientific impact and collaboration and uses a more transparent methodology. The Leiden Ranking does not rely on highly subjective data obtained from reputational surveys or on data provided by universities themselves. Also, the Leiden Ranking refrains from aggregating different dimensions of university performance into a single overall indicator.
Website
The Leiden Ranking is available at www.leidenranking.com.
========================================================
Nees Jan van Eck PhD
Researcher
Head of ICT
Centre for Science and Technology Studies
Leiden University
P.O. Box 905
2300 AX Leiden
The Netherlands
Willem Einthoven Building, Room B5-35
Tel: +31 (0)71 527 6445
Fax: +31 (0)71 527 3911
E-mail: ecknjpvan at cwts.leidenuniv.nl
Homepage: www.neesjanvaneck.nl
VOSviewer: www.vosviewer.com
CitNetExplorer: www.citnetexplorer.nl
========================================================
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://mail.asis.org/pipermail/sigmetrics/attachments/20150522/93f8afcd/attachment.html>
-------------- next part --------------
A non-text attachment was scrubbed...
Name: image004.png
Type: image/png
Size: 134073 bytes
Desc: image004.png
URL: <http://mail.asis.org/pipermail/sigmetrics/attachments/20150522/93f8afcd/attachment.png>
-------------- next part --------------
A non-text attachment was scrubbed...
Name: image002.png
Type: image/png
Size: 191799 bytes
Desc: image002.png
URL: <http://mail.asis.org/pipermail/sigmetrics/attachments/20150522/93f8afcd/attachment-0001.png>
More information about the SIGMETRICS
mailing list