Saturday, April 23, 2016

America's Sleepiest Colleges



Babies, Three, Sleep, Eyes Closed
A ranking from JAWBONE reproduced at THE Tab lists American colleges and universities according to the hours of sleep that students get. Apparently there is a significant correlation between the ranking of the institution according to the US News and late weekday bedtime.

Here are the top ten universities according to the number of hours slept per night (inverting the JAWBONE list).

1.   Colorado
2=  Tulane
2=  Vermont
4.    Auburn
5.    Arkansas
6=  South Carolina
6=  Iowa
8=  Brown
8=  Florida State
8=  Clemson

Please do not share this too much or demented bureaucrats will start incorporating caffeine injections into their rankings strategy.

Friday, April 22, 2016

Aussies are falling out of love with THE



Matthew Knott in the Sydney Morning Herald reports that Australian education experts are no longer impressed by the Times Higher Education World University Rankings.

"Grattan Institute higher education program director Andrew Norton says The Times rankings are not "terribly high quality".

"They should not be used as a guide for which university to go to and they shouldn't be used as a guide to higher education policy," he says.

In particular, he warns that movements up or down the league table – especially small ones – should not be used a reliable verdict on whether a university is improving or declining.

And he's not alone.

Australian higher education academic Simon Marginson, one of the leading experts on university rankings, is even more damning.

"In social science terms they are rubbish," he told an academic conference last year."

THE will probably not be bothered too much. After all, when you have been declared "education secretary of the world" in China, who cares if an Australian journalist compares you to dead sticks? 

No more excellence at the University of Missouri




Image result for free image university missouri


No, it is not a new initiative to identify more inclusive measures of achievement.


From Peter Woods at Minding the Campus:

"The University of Missouri has eliminated Respect and Excellence.  I have to write this in a hurry because it won’t be long before others will seize on this gift.  Respect and Excellence are the names for two residence halls at the University.  They are being closed because the University suddenly finds that its enrollments are plummeting.  Two other dorms were closed already in light of the crisis.
Let’s bask in the irony for a moment or two longer.  The University of Missouri arrived at this juncture by cravenly submitting to the demands of activists and the threats of football players who decided to abet the activists.  On November 9, System President Tim Wolfe and Chancellor R. Bowen Loftin resigned rather than face down those threats."
There is definitely something wrong with the administration if they did not anticipate the reaction to this.

Trinity College Dublin: A Case of Rankings Abuse



Trinity College Dublin (TCD) is giving itself a public flogging over its fall in the rankings. This is rather odd since in fact it has been doing pretty well over the last few years with one exception.

An article in the Irish Times reveals that the Provost of TCD, Patrick Prendergast, is planning a new business school and student accommodation but finds it difficult to raise the necessary cash.


"The college’s planned expansion comes at a time when many in higher education say the sector faces a funding crisis. Rising student numbers, declining state funding and restrictions on staff recruitment mean that many have had to make dramatic cuts to make ends meet.
Trinity and some of the bigger universities have at least been able to plug many of the funding gaps with private income, such as international students’ fees, research and other commercial sources."
According to the provost things are getting pretty rough.
“We have overcrowded classrooms. Our staff-student ratio is a way out of kilter. The universities have been very resilient; they have managed to keep going successfully. But is it sustainable? I and other university presidents don’t think it is sustainable at current funding levels.”
What does this have to do with rankings.?
"University rankings are considered vital to attracting international students and research funding. However, Trinity, like many Irish universities, has slid down world rankings in recent years as it copes with an increase in students and a reduction in funding.
Last month, TCD found itself at the centre of controversy when one of the main ranking agencies, QS, accused the college of violating its rules by influencing academics involved in its annual survey."
The provost claims:


“If Ireland really wants to be an island known for the talent of its people, and have companies locate here, then we can’t afford to have that one global indicator of the quality of education systems – rankings – decline."

But is it true that TCD is sliding down the rankings?

Let's take a look at the Shanghai Academic Ranking of World Universities (ARWU). This measures research in the natural and social sciences at the highest level. The methodology of these rankings has remained largely stable with some minor exceptions.  In 2004  there were some changes to help social science institutions a bit and in 2014 they began to use a new list of highly cited researchers. The latter move helped TCD a bit in 2014 and, assuming the highly cited researchers do not leave, will help a bit more later this year. 

Therefore, if there has been a significant change in scores in these rankings it is a reasonable assumption that this does actually reflect a change in quality.

Looking at the scores for the separate indicators in the Shanghai rankings, the score for alumni with Nobel and Fields awards fell from 15.4 to 10.3, for faculty with awards from 14.4 to 13.3, and for papers in Nature and Science very slightly from 13.2 to 13.1 between 2004 and 2015.

In contrast, TCD did much better for highly cited researchers (from zero to 12.3), publications in Web of Science journals (27.1 to 31.0) and productivity per capita (the sum of the above indicators divided by number of faculty). For all criteria, the top scoring university, Caltech for productivity and Harvard for everything else, gets 100.

Overall, TCD moved from the 202-300 band to the 151-200 band. A rough guess is that TCD has moved up  about 15 places altogether. About five of these places were gained because of the new highly cited researchers list. At this rate it could be in the top 100 in another century or so. That is very long term but as Basil Fawlty said, "these things take time".

Of course, the Shanghai rankings do not measure teaching, quality of students, income or internationalisation. For a detailed look at a more diverse set of criteria we can turn to the Round University Rankings. These are published by a Russian company but use data from Thomson Reuters, the same source that Times Higher Education (THE) used from 2010 to 2014.  There are, however, 20 indicators compared to the THE's 13 and they cover the period from 2010 to 2016.

Overall, TCD improved significantly, going from 174th to 102th place.

In the teaching dimension, TCD rose slightly from 207th place to 197th, doing slightly better for two doctoral degree indicators, quite a bit worse for two academic staff indicators and staying in exactly the same place for teaching reputation (187th).

For research, TCD improved significantly from 2010 to 2016, going from 193rd to 67th, did much better for papers, citations, completed doctoral degrees and research reputation but did worse for normalised citation impact.

TCD did very well for international diversity in 2010 (31st) but slipped back a bit by 2016 (46th).

For financial sustainability, TCD 's relative position worsened considerably falling from 229th to 412th.

So the data from RUR suggests that TCD's income and faculty resources  were declining relative to other universities over this period.  But so far this has had no significant effect on TCD's teaching profile while research has improved noticeably both in quality and quality.

TCD could quite plausibly claim to be the university with a tiger in its tank, getting more research and educating more students with limited financial and faculty resources.

Turning to the QS scores, in 2004 TCD was ranked 87th in the world and by 2014 had risen to 71st place, but fell back to 78th in 2015. They were still doing better than in 2016. It is not a good idea to draw any conclusions from the decline between 2014 and 2015 because in 2015 QS introduced a number of substantial methodological changes.

To summarise, TCD were doing better over fairly long periods in  the Shanghai,  RUR and QS rankings. Possibly, the fall in the QS rankings between 2014 and 2015 might portend difficulty ahead and perhaps the fall in income documented in RUR may eventually have a knock on effect although so far it has not. Still, it seems that  TCD has on the whole  been doing well. So what  is the Provost talking about?

It is the THE rankings that appear to show TCD in a bad light. In 2010 TCD was in 76th place overall and by 2015 had fallen to 160th. It is difficult to tell exactly what happened because 11 of the 13 indicators are bundled into three super-indicators and it is not clear exactly what contributed to rising or falling scores. In 2015 TCD had higher scores for  international orientation and lower scores for Teaching, Research, Citations and Income from Industry. The biggest  decline was in the Research score, from 45.3 to 30.8.

Clearly, THE is the odd man out as far as TCD is concerned and should not be taken too seriously. Firstly, there were major methodological changes last year which produced upheavals for universities around the world, including those in France, Korea and Turkey. There was another batch of changes in 2011. In addition, these rankings generate a lot of noise because of exchange rate fluctuations, the use of surveys which can be quite volatile outside the top fifty or so, and a citations indicator where (until last year) a single paper or adjunct faculty could produce an enormous change in scores.

THE have said  -- and here they must be given credit -- that:  "Because of changes in the underlying data, we strongly advise against direct comparisons with previous years’ World University Rankings."

It seems that TCD is doing the academic equivalent of taking a dive.

Monday, April 18, 2016

Off Topic: The Independent Gets Really Creative

The Independent has just posted a creativity test composed of one question. I can remember this question from a non-examination social studies class in grammar school several decades ago.

Surely the voice of the British intelligentsia can be more creative than that.

Round University Rankings


The latest Round University Rankings have been released by the Russian company, RUR Rankings Agency. These are essentially holistic rankings that attempt to go beyond the measurement of research output and quality. There are twenty indicators, although some of them such as Teaching Reputation, International Teaching Reputation and Research Reputation and International Students and International Bachelors are so similar that the information they provide is limited.

Basically these rankings cover much the same ground as the Times Higher Education (THE) World University Rankings. The income from industry indicator is not included but there are an additional eight indicators. The data is taken from Thomson Reuters' Global Institutional Profiles Project (GIPP) which was used by THE for their rankings from 2010 to 2014.

Unlike THE, which lumps its indicators together into groups,  the scores in the RUR are listed separately in the profiles. In addition, the rankings provide data for seven continuous years from 2010 to 2016. This provides an unusual opportunity to examine in detail the development of universities over a period of seven years, measured by 20 indicators. This is not the case with other rankings which have fewer indicators or which have changed their methodology.

It should be noted that participation in the GIPP is voluntary and therefore the universities in each edition could be different. For example, in 2015 100 universities dropped out of the project and 62 joined.

It is, however,  possible to examine a number of claims that have been made about changes in university quality over the last few years. I will  take a look at these in the next few posts.

For the moment, here are the top five in the overall rankings and the dimension rankings.

Overall
1.   Caltech
2.   Harvard
3.   Stanford
4.   MIT
5.   Chicago


Teaching
1.   Caltech
2.   Harvard
3.   Stanford
4.   MIT
5.   Duke

Research
1.   Caltech
2.   Harvard
3.   Stanford
4.   Northwestern University
5.   Erasmus University Rotterdam

International Diversity
1.   EPF Lausanne
2.   Imperial College London
3.   National University of Singapore
4.   University College London
5.   Oxford

Financial Sustainability
1.   Caltech
2.   Harvard
3.   Scuola Normale Superiore Pisa
4.   Pohang University of Science and Technology
5.   Karolinska Institute

Unfortunately these rankings have received little or no recognition outside Russia. Here are some examples.


MIPT entered the top four universities in Russia according to the Round University Ranking

Russian Universities in the lead in terms of growth in the international ranking of Round University Ranking

TSU [Tomsk State University]  has entered the 100 best universities for the quality of teaching

[St Petersburg]

Russian universities to top intl rankings by 2020 – Education Minister Livanov to RT


Saturday, April 16, 2016

Some fairly new news from THE

The latest of many spin offs from the THE world university rankings is a list of 150 universities less than 50 years old. The data is extracted from the world rankings released at the end of last year. THE have, however, reduced the weighting given to their academic reputation survey and so the results are a little different.

Is it possible that THE are thinking about reducing the weighting for the reputation survey in this year's world rankings?

Here are the top ten new universities with their overall scores and then the overall scores in the World University Rankings in brackets. It can be seen that changing the weighting for this indicator does in some cases make a difference, although not a spectacular one.


Place
University
Overall score

1Ecole Polytechnique Federale Lausanne
Switzerland
76.8  (76.1)   
2Nanyang Technological University
Singapore
72.5  (68.2)
Hong Kong University of Science and Technology70.7  (67.2)
4Maastricht University
Netherlands 
66.1  (59.9)
5Pohang University of Science and Technology
Korea
65.5  (56.9)
6Korean Advanced Institute of Science and Technology60.8  (53)
7University of Konstanz
Germany
58.9  (50.8)
8Karlsruhe Institute of Technology
Germany
58.6 (54.5)
9Pierre and Marie Curie University
France
58.2  (57)
10Scuola Superiore Normale Pisa
Italy
57.3  (50.2)

Most of the 150 universities were outside the top 200 in the world rankings and did not receive an overall score (although it could be calculated easily enough) so there is some new data here. 

As expected the young universities rankings have received a lot of publicity from the world media. Here's a sample.

University of Calgary ranks again as a top young global university by Times Higher Education










Thursday, April 14, 2016

Are there any more rankings left?

There seems to be an unending stream of new rankings. So far we have had from the big three or four subject rankings, field rankings, European, Asian, African, Latin American, Middle East and North Africa rankings, BRICS rankings. BRICS and emerging economies rankings, reputation rankings, young universities, old universities, most international universities rankings, research income from industry rankings.

From outside the charmed triangle or square we have had random rankings, length of name rankings,green rankings, twitter and LinkedIn rankings and rich universities rankings and of course in the USA a mixed bag of best universities for squirrels, gay friendly, top party schools and so on. I am a keen follower of the latter: when the US Air Force Academy gets in the top ten I shall pack up and move to Antarctica.

So are there any more international university rankings in the pipeline?

A few suggestions. Commonwealth universities, OIC universities, cold universities, high universities (altitude that is), poor universities, fertile universities (measured by branch campuses).

One that would be fun to see would be a Research Impact ranking based on those universities that have achieved a top  placing in the THE year- and field- normalised, regionally modified, standardised, citations ranking.

Some notable inclusions would be St. George's University of London, Rice University, Tokyo Metropolitan University, Federico Santa Maria Technical University, Florida Institute of Technology, National Research Nuclear University MEPhI and Alexandria University.



Monday, April 11, 2016

Ranking Rankers' Twitter Accounts

Over at Campus Morning Mail, Steven Matchett reflects on the limited attention that U-Multirank has received compared with THE's ranking of 150 young universities. He notes that THE has 192,000 followers on Twitter, while U-Multirank has only 2,390.

Digging further, here are the number of twitter followers for various people and organisations that have something to do with international university rankings.


World Uni Ranking (THE) 25,400
Phil Baty (THE) 15,300
QS 3,667
Bob Morse (US News) 2,122
Isidro Aguillo (Webometrics) 2,221
Ellie Bothwell, (THE) 2,049
World Univ Ranking (Shanghai) 1,226
Ben Sowter (QS) 1,115
CWTS Leiden 641
Proyecto U-Ranking (Spain) 561
RUR ranking (Russia)   532
5 top 100 (Russia)  355
Centre for Higher Education (Germany) 308
Richard Holmes 175

Sunday, April 10, 2016

Interview with Round University Ranking

See here for the text of an interview with Round University Ranking of Russia.


How to Survive Changes in Ranking Methodology

How to survive changes in ranking methodology

Thursday, April 07, 2016

How to Read International University Rankings


The London Daily Telegraph has published an article on 'How to Read the Different University Rankings' which refers to two international rankings, QS and THE and their spin-offs and some national ones. I assume this was intended for British students who might like to get a different perspective on UK university quality or who might be thinking of venturing abroad.

The article is not very satisfactory as it refers only to the QS and THE rankings and is uncritical.

So here is a brief survey of some global rankings for prospective students.

International university rankings fall into three groups: Internet-based like Webometrics and ic4u, research-based like the Academic Ranking of World Universities (ARWU) produced by the Shanghai Ranking Consultancy and "holistic" rankings such as the Quacquarelli Symonds (QS) and Times Higher Education (THE) world rankings which claim to measure reputation, teaching quality, internationalisation and other factors and combine indicators into a single index.

The first thing that you need to know is where an institution stands in the global university hierarchy. Most rankings cover only a fraction of the world's higher education institutions. Webometrics, however, rates the presence, impact, openness and research quality, measured by Google Scholar Citations, of nearly 24,000 universities. Only 5,483 of these get a score for quality so that a university with a rank of 5,484  is not doing anything resembling research at all.

While the Webometrics rankings are quite volatile at the top, the distinction between a university in 200th and one in 2,000th place is significant, even more so between 1,000th and 10,000th.

Webometrics can also help determine whether an institution is in fact a university in any sense of the word. If it isn't there the chances are that it isn't really a university. I recently heard of someone who claimed degrees from Gordon University in Florida.  A search of Webometrics revealed nothing so it is very likely that this is not a reputable institution.

Here ic4u is less helpful since it ranks only 11,606 places and does not provide information about specific indicators.

Looking at the research rankings, the oldest and most respected by academics is the ARWU published by Shanghai Ranking Consultancy. This has six criteria, all of them related to research. The methodology is straightforward and perhaps unsophisticated by today's standards. The humanities are excluded, it has nothing to say about teaching and it favours large, old and rich universities. It also privileges medical schools, as shown by last year's world rankings which put the University of California at San Francisco, a medical school, in eighteenth place in the world.

On the other hand, it is stable and sensible and by default the ranking of choice for research students and faculty.

ARWU also publishes subject rankings which should be checked. But it is worth noting that ARWU has proved vulnerable to the tactics of King Abdulaziz University (KAU) in Jeddah which hands out contracts to over 200 adjunct faculty members who list KAU as a secondary affiliation. This has put the university  in the world's top ten in the ARWU mathematics rankings.

There are a number of other global research rankings that could be consulted since they produce results that may be different from Shanghai. These include the National Taiwan University Rankings, the Best Global Universities published by US News, University Ranking by Academic Performance produced by Middle East Technical University and the CWTS Leiden Ranking. These, especially the Leiden Ranking, reach a high level of technical sophistication and should be consulted by anyone thinking of postgraduate study anywhere. The Leiden Ranking is very helpful in providing stability intervals for each score.

Since 2004 a number of rankings have appeared that attempt to go beyond simply measuring research output and quality. These are problematical in many ways. It is very difficult to find data that is comparable across international borders and their methodology can change. In addition, they rely on reputation surveys which can be biased and unreliable. At the moment the two best known international rankings of this type are the World University Rankings published by QS and THE.

It must be pointed that even these are still top heavy with research indicators. The QS rankings have a 40% weighting for an opinion survey of research quality, and another 20 percent for citations. Even the faculty student ratio indicator is sometimes a measure of research rather than teaching since it can be improved by adding research-only staff to the faculty. The THE rankings allot 30% to five research indicators and another 30% to citations, 2.5% to international research collaborations and 2.5% to research income from industry.

You should bear in mind that these rankings have been accused, not without justification, of national bias. A paper by Christopher Claassen of the University of Glasgow has found that the QS and THE rankings are seriously biased towards UK universities.

The metrics that do attempt to measure teaching quality in the THE and QS rankings are not very helpful. Both have faculty student ratio data but this is a very imprecise proxy for teaching resources. The THE rankings include five indicators in their super-indicator "Teaching: the Learning Environment", two of which measure the number of doctoral students or doctoral degrees, which does not say much about undergraduate instruction.

It is also a good idea to check the scores for the criteria that are combined to make up the composite score. If a university has a disproportionately high score for an indicator with a high weighting like QS's academic opinion survey (40%) or THE's citations indicator (30%) then alarm bells should start ringing.

In some ways the new Round University Ranking from Russia is an improvement. It uses data from Thomson Reuters as THE did until last year. It does nearly everything that THE and QS do and a few more things besides. Altogether there are 20 indicators, although some of these, such as three reputation indicators,  are so similar that they are effectively redundant.

Recently a consortium of European organisations and universities created U-Multirank, which takes a different approach. This is basically an online evaluation tool that allows users to choose how they wish to compare and sort universities. Unfortunately, its coverage is rather uneven. Data about teaching and learning is limited for Asia, North America and the UK although good for Western
Europe.

International rankings are generally not helpful in providing information about graduate employability although QS do include a reputation survey of employers and the Jeddah based Center for World University Rankings counts alumni who become CEOs of major companies.

The main global rankers also publish regional and specialist spin-offs: Latin American, Asian and European rankings, new university rankings, subject rankings. These should be treated with scepticism since they depend on a relatively small number of data points and consequently can be unreliable.

To summarise, these are things to remember when using global rankings:

  • first check with Webometrics to find out the approximate standing of a university in the world
  • for prospective postgraduate students in science and medicine, the Shanghai rankings should be consulted and confirmed by looking at other research rankings
  • potential undergraduate students should look at more than one "holistic" ranking and should always check the scores for specific indicators
  • be very sceptical of universities that get a good score for only one or two indicators or that do well in only one ranking
  • always look at scores for subjects if available but remember that they may be based on limited and unreliable data
  • for teaching related information the best source is probably U-Multirank but this is available only for some universities.







Friday, April 01, 2016

Really Useful Rankings: The Lowest Ranked Universities in the World

University rankings are getting a bit predictable, especially at the top. The top university is either in California or Cambridge, Massachusetts (or once in another Cambridge somewhere). The top twenty will be nearly all US places with a few British. Some of the established rankers do try to liven things up a bit: Nanyang Technological University in the top twenty, Middle East Technical University in the top hundred, but that does not do much to relieve the tedium.

So I thought it might be interesting to find the lowest ranked universities in the world. The trouble is most rankings do not rank many institutions so the best place to look is Webometrics which tries to rank all the universities on the planet. Here are the 22 institutions ranked 23,892 by Webometrics which means they have no Presence, no Impact, no Openness and not one Google Scholar Citation. Whether their existence goes beyond a website to include buildings and people I'll leave for you to find out.

Instituto Oswaldo Cruz de Certificacao, Brazil
Faculdade des Americas, Brazil
Escuela Normal Prescolar Adolfo Viguri Viguri, Mexico
Universidad Continente Americano Celaya, Mexico
Colegio Universitario y Tecnologico del Noreste, Mexico
Benemerita Escuela Normal Federalizada de Tamaulipas, Mexico
Universidad Virtual Latinoamericano, Venezuela
Faculdade Zumbi dos Palmares, Brazil
Faculdade Vizcaya, Brazil
Mugla Vocational Higher School, Turkey
Higher Polytechnic School Zagreb, Croatia
Sadat Institute of Higher Education, Afghanistan
Saint Paul Institute, Cambodia
Zhangzhou Institute of Technology, China
Technological University Thanlyin, Myanmar
Abasyn University, Pakistan
Salipur College, India
Adusumilli Vijaya Group of Colleges, India
Kaboora Institute of Higher Education, Afghanistan
Madina Engineering College, India
Universidade Lueji A'Nkonde, Angola
Reseau Marwan, Morocco