Or how to go beyond a league table position in evaluating a UK university department
Every five years or so all UK university departments get their research assessed in one gigantic peer review exercise, which is now called the REF. Each discipline is assessed separately, and scores can be used to compile a league table. The exercise has direct financial implications: the better the research, the more money universities get from the government. But if you know what academics are like, you will not be surprised to learn that those in the UK obsess about this exercise and its results to a far greater extent than the money involved would justify. The results of the latest exercise have just been published, and turned into league tables by Times Higher Education (THE) here.
You could say that the REF now provides the same sort of incentive system for UK universities as profit does for a firm. In some cases academics whose research is below their departmental average are put under pressure to leave by one means or another, and most academics feel acutely the pressure to improve on how their own output will be assessed by this exercise. In contrast, poor performance in teaching or administration is not nearly such a serious issue.
Many academics complain bitterly about the indignity of all this. An alternative system would be one where getting tenure was the last performance hurdle an academic had to pass, and from then on they were free to do what they liked. Research money could all be allocated on a project by project basis. I personally doubt that would be a better system from society’s point of view, and I do find it annoying how academics can complain so much about pressures that are taken for granted elsewhere.
It would be a mistake, however, to think that the position in some REF league table told you all you needed to know to evaluate the quality of research in a department. The REF releases a wealth of data, and going beyond the headline number (usually the GPA score) can be informative. In the latest exercise departments were evaluated under three headings: outputs, environment and impact. Details about what is involved for each category can be found here.
Outputs, which has the highest weight in the total (65%), looks at the quality of the four best recent publications of each submitted member of staff. The key word to note here is ‘submitted’. A department/university can choose not to submit all its staff to the REF, and by not submitting staff that a department/university considers are well below average it can raise its GPA score (if it gets its assessments right). So to the extent that staff are not submitted, the GPA will overestimate the average quality of the research done in that department. As I said, league tables normally just look at the GPA score [1].
To some it may seem strange that this is allowed, but there are arguments to justify it. Departments do pay a significant financial penalty for leaving staff out - they only get money for submitted staff. To get a guide to the total amount of quality adjusted research done in a department, simply multiply the GPA score by the number of people submitted (called ‘power’ by THE).
The decision about whether not to submit a member of staff is an agonising one [2] that involves many difficult trade-offs. To the individual not being submitted it is a nasty slap in the face. For the department, the perceived benefits in getting a higher position in GPA based league tables may outweigh the financial cost of not submitting staff members. Decisions on this front do vary from university to university, and from department to department: in economics, compare the third and fourth columns of the table below.
Although it only counts for 15% of the total GPA score, the ‘environment’ heading may be of particular interest to potential PhD students. It is based on a number of different criteria, including the number of PhDs, the support provided for research, and research income from outside grants. Only three economics departments had all elements of environment judged to be of the highest (4*) quality this time: UCL, LSE and Oxford.
Impact is a new category, accounting for 20% of the total. It is based on case studies where research has engaged with public, private and third sector organisations, or directly with the public. For example, one of Oxford’s case studies for economics was my own work on fiscal councils. A quick look at the results suggests that this new element has had a significant influence on the overall results. In economics, for example, the only department where all the submitted case studies were judged to be of the highest quality was Bristol. So while Bristol only came 12th= on published outputs, a strong impact and environment score lifted them to 6th in the overall ranking.
As with any evaluation system, there are difficult judgements to make on the details, and these can lead to possibilities to ‘play the system’. Chris Bertram focuses on one particular issue at Crooked Timber. Each iteration of the assessment exercise attempts to change the details of the rules to avoid this, only to allow some new possibility to exploit the system. Partly as a result, after each exercise many academics feel that there must be a better and less time consuming way to judge the quality of research produced by individual academics or departments, but perhaps the fact that we keep returning to the same basic procedure suggests otherwise.
REF 2014 results: economics and econometrics
University | GPA Score | No. of staff submitted | Eligible staff | Power | % 4* Outputs | % 4* Environ. | % 4* Impact |
UCL | 3.78 | 37 | 45 | 139 | 70 | 100 | 92 |
LSE | 3.55 | 51 | 56 | 182 | 56 | 100 | 87 |
Oxford | 3.44 | 84 | 97 | 289 | 43 | 100 | 64 |
Cambridge | 3.42 | 27 | 38 | 92 | 55 | 13 | 50 |
Warwick | 3.41 | 42 | 52 | 142 | 43 | 38 | 60 |
Bristol | 3.32 | 19 | 25 | 62 | 22 | 63 | 100 |
Essex | 3.25 | 33 | 40 | 108 | 29 | 63 | 20 |
Edinburgh | 3.14 | 18 | 28 | 55 | 31 | 50 | 13 |
Royal Holloway | 3.11 | 14 | 23 | 45 | 35 | 0 | 60 |
Nottingham | 3.05 | 35 | 46 | 107 | 20 | 13 | 18 |
UEA | 3.04 | 14 | 22 | 43 | 20 | 0 | 20 |
Surrey | 3.01 | 21 | 25 | 62 | 27 | 13 | 0 |
Queen Mary | 2.98 | 24 | 31 | 73 | 20 | 13 | 13 |
York | 2.93 | 28 | 46 | 82 | 14 | 0 | 40 |
St. Andrews | 2.92 | 21 | 31 | 60 | 24 | 0 | 0 |
Manchester | 2.89 | 33 | 45 | 96 | 11 | 13 | 40 |
Glasgow | 2.86 | 24 | 30 | 68 | 18 | 0 | 0 |
Sussex | 2.84 | 17 | 24 | 49 | 15 | 0 | 37 |
Exeter | 2.78 | 25 | 31 | 68 | 13 | 25 | 13 |
Birmingham | 2.78 | 24 | 27 | 67 | 8 | 0 | 27 |
Southampton | 2.70 | 22 | 28 | 59 | 22 | 0 | 10 |
Birkbeck | 2.60 | 25 | 32 | 65 | 10 | 0 | 0 |
Leicester | 2.59 | 22 | 29 | 58 | 19 | 0 | 0 |
Sheffield | 2.58 | 15 | 26 | 38 | 8 | 0 | 40 |
Aberdeen | 2.48 | 19 | 26 | 48 | 5 | 0 | 0 |
City | 2.44 | 14 | 26 | 33 | 17 | 0 | 20 |
Kent | 2.32 | 22 | 26 | 51 | 3 | 0 | 13 |
Brunel | 2.20 | 26 | 29 | 58 | 2 | 0 | 0 |
Note that many economics departments are assessed under Business and Management, and are not included here. Sources: columns 2,3 and 5: Times Higher Education, column 4: HESA, columns 6-8, REF.
[1] THE publishes an alternative university wide ranking (aggregated across departments) that multiplies the GPA by the proportion of staff submitted, but that implicitly gives the research of non-submitted staff a score of zero, which is likely to be too extreme. It is better to simply note either the power score, or the proportion of staff submitted. Approximate information on the number of staff eligible for submission by department can be found here.
[2] This is based on my own experience at my previous university, where I was research director for the school. Thankfully I have played no role in these decisions at Oxford!
Artikel keren lainnya:
Belum ada tanggapan untuk "How to reference the REF"
Posting Komentar