• Aucun résultat trouvé

World university rankings: use of quantitative indicators and results from student surveys

N/A
N/A
Protected

Academic year: 2021

Partager "World university rankings: use of quantitative indicators and results from student surveys"

Copied!
6
0
0

Texte intégral

(1)

HAL Id: hal-01340322

https://hal.archives-ouvertes.fr/hal-01340322

Preprint submitted on 14 Jul 2016

HAL is a multi-disciplinary open access archive for the deposit and dissemination of sci- entific research documents, whether they are pub- lished or not. The documents may come from teaching and research institutions in France or abroad, or from public or private research centers.

L’archive ouverte pluridisciplinaire HAL, est destinée au dépôt et à la diffusion de documents scientifiques de niveau recherche, publiés ou non, émanant des établissements d’enseignement et de recherche français ou étrangers, des laboratoires publics ou privés.

World university rankings: use of quantitative indicators and results from student surveys

Philippe Vidal

To cite this version:

Philippe Vidal. World university rankings: use of quantitative indicators and results from student

surveys. 2016. �hal-01340322�

(2)

World University rankings:

use of quantitative indicators and results from student surveys Philippe Vidal

OST du HCERES, France

Université Blaise Pascal, Clermont-Ferrand, France

Summary : World University rankings display their results as league tables (ARWU, THE, QS) or as performance groups (U-Multirank). Although a priori the latter is more satisfactory, it does have serious limitations. Alternative propositions are explored.

Introduction

One major criticism of league tables is that they tend to emphasise the differences in ranks. As an example, in the ARWU ranking, where Harvard university is the reference (global composite indicator

=100), the ranked number 2 university has a score around 70, which makes a difference of 30. In contrast, the difference between universities ranked 50 and 51 is around 0.5 and between universities ranked 99 and 100 the difference is around 0.1. That is why all the league table rankings display the data, after the rank 100, by rank groups where the Higher Education Institutions (HEIs’) are listed in alphabetic order (for ARWU, 100-150, 150-200, 200-300…). In contrast, U-Multirank (UMR) and the CHE University-ranking on which it is based, display the HEI’s results in the format of groups, where the HEIs’ are also listed in alphabetic order. So, unlike ARWU etc. rankings, where the order is simply given by the rank of the HEIs’

global composite indicators, the CHE-UMR group construction is based on a more complex methodology.

A feature of CHE-UMR rankings is student satisfaction surveys. These, like other indicators are presented in the form of groups which are based on a rather complex methodology.

The aim of this discussion is to demonstrate the flaws of those methodologies and to present alternative possibilities.

I. Interpretation of quantitative indicators

Using the distribution of indicator values, the UMR method produces a spread of points which are divided into 5 groups of uneven size. The principle is to plot the indicator values as y and the HEIs as x, and from this identify the x median value from which the corresponding y median is determined graphically. The guidelines for subdividing the values into groups is as follows:

Group A: where the indicator value is at least 25% above the median indicator.

Group B: where the indicator value is between the median indicator and 25% above the median indicator Group C: where the indicator value is between the median indicator and 25% below the median indicator Group D: where the indicator value is at least 25% below the median indicator

Group E: where the indicator value = 0

A full description of the methodology can be found in:

www.umultirank.org/cms/wp-content/uploads/2016/03/Rank-group-calculation-in-U-Multirank-2016.pdf .

This method would make sense if all the data points plotted on a straight line, whatever its slope value. As

seen in the following example, this situation is not what is observed when real data are used.

(3)

Exercise carried out on the real data

The UMR methodology described above (but where Group D is discarded) was applied to a set of data from about fifty French HEIs’ (Fig. 1). The test is focused on eight indicators (Citation rate, External research income, etc.) each of which is divided into 4 categories as indicated by color differences in Fig. 1.

The figure shows that there is a considerable size variation between the groups. For example, compare the size distribution of the groups Citation rate and External research income in Fig.1. In addition, since in the majority of cases there are no obvious breaks in the data spread, the choice of divisions between adjacent categories becomes rather arbitrary, being based on a strict arithmetic exercise rather than clear differences. As a consequence, some HEIs’ can either profit or be penalized depending on minor data differences. The most extreme example is the "Master graduation on time" category, in which the bulk of the dominating (green) group have indicator values close to 100% resulting in almost all of this HEIs’

grouping being classified as B. Because of this, there are no HEIs’ in group A.

0

50 100 150 200 250

0 10 20 30 40 50

External research income

Red = A; Green = B ; Black = C ; Blue = D

0 2 4 6 8 10 12 14 16 18

0 10 20 30 40

Patents awarded (size-normalized)

Red = A; Green = B ; Black = C ; Blue = D 0

0,2 0,4 0,6 0,8 1 1,2 1,4 1,6 1,8 2

0 20 40 60

Citation rate

Red = A; Green = B ; Black = C ; Blue = D

0 20 40 60 80 100 120

0 10 20 30 40 50

Graduating on time masters

Red = A; Green = B ; Black = C ; Blue = D

0 2 4 6 8 10 12 14 16

0 10 20 30 40 50 60

Interdisciplinary publications

Red = A; Green = B ; Black = C ; Blue = D

0 5 10 15 20

0 20 40 60

Copublications with industrial partners

Red = A; Green = B ; Black = C ; Blue = D

0 20 40 60 80 100 120

0 10 20 30

Graduating on time bachelors

Red = A; Green = B ; Black = C ; Blue = D 0

2 4 6 8 10 12 14 16 18 20

0 20 40 60

Red = A; Green = B ; Black = C ; Blue = D Top cited publications

Figure 1 Data from a set of French HEIs. The indicator values are reported as y. The letters A, B etc. correspond to the U-Multirank grouping methodology.

Figure 1: Data from a set of French HEIs’. The indicator values are reported as y. The letters A,

B etc. correspond to the U-Multirank grouping methodology.

(4)

Figure 2: Positioning (red dots) for 4 indicators (Bachelor graduation rate, Master graduation rate, Bachelors: graduating on time, Master: graduating on time) of Université Catherine de Parthenay (fictitious) Department of Psychology regarding the other Departments of Psychology (from demo- ceres.obs-ost.fr).

Conclusions

Although the CHE national ranking scheme, which is based on 3 groups (Top group with 25% of HEIs’, Medium with 50, and Bottom with 25) is a more satisfactory and fairer method than the 5 grouping of UMR, both grouping methodologies give the illusion of the existence of very different specific groups, each having their own characteristics. In fact the reverse is true since most of the data form a continuum.

A minimal improvement of the methodology would be to create groups only when there is a minimum gap between the extreme values (10%, 20%, more?).

As an alternative it should be possible to show in graphic form the position of each HEI with respect to their position for each indicator. An example is shown in Fig. 2.

II. Interpretation of results from student surveys.

Student surveys are based on a series of questions where the student gives an opinion based on a Likert scale of 1 (best) to 6. Given that the student answers reflect their individual opinion rather than

quantifiable measurement it is reasonable to ask how their answers can be presented and evaluated in a logical and rigorous way.

Ranking by groups: The CHE-UMR Method

An important aspect of the CHE-UMR evaluation of student satisfaction surveys is their use of groups to subdivide and grade student answers (see www.umultirank.org/cms/wp-content/uploads/2016/03/Rank- group-calculation-in-U-Multirank-2016.pdf).

Faced with the problem that the spread of student answers within an individual HEI is generally

substantially higher than the spread of other HEIs’ average values, the authors of CHE-UMR solve the

problem by assuming a real value for each HEI. From this they calculate the measurement error using

(5)

conventional metrology methods, which attempts to estimate, based on a limited number of measurements, the error bar in which the real value (length, weight etc.) is located.

From this a confidence interval stated at the 95% confidence level is calculated. Its value is given by 2 /( √ n) (where  is the standard deviation and n is the number of measurements. The standard deviation should be actually multiplied by the Student-Fisher coefficient, which rapidly becomes very close to 2 for n> 10). It follows that dividing the standard deviation by √n produces an error bar that is unreasonably short. Clearly, this conventional technique, which minimizes the real scatter of the response values, is appealing since it produces groups more homogeneous in appearance, which in turn are easier to use in a ranking system.

In brief, given that the answers of the students reflect the scatter of their opinions and that by definition there are no measurement errors, the use of a statistical evaluation approach is misleading and gives the user a false sense of reality.

The CHE-UMR methodology outlined above has two additional major drawbacks:

1) it is built on relative positioning and therefore the groups are created regardless of the degree of scattering in the distribution. When the data are highly variable, it is not too difficult to define classes but when data are rather homogeneous, grouping is more difficult to justify.

2) the method may not identify a top group when most of the answers are largely positive.

Alternative rankings

The problem posed by student surveys can be approached in other ways. For example:

a) Order by average: Penn State and Ohio universities

http://studentaffairs.psu.edu/assessment/satisfaction.shtml

This classification method has the great disadvantage of not considering the dispersion around the average value (Fig.1). However the reports do display all the Likert scales data, which could be the basis for a better use (see below).

b) Order by positive responses: University of Geneva

Only the % of satisfactory answers higher than 80% is taken into account. Here again there is some arbitrariness in the division into classes above and below 80%.

c) National Survey of Student Engagement: USA http://nsse.indiana.edu

Here the approach is to use quartiles. The raw data are presented on box plots and further evaluated using comparisons of average values with (1) the overall average (2) the best 10% (3) 50% better. This is an interesting exercise but not easy to use.

d) Order by % of positive responses: United Kingdom http://unistats.direct.gov.uk

Based on annual National Student Surveys, this method uses the % contributed by the top two groups of the Likert scale (i.e. groups 1+2). This presentation is easy to understand and does not suffer any of the flaws described above.

Student surveys: Conclusions

The methodology used in UMR is unsatisfactory because it uses a statistically invalid assumption and

then minimizes the dispersion of the answers. The methodology requires creating groups which

ignores the situation where there are no clear breaks in the data. The British methodology is clearer

(6)

and the easiest to implement. It can be completed by reporting all the data (including average values, median, standard deviation) in a separate document.

General Conclusions

The use of formal groups by CHE-UMR when evaluating performance indicators for HEIs (including student satisfaction surveys) is not the most appropriate method when dealing with the matter. The danger is that it may present an erroneous picture of the real distribution of the data. One underlying question is that UMR and CHE rankings are basically based on relative, and not absolute, HEIs positioning.

Besides not being statistically pertinent, this, in turn tends to distort the resulting rankings since this can (1) create artificial boundaries within sets of rather homogeneous data, and (2) create quality scores which do not match reality.

Acknowledgements

The author thanks Ghislaine Filliatreau and Robert W. Nesbitt for their helpful comments, and Marie-Laure Taillibert for her appreciable technical help.

Philippe Vidal, philippe.vidal@obs-ost.fr and phvidal@orange.fr

Références

Documents relatifs

We prove that an extreme Kerr initial data set is a unique absolute minimum of the total mass in a (physically relevant) class of vacuum, maximal, asymptotically flat, axisymmetric

If the survey focus is household work, leisure, personal time, or occupational time where social control over the length of that time is weak (time used by teachers to prepare

Dans les enquêtes emploi du temps où les deux conjoints ont rempli un carnet, il est possible d’aller au-delà de l’étude des interdépendances entre les durées moyennes

L’archive ouverte pluridisciplinaire HAL, est destinée au dépôt et à la diffusion de documents scientifiques de niveau recherche, publiés ou non, émanant des

(2007) show that neighbourhood is an important driver of the location of organic fields at the plot level. Based on such a plot level analysis, it could be interesting

‰ Village leaders not accountable for all villagers.. How to mitigate the vertical constraints?. „ Step by step towards a dialogue between villagers & National

Finally, sections 4, 5, 6, 7 and 8 are the proofs respectively of Theorem 2.1 (about the strong semigroup property of the linear part of equation (1.3)), Proposition 2.2 (an a

In particular we note the following journals: (i) Journal of the Associa- tion for Information Science and Technology by Wiley, (ii) Scientometrics by Springer, (iii)