360 likes | 496 Views
Development and Approbation of a Template Methodology for National Ranking of Higher Education Institutions: 2011 outputs and next steps. Larionova Marina ( NTF ) Perfilieva Olga ( NRU HSE International Organizations Research Institute )
E N D
Development and Approbation of a Template Methodology for National Ranking of Higher Education Institutions: 2011 outputs and next steps Larionova Marina (NTF) Perfilieva Olga (NRU HSEInternational Organizations Research Institute) Project “Development and Approbation of a Template Methodology for National Ranking of Higher Education Institutions” 2011-2013
I. Project background Project “Development and Approbation of a Template Methodology for National Ranking of Higher Education Institutions” 2011-2013 implemented by National Training Foundation in collaboration with International Organizations Research Institute of the National Research University Higher School of Economics at the request of the Russian Ministry of Education and Science.
Project goal To develop and approbate a template methodology for ranking of Russian higher education institutions through: • comparative analysis of global, national and specialized rankings; national approaches to evaluation of higher education institutions performance • public and expert discussions of the draft methodology • approbation of the draft methodology • processing and discussion of the approbation outcomes • consultations with IREG experts
Project tasks Analyzing the Russian approaches (methodologies and indicators)used to evaluate performance of higher education institutions. Conducting a comparative analysis of global, national, and specialized rankings; identifying their strengths and weaknesses. Carrying out a comparative analysis of the methodologies and approaches used in international /foreign and Russian practices. Developing a template methodology for national ranking of higher education institutions. Approbation of the developed methodology. Processing the approbation results Organizing public and expert discussions of the approbation results. Consulting with IREG experts to audit the methodology for national ranking of the higher education institutions. Amending the draft methodology in accordance with the results of the discussions. Developing recommendations on the application of the template methodology for national ranking of higher education institutions. Organizing an international conference to discuss the template methodology for national ranking of higher education institutions and the approbation outcomes. 2011 2012 2013
Key Outcomes The results of a comparative analysis of the methodologies and approaches used in international /foreign practices. The results of a comparative analysis of the methodologies and approaches used in the Russian practice. The results of a comparative analysis of the methodologies and approaches used in international/foreign practices and methodologies and approaches used in the Russian practice. A template methodology for national ranking of higher education institutions. The methodology approbation outcomes. The outcomes of expert and public discussions. The outcomes of consultations with the IREG experts. The draft methodology for ranking of Russian higher education institutions. The recommendations for application of the draft methodology for ranking of Russian higher education institutions. . The outcomes of an international conference on the methodology for national ranking of higher education institutions. 2011 2012 2013
II. General approaches to the methodology of developing a template methodology for HEIs ranking Key principles: • The methodology should provide reliable information on performance of higher education institutions and their position in rankings. • The methodology should inform users of educational services on diversity of higher education institutions and education programmes providing friendly and easy-to-use information • The methodology should facilitate improvement of quality and competitiveness of higher education institutions • The methodology should be a source of reliable data for global and regional rankings
Key principles The methodology should take into account: • Experience and achievements of the Russian higher education institutions in the area • Objectives to improve competitiveness and facilitate integration of the Russian higher education institutions into global education and research environment • Increasing number of the Russian higher education institutions that participate/will participate in global rankings • Pragmatic approach to the methodology: data collected for national ranking should correlate with the data universities provide for global rankings • Strengths of quantitative indicator. • Strengths of global ranking methodologies
Methodology for a comparative analysis 3 levels of analysis • 1 level:Comparative analysis of methodologies on key selected parameters (target groups, key objectives, areas of evaluation, frequency, method of data collection and processing etc). • 2 level:Assessing ranking methodologies against Berlin principles on ranking of Higher Education Institutions and the IREG audit criteria and identifying key quantitative indicators • 3 level:Assessing the indicators against criteria of relevance to the Russian education system development objectives, reliability and feasibility of data collection
1 level:Comparing methodologies using common parameters Criteria for selection of ranking methodologies: • Rankings in with Russian universities participate or are expected to take part • Most popular rankings, which top listing is perceived as “signal” of universities competitiveness in international education and research area. • Rankings with methodologies available in open access to ensure transparency and understanding of indicators’ relevance and reliability of the obtained results.
Criteria for selection of ranking methodologies (continued) • possibility of assessing ranking methodologies against Berlin principles on ranking Higher Education institutions • possibility of assessing ranking methodologies against IREG audit criteria • account of diverse practices of various countries • inclusion into analysis of different methodologies • feasibility of application for the national HEIs ranking
Methodologies analyzed Global rankings • The Academic Ranking of World Universities (Shanghai Jiao Tong University ranking) • Times Higher Education Supplement (THES). • World’s Best Universities Rankings - U.S. News and World Report • U-Multirank • Leiden Ranking • QS World University Rankings Regional rankings • U-Map
Methodologies analyzed National rankings UK • Time Good Education Guide (UK) • Guardian Ranking, Universities league table • Guardian Ranking, Specialist institutions USA • Carnegie classification • Forbes Germany • The CHE University Ranking • CHE Excellence Ranking • The CHE Research Ranking • CHE Employability Rating
Methodologies analyzed Specialized rankings • Financial Times ranking of business schools • Blumberg Business Week ranking • The Economist MBA ranking • Indicators for Mapping and Profiling Internationalization
Definitions • Single-dimensional ranking displays vertical diversity in terms of performance by using indicators. Most existing higher education rankings take the form of a league table. • League table – a single-dimensional list going from “best” to “worst”, assigning ordinal numbers to the entities which relate only to rank and scales of difference. • Multi-dimensional ranking does not try to combine education and research rankings, for example, into single, composite measure and is often user-driven because it enables interactive display of data. • Classificationis a system that allocates objects to groups on the basis of their characteristics. It shows horizontal diversity, where differences do not imply ordinary scales of “more”, “bigger” or “better”. It is aimed at showing diversity of higher education institutions. van Vught F.A., Westerheijden D.F. Multidimensional Ranking: a new transparency tool for higher education and research // Higher Education Management and Policy. 2010. Vol. 22/3.
Development of ranking systems Source: Shin J.Ch., Toutkoushian R.K., Teichler U. (eds.) University Rankings: Theoretical Basis, Methodology and Impacts on Global Higher Education. Springer, 2011. P.14.
1 level of analysis: Parameters for comparing ranking methodologies • Type (global, regional, national, specialized ranking). • Focus (mission, goal, objectives). • Target groups: - Target customers (prospective students, students, parents, academics); - General public (higher education institutions, employers, media, ministries, NGOand charities, expert community). • Status (independent, who is the founder etc) • Publication frequency (once a year, etc.). • Basic requirements to participants • Number of institutions ranked • Subject fields (Arts and Humanities, Engineering and Technology, Life Sciences etc.) • Geographical scope
1 level: Common parameters for comparing ranking methodologies • Education levels (undergraduate/postgraduate). • Ranking methodology: - key principles; - indicators, weights; - procedures for data collection; - data processing methods; - datatransformation into ranking. • Ranking outcome (league table, multi-dimensional ranking, clusters of universities). • Criticism and controversy of indicators used in a ranking. • Position of the Russian universities in global rankings. • Ranking reputation.
2 level:Assessing ranking methodologies against Berlin principles and the IREG audit criteria Criteria for assessing ranking methodologies strengths and weaknesses • Berlin principles on ranking of Higher Education Institutions • IREG Ranking Audit Criteria for assessing ranking methodologies
Logical framework for assessing ranking methodologies against the IREG audit criteria Scale for assessment of methodologies against IREGcriteria 0 – criterion is not applicable/data is not available 1 – does not comply with the criteria 2 – partially complies with the criteria 3 – fully complies with the criteria
2 level:Assessing ranking methodologies againstthe IREG audit criteria Position of methodologies: mean score/normalized mean score 26
2 level:Assessing ranking methodologies againstthe IREG audit criteria Fulfilment the IREG audit criteria 27
Limitations of methodologies (U-multirank, CHE University) Resource intensity Lack of comparable data on HEIs performance Complexity of indicators and procedures used for data collection Challenges of ensuring reliability of data received from surveys Challenges of ensuring reliability and quality of collecting large volume of data Difficulties associated with processing of large volume of data 28
3 level of analysis: Assessing the indicators against criteria of relevance to the Russian education system development objectives Identification of similar/repeatable/most frequently used and relevant quantitative indicators Analysis of the most frequently used quantitative indicators on the merits of • data availability • distribution of quantitative indicatorsto areas of evaluation • indicators’ weights Critical assessment most frequently used quantitative indicators against criteria ofrelevance to the Russian education system development objectives, reliability, feasibility of data collection (see the logical framework for assessment of identified indicators in the next slide)
Logical frameworkfor assessment of identified indicatorsfor ranking HEIs methodology
3level: List of output indicators included into experts’ assessment: Research: Number of citations per staff Number of citations per publication Field-normalized citations score. SСPP/SFCSm, SFCSm – field mean citations score Number of cited publications Total number of publications Number of publications in Nature and Science Number of publications in SSCI Number of publications in 40 specified journals Number of professional publications MNCS1 = S (СPP/FCSm) – mean-normalized citation score Research income Number of research grants won Ratio of staff with PhD degrees to the number of completed PhD theses International awards and prizes won Doctorate production Peer reviewed other research products
3level: List of output indicators included into experts’ assessment(continued): Teaching/learning: Alumni of an institution winning Nobel Prizes and Fields Medals Staff of an institution winning Nobel Prizes and Fields Medals Number of awardedPhDdegrees per staff Ratio of awarded PhD degrees to awarded Bachelor degrees Income per staff Proportion of students graduated with first or second class degree Graduation rate Four-year graduation rate Freshmen to Sophomore retention rate Number of students winning nationally competitive awards Listings of Alumni in Who is Who Alumni in Forbes/CCAP Corporate Officers lists Graduate unemployment rate Percentage graduating within norm period
3level: List of output indicators included into experts’ assessment(continued): International orientation: Ratio of international joint research publications per year to total number of publications per year Number of joint research publications International doctorate graduation rate Number of international research grants Income from international sources (teaching, research, contracts with international organisations) Knowledge transfer: Income from business and industry Income from knowledge transfer activities Number of university-industry joint research publications Number of CPD courses per staff Number of start-ups per staff Number of joint university-industry patents Number of cultural events (exhibitions, conferences, concerts), organised by an institution
3level: List of output indicators included into experts’ assessment(continued): Regional engagement: Proportion of income from local/regional sources Percentage of graduates working in the region Number of research contracts with regional partners Number of regional joint research publications Percentage of students in internships in local enterprises Number of theses in cooperation with region Employability: Percentage of employed graduates (six month after graduation) Percentage of graduates employed through career centres Graduates salary Graduates salary increase
IV. Next steps: • Discuss key elements of a template methodology with experts and stakeholders, including areas of evaluation, methods of data collection and indicators’ weights • Approbate the template methodology for ranking of higher education institutions • Conduct expert and public discussions of the approbation results • Conduct consultations with IREG experts on the methodology for ranking of higher education institutions
Thank you for your attention! National Training Foundation www.ntf.ru