The Arab University Rankings 2026 will be published on 26 November 2025.
ճ̽ƵWorld University Rankingsare the only global performance tables that judge research-intensive universities across all their core missions: teaching, research, knowledge transfer and international outlook.
This year the methodology behind the Arab University Rankings has been brought in line with the flagship World University Rankings, using broadly the same comprehensive and trusted set of performance indicators (listed below). However, the weightings have been recalibrated to reflect the priorities of Arab institutions.
This year’s release is named the Arab University Rankings 2026 because it uses data from the World University Rankings 2026 edition and is part of the same annual data cycle. Last year’s Arab rankings was named the 2024 edition because it used the previous methodology but no years have been skipped since the inaugural Arab University Rankings in 2021.
The methodology behind the Arab University Rankings is based on the samecomprehensive and trusted framework as the global tablebut some important adjustments have been made to reflect the features of universities in the Arab region.
̽Ƶ
Discover how universities are ranked, the indicators that are used, and why the THE Arab University Rankings is a trusted benchmark for university performance in the Arab World.
Download the full methodology as a PDF at the bottom of this page.
̽Ƶ

Key criteria for theArab University Rankings
Five core pillars of evaluation
- Teaching(the learning environment)
- Research environment(volume, income and reputation)
- Research quality(research strength, research excellence and research influence)
- International outlook(staff, students and research)
- Industry(income and patents)
How are the Arab University Rankings calculated?
Metrics included
Teaching (the learning environment): 29.5%
- Teaching reputation: 15%
- Doctorates awarded-to-academic-staff ratio: 5.5%
- Academic staff-to-student ratio: 4.5%
- Doctorates awarded-to-undergraduate-degrees-awarded ratio: 2%
- Institutional income per academic staff: 2.5%
The Academic Reputation Survey that underpins this pillar was carried out between November 2024 and January 2025. This exercise examined the perceived prestige of institutions in teaching and research. The teaching reputation metric is based on the number of teaching votes obtained from the survey. We have run the survey to ensure a balanced spread of responses across disciplines and countries. Where disciplines or countries were over- or under-represented, THE’s data team weighted the responses to fully reflect the global distribution of scholars. The 2025 data are combined with the results of the 2024 survey, giving more than 108,000 responses globally. Universities that received no votes score zero for this metric.
As well as giving a sense of how committed an institution is to nurturing the next generation of academics, a high proportion of postgraduate research students also suggests the provision of teaching at the highest level that is thus attractive to graduates and effective at developing them. This metric isgenerated by dividing the total subject-weighted number of doctorates by the total subject-weighted number of academic staff (because the volume of doctoral awards varies bydiscipline). The metric is normalised after calculation.
The academic staff-to-student ratio is defined as the total full-time equivalent (FTE) number of staff employed in an academic post divided byFTE number of students in all years and of all programmes that lead to adegree, certificate, university credit or other qualification. This variable andthe doctorates-awarded-to-undergraduate-degrees-awarded ratio are normalised after calculation.
Institutional income indicates an institution’s general status and gives abroad sense of the infrastructure and facilities available to students and staff. This metric is generated by dividing the institutional income, adjusted to purchasing power parity (PPP), by the total number of academic staff. This variable is normalised after calculation
Research environment: 29%
- Research reputation: 18%
- Research productivity: 5.5%
- Research income per academic staff: 5.5%
The most prominent indicator in this category looks at a university’s reputation for research excellence among its peers, based on the responses to our Academic Reputation Survey (see above).
To measure productivity, we count the number of publications published in the academic journals indexed by Elsevier’s Scopus database per scholar, scaled for institutional size and weighted by subject. This gives a sense of the university’s ability to get papers published in quality peer-reviewed journals. This measure includes a method to give credit for cross-subject research that results in papers being published in subjects where a university declares no staff.
Research income is scaled against academic staff numbers and adjusted for PPP. This is a somewhat controversial indicator because it can be influenced by national policy and economic circumstances. But income is crucial to the development of world-class research, and because much of it is subject to competition and judged by peer review, our experts suggested that it was a valid measure. This indicator takes account of each university’s distinct subject profile, reflecting the fact that research grants in science subjects are often bigger than those awarded for the highest-quality social science, arts and humanities research.
Research quality: 30%
- Research strength: 15%
- Research excellence: 7.5%
- Research influence: 7.5%
Our research quality pillar looks at universities’ role inspreading new knowledge and ideas.
̽Ƶ
We examine research quality by capturing the number of times a university’s published work is cited by scholars globally. This year, our bibliometric data supplier Elsevier provided more than 174.9million citations to 18.7million journal articles, article reviews, conference proceedings, books and book chapters published over five years. The data include more than 28,700 active peer-reviewed journals indexed by Elsevier’s Scopus database and all indexed publications between 2020 and 2024. Citations to these publications made in the six years from 2020 to 2025 are also collected.
We consider the field-weighted citation impact (FWCI) for each institution, per subject and overall. The research strength score used in the ranking is determined by calculating the 75th percentile FWCIscore of all papers published by each institution. We believe that this gives a more stable measure over time. The data are also normalised to reflect variations incitation volume between different subject areas.
Two new citation measures were introduced in 2023. Our research excellence metric recognises an institution’s contribution to the best research in each subject and overall. Excellence is measured by capturing the total number of publications by an institution that are among the top 10per cent of publications worldwide by FWCI. We adjust this number by year, subject and the total number of academic and research staff.
Finally, research influence examines the influence of an institution’s publications by analysing their corresponding citations. The importance of a publication is determined based on the importance of other papers citing it. We adjust this number by year, subject and the total number ofacademic and research staff.
̽Ƶ
International outlook: 7.5%
- Proportion of international students: 2.5%
- Proportion of international staff: 2.5%
- International co-authorship: 2.5%
The ability of a university to attract undergraduates, postgraduates and faculty from all over the planet is key to its success on the world stage.
International students and staff are defined as those whose nationality differs from the country where the institution is based. The first two metrics are calculated as the total FTE number of international students or staff divided by the total FTE number of students orstaff.
In the third international indicator, we calculate the proportion of a university’s total research journal publications that have at least one international co-author and reward higher volumes. This metric accounts for an institution’s subject mix and uses the same five-year window as the “Research quality” category.
Historically, large countries have been disadvantaged compared with small countries in our international metrics, in that it is “easier” for staff and students in small countries to work/study abroad. This led us to change our normalisation approach for the four measures in 2023, henceforth taking into consideration the population of a country when evaluating these metrics.
A study abroad metric – assessing the provision of international learning opportunities for domestic students – complements the International outlook pillar, but is currently given a weight of 0percent.
Իٰܲ: 4%
- Industry income per academic staff: 2%
- ʲٱԳٲ: 2%
A university’s ability to help industry with innovations, inventions and consultancy has become a core mission of the contemporary global academy. The industry income metric seeks to capture such knowledge transfer activity by looking at how much research income an institution earns from industry (adjusted forPPP), scaled against the number of academic staff it employs. This suggests the extent to which businesses are willing to pay for research and a university’s ability to attract funding in the commercial marketplace – useful indicators of institutional quality.
But the extent to which universities are supporting their national economies through technology transfer is an area that deserves greater recognition. The patents metric, introduced in the Arab University Rankings for the first time this year, is defined as the number of patents from any source that cite research conducted by the university. The data is provided by Elsevier and relates to patents published between 2020 and 2024 (not research published between these dates). Sources for patents include the World Intellectual Property Organisation, the European Patent Office and the patent offices of the US, the UK and Japan, as well as more than 100 patent offices around the world. In total, 43 are relevant for the time period. This measure is subject-weighted to avoid penalising universities producing research in fields low in patents, and scaled for institutional size.
Producing the overall ranking
Data collection
Institutions provide and sign off their institutional data for use in the rankings. On the rare occasions when a particular datapoint is not provided, we enter a conservative estimate for the affected metric. By doing this, we avoid penalising an institution too harshly with a “zero” value for data that it overlooks or doesnot provide, but we donot reward it for withholding them.
Getting to the final result
Moving from a series of specific datapoints to indicators and then to a total score for an institution requires us to match values that represent fundamentally different data. To do this, we use a standardisation approach for each indicator, and then combine the indicators in the proportions shown above.
The standardisation approach we use is based on the distribution of data within a particular indicator, where we calculate a cumulative probability function, and evaluate where a particular institution’s indicator sits within that function.
For most indicators, we calculate a normal cumulative probability function. The distribution of data in the metrics on teaching reputation, research reputation, research strength, research excellence and patents requires us to use an exponential scoring function.
Inclusion criteria for Arab University Rankings
Universities must supply data to be included in the ranking. They must also have published more than 500research publications between 2020 and2024.
The Arab University Rankings considers only institutions based in the following countries: Algeria, Bahrain, Comoros, Djibouti, Egypt, Iraq, Jordan, Kuwait, Lebanon, Libya, Mauritania, Morocco, Oman, Palestine, Qatar, Saudi Arabia, Somalia, Sudan, Syria, Tunisia, the United Arab Emirates and Yemen.
Understanding the results
Banded institutions
Precise ranks and overall scores are shown for the institutions ranked in the top 100. We then display banded ranks and overall scores for institutions in the rest of the table because the difference between their scores is not statistically significant.
̽Ƶ
Reporter institutions
A small number of institutions have“reporter” statusand are listed at the bottom of the table. This means that they provided data but did not meet our eligibility criteria to receive arank.
Register to continue
Why register?
- Registration is free and only takes a moment
- Once registered, you can read 3 articles a month
- Sign up for our newsletter
Subscribe
Or subscribe for unlimited access to:
- Unlimited access to news, views, insights & reviews
- Digital editions
- Digital access to ձᷡ’s university and college rankings analysis
Already registered or a current subscriber?




