The Glasgow Information Retrieval Group within the School of Computing Science at the University of Glasgow was founded 32 years ago in 1986 by Professor C. J. ‘Keith’ van Rijsbergen, often considered one of the founders of modern Information Retrieval (IR). From its outset, the Glasgow IR group has focused on improving the effectiveness of IR systems, inventing new logic & probabilistic retrieval models in the 90's and early 2000's, followed by the development of adaptive query expansion techniques, interactive multimedia models, the Divergence From Randomness framework, as well as leading research into quantum, expertise search and search result diversification models in the late 2000's. Since then, the Information Retrieval group embraced emerging machine learning and deep learning technologies for very large corpora and data streams, and have been at the forefront of research, development and application of those technologies for search and recommendation use-cases in a manner that ensures both effectiveness and efficiency.

The Glasgow IR Group has a strong research track record. Indeed, the ACM Digital Library shows that the group is ranked first by number of papers (429) at the SIGIR conference (the top CORE A* conference in the IR field). Meanwhile, a recent study by Microsoft Research of the 40 years of SIGIR showed the University of Glasgow as the 5th most cited university at the conference and the 1st in Europe. The group is also renowned for developing the popular open source IR platform,, which has been downloaded over 60,000 since its first release in 2004 and is cited by over 3500 research papers. Furthermore, the group has a long history of engagement with the public and industry sectors from small SMEs to multinational corporations.

The Informer magazine of BCS's Information Retrieval Specialist Group carried a recent profile on the Glasgow Information Retrieval Group.


As the most active Information Retrieval group by publications in Europe and one of the longest running, our research covers the full-spectrum of topics that are relevant to the development of IR systems:

IR & Recommender Systems Models

  • Theoretical modelling of IR systems
  • Machine learning and deep learning for information retrieval and recommender systems
  • Interactive information retrieval (personalised IR, emotion based search, user modelling for IR, gestural IR)
  • User modelling and personal information access
  • Topic modeling; Entity search; Natural language processing for IR
  • Recommender systems; Context-aware venue suggestion

Large-scale IR & Efficient IR

  • Web information retrieval; Big data and information retrieval
  • Efficient architecture for large-scale IR systems; Data stream processing architectures

Data Streams & IR

  • Real-time information retrieval
  • Search in social and sensor networks

Artificial Intelligence & IR

  • Conversational information seeking and dialogue systems
  • Information credibility, transparency, explainability and verification in IR systems
  • Fairness in information retrieval & recommender systems

Natural Language Processing & IR

  • Information extraction including entity and relation extraction
  • Automatic knowledge graph construction
  • Multi-task models, joint models and summarization


  • Multimedia information retrieval
  • Domain-specific information retrieval: smart cities; health; news; eDiscovery; sensitivity review
  • Emergency management and crisis informatics
  • Politics and Media


  • Test collections and evaluation metrics
  • Evaluation of IR systems and crowdsourcing for IR
  • Online and Offline Evaluation of IR and Recommender Systems
  • Eye-tracking and physiological approaches, such as fMRI


Active Projects:


Recent Projects:

Current staff and students

Academic Staff:


Current Research Assistants and Research Students:

  • Javier Sanz-Cruzado Puig
  • Ting Su
  • Xi Wang
  • Xiao Wang
  • Siwei Lu
  • JingMin Huang
  • Yashon Wu
  • Xin Xin
  • Carlos Gemmel
  • Federico Rossetto
  • Sarawoot Kongyoung
  • Alexander Hepburn
  • Ian Mackie
  • Jun Choi Hyun
  • Hitarth Narvala
  • Jijun Long
  • Maria Vlachou
  • Sasha Petrov
  • Thomas Janich
  • Zixuan Yi
  • Erland Frayling
  • Edward Richards


Recent Graduates

  • Jarana Manotumruksa (2019), University College London, Researcher
  • Anjie Fang (2019), Amazon, Applied Scientist
  • Jorge David Gonzalez Paule  (2019), Jobandtalent Espana, Data Scientist
  • Colin Wilkie (2019), Siemens, Data Engineer
  • David Maxwell (2019), University of Deft, Data Engineer
  • Graham McDonald (2019), University of Glasgow, Lecturer
  • James McMinn (2018), ScoopAnalytics, Co-Founder
  • Stuart Mackie (2018), BiP Solutions/Strathclyde Uni, Data Scientist
  • Horatiu Bota (2018), Prodsight, Data Scientist
  • Jesus Alberto Rodriquez Perez (2018), University of Glasgow, Postdoctoral Researcher
  • Fajie Yuan (2018), Tencent, Senior Researcher


Notable Alumni

  • Ryen White (Research Manager, Microsoft Research AI)
  • Mark Sanderson (Professor, Royal Melbourne Institute of Technology)
  • Mounia Lalmas (Head of Tech Research, Spotify)
  • Ian Ruthven (Professor, Strathclyde University)
  • Fabio Crestani (Professor, University of Lugano)
  • Vassilis Plachouras (Software Engineering, Facebook)
  • Leif Azzopardi (Chancellor's Fellow, Strathclyde University)
  • Rodrygo Santos (Assistant Professor, Federal University of Minas Gerais)
  • Eugene Kharitonov (Research Engineer, Facebook)
  • Saul Vargas (Senior Machine Learning Scientist, ASOS)
  • Dyaa Albakour (Lead Data Scientist, Signal Media)
  • Nut Limsopatham (Senior Researcher, Microsoft AI)
  • Amir Jadidinejad (AI Engineer, Glaxo Smith Kline)
  • Zaiqiao Meng (Researcher, Cambridge University)

Terrier IR platform

Terrier is a highly flexible, efficient, and effective open source search engine, readily deployable on large-scale collections of documents developed by the IR group. Terrier implements state-of-the-art indexing and retrieval functionalities, and provides an ideal platform for the rapid development and evaluation of large-scale retrieval applications. Indeed, Terrier is used internationally, with over 60,000 downloads since its first release in 2004. Terrier is is used widely by the research community, with over 3700 citations in research papers according to Google Scholar.

Visit the website at to learn more and download Terrier for free.

Popular resources

For those new to the Information Retrieval field, the group maintains a useful set of common resources for researchers and practitioners:

  • Information Retrieval Test Collections: On this page are a list of publically available IR test collections. Some are held locally and some are pointers to remote sites.
  • Collections of text and corpora: What's the difference between a test collection and a text collection? Well a test collection has to have associated queries and relevance judgements. The things in here are simply document collections.
  • Language reference works: This page contains links to online language reference works, such as dictionaries, thesauri etc.
  • IR systems: A list of links to some sites that have information about IR systems.
  • Linguistic utilities: Bits of IR language related utilities like stemmers, stop words lists, morphological taggers, etc.
  • IR Journals: Various table of contents and abstracts of the papers in a number of well known IR journals.
  • IR Organisations: Various IR groups and more formal organisations.
  • Books: Supplements of books or whole books online.

Upcoming events

Special IR Seminar & Panel Discussion on LLMs in IR

Group: Information Retrieval (IR)
Speaker: Eugene Yang (Johns Hopkins University) and Maik Fröbe (University of Jena)
Date: 31 March, 2023
Time: 14:00 - 16:30
Location: Sir Alwyn Williams Building, 422 Seminar Room

Mark your calendars! Dr. Eugene Yang and Maik Fröbe will be visiting us for a special IR Seminar. Each will give a talk on a timely topic in the field, and we'll wrap up with a panel discussion on the effect that large language models (like ChatGPT) will have on the field.


14:00-15:00 - Eugene Yang
From Monolingual Neural IR to Cross-Language and to Multi-Language IR

Eugene Yang is a postdoctoral fellow at the Human Language Technology Center of Excellence at Johns Hopkins University. He received his PhD from Georgetown University, where he specialized in High Recall Retrieval. Eugene's current research is focused on cross-language and multilingual retrieval. He is also a co-organizer of the TREC NeuCLIR track since 2022.


15:00-16:00 - Maik Fröbe
TIREx: The Information Retrieval Experiment Platform: Towards Reproducible Shared Tasks in IR

Maik Fröbe is a Ph.D. student under the supervision of Matthias Hagen (University of Halle: 2019 to 2022, University of Jena: 2022 till today) and part of the webis network. Maik received his Bachelor's and Masters's in computer science at the University of Leipzig. His current research interests lie in Information Retrieval, particularly learning to rank, web archive mining, and near-duplicate detection and its impact on IR evaluation. Maik is an active developer of the platform, which improved the reproducibility of a number of shared tasks and has an archive of more than 500 research prototypes.


Panel Discussion on large language models in IR

Pretraining, Instruction Tuning, Alignment, Specialization: On the Source of Large Language Model Abilities

Group: Information Retrieval (IR)
Speaker: Yao Fu, University of Edinburgh
Date: 17 April, 2023
Time: 15:00 - 16:00
Location: Sir Alwyn Williams Building, 422 Seminar Room


Recently, the field has been greatly impressed and inspired by Large Language Models (LLMs) like GPT-3.5. The LLMs' multi-dimensional abilities are significantly beyond many NLP researchers’ and practitioners’ expectations and thus reshaping the research paradigm of NLP. A natural question is how LLMs get there, and where these fantastic abilities come from. In this talk we try to dissect the strong LLM abilities and trace them to their sources, hoping to give a comprehensive roadmap about the evolution of LLMs.



Yao Fu is a Ph.D. student at the University of Edinburgh. Previously he finished his M.S. in Columbia University and B.S. in Peking University. Yao studies large scale probabilistic generative models for human language. His research interests include language model evolution, complex reasoning, and how to inject strong abilities to language models from first principles.