Russel Shawn Dsouza

Russel Dsouza

School of Computer Science
Research Fellow

Contact details

Address
University of Birmingham
Edgbaston
Birmingham
B15 2TT
UK

Russel is a Research Fellow at the Institute for Data and AI. His research focuses on data-efficient learning, interpretability, and the mechanistic understanding of alignment in large language models, alongside metascience and human-model disagreement.

Qualifications

  • MSc Artificial Intelligence and Machine Learning, University of Birmingham, 2023
  • BTech Electronics and Communications Engineering, National Institute of Technology Karnataka - India, 2017

Biography

Russel Dsouza's research focuses on making large language models more efficient, reliable, and understandable. At the Institute for Data and AI at the University of Birmingham, he  contributes to this goal while also helping to shape research priorities and supporting data-intensive research proposals across the University.

Russel's individual research investigates the core challenges of model alignment and factuality. This includes analysing the sources of human-model disagreement in the data used to align large language models and contributing to the development of optimised fact-checking pipelines. This work is complemented by interdisciplinary collaborations. With colleagues in Psychology, he has explored the differences between human and model reasoning, leading to a system for automating the scoring of psychological assessments. With a team in Education, Russel's work involved analysing the impact of student-AI interactions.

Russel holds a Master of Science in Machine Learning from the University of Birmingham, where his dissertation focused on evaluating adversarial alignment subversion in large language models. He also holds a Bachelor of Technology in Electronics and Communications Engineering; his dissertation involved developing a deep learning system for the pixel-wise classification of land-use from satellite imagery. Russel's prior research experience includes developing machine learning models for applications in satellite imaging and computational histopathology.

Teaching

  • Natural Language Processing

Research

Russel Dsouza's research aims to make large language models more efficient, reliable, and understandable. He is particularly interested in data efficiency, exploring methods for sample-efficient continual learning and token-efficient reasoning to build models that can learn and adapt without requiring massive computational resources.

This goal of efficiency is linked to the challenge of ensuring models are trustworthy. Russel's work on faithfulness and factuality addresses the need for models to produce truthful and verifiable information. This involves developing frameworks that can robustly check facts and reason about the world, moving beyond simple pattern matching.

Underpinning all of this is his interest in interpretability. To truly trust these systems, we must understand their internal workings. Russel focuses on the mechanistic interpretability of model alignment, seeking to understand how and why models behave as they do, which is crucial for ensuring their safety and preventing failures like alignment collapse.

Publications

Recent publications

Conference contribution

Wang, Y, Dsouza, R, Lee, R, Apperly, I, Devine, RT, van der Kleij, S & Lee, M 2025, Automatic Scoring of an Open-Response Measure of Advanced Mind-Reading Using Large Language Models. in Proceedings of the 10th Workshop on Computational Linguistics and Clinical Psychology (CLPsych 2025)). Association for Computational Linguistics, ACL, pp. 79–89, The Workshop on Computational Linguistics and Clinical Psychology, Albuquerque, New Mexico, United States, 3/05/25. https://doi.org/10.18653/v1/2025.clpsych-1.7

Ftouhi, F, Dsouza, R, Gamboa, LC, Liu, J, Abbas, A, Feng, Y, Ali, M, Lee, M & Kovatchev, V 2025, OldJoe at AVeriTeC: In-context learning for fact-checking. in The Eighth FEVER Workshop at The 63rd Annual Meeting of the Association for Computational Linguistics. 63rd Annual Meeting of the Association for Computational Linguistics, Vienna, Austria, 27/07/25.

Kovatchev, V & Dsouza, R 2025, Sources of Disagreement in Data for LLM Instruction Tuning. in CoMeDi: Context and Meaning - Navigating Disagreements in NLP Annotations. CoMeDi, Abu Dhabi, United Arab Emirates, 19/01/25.

View all publications in research portal