My research is concerned with the development of Safe AI, i.e, developing methods to ensure deployed artificial intelligence models do not pose a threat in high stakes environments. To address this, my work has focussed primarily on adversarial attacks (and how we can defend against them) in the Natural Language Processing (NLP) domain. My other works have explored other Safe AI related topics including: uncertainty for out of distribution handling; biases and shortcut learning.
My research has been applied to a range of tasks: standard NLP classification tasks (e.g. entailment and sentiment classification); grammatical error correction; neural machine translation, spoken language assessment, weather tabular data and standard image classification (object recognition) tasks.
Gender Bias and Universal Substitution Adversarial Attacks on Grammatical Error Correction Systems for Automated Assessment
Vyas Raina, Mark Gales
U.K. Speech 2022
Paper Poster CodeResidue-Based Natural Language Adversarial Attack Detection
Vyas Raina, Mark Gales
North American Chapter of the Association for Computational Linguistics (NAACL) 2022
Paper Poster CodeShifts: A dataset of real distributional shift across multiple large-scale tasks
Andrey Malinin, Neil Band, German Chesnokov, Yarin Gal, Mark JF Gales, Alexey Noskov, Andrey Ploskonosov, Liudmila Prokhorenkova, Ivan Provilkov, Vatsal Raina, Vyas Raina, Mariya Shmatova, Panos Tigas, Boris Yangel
Advances in Neural Information Processing Systems (NeurIPS), Datasets and Benchmarks Track 2021
Paper Presentation CodeUniversal adversarial attacks on spoken language assessment systems
Vyas Raina, Mark Gales, Katherine Knill
INTERSPEECH 2020
Paper Presentation CodeNeurIPS 2021
The Shifts Challenge aims to provide a standardized set of benchmark datasets and baseline models across a range of modalities to assess the impact of distributional shift in the wild. Often deployed systems will fail when used in domains where there exist a statistical shift from the source training domain. The aim of this challenge is two-fold: the development of models that are robust to real-life distributional shifts AND the ability of models to give a meaningful uncertainty measure for their predictions, i.e., the models should know when they are likely to be wrong, so that human intervention can be provided in such settings.
My work in this challenge focused on development of the Weather track, where I designed the data splits, augmentation of data and model baseline training. Further, I worked on refining and assessing uncertainty measures to be used for model evaluation. Finally, I actively helped in the organization, tutorial writing and running of the challenge at NeurIPS 2021.
The Shifts Weather Prediction dataset contains both a scalar regression and a multi-class classification task. Specifically, at a particular latitude, longitude, and timestamp, one must predict either the air temperature at two meters above the ground or the precipitation class, given targets and features derived from weather station measurements and weather forecast models. This data is used by Yandex for real-time weather forecasts and represents a real industrial application.
Paper Challenge Talks Code