Back to all members...

Ilia Shumailov

Associate Member (Senior Research Fellow), started 2022

Ilia is a Research Scientist at Google DeepMind. His interests lie in ML and Computer Security. Prior to DeepMind, Ilia was a Junior Research Fellow at Christ Church and a Fellow of the Vector Institute. He completed his PhD at the University of Cambridge.


Publications while at OATMLNews items mentioning Ilia ShumailovReproducibility and CodeBlog Posts

Publications while at OATML:

AI models collapse when trained on recursively generated data

Stable diffusion revolutionized image creation from descriptive text. GPT-2, GPT-3(.5) and GPT-4 demonstrated high performance across a variety of language tasks. ChatGPT introduced such language models to the public. It is now clear that generative artificial intelligence (AI) such as large language models (LLMs) is here to stay and will substantially change the ecosystem of online text and images. Here we consider what may happen to GPT-{n} once LLMs contribute much of the text found online. We find that indiscriminate use of model-generated content in training causes irreversible defects in the resulting models, in which tails of the original content distribution disappear. We refer to this effect as ‘model collapse’ and show that it can occur in LLMs as well as in variational autoencoders (VAEs) and Gaussian mixture models (GMMs). We build theoretical intuition behind the phenomenon and portray its ubiquity among all learned generative models. We demonstrate that it must be tak... [full abstract]


Ilia Shumailov, Zakhar Shumaylov, Yiren Zhao, Nicolas Papernot, Ross Anderson, Yarin Gal
Nature
[paper]
More publications on Google Scholar.

News items mentioning Ilia Shumailov:

Group work on model collapse in LLMs published in Nature

Group work on model collapse in LLMs published in Nature

24 Jul 2024

OATML group members Ilia Shumailov and Yarin Gal published a study in Nature which finds that indiscriminate use of model-generated content in training causes irreversible defects in the resulting models, in which tails of the original content distribution disappear. You can read the paper here.

Link to this news item


Reproducibility and Code

AI models collapse when trained on recursively generated data

This is the code repository for the paper “AI models collapse when trained on recursively generated data”.

Code
Ilia Shumailov, Zakhar Shumaylov, Yiren Zhao, Nicolas Papernot, Ross Anderson, Yarin Gal


Blog Posts

OATML Conference papers at NeurIPS 2022

OATML group members and collaborators are proud to present 8 papers at NeurIPS 2022 main conference, and 11 workshop papers. …

Full post...


Yarin Gal, Freddie Kalaitzis, Shreshth Malik, Lorenz Kuhn, Gunshi Gupta, Jannik Kossen, Pascal Notin, Andrew Jesson, Panagiotis Tigas, Tim G. J. Rudner, Sebastian Farquhar, Ilia Shumailov, 25 Nov 2022

Are you looking to do a PhD in machine learning? Did you do a PhD in another field and want to do a postdoc in machine learning? Would you like to visit the group?

How to apply


Contact

We are located at
Department of Computer Science, University of Oxford
Wolfson Building
Parks Road
OXFORD
OX1 3QD
UK
Twitter: @OATML_Oxford
Github: OATML
Email: oatml@cs.ox.ac.uk