Back to all members...
Ilia Shumailov
Associate Member (Senior Research Fellow), started 2022
Ilia is a Research Scientist at Google DeepMind. His interests lie in ML and Computer Security. Prior to DeepMind, Ilia was a Junior Research Fellow at Christ Church and a Fellow of the Vector Institute. He completed his PhD at the University of Cambridge.
Publications while at OATML • News items mentioning Ilia Shumailov • Reproducibility and Code • Blog Posts
Publications while at OATML:
AI models collapse when trained on recursively generated data
Stable diffusion revolutionized image creation from descriptive text. GPT-2, GPT-3(.5) and GPT-4 demonstrated high performance across a variety of language tasks. ChatGPT introduced such language models to the public. It is now clear that generative artificial intelligence (AI) such as large language models (LLMs) is here to stay and will substantially change the ecosystem of online text and images. Here we consider what may happen to GPT-{n} once LLMs contribute much of the text found online. We find that indiscriminate use of model-generated content in training causes irreversible defects in the resulting models, in which tails of the original content distribution disappear. We refer to this effect as ‘model collapse’ and show that it can occur in LLMs as well as in variational autoencoders (VAEs) and Gaussian mixture models (GMMs). We build theoretical intuition behind the phenomenon and portray its ubiquity among all learned generative models. We demonstrate that it must be tak... [full abstract]
Ilia Shumailov, Zakhar Shumaylov, Yiren Zhao, Nicolas Papernot, Ross Anderson, Yarin Gal
Nature
[paper]
News items mentioning Ilia Shumailov:
Group work on model collapse in LLMs published in Nature
24 Jul 2024
OATML group members Ilia Shumailov and Yarin Gal published a study in Nature which finds that indiscriminate use of model-generated content in training causes irreversible defects in the resulting models, in which tails of the original content distribution disappear. You can read the paper here.
Reproducibility and Code
AI models collapse when trained on recursively generated data
This is the code repository for the paper “AI models collapse when trained on recursively generated data”.
CodeIlia Shumailov, Zakhar Shumaylov, Yiren Zhao, Nicolas Papernot, Ross Anderson, Yarin Gal
Blog Posts
OATML Conference papers at NeurIPS 2022
OATML group members and collaborators are proud to present 8 papers at NeurIPS 2022 main conference, and 11 workshop papers. …
Full post...Yarin Gal, Freddie Kalaitzis, Shreshth Malik, Lorenz Kuhn, Gunshi Gupta, Jannik Kossen, Pascal Notin, Andrew Jesson, Panagiotis Tigas, Tim G. J. Rudner, Sebastian Farquhar, Ilia Shumailov, 25 Nov 2022