- Feb 2025: I won a UK AISI Grant to study the distortive effects of AI writing assistance.
- Dec 2024: Our PRISM alignment dataset won Best Paper (D&B) at NeurIPS 2024 🏆
- Dec 2024: I was elected Associate Member of Nuffield College at the University of Oxford.
- Nov 2024: I joined an EU JRC expert panel to consult on the implementation of the EU AI Act.
- Aug 2024: Our work on LLM values and opinions won Outstanding Paper at ACL 2024 🏆
- Jun 2024: I am at NAACL 2024 in Mexico City to present XSTest and organise WOAH.
- Jan 2024: I published SafetyPrompts.com, a living catalogue of open datasets for LLM safety.
- Jul 2023: The Sexism Detection Task we organised won Best Paper at SemEval 2023 🏆
- Jun 2023: I joined Dirk Hovy's MilaNLP Lab as a postdoctoral researcher.
- May 2023: I defended my PhD thesis in Oxford, assessed by Scott Hale and Maarten Sap.
- May 2023: The HateCheck project that I led won the Stanford AI Audit Challenge 🏆
- Mar 2023: My work on OpenAI's red team for GPT-4 was covered by various media outlets.
- Mar 2023: The AI start-up Rewire that I co-founded in 2021 was acquired by ActiveFence.
- ...
I am a Postdoctoral Researcher in the MilaNLP Lab at Bocconi University and an Associate Member of Nuffield College at the University of Oxford.
I work at the intersection of computation, language, and society.
In my current research, I focus on evaluating and improving the safety of large language models (LLMs), as well as studying their societal impacts.
News
Research
I am very excited about large language models (LLMs). This is why I work on improving their safety and understanding their societal impacts. For my recent work in this area, I won an Outstanding Paper Award at ACL and a Best Paper Award at NeurIPS D&B.
LLMs are now being used globally by hundreds of millions of people. I am particularly interested in the risks created by this widespread adoption, and how we can mitigate them.
For a complete record of my publications, please visit my Google Scholar profile.
Supervision
If you are a student at Bocconi or Oxford, and you are interested in working together on a topic related to my research, I am happy to discuss potential thesis supervision. Just get in touch :)
Press
I enjoy talking about my work, and I have been fortunate to have it featured across many different media outlets. Here are a few examples:
- - My work on OpenAI's red team for GPT-4 was covered by the FT, the Times, and Sifted.
- - My commentary on DeepSeek / AI safety was featured in Die Zeit, SZ, and Der Standard.
- - The HateCheck project I led was covered by MIT Tech Review, the WSJ and VentureBeat.