Stephen Roller

About

I am currently a Research Scientist at Google DeepMind. My research and work primarily focuses on pretraining the largest-scale language models.

Previously, I worked as a Member of the Technical Staff at Character.AI, where I lead the pre-training and training infrastructure efforts, and as a Research Engineer at Meta's Fundamental AI Research (FAIR), where I was a leading member of the Large Language Model and dialogue teams. Much of my work from FAIR is open source.

Before FAIR, I completed my PhD in Computer Science at the University of Texas at Austin under the supervision of Katrin Erk.

Contact

I'm most easily reached via my personal email, me@stephenroller.com.

Publications

My scholarly publications appear below in reverse chronological order. Stars by names indicate joint first authorship. Due to the sensitivity of my work in pretraining, my work has been unpublished since I left FAIR.

FAIR (2017 — 2023)

PhD (2010 — 2017)

Undergrad (2010)