Andrej Risteski

Bio

I am an Assistant Professor at the Machine Learning Department in Carnegie Mellon University. Prior to that, I was a Norbert Wiener Research Fellow jointly in the Applied Math department and IDSS at MIT. I received my PhD in the Computer Science Department at Princeton University under the advisement of Sanjeev Arora.

My research interests lie in the intersection of machine learning, statistics, and theoretical computer science, spanning topics like (probabilistic) generative models, algorithmic tools for learning and inference, representation and self-supervised learning, out-of-distribution generalization and applications of neural approaches to natural language processing and scientific domains. More broadly, the goal of my research is principled and mathematical understanding of statistical and algorithmic problems arising in modern machine learning paradigms.

I am the recipient of an NSF CAREER Award, an Amazon Research Award and a Google Research Award. I am also in part supported by several NSF and DoD awards, as well as an OpenAI Superalignment grant.

News & Events

June 2025

Invited talk at IMSI Workshop on Statistical and Computational Challenges in Probabilistic Scientific Machine Learning (SciML)

Link for more info: https://www.imsi.institute/activities/statistical-and-computational-challenges-in-probabilistic-scientific-machine-learning-sciml/

May 2025

Stephen Huan awarded DoE CSGF fellowship

Stephen has been awarded a Department of Energy (DOE) Computational Science Graduate Fellowship for the 2025–26 academic year. He is one of only 30 students nationwide selected for the fellowship, which supports researchers applying high-performance computing to problems in science and engineering.

Link for more info: https://www.cs.cmu.edu/news/2025/huan-doe-fellowship

May 2025

COLT workshop on Theory of AI for Scientific Computing (TaSC)

Myself, former lab member Tanya Marwah, Misha Khodak, Nick Boffi and Jianfeng Lu are co-organizing a COLT workshop on Theory of AI for Scientific Computing. We are soliciting contributed work at the interface of learning theory, statistics, numerical methods, algorithm design, and the physical sciences.

Link for more info: https://tasc-workshop.github.io/

February 2025

Tanya Marwah successfully depended her thesis, and will join Polymathic AI and Flatiron Institute as a Research Fellow. Congrats Tanya!

Academic Positions

Assistant Professor

Carnegie Mellon University: Machine Learning Department

2019 - present

Norbert Wiener Fellow

MIT: IDSS and Applied Mathematics

2017 - 2019

Education

Ph.D. in Computer Science

Princeton University

Advised by Sanjeev Arora

2012 - 2017

B.Sc. in Computer Science

Princeton University

2008 - 2012

Service

Senior Area Chair

NeurIPS 2024

Area Chair

ICLR 2023, 2024

NeurIPS 2020, 2021, 2022, 2023

COLT 2022

UAI 2022

Program Committee

FOCS 2020, 2022

Awards

Google Research Award

2024

Algorithmic Foundations for Generative AI: Inference, Distillation and Non-Autoregressive Generation

NSF CAREER Award

2023

Theoretical Foundations of Modern Machine Learning Paradigms: Generative and Out-of-Distribution

OpenAI Superalignment Award

2023

Research support for AI alignment work

Amazon Research Award

2022

Causal + Deep Out-of-Distribution Learning

Research Areas

Generative Models

Generative Models

Algorithmic and statistical foundations of generative models, from GANs to diffusion models.

Representation learning

Representation learning

Methodological foundations for learning, understanding and interpreting representations.

Out-of-Distribution Generalization

Out-of-Distribution Generalization

Robustness and adaptation to benign and adversarial distribution shifts.

Neural Language Models

Neural Language Models

Theoretical and practical aspects of deep learning approaches to natural language processing.

AI for Scientific Applications

AI for Scientific Applications

Machine learning approaches for the sciences.

Sampling and Optimization

Sampling and Optimization

Algorithms and theory for computationally and statistically efficient sampling and optimization.

Publications

Curated selection of key publications across research areas:

With Frederic Koehler and Alexander Heckett

ICLR 2023, Oral Equivalent, Top 5% of Papers

With Tanya Marwah and Zachary C. Lipton

NeurIPS 2021, Spotlight

With Elan Rosenfeld and Pradeep Ravikumar

ICLR 2021

With Sanjeev Arora and Yi Zhang

ICLR 2018

With Sanjeev Arora, Yuanzhi Li, Yingyu Liang and Tengyu Ma

Transactions of the Association for Computational Linguistics (TACL), 2016

Talks

Talks from the last 2 years (2023-2025):

Theoretical perspectives on modern machine learning paradigms: generative, scientific and out-of-distribution

2025

UC Berkeley CS seminar, 03/2025

The statistical cost of score-based losses

2024

Simons Institute for the Theory of Computing, Moderns Paradigms in Generalization Boot Camp, 09/2024

Neural networks for PDEs: representational power and inductive biases

2024

Workshop on Mathematics of Data, Institute for Mathematical Sciences at NUS, 01/2024

From algorithms to neural networks and back

2024

ETH Seminar, 07/2024

The statistical cost of score-based losses

2024

Oxford University, Computational Statistics and Machine Learning Seminar, 07/2024

Discernible patterns in trained Transformers: a view from simple linguistic sandboxes

2024

Theory of Interpretable AI seminar, 11/2024

Neural networks for PDEs: representational power and inductive biases

2024

CNLS Annual Conference on Physics-Informed Machine Learning, 10/2024

The statistical cost of score-based losses

2024

Duke University Computer Science Department Colloquium, 11/2024

The statistical cost of score matching

2024

Joint Mathematics Meetings, 01/2024

Neural networks for PDEs: representational power and inductive biases

2024

Meeting in Mathematical Statistics, CIRM, 12/2024

The statistical cost of score-based losses

2024

Newton Institute program 'Diffusion Models in Machine Learning: Foundations, generative models and non-convex optimisation', 07/2024

The statistical cost of score-based losses

2024

SIAM Mathematics of Data Science (MDS), 10/2024

The statistical cost of score-based losses

2024

University of College London (UCL), 07/2024

The statistical cost of score-based losses

2024

Webinar for the Section on Statistical Learning and Data Science (SLDS) of the American Statistical Association (ASA), 05/2024

From algorithms to neural networks and back

2023

Mathematics of Modern Machine Learning (M3L) Workshop at NeurIPS 2023, 12/2023 (talk starts at 22:15)

The statistical cost of score matching

2023

IDSS Stochastics and Statistics Seminar, MIT, 02/2023

Fit like you sample: sample-efficient score matching from fast mixing diffusions

2023

Allerton Conference, 09/2023

The statistical cost of score matching

2023

Microsoft Research Redmond, 08/2023

The statistical cost of score matching

2023

Yale FDS Workshop 'Theory and Practice of Foundation Models', 10/2023

The statistical cost of score matching

2023

Meeting in Mathematical Statistics, CIRM, 12/2023

Students

Current Students

Niki Hasrati
Niki Hasrati

PhD in Machine Learning

Stephen Huan
Stephen Huan

PhD in Computer Science

Yuchen Li
Yuchen Li

PhD in Machine Learning

Anna Wei
Anna Wei

PhD in Machine Learning

Alumni

Tanya Marwah
Tanya Marwah

PhD in Machine Learning (2025)

Research Fellow at the Flatiron Institute and Polymathic AI

Bingbin Liu
Bingbin Liu

PhD in Machine Learning (2024)

Kempner Institute Fellow since January 2025

Elan Rosenfeld
Elan Rosenfeld

PhD in Machine Learning (2024)

Research Scientist at Google Research

Kaiyue Wen
Kaiyue Wen

Visiting undergraduate student from Tsinghua University (Spring 2023)

PhD student at Stanford University

Yilong Qin
Yilong Qin

Masters in Machine Learning (MSML)

OpenAI, on leave from Stanford CS PhD program

Edoardo Botta
Edoardo Botta

Masters in Machine Learning (MSML)

Pinterest

Aashay Mehta
Aashay Mehta

Masters in Machine Learning (MSML)

Deep Karkhanis
Deep Karkhanis

Masters in Machine Learning (MSML)

Soundarya Krishnan
Soundarya Krishnan

Masters in Machine Learning (MSML)

Apple

Chirag Pabbaraju
Chirag Pabbaraju

Masters in Machine Learning (MSML)

PhD student at Stanford University

Divyansh Pareek
Divyansh Pareek

Masters in Machine Learning (MSML)

PhD student at University of Washington

Rares Darius-Buhai
Rares Darius-Buhai

Undergraduate at MIT

PhD student at ETH Zurich

Teaching

10.708 (Probabilistic Graphical Models)

10.707 (Advanced Deep Learning)

10.417 (Intermediate Deep Learning)

18.200A (Principles of Discrete and Applied Mathematics) at MIT

Fall 2017/18
Fall 2018/19