Omar Chehab

Bonjour! Welcome to my website. I am a postdoc at Carnegie Mellon University, in the Machine Learning Department , working with Pradeep Ravikumar .

I completed my graduate studies in France: this includes a PhD in Mathematical Computer Science at Inria with Aapo Hyvärinen and Alexandre Gramfort, and a postdoc in the Statistics Department of CREST-ENSAE with Anna Korba .

More details are listed in my CV and on this profile page. You can reach me via email, find my code on GitHub, browse my publications on Google Scholar, or connect on LinkedIn.

Research Talks Notes Teaching Service
profile photo

Research

My research is in machine learning, with the following goals:

  • Learning likelihoods from data
  • Generating samples from an unnormalized distribution and/or data
  • Learning representations from data
  • Learning causal structure from data

The algorithms I work with to achieve these goals range from diffusions and flows to multi-view independent component analysis.

All of the above goals can be framed as learning a probability distribution. A central question guiding my work is: “How much compute and data are required to achieve a given level of accuracy?”. For which distributions does this work or fail. And how does the algorithmic design impact this. In other words, I aim to quantify the computational and sample efficiency of these algorithms.

I study these questions across multiple data modalities: synthetic, image, and brain data. The synthetic data is designed to have certain statistical properties (e.g., multimodality); the standard image datasets are commonly used as benchmarks in the academic community, and the non-invasive brain recordings were central to my PhD work and remain an active research direction.

Energy-Based Models Diffusion and Flow Models Sampling Algorithms Score Matching Density Ratio Estimation Causal Discovery Representation Learning Brain Imaging Multi-View Independent Component Analysis

Sampling from Energy-Based Statistical Models

project image
Sampling from multi-modal distributions with polynomial query complexity in fixed dimension via reverse diffusion
Adrien Vacher, Omar Chehab, Anna Korba
Conference on Neural Information Processing Systems (NeurIPS), 2025

Time-reversed diffusions are state-of-the-art for sampling multi-modal distributions, but they rely on score estimates. We analyze how estimation errors affect the final samples.

project image
Provable Convergence and Limitations of Geometric Tempering for Langevin Dynamics
Omar Chehab, Anna Korba, Austin Stromme, Adrien Vacher
International Conference on Learning Representations (ICLR), 2025

Annealed MCMC tries to approximate a prescribed path of distributions. We show that the popular geometric mean path with a Gaussian has unfavorable geometry. Presented at the Yale workshop on sampling.

project image
A Practical Diffusion Path for Sampling
Omar Chehab, Anna Korba
Workshop on Structured Probabilistic Inference & Generative Modeling, International Conference on Machine Learning (ICML), 2024

Time-reversed diffusions are state-of-the-art in sampling but rely on score estimates. We aim to reduce their variance.

Estimating Energy-Based Statistical Models

project image
Conditional Noise-Contrastive Estimation of Energy-Based Models by Jumping Between Modes
Hanlin Yu, Michael U. Gutmann, Arto Klami, Omar Chehab
Workshop on Principles of Generative Modeling, EurIPS, 2025

We explore the design choices of a method called CNCE for learning energy-based models.

project image
Density Ratio Estimation with Conditional Probability Paths
Hanlin Yu, Arto Klami, Aapo Hyvärinen, Anna Korba, Omar Chehab
International Conference on Machine Learning (ICML), 2025

A density ratio can be obtained by integrating the time score of a probability path. We present an efficient way to estimate the time score.

project image
Optimizing the Noise in Self-Supervised Learning: from Importance Sampling to Noise-Contrastive Estimation
Omar Chehab, Alexandre Gramfort, Aapo Hyvärinen
arXiv, 2023
project image
Provable benefits of annealing for estimating normalizing constants: Importance Sampling, Noise-Contrastive Estimation, and beyond
Omar Chehab, Aapo Hyvärinen, Andrej Risteski
Spotlight, Conference on Neural Information Processing Systems (NeurIPS), 2023

Annealed Importance Sampling uses a prescribed path of distributions to compute an estimate of a normalizing constant. We quantify how the choice of path impacts the estimation error.

project image
The Optimal Noise in Noise-Contrastive Learning Is Not What You Think
Omar Chehab, Alexandre Gramfort, Aapo Hyvärinen
Conference on Uncertainty in Artificial Intelligence (UAI), 2022

NCE estimates the data density by minimizing a binary classification loss, between data and noise samples. We find the optimal noise distribution that minimizes the estimation error.

Learning Representations of Brain Activity

project image
Multi-View Causal Discovery without Non-Gaussianity: Identifiability and Algorithms
Ambroise Heurtebise, Omar Chehab, Pierre Ablin, Alexandre Gramfort, Aapo Hyvärinen
Oral, Workshop on Causality for Impact - Practical challenges for real-world applications of causal methods, EurIPS, 2025

We learn causal relationships (Directed Acyclic Graph) between random variables collected from different environments.

project image
MVICAD2: Multi-View Independent Component Analysis with Delays and Dilations
Ambroise Heurtebise, Omar Chehab, Pierre Ablin, Alexandre Gramfort
IEEE Transactions on Biomedical Engineering, 2025

Independent Component Analysis (ICA) is a popular algorithm for learning a representation of data. We propose a version that handles data collected from different contexts, and whose representations differ only by temporal delays or dilations.

project image
Deep Recurrent Encoder: an end-to-end network to model magnetoencephalography at scale
Omar Chehab*, Alexandre Defossez*, Jean-Christophe Loiseau, Alexandre Gramfort, Jean-Remi King
Journal of Neurons, Behavior, Data analysis, and Theory, 2022

We compare different models for predicting the brain’s response to external stimuli. Our model, based on a deep neural network, is more accurate and interpretable.

project image
Learning with self-supervision on EEG data
Alexandre Gramfort, Hubert Banville, Omar Chehab, Aapo Hyvärinen, Denis Engemann
IEEE workshop on Brain-Computer Interface, 2021

We learn rich representations of EEG brain activity using a self-supervised loss.

project image
Uncovering the structure of clinical EEG signals with self-supervised learning
Hubert Banville, Omar Chehab, Aapo Hyvärinen, Denis Engemann, Alexandre Gramfort
Journal of Neural Engineering, 2021

We learn rich representations of EEG brain activity using a self-supervised loss.

project image
A mean-field approach to the dynamics of networks of complex neurons, from nonlinear Integrate-and-Fire to Hodgkin–Huxley models
Mallory Carlu, Omar Chehab, Leonardo Dalla Porta, Damien Depannemaecker, Charlotte Héricé, Maciej Jedynak, Elif Köksal Ersöz, Paulo Muratore, Selma Souihe, Cristiano Capone, Yann Zerlaut, Alain Destexhe, Matteo di Volo
Journal of Neurophysiology, 2020

Our theory predicts the average behavior of neuronal populations that fire asynchronously.




Talks

Notes

Cheatsheet of computer commands

Teaching

I was a teaching assistant for the following Master’s courses.

Optimization for Data Science 2021-2023
Institut Polytechnique de Paris
Professors: Alexandre Gramfort, Pierre Ablin
Advanced Machine Learning 2020-2022
CentraleSupelec, Universite Paris-Saclay
Professors: Emilie Chouzenoux, Frederic Pascal
Optimization 2020-2021
CentraleSupelec, Universite Paris-Saclay
Professors: Jean-Christophe Pesquet, Sorin Olaru, Stephane Font

Service

I review submissions to the following machine learning conferences: NeurIPS, ICML, ICLR and AISTATS.

I also occasionally review submissions to these journals of statistics or machine learning: JMLR, TMRL, AISM, and Mathematical Methods of Statistics.

I am grateful to have been recognized as a "top reviewer" (AISTATS 2022, NeurIPS 2022-23-24).


Design and source code from Jon Barron's website