Logo

Yeshwanth Cherapanamjeri

Postdoc • CSAILMIT

I am a postdoc at MIT where I am fortunate to be mentored by Costis Daskalakis. I completed my Ph.D at U.C. Berkeley where I was adivsed by Peter Bartlett. Previously, I spent two wonderful years at MSR Bangalore where I worked with Prateek Jain, Nagarajan Natarajan, and Praneeth Netrapalli. Before that, I was advised by Ganesh Ramakrishnan and Soumen Chakrabarti for my undergraduate thesis at IIT Bombay.

I am on the 2024–25 academic job market

My research aims to build reliable machine learning algorithms which can cope with unprecendented challenges in the modern data production pipeline. Here are some directions that I am particularly excited about illustrating this theme:

Inference with Outliers: What are the statistical limits of estimation with outliers in the data? ([CT+22] [CC23] ) and Can we realize these limits computationally? ([CF19] [CH+20] [CM+20] [CA+20] ) These ideas have implications for other questions in statistics and theoretical computer science including classical PAC learning theory ([AC+23] ), and streaming data structures ([CN20] [CN21] [CN22] ).

Data from Strategic Interactions: Sometimes, data is systematically biased due to the strategic behavior of the data-generating agents. For instance, data collected from online marketplaces is not a true reflection of the participants’ utilities. How should we account for this strategic behavior to design statistically and computationally efficient estimators? ([CD+22] [CD+23] )

Dataset Design: Modern datasets are collected from a range of sources of varying data quality with low-quality data being cheap while high-quality data is vastly more expensive. This raises several questions: How should a practitioner prioritize between these sources? Are there inherent limitations of the use of certain data sources? What are the statistical costs of low-quality data? ([DC+24] [CD+25] )

Selected Publications * indicates equal contribution

Are Pairwise Comparisions Enough for Preference Learning?
Yeshwanth Cherapanamjeri*, Constantinos Daskalakis, Gabriele Farina, Sobhan Mohammadpour*
Under Preparation

How Much is a Noisy Image Worth? Data Scaling Laws for Ambient Diffusion
Giannis Daras*, Yeshwanth Cherapanamjeri*, Constantinos Daskalakis
Under Submission

Statistical Barriers to Affine-equivariant Estimation
Zihao Chen, Yeshwanth Cherapanamjeri
Under Submission

Optimal PAC Bounds without Uniform Convergence
Ishaq Aden-Ali, Yeshwanth Cherapanamjeri, Abhishek Shetty, Nikita Zhivotovskiy
Invited to SICOMP Special Issue for FOCS 2023
FOCS 2023

What Makes a Good Fisherman? Linear Regression under Self-Selection Bias
Yeshwanth Cherapanamjeri, Constantinos Daskalakis, Andrew Ilyas, Manolis Zampetakis
STOC 2023

Estimation of Standard Auction Models
Yeshwanth Cherapanamjeri, Constantinos Daskalakis, Andrew Ilyas, Manolis Zampetakis
EC 2022

Algorithms for Heavy-Tailed Statistics: Regression, Covariance Estimation, and Beyond
Yeshwanth Cherapanamjeri, Samuel B. Hopkins, Tarun Kathuria, Prasad Raghavendra, Nilesh Tripuraneni
STOC 2020

Fast Mean Estimation with Sub-Gaussian Rates
Yeshwanth Cherapanamjeri, Nicolas Flammarion, Peter L. Bartlett
COLT 2019

More Publications

Heavy-tailed Contamination is Easier than Adversarial Contamination
Yeshwanth Cherapanamjeri, Daniel Lee
Under Submission

Computing Approximate Centerpoints in Polynomial Time
Yeshwanth Cherapanamjeri
FOCS 2024

The Space Complexity of Learning-Unlearning Algorithms
Yeshwanth Cherapanamjeri, Sumegha Garg, Nived Rajaraman, Ayush Sekhari, Abhishek Shetty
Under Submission

Efficient Automated Circuit Discovery in Transformers using Contextual Decomposition
Aliyah R. Hsu, Georgia Zhou, Yeshwanth Cherapanamjeri, Yaxuan Huang, Anobel Odisho, Peter Carroll, Bin Yu
Under Submission

Diagnosing Transformers: Illuminating Feature Spaces for Clinical Decision-Making
Aliyah R. Hsu, Yeshwanth Cherapanamjeri, Briton Park, Tristan Naumann, Anobel Y. Odisho, Bin Yu
ICLR 2024

The One-Inclusion Graph Algorithm is not Always Optimal
Ishaq Aden-Ali, Yeshwanth Cherapanamjeri, Abhishek Shetty, Nikita Zhivotovskiy
COLT 2023

Optimal Algorithms for Linear Algebra in the Current Matrix Multiplication Time
Yeshwanth Cherapanamjeri, Sandeep Silwal, David P. Woodruff, Samson Zhou
SODA 2023

Robust Algorithms on Adaptive Inputs from Bounded Adversaries
Yeshwanth Cherapanamjeri, Sandeep Silwal, David P. Woodruff, Fred Zhang, Qiuyi Zhang, Samson Zhou
ICLR 2023

Uniform Approximations for Randomized Hadamard Transforms with Applications
Yeshwanth Cherapanamjeri, Jelani Nelson
STOC 2022

Adversarial Examples in Multi-Layer Random ReLU Networks
Peter L. Bartlett, Sébastien Bubeck, Yeshwanth Cherapanamjeri
NeurIPS 2021

A single gradient step finds adversarial examples on random two-layers neural networks
Sébastien Bubeck, Yeshwanth Cherapanamjeri, Gauthier Gidel, Rémi Tachet des Combes
Spotlight Presentation
NeurIPS 2021

Terminal Embeddings in Sublinear Time
Yeshwanth Cherapanamjeri, Jelani Nelson
FOCS 2021

On Adaptive Distance Estimation
Yeshwanth Cherapanamjeri, Jelani Nelson
Spotlight Presentation
NeurIPS 2020

Optimal Robust Linear Regression in Nearly Linear Time
Yeshwanth Cherapanamjeri, Efe Aras, Nilesh Tripuraneni, Michael I. Jordan, Nicolas Flammarion, Peter L. Bartlett
Under Submission

List Decodable Mean Estimation in Nearly Linear Time
Yeshwanth Cherapanamjeri, Sidhanth Mohanty, Morris Yau
FOCS 2020

Optimal Mean Estimation without a Variance
Yeshwanth Cherapanamjeri, Nilesh Tripuraneni, Peter L. Bartlett, Michael I. Jordan
COLT 2022

Testing Markov Chains Without Hitting
Yeshwanth Cherapanamjeri, Peter L. Bartlett
COLT 2019

Thresholding based Efficient Outlier Robust PCA
Yeshwanth Cherapanamjeri, Prateek Jain, Praneeth Netrapalli
COLT 2017

Nearly Optimal Robust Matrix Completion
Yeshwanth Cherapanamjeri, Kartik Gupta, Prateek Jain
ICML 2017