Hi, Thanks for your interest!
I am an assistant professor at the Department of Computer Science, University of Illinois (UIC), Chicago, since Fall, 2025. I am a member of UIC CS Theory group, as well as IDEAL Institute. Previously, I was a research scientist at Apple MLR, broadly working in the area of Machine Learning theory. I completed my postdoc at Microsoft Research (NYC) and PhD from the department of Computer Science, Indian Institute of Science, Bangalore. I was fortunate to be advised by Aditya Gopalan and Chiranjib Bhattacharyya during my PhD, and to have been a part of several internships, including Microsoft Research (In), Inria (Paris), and Google AI (MTV).
Research Focus: Building Personalized Educator (through Human-Aligned Learning)Main Research Area: AI-Alignment with Reinforcement Learning with Human Feedback (RLHF)
Related Research Areas: Machine Learning (esp. Online Learning Bandits, RL theory), Optimization, Federated Learning, Quantum Information Theory, Differential Privacy, and Mechanism Design.
Brief Research Summary
My primary research focuses on developing, improving, and theoretically analyzing reinforcement learning with human feedback (RLHF) algorithms to enhance personalized machine assistance for humanity. Applications include language models, assistive robotics, autonomous driving, and personalized systems—-almost any system that can improve itself through user interaction.
Specifically, I am deeply motivated by the tremendous potential of AI to democratize learning—reshaping our current education system into a truly adaptive, accessible, and personalized experience for every learner! Driven by this transformative power of generative AI and language models, we hope to design foundations for equitable, intelligent education systems that turn this vision into reality. My research focuses on developing futuristic educational models by leveraging RLHF theory alongside tools from Machine Learning (Online Learning, Bandits, and RL theory), Optimization, Federated Learning, Differential Privacy, and Mechanism Design.
A bit more generally, my work focuses on building large-scale, robust, personalized AI models for sequential decision-making under partial or restricted feedback, such as user interactions, preferences, demonstrations, proxy observations, and rankings. In the past, I have also explored combinatorial decision spaces, dynamic regret, multiplayer games, and distributed optimization. My work falls under various interdisciplinary research areas, including ML, learning theory, optimization, operations research, mechanism design, privacy, federated learning, and algorithmic fairness. Recently, I have also been spending considerable time exploring Quantum Information Theory and its potential to improve algorithmic efficiency across different classes of ML algorithms. Feel free to reach out if you'd like to discuss any of these topics!
[Selected Papers] [Full List] [Google Scholar] [DBLP] [arXiv]
Prospective Students: I am looking to work with ML enthusiasts and motivated students! If you are interested in joining our research group, please send me your resume and research statement. I also encourage you to apply to the UIC-CS PhD program (and do mention my name in your application).
Selected Papers:
- Efficient and Near-Optimal Algorithm for General Contextual Dueling Bandits with Offline Regression Oracles [Arxiv version]
Aadirupa Saha, Robert Schapire
In Neural Information Processing Systems, NeurIPS 2025 - Dueling Convex Optimization for General Preferences: An Unified Framework for Optimal Convergence Rates. [Arxiv Version]
Aadirupa Saha, Tomer Koren, Yishay Mansour
International Conference on Machine Learning, ICML 2025 - Stop Relying on No-Choice and Do not Repeat the Moves: Optimal, Efficient and Practical Algorithms for Assortment Optimization
Aadirupa Saha, Pierre Gaillard
International Conference on Learning Representations (ICLR), 2025 - Only Pay for What Is Uncertain: Variance-Adaptive Thompson Sampling. [Arxiv version]
Aadirupa Saha, Branislav Kveton
International Conference on Learning Representations (ICLR), 2024 - Federated Online and Bandit Convex Optimization [Arxiv Version]
Kumar Kshitij Patel, Lingxiao Wang, Aadirupa Saha, Nathan Srebro
In International Conference on Machine Learning, ICML 2023 - ANACONDA: Improved Dynamic Regret Algorithm for Adaptive Non-Stationary Dueling Bandits [Arxiv Version]
Thomas Kleine Buening, Aadirupa Saha
In International Conference on Artificial Intelligence and Statistics, AIStats 2023 - Versatile Dueling Bandits [Arxiv Version]
Aadirupa Saha, Pierre Gaillard
In International Conference on Machine Learning, ICML 2022 - Efficient and Optimal Algorithms for Contextual Dueling Bandits under Realizability [Arxiv Version]
Aadirupa Saha, Akshay Krishnamurthy
In Algorithmic Learning Theory, ALT 2022 - Optimal Algorithms for Stochastic Contextual Dueling Bandits
Aadirupa Saha
In Neural Information Processing Systems, NeurIPS 2021 - From PAC to Instance-Optimal Sample Complexity in the Plackett-Luce Model [Arxiv Version]
Aadirupa Saha, Aditya Gopalan
In International Conference on Machine Learning, ICML 2020
