Eric Mazumdar

Orcid: 0000-0002-1815-269X

Affiliations:
  • California Institute of Technology, Pasadena, CA, USA


According to our database1, Eric Mazumdar authored at least 42 papers between 2016 and 2024.

Collaborative distances:

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

Online presence:

On csauthors.net:

Bibliography

2024
Rethinking Scaling Laws for Learning in Strategic Environments.
CoRR, 2024

2023
Two-Timescale Q-Learning with Function Approximation in Zero-Sum Stochastic Games.
CoRR, 2023

Coupled Gradient Flows for Strategic Non-Local Distribution Shift.
CoRR, 2023

Convergent First-Order Methods for Bi-level Optimization and Stackelberg Games.
CoRR, 2023

Strategic Distribution Shift of Interacting Agents via Coupled Gradient Flows.
Proceedings of the Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, 2023

A Finite-Sample Analysis of Payoff-Based Independent Learning in Zero-Sum Stochastic Games.
Proceedings of the Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, 2023

Designing System Level Synthesis Controllers for Nonlinear Systems with Stability Guarantees.
Proceedings of the Learning for Dynamics and Control Conference, 2023

Synthesizing Reactive Test Environments for Autonomous Systems: Testing Reach-Avoid Specifications with Multi-Commodity Flows.
Proceedings of the IEEE International Conference on Robotics and Automation, 2023

Algorithmic Collective Action in Machine Learning.
Proceedings of the International Conference on Machine Learning, 2023

2022
A Note on Zeroth-Order Optimization on the Simplex.
CoRR, 2022

Decentralized, Communication- and Coordination-free Learning in Structured Matching Markets.
Proceedings of the Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, 2022

Langevin Monte Carlo for Contextual Bandits.
Proceedings of the International Conference on Machine Learning, 2022

Nonlinear System Level Synthesis for Polynomial Dynamical Systems.
Proceedings of the 61st IEEE Conference on Decision and Control, 2022

Fast Distributionally Robust Learning with Variance-Reduced Min-Max Optimization.
Proceedings of the International Conference on Artificial Intelligence and Statistics, 2022

Zeroth-Order Methods for Convex-Concave Min-max Problems: Applications to Decision-Dependent Risk Minimization.
Proceedings of the International Conference on Artificial Intelligence and Statistics, 2022

2021
Zeroth-Order Methods for Convex-Concave Minmax Problems: Applications to Decision-Dependent Risk Minimization.
CoRR, 2021

Who Leads and Who Follows in Strategic Classification?
Proceedings of the Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, 2021

Global Convergence to Local Minmax Equilibrium in Classes of Nonconvex Zero-Sum Games.
Proceedings of the Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, 2021

2020
Inverse Risk-Sensitive Reinforcement Learning.
IEEE Trans. Autom. Control., 2020

On Gradient-Based Learning in Continuous Games.
SIAM J. Math. Data Sci., 2020

Technical Report: Adaptive Control for Linearizable Systems Using On-Policy Reinforcement Learning.
CoRR, 2020

On Thompson Sampling with Langevin Algorithms.
CoRR, 2020

Feedback Linearization for Uncertain Systems via Reinforcement Learning.
Proceedings of the 2020 IEEE International Conference on Robotics and Automation, 2020

On Approximate Thompson Sampling with Langevin Algorithms.
Proceedings of the 37th International Conference on Machine Learning, 2020

Adaptive Control for Linearizable Systems Using On-Policy Reinforcement Learning.
Proceedings of the 59th IEEE Conference on Decision and Control, 2020

Expert Selection in High-Dimensional Markov Decision Processes.
Proceedings of the 59th IEEE Conference on Decision and Control, 2020

High Confidence Sets for Trajectories of Stochastic Time-Varying Nonlinear Systems.
Proceedings of the 59th IEEE Conference on Decision and Control, 2020

Policy-Gradient Algorithms Have No Guarantees of Convergence in Linear Quadratic Games.
Proceedings of the 19th International Conference on Autonomous Agents and Multiagent Systems, 2020

2019
Feedback Linearization for Unknown Systems via Reinforcement Learning.
CoRR, 2019

Policy-Gradient Algorithms Have No Guarantees of Convergence in Continuous Action and State Multi-Agent Settings.
CoRR, 2019

Convergence Analysis of Gradient-Based Learning with Non-Uniform Learning Rates in Non-Cooperative Multi-Agent Settings.
CoRR, 2019

On Finding Local Nash Equilibria (and Only Local Nash Equilibria) in Zero-Sum Games.
CoRR, 2019

Convergence Analysis of Gradient-Based Learning in Continuous Games.
Proceedings of the Thirty-Fifth Conference on Uncertainty in Artificial Intelligence, 2019

Local Nash Equilibria are Isolated, Strict Local Nash Equilibria in 'Almost All' Zero-Sum Continuous Games.
Proceedings of the 58th IEEE Conference on Decision and Control, 2019

2018
On the Convergence of Competitive, Multi-Agent Gradient-Based Learning.
CoRR, 2018

On the Analysis of Cyclic Drug Schedules for Cancer Treatment using Switched Dynamical Systems.
Proceedings of the 57th IEEE Conference on Decision and Control, 2018

2017
Risk-Sensitive Inverse Reinforcement Learning via Gradient Methods.
CoRR, 2017

A Multi-Armed Bandit Approach for Online Expert Selection in Markov Decision Processes.
CoRR, 2017

Optimal Causal Imputation for Control.
CoRR, 2017

Gradient-based inverse risk-sensitive reinforcement learning.
Proceedings of the 56th IEEE Annual Conference on Decision and Control, 2017

2016
To observe or not to observe: Queuing game framework for urban parking.
Proceedings of the 55th IEEE Conference on Decision and Control, 2016

Understanding the impact of parking on urban mobility via routing games on queue-flow networks.
Proceedings of the 55th IEEE Conference on Decision and Control, 2016


  Loading...