Kavosh Asadi

According to our database1, Kavosh Asadi authored at least 29 papers between 2016 and 2023.

Collaborative distances:
  • Dijkstra number2 of four.
  • Erdős number3 of four.

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

On csauthors.net:

Bibliography

2023
TAIL: Task-specific Adapters for Imitation Learning with Large Pretrained Models.
CoRR, 2023

TD Convergence: An Optimization Perspective.
Proceedings of the Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, 2023

Resetting the Optimizer in Deep RL: An Empirical Study.
Proceedings of the Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, 2023

Coarse-Grained Smoothness for Reinforcement Learning in Metric Spaces.
Proceedings of the International Conference on Artificial Intelligence and Statistics, 2023

2022
Characterizing the Action-Generalization Gap in Deep Q-Learning.
CoRR, 2022

Adaptive Interest for Emphatic Reinforcement Learning.
Proceedings of the Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, 2022

Faster Deep Reinforcement Learning with Slower Online Network.
Proceedings of the Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, 2022

2021
Smoothness in Reinforcement Learning with Large State and Action Spaces.
PhD thesis, 2021

Deep Q-Network with Proximal Iteration.
CoRR, 2021

Coarse-Grained Smoothness for RL in Metric Spaces.
CoRR, 2021

Convergence of a Human-in-the-Loop Policy-Gradient Algorithm With Eligibility Trace Under Reward, Policy, and Advantage Feedback.
CoRR, 2021

Continuous Doubly Constrained Batch Reinforcement Learning.
Proceedings of the Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, 2021

Lipschitz Lifelong Reinforcement Learning.
Proceedings of the Thirty-Fifth AAAI Conference on Artificial Intelligence, 2021

Deep Radial-Basis Value Functions for Continuous Control.
Proceedings of the Thirty-Fifth AAAI Conference on Artificial Intelligence, 2021

2020
Learning State Abstractions for Transfer in Continuous Control.
CoRR, 2020

Deep RBF Value Functions for Continuous Control.
CoRR, 2020

2019
Combating the Compounding-Error Problem with a Multi-step Model.
CoRR, 2019

DeepMellow: Removing the Need for a Target Network in Deep Q-Learning.
Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, 2019

Removing the Target Network from Deep Q-Networks with the Mellowmax Operator.
Proceedings of the 18th International Conference on Autonomous Agents and MultiAgent Systems, 2019

State Abstraction as Compression in Apprenticeship Learning.
Proceedings of the Thirty-Third AAAI Conference on Artificial Intelligence, 2019

2018
Mitigating Planner Overfitting in Model-Based Reinforcement Learning.
CoRR, 2018

Towards a Simple Approach to Multi-step Model-based Reinforcement Learning.
CoRR, 2018

Equivalence Between Wasserstein and Value-Aware Model-based Reinforcement Learning.
CoRR, 2018

Lipschitz Continuity in Model-based Reinforcement Learning.
Proceedings of the 35th International Conference on Machine Learning, 2018

2017
Mean Actor Critic.
CoRR, 2017

An Alternative Softmax Operator for Reinforcement Learning.
Proceedings of the 34th International Conference on Machine Learning, 2017

Hybrid Code Networks: practical and efficient end-to-end dialog control with supervised and reinforcement learning.
Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics, 2017

2016
Sample-efficient Deep Reinforcement Learning for Dialog Control.
CoRR, 2016

A New Softmax Operator for Reinforcement Learning.
CoRR, 2016


  Loading...