Alberto Maria Metelli

Orcid: 0000-0002-3424-5212

According to our database1, Alberto Maria Metelli authored at least 57 papers between 2017 and 2024.

Collaborative distances:

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

Online presence:

On csauthors.net:

Bibliography

2024
Offline Inverse RL: New Solution Concepts and Provably Efficient Algorithms.
CoRR, 2024

Performance Improvement Bounds for Lipschitz Configurable Markov Decision Processes.
CoRR, 2024

Information Capacity Regret Bounds for Bandits with Mediator Feedback.
CoRR, 2024

No-Regret Reinforcement Learning in Smooth MDPs.
CoRR, 2024

Inverse Reinforcement Learning with Sub-optimal Experts.
CoRR, 2024

Parameterized Projected Bellman Operator.
Proceedings of the Thirty-Eighth AAAI Conference on Artificial Intelligence, 2024

Recent Advancements in Inverse Reinforcement Learning.
Proceedings of the Thirty-Eighth AAAI Conference on Artificial Intelligence, 2024

2023
IWDA: Importance Weighting for Drift Adaptation in Streaming Supervised Learning Problems.
IEEE Trans. Neural Networks Learn. Syst., October, 2023

ARLO: A framework for Automated Reinforcement Learning.
Expert Syst. Appl., August, 2023

Causal Feature Selection via Transfer Entropy.
CoRR, 2023

Towards Fully Adaptive Regret Minimization in Heavy-Tailed Bandits.
CoRR, 2023

Pure Exploration under Mediators' Feedback.
CoRR, 2023

Nonlinear Feature Aggregation: Two Algorithms driven by Theory.
CoRR, 2023

An Option-Dependent Analysis of Regret Minimization Algorithms in Finite-Horizon Semi-Markov Decision Processes.
CoRR, 2023

Interpretable Linear Dimensionality Reduction based on Bias-Variance Analysis.
CoRR, 2023

Best Arm Identification for Stochastic Rising Bandits.
CoRR, 2023

On the Relation between Policy Improvement and Off-Policy Minimum-Variance Policy Evaluation.
Proceedings of the Uncertainty in Artificial Intelligence, 2023

Distributional Policy Evaluation: a Maximum Entropy approach to Representation Learning.
Proceedings of the Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, 2023

Truncating Trajectories in Monte Carlo Policy Evaluation: an Adaptive Approach.
Proceedings of the Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, 2023

Information-Theoretic Regret Bounds for Bandits with Fixed Expert Advice.
Proceedings of the IEEE Information Theory Workshop, 2023

Truncating Trajectories in Monte Carlo Reinforcement Learning.
Proceedings of the International Conference on Machine Learning, 2023

Dynamical Linear Bandits.
Proceedings of the International Conference on Machine Learning, 2023

Towards Theoretical Understanding of Inverse Reinforcement Learning.
Proceedings of the International Conference on Machine Learning, 2023

A Tale of Sampling and Estimation in Discounted Reinforcement Learning.
Proceedings of the International Conference on Artificial Intelligence and Statistics, 2023

Simultaneously Updating All Persistence Values in Reinforcement Learning.
Proceedings of the Thirty-Seventh AAAI Conference on Artificial Intelligence, 2023

Tight Performance Guarantees of Imitator Policies with Continuous Actions.
Proceedings of the Thirty-Seventh AAAI Conference on Artificial Intelligence, 2023

Wasserstein Actor-Critic: Directed Exploration via Optimism for Continuous-Actions Control.
Proceedings of the Thirty-Seventh AAAI Conference on Artificial Intelligence, 2023

2022
Exploiting environment configurability in reinforcement learning.
Frontiers in Artificial Intelligence and Applications 361, IOS Press, ISBN: 978-1-64368-363-8, 2022

Policy space identification in configurable environments.
Mach. Learn., 2022

A unified view of configurable Markov Decision Processes: Solution concepts, value functions, and operators.
Intelligenza Artificiale, 2022

Autoregressive Bandits.
CoRR, 2022

Optimizing Empty Container Repositioning and Fleet Deployment via Configurable Semi-POMDPs.
CoRR, 2022

Multi-Fidelity Best-Arm Identification.
Proceedings of the Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, 2022

Storehouse: a Reinforcement Learning Environment for Optimizing Warehouse Management.
Proceedings of the International Joint Conference on Neural Networks, 2022

Stochastic Rising Bandits.
Proceedings of the International Conference on Machine Learning, 2022

Balancing Sample Efficiency and Suboptimality in Inverse Reinforcement Learning.
Proceedings of the International Conference on Machine Learning, 2022

Trust Region Meta Learning for Policy Optimization.
Proceedings of the ECML/PKDD Workshop on Meta-Knowledge Transfer, 2022

Lifelong Hyper-Policy Optimization with Multiple Importance Sampling Regularization.
Proceedings of the Thirty-Sixth AAAI Conference on Artificial Intelligence, 2022

2021
Dealing with multiple experts and non-stationarity in inverse reinforcement learning: an application to real-life problems.
Mach. Learn., 2021

Safe Policy Iteration: A Monotonically Improving Approximate Policy Iteration Approach.
J. Mach. Learn. Res., 2021

Learning in Non-Cooperative Configurable Markov Decision Processes.
Proceedings of the Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, 2021

Subgaussian and Differentiable Importance Sampling for Off-Policy Evaluation and Learning.
Proceedings of the Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, 2021

Provably Efficient Learning of Transferable Rewards.
Proceedings of the 38th International Conference on Machine Learning, 2021

Policy Optimization as Online Learning with Mediator Feedback.
Proceedings of the Thirty-Fifth AAAI Conference on Artificial Intelligence, 2021

2020
Combining reinforcement learning with rule-based controllers for transparent and general decision-making in autonomous driving.
Robotics Auton. Syst., 2020

Importance Sampling Techniques for Policy Optimization.
J. Mach. Learn. Res., 2020

On the use of the policy gradient and Hessian in inverse reinforcement learning.
Intelligenza Artificiale, 2020

Control Frequency Adaptation via Action Persistence in Batch Reinforcement Learning.
Proceedings of the 37th International Conference on Machine Learning, 2020

Truly Batch Model-Free Inverse Reinforcement Learning about Multiple Intentions.
Proceedings of the 23rd International Conference on Artificial Intelligence and Statistics, 2020

Gradient-Aware Model-Based Policy Search.
Proceedings of the Thirty-Fourth AAAI Conference on Artificial Intelligence, 2020

2019
Propagating Uncertainty in Reinforcement Learning via Wasserstein Barycenters.
Proceedings of the Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, 2019

Feature Selection via Mutual Information: New Theoretical Insights.
Proceedings of the International Joint Conference on Neural Networks, 2019

Optimistic Policy Optimization via Multiple Importance Sampling.
Proceedings of the 36th International Conference on Machine Learning, 2019

Reinforcement Learning in Configurable Continuous Environments.
Proceedings of the 36th International Conference on Machine Learning, 2019

2018
Policy Optimization via Importance Sampling.
Proceedings of the Advances in Neural Information Processing Systems 31: Annual Conference on Neural Information Processing Systems 2018, 2018

Configurable Markov Decision Processes.
Proceedings of the 35th International Conference on Machine Learning, 2018

2017
Compatible Reward Inverse Reinforcement Learning.
Proceedings of the Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, 2017


  Loading...