Olivier Pietquin

Orcid: 0000-0002-5386-465X

Affiliations:
  • Google DeepMind
  • University Lille 1, France


According to our database1, Olivier Pietquin authored at least 204 papers between 2002 and 2024.

Collaborative distances:

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

Online presence:

On csauthors.net:

Bibliography

2024
Language Evolution with Deep Learning.
CoRR, 2024

Population-aware Online Mirror Descent for Mean-Field Games by Deep Reinforcement Learning.
CoRR, 2024

Back to Basics: Revisiting REINFORCE Style Optimization for Learning from Human Feedback in LLMs.
CoRR, 2024

MusicRL: Aligning Music Generation to Human Preferences.
CoRR, 2024

Learning Discrete-Time Major-Minor Mean Field Games.
Proceedings of the Thirty-Eighth AAAI Conference on Artificial Intelligence, 2024

2023
AudioLM: A Language Modeling Approach to Audio Generation.
IEEE ACM Trans. Audio Speech Lang. Process., 2023

Get Back Here: Robust Imitation by Return-to-Distribution Planning.
CoRR, 2023

Speak, Read and Prompt: High-Fidelity Text-to-Speech with Minimal Supervision.
CoRR, 2023

SingSong: Generating musical accompaniments from singing.
CoRR, 2023

On Imitation in Mean-field Games.
Proceedings of the Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, 2023

Regularization and Variance-Weighted Regression Achieves Minimax Optimality in Linear MDPs: Theory and Practice.
Proceedings of the International Conference on Machine Learning, 2023

Factually Consistent Summarization via Reinforcement Learning with Textual Entailment Feedback.
Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2023

2022
C3PO: Learning to Achieve Arbitrary Goals via Massively Entropic Pretraining.
CoRR, 2022

vec2text with Round-Trip Translations.
CoRR, 2022

AudioLM: a Language Modeling Approach to Audio Generation.
CoRR, 2022

Learning Correlated Equilibria in Mean-Field Games.
CoRR, 2022

KL-Entropy-Regularized RL with a Generative Model is Minimax Optimal.
CoRR, 2022

Learning Mean Field Games: A Survey.
CoRR, 2022

Lazy-MDPs: Towards Interpretable Reinforcement Learning by Learning When to Act.
CoRR, 2022

Emergent Communication: Generalization and Overfitting in Lewis Games.
Proceedings of the Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, 2022

Learning Natural Language Generation with Truncated Reinforcement Learning.
Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 2022

Scalable Deep Reinforcement Learning Algorithms for Mean Field Games.
Proceedings of the International Conference on Machine Learning, 2022

Continuous Control with Action Quantization from Demonstrations.
Proceedings of the International Conference on Machine Learning, 2022

On the role of population heterogeneity in emergent communication.
Proceedings of the Tenth International Conference on Learning Representations, 2022

Scaling Mean Field Games by Online Mirror Descent.
Proceedings of the 21st International Conference on Autonomous Agents and Multiagent Systems, 2022

Learning Equilibria in Mean-Field Games: Introducing Mean-Field PSRO.
Proceedings of the 21st International Conference on Autonomous Agents and Multiagent Systems, 2022

Lazy-MDPs: Towards Interpretable RL by Learning When to Act.
Proceedings of the 21st International Conference on Autonomous Agents and Multiagent Systems, 2022

Concave Utility Reinforcement Learning: The Mean-field Game Viewpoint.
Proceedings of the 21st International Conference on Autonomous Agents and Multiagent Systems, 2022

Solving N-Player Dynamic Routing Games with Congestion: A Mean-Field Approach.
Proceedings of the 21st International Conference on Autonomous Agents and Multiagent Systems, 2022

Implicitly Regularized RL with Implicit Q-values.
Proceedings of the International Conference on Artificial Intelligence and Statistics, 2022

Offline Reinforcement Learning as Anti-exploration.
Proceedings of the Thirty-Sixth AAAI Conference on Artificial Intelligence, 2022

Generalization in Mean Field Games by Learning Master Policies.
Proceedings of the Thirty-Sixth AAAI Conference on Artificial Intelligence, 2022

2021
RLDS: an Ecosystem to Generate, Share and Use Datasets in Reinforcement Learning.
CoRR, 2021

Learning Natural Language Generation from Scratch.
CoRR, 2021

Scaling up Mean Field Games with Online Mirror Descent.
CoRR, 2021

What Matters for Adversarial Imitation Learning?
Proceedings of the Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, 2021

There Is No Turning Back: A Self-Supervised Approach for Reversibility-Aware Reinforcement Learning.
Proceedings of the Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, 2021

Don't Do What Doesn't Matter: Intrinsic Motivation with Action Usefulness.
Proceedings of the Thirtieth International Joint Conference on Artificial Intelligence, 2021

Mean Field Games Flock! The Reinforcement Learning Way.
Proceedings of the Thirtieth International Joint Conference on Artificial Intelligence, 2021

Hyperparameter Selection for Imitation Learning.
Proceedings of the 38th International Conference on Machine Learning, 2021

Offline Reinforcement Learning with Pseudometric Learning.
Proceedings of the 38th International Conference on Machine Learning, 2021

Adversarially Guided Actor-Critic.
Proceedings of the 9th International Conference on Learning Representations, 2021

Primal Wasserstein Imitation Learning.
Proceedings of the 9th International Conference on Learning Representations, 2021

What Matters for On-Policy Deep Actor-Critic Methods? A Large-Scale Study.
Proceedings of the 9th International Conference on Learning Representations, 2021

Learning From Heterogeneous Eeg Signals with Differentiable Channel Reordering.
Proceedings of the IEEE International Conference on Acoustics, 2021

Show Me the Way: Intrinsic Motivation from Demonstrations.
Proceedings of the AAMAS '21: 20th International Conference on Autonomous Agents and Multiagent Systems, 2021

Self-Imitation Advantage Learning.
Proceedings of the AAMAS '21: 20th International Conference on Autonomous Agents and Multiagent Systems, 2021

2020
The Monte Carlo Transformer: a stochastic self-attention model for sequence prediction.
CoRR, 2020

What Matters In On-Policy Reinforcement Learning? A Large-Scale Empirical Study.
CoRR, 2020

Reinforcement Learning.
CoRR, 2020

Leverage the Average: an Analysis of Regularization in RL.
CoRR, 2020

HIGhER: Improving instruction following with Hindsight Generation for Experience Replay.
Proceedings of the 2020 IEEE Symposium Series on Computational Intelligence, 2020

Munchausen Reinforcement Learning.
Proceedings of the Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, 2020

Leverage the Average: an Analysis of KL Regularization in Reinforcement Learning.
Proceedings of the Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, 2020

Fictitious Play for Mean Field Games: Continuous Time Analysis and Applications.
Proceedings of the Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, 2020

A Machine of Few Words: Interactive Speaker Recognition with Reinforcement Learning.
Proceedings of the Interspeech 2020, 2020

"I'm Sorry Dave, I'm Afraid I Can't Do That" Deep Q-Learning from Forbidden Actions.
Proceedings of the 2020 International Joint Conference on Neural Networks, 2020

Self-Attentional Credit Assignment for Transfer in Reinforcement Learning.
Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence, 2020

Countering Language Drift with Seeded Iterated Learning.
Proceedings of the 37th International Conference on Machine Learning, 2020

Supervised Seeded Iterated Learning for Interactive Language Learning.
Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing, 2020

CopyCAT: : Taking Control of Neural Policies with Constant Attacks.
Proceedings of the 19th International Conference on Autonomous Agents and Multiagent Systems, 2020

Momentum in Reinforcement Learning.
Proceedings of the 23rd International Conference on Artificial Intelligence and Statistics, 2020

Foolproof Cooperative Learning.
Proceedings of The 12th Asian Conference on Machine Learning, 2020

Deep Conservative Policy Iteration.
Proceedings of the Thirty-Fourth AAAI Conference on Artificial Intelligence, 2020

On the Convergence of Model Free Learning in Mean Field Games.
Proceedings of the Thirty-Fourth AAAI Conference on Artificial Intelligence, 2020

2019
On Connections between Constrained Optimization and Reinforcement Learning.
CoRR, 2019

"I'm sorry Dave, I'm afraid I can't do that" Deep Q-learning from forbidden action.
CoRR, 2019

Credit Assignment as a Proxy for Transfer in Reinforcement Learning.
CoRR, 2019

Approximate Fictitious Play for Mean Field Games.
CoRR, 2019

MULEX: Disentangling Exploitation from Exploration in Deep RL.
CoRR, 2019

Targeted Attacks on Deep Reinforcement Learning Agents through Adversarial Observations.
CoRR, 2019

Scaling up budgeted reinforcement learning.
CoRR, 2019

Self-Educated Language Agent with Hindsight Experience Replay for Instruction Following.
Proceedings of the Visually Grounded Interaction and Language (ViGIL), 2019

Budgeted Reinforcement Learning in Continuous State Space.
Proceedings of the Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, 2019

Learning from a Learner.
Proceedings of the 36th International Conference on Machine Learning, 2019

A Theory of Regularized Markov Decision Processes.
Proceedings of the 36th International Conference on Machine Learning, 2019

Observational Learning by Reinforcement Learning.
Proceedings of the 18th International Conference on Autonomous Agents and MultiAgent Systems, 2019

2018
Playing the Game of Universal Adversarial Perturbations.
CoRR, 2018

Observe and Look Further: Achieving Consistent Performance on Atari.
CoRR, 2018

Noisy Networks For Exploration.
Proceedings of the 6th International Conference on Learning Representations, 2018

End-to-End Automatic Speech Translation of Audiobooks.
Proceedings of the 2018 IEEE International Conference on Acoustics, 2018

Visual Reasoning with Multi-hop Feature Modulation.
Proceedings of the Computer Vision - ECCV 2018, 2018

Training Dialogue Systems With Human Advice.
Proceedings of the 17th International Conference on Autonomous Agents and MultiAgent Systems, 2018

Actor-Critic Fictitious Play in Simultaneous Move Multistage Games.
Proceedings of the International Conference on Artificial Intelligence and Statistics, 2018

Deep Q-learning From Demonstrations.
Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence, 2018

Deep learning for multisensorial and multimodal interaction.
Proceedings of the Handbook of Multimodal-Multisensor Interfaces: Foundations, User Modeling, and Common Modality Combinations, 2018

2017
Bridging the Gap Between Imitation Learning and Inverse Reinforcement Learning.
IEEE Trans. Neural Networks Learn. Syst., 2017

Leveraging Demonstrations for Deep Reinforcement Learning on Robotics Problems with Sparse Rewards.
CoRR, 2017

Learning from Demonstrations for Real World Reinforcement Learning.
CoRR, 2017

Noisy Networks for Exploration.
CoRR, 2017

Observational Learning by Reinforcement Learning.
CoRR, 2017

LIG-CRIStAL System for the WMT17 Automatic Post-Editing Task.
CoRR, 2017

LIG-CRIStAL Submission for the WMT 2017 Automatic Post-Editing Task.
Proceedings of the Second Conference on Machine Translation, 2017

Modulating early visual processing by language.
Proceedings of the Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, 2017

Is the Bellman residual a bad proxy?
Proceedings of the Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, 2017

End-to-end optimization of goal-driven and visually grounded dialogue systems.
Proceedings of the Twenty-Sixth International Joint Conference on Artificial Intelligence, 2017

GuessWhat?! Visual Object Discovery through Multi-modal Dialogue.
Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition, 2017

Learning Nash Equilibrium for General-Sum Markov Games from Batch Data.
Proceedings of the 20th International Conference on Artificial Intelligence and Statistics, 2017

2016
Laughter Research: A Review of the ILHAIRE Project.
Proceedings of the Toward Robotic Socially Believable Behaving Systems - Volume I, 2016

Difference of Convex Functions Programming Applied to Control with Expert Data.
CoRR, 2016

Should one minimize the expected Bellman residual or maximize the mean value?
CoRR, 2016

Listen and Translate: A Proof of Concept for End-to-End Speech-to-Text Translation.
CoRR, 2016

Learning dialogue dynamics with the method of moments.
Proceedings of the 2016 IEEE Spoken Language Technology Workshop, 2016

MultiVec: a Multilingual and Multilevel Representation Learning Toolkit for NLP.
Proceedings of the Tenth International Conference on Language Resources and Evaluation LREC 2016, 2016

Compact and Interpretable Dialogue State Representation with Genetic Sparse Distributed Memory.
Proceedings of the Dialogues with Social Robots, 2016

A Stochastic Model for Computer-Aided Human-Human Dialogue.
Proceedings of the Interspeech 2016, 2016

Softened Approximate Policy Iteration for Markov Games.
Proceedings of the 33nd International Conference on Machine Learning, 2016

PAC learning of Probabilistic Automaton based on the Method of Moments.
Proceedings of the 33nd International Conference on Machine Learning, 2016

Score-based Inverse Reinforcement Learning.
Proceedings of the 2016 International Conference on Autonomous Agents & Multiagent Systems, 2016

On the Use of Non-Stationary Strategies for Solving Two-Player Zero-Sum Markov Games.
Proceedings of the 19th International Conference on Artificial Intelligence and Statistics, 2016

2015
Optimism in Active Learning.
Comput. Intell. Neurosci., 2015

Bayesian Credible Intervals for Online and Active Learning of Classification Trees.
Proceedings of the IEEE Symposium Series on Computational Intelligence, 2015

Human-Machine Dialogue as a Stochastic Game.
Proceedings of the SIGDIAL 2015 Conference, 2015

Learning of scanning strategies for electronic support using predictive state representations.
Proceedings of the 25th IEEE International Workshop on Machine Learning for Signal Processing, 2015

Inverse Reinforcement Learning in Relational Domains.
Proceedings of the Twenty-Fourth International Joint Conference on Artificial Intelligence, 2015

Non-negative Spectral Learning for Linear Sequential Systems.
Proceedings of the Neural Information Processing - 22nd International Conference, 2015

Optimism in Active Learning with Gaussian Processes.
Proceedings of the Neural Information Processing - 22nd International Conference, 2015

Imitation Learning Applied to Embodied Conversational Agents.
Proceedings of the 4th Workshop on Machine Learning for Interactive Systems, 2015

Approximate Dynamic Programming for Two-Player Zero-Sum Markov Games.
Proceedings of the 32nd International Conference on Machine Learning, 2015

Proceedings of the 4th Workshop on Machine Learning for Interactive Systems (MLIS-2015).
Proceedings of the 4th Workshop on Machine Learning for Interactive Systems, 2015

Spectral learning with non negative probabilities for finite state automaton.
Proceedings of the 2015 IEEE Workshop on Automatic Speech Recognition and Understanding, 2015

2014
Boosted Bellman Residual Minimization Handling Expert Demonstrations.
Proceedings of the Machine Learning and Knowledge Discovery in Databases, 2014

Difference of Convex Functions Programming for Reinforcement Learning.
Proceedings of the Advances in Neural Information Processing Systems 27: Annual Conference on Neural Information Processing Systems 2014, 2014

DINASTI: Dialogues with a Negotiating Appointment Setting Interface.
Proceedings of the Ninth International Conference on Language Resources and Evaluation, 2014

NASTIA: Negotiating Appointment Setting Interface.
Proceedings of the Ninth International Conference on Language Resources and Evaluation, 2014

Predicting when to laugh with structured classification.
Proceedings of the INTERSPEECH 2014, 2014

Ordinal regression for interaction quality prediction.
Proceedings of the IEEE International Conference on Acoustics, 2014

Boosted and reward-regularized classification for apprenticeship learning.
Proceedings of the International conference on Autonomous Agents and Multi-Agent Systems, 2014

Subspace identification for predictive state representation by nuclear norm minimization.
Proceedings of the 2014 IEEE Symposium on Adaptive Dynamic Programming and Reinforcement Learning, 2014

Active learning for classification: An optimistic approach.
Proceedings of the 2014 IEEE Symposium on Adaptive Dynamic Programming and Reinforcement Learning, 2014

2013
Algorithmic Survey of Parametric Value Function Approximation.
IEEE Trans. Neural Networks Learn. Syst., 2013

Classification structurée pour l'apprentissage par renforcement inverse.
Rev. d'Intelligence Artif., 2013

A survey on metrics for the evaluation of user simulations.
Knowl. Eng. Rev., 2013

Reward Shaping for Statistical Optimisation of Dialogue Management.
Proceedings of the Statistical Language and Speech Processing, 2013

Model-free POMDP optimisation of tutoring systems with echo-state networks.
Proceedings of the SIGDIAL 2013 Conference, 2013

Learning from Demonstrations: Is It Worth Estimating a Reward Function?
Proceedings of the Machine Learning and Knowledge Discovery in Databases, 2013

A Cascaded Supervised Learning Approach to Inverse Reinforcement Learning.
Proceedings of the Machine Learning and Knowledge Discovery in Databases, 2013

Particle swarm optimisation of spoken dialogue system strategies.
Proceedings of the INTERSPEECH 2013, 2013

Inverse reinforcement learning for interactive systems.
Proceedings of the 2nd Workshop on Machine Learning for Interactive Systems, 2013


Random projections: A remedy for overfitting issues in time series prediction with echo state networks.
Proceedings of the IEEE International Conference on Acoustics, 2013

Laugh-aware virtual agent and its impact on user amusement.
Proceedings of the International conference on Autonomous Agents and Multi-Agent Systems, 2013

2012
Introduction to the Issue on Advances in Spoken Dialogue Systems and Mobile Interface.
IEEE J. Sel. Top. Signal Process., 2012

A Comprehensive Reinforcement Learning Framework for Dialogue Management Optimization.
IEEE J. Sel. Top. Signal Process., 2012

Optimisation d'un tuteur intelligent à partir d'un jeu de données fixé (Optimization of a tutoring system from a fixed set of data) [in French].
Proceedings of the Joint Conference JEP-TALN-RECITAL 2012, 2012

Reward Function Learning for Dialogue Management.
Proceedings of the STAIRS 2012, 2012

Inverse Reinforcement Learning through Structured Classification.
Proceedings of the Advances in Neural Information Processing Systems 25: 26th Annual Conference on Neural Information Processing Systems 2012. Proceedings of a meeting held December 3-6, 2012

Statistical User Simulation for Spoken Dialogue Systems: What for, Which Data, Which Future?
Proceedings of the Workshop on Future directions and needs in the Spoken Dialog Community: Tools and Data, 2012

Co-adaptation in Spoken Dialogue Systems.
Proceedings of the Natural Interaction with Robots, 2012

Off-policy learning in large-scale POMDP-based dialogue systems.
Proceedings of the 2012 IEEE International Conference on Acoustics, 2012

Clustering behaviors of Spoken Dialogue Systems users.
Proceedings of the 2012 IEEE International Conference on Acoustics, 2012

A Reinforcement Learning Approach to Optimize the longitudinal Behavior of a Partial Autonomous Driving Assistance System.
Proceedings of the ECAI 2012, 2012

Behavior Specific User Simulation in Spoken Dialogue Systems.
Proceedings of the 10th ITG Conference on Speech Communication, 2012

2011
Sample-efficient batch reinforcement learning for dialogue management optimization.
ACM Trans. Speech Lang. Process., 2011

Introduction to special issue on machine learning for adaptivity in spoken dialogue systems.
ACM Trans. Speech Lang. Process., 2011

Functional Segmentation of Renal DCE-MRI Sequences Using Vector Quantization Algorithms.
Neural Process. Lett., 2011

Managing Uncertainty within KTD.
Proceedings of the Active Learning and Experimental Design workshop, 2011

Optimization of a tutoring system from a fixed set of data.
Proceedings of the ISCA International Workshop on Speech and Language Technology in Education, 2011

Uncertainty Management for On-Line Optimisation of a POMDP-Based Large-Scale Spoken Dialogue System.
Proceedings of the INTERSPEECH 2011, 2011

User Simulation in Dialogue Systems Using Inverse Reinforcement Learning.
Proceedings of the INTERSPEECH 2011, 2011

Training a BN-based user model for dialogue simulation with missing data.
Proceedings of the Fifth International Joint Conference on Natural Language Processing, 2011

Sample Efficient On-Line Learning of Optimal Dialogue Policies with Kalman Temporal Differences.
Proceedings of the IJCAI 2011, 2011

A Non-parametric Approach to Approximate Dynamic Programming.
Proceedings of the 10th International Conference on Machine Learning and Applications and Workshops, 2011

Automation Effects on Driver's Behaviour When Integrating a PADAS and a Distraction Classifier.
Proceedings of the Digital Human Modeling, 2011

Batch, Off-Policy and Model-Free Apprenticeship Learning.
Proceedings of the Recent Advances in Reinforcement Learning - 9th European Workshop, 2011

Single-trial P300 detection with Kalman filtering and SVMs.
Proceedings of the 19th European Symposium on Artificial Neural Networks, 2011

Batch reinforcement learning for optimizing longitudinal driving assistance strategies.
Proceedings of the 2011 IEEE Symposium on Computational Intelligence in Vehicles and Transportation Systems, 2011

Dynamic neural field optimization using the unscented Kalman filter.
Proceedings of the 2011 IEEE Symposium on Computational Intelligence, 2011

Parametric value function approximation: A unified view.
Proceedings of the 2011 IEEE Symposium on Adaptive Dynamic Programming And Reinforcement Learning, 2011

De l'Apprentissage Statistique pour le Contrôle Optimal et le Traitement du Signal.
, 2011

2010
Nonlinear Bayesian Filtering for Denoising of Electrocardiograms Acquired in a Magnetic Resonance Environment.
IEEE Trans. Biomed. Eng., 2010

Différences temporelles de Kalman. Cas déterministe.
Rev. d'Intelligence Artif., 2010

Kalman Temporal Differences.
J. Artif. Intell. Res., 2010

Sparse Approximate Dynamic Programming for Dialog Management.
Proceedings of the SIGDIAL 2010 Conference, 2010

Revisiting Natural Actor-Critics with Value Function Approximation.
Proceedings of the Modeling Decisions for Artificial Intelligence, 2010

Simulation of the Grounding Process in Spoken Dialog Systems with Bayesian Networks.
Proceedings of the Spoken Dialogue Systems for Ambient Environments, 2010

User and Noise Adaptive Dialogue Management Using Hybrid System Actions.
Proceedings of the Spoken Dialogue Systems for Ambient Environments, 2010

Single-speaker/multi-speaker co-channel speech classification.
Proceedings of the INTERSPEECH 2010, 2010

Optimizing spoken dialogue management with fitted value iteration.
Proceedings of the INTERSPEECH 2010, 2010

Eligibility traces through colored noises.
Proceedings of the International Conference on Ultra Modern Telecommunications, 2010

Statistically linearized least-squares temporal differences.
Proceedings of the International Conference on Ultra Modern Telecommunications, 2010

Bayesian framework for artifact reduction on ECG IN MRI.
Proceedings of the IEEE International Conference on Acoustics, 2010

Objective assessment of renal DCE-MRI image segmentation.
Proceedings of the 18th European Signal Processing Conference, 2010

Online speaker diarization with a size-monitored growing neural gas algorithm.
Proceedings of the 18th European Symposium on Artificial Neural Networks, 2010

2009
Tracking in Reinforcement Learning.
Proceedings of the Neural Information Processing, 16th International Conference, 2009

A specific QRS detector for electrocardiography during MRI: Using wavelets and local regularity characterization.
Proceedings of the IEEE International Conference on Acoustics, 2009

Kernelizing Vector Quantization Algorithms.
Proceedings of the 17th European Symposium on Artificial Neural Networks, 2009

Kalman Temporal Differences: The deterministic case.
Proceedings of the IEEE Symposium on Adaptive Dynamic Programming and Reinforcement Learning, 2009

2008
Adaptive RR prediction for cardiac MRI.
Proceedings of the IEEE International Conference on Acoustics, 2008

Functional semi-automated segmentation of renal DCE-MRI sequences.
Proceedings of the IEEE International Conference on Acoustics, 2008

Bayesian Reward Filtering.
Proceedings of the Recent Advances in Reinforcement Learning, 8th European Workshop, 2008

Functional semi-automated segmentation of renal DCE-MRI sequences using a Growing Neural Gas algorithm.
Proceedings of the 2008 16th European Signal Processing Conference, 2008

2007
Machine learning for spoken dialogue systems.
Proceedings of the INTERSPEECH 2007, 2007

Learning to Ground in Spoken Dialogue Systems.
Proceedings of the IEEE International Conference on Acoustics, 2007

2006
A probabilistic framework for dialog simulation and optimal strategy learning.
IEEE Trans. Speech Audio Process., 2006

Consistent Goal-Directed User Model for Realisitc Man-Machine Task-Oriented Spoken Dialogue Simulation.
Proceedings of the 2006 IEEE International Conference on Multimedia and Expo, 2006

Dynamic Bayesian Networks for NLU Simulation with Applications to Dialog Optimal Strategy Learning.
Proceedings of the 2006 IEEE International Conference on Acoustics Speech and Signal Processing, 2006

Machine Learning for Spoken Dialogue Management: An Experiment with Speech-Based Database Querying.
Proceedings of the Artificial Intelligence: Methodology, 2006

2005
Réseau bayesien pour un modèle d'utilisateur et un module de compréhension pour l'optimisation des systèmes de dialogues.
Proceedings of the Actes de la 12ème conférence sur le Traitement Automatique des Langues Naturelles. Articles courts, 2005

Comparing ASR modeling methods for spoken dialogue simulation and optimal strategy learning.
Proceedings of the INTERSPEECH 2005, 2005

A Probabilistic Description of Man-Machine Spoken Communication.
Proceedings of the 2005 IEEE International Conference on Multimedia and Expo, 2005

2004
Une description probabiliste de la communication parlée entre homme et machine.
Proceedings of the 16th conference on Association Francophone d'Interaction Homme-Machine, 2004

2003
Aided design of finite-state dialogue management systems.
Proceedings of the 2003 IEEE International Conference on Multimedia and Expo, 2003

2002
ASR system modeling for automatic evaluation and optimization of dialogue systems.
Proceedings of the IEEE International Conference on Acoustics, 2002


  Loading...