Eiji Uchibe

According to our database1, Eiji Uchibe authored at least 64 papers between 1994 and 2020.

Collaborative distances:

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Other 

Links

Online presence:

On csauthors.net:

Bibliography

2020
Imitation learning based on entropy-regularized forward and inverse reinforcement learning.
CoRR, 2020

2019
Deep reinforcement learning with smooth policy update: Application to robotic cloth manipulation.
Robotics Auton. Syst., 2019

Constrained Deep Q-Learning Gradually Approaching Ordinary Q-Learning.
Frontiers Neurorobotics, 2019

Theoretical Analysis of Efficiency and Robustness of Softmax and Gap-Increasing Operators in Reinforcement Learning.
Proceedings of the 22nd International Conference on Artificial Intelligence and Statistics, 2019

2018
Model-Free Deep Inverse Reinforcement Learning by Logistic Regression.
Neural Process. Lett., 2018

Sigmoid-weighted linear units for neural network function approximation in reinforcement learning.
Neural Networks, 2018

Cooperative and Competitive Reinforcement and Imitation Learning for a Mixture of Heterogeneous Learning Modules.
Frontiers Neurorobotics, 2018

Unbounded Output Networks for Classification.
CoRR, 2018

Robustness of linearly solvable Markov games employing inaccurate dynamics model.
Artif. Life Robotics, 2018

Deep Reinforcement Learning by Parallelizing Reward and Punishment using the MaxPain Architecture.
Proceedings of the 2018 Joint IEEE 8th International Conference on Development and Learning and Epigenetic Robotics, 2018

Efficient sample reuse in policy search by multiple importance sampling.
Proceedings of the Genetic and Evolutionary Computation Conference, 2018

Online meta-learning by parallel algorithm competition.
Proceedings of the Genetic and Evolutionary Computation Conference, 2018

2017
Adaptive Baseline Enhances EM-Based Policy Search: Validation in a View-Based Positioning Task of a Smartphone Balancer.
Frontiers Neurorobotics, 2017

Unifying Value Iteration, Advantage Learning, and Dynamic Policy Programming.
CoRR, 2017

Deep dynamic policy programming for robot control with raw images.
Proceedings of the 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems, 2017

Average Reward Optimization with Multiple Discounting Reinforcement Learners.
Proceedings of the Neural Information Processing - 24th International Conference, 2017

2016
From free energy to expected energy: Improving energy-based value function approximation in reinforcement learning.
Neural Networks, 2016

EM-based policy hyper parameter exploration: application to standing and balancing of a two-wheeled smartphone robot.
Artif. Life Robotics, 2016

Deep Inverse Reinforcement Learning by Logistic Regression.
Proceedings of the Neural Information Processing - 23rd International Conference, 2016

Emergence of communication among reinforcement learning agents under coordination environment.
Proceedings of the 2016 Joint IEEE International Conference on Development and Learning and Epigenetic Robotics, 2016

2015
Expected energy-based restricted Boltzmann machine for classification.
Neural Networks, 2015

2014
Combining learned controllers to achieve new goals based on linearly solvable MDPs.
Proceedings of the 2014 IEEE International Conference on Robotics and Automation, 2014

Inverse reinforcement learning using Dynamic Policy Programming.
Proceedings of the 4th International Conference on Development and Learning and on Epigenetic Robotics, 2014

2013
Evaluation of linearly solvable Markov decision process with dynamic model learning in a mobile robot navigation task.
Frontiers Neurorobotics, 2013

Scaled free-energy based reinforcement learning for robust and efficient learning in high-dimensional state spaces.
Frontiers Neurorobotics, 2013

Reinforcement learning with state-dependent discount factor.
Proceedings of the 2013 IEEE Third Joint International Conference on Development and Learning and Epigenetic Robotics, 2013

2011
Darwinian embodied evolution of the learning ability for survival.
Adapt. Behav., 2011

2010
Derivatives of Logarithmic Stationary Distributions for Policy Gradient Reinforcement Learning.
Neural Comput., 2010

Free-Energy Based Reinforcement Learning for Vision-Based Navigation with High-Dimensional Sensory Inputs.
Proceedings of the Neural Information Processing. Theory and Algorithms, 2010

2009
Co-evolution of Rewards and Meta-parameters in Embodied Evolution.
Proceedings of the Creating Brain-Like Intelligence: From Basic Principles to Complex Intelligent Systems, 2009

A Generalized Natural Actor-Critic Algorithm.
Proceedings of the Advances in Neural Information Processing Systems 22: 23rd Annual Conference on Neural Information Processing Systems 2009. Proceedings of a meeting held 7-10 December 2009, 2009

Emergence of Different Mating Strategies in Artificial Embodied Evolution.
Proceedings of the Neural Information Processing, 16th International Conference, 2009

2008
Finding intrinsic rewards by embodied evolution and constrained reinforcement learning.
Neural Networks, 2008

Learning how, what, and whether to communicate: emergence of protocommunication in reinforcement learning agents.
Artif. Life Robotics, 2008

Natural actor-critic with baseline adjustment for variance reduction.
Artif. Life Robotics, 2008

Co-evolution of Shaping Rewards and Meta-Parameters in Reinforcement Learning.
Adapt. Behav., 2008

A New Natural Policy Gradient by Stationary Distribution Metric.
Proceedings of the Machine Learning and Knowledge Discovery in Databases, 2008

NeuroEvolution Based on Reusable and Hierarchical Modular Representation.
Proceedings of the Advances in Neuro-Information Processing, 15th International Conference, 2008

2007
Evolutionary Development of Hierarchical Learning Structures.
IEEE Trans. Evol. Comput., 2007

Finding Exploratory Rewards by Embodied Evolution and Constrained Reinforcement Learning in the Cyber Rodents.
Proceedings of the Neural Information Processing, 14th International Conference, 2007

2006
Incremental Coevolution With Competitive and Cooperative Tasks in a Multirobot Environment.
Proc. IEEE, 2006

2005
The Cyber Rodent Project: Exploration of Adaptive Mechanisms for Self-Preservation and Self-Reproduction.
Adapt. Behav., 2005

Biologically inspired embodied evolution of survival.
Proceedings of the IEEE Congress on Evolutionary Computation, 2005

2004
Multi-agent reinforcement learning: using macro actions to learn a mating task.
Proceedings of the 2004 IEEE/RSJ International Conference on Intelligent Robots and Systems, Sendai, Japan, September 28, 2004

2003
An Evolutionary Approach to Automatic Construction of the Structure in Hierarchical Reinforcement Learning.
Proceedings of the Genetic and Evolutionary Computation, 2003

2002
Behavior generation for a mobile robot based on the adaptive fitness function.
Robotics Auton. Syst., 2002

2001
Multiagent Learning towards RoboCup.
New Gener. Comput., 2001

Evolutionary Behavior Selection with Activation/Termination Constraints.
Proceedings of the RoboCup 2001: Robot Soccer World Cup V, 2001

Dynamic Task Assignment in a Multiagent/Multitask Environment based on Module Conflict Resolution.
Proceedings of the 2001 IEEE International Conference on Robotics and Automation, 2001

2000
Osaka University "Trackies 2000".
Proceedings of the RoboCup 2000: Robot Soccer World Cup IV, 2000

1999
Cooperative Behavior Acquisition for Mobile Robots in Dynamically Changing Real Worlds Via Vision-Based Reinforcement Learning and Development.
Artif. Intell., 1999

Multiple Reward Criterion for Cooperative Behavior Acquisition in a Muliagent Environment.
Proceedings of the RoboCup-99: Robot Soccer World Cup III, 1999

The Team Description of Osaka University "Trackies-99".
Proceedings of the RoboCup-99: Robot Soccer World Cup III, 1999

1998
TRACKIES: RoboCup-97 Middle-Size League World Cochampion.
AI Mag., 1998

Cooperative Behavior Acquisition in a Multiple Mobile Robot Environment by Co-evolution.
Proceedings of the RoboCup-98: Robot Soccer World Cup II, 1998

An Application of Vision-Based Learning in RoboCup for a Real Robot with an Omnidirectional Vision System and the Team Description of Osaka University "Trackies".
Proceedings of the RoboCup-98: Robot Soccer World Cup II, 1998

Co-evolution for cooperative behavior acquisition in a multiple mobile robot environment.
Proceedings of the Proceedings 1998 IEEE/RSJ International Conference on Intelligent Robots and Systems. Innovations in Theory, 1998

Environmental Complexity Control for Vision-Based Learning Mobile Robot.
Proceedings of the IEEE International Conference on Robotics and Automation, 1998

Cooperative Behavior Acquisition in Multi Mobile Robots Environment by Reinforcement Learning Based on State Vector Estimation.
Proceedings of the IEEE International Conference on Robotics and Automation, 1998

State Space Construction for Behavior Acquisition in Multi Agent Environments with Vision and Action.
Proceedings of the Sixth International Conference on Computer Vision (ICCV-98), 1998

1997
Vision-Based Robot Learning Towards RoboCup: Osaka University "Trackies".
Proceedings of the RoboCup-97: Robot Soccer World Cup I, 1997

Vision Based State Space Construction for Learning Mobile Robots in Multi-agent Environments.
Proceedings of the Learning Robots, 6th European Workshop, 1997

1996
Behavior coordination for a mobile robot using modular reinforcement learning.
Proceedings of IEEE/RSJ International Conference on Intelligent Robots and Systems. IROS 1996, 1996

1994
Coordination of multiple behaviors acquired by a vision-based reinforcement learning.
Proceedings of IEEE/RSJ International Conference on Intelligent Robots and Systems, 1994


  Loading...