A copy of this work was available on the public web and has been preserved in the Wayback Machine. The capture dates from 2021; you can also visit the original URL.
The file type is application/pdf
.
Filters
Robust Biped Locomotion Using Deep Reinforcement Learning on Top of an Analytical Control Approach
[article]
2021
arXiv
pre-print
This paper proposes a modular framework to generate robust biped locomotion using a tight coupling between an analytical walking approach and deep reinforcement learning. ...
This dynamics model is used to design an adaptive reference trajectories planner and an optimal controller which are fully parametric. ...
Conclusion In this paper, we have tackled the problem of developing a robust biped locomotion framework by proposing a tight coupling between an analytical control approach and a reinforcement learning ...
arXiv:2104.10592v1
fatcat:ulkxmgotivdsrjdnysqimer7b4
Learning Hybrid Locomotion Skills – Learn to Exploit Residual Dynamics and Modulate Model-based Gait Control
[article]
2022
arXiv
pre-print
The framework embeds a kernel which is a fully parametric closed-loop gait generator based on analytical control. ...
This work aims to combine machine learning and control approaches for legged robots, and developed a hybrid framework to achieve new capabilities of balancing against external perturbations. ...
CONCLUSION In this paper, we proposed a locomotion framework based on a tight coupling between analytical control and deep reinforcement learning to combine the potential of both approaches. ...
arXiv:2011.13798v3
fatcat:dgrinscvufetbfmkyey66zbfnm
Learning locomotion skills using DeepRL
2017
Proceedings of the ACM SIGGRAPH / Eurographics Symposium on Computer Animation - SCA '17
The use of deep reinforcement learning allows for high-dimensional state descriptors, but little is known about how the choice of action representation impacts the learning difficulty and the resulting ...
We compare the impact of four different action parameterizations (torques, muscle-activations, target joint angles, and target joint-angle velocities) in terms of learning time, policy robustness, motion ...
INTRODUCTION The introduction of deep learning models to reinforcement learning (RL) has enabled policies to operate directly on high-dimensional, low-level state features. ...
doi:10.1145/3099564.3099567
dblp:conf/sca/PengP17
fatcat:crcfk4lctzbatotjsejrkt2qle
Deep Reinforcement Learning for Tensegrity Robot Locomotion
[article]
2017
arXiv
pre-print
the effectiveness of our approach on tensegrity robot locomotion. ...
In this work, we show how locomotion gaits can be learned automatically using a novel extension of mirror descent guided policy search (MDGPS) applied to periodic locomotion movements, and we demonstrate ...
Acknowledgments: This work was supported in part by the DARPA SIMPLEX program and an NSF CAREER award. ...
arXiv:1609.09049v3
fatcat:3xpdqdabgjbx5krke35kbxpnxu
Robust Dynamic Locomotion via Reinforcement Learning and Novel Whole Body Controller
[article]
2017
arXiv
pre-print
Our locomotion strategy relies on devising a reinforcement learning (RL) approach for robust walking. ...
We propose a robust dynamic walking controller consisting of a dynamic locomotion planner, a reinforcement learning process for robustness, and a novel whole-body locomotion controller (WBLC). ...
ACKNOWLEDGMENT The authors would like to thank the members of the Human Centered Robotics Laboratory at The University of Texas at Austin for their great help and support. ...
arXiv:1708.02205v1
fatcat:w3ltjoku6fg6nkul35263iq3o4
Learning Locomotion Controllers for Walking Using Deep FBSDE
[article]
2021
arXiv
pre-print
The efficacy of our approach is shown on a linear inverted pendulum model (LIPM) for walking. ...
In this paper, we propose a deep forward-backward stochastic differential equation (FBSDE) based control algorithm for locomotion tasks. ...
Recently, deep reinforcement learning has been widely applied to locomotion tasks. ...
arXiv:2107.07931v1
fatcat:lnacb2pjmjeunblqc7hbpvx6wq
Deep Reinforcement Learning for Industrial Insertion Tasks with Visual Inputs and Natural Rewards
[article]
2019
arXiv
pre-print
Reinforcement learning (RL) methods have been demonstrated to be capable of learning controllers in such environments from autonomous interaction with the environment, but running RL algorithms in the ...
We show that methods that combine RL with prior information, such as classical controllers or demonstrations, can solve these tasks from a reasonable amount of real-world interaction. ...
To handle this challenge, we extend the residual RL approach [8, 9] , which learns a parametric policy on top of a fixed, hand-specified controller, to the setting of vision-based manipulation. ...
arXiv:1906.05841v2
fatcat:undptyk6h5gmtcazeyialzwzia
Detection of Landmines Using Nuclear Quadrupole Resonance (NQR): Signal Processing to Aid Classification
[chapter]
2006
Climbing and Walking Robots
These design drivers include: Lightweight Robustness Autonomy Controllable at high-level of abstraction Optimized locomotion Emphasis on collision avoidance These driving factors are described in more ...
The paper deals with a pattern recognition approach to detect and classify falls of bipedal robots according to intensity and direction. ...
Locomotion system is highly dependant on the environment nature. In some fields the most used solutions (wheels, legs or tracks) often show weakness. ...
doi:10.1007/3-540-26415-9_100
dblp:conf/clawar/SomasundaramASS05
fatcat:4vu25zlx65dszgbcxkusxozfve
Morphological Properties of Mass–Spring Networks for Optimal Locomotion Learning
2017
Frontiers in Neurorobotics
On the one hand, the use of flexible materials can lead to higher power efficiency and more fluent and robust motions. ...
We first investigate the influence of the network size and compliance on locomotion quality and energy efficiency by optimizing an external open-loop controller using evolutionary algorithms. ...
The data were analyzed by GU with help of FW, JDegrave, and JDambre. The manuscript was mostly written by GU, with comments and corrections from FW and JDambre. ...
doi:10.3389/fnbot.2017.00016
pmid:28396634
pmcid:PMC5366341
fatcat:iz2jxwlxgfaxrcldq23gvilvvy
Deep Reinforcement Learning, a textbook
[article]
2022
arXiv
pre-print
The way in which deep reinforcement learning explores complex environments reminds us of how children learn, by playfully trying out things, getting feedback, and trying again. ...
The aim of this book is to provide a comprehensive overview of the field of deep reinforcement learning. ...
What is Deep Reinforcement Learning? Deep reinforcement learning is the combination of deep learning and reinforcement learning. ...
arXiv:2201.02135v2
fatcat:3icsopexerfzxa3eblpu5oal64
Policy Transfer via Kinematic Domain Randomization and Adaptation
[article]
2021
arXiv
pre-print
Our algorithm, Multi-Policy Bayesian Optimization, trains an ensemble of universal policies conditioned on virtual kinematic parameters and efficiently adapts to the target environment using a limited ...
Transferring reinforcement learning policies trained in physics simulation to the real hardware remains a challenge, known as the "sim-to-real" gap. ...
INTRODUCTION The advent of Deep Reinforcement Learning (DRL) has demonstrated a promising approach to design robotic controllers for diverse robot motor skills [1] , [2] , [3] , [4] . ...
arXiv:2011.01891v3
fatcat:jmsbzzw4lzh43kisclw57vmvxm
Mapless Humanoid Navigation Using Learned Latent Dynamics
[article]
2021
arXiv
pre-print
In this paper, we propose a novel Deep Reinforcement Learning approach to address the mapless navigation problem, in which the locomotion actions of a humanoid robot are taken online based on the knowledge ...
In addition, we incorporate a termination likelihood predictor model as an auxiliary loss function of the control policy, which enables the agent to anticipate terminal states of success and failure. ...
[21] investigate visual navigation on a bipedal platform and learned a control policy by using DDPG. ...
arXiv:2108.03866v1
fatcat:5ll7nbvus5hvvhunz2stgjmt4y
A posture sequence learning system for an anthropomorphic robotic hand
2004
Robotics and Autonomous Systems
Work in that area tackles the development of robust algorithms for motor control, motor learning, gesture recognition and visuo-motor integration. ...
Studies of imitation in biological systems stressed a number of key issues that are, also, of concerns to Robotics. ...
The work is being done with the support of CYCIT, Spanish Research Agency, under the projects: DPI2001-0822 and DPI2002-04286-C2-02 Acknowledgments: Among many people who contributed to the robot system ...
doi:10.1016/j.robot.2004.03.008
fatcat:fefnr5l7gnhnhaca5nw52f3kru
Safe Value Functions
[article]
2021
arXiv
pre-print
Although these criteria are often solved separately with different tools to maintain formal guarantees, it is also common practice in reinforcement learning to simply modify reward functions by penalizing ...
Safety constraints and optimality are important, but sometimes conflicting criteria for controllers. ...
Levine, “Learning to walk via deep reinforcement
“A learnable safety measure,” in Conference on Robot learning,” in Proceedings of Robotics: Science and
Learning (CoRL), PMLR, 2019 ...
arXiv:2105.12204v2
fatcat:5tc7aqflc5etljp3ceux55vmvy
DeepMind Control Suite
[article]
2018
arXiv
pre-print
The DeepMind Control Suite is a set of continuous control tasks with a standardised structure and interpretable rewards, intended to serve as performance benchmarks for reinforcement learning agents. ...
The tasks are written in Python and powered by the MuJoCo physics engine, making them easy to use and modify. We include benchmarks for several learning algorithms. ...
However, continuous control domains cannot be solved by humans, so a different approach must be taken. In order to tackle both of these challenges, we ran variety of learning agents (e.g. ...
arXiv:1801.00690v1
fatcat:nsiln5qimje7pggv3ysh3k5swu
« Previous
Showing results 1 — 15 out of 86 results