Exploring reward shaping in discrete and continuous action spaces: A deep reinforcement learning study on Turtlebot3

Abstract: In robotics, reinforcement learning can train controllers or agents to find optimal solutions for complex tasks by enabling the robot to interact repeatedly with the environment. The reward function is an important aspect that guides any reinforcement algorithm to find the desired solution successfully. This work examines two deep reinforcement learning approaches, one uses a Deep Q‐Network (DQN) and the other a deep deterministic policy gradient (DDPG) algorithm, applicable to navigation tasks for mobile robots. Comparison between different reward schemes for both algorithms is one of the main focuses in this work. The methodology is implemented in the simulation for a mobile robot called TurtleBot3. The task for the robot is to navigate through obstacles from an initial location to a goal position. Finally, the trained end‐to‐end navigation stack is also implemented on the actual TurtleBot3 in a real environment. The robot uses a Lidar sensor to detect obstacles. The Lidar measurements and the relative position and angle of the robot to the target location are the inputs to the controller. The TurtleBot3 also utilizes distance information from its Lidar sensor to create an environmental map using the simultaneous localization and mapping (SLAM) technique. Additionally, given an initial position, the robot employs its inertial measurement unit (IMU) sensor and encoders for precise localization.

Location
Deutsche Nationalbibliothek Frankfurt am Main
Extent
Online-Ressource
Language
Englisch

Bibliographic citation
Exploring reward shaping in discrete and continuous action spaces: A deep reinforcement learning study on Turtlebot3 ; day:04 ; month:10 ; year:2024 ; extent:13
Proceedings in applied mathematics and mechanics ; (04.10.2024) (gesamt 13)

Creator

DOI
10.1002/pamm.202400169
URN
urn:nbn:de:101:1-2410041432108.185400013998
Rights
Open Access; Der Zugriff auf das Objekt ist unbeschränkt möglich.
Last update
15.08.2025, 7:31 AM CEST

Data provider

This object is provided by:
Deutsche Nationalbibliothek. If you have any questions about the object, please contact the data provider.

Associated

Other Objects (12)