Robot Motion planning using Deep Reinforcement Learning / (Record no. 614873)

000 -LEADER
fixed length control field 02215nam a22001577a 4500
082 ## - DEWEY DECIMAL CLASSIFICATION NUMBER
Classification number 629.8
100 ## - MAIN ENTRY--PERSONAL NAME
Personal name Khan, Zahra
245 ## - TITLE STATEMENT
Title Robot Motion planning using Deep Reinforcement Learning /
Statement of responsibility, etc. Zahra Khan
264 ## - PRODUCTION, PUBLICATION, DISTRIBUTION, MANUFACTURE, AND COPYRIGHT NOTICE
Place of production, publication, distribution, manufacture Islamabad :
Name of producer, publisher, distributor, manufacturer SMME- NUST;
Date of production, publication, distribution, manufacture, or copyright notice 2025.
300 ## - PHYSICAL DESCRIPTION
Extent 62p.
Other physical details Soft Copy
Dimensions 30cm
500 ## - GENERAL NOTE
General note Effective motion planning is needed for autonomous robots in challenging environments.<br/>Sampling-based algorithms sample random points in high-dimensional spaces but struggle<br/>to work well in complex environments due to slow convergence and inefficiencies.<br/>Improvements in Deep Reinforcement Learning (DRL) overcome this strategy through<br/>learning optimal policies from acting in the environment, reducing reliance on<br/>environmental data and faster rates of convergence. But DRL resorts to sparse reward<br/>functions to produce suboptimal paths and poor exploration. To advance beyond such<br/>shortcomings, we propose an active SLAM-sourced information reward function. SLAMweighted reward enhances navigation efficiency with richer environment perception and<br/>robustness in unexplored areas. It comprises distance-to-target and smoothness terms over<br/>trajectory that encourage reduced distances, reduced oscillation, and more stable robot<br/>performance. We use the Soft Actor-Critic (SAC) algorithm in our reward function,<br/>because it is best to perform well in new environments. Comparison experiments using<br/>Twin Delayed Deep Deterministic Policy Gradient (TD3) and Deep Deterministic Policy<br/>Gradient (DDPG) in cluttered and sparse environments demonstrated the superior<br/>performance of SAC. In sparse settings, SAC was 100% successful, performing 9.4%<br/>better than TD3 and 14.1% better than DDPG, and using 35% fewer steps than TD3 and<br/>63% fewer than DDPG. In chaotic settings, SAC was 87.5% successful, performing 40.6%<br/>better than TD3 and 71.9% better than DDPG. These performances attest to the<br/>unprecedented effectiveness of the SAC algorithm in the direction of autonomous robots.
650 ## - SUBJECT ADDED ENTRY--TOPICAL TERM
Topical term or geographic name entry element MS Robotics and Intelligent Machine Engineering
700 ## - ADDED ENTRY--PERSONAL NAME
Personal name Supervisor : Dr. Fahad Iqbal
856 ## - ELECTRONIC LOCATION AND ACCESS
Uniform Resource Identifier <a href="http://10.250.8.41:8080/xmlui/handle/123456789/55263">http://10.250.8.41:8080/xmlui/handle/123456789/55263</a>
942 ## - ADDED ENTRY ELEMENTS (KOHA)
Source of classification or shelving scheme
Koha item type Thesis
Holdings
Withdrawn status Permanent Location Current Location Shelving location Date acquired Full call number Barcode Koha item type
  School of Mechanical & Manufacturing Engineering (SMME) School of Mechanical & Manufacturing Engineering (SMME) E-Books 09/25/2025 629.8 SMME-TH-1179 Thesis
© 2023 Central Library, National University of Sciences and Technology. All Rights Reserved.