<?xml version="1.0" encoding="UTF-8"?>
<record
    xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
    xsi:schemaLocation="http://www.loc.gov/MARC21/slim http://www.loc.gov/standards/marcxml/schema/MARC21slim.xsd"
    xmlns="http://www.loc.gov/MARC21/slim">

  <leader>02215nam a22001577a 4500</leader>
  <datafield tag="082" ind1=" " ind2=" ">
    <subfield code="a">629.8</subfield>
  </datafield>
  <datafield tag="100" ind1=" " ind2=" ">
    <subfield code="a">Khan, Zahra </subfield>
    <subfield code="9">34408</subfield>
  </datafield>
  <datafield tag="245" ind1=" " ind2=" ">
    <subfield code="a">Robot Motion planning using Deep Reinforcement Learning /</subfield>
    <subfield code="c">Zahra Khan</subfield>
  </datafield>
  <datafield tag="264" ind1=" " ind2=" ">
    <subfield code="a">Islamabad : </subfield>
    <subfield code="b">SMME- NUST; </subfield>
    <subfield code="c">2025.</subfield>
  </datafield>
  <datafield tag="300" ind1=" " ind2=" ">
    <subfield code="a">62p.</subfield>
    <subfield code="b">Soft Copy</subfield>
    <subfield code="c">30cm</subfield>
  </datafield>
  <datafield tag="500" ind1=" " ind2=" ">
    <subfield code="a">Effective motion planning is needed for autonomous robots in challenging environments.
Sampling-based algorithms sample random points in high-dimensional spaces but struggle
to work well in complex environments due to slow convergence and inefficiencies.
Improvements in Deep Reinforcement Learning (DRL) overcome this strategy through
learning optimal policies from acting in the environment, reducing reliance on
environmental data and faster rates of convergence. But DRL resorts to sparse reward
functions to produce suboptimal paths and poor exploration. To advance beyond such
shortcomings, we propose an active SLAM-sourced information reward function. SLAMweighted reward enhances navigation efficiency with richer environment perception and
robustness in unexplored areas. It comprises distance-to-target and smoothness terms over
trajectory that encourage reduced distances, reduced oscillation, and more stable robot
performance. We use the Soft Actor-Critic (SAC) algorithm in our reward function,
because it is best to perform well in new environments. Comparison experiments using
Twin Delayed Deep Deterministic Policy Gradient (TD3) and Deep Deterministic Policy
Gradient (DDPG) in cluttered and sparse environments demonstrated the superior
performance of SAC. In sparse settings, SAC was 100% successful, performing 9.4%
better than TD3 and 14.1% better than DDPG, and using 35% fewer steps than TD3 and
63% fewer than DDPG. In chaotic settings, SAC was 87.5% successful, performing 40.6%
better than TD3 and 71.9% better than DDPG. These performances attest to the
unprecedented effectiveness of the SAC algorithm in the direction of autonomous robots.</subfield>
  </datafield>
  <datafield tag="650" ind1=" " ind2=" ">
    <subfield code="a">MS Robotics and Intelligent Machine Engineering </subfield>
    <subfield code="9">119486</subfield>
  </datafield>
  <datafield tag="700" ind1=" " ind2=" ">
    <subfield code="a">Supervisor : Dr. Fahad Iqbal</subfield>
    <subfield code="9">130707</subfield>
  </datafield>
  <datafield tag="856" ind1=" " ind2=" ">
    <subfield code="u">http://10.250.8.41:8080/xmlui/handle/123456789/55263</subfield>
  </datafield>
  <datafield tag="942" ind1=" " ind2=" ">
    <subfield code="2">ddc</subfield>
    <subfield code="c">THE</subfield>
  </datafield>
  <datafield tag="999" ind1=" " ind2=" ">
    <subfield code="c">614873</subfield>
    <subfield code="d">614873</subfield>
  </datafield>
  <datafield tag="952" ind1=" " ind2=" ">
    <subfield code="0">0</subfield>
    <subfield code="1">0</subfield>
    <subfield code="4">0</subfield>
    <subfield code="7">0</subfield>
    <subfield code="a">SMME</subfield>
    <subfield code="b">SMME</subfield>
    <subfield code="c">EB</subfield>
    <subfield code="d">2025-09-25</subfield>
    <subfield code="l">0</subfield>
    <subfield code="o">629.8</subfield>
    <subfield code="p">SMME-TH-1179</subfield>
    <subfield code="r">2025-09-25</subfield>
    <subfield code="w">2025-09-25</subfield>
    <subfield code="y">THE</subfield>
  </datafield>
</record>
