Asad
Asad
Rescue robot
Image courtesy: www.designnews.com
Department of Electrical Engineering
EXPLORATION OF AN UNKNOWN ENVIRONMENT 4
USING DEEP REINFORCEMENT LEARNING
Problem, Solution, Objective
Problem
Conventional methods are computationally expensive, require more
time and memory due to which navigation becomes difficult in large
environments.
Solution
The robot should learn to navigate/explore the environment while
avoiding obstacles without a map.
Objective
To design and analyze the navigation/ exploration system of a mobile
robot in a map-less environment utilizing learning-based methods.
Supervised Learning
Image courtesy: https://www.javatpoint.com
Reinforcement Learning
Unsupervised Learning Image courtesy: https://www.javatpoint.com
Image courtesy: https://www.javatpoint.com
EXPLORATION OF AN UNKNOWN ENVIRONMENT
Department of Electrical Engineering 6
USING DEEP REINFORCEMENT LEARNING
Reinforcement Learning
The Agent-Environment interaction in RL [1].
Agent-environment interaction.
Image courtesy: [1]
Main goal is to maximize the reward.
∞
𝐺𝑡 = 𝑚𝑎𝑥 𝛾 𝑘 𝑅𝑡+𝑘+1
𝑘=0
• In this equation,
• 𝛾 is discount factor ↑ ↓ ← →
• 𝛼 is learning rate S1 0 -1 -2 1
• Q table works as policy.
• Is Q learning applicable for large S2 2 0 -1 0
state space? S3 0 0 3 -2
S4 -3 1 -1 0
state’s Q-Value
St
(St, At) (St+1)
(St,At,Rt,St+1)
[3]
[5]
Fig: The result of Double DQN in obstacle Fig: The result of Double DQN in obstacle free
free environment (Rewards) environment( Steps)
Fig: The result of Double DQN Fig: The result of Double DQN
algorithm with LIDAR ( Rewards) algorithm with LIDAR ( Steps)
Simulation 14-May-22 30
Improvement 13-Jun-22 34