Maze Solver (Reinforcement Learning)

Maze Solving using Value iterations, Dynamic Programming

You are now following this Submission

Refer to 4.1, Reinforcement learning: An introduction, RS Sutton, AG Barto , MIT press
Value Iterations:
Algorithms of dynamic programming to solve finite MDPs. Policy evaluation refers to the (typically) iterative computation of the value functions for a given policy. Policy improvement refers to the computation of an improved policy given the value function for that policy. Putting these two computations together, we obtain policy iteration and value iteration, the two most popular DP methods. Either of these can be used to reliably compute optimal policies and value functions for finite MDPs given complete knowledge of the MDP.

◮ Problem: find optimal policy π
◮ Solution: iterative application of Bellman optimality backup
◮ v1 → v2 → ... → v∗
◮ Using synchronous backups, At each iteration k + 1 For all states s ∈ S : Update v_{k+1}(s) from v_{k}(s')
◮ Convergence to v∗ will be proven later
◮ Unlike policy iteration, there is no explicit policy
◮ Intermediate value functions may not correspond to any policy

Cite As

Bhartendu (2026). Maze Solver (Reinforcement Learning) (https://au.mathworks.com/matlabcentral/fileexchange/63062-maze-solver-reinforcement-learning), MATLAB Central File Exchange. Retrieved .

General Information

MATLAB Release Compatibility

  • Compatible with any release

Platform Compatibility

  • Windows
  • macOS
  • Linux
Version Published Release Notes Action
1.0.0.0