Project 5: Q-Table Reinforcement Learning Maze Solver

Results

The reinforcement learning maze solver highlights its ability to train an agent for efficient maze navigation. Despite the challenges faced, this project helped to understand the nuances of RL model training. I learned that Q-Learning is a very helpful technique to help an agent solve a known problem but it becomes hard to use the same technique to solve "new" problems. For example it can easily learn how to solve complexe mazes as long as it knows the mazes and had a chance to train on them. It does not achieve well in new mazes it has never seen before. I think Q-Learning can be very helpful to detect new patters when solving an issue but not as a general agent that can be used to solve unknonw challenges.





Key Takeaways

  • Q-Learning can perform well in known environments but is hard to tweak for new environments.
  • Reward and penalty systems effectively guided the agent but required careful adjustment.
  • Dynamic exploration strategies provided a balance between discovering new paths and exploiting known optimal paths.
  • The agent demonstrated some ability to generalize and solve unseen test mazes, though consistency remains a challenge and complex mazes can not be solved.

Conclusion

This project demonstrates the potential of reinforcement learning for solving navigation problems, with a focus on achieving generalization over different maze sizes. While small mazes and parkours were navigated successfully, scaling up remains an open challenge. I am currently thinking of heading into the direction of using Deep Q-Networks (DQN) which could be a better fit to train on random mazes to achieve a general maze solver agent.