Offered By: IBM
Playing TicTacToe with Reinforcement Learning and OpenAI Gym
Learn how to create and teach an agent that never loses to play TicTacToe using a Reinforcement Learning algorithm called Temporal Difference Learning and Open AI Gym
Continue readingGPXX0XENEN
Data Science
668 EnrolledAt a Glance
Learn how to create and teach an agent that never loses to play TicTacToe using a Reinforcement Learning algorithm called Temporal Difference Learning and Open AI Gym
In this Guided Project, you will learn how to interact with the OpenAI Gym environment. We will be working with a custom environment created to play TicTacToe so you will also learn how to install custom environments. Additionally, we will learn about Reinforcement Learning and the algorithm Temporal Difference Learning, and how to implement an Agent using Temporal Difference Learning to play TicTacToe. Finally, we will play TicTacToe with our trained agent and environment and see an example of a TicTacToe game with a graphical user interface.
Learn by Doing
A Look at the Project Ahead
- Install a custom OpenAI Gym environment
- Work with an OpenAI Gym environment and the TicTacToe environment
- Explain what Reinforcement Learning is
- Explain what Temporal Difference Learning is
- Create an agent that uses Temporal Difference Learning to play TicTacToe
- Train and Test the agents using the TicTacToe environment
- Play some games against the trained agent
What You’ll Need
Your Instructor
Estimated Effort
45 minutes
Level
Intermediate
Skills You Will Learn
Python, Data Science, Machine Learning, Artificial Intelligence
Language
English
Instructors
Contributors
Arnav Shah
Software Developer Intern @ IBM
15 y/o SWD Intern @ IBM | ML researcher.
Read moreArtem Arutyunov
Data Scientist
Hey, Artem here! I am excited about answering new challenges with data science, machine learning and especially Reinforcement Learning. Love helping people to learn, and learn myself. Studying Math and Stats at University of Toronto, hit me up if you are from there as well.
Read more