How might we amend the reinforcement learning algorithm


Problem

Symmetries Many Tic-Tac-Toe positions appear different but are really the same because of symmetries. How might we amend the reinforcement learning algorithm described above to take advantage of this? In what ways would this improve it? Now think again, suppose the opponent did not take advantage of symmetries. In that case, should we? It is true then that symmetrically equivalent positions should necessarily have the same value?

Request for Solution File

Ask an Expert for Answer!!
Data Structure & Algorithms: How might we amend the reinforcement learning algorithm
Reference No:- TGS02639169

Expected delivery within 24 Hours