Part of Advances in Neural Information Processing Systems 8 (NIPS 1995)
Philip N. Sabes, Michael Jordan
We present a new algorithm for associative reinforcement learn(cid:173) ing. The algorithm is based upon the idea of matching a network's output probability with a probability distribution derived from the environment's reward signal. This Probability Matching algorithm is shown to perform faster and be less susceptible to local minima than previously existing algorithms. We use Probability Match(cid:173) ing to train mixture of experts networks, an architecture for which other reinforcement learning rules fail to converge reliably on even simple problems. This architecture is particularly well suited for our algorithm as it can compute arbitrarily complex functions yet calculation of the output probability is simple.