Teacher forcing explained

Teacher forcing is an algorithm for training the weights of recurrent neural networks (RNNs).[1] It involves feeding observed sequence values (i.e. ground-truth samples) back into the RNN after each step, thus forcing the RNN to stay close to the ground-truth sequence.[2]

The term "teacher forcing" can be motivated by comparing the RNN to a human student taking a multi-part exam where the answer to each part (for example a mathematical calculation) depends on the answer to the preceding part. In this analogy, rather than grading every answer in the end, with the risk that the student fails every single part even though they only made a mistake in the first one, a teacher records the score for each individual part and then tells the student the correct answer, to be used in the next part.[3]

The use of an external teacher signal is in contrast to real-time recurrent learning (RTRL).[4] Teacher signals are known from oscillator networks.[5] The promise is, that teacher forcing helps to reduce the training time.[6]

The term "teacher forcing" was introduced in 1989 by Ronald J. Williams and David Zipser, who reported that the technique was already being "frequently used in dynamical supervised learning tasks" around that time.[7]

A NeurIPS 2016 paper introduced the related method of "professor forcing".

See also

Notes and References

  1. Book: John F. Kolen. Stefan C. Kremer. A Field Guide to Dynamical Recurrent Networks. 15 January 2001. John Wiley & Sons. 978-0-7803-5369-5. 202–.
  2. Lamb . Alex M . Goyal . Anirudh . Zhang . Ying . Zhang . Saizheng . Courville . Aaron C . Bengio . Yoshua . Yoshua Bengio . 2016 . Professor Forcing: A New Algorithm for Training Recurrent Networks . Advances in Neural Information Processing Systems . Curran Associates, Inc. . 29. 1610.09038 .
  3. Web site: Wong . Wanshun . 2019-10-15 . What is Teacher Forcing? . 2022-03-25 . Towards Data Science . en.
  4. Book: Zhang, Ming. Artificial Higher Order Neural Networks for Economics and Business. 31 July 2008. IGI Global. 978-1-59904-898-7. 195–.
  5. Book: Yves Chauvin. David E. Rumelhart. Backpropagation: Theory, Architectures, and Applications. 1 February 2013. Psychology Press. 978-1-134-77581-1. 473–.
  6. Book: George Bekey. Kenneth Y. Goldberg. Neural Networks in Robotics. 30 November 1992. Springer Science & Business Media. 978-0-7923-9268-2. 247–.
  7. Williams . Ronald J. . Zipser . David . June 1989 . A Learning Algorithm for Continually Running Fully Recurrent Neural Networks . Neural Computation . 1 . 2 . 270–280 . 10.1162/neco.1989.1.2.270 . 10.1.1.52.9724 . 14711886 . 0899-7667.