Inductive bias explained

The inductive bias (also known as learning bias) of a learning algorithm is the set of assumptions that the learner uses to predict outputs of given inputs that it has not encountered.Inductive bias is anything which makes the algorithm learn one pattern instead of another pattern (e.g. step-functions in decision trees instead of continuous function in a linear regression model). Learning is the process of apprehending useful knowledge by observing and interacting with the world.[1] It involves searching a space of solutions for one expected to provide a better explanation of the data or to achieve higher rewards. But in many cases, there are multiple solutions which are equally good.[2] An inductive bias allows a learning algorithm to prioritize one solution (or interpretation) over another, independent of the observed data.[3]

In machine learning, one aims to construct algorithms that are able to learn to predict a certain target output. To achieve this, the learning algorithm is presented some training examples that demonstrate the intended relation of input and output values. Then the learner is supposed to approximate the correct output, even for examples that have not been shown during training. Without any additional assumptions, this problem cannot be solved since unseen situations might have an arbitrary output value. The kind of necessary assumptions about the nature of the target function are subsumed in the phrase inductive bias.

A classical example of an inductive bias is Occam's razor, assuming that the simplest consistent hypothesis about the target function is actually the best. Here consistent means that the hypothesis of the learner yields correct outputs for all of the examples that have been given to the algorithm.

Approaches to a more formal definition of inductive bias are based on mathematical logic. Here, the inductive bias is a logical formula that, together with the training data, logically entails the hypothesis generated by the learner. However, this strict formalism fails in many practical cases, where the inductive bias can only be given as a rough description (e.g. in the case of artificial neural networks), or not at all.

Types

The following is a list of common inductive biases in machine learning algorithms.

Shift of bias

Although most learning algorithms have a static bias, some algorithms are designed to shift their bias as they acquire more data. This does not avoid bias, since the bias shifting process itself must have a bias.

See also

Notes and References

  1. Battaglia . Peter W. . Hamrick . Bapst . Sanchez-Gonzalez . Zambaldi . Malinowski . Tacchetti . Raposo . Santoro . Faulkner . 2018 . Relational inductive biases, deep learning, and graph networks . cs.LG . 1806.01261.
  2. Book: Goodman, Nelson . Fact, Fiction, and Forecast . 1955 . Harvard University Press . 1955 . 978-0-674-29071-6 . 59–83 . The new riddle of induction.
  3. Mitchell . Tom M . 1980 . The need for biases in learning generalizations . Rutgers University Technical Report CBM-TR-117 . 184–191.