Domain adaptation explained

Domain adaptation[1] [2] [3] is a field associated with machine learning and transfer learning. This scenario arises when we aim at learning a model from a source data distribution and applying that model on a different (but related) target data distribution. For instance, one of the tasks of the common spam filtering problem consists in adapting a model from one user (the source distribution) to a new user who receives significantly different emails (the target distribution). Domain adaptation has also been shown to be beneficial to learning unrelated sources.[4] Note that, when more than one source distribution is available the problem is referred to as multi-source domain adaptation.[5]

Overview

Domain adaptation is the ability to apply an algorithm trained in one or more "source domains" to a different (but related) "target domain". Domain adaptation is a subcategory of transfer learning. In domain adaptation, the source and target domains all have the same feature space (but different distributions); in contrast, transfer learning includes cases where the target domain's feature space is different from the source feature space or spaces.[6]

Domain shift

A domain shift,[7] or distributional shift,[8] is a change in the data distribution between an algorithm's training dataset, and a dataset it encounters when deployed. These domain shifts are common in practical applications of artificial intelligence. Conventional machine-learning algorithms often adapt poorly to domain shifts. The modern machine-learning community has many different strategies to attempt to gain better domain adaptation.[7]

Examples

Other applications include wifi localization detection and many aspects of computer vision.[6]

Formalization

Let

X

be the input space (or description space) and let

Y

be the output space (or label space). The objective of a machine learning algorithm is to learn a mathematical model (a hypothesis)

h:X\toY

able to attach a label from

Y

to an example from

X

. This model is learned from a learning sample

S=\{(xi,yi)\in(X x

m
Y)\}
i=1
.

Usually in supervised learning (without domain adaptation), we suppose that the examples

(xi,yi)\inS

are drawn i.i.d. from a distribution

DS

of support

X x Y

(unknown and fixed). The objective is then to learn

h

(from

S

) such that it commits the least error possible for labelling new examples coming from the distribution

DS

.

The main difference between supervised learning and domain adaptation is that in the latter situation we study two different (but related) distributions

DS

and

DT

on

X x Y

. The domain adaptation task then consists of the transfer of knowledge from the source domain

DS

to the target one

DT

. The goal is then to learn

h

(from labeled or unlabelled samples coming from the two domains) such that it commits as little error as possible on the target domain

DT

.

The major issue is the following: if a model is learned from a source domain, what is its capacity to correctly label data coming from the target domain?

The different types of domain adaptation

There are several contexts of domain adaptation. They differ in the information considered for the target task.

  1. The unsupervised domain adaptation: the learning sample contains a set of labeled source examples, a set of unlabeled source examples and a set of unlabeled target examples.
  2. The semi-supervised domain adaptation: in this situation, we also consider a "small" set of labeled target examples.
  3. The supervised domain adaptation: all the examples considered are supposed to be labeled.

Four algorithmic principles

Reweighting algorithms

The objective is to reweight the source labeled sample such that it "looks like" the target sample (in terms of the error measure considered).[14] [15]

Iterative algorithms

A method for adapting consists in iteratively "auto-labeling" the target examples.[16] The principle is simple:

  1. a model

h

is learned from the labeled examples;

h

automatically labels some target examples;
  1. a new model is learned from the new labeled examples.

Note that there exist other iterative approaches, but they usually need target labeled examples.[17] [18]

Search of a common representation space

The goal is to find or construct a common representation space for the two domains. The objective is to obtain a space in which the domains are close to each other while keeping good performances on the source labeling task.This can be achieved through the use of Adversarial machine learning techniques where feature representations from samples in different domains are encouraged to be indistinguishable.[19] [20]

Hierarchical Bayesian Model

p(n)

, which is essentially a factorization model for counts

n

, to derive domain-dependent latent representations allowing both domain-specific and globally shared latent factors.[4]

Softwares

Several compilations of domain adaptation and transfer learning algorithms have been implemented over the past decades:

Notes and References

  1. Book: Redko . Ievgen . Morvant . Emilie . Habrard . Amaury . Sebban . Marc . Bennani . Younès . Advances in Domain Adaptation Theory . 2019 . ISTE Press - Elsevier . 9781785482366 . 187 .
  2. Book: Bridle. John S.. Cox. Stephen J. RecNorm: Simultaneous normalisation and classification applied to speech recognition. Conference on Neural Information Processing Systems (NIPS). 1990. 234–240. http://papers.nips.cc/paper/328-recnorm-simultaneous-normalisation-and-classification-applied-to-speech-recognition.pdf.
  3. Ben-David. Shai. Blitzer. John. Crammer. Koby. Kulesza. Alex. Pereira. Fernando. Wortman Vaughan. Jennifer. A theory of learning from different domains. Machine Learning. 2010. 79. 1–2. 151–175. 10.1007/s10994-009-5152-4. free.
  4. 1810.09433. Hajiramezanali. Ehsan. Bayesian multi-domain learning for cancer subtype discovery from next-generation sequencing count data. Siamak Zamani Dadaneh. Karbalayghareh. Alireza. Zhou. Mingyuan. Qian. Xiaoning. stat.ML. 2018.
  5. Crammer. Koby. Kearns. Michael. Wortman. Jeniifer. Learning from Multiple Sources. Journal of Machine Learning Research. 2008. 9. 1757–1774.
  6. Sun . Shiliang . Shi . Honglei . Wu . Yuanbin . A survey of multi-source domain adaptation . Information Fusion . July 2015 . 24 . 84–92 . 10.1016/j.inffus.2014.12.003. 18385140 .
  7. Sun, Baochen, Jiashi Feng, and Kate Saenko. "Return of frustratingly easy domain adaptation." In Thirtieth AAAI Conference on Artificial Intelligence. 2016.
  8. Amodei, Dario, Chris Olah, Jacob Steinhardt, Paul Christiano, John Schulman, and Dan Mané. "Concrete problems in AI safety." arXiv preprint arXiv:1606.06565 (2016).
  9. Daumé III, Hal. "Frustratingly easy domain adaptation." arXiv preprint arXiv:0907.1815 (2009).
  10. Ben-David, Shai, John Blitzer, Koby Crammer, and Fernando Pereira. "Analysis of representations for domain adaptation." In Advances in neural information processing systems, pp. 137-144. 2007.
  11. Hu . Yipeng . Jacob . Joseph . Parker . Geoffrey J. M. . Hawkes . David J. . Hurst . John R. . Stoyanov . Danail . The challenges of deploying artificial intelligence models in a rapidly evolving pandemic . Nature Machine Intelligence . June 2020 . 2 . 6 . 298–300 . 10.1038/s42256-020-0185-2. 2005.12137 . en . 2522-5839. free .
  12. News: Matthews . Dylan . AI disaster won't look like the Terminator. It'll be creepier. . 21 June 2020 . Vox . 26 March 2019 . en.
  13. News: Our weird behavior during the pandemic is messing with AI models . 21 June 2020 . MIT Technology Review . 11 May 2020 . en.
  14. Book: Huang. Jiayuan. Smola. Alexander J.. Gretton. Arthur. Borgwardt. Karster M.. Schölkopf. Bernhard. Correcting Sample Selection Bias by Unlabeled Data. Conference on Neural Information Processing Systems (NIPS). 2006. 601–608. http://papers.nips.cc/paper/3075-correcting-sample-selection-bias-by-unlabeled-data.pdf.
  15. Shimodaira. Hidetoshi. Improving predictive inference under covariate shift by weighting the log-likelihood function. Journal of Statistical Planning and Inference. 2000. 90. 2. 227–244. 10.1016/S0378-3758(00)00115-4. 9238949 .
  16. Gallego . A.J. . Calvo-Zaragoza . J. . Fisher . R.B. . 2020 . Incremental Unsupervised Domain-Adversarial Training of Neural Networks . IEEE Transactions on Neural Networks and Learning Systems . PP . 11 . 4864–4878 . 10.1109/TNNLS.2020.3025954 . 33027004 . 20.500.11820/72ba0443-8a7d-4cdd-8212-38682d4f0730 . 210164756 . free .
  17. Arief-Ang. I.B.. Salim. F.D.. Hamilton. M. . 2017-11-08. DA-HOC: semi-supervised domain adaptation for room occupancy prediction using CO2 sensor data. 4th ACM International Conference on Systems for Energy-Efficient Built Environments (BuildSys). 1–10. 10.1145/3137133.3137146. Delft, Netherlands. 978-1-4503-5544-5.
  18. Arief-Ang . I.B. . Hamilton . M. . Salim . F.D. . 2018-12-01 . A Scalable Room Occupancy Prediction with Transferable Time Series Decomposition of CO2 Sensor Data . ACM Transactions on Sensor Networks . 14 . 3–4 . 21:1–21:28 . 10.1145/3217214 . 54066723 .
  19. Ganin . Yaroslav . Ustinova . Evgeniya . Ajakan . Hana . Germain . Pascal . Larochelle . Hugo . Laviolette . François . Marchand . Mario . Lempitsky . Victor . 2016 . Domain-Adversarial Training of Neural Networks . Journal of Machine Learning Research . 17 . 1–35 .
  20. 1703.01461. Hajiramezanali. Ehsan. Addressing Appearance Change in Outdoor Robotics with Adversarial Domain Adaptation. Siamak Zamani Dadaneh. Karbalayghareh. Alireza. Zhou. Mingyuan. Qian. Xiaoning. cs.RO. 2017.
  21. de Mathelin, Antoine and Deheeger, François and Richard, Guillaume and Mougeot, Mathilde and Vayatis, Nicolas (2020) "ADAPT: Awesome Domain Adaptation Python Toolbox"
  22. Mingsheng Long Junguang Jiang, Bo Fu. (2020) "Transfer-learning-library"
  23. Ke Yan. (2016) "Domain adaptation toolbox"