Transfer learning explained
Transfer learning (TL) is a technique in machine learning (ML) in which knowledge learned from a task is re-used in order to boost performance on a related task.[1] For example, for image classification, knowledge gained while learning to recognize cars could be applied when trying to recognize trucks. This topic is related to the psychological literature on transfer of learning, although practical ties between the two fields are limited. Reusing/transferring information from previously learned tasks to new tasks has the potential to significantly improve learning efficiency.[2]
Since transfer learning makes use of training with multiple objective functions it is related to cost-sensitive machine learning and multi-objective optimization.[3]
History
In 1976, Bozinovski and Fulgosi published a paper addressing transfer learning in neural network training.[4] [5] The paper gives a mathematical and geometrical model of the topic. In 1981, a report considered the application of transfer learning to a dataset of images representing letters of computer terminals, experimentally demonstrating positive and negative transfer learning.[6]
In 1992, Lorien Pratt formulated the discriminability-based transfer (DBT) algorithm.[7]
By 1998, the field had advanced to include multi-task learning,[8] along with more formal theoretical foundations.[9] Influential publications on transfer learning include the book Learning to Learn in 1998, a 2009 survey[10] and a 2019 survey.[11]
Ng said in his NIPS 2016 tutorial[12] that TL would become the next driver of machine learning commercial success after supervised learning.
In the 2020 paper, "Rethinking Pre-Training and self-training",[13] Zoph et al. reported that pre-training can hurt accuracy, and advocate self-training instead.
Definition
The definition of transfer learning is given in terms of domains and tasks. A domain
consists of: a feature space
and a
marginal probability distribution
, where
. Given a specific domain,
, a task consists of two components: a label space
and an objective predictive function
. The function
is used to predict the corresponding label
of a new instance
. This task, denoted by
, is learned from the training data consisting of pairs
, where
and
.
[14] Given a source domain
and learning task
, a target domain
and learning task
, where
, or
, transfer learning aims to help improve the learning of the target predictive function
in
using the knowledge in
and
.
[14] Applications
Algorithms are available for transfer learning in Markov logic networks and Bayesian networks. Transfer learning has been applied to cancer subtype discovery,[15] building utilization,[16] [17] general game playing,[18] text classification,[19] [20] digit recognition,[21] medical imaging and spam filtering.[22]
In 2020, it was discovered that, due to their similar physical natures, transfer learning is possible between electromyographic (EMG) signals from the muscles and classifying the behaviors of electroencephalographic (EEG) brainwaves, from the gesture recognition domain to the mental state recognition domain. It was noted that this relationship worked in both directions, showing that electroencephalographic can likewise be used to classify EMG.[23] The experiments noted that the accuracy of neural networks and convolutional neural networks were improved[24] through transfer learning both prior to any learning (compared to standard random weight distribution) and at the end of the learning process (asymptote). That is, results are improved by exposure to another domain. Moreover, the end-user of a pre-trained model can change the structure of fully-connected layers to improve performance.[25]
Software
Several compilations of transfer learning and domain adaptation algorithms have been implemented:
- ADAPT[26] (Python)
- TLlib[27] (Python)
- Domain-Adaptation-Toolbox[28] (Matlab)
See also
Sources
- Book: [{{google books|plainurl=y|id=X_jpBwAAQBAJ}} Learning to Learn]. Thrun. Sebastian. Pratt. Lorien. 6 December 2012. Springer Science & Business Media. 978-1-4615-5529-2.
Notes and References
- Web site: West . Jeremy . Dan . Ventura . Sean . Warnick . Spring Research Presentation: A Theoretical Foundation for Inductive Transfer . Brigham Young University, College of Physical and Mathematical Sciences . 2007 . 2007-08-05 . dead . https://web.archive.org/web/20070801120743/http://cpms.byu.edu/springresearch/abstract-entry?id=861 . 2007-08-01 .
- George Karimpanal. Thommen. Bouffanais. Roland. 2019. Self-organizing maps for storage and transfer of knowledge in reinforcement learning. Adaptive Behavior. 27. 2. 111–126. 10.1177/1059712318818568. 1059-7123. 1811.08318. 53774629.
- Cost-Sensitive Machine Learning. (2011). USA: CRC Press, Page 63, https://books.google.de/books?id=8TrNBQAAQBAJ&pg=PA63
- Stevo. Bozinovski and Ante Fulgosi (1976). "The influence of pattern similarity and transfer learning upon the training of a base perceptron B2." (original in Croatian) Proceedings of Symposium Informatica 3-121-5, Bled.
- Stevo Bozinovski (2020) "Reminder of the first paper on transfer learning in neural networks, 1976". Informatica 44: 291–302.
- S. Bozinovski (1981). "Teaching space: A representation concept for adaptive pattern classification." COINS Technical Report, the University of Massachusetts at Amherst, No 81-28 [available online: UM-CS-1981-028.pdf]
- Book: Pratt, L. Y. . [{{google books|plainurl=y|id=6tGHlwEACAAJ|page=204}} NIPS Conference: Advances in Neural Information Processing Systems 5 ]. Morgan Kaufmann Publishers . 1992 . 204–211 . Discriminability-based transfer between neural networks . https://proceedings.neurips.cc/paper/1992/file/67e103b0761e60683e83c559be18d40c-Paper.pdf.
- Caruana, R., "Multitask Learning", pp. 95-134 in
- Baxter, J., "Theoretical Models of Learning to Learn", pp. 71-95
- Pan . Sinno Jialin . Yang . Qiang . 2009 . A Survey on Transfer Learning . IEEE.
- 2019 . A Comprehensive Survey on Transfer Learning . IEEE . 1911.02685 . Zhuang . Fuzhen . Qi . Zhiyuan . Duan . Keyu . Xi . Dongbo . Zhu . Yongchun . Zhu . Hengshu . Xiong . Hui . He . Qing .
- Web site: Nuts and bolts of building AI applications using Deep Learning, slides.
- Zoph . Barret . Rethinking pre-training and self-training . Advances in Neural Information Processing Systems . 2020 . 33 . 3833–3845 . 2006.06882 . 2022-12-20.
- Lin . Yuan-Pin . Jung . Tzyy-Ping . Improving EEG-Based Emotion Classification Using Conditional Transfer Learning . Frontiers in Human Neuroscience . 27 June 2017 . 11 . 334 . 10.3389/fnhum.2017.00334. 28701938 . 5486154 . free . Material was copied from this source, which is available under a Creative Commons Attribution 4.0 International License.
- Hajiramezanali, E. & Dadaneh, S. Z. & Karbalayghareh, A. & Zhou, Z. & Qian, X. Bayesian multi-domain learning for cancer subtype discovery from next-generation sequencing count data. 32nd Conference on Neural Information Processing Systems (NeurIPS 2018), Montréal, Canada.
- Arief-Ang. I.B.. Salim. F.D.. Hamilton. M.. 2017-11-08. DA-HOC: semi-supervised domain adaptation for room occupancy prediction using CO2 sensor data. 4th ACM International Conference on Systems for Energy-Efficient Built Environments (BuildSys). Delft, Netherlands. 1–10. 10.1145/3137133.3137146. 978-1-4503-5544-5.
- Arief-Ang . I.B. . Hamilton . M. . Salim . F.D. . 2018-12-01 . A Scalable Room Occupancy Prediction with Transferable Time Series Decomposition of CO2 Sensor Data . ACM Transactions on Sensor Networks . 14 . 3–4 . 21:1–21:28 . 10.1145/3217214 . 54066723 .
- Banerjee, Bikramjit, and Peter Stone. "General Game Learning Using Knowledge Transfer." IJCAI. 2007.
- Do. Chuong B.. Ng. Andrew Y.. 2005. Neural Information Processing Systems Foundation, NIPS*2005. 2007-08-05. Transfer learning for text classification.
- Rajat. Raina. Ng. Andrew Y.. Koller. Daphne. 2006. Twenty-third International Conference on Machine Learning. 2007-08-05. Constructing Informative Priors using Transfer Learning.
- Book: Maitra. D. S.. Bhattacharya. U.. Parui. S. K.. 2015 13th International Conference on Document Analysis and Recognition (ICDAR) . CNN based common approach to handwritten character recognition of multiple scripts . August 2015. 1021–1025. 10.1109/ICDAR.2015.7333916. 978-1-4799-1805-8. 25739012.
- Bickel. Steffen. 2006. ECML-PKDD Discovery Challenge Workshop. 2007-08-05. ECML-PKDD Discovery Challenge 2006 Overview.
- Bird . Jordan J. . Kobylarz . Jhonatan . Faria . Diego R. . Ekart . Aniko . Ribeiro . Eduardo P. . Cross-Domain MLP and CNN Transfer Learning for Biological Signal Processing: EEG and EMG . IEEE Access . Institute of Electrical and Electronics Engineers (IEEE) . 8 . 2020 . 2169-3536 . 10.1109/access.2020.2979074 . 54789–54801. free . 2020IEEEA...854789B .
- Book: Maitra. Durjoy Sen. Bhattacharya. Ujjwal. Parui. Swapan K.. 2015 13th International Conference on Document Analysis and Recognition (ICDAR) . CNN based common approach to handwritten character recognition of multiple scripts . August 2015. https://ieeexplore.ieee.org/document/7333916. 1021–1025. 10.1109/ICDAR.2015.7333916. 978-1-4799-1805-8. 25739012.
- SpinalNet: Deep Neural Network with Gradual Input. H. M. Dipu. Kabir. Moloud. Abdar. Seyed Mohammad Jafar. Jalali. Abbas. Khosravi. Amir F.. Atiya. Saeid. Nahavandi. Dipti. Srinivasan. Dipti Srinivasan. January 7, 2022. IEEE Transactions on Artificial Intelligence. 4 . 5 . 1165–1177 . 10.1109/TAI.2022.3185179. 2007.03347 . 220381239 .
- de Mathelin, Antoine and Deheeger, François and Richard, Guillaume and Mougeot, Mathilde and Vayatis, Nicolas (2020) "ADAPT: Awesome Domain Adaptation Python Toolbox"
- Mingsheng Long Junguang Jiang, Bo Fu. (2020) "Transfer-learning-library"
- Ke Yan. (2016) "Domain adaptation toolbox"