TY - JOUR
T1 - A data-centric review of deep transfer learning with applications to text data
AU - Bashath, Samar
AU - Perera, Nadeesha
AU - Tripathi, Shailesh
AU - Manjang, Kalifa
AU - Dehmer, Matthias
AU - Streib, Frank Emmert
N1 - Funding Information:
MD thanks the Austrian Science Funds for supporting this work (project P30031).
Publisher Copyright:
© 2021 The Author(s)
PY - 2022/3
Y1 - 2022/3
N2 - In recent years, many applications are using various forms of deep learning models. Such methods are usually based on traditional learning paradigms requiring the consistency of properties among the feature spaces of the training and test data and also the availability of large amounts of training data, e.g., for performing supervised learning tasks. However, many real-world data do not adhere to such assumptions. In such situations transfer learning can provide feasible solutions, e.g., by simultaneously learning from data-rich source data and data-sparse target data to transfer information for learning a target task. In this paper, we survey deep transfer learning models with a focus on applications to text data. First, we review the terminology used in the literature and introduce a new nomenclature allowing the unequivocal description of a transfer learning model. Second, we introduce a visual taxonomy of deep learning approaches that provides a systematic structure to the many diverse models introduced until now. Furthermore, we provide comprehensive information about text data that have been used for studying such models because only by the application of methods to data, performance measures can be estimated and models assessed.
AB - In recent years, many applications are using various forms of deep learning models. Such methods are usually based on traditional learning paradigms requiring the consistency of properties among the feature spaces of the training and test data and also the availability of large amounts of training data, e.g., for performing supervised learning tasks. However, many real-world data do not adhere to such assumptions. In such situations transfer learning can provide feasible solutions, e.g., by simultaneously learning from data-rich source data and data-sparse target data to transfer information for learning a target task. In this paper, we survey deep transfer learning models with a focus on applications to text data. First, we review the terminology used in the literature and introduce a new nomenclature allowing the unequivocal description of a transfer learning model. Second, we introduce a visual taxonomy of deep learning approaches that provides a systematic structure to the many diverse models introduced until now. Furthermore, we provide comprehensive information about text data that have been used for studying such models because only by the application of methods to data, performance measures can be estimated and models assessed.
KW - Deep learning
KW - Domain adaptation
KW - Machine learning
KW - Natural language processing
KW - Transfer learning
UR - http://www.scopus.com/inward/record.url?scp=85120668947&partnerID=8YFLogxK
U2 - 10.1016/j.ins.2021.11.061
DO - 10.1016/j.ins.2021.11.061
M3 - Article
AN - SCOPUS:85120668947
SN - 0020-0255
VL - 585
SP - 498
EP - 528
JO - Information Sciences
JF - Information Sciences
ER -