Multi-Transfer: Transfer Learning with Multiple Views and Multiple Sources

被引:39
作者
Tan, Ben [1 ]
Zhong, Erheng [1 ]
Xiang, Evan Wei [1 ]
Yang, Qiang [1 ]
机构
[1] Hong Kong Univ Sci & Technol, Dept Comp Sci & Engn, Hong Kong, Hong Kong, Peoples R China
关键词
transfer learning; multi-view learning; multiple data sources;
D O I
10.1002/sam.11226
中图分类号
TP18 [人工智能理论];
学科分类号
140502 [人工智能];
摘要
Transfer learning, which aims to help learning tasks in a target domain by leveraging knowledge from auxiliary domains, has been demonstrated to be effective in different applications such as text mining, sentiment analysis, and so on. In addition, in many real-world applications, auxiliary data are described from multiple perspectives and usually carried by multiple sources. For example, to help classify videos on Youtube, which include three perspectives: image, voice and subtitles, one may borrow data from Flickr, Last. FM and Google News. Although any single instance in these domains can only cover a part of the views available on Youtube, the piece of information carried by them may compensate one another. If we can exploit these auxiliary domains in a collective manner, and transfer the knowledge to the target domain, we can improve the target model building from multiple perspectives. In this article, we consider this transfer learning problem as Transfer Learning with Multiple Views and Multiple Sources. As different sources may have different probability distributions and different views may compensate or be inconsistent with each other, merging all data in a simplistic manner will not give an optimal result. Thus, we propose a novel algorithm to leverage knowledge from different views and sources collaboratively, by letting different views from different sources complement each other through a co-training style framework, at the same time, it revises the distribution differences in different domains. We conduct empirical studies on several real-world datasets to show that the proposed approach can improve the classification accuracy by up to 8% against different kinds of state-of-the-art baselines. (C) 2014 Wiley Periodicals, Inc.
引用
收藏
页码:282 / 293
页数:12
相关论文
共 24 条
[1]
Anguita D., 2012, P INT WORKSH AMB ASS, P216
[2]
[Anonymous], 2011, P 22 INT JOINT C ART
[3]
[Anonymous], 2012, ACM Trans. Knowl. Discov. Data, DOI DOI 10.1145/2382577.2382582
[4]
[Anonymous], 2003, P 20 INT C MACH LEAR
[5]
[Anonymous], 2010, SDM
[6]
Belkin M, 2006, J MACH LEARN RES, V7, P2399
[7]
Blum A., 1998, Proceedings of the Eleventh Annual Conference on Computational Learning Theory, P92, DOI 10.1145/279943.279962
[8]
Chen M., 2011, NIPS, V24, P2456
[9]
Dai W., 2007, P 24 INT C MACH LEAR, P193, DOI [10.1145/1273496.1273521, DOI 10.1145/1273496.1273521]
[10]
Duan J, 2012, IEEE IC COMP COM NET