A Selective Multiple Instance Transfer Learning Method for Text Categorization Problems

作者:

Highlights:

摘要

Multiple instance learning (MIL) is a generalization of supervised learning which attempts to learn a distinctive classifier from bags of instances. This paper addresses the problem of the transfer learning-based multiple instance method for text categorization problem. To provide a safe transfer of knowledge from a source task to a target task, this paper proposes a new approach, called selective multiple instance transfer learning (SMITL), which selects the case that the multiple instance transfer learning will work in step one, and then builds a multiple instance transfer learning classifier in step two. Specifically, in the first step, we measure whether the source task and the target task are related or not by investigating the similarity of the positive features of both tasks. In the second step, we construct a transfer learning-based multiple instance method to transfer knowledge from a source task to a target task if both tasks are found to be related in the first step. Our proposed approach explicitly addresses the problem of safe transfer of knowledge for multiple instance learning on the text classification problem. Extensive experiments have shown that SMITL can determine whether the two tasks are related for most data sets, and outperforms classic multiple instance learning methods.

论文关键词:Data Mining,Transfer Learning

论文评审过程:Received 2 December 2016, Revised 10 November 2017, Accepted 13 November 2017, Available online 14 November 2017, Version of Record 19 December 2017.

论文官网地址:https://doi.org/10.1016/j.knosys.2017.11.019