Multi-task learning with one-class SVM
Résumé
Multi-task learning technologies have been developed to be an effective way to improve the generalization performance by training multiple related tasks simultaneously. The determination of the relatedness between tasks is usu- ally the key to the formulation of a multi-task learning method. In this paper, we make the assumption that when tasks are related to each other, usually their models are close enough, that is, their models or their model parameters are close to a certain mean function. Following this task relat- edness assumption, two multi-task learning formulations based on one-class Support Vector Machines (one-class SVM) are presented. With the help of new kernel design, both multi-task learning methods can be solved by the optimization program of a single one-class SVM. Experiments conducted on both low dimensional nonlinear toy dataset and high dimensional textured images show that our approaches lead to very encouraging results.
Origine | Fichiers produits par l'(les) auteur(s) |
---|