|
MULTITASK LEARNING
by Rich Caruana
Multitask Learning is an approach to inductive transfer that improves
generalization by using the domain information contained in the
training signals of related tasks as an inductive bias. It does this
by learning tasks in parallel while using a shared representation;
what is learned for each task can help other tasks be learned better.
This paper reviews prior work on MTL, presents new evidence that MTL
in backprop nets discovers task relatedness without the need of
supervisory signals, and presents new results for MTL with k-nearest
neighbor and kernel regression. In this paper we demonstrate
multitask learning in three domains. We explain how multitask
learning works, and show that there are many opportunities for
multitask learning in real domains. We present an algorithm and
results for multitask learning with case-based methods like k-nearest
neighbor and kernel regression, and sketch an algorithm for multitask
learning in decision trees. Because multitask learning works, can be
applied to many different kinds of domains, and can be used with
different learning algorithms, we conjecture there will be many
opportunities for its use on real-world problems.
|