Multi-task Active Learning with Output Constraints (2010)

Yi Zhang


Many problems in information extraction, text mining, natural language processing and other fields exhibit the same property: multiple prediction tasks are related in the sense that their outputs (labels) satisfy certain constraints. In this paper, we propose an active learning framework exploiting such relations among tasks. Intuitively, with task outputs coupled by constraints, active learning can utilize not only the uncertainty of the prediction in a single task but also the inconsistency of predictions across tasks. We formalize this idea as a cross-task value of information criteria, in which the reward of a labeling assignment is propagated and measured over all relevant tasks reachable through constraints. A specific example of our framework leads to the cross entropy measure on the predictions of coupled tasks, which generalizes the entropy in the classical single-task uncertain sampling. We conduct experiments on two real-world problems: web information extraction and document classification. Empirical results demonstrate the effectiveness of our framework in actively collecting labeled examples for multiple related tasks.

Full text

Download (application/pdf, 108.6 kB)

Approximate BibTeX Entry

    Year = {2010},
    Booktitle = {AAAI 2010},
    Author = { Yi Zhang },
    Title = {Multi-task Active Learning with Output Constraints}

Copyright 2010, Carnegie Mellon University, Auton Lab. All Rights Reserved.