Metric Learning for Multi-Output Tasks

Publication Type:
Journal Article
IEEE Transactions on Pattern Analysis and Machine Intelligence, 2019, 41 (2), pp. 408 - 422
Issue Date:
Filename Description Size
08263148.pdfPublished Version524.29 kB
Adobe PDF
Full metadata record
© 1979-2012 IEEE. Multi-output learning with the task of simultaneously predicting multiple outputs for an input has increasingly attracted interest from researchers due to its wide application. The k nearest neighbor (k \text{NN}) algorithm is one of the most popular frameworks for handling multi-output problems. The performance of k \text{NN} depends crucially on the metric used to compute the distance between different instances. However, our experiment results show that the existing advanced metric learning technique cannot provide an appropriate distance metric for multi-output tasks. This paper systematically studies how to efficiently learn an appropriate distance metric for multi-output problems with provable guarantee. In particular, we present a novel large margin metric learning paradigm for multi-output tasks, which projects both the input and output into the same embedding space and then learns a distance metric to discover output dependency such that instances with very different multiple outputs will be moved far away. Several strategies are then proposed to speed up the training and testing time. Moreover, we study the generalization error bound of our method for three learning tasks, which shows that our method converges to the optimal solutions. Experiments on three multi-output learning tasks (multi-label classification, multi-target regression, and multi-concept retrieval) validate the effectiveness and scalability of the proposed method.
Please use this identifier to cite or link to this item: