Please use this identifier to cite or link to this item: http://hdl.handle.net/20.500.12188/17149
Title: Two stage architecture for multi-label learning
Authors: Madjarov, Gjorgji 
GJorgjevikj, Dejan 
Džeroski, Sašo
Issue Date: Mar-2012
Publisher: Elsevier BV
Journal: Pattern Recognition 
Abstract: A common approach to solving multi-label learning problems is to use problem transformation methods and dichotomizing classifiers as in the pair-wise decomposition strategy. One of the problems with this strategy is the need for querying a quadratic number of binary classifiers for making a prediction that can be quite time consuming, especially in learning problems with a large number of labels. To tackle this problem, we propose a Two Stage Architecture (TSA) for efficient multi-label learning. We analyze three implementations of this architecture the Two Stage Voting Method (TSVM), the Two Stage Classifier Chain Method (TSCCM) and the Two Stage Pruned Classifier Chain Method (TSPCCM). Eight different real-world datasets are used to evaluate the performance of the proposed methods. The performance of our approaches is compared with the performance of two algorithm adaptation methods (Multi-Label k-NN and Multi-Label C4.5) and five problem transformation methods (Binary Relevance, Classifier Chain, Calibrated Label Ranking with majority voting, the Quick Weighted method for pair-wise multi-label learning and the Label Powerset method). The results suggest that TSCCM and TSPCCM outperform the competing algorithms in terms of predictive accuracy, while TSVM has comparable predictive performance. In terms of testing speed, all three methods show better performance as compared to the pair-wise methods for multi-label learning.
URI: http://hdl.handle.net/20.500.12188/17149
DOI: 10.1016/j.patcog.2011.08.011
Appears in Collections:Faculty of Computer Science and Engineering: Journal Articles

Show full item record

Page view(s)

24
checked on Apr 26, 2024

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.