Abstract
Kernel Generalised Learning Vector Quantisation (KGLVQ) was proposed to extend Generalised Learning Vector Quantisation into the kernel feature space to deal with complex class boundaries and thus yielded promising performance for complex classification tasks in pattern recognition. However KGLVQ does not follow the maximal margin principle, which is crucial for kernel-based learning methods. In this paper we propose a maximal margin approach (MLVQ) to the KGLVQ algorithm. MLVQ inherits the merits of KGLVQ and also follows the maximal margin principle to improve the generalisation capability. Experiments performed on the well-known data sets available in UCI repository show promising classification results for the proposed method
Original language | English |
---|---|
Title of host publication | The 2013 International Joint Conference on Neural Networks (IJCNN) |
Editors | Plamen Angelov, Daniel Levine |
Place of Publication | USA |
Publisher | IEEE, Institute of Electrical and Electronics Engineers |
Pages | 1668-1673 |
Number of pages | 6 |
Volume | 1 |
ISBN (Print) | 9781467361293 |
DOIs | |
Publication status | Published - 2013 |
Event | 2013 International Joint Conference on Neural Networks (IJCNN) - Dallas, Texas, United States Duration: 4 Aug 2013 → 9 Aug 2013 |
Conference
Conference | 2013 International Joint Conference on Neural Networks (IJCNN) |
---|---|
Country/Territory | United States |
City | Texas |
Period | 4/08/13 → 9/08/13 |