Publication Type
Conference Proceeding Article
Version
submittedVersion
Publication Date
6-2006
Abstract
In supervised learning, most single solution neural networks such as constructive backpropagation give good results when used with some datasets but not with others. Others such as probabilistic neural networks (PNN) fit a curve to perfection but need to be manually tuned in the case of noisy data. Recursive percentage based hybrid pattern training (RPHP) overcomes this problem by recursively training subsets of the data, thereby using several neural networks. MultiLearner based recursive training (MLRT) is an extension of this approach, where a combination of existing and new learners are used and subsets are trained using the weak learner which is best suited for this subset. We observed that empirically, MLRT performs considerably well as compared to RPHP and other systems on benchmark data with 11% improvement in accuracy on the spam dataset and comparable performances on the vowel and the two-spiral problems
Keywords
Neural Networks, Supervised Learning, Probabilistic Neural Networks (PNN), Backpropagation
Discipline
Databases and Information Systems
Research Areas
Data Science and Engineering
Publication
Proceedings of the 2006 IEEE Conference on Cybernetics and Intelligent Systems, Bangkok, Thailand, June 7-9
ISBN
1424400228
Identifier
10.1109/ICCIS.2006.252267
Publisher
IEEE
City or Country
Bangkok, Thailand
Citation
RAMANATHAN, Kiruthika; GUAN, Sheng Uei; and IYER, Laxmi R..
MultiLearner based recursive supervised training. (2006). Proceedings of the 2006 IEEE Conference on Cybernetics and Intelligent Systems, Bangkok, Thailand, June 7-9.
Available at: https://ink.library.smu.edu.sg/sis_research/7393
Creative Commons License
This work is licensed under a Creative Commons Attribution-NonCommercial-No Derivative Works 4.0 International License.
Additional URL
http://doi.org/10.1109/ICCIS.2006.252267