Commenced in January 2007
Paper Count: 31108
Performance Comparison and Evaluation of AdaBoost and SoftBoost Algorithms on Generic Object Recognition
Abstract:SoftBoost is a recently presented boosting algorithm, which trades off the size of achieved classification margin and generalization performance. This paper presents a performance evaluation of SoftBoost algorithm on the generic object recognition problem. An appearance-based generic object recognition model is used. The evaluation experiments are performed using a difficult object recognition benchmark. An assessment with respect to different degrees of label noise as well as a comparison to the well known AdaBoost algorithm is performed. The obtained results reveal that SoftBoost is encouraged to be used in cases when the training data is known to have a high degree of noise. Otherwise, using Adaboost can achieve better performance.
Digital Object Identifier (DOI): doi.org/10.5281/zenodo.1060339Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1530
 Ayhan Demiriz, Kristin P. Bennett, and John Shawe-Taylor. Linear programming boosting via column generation. Machine Learning, 46(1- 3):225-254, 2002.
 Yoav. Freund and Robert E. Schapire. A decision theoretic generalization of online learning and application to boosting. Computer and System Sciences, 55(1):119-139, 1997.
 Doaa Hegazy and Joachim Denzler. Boosting local colored features for generic object recognition. Pattern Recognition and Image Understanding, 18:323-327, 2008.
 David G. Lowe. Distinctive image features from scale-invariant keypoints. International Journal of Computer Vision, 60:91-110, 2004.
 Krystian Mikolajczyk and Cordelia Schmid. An affine invariant interest point detector. In 7th European Conference on Computer Vision ECCV02, pages 128-142, 2002.
 Pierre Moreels and Pietro Perona. Evaluation of features detectors and descriptors based on 3d objects. Int. J. Comput. Vision, 73(3):263-284, 2007.
 Andreas Opelt, Axel Pinz, Michael Fussenegger, and Peter Auer. Generic object recognirion with boosting. IEEE Transactions on Pattern Analysis and Machine Intelligence, 28(3):416-431, 2006.
 Gunnar. R┬¿atsch, T. Onoda, and K.-R. M┬¿uller. Soft margins for AdaBoost. Machine Learning, 42(3):287-320, March 2001.
 Gunnar R┬¿atsch and Manfred K. Warmuth. Efficient margin maximizing with boosting. Journal of Machine Learning Research, 6:2131-2152, 2002.
 Robert E. Schapire, Yoav Freund, Peter Bartlett, and Wee Sun Lee. Boosting the margin: a new explanation for the effectiveness of voting methods. In Proc. 14th International Conference on Machine Learning, pages 322-330. Morgan Kaufmann, 1997.
 Robert E. Schapire and Yoram Singer. Improved boosting algorithms using confidence-rated predictions. Machine Learning, 37:297-336, 1999.
 Zehang Sun, George Bebis, and Ronald Miller. Boosting object detection using feature selection. page 290, 2003.
 Joost van de Weijer and Cordelia Schmid. Coloring Local Feature Extraction. In 8th European Conference on Computer Vision ECCV06, volume 2, pages 334-348, 2006.
 Paul Viola and Michael Jones. Rapid object detection uisng a boosted cascade of simple features. In IEEE Computer Scociety Conference on Computer Vision and Pattern Recognition CVPR01, volume 1, pages 511-518, 2001.
 Manfred K. Warmuth, Karen Glocer, and Gunnar R┬¿atsch. Boosting algorithms for maximizining the soft margin. Advances in Neural Information Processing Systems (NIPS-08), 2008.
 Manfred K. Warmuth, Jun Liao, and Gunnar R┬¿atsch. Totally corrective boosting algorithms that maximize the margin. In ICML -06: Proceedings of the 23rd international conference on Machine learning, pages 1001-1008, New York, NY, USA, 2006. ACM.