Movatterモバイル変換


[0]ホーム

URL:


Skip to main content

Advertisement

Springer Nature Link
Log in

Sample Reduction Using\(\ell _1\)-Norm Twin Bounded Support Vector Machine

  • Conference paper
  • First Online:

Abstract

Twin support machine (TSVM) has a lower time complexity than support vector machine (SVM), but it has a poor ability to perform sample reduction. In order to improve the ability of TSVM to reduce sample, we propose an\(\ell _1\)-norm twin bounded support machine (\(\ell _1\)-TBSVM) inspired by the sparsity of\(\ell _1\)-norm in feature space. The objective function of\(\ell _1\)-TBSVM contains the hinge loss and the\(\ell _1\)-norm terms, both which can induce sparsity. We solve the primal programming problems of\(\ell _1\)-TBSVM to prevent the disappearance of sparsity and avoid the situation that the inverse of matrix does not exist. Thus,\(\ell _1\)-TBSVM has a good sparsity, or a good ability to reduce sample. Experimental results on synthetic and UCI datasets indicate that\(\ell _1\)-TBSVM has a good ability to perform sample reduction and simultaneously enhances the classification performance.

This is a preview of subscription content,log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
¥17,985 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
JPY 3498
Price includes VAT (Japan)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
JPY 16015
Price includes VAT (Japan)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
JPY 20019
Price includes VAT (Japan)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide -see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Similar content being viewed by others

References

  1. Bi, J., Bennett, K.P., Embrechts, M., Breneman, C.M., Song, M.: Dimensionality reduction via sparse support vector machines. J. Mach. Learn. Res.3, 1229–1243 (2003)

    MATH  Google Scholar 

  2. Cortes, C., Vapnik, V.: Support vector networks. Mach. Learn.20(3), 273–297 (1995)

    MATH  Google Scholar 

  3. Cristianini, N., Shawe-Taylor, J.: An Introduction to Support Vector Machines and Other Kernel-Based Learning Methods. Cambridge University, Cambridge (2000)

    Book  Google Scholar 

  4. Dheeru, D., Karra Taniskidou, E.: UCI machine learning repository (2017).http://archive.ics.uci.edu/ml

  5. Duda, R.O., Hart, P.E., Stork, D.G.: Pattern Classification, 2nd edn. Wiley, New York (2001)

    MATH  Google Scholar 

  6. Floyd, S., Warmuth, M.: Sample compression, learnability, and the Vapnik-Chervonenkis dimension. Mach. Learn.21(3), 269–304 (1995)

    Google Scholar 

  7. Gao, S., Ye, Q., Ye, N.: 1-norm least squares twin support vector machines. Neurocomputing74(17), 3590–3597 (2011)

    Article  Google Scholar 

  8. Graepel, T., Herbrich, R., Shawe-Taylor, J., Holloway, R.: Generalisation error bounds for sparse linear classifiers. In: Proceedings of the Thirteenth Annual Conference on Computational Learning Theory, pp. 298–303 (2000)

    Google Scholar 

  9. Huffener, F., Niedermeier, R., Wernicke, S.: Techniques for practical fixed-parameter algorithms. Comput. J.51(1), 7–25 (2008)

    Article  Google Scholar 

  10. Jayadeva, Khemchandani, R., Chandra, S.: Twin support vector machine for pattern classification. IEEE Trans. Pattern Anal. Mach. Intell.29(5), 905–910 (2007)

    Google Scholar 

  11. Kumar, M.A., Gopall, M.: Least squares twin support vector machine for pattern classification. Expert Syst. Appl.36, 7535–7543 (2009)

    Article  Google Scholar 

  12. Mangasarian, O.L., Wild, E.: Multisurface proximal support vector machine classification via generalized eigenvalues. IEEE Trans. Pattern Anal. Mach. Intell.28(1), 69–74 (2006)

    Article  Google Scholar 

  13. Niedermeier, R.: Invitation to Fixed-Parameter Algorithms. Oxford University Press, Oxford (2006)

    Google Scholar 

  14. Peng, X.: Building sparse twin support vector machine classifiers in primal space. Inf. Sci.181(18), 3967–3980 (2011)

    Article  Google Scholar 

  15. Peng, X.: TPSVM: a novel twin parametric-margin support vector machine for pattern recognition. Pattern Recogn.44, 2678–2692 (2011)

    Article  Google Scholar 

  16. Shao, Y., Zhang, C., Wang, X., Deng, N.: Improvements on twin support vector machine. IEEE Trans. Neural Netw.22(6), 962–968 (2011)

    Article  Google Scholar 

  17. Tanveer, M.: Robust and sparse linear programming twin support vector machines. Cogn. Comput.7(1), 137–149 (2015)

    Article  Google Scholar 

  18. Tian, Y., Ju, X., Qi, Z.: Efficient sparse nonparallel support vector machines for classification. Neural Comput. Appl.24(5), 1089–1099 (2014)

    Article  Google Scholar 

  19. Vapnik, V.N.: The Nature of Statistical Learning Theory. Springer, New York (2000).https://doi.org/10.1007/978-1-4757-3264-1

  20. Xu, Y., Wang, L., Zhong, P.: A rough margin-based v-twin support vector machine. Neural Comput. Appl.21, 1307–1317 (2012)

    Article  Google Scholar 

  21. Zhang, L., Zhou, W.: On the sparseness of 1-norm support vector machines. Neural Netw.23(3), 373–385 (2010)

    Article  Google Scholar 

  22. Zhang, L., et al.: Kernel sparse representation-based classifier. IEEE Trans. Sig. Process.60(4), 1684–1695 (2012)

    Article MathSciNet  Google Scholar 

  23. Zhang, L., Zhou, W., Jiao, L.: Hidden space support vector machine. IEEE Trans. Neural Netw.15(6), 1424–1434 (2004)

    Article  Google Scholar 

  24. Zhang, Z., Zhen, L., Deng, N., Tan, J.: Sparse least square twin support vector machine with adaptive norm. Appl. Intell.41(4), 1097–1107 (2014)

    Article  Google Scholar 

  25. Zhou, W., Zhang, L., Jiao, L.: Linear programming support vector machines. Pattern Recogn.35(12), 2927–2936 (2002)

    Article  Google Scholar 

Download references

Acknowledgment

This work was supported in part by the Natural Science Foundation of the Jiangsu Higher Education Institutions of China under Grant No. 19KJA550002, by the Six Talent Peak Project of Jiangsu Province of China under Grant No. XYDXX-054, by the Priority Academic Program Development of Jiangsu Higher Education Institutions, and by the Collaborative Innovation Center of Novel Software Technology and Industrialization.

Author information

Authors and Affiliations

  1. School of Computer Science and Technology, Joint International Research Laboratory of Machine Learning and Neuromorphic Computing, Soochow University, Suzhou, 215006, China

    Xiaohan Zheng, Li Zhang & Leilei Yan

  2. Provincial Key Laboratory for Computer Information Processing Technology, Soochow University, Suzhou, 215006, China

    Li Zhang

Authors
  1. Xiaohan Zheng

    You can also search for this author inPubMed Google Scholar

  2. Li Zhang

    You can also search for this author inPubMed Google Scholar

  3. Leilei Yan

    You can also search for this author inPubMed Google Scholar

Corresponding author

Correspondence toLi Zhang.

Editor information

Editors and Affiliations

  1. Harbin Institute of Technology, Shenzhen, China

    Haijun Zhang

  2. Nanfang College of Sun Yat-sen University, Guangzhou, China

    Zhi Yang

  3. Hefei University of Technology, Hefei, China

    Zhao Zhang

  4. Chongqing University, Chongqing, China

    Zhou Wu

  5. South China Normal University, Guangzhou, China

    Tianyong Hao

Rights and permissions

Copyright information

© 2021 Springer Nature Singapore Pte Ltd.

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Zheng, X., Zhang, L., Yan, L. (2021). Sample Reduction Using\(\ell _1\)-Norm Twin Bounded Support Vector Machine. In: Zhang, H., Yang, Z., Zhang, Z., Wu, Z., Hao, T. (eds) Neural Computing for Advanced Applications. NCAA 2021. Communications in Computer and Information Science, vol 1449. Springer, Singapore. https://doi.org/10.1007/978-981-16-5188-5_11

Download citation

Publish with us

Access this chapter

Subscribe and save

Springer+ Basic
¥17,985 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
JPY 3498
Price includes VAT (Japan)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
JPY 16015
Price includes VAT (Japan)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
JPY 20019
Price includes VAT (Japan)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide -see info

Tax calculation will be finalised at checkout

Purchases are for personal use only


[8]ページ先頭

©2009-2025 Movatter.jp