DOI QR코드

DOI QR Code

A transductive least squares support vector machine with the difference convex algorithm

  • Received : 2014.02.14
  • Accepted : 2014.03.07
  • Published : 2014.03.31

Abstract

Unlabeled examples are easier and less expensive to obtain than labeled examples. Semisupervised approaches are used to utilize such examples in an eort to boost the predictive performance. This paper proposes a novel semisupervised classication method named transductive least squares support vector machine (TLS-SVM), which is based on the least squares support vector machine. The proposed method utilizes the dierence convex algorithm to derive nonconvex minimization solutions for the TLS-SVM. A generalized cross validation method is also developed to choose the hyperparameters that aect the performance of the TLS-SVM. The experimental results conrm the successful performance of the proposed TLS-SVM.

Keywords

References

  1. Adankon, M., Cheriet, M. and Biem, A. (2009). Semisupervised least squares support vector machine. IEEE Tranactions on Neural Networks, 20, 1858-1870. https://doi.org/10.1109/TNN.2009.2031143
  2. An, L. and Tao, P. (1997). Solving a class of linearly constrained indefinite quadratic problems by D.C. algorithms. Journal of Global Optimization, 11, 253-285. https://doi.org/10.1023/A:1008288411710
  3. Astotino, A. and Fuduli, A. (2005). Nonsmooth optimization technique for semisupervised classificatin. IEEE Transactions on Pattern Analysis and Machine Intelligence, 29, 2135-2142.
  4. Blum, A. and Mitchell, T. (1998). Combining labeled and unlabeled data with co-training. Proceedings of the 11th Annual Conference on Computational Learning Theory, Madison, Wisconsin, 92-100.
  5. Blum, A. and Chawla, S. (2001). Learning from labeled and unlabeled data using graph minicuts. In Proceedings of ICML 2001, 19-26.
  6. Chapelle, O. and Zien, A. (2005). Semi-supervised classification by low density separarion. In Proceedings of tht 10th International Workshop on Artificial Intelligence and Statistics, 57-64.
  7. Chapelle, O., Sindhwani, V. and Keerthi, S. (2008). Optimization techniques for semisupervised support vector machines. Journal of Machine Learning Research, 9, 203-233.
  8. Chen, Y., Wang, G. and Dong, S. (2002). Learning with progressive transductive support vector machine. Proceedings of International Conference on Data Mining, Maebashi City, Japan, 67-74.
  9. Craven, P. and Wahba, G. (1979). Smothing noisy data with spline functions. Numerical Mathematics, 31, 377-403.
  10. Joachims, T. (1999). Transductive inference for text classification using support vecter machines. Proceedings of International conference on Machine Learning, Bled, Slovenia, 200-209.
  11. Mercer, J. (1909). Functions of positive and negative type and their connection with theory of integral equations. Philosophical Transactions of Royal Society A, 415-446.
  12. Seok, K. (2012). Study on semi-supervised local constant regression estimation. Journal of the Korean Data & Information Science Society, 23, 579-585. https://doi.org/10.7465/jkdi.2012.23.3.579
  13. Seok, K. (2013). A study on semi-supervised kernel ridge regression estimation. Journal of the Korean Data & Information Science Society, 24, 341-353. https://doi.org/10.7465/jkdi.2013.24.2.341
  14. Suykens, J. A. K. (2000). Least squares support vector machine for classication and nonlinear modeling. Neural Network World, Special issue on PASE 2000, 10, 29-48.
  15. Suykens, J. A. K. and Vanderwalle, J. (1999). Least square support vector machine classifier, Neural Processing Letters, 9, 293-300. https://doi.org/10.1023/A:1018628609742
  16. Suykens, J. A. K., Van Gestel, T., De Brabanter, J., De Moor, B. and Vanthienen, J. (2002). Least squares support vector machines, World Scientific, Singapore.
  17. Vapnik, V. N. (1995). The nature of statistical learning theory, Springer, New York.
  18. Vapnik, V. N. (1998). Statistical learning theory, John Wiley, New York.
  19. Wang, J., Shen, X. and Pan, W. (2007). On transductive support vector machine. Contemporary Mathematics, 43, 7-19.
  20. Xu, S., An. X., Qiao, X., Zhu, L. and Li, L. (2011) Semisupervised least squares support vector regression machines. Journal of Information & Computational Science, 8, 885-892.
  21. Xu, Z., King, I. and Lyu, M. R. (2010). More than semi-supervised learning, LAP LAMBERT Academic Publishing, Germany.
  22. Zhang, R., Wang, W., Ma, Y. and Men, C. (2009). Least square transduction support vector machine. Neural Processing Letters, 29, 133-142. https://doi.org/10.1007/s11063-009-9099-z
  23. Zhu, X. and Goldberg, A. (2009). Introduction to semi-supervised learning, Morgan & Claypool Publishers, CA.

Cited by

  1. Geographically weighted kernel logistic regression for small area proportion estimation vol.27, pp.2, 2016, https://doi.org/10.7465/jkdi.2016.27.2.531
  2. Estimation and variable selection in censored regression model with smoothly clipped absolute deviation penalty vol.27, pp.6, 2016, https://doi.org/10.7465/jkdi.2016.27.6.1653
  3. Multioutput LS-SVR based residual MCUSUM control chart for autocorrelated process vol.27, pp.2, 2016, https://doi.org/10.7465/jkdi.2016.27.2.523
  4. Deep LS-SVM for regression vol.27, pp.3, 2016, https://doi.org/10.7465/jkdi.2016.27.3.827
  5. SVQR with asymmetric quadratic loss function vol.26, pp.6, 2015, https://doi.org/10.7465/jkdi.2015.26.6.1537
  6. LS-SVM for large data sets vol.27, pp.2, 2016, https://doi.org/10.7465/jkdi.2016.27.2.549
  7. Adaptive ridge procedure for L0-penalized weighted support vector machines vol.28, pp.6, 2014, https://doi.org/10.7465/jkdi.2017.28.6.1271