DOI QR코드

DOI QR Code

A Co-Evolutionary Computing for Statistical Learning Theory

  • Jun Sung-Hae (Department of Statistics, Cheongju University)
  • Published : 2005.12.01

Abstract

Learning and evolving are two basics for data mining. As compared with classical learning theory based on objective function with minimizing training errors, the recently evolutionary computing has had an efficient approach for constructing optimal model without the minimizing training errors. The global search of evolutionary computing in solution space can settle the local optima problems of learning models. In this research, combining co-evolving algorithm into statistical learning theory, we propose an co-evolutionary computing for statistical learning theory for overcoming local optima problems of statistical learning theory. We apply proposed model to classification and prediction problems of the learning. In the experimental results, we verify the improved performance of our model using the data sets from UCI machine learning repository and KDD Cup 2000.

Keywords

References

  1. G. Casella, R. L. Berger, Statistical Inference, Duxbury Press, 1990
  2. V. Cherkassky, F. Mulier, Learning From Data Concepts, Theory, and Methods, John Wiley & Sons, 1998
  3. C. Cortes, V. Vapnik, 'Support Vector Networks,' Machine Learning, vol. 20, pp. 273-297, 1995
  4. A. E. Eiben, J. E., Smith, Introduction to Evolutionary Computing, Springer, 2003
  5. R. Fletcher, Practical Methods of Optimization, John Wiley & Sons, 1989
  6. J. Han, M. Kamber, Data Mining: Concepts and Techniques, Morgan Kaufmann Publishers, 2001
  7. W. D. Hill, 'Co-evolving parasites improve simulated evolution as an optimization procedure,' Physica, vol. 42, pp. 228-234, 1992
  8. S. H. Jun, M. Gen, 'Evolutionary Computation for Statistical Learning Theory', Proceedings of Joint 2nd International Conference on Soft Computing and Intelligent Systems and 5th International Symposium on Advanced Intelligent Systems, 2004
  9. S. H. Jun, 'Web Usage Mining Using Support Vector Machine', Lecture Note in Computer Science, vol. 3512, pp. 349-356, Springer, 2005 https://doi.org/10.1007/11494669_43
  10. R. J. A. Little, D. B. Rubin, Statistical Analysis with Missing Data, John Wiley & Sons, 2002
  11. T. M. Mitchell, Machine Learning, WCB McGraw Hill, 1997
  12. T. M. Mitchell, An introduction to Genetic Algorithms, MIT Press, 1998
  13. D. J. Montana, L. Davis, 'Training Feedforward Neural Network Using Genetic Algorithm,' Proceedings of the International Conference on Artificial Intelligence, pp. 763-767, 1989
  14. R. H. Myers, Classical and Modern Regression with applications, Duxbury, 1990
  15. A. S. Pandya, R. B. Macy, Pattern Recognition with Neural Networks in C++, CRC Press, 1995
  16. M. J. Park, Competitive Co-Evolving Neural Network: Host and Parasites, Master's thesis, Sogang University, 2002
  17. S. H. Ross. Introductory Statistics, McGraw-Hill, 1996
  18. A. J. Smola, Regression estimation with support vector learning machines, Master's thesis, Technische University, 1996
  19. V. N. Vapnik, The Nature of Statistical Learning Theory, Springer, 1995
  20. V. N. Vapnik, Statistical Learning Theory, John Wiley & Sons, 1998
  21. V. N. Vapnik, S. Golowich, A. Smola, 'Support vector method for function approximation, regres-sion estimation, and signal processing,' Advanced in Neural Information Processing Systems, Cambridge, MA, 1997
  22. X, Yao, 'A review of evolutionary artificial neural networks,' International Journal Intelligent System, vol. 8, no. 4, pp. 539-567, 1993 https://doi.org/10.1002/int.4550080406
  23. X. Yao, 'Evolving Artificial Neural Networks,' Proceeding of the IEEE, vol. 87, issue 9, pp. 1423-1447, 1999 https://doi.org/10.1109/5.784219
  24. KDD Cup 2000, www.ecn.purdue.edu /KDDCUP
  25. UCI M. L. Repository, www.ics.uci.edu/~mleam