• Title/Summary/Keyword: Support Vector Machine

Search Result 1,774, Processing Time 0.028 seconds

Early Software Quality Prediction Using Support Vector Machine (Support Vector Machine을 이용한 초기 소프트웨어 품질 예측)

  • Hong, Euy-Seok
    • Journal of Information Technology Services
    • /
    • v.10 no.2
    • /
    • pp.235-245
    • /
    • 2011
  • Early criticality prediction models that determine whether a design entity is fault-prone or not are becoming more and more important as software development projects are getting larger. Effective predictions can reduce the system development cost and improve software quality by identifying trouble-spots at early phases and proper allocation of effort and resources. Many prediction models have been proposed using statistical and machine learning methods. This paper builds a prediction model using Support Vector Machine(SVM) which is one of the most popular modern classification methods and compares its prediction performance with a well-known prediction model, BackPropagation neural network Model(BPM). SVM is known to generalize well even in high dimensional spaces under small training data conditions. In prediction performance evaluation experiments, dimensionality reduction techniques for data set are not used because the dimension of input data is too small. Experimental results show that the prediction performance of SVM model is slightly better than that of BPM and polynomial kernel function achieves better performance than other SVM kernel functions.

Support Vector Machine based on Stratified Sampling

  • Jun, Sung-Hae
    • International Journal of Fuzzy Logic and Intelligent Systems
    • /
    • v.9 no.2
    • /
    • pp.141-146
    • /
    • 2009
  • Support vector machine is a classification algorithm based on statistical learning theory. It has shown many results with good performances in the data mining fields. But there are some problems in the algorithm. One of the problems is its heavy computing cost. So we have been difficult to use the support vector machine in the dynamic and online systems. To overcome this problem we propose to use stratified sampling of statistical sampling theory. The usage of stratified sampling supports to reduce the size of training data. In our paper, though the size of data is small, the performance accuracy is maintained. We verify our improved performance by experimental results using data sets from UCI machine learning repository.

An analysis of satisfaction index on computer education of university using kernel machine (커널머신을 이용한 대학의 컴퓨터교육 만족도 분석)

  • Pi, Su-Young;Park, Hye-Jung;Ryu, Kyung-Hyun
    • Journal of the Korean Data and Information Science Society
    • /
    • v.22 no.5
    • /
    • pp.921-929
    • /
    • 2011
  • In Information age, the academic liberal art Computer education course set up goals for promoting computer literacy and for developing the ability to cope actively with in Information Society and for improving productivity and competition among nations. In this paper, we analyze on discovering of decisive property and satisfaction index to have a influence on computer education on university students. As a preprocessing method, the proposed method select optimum property using correlation feature selection of machine learning tool based on Java and then we use multiclass least square support vector machine based on statistical learning theory. After applying that compare with multiclass support vector machine and multiclass least square support vector machine, we can see the fact that the proposed method have a excellent result like multiclass support vector machine in analysis of the academic liberal art computer education satisfaction index data.

The Use of MSVM and HMM for Sentence Alignment

  • Fattah, Mohamed Abdel
    • Journal of Information Processing Systems
    • /
    • v.8 no.2
    • /
    • pp.301-314
    • /
    • 2012
  • In this paper, two new approaches to align English-Arabic sentences in bilingual parallel corpora based on the Multi-Class Support Vector Machine (MSVM) and the Hidden Markov Model (HMM) classifiers are presented. A feature vector is extracted from the text pair that is under consideration. This vector contains text features such as length, punctuation score, and cognate score values. A set of manually prepared training data was assigned to train the Multi-Class Support Vector Machine and Hidden Markov Model. Another set of data was used for testing. The results of the MSVM and HMM outperform the results of the length based approach. Moreover these new approaches are valid for any language pairs and are quite flexible since the feature vector may contain less, more, or different features, such as a lexical matching feature and Hanzi characters in Japanese-Chinese texts, than the ones used in the current research.

Semi-supervised regression based on support vector machine

  • Seok, Kyungha
    • Journal of the Korean Data and Information Science Society
    • /
    • v.25 no.2
    • /
    • pp.447-454
    • /
    • 2014
  • In many practical machine learning and data mining applications, unlabeled training examples are readily available but labeled ones are fairly expensive to obtain. Therefore semi-supervised learning algorithms have attracted much attentions. However, previous research mainly focuses on classication problems. In this paper, a semi-supervised regression method based on support vector regression (SVR) formulation that is proposed. The estimator is easily obtained via the dual formulation of the optimization problem. The experimental results with simulated and real data suggest superior performance of the our proposed method compared with standard SVR.

Incremental Support Vector Learning Method for Function Approximation (함수 근사를 위한 점증적 서포트 벡터 학습 방법)

  • 임채환;박주영
    • Proceedings of the IEEK Conference
    • /
    • 2002.06c
    • /
    • pp.135-138
    • /
    • 2002
  • This paper addresses incremental learning method for regression. SVM(support vector machine) is a recently proposed learning method. In general training a support vector machine requires solving a QP (quadratic programing) problem. For very large dataset or incremental dataset, solving QP problems may be inconvenient. So this paper presents an incremental support vector learning method for function approximation problems.

  • PDF

Empirical Choice of the Shape Parameter for Robust Support Vector Machines

  • Pak, Ro-Jin
    • Communications for Statistical Applications and Methods
    • /
    • v.15 no.4
    • /
    • pp.543-549
    • /
    • 2008
  • Inspired by using a robust loss function in the support vector machine regression to control training error and the idea of robust template matching with M-estimator, Chen (2004) applies M-estimator techniques to gaussian radial basis functions and form a new class of robust kernels for the support vector machines. We are specially interested in the shape of the Huber's M-estimator in this context and propose a way to find the shape parameter of the Huber's M-estimating function. For simplicity, only the two-class classification problem is considered.

Semisupervised support vector quantile regression

  • Seok, Kyungha
    • Journal of the Korean Data and Information Science Society
    • /
    • v.26 no.2
    • /
    • pp.517-524
    • /
    • 2015
  • Unlabeled examples are easier and less expensive to be obtained than labeled examples. In this paper semisupervised approach is used to utilize such examples in an effort to enhance the predictive performance of nonlinear quantile regression problems. We propose a semisupervised quantile regression method named semisupervised support vector quantile regression, which is based on support vector machine. A generalized approximate cross validation method is used to choose the hyper-parameters that affect the performance of estimator. The experimental results confirm the successful performance of the proposed S2SVQR.

A Study on Support Vectors of Least Squares Support Vector Machine

  • Seok, Kyungha;Cho, Daehyun
    • Communications for Statistical Applications and Methods
    • /
    • v.10 no.3
    • /
    • pp.873-878
    • /
    • 2003
  • LS-SVM(Least-Squares Support Vector Machine) has been used as a promising method for regression as well as classification. Suykens et al.(2000) used only the magnitude of residuals to obtain SVs(Support Vectors). Suykens' method behaves well for homogeneous model. But in a heteroscedastic model, the method shows a poor behavior. The present paper proposes a new method to get SVs. The proposed method uses the variance of noise as well as the magnitude of residuals to obtain support vectors. Through the simulation study we justified excellence of our proposed method.

Modifying linearly non-separable support vector machine binary classifier to account for the centroid mean vector

  • Mubarak Al-Shukeili;Ronald Wesonga
    • Communications for Statistical Applications and Methods
    • /
    • v.30 no.3
    • /
    • pp.245-258
    • /
    • 2023
  • This study proposes a modification to the objective function of the support vector machine for the linearly non-separable case of a binary classifier yi ∈ {-1, 1}. The modification takes into account the position of each data item xi from its corresponding class centroid. The resulting optimization function involves the centroid mean vector, and the spread of data besides the support vectors, which should be minimized by the choice of hyper-plane β. Theoretical assumptions have been tested to derive an optimal separable hyperplane that yields the minimal misclassification rate. The proposed method has been evaluated using simulation studies and real-life COVID-19 patient outcome hospitalization data. Results show that the proposed method performs better than the classical linear SVM classifier as the sample size increases and is preferred in the presence of correlations among predictors as well as among extreme values.