• Title/Summary/Keyword: K-nearest neighbors

Search Result 192, Processing Time 0.033 seconds

On the Use of Modified Adaptive Nearest Neighbors for Classification (수정된 적응 최근접 방법을 활용한 판별분류방법에 대한 연구)

  • Maeng, Jin-Woo;Bang, Sung-Wan;Jhun, Myoung-Shic
    • The Korean Journal of Applied Statistics
    • /
    • v.23 no.6
    • /
    • pp.1093-1102
    • /
    • 2010
  • Even though the k-Nearest Neighbors Classification(KNNC) is one of the popular non-parametric classification methods, it does not consider the local features and class information for each observation. In order to overcome such limitations, several methods have been developed such as Adaptive Nearest Neighbors Classification(ANNC) and Modified k-Nearest Neighbors Classification(MKNNC). In this paper, we propose the Modified Adaptive Nearest Neighbors Classification(MANNC) that employs the advantages of both the ANNC and MKNNC. Through a real data analysis and a simulation study, we show that the proposed MANNC outperforms other methods in terms of classification accuracy.

On the use of weighted adaptive nearest neighbors for missing value imputation (가중 적응 최근접 이웃을 이용한 결측치 대치)

  • Yum, Yunjin;Kim, Dongjae
    • The Korean Journal of Applied Statistics
    • /
    • v.31 no.4
    • /
    • pp.507-516
    • /
    • 2018
  • Widely used among the various single imputation methods is k-nearest neighbors (KNN) imputation due to its robustness even when a parametric model such as multivariate normality is not satisfied. We propose a weighted adaptive nearest neighbors imputation method that combines the adaptive nearest neighbors imputation method that accounts for the local features of the data in the KNN imputation method and weighted k-nearest neighbors method that are less sensitive to extreme value or outlier among k-nearest neighbors. We conducted a Monte Carlo simulation study to compare the performance of the proposed imputation method with previous imputation methods.

Adaptive Nearest Neighbors를 활용한 결측치 대치

  • 전명식;정형철
    • Proceedings of the Korean Statistical Society Conference
    • /
    • 2004.11a
    • /
    • pp.185-190
    • /
    • 2004
  • 비모수적 결측치 대치 방법으로 널리 사용되는 k-nearest neighbors(KNN) 방법은 자료의 국소적(local) 특징을 고려하지 않고 전체 자료에 대해 균일한 이웃의 개수 k를 사용하는 단점이 있다. 본 연구에서는 KNN의 대안으로 자료의 국소적 특징을 고려하는 adaptive nearest neighbors(ANN) 방법을 제안하였다. 나아가 microarray 자료의 경우에 대하여 결측치 대치를 통해 KNN과 ANN의 성능을 비교하였다.

  • PDF

Adaptive Nearest Neighbors for Classification (Adaptive Nearest Neighbors를 활용한 판별분류방법)

  • Jhun, Myoung-Shic;Choi, In-Kyung
    • The Korean Journal of Applied Statistics
    • /
    • v.22 no.3
    • /
    • pp.479-488
    • /
    • 2009
  • The ${\kappa}$-Nearest Neighbors Classification(KNNC) is a popular non-parametric classification method which assigns a fixed number ${\kappa}$ of neighbors to every observation without consideration of the local feature of the each observation. In this paper, we propose an Adaptive Nearest Neighbors Classification(ANNC) as an alternative to KNNC. The proposed ANNC method adapts the number of neighbors according to the local feature of the observation such as density of data. To verify characteristics of ANNC, we compare the number of misclassified observation with KNNC by Monte Carlo study and confirm the potential performance of ANNC method.

On the Use of Sequential Adaptive Nearest Neighbors for Missing Value Imputation (순차 적응 최근접 이웃을 활용한 결측값 대치법)

  • Park, So-Hyun;Bang, Sung-Wan;Jhun, Myoung-Shic
    • The Korean Journal of Applied Statistics
    • /
    • v.24 no.6
    • /
    • pp.1249-1257
    • /
    • 2011
  • In this paper, we propose a Sequential Adaptive Nearest Neighbor(SANN) imputation method that combines the Adaptive Nearest Neighbor(ANN) method and the Sequential k-Nearest Neighbor(SKNN) method. When choosing the nearest neighbors of missing observations, the proposed SANN method takes the local feature of the missing observations into account as well as reutilizes the imputed observations in a sequential manner. By using a Monte Carlo study and a real data example, we demonstrate the characteristics of the SANN method and its potential performance.

On the Use of Weighted k-Nearest Neighbors for Missing Value Imputation (Weighted k-Nearest Neighbors를 이용한 결측치 대치)

  • Lim, Chanhui;Kim, Dongjae
    • The Korean Journal of Applied Statistics
    • /
    • v.28 no.1
    • /
    • pp.23-31
    • /
    • 2015
  • A conventional missing value problem in the statistical analysis k-Nearest Neighbor(KNN) method are used for a simple imputation method. When one of the k-nearest neighbors is an extreme value or outlier, the KNN method can create a bias. In this paper, we propose a Weighted k-Nearest Neighbors(WKNN) imputation method that can supplement KNN's faults. A Monte-Carlo simulation study is also adapted to compare the WKNN method and KNN method using real data set.

Robust Similarity Measure for Spectral Clustering Based on Shared Neighbors

  • Ye, Xiucai;Sakurai, Tetsuya
    • ETRI Journal
    • /
    • v.38 no.3
    • /
    • pp.540-550
    • /
    • 2016
  • Spectral clustering is a powerful tool for exploratory data analysis. Many existing spectral clustering algorithms typically measure the similarity by using a Gaussian kernel function or an undirected k-nearest neighbor (kNN) graph, which cannot reveal the real clusters when the data are not well separated. In this paper, to improve the spectral clustering, we consider a robust similarity measure based on the shared nearest neighbors in a directed kNN graph. We propose two novel algorithms for spectral clustering: one based on the number of shared nearest neighbors, and one based on their closeness. The proposed algorithms are able to explore the underlying similarity relationships between data points, and are robust to datasets that are not well separated. Moreover, the proposed algorithms have only one parameter, k. We evaluated the proposed algorithms using synthetic and real-world datasets. The experimental results demonstrate that the proposed algorithms not only achieve a good level of performance, they also outperform the traditional spectral clustering algorithms.

Missing values imputation for time course gene expression data using the pattern consistency index adaptive nearest neighbors (시간경로 유전자 발현자료에서 패턴일치지수와 적응 최근접 이웃을 활용한 결측값 대치법)

  • Shin, Heyseo;Kim, Dongjae
    • The Korean Journal of Applied Statistics
    • /
    • v.33 no.3
    • /
    • pp.269-280
    • /
    • 2020
  • Time course gene expression data is a large amount of data observed over time in microarray experiments. This data can also simultaneously identify the level of gene expression. However, the experiment process is complex, resulting in frequent missing values due to various causes. In this paper, we propose a pattern consistency index adaptive nearest neighbors as a method of missing value imputation. This method combines the adaptive nearest neighbors (ANN) method that reflects local characteristics and the pattern consistency index that considers consistent degree for gene expression between observations over time points. We conducted a Monte Carlo simulation study to evaluate the usefulness of proposed the pattern consistency index adaptive nearest neighbors (PANN) method for two yeast time course data.

Design of Fuzzy k-Nearest Neighbors Classifiers based on Feature Extraction by using Stacked Autoencoder (Stacked Autoencoder를 이용한 특징 추출 기반 Fuzzy k-Nearest Neighbors 패턴 분류기 설계)

  • Rho, Suck-Bum;Oh, Sung-Kwun
    • The Transactions of The Korean Institute of Electrical Engineers
    • /
    • v.64 no.1
    • /
    • pp.113-120
    • /
    • 2015
  • In this paper, we propose a feature extraction method using the stacked autoencoders which consist of restricted Boltzmann machines. The stacked autoencoders is a sort of deep networks. Restricted Boltzmann machines (RBMs) are probabilistic graphical models that can be interpreted as stochastic neural networks. In terms of pattern classification problem, the feature extraction is a key issue. We use the stacked autoencoders networks to extract new features which have a good influence on the improvement of the classification performance. After feature extraction, fuzzy k-nearest neighbors algorithm is used for a classifier which classifies the new extracted data set. To evaluate the classification ability of the proposed pattern classifier, we make some experiments with several machine learning data sets.

Magnetic Property of Oxide with the Perovskite Structure, $A_2Fe(III)BO_6$ (A = Ca, Sr, Ba and B = Sb, Bi)

  • 이성옥;조태연;변송호
    • Bulletin of the Korean Chemical Society
    • /
    • v.18 no.1
    • /
    • pp.91-97
    • /
    • 1997
  • In the course of magnetic study on several perovskite-type oxides, A2Fe(Ⅲ)BO6 (A = Ca, Sr, Ba and B = Sb, Bi), we have observed a strong irreversibility in their dc-magnetizations. When the structural data and the Mossbauer spectra are considered, such an irreversibility is to be associated with some competitions between the nearest-neighbors (nn) and the next-nearest-neighbors (nnn) in their magnetic sublattices. Particularly, the Mossbauer spectra indicate that Sr2FeBiO6 of cubic perovskite-structure is apparently well ordered crystalline compound. Nontheless this antiferromagnet shows a magnetic property which resembles that of a spin-glass. The strong history dependence is observed below 91 K and the irreversible magnetic behavior is also observed from the measurement of hysteresis loops at 10 K after zero-field-cooled (zfc) and field-cooled (fc) processes. Considering the nn and the nnn superexchanges of almost same order in ordered perovskite, it is proposed that there exists a competition and cancellation of antiferromagnetic and ferromagnetic superexchange between the nearest-neighbors and the next-nearest-neighbors, thus introducing a certain degree of frustration.