• Title/Summary/Keyword: convolution-based

Search Result 584, Processing Time 0.024 seconds

A Proposal of Shuffle Graph Convolutional Network for Skeleton-based Action Recognition

  • Jang, Sungjun;Bae, Han Byeol;Lee, HeanSung;Lee, Sangyoun
    • The Journal of Korea Institute of Information, Electronics, and Communication Technology
    • /
    • v.14 no.4
    • /
    • pp.314-322
    • /
    • 2021
  • Skeleton-based action recognition has attracted considerable attention in human action recognition. Recent methods for skeleton-based action recognition employ spatiotemporal graph convolutional networks (GCNs) and have remarkable performance. However, most of them have heavy computational complexity for robust action recognition. To solve this problem, we propose a shuffle graph convolutional network (SGCN) which is a lightweight graph convolutional network using pointwise group convolution rather than pointwise convolution to reduce computational cost. Our SGCN is composed of spatial and temporal GCN. The spatial shuffle GCN contains pointwise group convolution and part shuffle module which enhances local and global information between correlated joints. In addition, the temporal shuffle GCN contains depthwise convolution to maintain a large receptive field. Our model achieves comparable performance with lowest computational cost and exceeds the performance of baseline at 0.3% and 1.2% on NTU RGB+D and NTU RGB+D 120 datasets, respectively.

An Implementation of a Convolutional Accelerator based on a GPGPU for a Deep Learning (Deep Learning을 위한 GPGPU 기반 Convolution 가속기 구현)

  • Jeon, Hee-Kyeong;Lee, Kwang-yeob;Kim, Chi-yong
    • Journal of IKEEE
    • /
    • v.20 no.3
    • /
    • pp.303-306
    • /
    • 2016
  • In this paper, we propose a method to accelerate convolutional neural network by utilizing a GPGPU. Convolutional neural network is a sort of the neural network learning features of images. Convolutional neural network is suitable for the image processing required to learn a lot of data such as images. The convolutional layer of the conventional CNN required a large number of multiplications and it is difficult to operate in the real-time on the embedded environment. In this paper, we reduce the number of multiplications through Winograd convolution operation and perform parallel processing of the convolution by utilizing SIMT-based GPGPU. The experiment was conducted using ModelSim and TestDrive, and the experimental results showed that the processing time was improved by about 17%, compared to the conventional convolution.

A Reconsideration of the Causality Requirement in Proving the z-Transform of a Discrete Convolution Sum (이산 Convolution 적산의 z변환의 증명을 위한 인과성의 필요에 대한 재고)

  • Chung Tae-Sang;Lee Jae Seok
    • The Transactions of the Korean Institute of Electrical Engineers D
    • /
    • v.52 no.1
    • /
    • pp.51-54
    • /
    • 2003
  • The z-transform method is a basic mathematical tool in analyzing and designing digital signal processing systems for discrete input and output signals. There are may cases where the output signal is in the form of a discrete convolution sum of an input function and a designed digital processing algorithm function. It is well known that the z-transform of the convolution sum becomes the product of the two z-transforms of the input function and the digital processing function, whose proofs require the causality of the digital signal processing function in the almost all the available references. However, not all of the convolution sum functions are based on the causality. Many digital signal processing systems such as image processing system may depend not on the time information but on the spatial information, which has nothing to do with causality requirement. Thus, the application of the causality-based z-transform theorem on the convolution sum cannot be used without difficulty in this case. This paper proves the z-transform theorem on the discrete convolution sum without causality requirement, and make it possible for the theorem to be used in analysis and desing for any cases.

Algorithm-based fault tolerant vector convolution on array processor (배열프로세서상에서 알고리즘 기반 결함허용 벡터 컨버루션)

  • 송기용
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.23 no.8
    • /
    • pp.1977-1983
    • /
    • 1998
  • An algorithm-based fault tolerant scheme for the vector convolution is proposed employing the positive and negative checksum vectors that are defined in this paper based on the encoder vector. The proposed scheme is implemented on the aray processor, and then the amount of redundancy is examined thrugh the complexity analysis.

  • PDF

A Study on the Classification of Surface Defect Based on Deep Convolution Network and Transfer-learning (신경망과 전이학습 기반 표면 결함 분류에 관한 연구)

  • Kim, Sung Joo;Kim, Gyung Bum
    • Journal of the Semiconductor & Display Technology
    • /
    • v.20 no.1
    • /
    • pp.64-69
    • /
    • 2021
  • In this paper, a method for improving the defect classification performance in low contrast, ununiformity and featureless steel plate surfaces has been studied based on deep convolution neural network and transfer-learning neural network. The steel plate surface images have low contrast, ununiformity, and featureless, so that the contrast between defect and defect-free regions are not discriminated. These characteristics make it difficult to extract the feature of the surface defect image. A classifier based on a deep convolution neural network is constructed to extract features automatically for effective classification of images with these characteristics. As results of the experiment, AlexNet-based transfer-learning classifier showed excellent classification performance of 99.43% with less than 160 seconds of training time. The proposed classification system showed excellent classification performance for low contrast, ununiformity, and featureless surface images.

Visualization of Convolution Operation Using Scalable Vector Graphics (SVG를 이용한 컨벌루션 연산의 시각화)

  • Kim, Yeong-Mi;Kang, Eui-Sung
    • The Journal of Korean Association of Computer Education
    • /
    • v.10 no.1
    • /
    • pp.97-105
    • /
    • 2007
  • In this paper, visualization of convolution operation is presented, which is implemented by scalable vector graphics (SVG). Convolution operation is one of the basic essential concepts in the area of signal and image processing. However, it is difficult for students to intuitively understand the operation of convolution since it is mainly based on mathematical representation. We present the visualization of convolution operation and its applications which are implemented by SVG. The effects of the proposed approach have been analyzed by interviews. It has been seen that the proposed visualization of convolution operation could be effectively applied to learn the convolution operation and its applications.

  • PDF

A Video Expression Recognition Method Based on Multi-mode Convolution Neural Network and Multiplicative Feature Fusion

  • Ren, Qun
    • Journal of Information Processing Systems
    • /
    • v.17 no.3
    • /
    • pp.556-570
    • /
    • 2021
  • The existing video expression recognition methods mainly focus on the spatial feature extraction of video expression images, but tend to ignore the dynamic features of video sequences. To solve this problem, a multi-mode convolution neural network method is proposed to effectively improve the performance of facial expression recognition in video. Firstly, OpenFace 2.0 is used to detect face images in video, and two deep convolution neural networks are used to extract spatiotemporal expression features. Furthermore, spatial convolution neural network is used to extract the spatial information features of each static expression image, and the dynamic information feature is extracted from the optical flow information of multiple expression images based on temporal convolution neural network. Then, the spatiotemporal features learned by the two deep convolution neural networks are fused by multiplication. Finally, the fused features are input into support vector machine to realize the facial expression classification. Experimental results show that the recognition accuracy of the proposed method can reach 64.57% and 60.89%, respectively on RML and Baum-ls datasets. It is better than that of other contrast methods.

Multi-site based earthquake event classification using graph convolution networks (그래프 합성곱 신경망을 이용한 다중 관측소 기반 지진 이벤트 분류)

  • Kim, Gwantae;Ku, Bonhwa;Ko, Hanseok
    • The Journal of the Acoustical Society of Korea
    • /
    • v.39 no.6
    • /
    • pp.615-621
    • /
    • 2020
  • In this paper, we propose a multi-site based earthquake event classification method using graph convolution networks. In the traditional earthquake event classification methods using deep learning, they used single-site observation to estimate seismic event class. However, to achieve robust and accurate earthquake event classification on the seismic observation network, the method using the information from the multi-site observations is needed, instead of using only single-site data. Firstly, our proposed model employs convolution neural networks to extract informative embedding features from the single-site observation. Secondly, graph convolution networks are used to integrate the features from several stations. To evaluate our model, we explore the model structure and the number of stations for ablation study. Finally, our multi-site based model outperforms up to 10 % accuracy and event recall rate compared to single-site based model.

Human Action Recognition Based on 3D Convolutional Neural Network from Hybrid Feature

  • Wu, Tingting;Lee, Eung-Joo
    • Journal of Korea Multimedia Society
    • /
    • v.22 no.12
    • /
    • pp.1457-1465
    • /
    • 2019
  • 3D convolution is to stack multiple consecutive frames to form a cube, and then apply the 3D convolution kernel in the cube. In this structure, each feature map of the convolutional layer is connected to multiple adjacent sequential frames in the previous layer, thus capturing the motion information. However, due to the changes of pedestrian posture, motion and position, the convolution at the same place is inappropriate, and when the 3D convolution kernel is convoluted in the time domain, only time domain features of three consecutive frames can be extracted, which is not a good enough to get action information. This paper proposes an action recognition method based on feature fusion of 3D convolutional neural network. Based on the VGG16 network model, sending a pre-acquired optical flow image for learning, then get the time domain features, and then the feature of the time domain is extracted from the features extracted by the 3D convolutional neural network. Finally, the behavior classification is done by the SVM classifier.

Network Traffic Classification Based on Deep Learning

  • Li, Junwei;Pan, Zhisong
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.14 no.11
    • /
    • pp.4246-4267
    • /
    • 2020
  • As the network goes deep into all aspects of people's lives, the number and the complexity of network traffic is increasing, and traffic classification becomes more and more important. How to classify them effectively is an important prerequisite for network management and planning, and ensuring network security. With the continuous development of deep learning, more and more traffic classification begins to use it as the main method, which achieves better results than traditional classification methods. In this paper, we provide a comprehensive review of network traffic classification based on deep learning. Firstly, we introduce the research background and progress of network traffic classification. Then, we summarize and compare traffic classification based on deep learning such as stack autoencoder, one-dimensional convolution neural network, two-dimensional convolution neural network, three-dimensional convolution neural network, long short-term memory network and Deep Belief Networks. In addition, we compare traffic classification based on deep learning with other methods such as based on port number, deep packets detection and machine learning. Finally, the future research directions of network traffic classification based on deep learning are prospected.