• Title/Summary/Keyword: Image Summarization Algorithm

Search Result 3, Processing Time 0.018 seconds

The Image Summarization Algorithm for Reviewing the Virtual Reality Experience (가상현실 경험을 복습시켜주는 사진 정리 알고리즘)

  • Kwak, Eun-Joo;Cho, Yong-Joo;Cho, Hyun-Sang;Park, Kyoung-Shin
    • The KIPS Transactions:PartB
    • /
    • v.15B no.3
    • /
    • pp.211-218
    • /
    • 2008
  • In this paper, we proposed a new image summarization algorithm designed for automatically summarizing user's snapshot photos taken in a virtual environment based on user's context information and educational contents, and then presenting a summarized photos shortly after user's virtual reality experience. While other image summarization algorithms used date, location, and keyword to effectively summarize a large amount of photos, this algorithm is intended to improve users' memory retention by recalling their interests and important educational contents. This paper first describes some criteria of extracting the meaningful images to improve learning effects and the identification rate calculations, followed by the system architecture that integrates the virtual environment and the viewer interface. It will also discuss a user study to model the algorithm's optimal identification rate and then future research directions.

An Efficient Face Region Detection for Content-based Video Summarization (내용기반 비디오 요약을 위한 효율적인 얼굴 객체 검출)

  • Kim Jong-Sung;Lee Sun-Ta;Baek Joong-Hwan
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.30 no.7C
    • /
    • pp.675-686
    • /
    • 2005
  • In this paper, we propose an efficient face region detection technique for the content-based video summarization. To segment video, shot changes are detected from a video sequence and key frames are selected from the shots. We select one frame that has the least difference between neighboring frames in each shot. The proposed face detection algorithm detects face region from selected key frames. And then, we provide user with summarized frames included face region that has an important meaning in dramas or movies. Using Bayes classification rule and statistical characteristic of the skin pixels, face regions are detected in the frames. After skin detection, we adopt the projection method to segment an image(frame) into face region and non-face region. The segmented regions are candidates of the face object and they include many false detected regions. So, we design a classifier to minimize false lesion using CART. From SGLD matrices, we extract the textual feature values such as Inertial, Inverse Difference, and Correlation. As a result of our experiment, proposed face detection algorithm shows a good performance for the key frames with a complex and variant background. And our system provides key frames included the face region for user as video summarized information.

Effective Hand Gesture Recognition by Key Frame Selection and 3D Neural Network

  • Hoang, Nguyen Ngoc;Lee, Guee-Sang;Kim, Soo-Hyung;Yang, Hyung-Jeong
    • Smart Media Journal
    • /
    • v.9 no.1
    • /
    • pp.23-29
    • /
    • 2020
  • This paper presents an approach for dynamic hand gesture recognition by using algorithm based on 3D Convolutional Neural Network (3D_CNN), which is later extended to 3D Residual Networks (3D_ResNet), and the neural network based key frame selection. Typically, 3D deep neural network is used to classify gestures from the input of image frames, randomly sampled from a video data. In this work, to improve the classification performance, we employ key frames which represent the overall video, as the input of the classification network. The key frames are extracted by SegNet instead of conventional clustering algorithms for video summarization (VSUMM) which require heavy computation. By using a deep neural network, key frame selection can be performed in a real-time system. Experiments are conducted using 3D convolutional kernels such as 3D_CNN, Inflated 3D_CNN (I3D) and 3D_ResNet for gesture classification. Our algorithm achieved up to 97.8% of classification accuracy on the Cambridge gesture dataset. The experimental results show that the proposed approach is efficient and outperforms existing methods.