• Title/Summary/Keyword: Keyframe analysis

Search Result 11, Processing Time 0.024 seconds

Improved Quality Keyframe Selection Method for HD Video

  • Yang, Hyeon Seok;Lee, Jong Min;Jeong, Woojin;Kim, Seung-Hee;Kim, Sun-Joong;Moon, Young Shik
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.13 no.6
    • /
    • pp.3074-3091
    • /
    • 2019
  • With the widespread use of the Internet, services for providing large-capacity multimedia data such as video-on-demand (VOD) services and video uploading sites have greatly increased. VOD service providers want to be able to provide users with high-quality keyframes of high quality videos within a few minutes after the broadcast ends. However, existing keyframe extraction tends to select keyframes whose quality as a keyframe is insufficiently considered, and it takes a long computation time because it does not consider an HD class image. In this paper, we propose a keyframe selection method that flexibly applies multiple keyframe quality metrics and improves the computation time. The main procedure is as follows. After shot boundary detection is performed, the first frames are extracted as initial keyframes. The user sets evaluation metrics and priorities by considering the genre and attributes of the video. According to the evaluation metrics and the priority, the low-quality keyframe is selected as a replacement target. The replacement target keyframe is replaced with a high-quality frame in the shot. The proposed method was subjectively evaluated by 23 votes. Approximately 45% of the replaced keyframes were improved and about 18% of the replaced keyframes were adversely affected. Also, it took about 10 minutes to complete the summary of one hour video, which resulted in a reduction of more than 44.5% of the execution time.

Comparison of big data image analysis techniques for user curation (사용자 큐레이션을 위한 빅데이터 영상 분석 기법 비교)

  • Lee, Hyoun-Sup;Kim, Jin-Deog
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2021.05a
    • /
    • pp.563-565
    • /
    • 2021
  • The most important feature of the recently increasing content providing service is that the amount of content increase over time is very large. Accordingly, the importance of user curation is increasing, and various techniques are used to implement it. In this paper, among the techniques for video recommendation, the analysis technique using voice data and subtitles and the video comparison technique based on keyframe extraction are compared with the results of implementing and applying the video content of real big data. In addition, through the comparison result, a video content environment to which each analysis technique can be applied is proposed.

  • PDF

Finding focused key frames of a given meaning on video data (영상의 특정 의미를 반영하는 Key Frame의 추출 방법)

  • Ha, Jong-Woo;Noh, Jung-Dam;Yoon, Soungwoong;Kim, Min-Soo;Ahn, Chang-Won
    • Proceedings of the Korean Society of Computer Information Conference
    • /
    • 2022.01a
    • /
    • pp.85-88
    • /
    • 2022
  • 영상을 구성하는 프레임 중에 키프레임은 일반적으로 영상 정보를 효과적으로 요약하거나 용이한 분석을 위해 선정된다. 화상이 가진 의미는 인물/사물 등의 객체탐지를 통해 추출되는데, 기존의 키프레임 관련 연구는 영상이 가지는 의미를 반영하는 키프레임을 찾아내기 어렵다. 본 논문에서는 영상이 가지는 특정 의미가 있다고 할 때 이를 반영하는 키프레임을 효과적으로 추출하는 방법을 실험적으로 탐구하였다. 구체적으로 영상을 통할하는 의미를 피로라고 가정하고 영상의 졸음 인식 관련 연구에 사용되는 DDD 데이터셋을 이용하여 효과적인 키프레임 추출 기법을 적용해 보았으며, 실험 결과 졸음이라는 특정 정보에 대한 해석을 도울 수 있는 의미 있는 요약을 제공하는 키프레임들을 효과적으로 추출하는 분석 기법을 찾아낼 수 있었다.

  • PDF

The Scene Analysis and Keyframe Extraction for Content-Based Indexing on Compressed Image Sequence (압축된 영상 시퀀스에서 내용 기반 색인을 위한 장면 분석 및 키 프레임 추출)

  • 오상헌;김상렬;김주도;이근영
    • Proceedings of the IEEK Conference
    • /
    • 1999.11a
    • /
    • pp.605-608
    • /
    • 1999
  • In this paper, we proposed several scene analysis algorithms. These algorithms using image difference and histogram operate on the sequence of DC coefficient which is extracted from Motion JPEG or MPEG without full-frame decompression. Since DC sequence has the most information of full frame while it has reduced data. Experimental results show less than 1/64 of full frame analysing complexity and exactly analyze scene changes and extract key frames.

  • PDF

Client-driven Animated Keyframe Generation System Using Music Analysis (음악 분석을 이용한 클라이언트 중심의 키프레임 생성 시스템)

  • Mujtaba, Ghulam;Kim, Seondae;Park, Eunsoo;Kim, Seunghwan;Ryu, Jaesung;Ryu, Eun-Seok
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 2019.06a
    • /
    • pp.173-175
    • /
    • 2019
  • Animated images formats such as WebP are highly portable graphics formats that are being used everywhere on the Internet. Despite their small sizes and duration, WebP image previews the video without watching the entire content with minimum bandwidth. This paper proposed a novel method to generate personalized WebP images in the client side using its computation resources. The proposed system automatically extracts the WebP image from climax point using music analysis. Based on user interest, the system predicts the genre using Convolutional Neural Network (CNN). The proposed method can easily integrate with streaming platforms such as YouTube, Netflix, Hulu, and others.

  • PDF

A Novel Video Copy Detection Method based on Statistical Analysis (통계적 분석 기반 불법 복제 비디오 영상 감식 방법)

  • Cho, Hye-Jeong;Kim, Ji-Eun;Sohn, Chae-Bong;Chung, Kwang-Sue;Oh, Seoung-Jun
    • Journal of Broadcast Engineering
    • /
    • v.14 no.6
    • /
    • pp.661-675
    • /
    • 2009
  • The carelessly and illegally copied contents are raising serious social problem as internet and multimedia technologies are advancing. Therefore, development of video copy detection system must be settled without delay. In this paper, we propose the hierarchical video copy detection method that estimates similarity using statistical characteristics between original video and manipulated(transformed) copy video. We rank according to luminance value of video to be robust to spacial transformation, and choose similar videos categorized as candidate segments in huge amount of database to reduce processing time and complexity. The copy videos generally insert black area in the edge of the image, so we remove rig black area and decide copy or not by using statistical characteristics of original video and copied video with center part of frame that contains important information of video. Experiment results show that the proposed method has similar keyframe accuracy to reference method, but we use less memory to save feature information than reference's, because the number of keyframes is less 61% than that of reference's. Also, the proposed method detects if the video is copied or not efficiently despite expansive spatial transformations such as blurring, contrast change, zoom in, zoom out, aspect ratio change, and caption insertion.

A Study on the Analysis of Technique and Artistic Expression Factors in the Process of 3D Animation Production (3D 애니메이션 제작 과정에서 기술 및 예술적 표현요소 분석에 관한 연구)

  • 백승만;조윤아
    • Archives of design research
    • /
    • v.16 no.1
    • /
    • pp.83-92
    • /
    • 2003
  • Recently 3D animation can produce animation dose to real picture owing to rapid technical development of hardware and software and is used as entertainment factors for movie, ads, game and digital pictures. The process of making 3D animation needs the various making techniques and expression factors unlike traditional animation. Production technique and expression factors in making 3D animation are not acted as individual factor, but they can achieve an animation with high perfection when they should be combined. Therefore this study analyzes the technical expression methods based on hardware and software for operation by computer process and the artistic expression factors based on aesthetic, figurative, design and pictorial factors and then seeks a method of making animation with technical perfection.

  • PDF

Automatic Poster Generation System Using Protagonist Face Analysis

  • Yeonhwi You;Sungjung Yong;Hyogyeong Park;Seoyoung Lee;Il-Young Moon
    • Journal of information and communication convergence engineering
    • /
    • v.21 no.4
    • /
    • pp.287-293
    • /
    • 2023
  • With the rapid development of domestic and international over-the-top markets, a large amount of video content is being created. As the volume of video content increases, consumers tend to increasingly check data concerning the videos before watching them. To address this demand, video summaries in the form of plot descriptions, thumbnails, posters, and other formats are provided to consumers. This study proposes an approach that automatically generates posters to effectively convey video content while reducing the cost of video summarization. In the automatic generation of posters, face recognition and clustering are used to gather and classify character data, and keyframes from the video are extracted to learn the overall atmosphere of the video. This study used the facial data of the characters and keyframes as training data and employed technologies such as DreamBooth, a text-to-image generation model, to automatically generate video posters. This process significantly reduces the time and cost of video-poster production.

Big Data Analysis Method for Recommendations of Educational Video Contents (사용자 추천을 위한 교육용 동영상의 빅데이터 분석 기법 비교)

  • Lee, Hyoun-Sup;Kim, JinDeog
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.25 no.12
    • /
    • pp.1716-1722
    • /
    • 2021
  • Recently, the capacity of video content delivery services has been increasing significantly. Therefore, the importance of user recommendation is increasing. In addition, these contents contain a variety of characteristics, making it difficult to express the characteristics of the content properly only with a few keywords(Elements used in the search, such as titles, tags, topics, words, etc.) specified by the user. Consequently, existing recommendation systems that use user-defined keywords have limitations that do not properly reflect the characteristics of objects. In this paper, we compare the efficiency of between a method using voice data-based subtitles and an image comparison method using keyframes of images in recommendation module of educational video service systems. Furthermore, we propose the types and environments of video content in which each analysis technique can be efficiently utilized through experimental results.

Annotation Method based on Face Area for Efficient Interactive Video Authoring (효과적인 인터랙티브 비디오 저작을 위한 얼굴영역 기반의 어노테이션 방법)

  • Yoon, Ui Nyoung;Ga, Myeong Hyeon;Jo, Geun-Sik
    • Journal of Intelligence and Information Systems
    • /
    • v.21 no.1
    • /
    • pp.83-98
    • /
    • 2015
  • Many TV viewers use mainly portal sites in order to retrieve information related to broadcast while watching TV. However retrieving information that people wanted needs a lot of time to retrieve the information because current internet presents too much information which is not required. Consequentially, this process can't satisfy users who want to consume information immediately. Interactive video is being actively investigated to solve this problem. An interactive video provides clickable objects, areas or hotspots to interact with users. When users click object on the interactive video, they can see additional information, related to video, instantly. The following shows the three basic procedures to make an interactive video using interactive video authoring tool: (1) Create an augmented object; (2) Set an object's area and time to be displayed on the video; (3) Set an interactive action which is related to pages or hyperlink; However users who use existing authoring tools such as Popcorn Maker and Zentrick spend a lot of time in step (2). If users use wireWAX then they can save sufficient time to set object's location and time to be displayed because wireWAX uses vision based annotation method. But they need to wait for time to detect and track object. Therefore, it is required to reduce the process time in step (2) using benefits of manual annotation method and vision-based annotation method effectively. This paper proposes a novel annotation method allows annotator to easily annotate based on face area. For proposing new annotation method, this paper presents two steps: pre-processing step and annotation step. The pre-processing is necessary because system detects shots for users who want to find contents of video easily. Pre-processing step is as follow: 1) Extract shots using color histogram based shot boundary detection method from frames of video; 2) Make shot clusters using similarities of shots and aligns as shot sequences; and 3) Detect and track faces from all shots of shot sequence metadata and save into the shot sequence metadata with each shot. After pre-processing, user can annotates object as follow: 1) Annotator selects a shot sequence, and then selects keyframe of shot in the shot sequence; 2) Annotator annotates objects on the relative position of the actor's face on the selected keyframe. Then same objects will be annotated automatically until the end of shot sequence which has detected face area; and 3) User assigns additional information to the annotated object. In addition, this paper designs the feedback model in order to compensate the defects which are wrong aligned shots, wrong detected faces problem and inaccurate location problem might occur after object annotation. Furthermore, users can use interpolation method to interpolate position of objects which is deleted by feedback. After feedback user can save annotated object data to the interactive object metadata. Finally, this paper shows interactive video authoring system implemented for verifying performance of proposed annotation method which uses presented models. In the experiment presents analysis of object annotation time, and user evaluation. First, result of object annotation average time shows our proposed tool is 2 times faster than existing authoring tools for object annotation. Sometimes, annotation time of proposed tool took longer than existing authoring tools, because wrong shots are detected in the pre-processing. The usefulness and convenience of the system were measured through the user evaluation which was aimed at users who have experienced in interactive video authoring system. Recruited 19 experts evaluates of 11 questions which is out of CSUQ(Computer System Usability Questionnaire). CSUQ is designed by IBM for evaluating system. Through the user evaluation, showed that proposed tool is useful for authoring interactive video than about 10% of the other interactive video authoring systems.