Go to the main menu
Skip to content
Go to bottom
REFERENCE LINKING PLATFORM OF KOREA S&T JOURNALS
> Journal Vol & Issue
KIISE Transactions on Computing Practices
Journal Basic Information
Journal DOI :
Korean Institute of Information Scientists and Engineers
Editor in Chief :
Volume & Issues
Volume 20, Issue 12 - Dec 2014
Volume 20, Issue 11 - Nov 2014
Volume 20, Issue 10 - Oct 2014
Volume 20, Issue 9 - Sep 2014
Selecting the target year
Big Data Preprocessing for Predicting Box Office Success
Jun, Hee-Gook ; Hyun, Geun-Soo ; Lim, Kyung-Bin ; Lee, Woo-Hyun ; Kim, Hyoung-Joo ;
KIISE Transactions on Computing Practices, volume 20, issue 12, 2014, Pages 615~622
DOI : 10.5626/KTCP.2014.20.12.615
The Korean film market has rapidly achieved an international scale, and this has led to a need for decision-making based on analytical methods that are more precise and appropriate. In this modern era, a highly advanced information environment can provide an overwhelming amount of data that is generated in real time, and this data must be properly handled and analyzed in order to extract useful information. In particular, the preprocessing of large data, which is the most time-consuming step, should be done in a reasonable amount of time. In this paper, we investigated a big data preprocessing method for predicting movie box office success. We analyzed the movie data characteristics for specialized preprocessing methods, and used the Hadoop MapReduce framework. The experimental results showed that the preprocessing methods using big data techniques are more effective than existing methods.
Performance Analysis of Open Source Based Distributed Deduplication File System
Jung, Sung-Ouk ; Choi, Hoon ;
KIISE Transactions on Computing Practices, volume 20, issue 12, 2014, Pages 623~631
DOI : 10.5626/KTCP.2014.20.12.623
Comparison of two representative deduplication file systems, LessFS and SDFS, shows that Lessfs is better in execution time and CPU utilization while SDFS is better in storage usage (around 1/8 less than general file systems). In this paper, a new system is proposed where the advantages of SDFS and Lessfs are combined. The new system uses multiple DFEs and one DSE to maintain the integrity and consistency of the data. An evaluation study to compare between Single DFE and Dual DFE indicates that the Dual DFE was better than the Single DFE. The Dual DFE reduced the CPU usage and provided fast deduplication time. This reveals that proposed system can be used to solve the problem of an increase in large data storage and power consumption.
A Study of File Outflow Monitoring Process using the File System and NDIS Driver
Lee, Mintae ;
KIISE Transactions on Computing Practices, volume 20, issue 12, 2014, Pages 632~639
DOI : 10.5626/KTCP.2014.20.12.632
In this paper, we propose a PC's self-monitoring system for protecting illegal outflow of important personal files, which are managed in the Windows environment PC. This paper is based on the idea that it should be a read (access) file operation in order to outflow files through the network. To compare the information (name, some content) obtained from all operations of the reading of the 'Windows File System Driver' and captured transmission packets of mini port of 'Windows NDIS (Network Driver Interface Specification) Driver', and, if two pieces of information match, this system will determine the transmission to user. In this paper, tentatively, we developed a function to gather file information about file read operation and developed a function for transmission packet capture. This demonstration implementation showed that the proposed process has validity and the proposed process verified that it does not significantly affect the PC's performance.
Design and Implementation of the Web-based Learning System for C Programming Language
Woo, Yeomyeong ; Bang, Jiwoong ; Song, Jaemin ; Yoo, Jinyeong ; Lee, Sangjun ;
KIISE Transactions on Computing Practices, volume 20, issue 12, 2014, Pages 640~645
DOI : 10.5626/KTCP.2014.20.12.640
We now live in a society that is highly information-oriented, and as a result programming education has become more important and is emphasized day by day. Currently, there is strong support for carrying out programming education in early childhood, and various Web-based services have launched to provide programming education. Currently, Web-based programming education services have not yet launched in Korea. In particular, Web-based education services for C language programming, which is used in various areas in industry, are lacking due to limitations in the development environments. In this paper, we design and implement a Web-based programming learning system, 'ICANC', to provide the necessary information to learn and practice the C language.
Histogram Equalization using Gamma Transformation
Chung, Soyoung ; Chung, Min Gyo ;
KIISE Transactions on Computing Practices, volume 20, issue 12, 2014, Pages 646~651
DOI : 10.5626/KTCP.2014.20.12.646
Histogram equalization generally has the disadvantage that if the distribution of the gray level of an image is concentrated in one place, then the range of the gray level in the output image is excessively expanded, which then produces a visually unnatural result. However, a gamma transformation can reduce such unnatural appearances since it operates under a nonlinear regime. Therefore, this paper proposes a new histogram equalization method that can improve image quality by using a gamma transformation. The proposed method 1) derives the proper form of the gamma transformation by using the average brightness of the input image, 2) linearly combines the earlier gamma transformation with a CDF (Cumulative Distribution Function) for the image in order to obtain a new CDF, and 3) to finally perform histogram equalization by using the new CDF. The experimental results show that relative to existing methods, the proposed method provides good performance in terms of quantitative measures, such as entropy, UIQ, SSIM, etc., and it also naturally enhances the image quality in visual perspective as well.
An Adaptive Priority-based Sequenced Route Query Processing Method in Road Networks
Ryu, Hyeongcheol ; Jung, Sungwon ;
KIISE Transactions on Computing Practices, volume 20, issue 12, 2014, Pages 652~657
DOI : 10.5626/KTCP.2014.20.12.652
Given a starting point, destination point and various Points Of Interest (POIs), which contain a full or partial order, for a user to visit we wish to create, a sequenced route from the starting point to the destination point that includes one member of each POI type in a particular order. This paper proposes a method for finding the approximate shortest route between the start point, destination point and one member of each POI type. There are currently two algorithms that perform this task but they both have weaknesses. One of the algorithms only considers the distance between the visited POI (or starting point) and POI to visit next. The other algorithm chooses candidate points near the straight-line distance between the start point and destination but does not consider the order of visits on the corresponding network path. This paper outlines an algorithm that chooses the candidate points that are nearer to the network path between the start point and destination using network search. The algorithm looks for routes using the candidate points and finds the approximate shortest route by assigning an adaptive priority to the route that visits more POIs in a short amount of time.
Development of a Testing Tool to Validate Integrity of a Constructed Cloud System based on OpenStack
Son, Seokho ; Kang, Dong-Jae ;
KIISE Transactions on Computing Practices, volume 20, issue 12, 2014, Pages 658~663
DOI : 10.5626/KTCP.2014.20.12.658
Due to the rapid development of open source based Cloud management platforms such as OpenStack, many and various individuals and institutions are trying to construct Cloud computing systems based on open source software projects. It is essential, for those who install, administrate or operate a Cloud, to validate the integrity of the constructed Cloud system. This paper, therefore, proposes the design of a testing tool for validating the integrity of a constructed OpenStack-based Cloud system. Especially, the proposed testing tool utilizes the Tempest project which is an open source project that consists of OpenStack test cases. The proposed testing tool is expected to help developing Cloud technology and Cloud users.
Variability-based Service Specification Method for Brokering Cloud Services
An, Youngmin ; Park, Joonseok ; Yeom, Keunhyuk ;
KIISE Transactions on Computing Practices, volume 20, issue 12, 2014, Pages 664~669
DOI : 10.5626/KTCP.2014.20.12.664
As the prevalence of cloud computing increases, various cloud service types have emerged, such as IaaS, PaaS, and SaaS. The growth and diversification of these cloud services has also resulted in the development of technology for cloud service brokers (CSBs), which serve as intermediate cloud services that can assist cloud tenants (users) in deploying services that fit their requirements. In order to broker cloud services, CSBs require the specification of structural models in order to facilitate the analysis and search for cloud services. In this study, we propose a variability-based service analysis model (SAM) that can be used to describe various cloud services. This model is based on the concept of variability in the software product line and represents the commonality and variability of cloud services by binding variants to each variation point that exists in the specification, quality, and pricing of the services. We also propose a virtual cloud bank architecture as a CSB that serves as an intermediate to provides tenants with appropriate cloud services based on the SAM.
Effectiveness Optimization for Metro-Style Graphical User Interfaces
Kim, Kangtae ; Kim, Kihyuk ; Lee, Sungkil ;
KIISE Transactions on Computing Practices, volume 20, issue 12, 2014, Pages 670~675
DOI : 10.5626/KTCP.2014.20.12.670
Graphical user interfaces (GUI) in modern software deliver information visually, and a well-designed interface can provide information to the use in an organized and intuitive manner while poorly-designed interfaces can cause visual inconvenience and confusion. In order to effectively deliver information to the user, visual attention should be placed on a prominent location in the image. This paper introduces a method based on a human visual system (HVS) that can improve Metro-style GUIs by reducing a user's workload to visually find information. Our method is designed with spatial mapping and color mapping for buttons in the Metro-style GUI. Also we define a metric for Metro-style GUI effectiveness, including an optimization algorithm. The results show that our method improves the performance of visual search tasks in a Metro-style GUI.
Proposing and Validating a Classification Method based on Knowledge Structure to Identify High-Quality Presentation Slides
Jung, Wonchul ; Kim, Seongchan ; Yi, Mun Y. ;
KIISE Transactions on Computing Practices, volume 20, issue 12, 2014, Pages 676~681
DOI : 10.5626/KTCP.2014.20.12.676
In order to discern and classify high-quality slides, our research proposes a classification method that utilizes a knowledge structure containing information on the presentation slides. After analyzing whether our knowledge structure captures the content's quality information, we developed a classification method based on the knowledge structure produced from the analysis results. With the proposed method, we compared results classified by quality of presentation slides. Through this comparison, we verified that the slides in the high quality group could be classified and were able to retrieve high quality slides. The results show that, by utilizing the cognitive model of a knowledge structure, our method can increase the effectiveness of classification when search or recommendation is conducted mainly with high-quality slides.
Improved First-Phoneme Searches Using an Extended Burrows-Wheeler Transform
Kim, Sung-Hwan ; Cho, Hwan-Gue ;
KIISE Transactions on Computing Practices, volume 20, issue 12, 2014, Pages 682~687
DOI : 10.5626/KTCP.2014.20.12.682
First phoneme queries are important functionalities that provide an improvement in the usability of interfaces that produce errors frequently due to their restricted input environment, such as in navigators and mobile devices. In this paper, we propose a time-space efficient data structure for Korean first phoneme queries that disassembles Korean strings in a phoneme-wise manner, rearranges them into circular strings, and finally, indexes them using the extended Burrows-Wheeler Transform. We also demonstrate that our proposed method can process more types of query using less space than previous methods. We also show it can improve the search time when the query length is shorter and the proportion of first phonemes is higher.
Resource Allocation and Control System for VoIP QoS Provision in Cognitive Radio Networks
Kim, Bosung ; Lee, Gyu-Min ; Roh, Byeong-Hee ; Choi, Geunkyung ; Oh, Ilhyuk ;
KIISE Transactions on Computing Practices, volume 20, issue 12, 2014, Pages 688~693
DOI : 10.5626/KTCP.2014.20.12.688
With the advent of ubiquitous environments, the smart phone has come into wide use and the demand for various content increases. Thus, in order to efficiently utilize limited resources cognitive radio technology is regarded as a possible solution. Besides spectrum sensing or access schemes, the provision of VoIP traffic service for secondary users with limited spectrum resources is a very important issue. In this paper, a resource allocation and control system for VoIP QoS provision in cognitive radio networks is proposed. Firstly, as the system model, the time structure of the network is addressed and, according to the structure, a bandwidth broker is proposed. In addition, based on available bandwidth estimated by the bandwidth broker, a connection admission control for secondary users is developed. It is demonstrated that the provision of VoIP QoS is greatly affected by channel utilization, the number of channels, and the length of timeslot.
Bypass Generation Mechanism using Mobility Prediction for Improving Delay of AODV in MANET
Youn, Byungseong ; Kim, Kwangsoo ; Kim, Hakwon ; Roh, Byeong-Hee ;
KIISE Transactions on Computing Practices, volume 20, issue 12, 2014, Pages 694~699
DOI : 10.5626/KTCP.2014.20.12.694
In mobile ad-hoc networks (MANET), the network topology and neighboring nodes change frequently, since MANET is composed of nodes that have mobility without a fixed network infrastructure. The AODV routing protocol is advantageous for MANET, but AODV has a delay in the transmission of data packets because AODV can not transmit data during route recovery. This paper proposes solving the above problem of AODV by using a bypass generation mechanism for data transmission during route recovery. For further improvement, additional mechanisms that coordinate the reception threshold of a hello packet are proposed in order to improve the accuracy of the information obtained from the neighboring nodes when the bypass is generated due to a link failure and the immediacy of the route recovery. Simulation results show that the proposed technique improves the performance in terms of the delay in transmission compared to traditional AODV.
Molecular Computing Simulation of Cognitive Anagram Solving
Chun, Hyo-Sun ; Lee, Ji-Hoon ; Ryu, Je-Hwan ; Baek, Christina ; Zhang, Byoung-Tak ;
KIISE Transactions on Computing Practices, volume 20, issue 12, 2014, Pages 700~705
DOI : 10.5626/KTCP.2014.20.12.700
An anagram is a form of word play to find a new word from a set of given alphabet letters. Good human anagram solvers use the strategy of bigrams. They explore a constraint satisfaction network in parallel and answers consequently pop out quickly. In this paper, we propose a molecular computational algorithm using the same process as this. We encoded letters into DNA sequences and made bigrams and then words by connecting the letter sequences. From letters and bigrams, we performed DNA hybridization, ligation, gel electrophoresis and finally, extraction and separation to extract bigrams. From the matched bigrams and words, we performed the four molecular operations again to distinguish between right and wrong results. Experimental results show that our molecular computer can identify cor rect answers and incorrect answers. Our work shows a new possibility for modeling the cognitive and parallel thinking process of a human.
Two-way DLNA Adaptor for Interconnecting Media Contents in Internet of Things
Yim, Hyung-Jun ; Lee, Kyu-Chul ;
KIISE Transactions on Computing Practices, volume 20, issue 12, 2014, Pages 706~710
DOI : 10.5626/KTCP.2014.20.12.706
The Internet of Things (IoT) is a concept that describes environments where various things are connected to the Internet. These objects can identify themselves and allow other devices the identify them as well. DLNA focuses on delivering interoperability guidelines based on open industry standards in order to complete cross-industry digital convergence. This paper proposes a two-way DLNA Adaptor and addresses the methodologies of service interoperation of the media contents. In this paper, the designs of the components of the DLNA Adaptor are presented, and the conclusion provides general remarks and a discussion of future work.
Sentence Similarity Measurement Method Using a Set-based POI Data Search
Ko, EunByul ; Lee, JongWoo ;
KIISE Transactions on Computing Practices, volume 20, issue 12, 2014, Pages 711~716
DOI : 10.5626/KTCP.2014.20.12.711
With the gradual increase of interest in plagiarism and intelligent file content search, the demand for similarity measuring between two sentences is increasing. There is a lot of researches for sentence similarity measurement methods in various directions such as n-gram, edit-distance and LSA. However, these methods have their own advantages and disadvantages. In this paper, we propose a new sentence similarity measurement method approaching from another direction. The proposed method uses the set-based POI data search that improves search performance compared to the existing hard matching method when data includes the inverse, omission, insertion and revision of characters. Using this method, we are able to measure the similarity between two sentences more accurately and more quickly. We modified the data loading and text search algorithm of the set-based POI data search. We also added a word operation algorithm and a similarity measure between two sentences expressed as a percentage. From the experimental results, we observe that our sentence similarity measurement method shows better performance than n-gram and the set-based POI data search.