• Title/Summary/Keyword: Cache Hoarding

Search Result 5, Processing Time 0.022 seconds

A Cache Hoarding Method Using Collaborative Filtering in Mobile Computing Environments (모바일 컴퓨팅 환경에서 협업추천 모형을 이용한 캐시 적재 기법)

  • Jun, Sung-Hae;Jung, Sung-Won;Oh, Kyung-Whan
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.14 no.6
    • /
    • pp.687-692
    • /
    • 2004
  • In this paper, we proposed an efficient cache hoarding method in mobile computing environments using collaborative filtering. This method is used for solving the difficult problem of mobile computing, which is the vacuum of information service depending on low bandwidth, long delay, and frequent network disconnection. Many previous researches have been studied a cache hoarding approach for solving these problems of mobile client. But, the research of history information of mobile client did not support all informative requests for mobile clients. In our research, collaborative filtering model using history information and location data of mobile client is proposed. This proposed model supports an efficient service of necessary items for client's requirement. For the performance evaluation of proposed model, we make an experiment of simulation data using SAS enterprise miner. According to objective evaluation using cache hit ratio, we show that our model has a good result.

Ubiquitous Computing and Statistics; What's the Connection?

  • Jun, Sung-Hae;Jorn, Hongsuk
    • Communications for Statistical Applications and Methods
    • /
    • v.11 no.2
    • /
    • pp.287-295
    • /
    • 2004
  • Mark Weiser introduced ubiquitous computing in his article titled 'The computer for 21st Century' in 1991. This has been new paradigm after internet. Now, the rapid development of mobile computer, wireless network, and intelligent system has supported ubiquitous computing environment. In the related area of information science, the researchers have studied on ubiquitous computing. But in the field of Korea statistics, this research has not been worked yet. So, we proposed the connection between statistics and ubiquitous computing in this paper. As an example, we showed an efficient cache hoarding for ubiquitous computing using statistical methods. In experimental results, we verified our proposed issue.

Data Mining Approach for Supporting Hoarding in Mobile Computing Environments

  • Jeon, Seong-Hae;Ryu, Je-Bok;Lee, Seung-Ju
    • Proceedings of the Korean Statistical Society Conference
    • /
    • 2003.05a
    • /
    • pp.13-17
    • /
    • 2003
  • 본 논문에서는 낮은 대역폭, 높은 지연, 그리고 잦은 네트워크 단절로 인한 모바일 컴퓨팅 환경의 문제점들을 해결하기 위한 효과적인 캐시 적재 기법으로서 협업 추천 기반의 데이터 마이닝 전략을 제안하였다. 캐시 적재가 모바일 클라이언트의 이러한 문제점들을 해결하기 위한 효율적인 방법이 된다는 기존의 연구는 많이 진행되어 왔다. 하지만 모바일 컴퓨터의 요구에 대한 이력 정보만을 이용한 기존의 연구는 모바일 클라이언트가 필요로 하는 모든 정보 요구를 만족하지 못하였다. 특히 저장 공간의 제약을 갖는 모바일 컴퓨터의 한계 때문에 더욱 큰 어려움을 갖게 되었다. 본 연구에서는 모바일 클라이언트의 이력 정보에 대하여 데이터 마이닝 기법을 적용한 캐시 적재 기법을 제안하여 적은 캐시 용량만으로도 모바일 클라이언트의 요구를 만족할 수 있는 아이템들을 효과적으로 서비스할 수 있도록 하였다. CSIM Simulator를 이용하여 모의 데이터를 생성하여, 제안 모형의 성능 평가를 위한 실험을 수행하였다. Cache hit ratio를 이용한 객관적인 성능 평가를 통하여 제안된 모형이 모바일 클라이언트의 캐시 적재 기법으로서 우수한 성능을 보임이 확인되었다.

  • PDF

Prefetching Mechanism using the User's File Access Pattern Profile in Mobile Computing Environment (이동 컴퓨팅 환경에서 사용자의 FAP 프로파일을 이용한 선인출 메커니즘)

  • Choi, Chang-Ho;Kim, Myung-Il;Kim, Sung-Jo
    • Journal of KIISE:Information Networking
    • /
    • v.27 no.2
    • /
    • pp.138-148
    • /
    • 2000
  • In the mobile computing environment, in order to make copies of important files available when being disconnected the mobile host(client) must store them in its local cache while the connection is maintained. In this paper, we propose the prefetching mechanism for the client to save files which may be accessed in the near future. Our mechanism utilizes analyzer, prefetch-list producer, and prefetch manager. The analyzer records file access patterns of the user in a FAP(File Access Patterns) profile. Using the profile, the prefetch-list producer creates the prefetch-list. The prefetch manager requests a file server to return this list. We set the parameter TRP(Threshold of Reference Probability) to ensure that only reasonably related files can be prefetched. The prefetch-list producer adds the files to a prefetch-list if their reference probability is greater than the TRP. We also use the parameter TACP(Threshold of Access Counter Probability) to reduce the hoarding size required to store a prefetch-list. Finally, we measure the metrics such as the cache hit ratio, the number of files referenced by the client after disconnection and the hoarding size. The simulation results show that the performance of our mechanism is superior to that of the LRU caching mechanism. Our results also show that prefetching with the TACP can reduce the hoard size while maintaining similar performance of prefetching without TACP.

  • PDF

Broadcasting and Caching Schemes for Location-dependent Queries in Urban Areas (도심환경에서 위치의존 질의를 위한 방송과 캐싱 기법)

  • Jung Il-dong;Yu Young-ho;Lee Jong-hwan;Kim Kyongsok
    • Journal of KIISE:Databases
    • /
    • v.32 no.1
    • /
    • pp.56-70
    • /
    • 2005
  • The results of location-dependent queries(LDQ) generally depend on the current locations of query issuers. Many mechanisms, e.g. broadcast scheme, hoarding, or racking policy, have been developed to improve the system peformance and provide better services, which are specialized for LDQs. Considering geographical adjacency of data and characteristics oi target area, caching policy and broadcast scheme affect the overall performance in LDQ. For this reason, we propose both the caching policy and broadcast scheme, which these features are reflected in. Based on the adjacency of data in LDQ, our broadcast scheme use Hilbert curve to cluster data. Moreover, in order to develop the caching policy suitable for LDQ on urban area, we apply the moving distance of a MH(Mobile Host) to our caching policy We evaluate the performance of the caching policy measuring the workload of MHs and the correctness of LDQ results and the performance of the broadcast scheme measuring the average setup-time of MHs in our experiments. Finally, we expect that our caching policy Provides more correct answers when executing LDQ in focal cache and leads significant improvement of the performance of MHs. It also seems quite probable that our broadcast scheme leads improvement of battery life of the MH.