• Title/Summary/Keyword: time memory data tradeoff

Search Result 4, Processing Time 0.021 seconds

PERFORMANCE COMPARISON OF CRYPTANALYTIC TIME MEMORY DATA TRADEOFF METHODS

  • Hong, Jin;Kim, Byoung-Il
    • Bulletin of the Korean Mathematical Society
    • /
    • v.53 no.5
    • /
    • pp.1439-1446
    • /
    • 2016
  • The execution complexities of the major time memory data tradeoff methods are analyzed in this paper. The multi-target tradeoffs covered are the classical Hellman, distinguished point, and fuzzy rainbow methods, both in their non-perfect and perfect table versions for the latter two methods. We show that their computational complexities are identical to those of the corresponding single-target methods executed under certain matching parameters and conclude that the perfect table fuzzy rainbow tradeoff method is most preferable.

Analysis on TMD-Tradeoff and State Entropy Loss of Stream Cipher MICKEY (스트림 암호 MICKEY의 TMD-Tradeoff와 내부 상태 엔트로피의 손실에 관한 분석)

  • Kim, Woo-Hwan;Hong, Jin
    • Journal of the Korea Institute of Information Security & Cryptology
    • /
    • v.17 no.2
    • /
    • pp.73-81
    • /
    • 2007
  • We give two weaknesses of a recently proposed streamcipher MICKEY. We show time-memory-data tradeoff is applicable. We also show that the state update function reduces entropy of the internal state as it is iterated, resulting in keystreams that start out differently but become merged together towards the end.

Efficient Accessing and Searching in a Sequence of Numbers

  • Seo, Jungjoo;Han, Myoungji;Park, Kunsoo
    • Journal of Computing Science and Engineering
    • /
    • v.9 no.1
    • /
    • pp.1-8
    • /
    • 2015
  • Accessing and searching in a sequence of numbers are fundamental operations in computing that are encountered in a wide range of applications. One of the applications of the problem is cryptanalytic time-memory tradeoff which is aimed at a one-way function. A rainbow table, which is a common method for the time-memory tradeoff, contains elements from an input domain of a hash function that are normally sorted integers. In this paper, we present a practical indexing method for a monotonically increasing static sequence of numbers where the access and search queries can be addressed efficiently in terms of both time and space complexity. For a sequence of n numbers from a universe $U=\{0,{\ldots},m-1\}$, our data structure requires n lg(m/n) + O(n) bits with constant average running time for both access and search queries. We also give an analysis of the time and space complexities of the data structure, supported by experiments with rainbow tables.

Non-Simultaneous Sampling Deactivation during the Parameter Approximation of a Topic Model

  • Jeong, Young-Seob;Jin, Sou-Young;Choi, Ho-Jin
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.7 no.1
    • /
    • pp.81-98
    • /
    • 2013
  • Since Probabilistic Latent Semantic Analysis (PLSA) and Latent Dirichlet Allocation (LDA) were introduced, many revised or extended topic models have appeared. Due to the intractable likelihood of these models, training any topic model requires to use some approximation algorithm such as variational approximation, Laplace approximation, or Markov chain Monte Carlo (MCMC). Although these approximation algorithms perform well, training a topic model is still computationally expensive given the large amount of data it requires. In this paper, we propose a new method, called non-simultaneous sampling deactivation, for efficient approximation of parameters in a topic model. While each random variable is normally sampled or obtained by a single predefined burn-in period in the traditional approximation algorithms, our new method is based on the observation that the random variable nodes in one topic model have all different periods of convergence. During the iterative approximation process, the proposed method allows each random variable node to be terminated or deactivated when it is converged. Therefore, compared to the traditional approximation ways in which usually every node is deactivated concurrently, the proposed method achieves the inference efficiency in terms of time and memory. We do not propose a new approximation algorithm, but a new process applicable to the existing approximation algorithms. Through experiments, we show the time and memory efficiency of the method, and discuss about the tradeoff between the efficiency of the approximation process and the parameter consistency.