Guesswork, Large Deviations, and Shannon Entropy

IEEE Transactions on Information Theory(2013)

引用 77|浏览28
暂无评分
摘要
How hard is it to guess a password? Massey showed that a simple function of the Shannon entropy of the distribution from which the password is selected is a lower bound on the expected number of guesses, but one which is not tight in general. In a series of subsequent papers under ever less restrictive stochastic assumptions, an asymptotic relationship as password length grows between scaled moments of the guesswork and specific Rényi entropy was identified. Here, we show that, when appropriately scaled, as the password length grows, the logarithm of the guesswork satisfies a large deviation principle (LDP), providing direct estimates of the guesswork distribution when passwords are long. The rate function governing the LDP possesses a specific, restrictive form that encapsulates underlying structure in the nature of guesswork. Returning to Massey's original observation, a corollary to the LDP shows that expectation of the logarithm of the guesswork is the specific Shannon entropy of the password selection process.
更多
查看译文
关键词
authorisation,entropy,Rényi entropy,Shannon entropy,asymptotic relationship,guesswork distribution,large deviation principle,password selection process,restrictive stochastic assumption,Guesswork,Rényi entropy,Shannon entropy,large deviations
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要