Affiliation:
1. Institute for Cross-Disciplinary Physics and Complex Systems IFISC (UIB-CSIC), Campus Universitat de les Illes Balears, E-07122 Palma de Mallorca, Spain
Abstract
Entropy estimation is a fundamental problem in information theory that has applications in various fields, including physics, biology, and computer science. Estimating the entropy of discrete sequences can be challenging due to limited data and the lack of unbiased estimators. Most existing entropy estimators are designed for sequences of independent events and their performances vary depending on the system being studied and the available data size. In this work, we compare different entropy estimators and their performance when applied to Markovian sequences. Specifically, we analyze both binary Markovian sequences and Markovian systems in the undersampled regime. We calculate the bias, standard deviation, and mean squared error for some of the most widely employed estimators. We discuss the limitations of entropy estimation as a function of the transition probabilities of the Markov processes and the sample size. Overall, this paper provides a comprehensive comparison of entropy estimators and their performance in estimating entropy for systems with memory, which can be useful for researchers and practitioners in various fields.
Funder
Agencia Estatal de Investigación
Reference73 articles.
1. Dobzhansky, T., Hecht, M.K., and Steere, W.C. (1972). Evolutionary Biology: Volume 6, Springer.
2. Stinson, D.R. (1995). Cryptography: Theory and Practice, CRC Press Inc.. [1st ed.].
3. Entropy and Information in Neural Spike Trains;Strong;Phys. Rev. Lett.,1998
4. Maximum Entropy Modeling of Short Sequence Motifs with Applications to RNA Splicing Signals;Yeo;J. Comput. Biol. J. Comput. Mol. Cell Biol.,2004
5. Cover, T., and Thomas, J. (2006). Elements of Information Theory, John Wiley and Sons.
Cited by
3 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献