<p>'Complexity' has several definitions in diverse fields. These measures are indicators of some aspects of the nature of the signal. Such measures are used to analyse and classify signals and as a signal diagnostics tool to distinguish between periodic, quasiperiodic, chaotic and random signals. Lempel-Ziv (LZ) complexity and approximate entropy (ApEn) are such popular complexity measures that are widely used for characterizing biological signals also. In this paper, we compare the utility of ApEn, LZ complexities and Shannon's entropy in characterizing data from a nonlinear chaotic map (logistic map). In this work, we show that LZ and ApEn complexity measures can characterize the data complexities correctly for data sequences as short as 20 in length while Shannon's entropy fails for length less than 50. In the case of noisy sequences with 10% uniform noise, Shannon's entropy works only for lengths greater than 200 while LZ and ApEn are successful with sequences of lengths greater than 30 and 20, respectively. © Indian Academy of Sciences.</p>
cited By 0
K. Balasubramanian, Nair, S. S., and Nagaraj, N., “Classification of periodic, chaotic and random sequences using approximate entropy and Lempel-Ziv complexity measures”, Pramana - Journal of Physics, vol. 84, pp. 365-372, 2015.