Some Characterization Results Based on Dynamic Survival and Failure Entropies

- Journal title : Communications for Statistical Applications and Methods
- Volume 18, Issue 6, 2011, pp.787-798
- Publisher : The Korean Statistical Society
- DOI : 10.5351/CKSS.2011.18.6.787

Title & Authors

Some Characterization Results Based on Dynamic Survival and Failure Entropies

Abbasnejad, Maliheh;

Abbasnejad, Maliheh;

Abstract

In this paper, we develop some characterization results in terms of survival entropy of the first order statistic. In addition, we generalize the cumulative entropy recently proposed by Di Crescenzo and Logobardi (2009) to a new measure of information (called the failure entropy) and study some properties of it and its dynamic version. Furthermore, power distribution is characterized based on dynamic failure entropy.

Keywords

First order statistic;power distribution;mean past life function;reversed Hazard function;

Language

English

Cited by

1.

2.

3.

4.

References

1.

Abbasnejad, M., Arghami, N. R., Morgenthaler, S. and Mohtashami Borzadaran, G. R. (2010). On the dynamic survival entropy, Statistics and Probability Letters, 80, 1962-1971.

2.

Abraham, B. and Sankaran, P. G. (2005). Penyi's entropy for residual lifetime distribution, Statistical Paper, 46, 17-30.

3.

Arnold, B. C., Balakrishnan, N. and Nagaraja, H. N. (1992). A First Course in Order Statistics, John Wiley & Sons, New York.

4.

Asadi, M., Ebrahimi, N. and Soofi, E. S. (2005). Dynamic generalized information measures, Statistics and Probability Letters, 71, 85-98.

5.

Asadi, M. and Zohrevand, Y. (2007). On the dynamic cumulative residual entropy, Journal of Statistical Planning and Inference, 137, 1931-1941.

6.

Baratpour, S. (2010). Characterization based on cumulative residual entropy of first order ststistics, Communications in Statistics: Theory and Methods, 39, 3645-3651.

7.

Baratpour, S., Ahmadi, J. and Arghami, N. R. (2008). Some characterization based on Renyi entropy of order statistics and record values, Journal of Statistical Planning and Inference, 138, 2544-2551.

8.

David, H. A. and Nagaraja, H. N. (2003). Order Statistics, John Wiley & Sons, New York.

9.

Di Crescenzo, A. and Longobardi, M. (2002). Entropy-based measure of uncertainty in past lifetime distributions, Journal of Applied Probability, 39, 434-440.

10.

Di Crescenzo, A. and Longobardi, M. (2004). A measure of discrimination between past lifetime distributions, Statistics and Probability Letters, 67, 173-182.

11.

Di Crescenzo, A. and Longobardi, M. (2009). On cumulative entropies, Journal of Statistical Planning and Inference, 139, 4072-4087.

12.

Ebrahimi, N. (1996). How to measure uncertainty in the residual lifetime distributions, Sankhya, 58, 48-57.

13.

Gertsbakh, I. and Kagan, A. (1999). Characterization of the Weibull distribution by properties of the Fisher information under type I censoring, Statistics and Probability Letters, 42, 99-105.

14.

Kamps, U. (1998). Characterizations of distributions by recurrence relations and identities for moments of order statistics, In Order Statistics: Theory and Methods. Handbook of Statistics, Balakrishnan, N., Rao, C. R., Eds. 16, Amesterdam: Elsevier, 291-311.

15.

Nanda, A. K. and Paul, P. (2006). Some properties of past entropy and their applications, Metrika, 64, 47-61.

16.

Nanda, A. K., Singh, H., Misra, N. and Paul, P. (2003). Reliability properties of reversed residual lifetime, Communications in Statistics: Theory and Methods, 32, 2031-2042.

17.

Rao, M. (2005). More on a new concept of netropy and information, Journal of Theoretical Probability, 18, 967-981.

18.

Rao, M., Chen, Y., Vemuri, B.C. and Wang, F. (2004). Cumulative residual entropy: A new measure of information, IEEE Transactions on Information Theoty, 50, 1220-1228.

19.

Renyi, A. (1961). On measures of entropy and information, In Proceeding of the Fourth Berkeley Symposium, I, UC Press, Berkeley, 547-561.

20.

21.

Wang, F. and Vemuri, B. C. (2007). Non-rigid multi-model image registration using cross-cumulative residual entropy, International Journal of Computer Vision, 74, 201-215.