KULLBACK-LEIBLER INFORMATION MEASURE FOR STUDYING CONVERGENCE-RATES OF DENSITIES AND DISTRIBUTIONS

Citation
Me. Meyer et Dv. Gokhale, KULLBACK-LEIBLER INFORMATION MEASURE FOR STUDYING CONVERGENCE-RATES OF DENSITIES AND DISTRIBUTIONS, IEEE transactions on information theory, 39(4), 1993, pp. 1401-1409
Citations number
11
Categorie Soggetti
Mathematics,"Engineering, Eletrical & Electronic
ISSN journal
00189448
Volume
39
Issue
4
Year of publication
1993
Pages
1401 - 1409
Database
ISI
SICI code
0018-9448(1993)39:4<1401:KIMFSC>2.0.ZU;2-2
Abstract
The objective is to propose the Kullback-Leibler (KL) information meas ure I(f1 : f2) as an index for studying rates of convergence of densit ies and distribution functions. To this end, upper bounds in terms of I(f1 : f2) for several distance functions for densities and for distri bution functions are obtained first. Many illustrations of the use of this technique are given. It is shown, for example, that the sequence of KL information measures converges to zero more slowly for a normali zed sequence of gamma random variables converging to its limiting norm al distribution than for a normalized sequence of largest order statis tics from an exponential distribution converging to its limiting extre me value distribution. Furthermore, a sequence of KL information measu res for log-normal random variables approaching normality converges mo re slowly to zero than for a sequence of normalized gamma random varia bles.