Polynomial learnability of stochastic rules with respect to the KL-divergence and quadratic distance

Citation
N. Abe et al., Polynomial learnability of stochastic rules with respect to the KL-divergence and quadratic distance, IEICE T INF, E84D(3), 2001, pp. 299-316
Citations number
18
Categorie Soggetti
Information Tecnology & Communication Systems
Journal title
IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS
ISSN journal
09168532 → ACNP
Volume
E84D
Issue
3
Year of publication
2001
Pages
299 - 316
Database
ISI
SICI code
0916-8532(200103)E84D:3<299:PLOSRW>2.0.ZU;2-B
Abstract
We consider the problem of efficient learning of probabilistic concepts (p- concepts) and more generally stochastic rules in the sense defined by Kearn s and Schapire [6] and by Yamanishi [18]. Their models extend the PAC-learn ing model of Valiant [16] to the learning scenario in which the target conc ept or function is stochastic rather than deterministic as in Valiant's ori ginal model. In this paper, we consider the learnability of stochastic rule s with respect to the classic 'Kullback-Leibler divergence' (KL divergence) as well as the quadratic distance as the distance measure between the rule s. First, we show that the notion of polynomial time learnability of p-conc epts and stochastic rules with fixed range sire using the KL divergence is in fact equivalent to the same notion using the quadratic distance, and hen ce any of the distances considered in [6] and [18]: the quadratic, variatio n, and Kellinger distances. As a corollary, it follows that a wide range of classes of p-concepts which were shown to be polynomially learnable with r espect to the quadratic distance in [6] are also learnable with respect to the KL divergence. The sample and time complexity of algorithms that would be obtained by the above general equivalence, however, are far from optimal . We present a polynomial learning algorithm with reasonable sample and tim e complexity far the important class of convex linear combinations of stoch astic rules. We also develop a simple and versatile technique for obtaining sample complexity bounds for learning classes of stochastic rules with res pect to the KL-divergence and quadratic distance, and apply them to produce bounds for the classes of probabilistic finite state accepters (automata), probabilistic decision lists, and convex linear combinations.