The goal of this paper is to investigate various language model smooth
ing techniques and decision tree based language model design algorithm
s. For this purpose, we build language models for printable characters
(letters), based on the Brown corpus. We consider two classes of mode
ls for the text generation process: the n-gram language model and vari
ous decision tree based language models. In the first part of the pape
r, we compare the most popular smoothing algorithms applied to the for
mer. We conclude that the bottom-up deleted interpolation algorithm pe
rforms the best in the task of n-gram letter language model smoothing,
significantly outperforming the back-off smoothing technique for larg
e values of n. In the second part of the paper, we consider various de
cision tree development algorithms. Among them, a K-means clustering t
ype algorithm for the design of the decision tree questions gives the
best results. However, the n-gram language model outperforms the decis
ion tree language models for letter language modeling. We believe that
this is due to the predictive nature of letter strings, which seems t
o be naturally modeled by n-grams, (C) 1998 Elsevier Science B.V. All
rights reserved.