We investigate a neural net-based algorithm for enhanced imaging throu
gh atmospheric turbulence. The concept is based on a standard model of
optical turbulence, according to which a short-exposure point-spread
function is a random superposition of speckles. This leads to a new me
thod of image processing called the Fourier division approach. The lat
ter requires the taking of two short-exposure images in rapid successi
on, which are picked up by an image-plane array, divided in Fourier sp
ace, and then processed by a minimum entropy-neural net approach. The
main task of the processing is to estimate the two short-exposure poin
t-spread functions that characterize the two images. Given these estim
ates, the two images may now be inverse filtered to produce two sharp
object-scene estimates. These have most of the turbulence degradation
removed, and are averaged to produce a single output image. The approa
ch shows promise, in computer simulations, of removing nearly all of t
he turbulence degradation very quickly (currently tens of seconds). A
further benefit arises from knowledge of the two short-exposure point-
spread functions. These should permit identification of the state of t
urbulence along the imaging Line of sight and, in particular, the pres
ence of wind shear.