Learning Efficient Codes for Natural Images

Part I. Michael S. Lewicki

collaborator: Terrence J. Sejnowski

We derive a learning algorithm for inferring an overcomplete basis, ie more basis functions than input variables, by viewing it as probabilistic model of the observed data. Redundancy in the representation is removed by the prior on the basis coefficients which assigns probabilities to alternative representations. A Laplacian prior leads to representations that are sparse and are a nonlinear function of the data. In contrast to traditional complete bases, such as Fourier or Wavelet, the vectors in an overcomplete basis can become specialized for a larger variety of features present in the entire ensemble of data. This framework provides a new view of overcomplete bases, placing emphasis on their greater flexibility in modeling the underlying statistical density of the data, which allows for better noise reduction properties and greater coding efficiency.

Part II. Bruno A. Olshausen

We apply a general technique for learning overcomplete bases (Lewicki & Sejnowski, 1997) to the problem of finding efficient image codes. The bases learned by the algorithm are localized, oriented, and bandpass, consistent with earlier results obtained using different methods Olshausen and Field (1996); Bell and Sejnowski (1996). We show that higher degrees of overcompleteness produce bases which have much greater likelihood and results in a Gabor-like basis with greater sampling density in position, orientation, and scale. This framework also allows different bases to be compared objectively by calculating their probability given the observed data. Compared to the complete and overcomplete Fourier and wavelet bases, the learned bases have much greater probability and thus have the potential to yield better coding efficiency. We demonstrate the improvement in the representation of the learned bases by showing superior noise reduction properties.
Back to the Natural Scenes Meeting Agenda.