Representations of images, in general, belong to three probabilistic families, developed for different regimes of data and tasks. (i) Descriptive models, originated from statistical physics, reproduce certain statistical regularities in data, and are often suitable for patterns in the high entropy regime, such as MRF, Gibbs and FRAME. (ii) Generative models, originated from harmonic analysis, seek latent variables and dictionaries to explain data in parsimonious representations, and are often more effective for the low entropy regime, such as sparse models and auto-encoders. (iii) Discriminative models are often trained by statistical regression for classification tasks. This talk will start with the Julesz quest on texture and texton representations in the 1960s, and then review the developments, interactions and integration of these model families in the recent deep learning era, such as the adversary and cooperative models. Then the talk will draw a unification of these models in a continuous entropy spectrum in terms of information scaling. Finally, the talk will discuss future directions in developing cognitive models for representations beyond deep learning, i.e. modeling the task-oriented cognitive aspects, such as functionality, physics, intents and causality, which are the invisible “dark matter”, by analogy to cosmology, in human intelligence.