Information Maximization in Face Processing

Size: px
Start display at page:

Download "Information Maximization in Face Processing"

Transcription

1 Information Maximization in Face Processing Marian Stewart Bartlett Institute for Neural Computation University of California, San Diego Abstract This perspective paper explores principles of unsupervised learning and how they relate to face recognition. Dependency coding and information maximization appear to be central principles in neural coding early in the visual system. These principles may be relevant to how we think about higher visual processes such as face recognition as well. The paper first reviews examples of dependency learning in biological vision, along with principles of optimal information transfer and information maximization. Next, we examine algorithms for face recognition by computer from a perspective of information maximization. The eigenface approach can be considered an unsupervised system that learns the first and second-order dependencies among face image pixels. Eigenfaces maximizes information transfer only in the case where the input distributions are Gaussian. Independent component analysis (ICA) learns high-order dependencies in addition to first and second-order relations, and maximizes information transfer for a more general set of input distributions. Face representations based on ICA gave better recognition performance than eigenfaces, supporting the theory that information maximization is a good strategy for high level visual functions such as face recognition. Finally, we review perceptual studies suggesting that dependency learning is relevant to human face perception as well, and present an information maximization account of perceptual effects such as the atypicality bias, and face adaptation aftereffects. Introduction Horace Barlow has argued that statistical dependencies in the sensory input contain structural information about the environment, and that a general strategy for sensory systems is to learn the expected dependencies. Figure 1 is an example from Barlow s 1989 paper illustrating the idea that the percept of structure is driven by the dependencies in the sensory input. The set of points on the left was randomly selected from a Gaussian distribution. In the figure on the right, half of the points were generated from a Gaussian distribution and the other half were generated by rotating those points 5 degrees about the centroid of the distribution. The dependence between pairs of dots gives this image a percept of structure. Figure 1. Example of visual structure from statistical depencencies. (From Barlow, 1989).

2 Adaptation. Adaptation mechanisms are examples of encoding dependencies in early visual processing. See Figure 2. As pointed out by Barlow (1989), a first order redundancy is mean luminance. Adaptation mechanisms take advantage of this redundancy by expressing lightness values relative to the mean. In this example, these two squares have the same grayvalue, but they are embedded in regions with different mean luminance, and the perception of brightness is different. A second order redundancy is luminance variance, which is a measure of contrast. Contrast gain control mechanisms express contrast relative to the expected contrast. Here is an example of contrast gain control. These two squares have the same contrast, but this one appears to have higher contrast when it is embedded in a region of low mean contrast. Figure 2. Simultaneous contrast. Top: Removing first order redundancy in the luminance signal. The inner squares have the same grayvalue. The percept of brightness is relative to the mean brightness in the region. Bottom: Removing second-order redundancy in the luminance signal. The inner squares have the same contrast. The percept of contrast is relative to the mean contrast in the region. Information maximization in neural codes. A more comprehensive way to encode the redundancy in the input signal, instead of just learning the means, is to learn the probability distribution of signals in the environment. Learning the redundancy means learning which values occur most frequently. Neurons with a limited dynamic range can increase the information that the response gives about the signal by placing the more steeply sloped portions of the transfer function in the regions of highest density, and shallower slopes at regions of low density. What we know from information theory is that the optimal response function for maximizing information transfer is to match the transfer function to the cumulative probability density of the input signal. See Figure 3. Figure 3. The response distribution is obtained by multiplying the input distribution by the transfer function. When the transfer function matches the cumulative probability density of the input, the output distribution is uniform and information transfer is maximized.

3 Simon Laughlin was one of the first people to explore this issue in neural response functions, and he coined the term infomax. Figure 4a shows an estimate of the probability of contrasts measured in the fly environment. Figure 4b shows the cumulative probability density of contrast in the fly s environment, and it is the response function that is predicted if the fly s visual system is performing information maximization. Superimposed is the observed transfer function. Green dots show the observed response of the fly LMC cells as a function of contrast. There is a close match between the observed transfer function and that predicted by information maximization. a. b. Figure 4. Information maximization in blowfly compound eye. Left: Estimated probability density of contrast in fly environment. Right: Solid line shows response function predicted by information maximization. Green dots give observed response of Large Monopolar Cells (LMC). From Laughlin (1981). There are numerous other examples of information maximization in neural codes (See Simoncelli and Olshausen, 2001, for a review). For example, Atick and Redlich (1991) showed this for contrast sensitivity as a function of spatial frequency in both Macaque retinal ganglion cells and psychophysical contrast sensitivity functions. Macleod and von der Twer have shown that information maximization of chromatic signals also account for spectral sensitivities in the primate color opponent system as well. Infomax principles have also been demonstrated in the temporal frequency response in cat LGN (Dan et al., 1996), primary visual cortex (Vinge & Gallant, 2000). It has also been shown that short-term adaptation reduces dependencies between neurons both physiologically (e.g. Carandini, 1998) and using psychophysics (e.g. Webster, 1996). Information theory: Here we review some basic concepts from information theory that are behind the information maximization principle. I(x) = -log(p(x)) (1) The information I in a signal x is inversely proportional to its probability. Low probability events contain more information. For example, if you tell me that you spoke to a student in your class who was 20 years old, that would not tell me much about who it was, but if you tell me you spoke to a student who was 43, that would give me more information about who it was. H(x) = E[I(x)] (2)

4 The entropy H is the expected value of the information over the whole probability distribution. Entropy is maximized by flat distributions and minimized by peaked distributions. The mutual information between an input x and an output y, I(x,y), is the expected value of the joint probability of x and y, but it can be expressed as the entropy of Y minus the entropy of y given x. I(x,y) = H(y) H(y x) (3) The entropy of y given x is the noise, since it is the information in the signal that is not accounted for by the input. Therefore, maximizing I(x,y) is equivalent to maximizing H(y). In other words, maximizing information transfer is equivalent to maximizing the entropy of the output. Hebbian Learning. These concepts relate to learning at the neuron level. Hebbian learning is a model for long-term potentiation in neurons, in which weights are increased when the input and output are simultaneously active. The weight update rule is typically formulated as the product of the input and the output activations. This simple rule learns the covariance structure of the input (i.e. the second-order relations). In the case of a single input and output unit, Hebbian learning maximizes the information transfer between the input and the output (Linkser, 1988). Hence long term potentiation is related to information maximization. For multiple inputs and outputs, however, Hebbian learning doesn t maximize information transfer except in the special case where all of the signal distributions are Gaussian. Many types of natural signals have been shown to have non-gaussian distributions, in which the distributions are much more steeply peaked (e.g. Lewicki, 2002). Independent component analysis. The independent component analysis learning rule developed by Bell & Sejnowski (1995) is a generalization of Linsker s information maximization principle to the multiple unit case, X=(x 1 x n ), Y = (y 1 y n ). In this case, the information transfer between the input X and output Y is maximized by maximizing the joint entropy of the output, H(Y). Maximizing H(Y) can minimize the mutual information between the individual outputs Y = (y 1 y n ). This can be seen as follows: The equation for the joint entropy of the output Y is the sum of the individual entropies minus the mutual information between them: H(y 1,y 2 ) = H(y 1 )+H(y 2 ) I(y 1,y 2 ). (4) Maximizing H(y 1,y 2 ) encourages I(y 1,y 2 ) to be small. The mutual information is guaranteed to be minimized when each transfer function matches the cumulative probability density of the corresponding source (Parga). So the take-home message is that information maximization and redundancy reduction are highly related. The independent component analysis learning rule has a Hebbian learning term, but instead of being a straight product between the input and the output activations, the Hebbian learning is between the input and the gradient of the output: W = α [(W T ) -1 + y x T ) (5)

5 where W is a weight matrix and α is the learning rate. 1 Sparse codes. For maximizing information with a single unit, the optimal output distribution is flat or uniform, but when maximizing information in a population code, the output distribution of the individual units is different. ICA maximizes the joint entropy of the output, but minimizes the marginal entropies (the entropy of the individual outputs). So in a population code, infomax makes the joint output distribution flat, but the distributions of the individual outputs are highly peaked, or sparse. See Figure 5. Figure 5. In a population code, a maximum entropy distribution (solid line), can be created from individual neurons with sparse distributions (dashed lines). This falls out of the equation for joint entropy, which is repeated here. The independent component solution to equation 4 which minimizes I(y 1,y 2 ) also minimizes H(y 1 ) and H(y 2 ). (We hold H(y 1,y 2 ) constant and minimize I(y 1,y 2 ). This in turn minimizes H(y 1 ) and H(y 2 ).) H(y 1,y 2 ) = H(y 1 )+H(y 2 ) I(y 1,y 2 ). Barlow (1989) advocated minimum entropy coding for this reason, where the individual neurons have minimum entropy, or sparse distributions. It can be shown that maximizing sparseness without information loss is equivalent to ICA and infomax (Barlow; Olshaussen & Field). (Again we hold H(y 1,y 2 ) constant, and this time minimize H(y 1 ) and H(y 2 ), which in turn minimizes I(y1,y2).) Neurons early in the visual system may do infomax within a neural response, whereas neurons higher in the visual system may do infomax in population codes. Certainly there is a trend for neural responses to become more sparse higher in the visual system (e.g. Rolls). Relationships of ICA to V1 receptive fields. A number of relationships have been shown between independent component analysis and the response properties of visual cortical neurons. For example, applying ICA to a set of natural images produces a set of learned weights that are local, spatially opponent edge filters (Bell & Sejnowski, 1996: Olshaussen & Field, 1996). Conversely, passing a set of natural scenes through a bank of Gabor filters produces outputs that are at least pairwise independent when you also include gain control mechanisms such as the ones proposed by David Heeger to model V1 responses (Field, 1991; Simoncelli, 1997). When ICA is 1 Another version of this learning rule employs the natural gradient, which is the gradient multiplied by W T W (Bell & Sejnowski, 1996), which has important regularization properties. Here the Hebbian learning term is between the input and the natural gradient of the output. W = α (I + y x T W T )W.

6 applied to chromatic natural scenes, the set of weights segments into color-opponent and broadband filters, where the color-opponent filters code for red-green and blue yellow (Wachtler, Lee & Sejnowski, 2001). Moreover, a two-layer ICA model learned translation invariant receptive fields related to complex cell responses (Hyvarinen & Hoyer, 2001). Application to face recognition by computer These learning principles have been applied to face images for face recognition. Eigenfaces (Turk & Pentland, 1991; Cottrell & Metcalfe, 1991) is essentially an unsupervised learning strategy that learns the second-order dependencies among the image pixels. It applies principal component analysis (PCA) to a set of face images. Principal component solutions can be learned in neural networks with simple Hebbian learning rules (e.g. Oja 1989). Hence one way to interpret Eigenfaces, albeit not the way it is usually presented in the computer vision literature, is that it applies a Hebbian learning rule to a set of image pixels. a b Figure 6. Eigenfaces. a: Example face space in which the first 3 pixel values of 15 face images is plotted The first two eigenvectors are depicted by dashed lines. b: 4 eigenvectors of a 60x90 pixel face space. Each image is a point in a high dimensional space defined by the grayvalue taken at each of the pixel locations. See Figure 6. Principal component analysis rotates the axes to point in the directions of maximum variance of the data. The faces are recoded by their coordinates with respect to the new axes. These coordinates are the a i in Figure 6. The axes are defined by the eigenvectors of the covariance matrix. These eigenimages form a basis set where each face is a weighted combination of the basis images. The new coordinates, or weights, form a representation on which identity classification is performed. This approach took the computer vision community by surprise in the early 1990 s. It performed considerably better than contemporary approaches that focused on measuring specific facial features and the geometric relationships between them. The approach seemed counterintuitive many in computer vision. Why should this representation give a good description of face images? The success of the approach may have been due to the fundamental strategy of learning the dependencies in a population of face images. Principal component analysis learns the second-order dependencies among the pixels but does not learn high order dependencies. For example, if we change the pixels so that the covariances stay the same, then PCA gives the same answer. Bartlett, Movellan, & Sejnowski (2002) developed a representation for face recognition based on independent component analysis. Independent component analysis learns the high order

7 dependencies in addition to the ones that PCA learns. We compared it to eigenfaces to see if encoding more of the dependencies would result in better recognition performance. Second order vs. high-order dependencies in face images. The second-order dependencies correspond to the amplitude spectrum of an image, and the high order dependencies correspond to the phase. More precisely, a theorem by Einstein and Wiener showed that the Fourier transform of the covariance matrix of an image it corresponds to the amplitude spectrum of the image. What is left over is the phase. Figure 7 illustrates the contribution of high order dependencies to face perception. Experiments such as those by Oppenheim and Lim suggest that the phase spectrum contains the structural information that drives drives human perception. If you do a Fourier transform on the face images in Figure 7, and then reconstruct a face using the amplitude spectrum of the male but the phase spectrum of the female, the face looks like the female. Likewise, if we use his phase spectrum and her power spectrum the image looks like him. a b Figure 7. Second-order (amplitude) and high-order (phase) information and its effect on face perception. Bartlett, Movellan, & Sejnowski described two ways to apply ICA to face images. One way is similar to the PCA example illustrated above. Each image is an observation in a high dimensional space where the dimensions are the pixels. (See Figure 8a). It is also possible to treat the images as observations and the pixels as the variables. Each pixel is plotted according to the grayvalue it takes on over a set of face images. (See Figure 8b). a b Figure 8. Applying ICA to face images. a. Architecture I. b. Architecture II.

8 The ICA model decomposes images as X=AS where A is a mixing matrix and S is a matrix of independent sources. In Archietecture I, S contains the basis images, and A contains the coefficients. Whereas in Architecture II, S contains the coefficients and A contains the basis images. The basis images learnedin Architecture I are spatially local, whereas the basis images learned in Architecture II are global, or configural. (See Figure 9). Figure 9. Image synthesis model for the two architectures. The ICA model decomposes images as X=AS, where A is a mixing matrix and S is a matrix of independent sources. In Architecture I (top), the independent sources S are basis images and A contains the coefficients, whereas in Architecture II (bottom) the independent sources S are face codes (the coefficients) and A contains the basis images. Bartlett, Movellan and Sejnowski (2003) compared face recognition performance of ICA to PCA on a set of FERET face images. This image set contained 425 individuals with up to four images each: Same day with a change of expression, a different day up to 2 years later, and a different day with a change of expression. Recognition was done by nearest neighbor using cosines as the similarity measure. The results are shown in Figure 10. Both ICA face representations outperformed PCA for the pictures taken on a different day, which are the much harder recognition conditions..

9 Figure 10. Face recognition performance on the FERET database. ICA architctures I and II are compared to PCA (eigenfaces). Gray bars show improvement in performance following classspecific subselection of basis dimensions. The ICA representations maximize information transmission in the presence of noise and thus they may be more rubust to sources of noise such as variations in lighting conditions, facial expressions, and changes in hair and make-up. When subsets of bases are selected, ICA and PCA define different subspaces. Bartlett, Movellan, & Sejnowski examined face recognition performance following subspace selection with both ICA and PCA. Bases were selected by class discriminability, which we defined as the ratio of the variance within faces to the variance between faces. The gray extensions in Figure 10 show the improvement by subspace selection. ICA defined subspaces encoded more information about facial identity than PCA-defined subspaces. We also explored subselection of PCA bases ordered by eigenvalue. ICA continued to outperform PCA. Moreover, combining the outputs of the two ICA representations gave better performance then either one alone. The two representations appeared to capture different information about the face images. These findings suggest that information maximization principles are an effective coding strategy for face recognition by computer. Namely, the more dependencies that were encoded, the better the recognition performance. Local representations versus factorial codes. Draper and colleagues (2003) conducted a comparison of ICA and PCA on a substantially larger set of FERET face images consisting of 1196 individuals. This study supported the finding that ICA outperformed PCA, and included a change in lighting condition which we had not previously tested. ICA with architecture II obtained 51% accuracy on 192 probes with changes in lighting, compared to the best PCA performance of 40% correct. An interesting finding to emerge from the Draper study is that the ICA representation with Architecture II outperformed Architecture I for identity recognition. See Figure 11. According to arguments by Barlow, Atick, and Field (Barlow, 1989; Atick, 1992a; Field, 1994), the sparse, factorial properties of the representation in Architecture II should be optimal for face coding. Architecture II provides a factorial face code, in that each element of the face representation is independent of the others (i.e. the coefficients are independent). Although the previous study showed no significant difference in recognition performance for the two architectures, there may have been insufficient training data for a difference to emerge. Bartlett (2001) predicted that The ICA-factorial representation may prove to have a greater advantage given a much larger training set of images. Indeed, this prediction was born out. When the task was changed to recognition of facial expressions, however, Draper et al found that the ICA representation from Architecture I outperformed the ICA representation from Architecture II. The task was to recognize 6 facial actions, which are individual facial movements approximately corresponding to individual facial muscles. Draper et al attributed their pattern of results to differences in local versus global processing requirements of the two tasks. Architecture I defines local face features whereas Architecture II defines more configural face features. A large body of literature in human face processing points to the importance of configural information for identity recognition, whereas the facial expression recognition task in this study may have greater emphasis on local information. This speaks

10 to the issue of separate basis sets for expression and identity. There is evidence in functional neuroscience for separate processing of identity and expression in the brain (e.g. Haxby, 2000.) Here we obtain better recognition performance when we define different basis sets in which we switch what is treated as an observation versus what is treated as an independent variable for the purposes of information maximization. Figure 11. Face recognition performance with a larger image set from Draper et al. (2003). Dependency learning and face perception There are a number of perceptual studies that support the relevance of dependency encoding to human face perception. A large body of work showed that unsupervised learning of 2 nd order dependencies successfully models a number of aspects of human face perception including similarity, typicality, recognition accuracy, and other-race effects (e.g. O Toole et al., 1994; Hancock et al., 1998; Calder). Moreover, one study found that ICA better accounts for human judgments of facial similarity than PCA, supporting the idea that the more dependencies are encoded, the better the model of human perception for some tasks (Hancock, 2000). There is also support from neurophysiology for information maximization principles in face coding. The response distributions of IT face cells are sparse and there is very little redundancy between cells (Rolls, 2004; Rolls & Tovee, 1995). A perceptual discrimination study by Parraga, Troscianko, and Tollhurst (2000) supported information maximization in object perception. Sensitivity to small perturbations from morphing was highest for pictures with a natural second-order statistics, and degraded as the second order statistics were made less natural. Adaptation aftereffects in face perception have also been demonstrated (e.g. Webster, 1999; Leopold, O Toole, Vetter, & Blantz, 2001). For example, after adapting to a distorted face, a neutral face appears distorted in the opposite direction. This is a high-level version of the adaptation processes discussed earlier in this paper. Similar effects have been shown for race, gender, and facial expressions (Kaping et al., 2002). Hence it appears that faces may be coded according to expected values along these dimensions. Interestingly, adaptation to a nondistorted face does not make the distorted face appear more distorted. This is consistent with an information maximization account of adaptation since adapting to a neutral face would not shift the population mean.

11 Tanaka and colleagues showed a related effect with typical and atypical faces, where instead of adaptation, it was life experience that was shaping the probability distributions of faces (Tanaka, Giles, Kreeman, & Simon, 1998). The Tanaka study showed that morphs between typical and atypical parents appear to be more similar to the atypical parent. Figure 12 shows an example 50/50 morph between typical and atypical faces. In a 2-alternative forced choice, subjects chose the atypical parent as more similar about 60% of the time. Typical 50% morph Atypical Figure 12. Sample morph between typical and atypical parent faces, from Tanaka, Giles, Kreeman, & Simon, (1998). Tanaka et al suggested an attractor network account for this effect, where the atypical parent has a larger basin of attraction than the typical parent. See Figure XX. Bartlett & Tanaka implemented an attractor network model of face perception, and indeed successfully modeled the atypicality bias. Inspection of this model provides further insights about the potential role of redundancy reduction in face perception. In this model, a decorrelation step is necessary in order to encode highly similar patterns in a Hopfield attractor network. Decorrelation was necessary in order for each face to take on a distinct pattern of sustained activity. This Hence the atypicality bias arose from a decorrelation process inherent to producing separate internal representations for highly similar faces. An alternative account of the atypicality bias is provided by information maximization and redundancy reduction. This account is compatible with the attractor network hypothesis but more parsimonious. Figure 13a shows the probability distribution of face property X such as nose size. A typical face has a value near the mean for this property, and an atypical face has a value that is in the tail. Figure 13b shows the transfer function between signal value and percept. Let us suppose it is adapted to match the cumulative probability density so that it performs information maximization. In this case, typical faces fall on a highly sloping region of the transfer function, whereas atypical faces fall on a flatter region. The 50% morph is mapped closer to the atypical face because of the shape of the transfer function. The infomax account makes an additional prediction. Although it is well known that faces rated as atypical tend to also be easier to recognize, this model predicts that subjects will be less sensitive to physical changes in the physical properties of atypical faces. Perceptual effects such as the other-race effect fall out of the same information maximization model. In Figure 13, simply replace typical and atypical with same race and other race. Since subjects have more experience with same-race faces, the physical properties of same-race faces would fall near the center of the probability density function. The infomax transfer function would be steeper near same-race stimulus properties, and shallower near other race stimulus properties. Walker and Tanaka (in press) recently showed that discrimination performance is superior for same-race than other-race faces, which is consistent with information maximization coding.

12 a b Figure 13. Information maximization account of the atypicality bias in the perception of morphed faces. a. Example probability distribution for a face property such as nose size, in which there is higher probability density near typical faces. b. Perceptual transfer function predicted by information maximization (the cumulative density of the distribution on the left). The percept of the 50% morph is mapped closer to the atypical face because of the shape of the transfer function. (See dashed lines). Information maximization account of adaptation aftereffects. This account also applies to face adaptation aftereffects. The neural mechanisms may not be the same, they operate over different time scales, but the information maximization principle may apply to both cases. Consider, for example, that the adapting face is a distorted face in the tail of the face distribution for face feature X such as nose size. See Figure 14a. After adapting, the short-term estimate of probability has increased density near the adapting stimulus, as shown by the dashed curve. Figure 14b shows the transfer functions predicted by information maximization for both pre- and post-adaptation probability densities. These are the cumulative probability density functions. Note that after adaptation, the neutral face is mapped to a distortion in the opposite direction of the adapting stimulus, which matches the psychophysical findings (e.g. Webster 1999). The increased slope near the adapting stimulus also predicts increased sensitivity near the adapting stimulus. An alternative possibility is that neural mechanisms do not implement unconstrained optimization such as Figure 14b, but that they do constrained optimization the with a family of functions such as a sigmoid. This possibility is illustrated in Figure 15a. Here a sigmoid was fitted to the cumulative probability density of the adaptation distribution using logistic regression. This model also predicts repulsion of the neutral face away from the adapting stimulus. These two models, however, give differing predictions on sensitivity changes, shown in Figure 15b. Sensitivity predictions were obtained by taking the first derivative of the transfer function. Unconstrained information maximization predicts increases in sensitivity near the adapting stimulus. The sigmoid fit to the optimal transfer function predicts a much smaller increase in sensitivity near the adapting stimulus, plus an overall shift in sensitivity towards the adapting stimulus.

13 a b Figure 14. a. Example probability density of a face property such as nose size before and after adaptation. Adaptation to a face with large noses changes the short-term probability density of nose size. Extra density is added near the adapting stimulus, and density decreases elsewhere since it must sum to one. B. Perceptual transfer function predicted by information maximization. (the cumulative pdf of the density following adaptation). a b Figure 15. a: approximation of the optimal transfer function following adaptation with a sigmoid. b. Sensitivity functions predicted by the two transfer functions shown in Figure XX and XX. Sensitivity predictions were obtained by taking the first derivative of the transfer function. Information maximization (red --) predicts increased sensitivity near the adapting stimulus. Alternatively, a sigmoid fit to the optimal transfer function (green -.-) predicts a shift in the sensitivity curve towards the adapting stimulus. Infomax in the spatiotemporal domain. Carrying these learning strategies into the spatiotemporal domain can help learn visual invariances. Viewpoint invariant representations can be obtained by learning temporal relationships between images as an object moves in the environment. There are a number of models of learning invariances from spatiotemporal dependencies (e.g. Foldiak, 1991; O Reilly, 1992; Becker 1999, Wallis & Rolls, 1997; Stone,

14 1996; Bartlett & Sejnowski, 1998). For example, Bartlett & Sejnowski (1998) modeled the development of representations that are independent of head pose by learning spatiotemporal relationships. Moreover, spatio-temporal ICA on movies of natural images results in space/time varying filters such as those in MT (Dan). The long open time of the NMDA channel provides a possible mechanism for learning spatiotemporal relationships in long-term potentiation (O Reilly, 1992). There is evidence that cells in the primate anterior temporal lobe can learn temporal relationships between visual inputs (Miyashita, 1988). After monkeys were presented a sequence of fractal patterns for a number of weeks, the AIT cells developed tuning curves for the position of the stimuli in the sequence. Responses to neighboring stimuli were correlated, and correlation reduced as the distance in the sequence increased. Summary Dependency coding and information maximization appear to be central principles in neural coding early in the visual system. Neural systems with limited dynamic range can increase the information that the response gives about the signal by placing the more steeply sloped portions of the transfer function in the regions of highest density, and shallower slopes at regions of low density. The function that maximizes information transfer is the one that matches the cumulative probability density of the input. There is a large body of evidence that neural codes in vision and other sensory modalities match the statistical structure of the environment, and hence maximize information about environmental signals to a degree. See (Simoncelli, 2001) for a review. This paper described how these principles may be relevant to how we think about higher visual processes such as face recognition as well. Here we examined algorithms for face recognition by computer from a perspective of information maximization. Principal component solutions can be learned in neural networks with simple Hebbian learning rules (Oja, 1989). Hence the Eigenface approach can be considered a form of Hebbian learning model, which performs information maximization under restricted conditions. Independent component analysis performs information maximization under more general conditions. The learning rule contains a Hebbian learning term, but it is between the input and the gradient of the output. Face representations derived from ICA gave better recognition performance than face representations based on PCA. This suggests that information maximization in early processing is an effective strategy for face recognition by computer. Next, we presented an information maximization account of perceptual effects including otherrace effects and typicality effects, and showed how face adaptation aftereffects (e.g. Kaping, 2002; Ng, 2003; Webster, 1999) are consistent with information maximization on short time scales. Two models of information maximization in adaptation were presented, one in which the visual system learns a high-order transfer functions that match the curves in the cumulative probability density, and another in which the cumulative probability density is approximated with the closest fitting sigmoid. These two models gave different predictions for sensitivity postadaptation. The extent to which information maximization principles account for perceptual learning and adaptation aftereffects in human face perception is an open question and an area of active research.

15 References Abbonizio, G., Langley, K. & Clifford, C.W.G. (2002). Contrast adaptation may enhance contrast discrimination. Spatial Vision, 16, Atick, J.J. (1992). Could information theory provide an ecological theory of sensory processing? Network, 3, Atick, J.J., & Redlich, A.N. (1992). What does the retina know about natural scenes? Neural Computation, 4, Ball, K., & Sekuler, R. (1987). Direction-specific improvement in motion discrimination. Vision Res, 27 (6), Barlow, H. (1994). What is the computational goal of the neocortex? In: C. Koch (Ed.) Large scale neuronal theories of the brain (pp ). Cambridge, MA: MIT Press. Barlow, H.B. (1989). Unsupervised learning. Neural Computation, 1, Bartlett, M. (2001a). Face Image Analysis by Unsupervised Learning. (Norwell, Massachusetts: Kluwer Academic Publishers. Bartlett, M.S. (2001b). Face Image Analysis by Unsupervised Learning. 612 (Boston: Kluwer Academic Publishers. Bartlett, M.S., Littlewort, G., Lainscsek, C., Fasel, I., & Movellan, J.R. (October 2004). Machine Learning Methods for Fully Automatic Recognition of Facial Expressions and Facial Actions. In: IEEE International Conference on Systems, Man & Cybernetics (The Hague, Netherlands. Bartlett, M.S., Movellan, J.R., & Sejnowski, T.J. (2002). Image representations for facial expression recognition. IEEE transactions on neural networks, 13 (6), Bartlett, M.S., & Sejnowski, T.J. (1998). Learning viewpoint invariant face representations from visual experience in an attractor network. Network, 9 (3), Bartlett, M.S., & Tanaka, J.W. (1998). An attractor field model of face representations: Effects of typicality and image morphing. In: Psychonomics Society Satellite Symposium on Object Perception and Memory (OPAM) Bell, A.J., & Sejnowski, T.J. (1995). An information-maximization approach to blind separation and blind deconvolution. Neural Computation, 7(6), Bell, A.J., & Sejnowski, T.J. (1997). The independent components of natural scenes are edge filters. Vision Research, 37 (23),

16 Bex, P.J., Bedingham, S., & Hammett, S.T. (1999). Apparent speed and speed sensitivity during adaptation to motion. Journal of the Optical Society of America A, 16 (12), Carandini, M.M., J., & Ferster, D. (1998). Pattern adaptation and cross orientation interactions in the primary visual cortex. Neuropharmacology, 37, Chaparro, A., Stromeyer, C.F., 3rd, Huang, E.P., Kronauer, R.E., & Eskew, R.T., Jr. (1993). Colour is what the eye sees best. Nature, 361 (6410), Clifford, C.W., & Wenderoth, P. (1999). Adaptation to temporal modulation can enhance differential speed sensitivity. Vision Res, 39 (26), Cottrell, G., Dailey, M., Padgett, C., & R, A. (2000). Computational, Geometric, and Process Perspectives on Facial Cognition: Contexts and Challenges. (Erlbaum. Cover, T.M., & Thomas, J.A. (1991). Elements of information theory. (New York: John Wiley \& Sons. Dan, Y., Atick, J. & Ried R. (1996). Efficient coding of natural scenes in the lateral geniculate nucleus: experimental Tests of a computaitonal theory. Journal of Neuroscience, 16, Donato, G., Bartlett, M., Hager, J., Ekman, P., & Sejnowski, T. (1999). Classifying Facial Actions. IEEE Transactions on Pattern Analysis and Machine Intelligence, 21 (10), Durgin, F.H., Gigone, K.M., & Schaffer, E. (2004). Improved visual speed discrimination while walking [abstract]. Vision Sciences Society Meeting, G109. Fasel, I.R., Fortenberry, B., & Movellan, J.R. (in press). {GB}oost: {A} Generative Framework for Boosting with Applications to Real-Time Eye Coding. Computer Vision and Image Understanding Field, D.J. (1987). Relations between the statistics of natural images and the response properties of cortical cells. Journal of the Optical Society of America, A, 4, Greenlee, M.H., F. (1988). The functional role of contrast adaptation. Vision Research, 28, Gros, B.L., Blake, R., & Hiris, E. (1998). Anisotropies in visual motion perception: a fresh look. J Opt Soc Am A Opt Image Sci Vis, 15 (8), Hancock, P. (2000). Alternative representations for faces. In: British Psychological Society, Cognitive Section (University of Essex, September 6-8.

17 Hancock, P.J.B., Burton, A.M., & Bruce, V. (1996). Face processing: human perception and principal components analysis. Memory and Cognition, 24, Hurvich, L.M. (1972). Color vision deficiencies. In: Handbook of Sensory Physiology, VII/4 (pp ). Berlin: Springer Verlag. Kaping, D., Duhamel, P., & Webster, M. (2002). Adaptation to Natural Face Categories. Journal of Vision, 2 (10), 128. Krukowski, A.E., Pirog, K.A., Beutter, B.R., Brooks, K.R., & Stone, L.S. (2003). Human discrimination of visual direction of motion with and without smooth pursuit eye movements. J Vis, 3 (11), Laughlin, S. (1981). A simple coding procedure enhances a neuron's information capacity. Z. Naturforsch, 36, Levinson, E., & Sekuler, R. (1976). Adaptation alters perceived direction of motion. Vision Res, 16 (7), Littlewort, G., Bartlett, M.S., Fasel, I., Susskind, J., & Movellan, J.R. (2004). Dynamics of Facial Expression Extracted Automatically from Video. In: IEEE Conference on Computer Vision and Pattern Recognition, Workshop on Face Processing in Video Loomis, J.M., & Berger, T. (1979). Effects of chromatic adaptation on color discrimination and color appearance. Vision Res, 19 (8), Macleod, D., & Twer, T.v.d. (1996). Optimal nonlinear codes Technical Report 28/96. (Universit{\"a}t Bielefeld, Zentrum f{\"u}r interdisziplin{\"a}re Froschung. MacLeod, D.I.A. (2003). Colour discrimination, colour constancy and natural scene statistics (The Verriest lecture). In: J.D. Mollon, J. Pokorny, & K. Knoblauch (Eds.), Normal & Defective Colour Vision (New York: Oxford University Press. Miyahara, E., Smith, V.C., & Pokorny, J. (1993). How surrounds affect chromaticity discrimination. J Opt Soc Am A, 10 (4), Neitz, J., Carroll, J., Yamauchi, Y., Neitz, M., & Williams, D.R. (2002). Color perception is mediated by a plastic neural mechanism that is adjustable in adults. Neuron, 35 (4), Ng, M., Kaping, D., Webster, M.A., Anstis, S., & Fine, I. (2003). Selective tuning of face perception [Abstract]. Journal of Vision, 3 (9), 106a. O`Toole, A., Deffenbacher, K., Valentin, D., & Abdi, H. (1994). Structural aspects of face recognition and the other race effect. Memory and Cognition, 22 (2),

18 Pickford, R.W. (1958). A review of some problems of colour vision and colour blindness. Advancement of Science, 15, Pokorny, J., & Smith, V.C. (1977). Evaluation of single-pigment shift model of anomalous trichromacy. Journal of the Optical Society of America, 67, Pokorny, J., Smith, V.C., & Verriest, G. (1979). Congenital Color Defects. In: J. Pokorny, V.C. Smith, G. Verriest, & A. Pinckers (Eds.), Congenital and Acquired Color Vision Defects. (New York: Grune and Stratton. Regan, D.B., K. (1985). Postadaptation Orientation Discrimination. Journal of the Optical Society of America A, 2, Rolls, E.T., Aggelopoulos,N.C., Franco,L., and Treves,A. (2004). Information encoding in the inferior temporal cortex: contributions of the firing rates and correlations between the firing of neurons. Biological Cybernetics, 90, Rolls, E.T., & Tovee, M.J. (1995). Sparseness of the neuronal representation of stimuli in the primate temporal visual cortex. Journal of Neurophysiology, 73(2), Schrater, P.R., & Simoncelli, E.P. (1998). Local velocity representation: evidence from motion adaptation. Vision Res, 38 (24), Sekuler, R., Watamaniuk, S.N.J., & Blake, R. (2002). Visual motion perception. In: S. Yantis (Ed.) Stevens' Handbook of Experimental Psychology: Vol. 1. Sensation and perception (New York: Wiley. Simoncelli, E.O., B. (2001). Natural Image Statistics and Neural Representation. Annual Review of Neuroscience, 24, Simoncelli, E.P. (November ). Statistical Models for Images: Compression, Restoration and Synthesis. In: 31st Asilomar Conference on Signals, Systems and Computers (Pacific Grove, CA. Tanaka, J.W., Giles, M., Kremen, S., & Simon, V. (1998). Mapping attractor fields in face space: The atypicality bias in face recognition. Cognition, 68 (3), Thornton, J.E., & Pugh, E.N., Jr. (1983). Red/Green color opponency at detection threshold. Science, 219 (4581), Verriest, G. (1960). Contribution a l'étude de la correlation entre la quotient d anomalie et l étendue des confusions colorimétriques dan les systèmes trichromatiques anormaux. Revue d'optique, 39, Vinje, W.G., J. (2000). Sparse coding and decorrelation in primary visual cortex during natural vision. Science, 287,

19 von der Twer, T., & MacLeod, D.I. (2001). Optimal nonlinear codes for the perception of natural colours. Network, 12 (3), Webster, M. (1996). Human colour perception and its adaptation. Network: Computation in Neural Systems, 7, Webster, M., Werner, J. & Field, D. (in press). Adaptation in the phenomenology of perception. In: C.C.G. Rhodes (Ed.) Fitting the Mind to the World: Adaptation and Aftereffects in High Level Vision, 2 (Oxford University Press. Webster, M.A., & Mollon, J.D. (1997). Adaptation and the color statistics of natural images. Vision Res, 37 (23), Webster, M.A.M., J.D. (1994). Adaptation and the Color Statistics of Natural Images. Vision Research, 37, Webster, M.M., O. (1999). Figural Aftereffects in the Perception of Faces. Psychonomic Bulletin Review, 6 (4), Werner, J.S. (1998). Aging through the eyes of Monet. In: R.K. W. Backhaus, Werner, J. (Ed.) Color Vision: Perspectives from Different Disciplines (pp. 3-41). Berlin: Walter de Gruyter & Co. Willis, M.P., & Farnsworth, D. (1952). Comparative Evaluation of Anomaloscopes. (Bureau of Medicine and Surgery, U.S. Navy.

Information Maximization in Face Processing

Information Maximization in Face Processing Information Maximization in Face Processing Marian Stewart Bartlett Institute for Neural Computation University of California, San Diego marni@salk.edu Abstract This perspective paper explores principles

More information

This article was originally published in a journal published by Elsevier, and the attached copy is provided by Elsevier for the author s benefit and for the benefit of the author s institution, for non-commercial

More information

Visual discrimination and adaptation using non-linear unsupervised learning

Visual discrimination and adaptation using non-linear unsupervised learning Visual discrimination and adaptation using non-linear unsupervised learning Sandra Jiménez, Valero Laparra and Jesús Malo Image Processing Lab, Universitat de València, Spain; http://isp.uv.es ABSTRACT

More information

Morton-Style Factorial Coding of Color in Primary Visual Cortex

Morton-Style Factorial Coding of Color in Primary Visual Cortex Morton-Style Factorial Coding of Color in Primary Visual Cortex Javier R. Movellan Institute for Neural Computation University of California San Diego La Jolla, CA 92093-0515 movellan@inc.ucsd.edu Thomas

More information

lateral organization: maps

lateral organization: maps lateral organization Lateral organization & computation cont d Why the organization? The level of abstraction? Keep similar features together for feedforward integration. Lateral computations to group

More information

CN530, Spring 2004 Final Report Independent Component Analysis and Receptive Fields. Madhusudana Shashanka

CN530, Spring 2004 Final Report Independent Component Analysis and Receptive Fields. Madhusudana Shashanka CN530, Spring 2004 Final Report Independent Component Analysis and Receptive Fields Madhusudana Shashanka shashanka@cns.bu.edu 1 Abstract This report surveys the literature that uses Independent Component

More information

Sensory Adaptation within a Bayesian Framework for Perception

Sensory Adaptation within a Bayesian Framework for Perception presented at: NIPS-05, Vancouver BC Canada, Dec 2005. published in: Advances in Neural Information Processing Systems eds. Y. Weiss, B. Schölkopf, and J. Platt volume 18, pages 1291-1298, May 2006 MIT

More information

A Neurally-Inspired Model for Detecting and Localizing Simple Motion Patterns in Image Sequences

A Neurally-Inspired Model for Detecting and Localizing Simple Motion Patterns in Image Sequences A Neurally-Inspired Model for Detecting and Localizing Simple Motion Patterns in Image Sequences Marc Pomplun 1, Yueju Liu 2, Julio Martinez-Trujillo 2, Evgueni Simine 2, and John K. Tsotsos 2 1 Department

More information

Fitting the world to the mind: Transforming images to mimic perceptual adaptation

Fitting the world to the mind: Transforming images to mimic perceptual adaptation Fitting the world to the mind: Transforming images to mimic perceptual adaptation Michael A. Webster 1, Kyle McDermott 1, and George Bebis 2 Departments of Psychology 1 and Computer Science and Electrical

More information

Facial Memory is Kernel Density Estimation (Almost)

Facial Memory is Kernel Density Estimation (Almost) Facial Memory is Kernel Density Estimation (Almost) Matthew N. Dailey Garrison W. Cottrell Department of Computer Science and Engineering U.C. San Diego La Jolla, CA 92093-0114 fmdailey,garyg@cs.ucsd.edu

More information

OPTO 5320 VISION SCIENCE I

OPTO 5320 VISION SCIENCE I OPTO 5320 VISION SCIENCE I Monocular Sensory Processes of Vision: Color Vision Mechanisms of Color Processing . Neural Mechanisms of Color Processing A. Parallel processing - M- & P- pathways B. Second

More information

Neuronal responses to plaids

Neuronal responses to plaids Vision Research 39 (1999) 2151 2156 Neuronal responses to plaids Bernt Christian Skottun * Skottun Research, 273 Mather Street, Piedmont, CA 94611-5154, USA Received 30 June 1998; received in revised form

More information

Just One View: Invariances in Inferotemporal Cell Tuning

Just One View: Invariances in Inferotemporal Cell Tuning Just One View: Invariances in Inferotemporal Cell Tuning Maximilian Riesenhuber Tomaso Poggio Center for Biological and Computational Learning and Department of Brain and Cognitive Sciences Massachusetts

More information

Facial Memory is Kernel Density Estimation (Almost)

Facial Memory is Kernel Density Estimation (Almost) Facial Memory is Kernel Density Estimation (Almost) Matthew N. Dailey Garrison W. Cottrell Department of Computer Science and Engineering U.C. San Diego La Jolla, CA 92093-0114 fmdailey,garyg@cs.ucsd.edu

More information

Statistical and Neural Methods for Vision-based Analysis of Facial Expressions and Gender

Statistical and Neural Methods for Vision-based Analysis of Facial Expressions and Gender Proc. IEEE Int. Conf. on Systems, Man and Cybernetics (SMC 2004), Den Haag, pp. 2203-2208, IEEE omnipress 2004 Statistical and Neural Methods for Vision-based Analysis of Facial Expressions and Gender

More information

A Vision-based Affective Computing System. Jieyu Zhao Ningbo University, China

A Vision-based Affective Computing System. Jieyu Zhao Ningbo University, China A Vision-based Affective Computing System Jieyu Zhao Ningbo University, China Outline Affective Computing A Dynamic 3D Morphable Model Facial Expression Recognition Probabilistic Graphical Models Some

More information

The Integration of Features in Visual Awareness : The Binding Problem. By Andrew Laguna, S.J.

The Integration of Features in Visual Awareness : The Binding Problem. By Andrew Laguna, S.J. The Integration of Features in Visual Awareness : The Binding Problem By Andrew Laguna, S.J. Outline I. Introduction II. The Visual System III. What is the Binding Problem? IV. Possible Theoretical Solutions

More information

Vision Seeing is in the mind

Vision Seeing is in the mind 1 Vision Seeing is in the mind Stimulus: Light 2 Light Characteristics 1. Wavelength (hue) 2. Intensity (brightness) 3. Saturation (purity) 3 4 Hue (color): dimension of color determined by wavelength

More information

Reading Assignments: Lecture 5: Introduction to Vision. None. Brain Theory and Artificial Intelligence

Reading Assignments: Lecture 5: Introduction to Vision. None. Brain Theory and Artificial Intelligence Brain Theory and Artificial Intelligence Lecture 5:. Reading Assignments: None 1 Projection 2 Projection 3 Convention: Visual Angle Rather than reporting two numbers (size of object and distance to observer),

More information

Competing Frameworks in Perception

Competing Frameworks in Perception Competing Frameworks in Perception Lesson II: Perception module 08 Perception.08. 1 Views on perception Perception as a cascade of information processing stages From sensation to percept Template vs. feature

More information

Competing Frameworks in Perception

Competing Frameworks in Perception Competing Frameworks in Perception Lesson II: Perception module 08 Perception.08. 1 Views on perception Perception as a cascade of information processing stages From sensation to percept Template vs. feature

More information

Adaptation and perceptual norms

Adaptation and perceptual norms Adaptation and perceptual norms Michael A. Webster*, Maiko Yasuda, Sara Haber, Deanne Leonard, Nicole Ballardini Department of Psychology / 296, University of Nevada-Reno, Reno NV 89557, USA ABSTRACT We

More information

NATURAL IMAGE STATISTICS AND NEURAL REPRESENTATION

NATURAL IMAGE STATISTICS AND NEURAL REPRESENTATION Annu. Rev. Neurosci. 2001. 24:1193 216 Copyright c 2001 by Annual Reviews. All rights reserved NATURAL IMAGE STATISTICS AND NEURAL REPRESENTATION Eero P Simoncelli Howard Hughes Medical Institute, Center

More information

Natural Images: Coding Efficiency

Natural Images: Coding Efficiency Natural Images: Coding Efficiency 19 Natural Images: Coding Efficiency D J Graham and D J Field, Cornell University, Ithaca, NY, USA ã 2009 Elsevier Ltd. All rights reserved. Introduction Images and movies

More information

Associative Decorrelation Dynamics: A Theory of Self-Organization and Optimization in Feedback Networks

Associative Decorrelation Dynamics: A Theory of Self-Organization and Optimization in Feedback Networks Associative Decorrelation Dynamics: A Theory of Self-Organization and Optimization in Feedback Networks Dawei W. Dong* Lawrence Berkeley Laboratory University of California Berkeley, CA 94720 Abstract

More information

Lighta part of the spectrum of Electromagnetic Energy. (the part that s visible to us!)

Lighta part of the spectrum of Electromagnetic Energy. (the part that s visible to us!) Introduction to Physiological Psychology Vision ksweeney@cogsci.ucsd.edu cogsci.ucsd.edu/~ /~ksweeney/psy260.html Lighta part of the spectrum of Electromagnetic Energy (the part that s visible to us!)

More information

16/05/2008 #1

16/05/2008  #1 16/05/2008 http://www.scholarpedia.org/wiki/index.php?title=sparse_coding&printable=yes #1 Sparse coding From Scholarpedia Peter Foldiak and Dominik Endres (2008), Scholarpedia, 3(1):2984. revision #37405

More information

Subjective randomness and natural scene statistics

Subjective randomness and natural scene statistics Psychonomic Bulletin & Review 2010, 17 (5), 624-629 doi:10.3758/pbr.17.5.624 Brief Reports Subjective randomness and natural scene statistics Anne S. Hsu University College London, London, England Thomas

More information

Spectrograms (revisited)

Spectrograms (revisited) Spectrograms (revisited) We begin the lecture by reviewing the units of spectrograms, which I had only glossed over when I covered spectrograms at the end of lecture 19. We then relate the blocks of a

More information

Analysis of in-vivo extracellular recordings. Ryan Morrill Bootcamp 9/10/2014

Analysis of in-vivo extracellular recordings. Ryan Morrill Bootcamp 9/10/2014 Analysis of in-vivo extracellular recordings Ryan Morrill Bootcamp 9/10/2014 Goals for the lecture Be able to: Conceptually understand some of the analysis and jargon encountered in a typical (sensory)

More information

Position invariant recognition in the visual system with cluttered environments

Position invariant recognition in the visual system with cluttered environments PERGAMON Neural Networks 13 (2000) 305 315 Contributed article Position invariant recognition in the visual system with cluttered environments S.M. Stringer, E.T. Rolls* Oxford University, Department of

More information

Facial Expression Recognition Using Principal Component Analysis

Facial Expression Recognition Using Principal Component Analysis Facial Expression Recognition Using Principal Component Analysis Ajit P. Gosavi, S. R. Khot Abstract Expression detection is useful as a non-invasive method of lie detection and behaviour prediction. However,

More information

Plasticity of Cerebral Cortex in Development

Plasticity of Cerebral Cortex in Development Plasticity of Cerebral Cortex in Development Jessica R. Newton and Mriganka Sur Department of Brain & Cognitive Sciences Picower Center for Learning & Memory Massachusetts Institute of Technology Cambridge,

More information

Significance of Natural Scene Statistics in Understanding the Anisotropies of Perceptual Filling-in at the Blind Spot

Significance of Natural Scene Statistics in Understanding the Anisotropies of Perceptual Filling-in at the Blind Spot Significance of Natural Scene Statistics in Understanding the Anisotropies of Perceptual Filling-in at the Blind Spot Rajani Raman* and Sandip Sarkar Saha Institute of Nuclear Physics, HBNI, 1/AF, Bidhannagar,

More information

Visual Object Recognition Computational Models and Neurophysiological Mechanisms Neurobiology 130/230. Harvard College/GSAS 78454

Visual Object Recognition Computational Models and Neurophysiological Mechanisms Neurobiology 130/230. Harvard College/GSAS 78454 Visual Object Recognition Computational Models and Neurophysiological Mechanisms Neurobiology 130/230. Harvard College/GSAS 78454 Web site: http://tinyurl.com/visionclass (Class notes, readings, etc) Location:

More information

Human colour perception and its adaptation

Human colour perception and its adaptation Network: Computation in Neural Systems 7 (1996) 587 634. Printed in the UK Human colour perception and its adaptation Michael A Webster Department of Psychology, University of Nevada, Reno, NV 89557-0062,

More information

Validating the Visual Saliency Model

Validating the Visual Saliency Model Validating the Visual Saliency Model Ali Alsam and Puneet Sharma Department of Informatics & e-learning (AITeL), Sør-Trøndelag University College (HiST), Trondheim, Norway er.puneetsharma@gmail.com Abstract.

More information

M Cells. Why parallel pathways? P Cells. Where from the retina? Cortical visual processing. Announcements. Main visual pathway from retina to V1

M Cells. Why parallel pathways? P Cells. Where from the retina? Cortical visual processing. Announcements. Main visual pathway from retina to V1 Announcements exam 1 this Thursday! review session: Wednesday, 5:00-6:30pm, Meliora 203 Bryce s office hours: Wednesday, 3:30-5:30pm, Gleason https://www.youtube.com/watch?v=zdw7pvgz0um M Cells M cells

More information

Information-theoretic stimulus design for neurophysiology & psychophysics

Information-theoretic stimulus design for neurophysiology & psychophysics Information-theoretic stimulus design for neurophysiology & psychophysics Christopher DiMattina, PhD Assistant Professor of Psychology Florida Gulf Coast University 2 Optimal experimental design Part 1

More information

Nonlinear processing in LGN neurons

Nonlinear processing in LGN neurons Nonlinear processing in LGN neurons Vincent Bonin *, Valerio Mante and Matteo Carandini Smith-Kettlewell Eye Research Institute 2318 Fillmore Street San Francisco, CA 94115, USA Institute of Neuroinformatics

More information

Self-Organization and Segmentation with Laterally Connected Spiking Neurons

Self-Organization and Segmentation with Laterally Connected Spiking Neurons Self-Organization and Segmentation with Laterally Connected Spiking Neurons Yoonsuck Choe Department of Computer Sciences The University of Texas at Austin Austin, TX 78712 USA Risto Miikkulainen Department

More information

Mammogram Analysis: Tumor Classification

Mammogram Analysis: Tumor Classification Mammogram Analysis: Tumor Classification Term Project Report Geethapriya Raghavan geeragh@mail.utexas.edu EE 381K - Multidimensional Digital Signal Processing Spring 2005 Abstract Breast cancer is the

More information

Early Stages of Vision Might Explain Data to Information Transformation

Early Stages of Vision Might Explain Data to Information Transformation Early Stages of Vision Might Explain Data to Information Transformation Baran Çürüklü Department of Computer Science and Engineering Mälardalen University Västerås S-721 23, Sweden Abstract. In this paper

More information

Group Redundancy Measures Reveal Redundancy Reduction in the Auditory Pathway

Group Redundancy Measures Reveal Redundancy Reduction in the Auditory Pathway Group Redundancy Measures Reveal Redundancy Reduction in the Auditory Pathway Gal Chechik Amir Globerson Naftali Tishby Institute of Computer Science and Engineering and The Interdisciplinary Center for

More information

ABSTRACT 1. INTRODUCTION 2. ARTIFACT REJECTION ON RAW DATA

ABSTRACT 1. INTRODUCTION 2. ARTIFACT REJECTION ON RAW DATA AUTOMATIC ARTIFACT REJECTION FOR EEG DATA USING HIGH-ORDER STATISTICS AND INDEPENDENT COMPONENT ANALYSIS A. Delorme, S. Makeig, T. Sejnowski CNL, Salk Institute 11 N. Torrey Pines Road La Jolla, CA 917,

More information

Introduction to Computational Neuroscience

Introduction to Computational Neuroscience Introduction to Computational Neuroscience Lecture 11: Attention & Decision making Lesson Title 1 Introduction 2 Structure and Function of the NS 3 Windows to the Brain 4 Data analysis 5 Data analysis

More information

ADAPTATION TO RACIAL CONTENT OF EMERGENT RACE FACES: GRADIENT SHIFT OR PEAK SHIFT?

ADAPTATION TO RACIAL CONTENT OF EMERGENT RACE FACES: GRADIENT SHIFT OR PEAK SHIFT? ADAPTATION TO RACIAL CONTENT OF EMERGENT RACE FACES: GRADIENT SHIFT OR PEAK SHIFT? Otto MacLin, Kim MacLin, and Dwight Peterson Department of Psychology, University of Northern Iowa otto.maclin@uni.edu

More information

Computational Cognitive Neuroscience

Computational Cognitive Neuroscience Computational Cognitive Neuroscience Computational Cognitive Neuroscience Computational Cognitive Neuroscience *Computer vision, *Pattern recognition, *Classification, *Picking the relevant information

More information

Continuous transformation learning of translation invariant representations

Continuous transformation learning of translation invariant representations Exp Brain Res (21) 24:255 27 DOI 1.17/s221-1-239- RESEARCH ARTICLE Continuous transformation learning of translation invariant representations G. Perry E. T. Rolls S. M. Stringer Received: 4 February 29

More information

Sparse Coding in Sparse Winner Networks

Sparse Coding in Sparse Winner Networks Sparse Coding in Sparse Winner Networks Janusz A. Starzyk 1, Yinyin Liu 1, David Vogel 2 1 School of Electrical Engineering & Computer Science Ohio University, Athens, OH 45701 {starzyk, yliu}@bobcat.ent.ohiou.edu

More information

A THEORY OF MCCOLLOUGH EFFECT AND. CHUN CHIANG Institute of Physics, Academia Sinica

A THEORY OF MCCOLLOUGH EFFECT AND. CHUN CHIANG Institute of Physics, Academia Sinica A THEORY OF MCCOLLOUGH EFFECT AND CONTINGENT AFTER-EFFECT CHUN CHIANG Institute of Physics, Academia Sinica A model is advanced to explain the McCollough effect and the contingent motion after-effect.

More information

arxiv: v1 [q-bio.nc] 13 Jan 2008

arxiv: v1 [q-bio.nc] 13 Jan 2008 Recurrent infomax generates cell assemblies, avalanches, and simple cell-like selectivity Takuma Tanaka 1, Takeshi Kaneko 1,2 & Toshio Aoyagi 2,3 1 Department of Morphological Brain Science, Graduate School

More information

Spontaneous Cortical Activity Reveals Hallmarks of an Optimal Internal Model of the Environment. Berkes, Orban, Lengyel, Fiser.

Spontaneous Cortical Activity Reveals Hallmarks of an Optimal Internal Model of the Environment. Berkes, Orban, Lengyel, Fiser. Statistically optimal perception and learning: from behavior to neural representations. Fiser, Berkes, Orban & Lengyel Trends in Cognitive Sciences (2010) Spontaneous Cortical Activity Reveals Hallmarks

More information

USING AUDITORY SALIENCY TO UNDERSTAND COMPLEX AUDITORY SCENES

USING AUDITORY SALIENCY TO UNDERSTAND COMPLEX AUDITORY SCENES USING AUDITORY SALIENCY TO UNDERSTAND COMPLEX AUDITORY SCENES Varinthira Duangudom and David V Anderson School of Electrical and Computer Engineering, Georgia Institute of Technology Atlanta, GA 30332

More information

Efficient Encoding of Natural Time Varying Images Produces Oriented Space-Time Receptive Fields

Efficient Encoding of Natural Time Varying Images Produces Oriented Space-Time Receptive Fields Efficient Encoding of Natural Time Varying Images Produces Oriented Space-Time Receptive Fields Rajesh P. N. Rao and Dana H. Ballard Department of Computer Science University of Rochester Rochester, NY

More information

Recognition of Faces of Different Species: A Developmental Study Between 5 and 8 Years of Age

Recognition of Faces of Different Species: A Developmental Study Between 5 and 8 Years of Age Infant and Child Development Inf. Child Dev. 10: 39 45 (2001) DOI: 10.1002/icd.245 Recognition of Faces of Different Species: A Developmental Study Between 5 and 8 Years of Age Olivier Pascalis a, *, Elisabeth

More information

Adventures into terra incognita

Adventures into terra incognita BEWARE: These are preliminary notes. In the future, they will become part of a textbook on Visual Object Recognition. Chapter VI. Adventures into terra incognita In primary visual cortex there are neurons

More information

Ch 5. Perception and Encoding

Ch 5. Perception and Encoding Ch 5. Perception and Encoding Cognitive Neuroscience: The Biology of the Mind, 2 nd Ed., M. S. Gazzaniga, R. B. Ivry, and G. R. Mangun, Norton, 2002. Summarized by Y.-J. Park, M.-H. Kim, and B.-T. Zhang

More information

Consciousness The final frontier!

Consciousness The final frontier! Consciousness The final frontier! How to Define it??? awareness perception - automatic and controlled memory - implicit and explicit ability to tell us about experiencing it attention. And the bottleneck

More information

Statistical Models of Natural Images and Cortical Visual Representation

Statistical Models of Natural Images and Cortical Visual Representation Topics in Cognitive Science 2 (2010) 251 264 Copyright Ó 2009 Cognitive Science Society, Inc. All rights reserved. ISSN: 1756-8757 print / 1756-8765 online DOI: 10.1111/j.1756-8765.2009.01057.x Statistical

More information

eye as a camera Kandel, Schwartz & Jessel (KSJ), Fig 27-3

eye as a camera Kandel, Schwartz & Jessel (KSJ), Fig 27-3 eye as a camera Kandel, Schwartz & Jessel (KSJ), Fig 27-3 retinal specialization fovea: highest density of photoreceptors, aimed at where you are looking -> highest acuity optic disk: cell-free area, where

More information

Supporting Information

Supporting Information 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 Supporting Information Variances and biases of absolute distributions were larger in the 2-line

More information

Functional Elements and Networks in fmri

Functional Elements and Networks in fmri Functional Elements and Networks in fmri Jarkko Ylipaavalniemi 1, Eerika Savia 1,2, Ricardo Vigário 1 and Samuel Kaski 1,2 1- Helsinki University of Technology - Adaptive Informatics Research Centre 2-

More information

Fundamentals of Psychophysics

Fundamentals of Psychophysics Fundamentals of Psychophysics John Greenwood Department of Experimental Psychology!! NEUR3045! Contact: john.greenwood@ucl.ac.uk 1 Visual neuroscience physiology stimulus How do we see the world? neuroimaging

More information

Ch 5. Perception and Encoding

Ch 5. Perception and Encoding Ch 5. Perception and Encoding Cognitive Neuroscience: The Biology of the Mind, 2 nd Ed., M. S. Gazzaniga,, R. B. Ivry,, and G. R. Mangun,, Norton, 2002. Summarized by Y.-J. Park, M.-H. Kim, and B.-T. Zhang

More information

2/3/17. Visual System I. I. Eye, color space, adaptation II. Receptive fields and lateral inhibition III. Thalamus and primary visual cortex

2/3/17. Visual System I. I. Eye, color space, adaptation II. Receptive fields and lateral inhibition III. Thalamus and primary visual cortex 1 Visual System I I. Eye, color space, adaptation II. Receptive fields and lateral inhibition III. Thalamus and primary visual cortex 2 1 2/3/17 Window of the Soul 3 Information Flow: From Photoreceptors

More information

Optimizing visual performance by adapting images to observers Michael A. Webster* a, Igor Juricevic b

Optimizing visual performance by adapting images to observers Michael A. Webster* a, Igor Juricevic b Optimizing visual performance by adapting images to observers Michael A. Webster* a, Igor Juricevic b a Dept. of Psychology/296, University of Nevada Reno, Reno, NV 89557, USA; b Dept. of Psychology, Indiana

More information

Adaptation and the Phenomenology of Perception

Adaptation and the Phenomenology of Perception Clifford-09.qxd 11/25/04 3:32 PM Page 241 9 Adaptation and the Phenomenology of Perception MICHAEL A. WEBSTER, JOHN S. WERNER, AND DAVID J. FIELD 9.1 Introduction To what extent do we have shared or unique

More information

Independent Component Analysis in Neuron Model

Independent Component Analysis in Neuron Model Independent Component Analysis in Neuron Model Sheng-Hsiou Hsu shh078@ucsd.edu Yae Lim Lee yaelimlee@gmail.com Department of Bioengineering University of California, San Diego La Jolla, CA 92093 Abstract

More information

THE ENCODING OF PARTS AND WHOLES

THE ENCODING OF PARTS AND WHOLES THE ENCODING OF PARTS AND WHOLES IN THE VISUAL CORTICAL HIERARCHY JOHAN WAGEMANS LABORATORY OF EXPERIMENTAL PSYCHOLOGY UNIVERSITY OF LEUVEN, BELGIUM DIPARTIMENTO DI PSICOLOGIA, UNIVERSITÀ DI MILANO-BICOCCA,

More information

Asymmetries in ecological and sensorimotor laws: towards a theory of subjective experience. James J. Clark

Asymmetries in ecological and sensorimotor laws: towards a theory of subjective experience. James J. Clark Asymmetries in ecological and sensorimotor laws: towards a theory of subjective experience James J. Clark Centre for Intelligent Machines McGill University This talk will motivate an ecological approach

More information

On the implementation of Visual Attention Architectures

On the implementation of Visual Attention Architectures On the implementation of Visual Attention Architectures KONSTANTINOS RAPANTZIKOS AND NICOLAS TSAPATSOULIS DEPARTMENT OF ELECTRICAL AND COMPUTER ENGINEERING NATIONAL TECHNICAL UNIVERSITY OF ATHENS 9, IROON

More information

Hierarchical Bayesian Modeling of Individual Differences in Texture Discrimination

Hierarchical Bayesian Modeling of Individual Differences in Texture Discrimination Hierarchical Bayesian Modeling of Individual Differences in Texture Discrimination Timothy N. Rubin (trubin@uci.edu) Michael D. Lee (mdlee@uci.edu) Charles F. Chubb (cchubb@uci.edu) Department of Cognitive

More information

To What Extent Can the Recognition of Unfamiliar Faces be Accounted for by the Direct Output of Simple Cells?

To What Extent Can the Recognition of Unfamiliar Faces be Accounted for by the Direct Output of Simple Cells? To What Extent Can the Recognition of Unfamiliar Faces be Accounted for by the Direct Output of Simple Cells? Peter Kalocsai, Irving Biederman, and Eric E. Cooper University of Southern California Hedco

More information

SUPPLEMENTARY INFORMATION. Table 1 Patient characteristics Preoperative. language testing

SUPPLEMENTARY INFORMATION. Table 1 Patient characteristics Preoperative. language testing Categorical Speech Representation in the Human Superior Temporal Gyrus Edward F. Chang, Jochem W. Rieger, Keith D. Johnson, Mitchel S. Berger, Nicholas M. Barbaro, Robert T. Knight SUPPLEMENTARY INFORMATION

More information

Parallel processing strategies of the primate visual system

Parallel processing strategies of the primate visual system Parallel processing strategies of the primate visual system Parallel pathways from the retina to the cortex Visual input is initially encoded in the retina as a 2D distribution of intensity. The retinal

More information

An Information Theoretic Model of Saliency and Visual Search

An Information Theoretic Model of Saliency and Visual Search An Information Theoretic Model of Saliency and Visual Search Neil D.B. Bruce and John K. Tsotsos Department of Computer Science and Engineering and Centre for Vision Research York University, Toronto,

More information

196 IEEE TRANSACTIONS ON AUTONOMOUS MENTAL DEVELOPMENT, VOL. 2, NO. 3, SEPTEMBER 2010

196 IEEE TRANSACTIONS ON AUTONOMOUS MENTAL DEVELOPMENT, VOL. 2, NO. 3, SEPTEMBER 2010 196 IEEE TRANSACTIONS ON AUTONOMOUS MENTAL DEVELOPMENT, VOL. 2, NO. 3, SEPTEMBER 2010 Top Down Gaze Movement Control in Target Search Using Population Cell Coding of Visual Context Jun Miao, Member, IEEE,

More information

Exploring the Functional Significance of Dendritic Inhibition In Cortical Pyramidal Cells

Exploring the Functional Significance of Dendritic Inhibition In Cortical Pyramidal Cells Neurocomputing, 5-5:389 95, 003. Exploring the Functional Significance of Dendritic Inhibition In Cortical Pyramidal Cells M. W. Spratling and M. H. Johnson Centre for Brain and Cognitive Development,

More information

Models of Attention. Models of Attention

Models of Attention. Models of Attention Models of Models of predictive: can we predict eye movements (bottom up attention)? [L. Itti and coll] pop out and saliency? [Z. Li] Readings: Maunsell & Cook, the role of attention in visual processing,

More information

Neuroinformatics. Ilmari Kurki, Urs Köster, Jukka Perkiö, (Shohei Shimizu) Interdisciplinary and interdepartmental

Neuroinformatics. Ilmari Kurki, Urs Köster, Jukka Perkiö, (Shohei Shimizu) Interdisciplinary and interdepartmental Neuroinformatics Aapo Hyvärinen, still Academy Research Fellow for a while Post-docs: Patrik Hoyer and Jarmo Hurri + possibly international post-docs PhD students Ilmari Kurki, Urs Köster, Jukka Perkiö,

More information

Lecturer: Rob van der Willigen 11/9/08

Lecturer: Rob van der Willigen 11/9/08 Auditory Perception - Detection versus Discrimination - Localization versus Discrimination - - Electrophysiological Measurements Psychophysical Measurements Three Approaches to Researching Audition physiology

More information

A MULTI-STAGE COLOR MODEL REVISITED: IMPLICATIONS FOR A GENE THERAPY CURE FOR RED-GREEN COLORBLINDNESS

A MULTI-STAGE COLOR MODEL REVISITED: IMPLICATIONS FOR A GENE THERAPY CURE FOR RED-GREEN COLORBLINDNESS Abstract for Chapter A MULTI-STAGE COLOR MODEL REVISITED: IMPLICATIONS FOR A GENE THERAPY CURE FOR RED-GREEN COLORBLINDNESS Katherine Mancuso 1, Matthew C. Mauck 2, James A. Kuchenbecker 1, Maureen Neitz

More information

SENSES: VISION. Chapter 5: Sensation AP Psychology Fall 2014

SENSES: VISION. Chapter 5: Sensation AP Psychology Fall 2014 SENSES: VISION Chapter 5: Sensation AP Psychology Fall 2014 Sensation versus Perception Top-Down Processing (Perception) Cerebral cortex/ Association Areas Expectations Experiences Memories Schemas Anticipation

More information

Prof. Greg Francis 7/31/15

Prof. Greg Francis 7/31/15 s PSY 200 Greg Francis Lecture 06 How do you recognize your grandmother? Action potential With enough excitatory input, a cell produces an action potential that sends a signal down its axon to other cells

More information

Goals. Visual behavior jobs of vision. The visual system: overview of a large-scale neural architecture. kersten.org

Goals. Visual behavior jobs of vision. The visual system: overview of a large-scale neural architecture. kersten.org The visual system: overview of a large-scale neural architecture Daniel Kersten Computational Vision Lab Psychology Department, U. Minnesota Goals Provide an overview of a major brain subsystem to help

More information

Lecturer: Rob van der Willigen 11/9/08

Lecturer: Rob van der Willigen 11/9/08 Auditory Perception - Detection versus Discrimination - Localization versus Discrimination - Electrophysiological Measurements - Psychophysical Measurements 1 Three Approaches to Researching Audition physiology

More information

Visual Categorization: How the Monkey Brain Does It

Visual Categorization: How the Monkey Brain Does It Visual Categorization: How the Monkey Brain Does It Ulf Knoblich 1, Maximilian Riesenhuber 1, David J. Freedman 2, Earl K. Miller 2, and Tomaso Poggio 1 1 Center for Biological and Computational Learning,

More information

Natural Scene Statistics and Perception. W.S. Geisler

Natural Scene Statistics and Perception. W.S. Geisler Natural Scene Statistics and Perception W.S. Geisler Some Important Visual Tasks Identification of objects and materials Navigation through the environment Estimation of motion trajectories and speeds

More information

Object recognition and hierarchical computation

Object recognition and hierarchical computation Object recognition and hierarchical computation Challenges in object recognition. Fukushima s Neocognitron View-based representations of objects Poggio s HMAX Forward and Feedback in visual hierarchy Hierarchical

More information

Concurrent measurement of perceived speed and speed discrimination threshold using the method of single stimuli

Concurrent measurement of perceived speed and speed discrimination threshold using the method of single stimuli Vision Research 39 (1999) 3849 3854 www.elsevier.com/locate/visres Concurrent measurement of perceived speed and speed discrimination threshold using the method of single stimuli A. Johnston a, *, C.P.

More information

using deep learning models to understand visual cortex

using deep learning models to understand visual cortex using deep learning models to understand visual cortex 11-785 Introduction to Deep Learning Fall 2017 Michael Tarr Department of Psychology Center for the Neural Basis of Cognition this lecture A bit out

More information

Information and neural computations

Information and neural computations Information and neural computations Why quantify information? We may want to know which feature of a spike train is most informative about a particular stimulus feature. We may want to know which feature

More information

Lateral Geniculate Nucleus (LGN)

Lateral Geniculate Nucleus (LGN) Lateral Geniculate Nucleus (LGN) What happens beyond the retina? What happens in Lateral Geniculate Nucleus (LGN)- 90% flow Visual cortex Information Flow Superior colliculus 10% flow Slide 2 Information

More information

Neural circuits PSY 310 Greg Francis. Lecture 05. Rods and cones

Neural circuits PSY 310 Greg Francis. Lecture 05. Rods and cones Neural circuits PSY 310 Greg Francis Lecture 05 Why do you need bright light to read? Rods and cones Photoreceptors are not evenly distributed across the retina 1 Rods and cones Cones are most dense in

More information

Seeing Color. Muller (1896) The Psychophysical Axioms. Brindley (1960) Psychophysical Linking Hypotheses

Seeing Color. Muller (1896) The Psychophysical Axioms. Brindley (1960) Psychophysical Linking Hypotheses Muller (1896) The Psychophysical Axioms The ground of every state of consciousness is a material process, a psychophysical process so-called, to whose occurrence the state of consciousness is joined To

More information

Photoreceptors Rods. Cones

Photoreceptors Rods. Cones Photoreceptors Rods Cones 120 000 000 Dim light Prefer wavelength of 505 nm Monochromatic Mainly in periphery of the eye 6 000 000 More light Different spectral sensitivities!long-wave receptors (558 nm)

More information

Simulations of adaptation and color appearance in observers with varying spectral sensitivity

Simulations of adaptation and color appearance in observers with varying spectral sensitivity Ophthal. Physiol. Opt. 2010 30: 602 610 Simulations of adaptation and color appearance in observers with varying spectral sensitivity Michael A. Webster, Igor Juricevic and Kyle C. McDermott Department

More information