We trained a neurocomputational model on six categories of photographic images that were used in a previous fMRI study of object and face processing. Multivariate pattern analyses of the activations elicited in the object-encoding layer of the model yielded results consistent with two previous, contradictory fMRI studies. Findings from one of the studies [Haxby, J. V., Gobbini, M. I., Furey, M. L., Ishai, A., Schouten, J. L., & Pietrini, P. Distributed and overlapping representations of faces and objects in ventral temporal cortex. Science, 293, 2425–2430, 2001] were interpreted as evidence for the object-form topography model. Findings from the other study [Spiridon, M., & Kanwisher, N. How distributed is visual category information in human occipito-temporal cortex? An fMRI study. Neuron, 35, 1157–1165, 2002] were interpreted as evidence for neural processing mechanisms in the fusiform face area that are specialized for faces. Because the model contains no special processing mechanism or specialized architecture for faces and yet it can reproduce the fMRI findings used to support the claim that there are specialized face-processing neurons, we argue that these fMRI results do not actually support that claim. Results from our neurocomputational model therefore constitute a cautionary tale for the interpretation of fMRI data.

What is the nature of the representations in visual cortex that underlie object processing? Empirical research has attempted to address this fundamental question for more than half a century, using a wide range of experimental techniques. Recently, two contrasting theoretical frameworks for understanding the functional organization of ventral visual cortex have emerged: The first is a modular view of object processing (e.g., Kanwisher, 2010; Spiridon & Kanwisher, 2002; Kanwisher, McDermott, & Chun, 1997), and the second is a “distributed processing” account (e.g., O'Toole, Jiang, Abdi, & Haxby, 2005; Hanson, Matsuka, & Haxby, 2004; Haxby et al., 2001; Ishai, Ungerleider, Martin, Schouten, & Haxby, 1999).

According to the modular view, there is cortical specialization for specific domains of visual recognition: Human inferotemporal cortex contains discrete regions dedicated to distinct object categories, such as faces (Kanwisher et al., 1997; McCarthy, Puce, Gore, & Allison, 1997), places (Epstein & Kanwisher, 1998), body parts (Downing, Jiang, Shuman, & Kanwisher, 2001), and word forms (Cohen et al., 2002; Petersen, Fox, Snyder, & Raichle, 1990). Such regions are proposed to have a selective role in the perception of a restricted object category and to employ qualitatively distinct processing mechanisms (Kanwisher, 2010; Spiridon & Kanwisher, 2002). In addition, it has been suggested that such modules may have evolutionary as well as developmental origins (Kanwisher, 2010).

In contrast, distributed processing theories of visual recognition claim that object representations are distributed across a broad expanse of inferotemporal cortex. The object form topography account, proposed by Haxby and colleagues (Haxby et al., 2001; Ishai, Ungerleider, Martin, & Haxby, 2000; Ishai et al., 1999), posits that ventral temporal cortex contains a continuous representation of object form, with a topological organization that reflects the distinctions between object categories. Arguing against the modular view, Ishai et al. (1999) suggest that the topology arises because information characteristic of objects within a category clusters together in cortex, producing a region that responds maximally to that category; this gives the appearance of a module, belying the continuous, distributed nature of the cortical representations.

Both the modular and distributed processing views of object processing have been supported primarily by data from fMRI studies, in which participants view pictures of objects and faces while the hemodynamic response in their ventral visual cortex is measured (e.g., Downing et al., 2001; Haxby et al., 2001; Ishai et al., 1999; Kanwisher et al., 1997; McCarthy et al., 1997). In this article, we focus on the debate over the existence of an anatomical module for faces by using a neurocomputational model to replicate studies in which multivoxel pattern analysis (MVPA) was used to interpret the fMRI data. Unlike traditional univariate analyses of fMRI data, MVPA examines patterns of activation across voxels in a search for combinations that maximally discriminate between behavioral conditions. Therefore, MVPA allows for the possibility of discovering a combinatorial representational code, which is likely to be utilized by the brain if representations are truly distributed. Using a neurocomputationally plausible model of object processing (Cowell, Huber, & Cottrell, 2009), we perform MVPA on the activation patterns in the layer of the model where stimulus representations reside. Our aim is to assist in the interpretation of fMRI results by examining the necessity of possible schemes of neural representation (e.g., modular vs. distributed) for producing particular patterns of BOLD activation. The method we present allows us to check whether the representational assumptions that are made in interpreting MVPA results are necessary to produce the data reported in empirical studies.

We examine the studies of Spiridon and Kanwisher (2002) and Haxby et al. (2001). In line with the idea that MVPA may be used to investigate both distributed and modular hypotheses, the authors of these two studies drew opposing conclusions from their results. Haxby et al. (2001) found that information about the category membership of an object was distributed across visual cortex, rather than confined to regions that were maximally active in response to that category, and therefore argued in favor of distributed object representations. Spiridon and Kanwisher (2002) also found that information was distributed but that the information in “face-selective” regions was more exclusively useful for making category decisions about faces than was the case for any other object-selective region. That is, the “house-selective” region was not as selective in its utility for categorizing houses, nor was the “bottle-selective” region superior in its categorization of bottles, and so on. The authors took this as evidence for specialized processing of faces by face-selective neurons.

We trained our neurocomputational model on six categories of the photographic images used by Haxby et al. (2001). MVPA of the activation patterns elicited in the object-encoding layer demonstrated that both the Haxby et al. (2001) results and the Spiridon and Kanwisher (2002) results were replicated by our model. Because the model contains no special processing mechanism or specialized architecture for faces and yet the above finding from Spiridon and Kanwisher (2002) was nonetheless produced in the model, we argue that this result from Spiridon and Kanwisher (2002) cannot be used to support the claim that there are specialized face-processing neurons in the fusiform face area (FFA).1 The findings from our neurocomputational model therefore constitute a cautionary tale for the interpretation of fMRI data. Moreover, this study introduces a novel computational method for testing the necessity of cognitive hypotheses for explaining fMRI data from studies of object processing.

A Neurocomputational Model of Object Processing in Visual Cortex

To simulate fMRI data, we require a model that is, on some level, neurocomputationally plausible. Because the target data are responses of voxels—each of which contains several million neurons (Logothetis, 2008)—we opt for a level of analysis in which the architecture and processing algorithm mimic cortical connectivity and processing across networks of neurons rather than in single units. An appropriate model is the Kohonen network (Kohonen, 1982), in which the learning algorithm is a computational abstraction of cortical mechanisms such as Hebbian learning and lateral inhibition. A Kohonen network self-organizes, meaning that learning is unsupervised, lending it biological plausibility. Learning of the feedforward weights is strongly influenced by lateral connectivity between units in the network (Kohonen, 1982, 1984); this architecture produces representations that preserve the similarity structure of the stimulus inputs, as in representations seen in mammalian sensory and motor cortex (e.g., Krubitzer & Calford, 1992; Cynader, Swindale, & Matsubara, 1987; Swindale, Matsubara, & Cynader, 1987; Aitkin, Merzenich, Irvine, Clarey, & Nelson, 1986; Lemon, 1981). This makes the network particularly well suited to our purposes, because stimulus representations are spatially situated across units and subsets of those units that are defined by stimulus preference are often spatially contiguous, similar to category-selective voxels seen in fMRI studies. These subsets of units can be used for MVPA, analogous to the subsets of voxels used in the fMRI studies. Moreover, because of this property, the method can potentially create a model “FFA,” unlike a backpropagation network in which the hidden unit representations are fully combinatorial without any spatial localization. In addition, a Kohonen network offers a system of representation that is somewhere between distributed, in that all inputs are connected to all representational units, and localized, in that lateral inhibition constrains representations to be somewhat localized in space. Thus, the network representations mimic those observed in fMRI studies of object-processing areas, possessing both distributed and localized properties, and the model has the potential to reproduce MVPA findings from both the “distributed account” and the “modular account”. By analyzing the activation patterns that the model contains, we can investigate whether both sets of results can emerge from a system of representation in which no modules are explicitly assumed.

Figure 1 shows the model architecture. First, input images are filtered using Gabor wavelets, modeling the receptive fields in V1. The Gabor filter outputs are reduced in dimensionality by principal component analysis (PCA), a process that can be accomplished by Hebbian learning (Sanger, 1989). Processing in these two early stages echoes the dimensionality expansion and reduction of stimulus representations that are thought to occur in striate and extrastriate visual cortex (Sanger, 1989; Daugman, 1985). At the next stage of the model, the stimulus representations resulting from Gabor filtering and PCA are input to a Kohonen network, in which topographically arranged representations of objects develop. In the final stage, representations in the topographic layer are associated, via the delta rule, with output nodes corresponding to six object categories, as a test of whether the information in the representation layer is sufficient to distinguish the categories.

Figure 1. 

Model architecture. Photographic images are Gabor filtered, and then the filter outputs are reduced in dimensionality through PCA. For each stimulus, projections onto the top 20 PCs are passed as inputs to the two-dimensional Kohonen network at the object encoding level of the model. In a final layer, units in the object encoding level are connected to six category-level output units, via weights that are adjustable through the delta rule. The activations of units at the category level are used to determine network behavior (i.e., categorization performance) during training to criterion. Activations of units in the object encoding level are used for the MVPA simulations.

Figure 1. 

Model architecture. Photographic images are Gabor filtered, and then the filter outputs are reduced in dimensionality through PCA. For each stimulus, projections onto the top 20 PCs are passed as inputs to the two-dimensional Kohonen network at the object encoding level of the model. In a final layer, units in the object encoding level are connected to six category-level output units, via weights that are adjustable through the delta rule. The activations of units at the category level are used to determine network behavior (i.e., categorization performance) during training to criterion. Activations of units in the object encoding level are used for the MVPA simulations.

Close modal

The use of a Kohonen network at the hidden layer is a departure from previous models of face and object processing we have developed (Tong, Joyce, & Cottrell, 2008; Dailey & Cottrell, 1999), in which the hidden layer was trained with the backpropagation algorithm. The Kohonen network introduces two key properties not possessed by a backpropagation network. First, the learning of object representations in the model is unsupervised and neurocomputationally plausible. Second, units in the Kohonen network occupy a fixed position relative to their neighbors within a two-dimensional grid, and object representations are arranged topographically across them, making it possible to find the units that are “selective for” a given object category.

The network is trained by presenting a series of stimuli and incrementally updating the weights from the structural encoding level to the object-encoding units at each presentation. The topography of stimulus representations emerges in the network owing to the neighborhood function governing the weight updates: the most active unit updates its weights most strongly to move closer to the stimulus; neighboring units in the grid are also updated in the direction of the stimulus, but less strongly; and the weights of distant units are not updated. Consequently, neighboring units learn similar weight values and come to represent items that are close in stimulus space.

MVPA of Simulated Data

We replicated the fMRI studies by treating the activations of units at the object encoding level as responses of voxels in an fMRI data set. We performed MVPA using correlation, measuring the discriminability of object categories by comparing the similarity of the representations elicited by each category. As in Spiridon and Kanwisher (2002) and Haxby et al. (2001), we analyzed the activation patterns not only across all voxels (units) that were responsive to objects but also across smaller subsets of those units. The subsets were defined, as in the fMRI studies, as populations of units that were more strongly activated by one particular stimulus category than the others, providing a computational equivalent of “face-selective” or “house-selective” regions in cortex.

Visual Categorization Training

Stimuli

We used grayscale photographic stimuli from the fMRI study of Haxby et al. (2001). Stimuli belonged to six categories: faces, houses, chairs, bottles, scissors, and shoes. This set comprised all the object categories of Haxby et al. (2001) except cats2 and scrambled images. All images were scaled and cropped to 64 × 64 pixels. We used 240 training images, 40 from each category, with each category set containing 10 exemplars pictured from four different views. An additional 48 images (eight per category: two distinct exemplars from four different views) were assigned to a holdout set for testing classification performance after every 10 training epochs.

Image Preprocessing

Stimuli were filtered using Gabor wavelets to transform the images into representations suitable for object recognition (Dailey & Cottrell, 1999). The Gabor filters were applied to 1,024 points in each image, evenly spaced on a 32 × 32 grid. This resulted in a vector of size 40,960 (1,024 sample points, at which eight orientations and five scales of the filter were applied), and we reduced the dimensionality of the patterns by performing PCA on all 288 images used in the study. For each image, we retained the projections onto the first 20 principal components to use as input to the Kohonen network. We did not normalize the vectors produced by the PCA to have equal variance (a common step for backpropagation networks) because unnormalized vectors enabled better learning in the Kohonen layer.

Training the Neural Network

Twelve networks were trained to classify objects into six categories. Learning of feedforward weights from the input units to the object-encoding units was unsupervised, that is, it was independent from the learning of weights from the object-encoding units to the output units. Unlike backpropagation, the only effect of the object-encoding-to-output weights on the input-to-object-encoding weights was that all training was terminated according to classification accuracy at the output units. All units in the model were sigmoidal. The weights from input units to object-encoding units were trained by the Kohonen learning rule as follows:
in which wji is the weight from input i to unit j, ai is the activation of input i, and f is a neighborhood weighting function on the learning rate. The neighborhood function is centered on the most active unit and is given by the following:
where dist is the distance from unit i to the maximally active unit in the grid and η is the learning rate, which starts at 1 and reduces over epochs, as η = epoch(−0.2). G is a Gaussian width parameter that decreases over the first 50 epochs, as G = 0.5 + 10*epoch(−0.3). Activations of sigmoidal units in the Kohonen layer were scaled by a factor of 10 in the exponential. The weights from object-encoding units to output units were trained using a delta rule with learning rate of 0.01. The object-encoding layer was a square grid with sides of length 40, 50, 60, or 70 units.

Classification accuracy was assessed after every 10 training epochs by recording the responses at the output nodes to stimuli in the holdout set. Training of each network was terminated when classification accuracy exceeded 85% on three successive presentations of the holdout set, but not before at least 40 training epochs were completed.

Procedures for MVPA of Empirical Data

Haxby et al. MVPA Methods

Participants viewed grayscale photographic images of eight categories: faces, houses, cats, bottles, shoes, scissors, chairs, and phase-scrambled images. For each participant, fMRI data were screened to find “object-selective” voxels, in which the BOLD signal to the different object categories differed significantly, as assessed by the omnibus effect of seven regressors used to model the seven meaningful categories. Only object-selective voxels were used in subsequent MVPA. Data for each participant were split into halves: odd- and even-numbered scans. The discriminability of brain images was determined by examining the similarity of activation patterns in the halves, with similarity measured as the voxelwise correlation between the patterns. For the pairwise category comparison between, for example, faces and shoes, the correlation between the mean response to faces on odd runs and that to faces on even runs was compared with the correlation between the mean response to faces on odd runs and that to shoes on even runs. If the face–face (within)-category correlation was higher than the face–shoe (between-category) correlation, the discrimination was scored as correct. In fact, there are four binary comparisons to be made for each pairwise category discrimination: faceodd–faceeven versus faceodd–shoeeven, faceodd–faceeven versus faceeven–shoeodd, shoeodd–shoeeven versus faceodd–shoeeven, and shoeodd–shoeeven versus faceeven–shoeodd. Averaging over four binary comparisons yields a possible score of 0%, 25%, 50%, 75%, or 100%, for each pairwise category discrimination. The overall discriminability of a category was determined by averaging over all seven pairwise category comparisons involving that category (e.g., for faces: face–house, face–cat, face–shoe, etc.).

Spiridon and Kanwisher MVPA Methods

These authors used MVPA procedures very similar to those of Haxby et al., except for the way in which scan data were split into halves. Spiridon and Kanwisher (2002) assayed three different splits: first, they divided the scan data so that brain images in the halves had been evoked by the identical set of images (the “identical images” condition); second, the halves of scan data were evoked by images of the same individuals in each category, but the individuals were not pictured from the same views in each half (“different views”); and third, the sets of images evoking responses in the halves of data contained different exemplars of each object category (“different exemplars”). Thus, the within-category correlation (WCC) between the halves of the data measured either the reliability of brain responses to a given stimulus, the reliability of responses to different views of the same individual, or the reliability of brain responses to different individuals from the same category. Importantly for our purposes, Spiridon and Kanwisher (2002) found no significant differences between the accuracy of pairwise category discriminations across the three conditions.

Procedures for MVPA of Simulated Data

Recording Activation Patterns

For all trained networks, we recorded the activation patterns in the object-encoding units elicited by all 240 training stimuli on the final epoch of training.3

Assigning Images to “Scans”

In the fMRI studies, participants viewed blocks of 12 (Haxby et al.) or 16 (Spiridon and Kanwisher) photographic images from a single category. Because the hemodynamic response is slow, the activation of voxels by any individual image was temporally smeared across the block such that the activation pattern recorded for a particular category on any scan was an amalgamation of the brain responses elicited by all images in that block. In contrast, in our simulations, we were able to extract the precise activation pattern elicited by each photographic image. To approximate the data averaging in the fMRI studies, we assigned stimulus images within each category to blocks and averaged across all patterns within a block. In each simulated “scan,” there was one such block per category.

In total, we used activation patterns caused by 240 stimuli, comprising 40 from each category, consisting of 10 individual exemplars pictured from four different viewpoints. (However, only for chairs and shoes were the four viewpoints consistent across individual exemplars; for other categories, each exemplar was pictured from four viewpoints, but not necessarily the same four viewpoints as used for other exemplars). We assigned the 40 images from each category to four scans of 10 images using a protocol that approximated those of the fMRI studies. In simulations, each “scan” contained two views each of five individuals. For each category, we randomly assigned five individuals to Scans 1 and 2, and the other five individuals to Scans 3 and 4. In addition, for all categories, we constrained Scans 1 and 3 to contain only two of the four views per individual and Scans 2 and 4 to contain the other two views.

Dividing the Data

For each category, we divided the activation patterns corresponding to the four scans into halves (two sets of two) and averaged over each half to obtain the two averaged activation patterns for measuring correlations. To split the scans, we used both the “different views” and the “different exemplars” methods of Spiridon and Kanwisher (2002). For “different views,” we assigned Scans 1 and 3 to the first half and Scans 2 and 4 to the second half, so that each half was generated by images of all possible individuals but only half of all possible views (albeit that not all categories comprised a standard set of four views). For “different exemplars,” we assigned Scans 1 and 2 to the first half and Scans 3 and 4 to the second half, so that each half contained all views of any individual but only half of all individuals. In all simulations, we ran the analysis with both types of split then averaged the discrimination scores from the two analyses into a single score. In general, we found that the type of split did not dramatically affect discrimination accuracy.

Determining Visually Active and Category-selective Units

We tested the object-encoding layer units in our model for visual activity and category selectivity with methods very similar to those of the fMRI studies, before performing MVPA. For Simulations 1–3 (replicating Haxby et al., 2001), we included all object-selective units. To determine object selectivity for each unit, we performed a one-way ANOVA on the unit's responses to each category and examined the omnibus effect of the six conditions (i.e., categories), using an uncorrected significance level of p < 10−6 as our criterion. (We found that, in a typical simulation, slightly more than 90% of model units passed this criterion.) Spiridon and Kanwisher (2002) first screened voxels to exclude any voxels that were not visually active, that is, possessed activations that were not significantly different from zero for any object category. In Simulation 4, we did the same with all object-encoding units and found that, unsurprisingly, all units were visually active. Next, we determined the category selectivity of each unit by performing an independent samples t test comparing the responses to the maximally activating object category with the responses to all other categories. We used the p values derived in this t test to rank order the voxels in terms of selectivity, to determine the “most selective” units (see below).

Selection of Units

In Simulation 1, we used all object-selective units. In Simulation 2, we used all object-selective units except those that were maximally activated by one of the two categories being discriminated. For example, when discriminating houses and scissors, we used all object-selective units except those maximally activated by either house images or scissor images. The units maximally activated by a given category were defined as those among the object-selective units (i.e., among units whose responses differed significantly by object category) for which the mean activation caused by that category was higher than the mean activations caused by all other categories. We report the overall discriminability of each category; for example, for “houses,” this is the average across all pairwise discriminations involving the category “houses” (houses vs. faces, houses vs. bottles, etc.).

In Simulation 3, we included only object-selective units that were maximally activated by a single category, with “maximally activated” being determined as in Simulation 2. We then used these six sets of units—maximally activated by each of the six categories—to perform six separate analyses. The number of units in the object-encoding grid that qualified as maximally activated by each category was not equal across categories.

Simulation 4 replicated a finding from Spiridon and Kanwisher (2002), which involved an analysis similar to Simulation 3 but with the number of category-selective voxels fixed at 30 for all categories. For each category, we selected the 30 units for which the independent samples t test measuring category selectivity had yielded the smallest p values. Following Spiridon and Kanwisher (2002), for all nonpreferred categories in each set of 30 units, we excluded all discriminations involving the preferred category. For example, in the “house units” analysis, we excluded all pairwise discriminations involving houses in the calculation of mean overall discrimination accuracy, doing so for all overall category discrimination scores except “house” (for which it is not possible to exclude all pairwise discriminations involving houses and still compute a score).

Discrimination Accuracy

Empirical BOLD data contain considerable noise, arising from both internal sources within the brains of scanned participants and external sources, such as variability in scanner measurements. Because of such noise, activation patterns in the halves of the scan data elicited by the same object category (such as faceeven–faceodd or shoeeven–shoeodd) produce voxelwise correlations considerably less than 1. This is true even in the “identical images” condition, in which brain responses in the halves of the scan data are elicited by the exact same set of images. In contrast, in the neurocomputational model, there is no noise: Each time a particular image is presented to the same trained network, the exact same activation pattern results. Consequently, if input images are assigned to “scans” in the halves of the simulated data according to the “identical images” protocol, all WCCs are 1. On the other hand, if input images are assigned to the halves of simulated data using a scheme similar to the “different views” or “different exemplars” conditions of Spiridon and Kanwisher (2002), natural variability in network responses to different instances of the same category produces WCCs of less than 1, which are more in line with those of fMRI data. Therefore, we divided the activation patterns such that the two data halves did not contain identical images, yielding WCC and between-category correlation (BCC) values that varied from −1 to +1.

We determined category discriminability by averaging over pairwise category discrimination scores, each of which was derived from four binary comparisons of WCC and BCC values. However, we modified the procedure for comparing correlation values from that used in the fMRI studies. Because the simulated activation patterns contained far less noise than empirical BOLD data, the WCC values were consistently higher than the BCC values, and so, averaging across the four binary comparisons of within- versus between-correlations consistently yielded scores of 100%. To avoid this ceiling effect, we derived a probabilistic pairwise category discrimination score from the WCC and BCC values as follows:
where i = 1, 2 is the pair of categories being compared; if i = 2, then j = 1, whereas if i = 1, then j = 2. Thus, rbetweeni,j is the correlation between patterns elicited by category i in the first half of the data and category j in the second half of the data, and rbetweenj,i is the correlation between patterns elicited by category i in the second half of the data and category j in the first half of the data. The constant β = 2 and was chosen such that discrimination was at 95% for high WCC and low BCC values, before applying it to the network data. Probability of correct choice was used as a proxy for accuracy of a pairwise category discrimination.

Exploratory Analyses

BCC and WCC

To determine which properties of the simulated object representations were driving the simulated MVPA results, we examined separately the two correlation values—BCC and WCC—that contribute to pairwise discriminations. To calculate BCC and WCC, activation patterns to all stimuli in a category were split into halves and averaged, as for the calculation of discrimination accuracy. For BCC values, the averaged patterns from the halves of the data were compared for all 15 pairings of two different categories (face–house, face–chair, face–bottle, …, house–chair, house–bottle, etc.). There were two ways of making each comparison, for example, faces in the first half of the data versus houses in the second half (face1–house2), as well as faces in the second half versus houses in the first half (face2–house1), and two ways of splitting the activation patterns into halves (“different views” and “different exemplars”). The mean BCC reported for each category is the average across all five pairwise comparisons involving that category, across both comparison types and both split types. Mirroring the calculation of discrimination accuracy in Simulation 4 (in which we followed the procedure of Spiridon & Kanwisher, 2002; see Selection of Units), for all nonpreferred categories in each set of 30 units, we excluded all BCCs involving the preferred category. To determine each WCC value, we compared the average patterns elicited by that category in the two halves of the data.

Dendrograms

To visualize the similarity of activation patterns in the different categories, we performed a hierarchical cluster analysis and plotted dendrograms using the MATLAB function dendrogram. Assuming some similarity space for all objects in a data set, a dendrogram depicts the clustering of objects in that space by using inverted U-shaped lines to connect objects in a hierarchical tree. We quantified distance as one minus the correlation between two activation patterns (using the MATLAB function pdist with method “correlation”) to mirror the similarity metric used in Simulation 4. We then used these distances to link similar pairs of objects into binary clusters (containing two objects) before linking these clusters to each other and to other objects, creating progressively bigger clusters until all activation patterns in the data set were linked together in a tree (using the MATLAB function linkage). Each tree was plotted as a dendrogram. The height of each inverted U indicates the distance between the objects, or groups of objects, that are connected by the two arms of the U. Where the two objects connected by an inverted U contain more than one activation pattern, the height of the U represents the average of distances between all possible pairs of objects across the groups.

Using a neurocomputationally plausible model (Figure 1), we trained 100 networks to classify photographic images into six object categories. To ensure that the network properties we investigated were not dependent on network size, we tested 25 networks at each of four sizes4 for the object representation layer: 40 × 40, 50 × 50, 60 × 60, and 70 × 70 units. Ninety-nine networks reached a prespecified performance criterion after between 40 and 200 training epochs; one network failed to reach criterion within a prespecified limit and so was replaced by an additional, successful simulation, giving 100 networks in the final data set. On completion of training, a topographic organization of object representations was typically seen across units in the object-encoding layer: Units that were maximally active to a particular object category tended to be spatially clustered in the two-dimensional grid of units. Figure 2 shows the category preference elicited by each unit in the object encoding layer in a typical simulation; Figure 3 shows the average activation value elicited in each unit by the set of patterns in each category. Figure 2 reflects numerical category preferences: Each unit is coded according to the category that elicited the numerically highest average activation in it, regardless of whether that activation value was statistically significantly higher than the activation caused by other categories. Units were subject to selectivity tests analogous to those used by the fMRI authors before being entered into MVPA simulations; therefore, in any given simulation, only a subset of the units shown in each patch may have been included in the MVPA.

Figure 2. 

Category preferences of units in the object encoding layer, in a typical network size of 40 × 40 units trained on six object categories. Each unit in the layer is depicted by a textured square at the xy position corresponding to its location in the model; texture indicates the object category that maximally activates the unit. Category preference for each unit is taken as the category that elicited the numerically highest activation in the unit, when comparing the activation patterns obtained by averaging all stimuli in each category.

Figure 2. 

Category preferences of units in the object encoding layer, in a typical network size of 40 × 40 units trained on six object categories. Each unit in the layer is depicted by a textured square at the xy position corresponding to its location in the model; texture indicates the object category that maximally activates the unit. Category preference for each unit is taken as the category that elicited the numerically highest activation in the unit, when comparing the activation patterns obtained by averaging all stimuli in each category.

Close modal
Figure 3. 

Activation of all units in the grid (for the same 40 × 40 network depicted in Figure 2), averaged over all stimulus patterns in a given category. As in Figure 2, each unit is depicted by a square at its xy position in the model. High activations are shown in red, whereas low activations are in blue, as indicated by the scale at the left.

Figure 3. 

Activation of all units in the grid (for the same 40 × 40 network depicted in Figure 2), averaged over all stimulus patterns in a given category. As in Figure 2, each unit is depicted by a square at its xy position in the model. High activations are shown in red, whereas low activations are in blue, as indicated by the scale at the left.

Close modal

Simulation 1: Category Discrimination Using All Object-selective Voxels

The first finding reported by Haxby et al. (2001) was that activation patterns across all object-selective voxels in ventral temporal cortex contained sufficient information to perform pairwise category discriminations between the object categories viewed, with an overall accuracy level of 96%. In addition, Haxby et al. (2001) found that the pairwise discrimination of activation patterns elicited by stimuli from two different categories was most accurate when one of the two categories was faces. Activation patterns caused by houses were also well discriminated from other categories. Pairwise discriminations for scissors, shoes, and bottles were, on average, less accurate.

Having trained networks to classify the six categories, we extracted activation patterns elicited in the object-encoding units by images from each category. In this analysis, which included all object-selective units, we performed MVPA on the model activation patterns using the methods of Haxby et al. (2001) for analyzing BOLD activation patterns. Simulated activation patterns, like human brain images, contained sufficient information to produce accurate pairwise discrimination of the six object categories included in our study (Table 1, top row). Overall accuracy was 85% with the parameters we used to determine discrimination performance (see Methods). In addition, we obtained a very similar qualitative trend in the accuracy of pairwise discriminations across object categories: Patterns elicited by faces were the most accurately discriminated from other categories, whereas scissors, shoes, and bottles were the most poorly discriminated classes.

Table 1. 

Mean ± SEM Pairwise Category Discrimination Accuracy for Each of the Six Object Categories in Simulation 1 (“All Category-selective Units”), Simulation 2 (“Minus Units Maximally Active…”), and Simulation 3 (“Units Maximally Active to…”)

Region
Percent Correct Discrimination
Faces
Houses
Chairs
Bottles
Scissors
Shoes
All category-selective units 91.0 ± 0.06 87.3 ± 0.09 87.2 ± 0.08 83.7 ± 0.1 81.1 ± 0.1 80.0 ± 0.06 
Minus units maximally active to compared categories 90.8 ± 0.12 85.3 ± 0.19 88.1 ± 0.14 85.8 ± 0.15 80.8 ± 0.13 80.5 ± 0.09 
Units maximally active to: 
 Faces 92.4 ± 0.07 80.1 ± 0.22 80.2 ± 0.35 82.9 ± 0.19 76.8 ± 0.26 73.7 ± 0.28 
 Houses 86.5 ± 0.16 86.2 ± 0.35 86.2 ± 0.17 81.3 ± 0.15 80.4 ± 0.15 79.7 ± 0.12 
 Small objects 88.0 ± 0.09 85.0 ± 0.15 84.3 ± 0.09 83.5 ± 0.13 81.7 ± 0.16 80.2 ± 0.08 
Region
Percent Correct Discrimination
Faces
Houses
Chairs
Bottles
Scissors
Shoes
All category-selective units 91.0 ± 0.06 87.3 ± 0.09 87.2 ± 0.08 83.7 ± 0.1 81.1 ± 0.1 80.0 ± 0.06 
Minus units maximally active to compared categories 90.8 ± 0.12 85.3 ± 0.19 88.1 ± 0.14 85.8 ± 0.15 80.8 ± 0.13 80.5 ± 0.09 
Units maximally active to: 
 Faces 92.4 ± 0.07 80.1 ± 0.22 80.2 ± 0.35 82.9 ± 0.19 76.8 ± 0.26 73.7 ± 0.28 
 Houses 86.5 ± 0.16 86.2 ± 0.35 86.2 ± 0.17 81.3 ± 0.15 80.4 ± 0.15 79.7 ± 0.12 
 Small objects 88.0 ± 0.09 85.0 ± 0.15 84.3 ± 0.09 83.5 ± 0.13 81.7 ± 0.16 80.2 ± 0.08 

In each row, responses of only those units belonging to a particular subset of units in the object encoding layer (defined on the left) were included in the analysis.

This simulation also closely approximates an analysis performed by Spiridon and Kanwisher (2002). These authors noted that, if the same set of images are used to generate the brain patterns in both halves of the data between which correlations were measured, good category discrimination performance may simply reflect the reliability of BOLD responses to low-level features, rather than the presence of category-level information. To establish the presence of category information, Spiridon and Kanwisher replicated the analysis of Haxby et al. using different stimuli from the same category to generate each half of the data; like Haxby et al., they found good discrimination performance. In Simulation 1, we followed Spiridon and Kanwisher's procedure for dividing the data, using nonidentical image sets to generate the halves (see Methods). Our finding of good classification performance despite nonidentical images sets therefore resembles the Spiridon and Kanwisher (2002) result. Moreover, Spiridon and Kanwisher replicated the finding of Haxby et al. that faces and houses were more accurately discriminated than other object categories (as did O'Toole et al., 2005, in a reanalysis of the data from Haxby et al., 2001) and argued that this indicated a “special” status for faces and houses in visual cortex. However, we were able to simulate this result in a computational model whose architecture and processing algorithms confer no special status on faces or houses. There was no face or house module in the model and no special process employed to learn these category representations.

Simulation 2: Category Discrimination Using Nonmaximally Active Voxels

We replicated a second finding of Haxby et al. (2001): Information about a particular object category is not contained solely in the cortical region responding maximally to that category. Haxby et al. assessed pairwise category discrimination using patterns of brain responses from which the voxels maximally responsive to the two categories being discriminated had been removed. For example, in the discrimination of houses and shoes, all voxels maximally activated by houses or by shoes were removed from the analysis. Category identification was still possible based on patterns containing only nonmaximal responses; the overall accuracy of pairwise discriminations (94%) was only slightly reduced relative to the case where all object-selective cortex was included in the analysis (96%). We replicated this analysis in the model and also found that good discrimination performance was maintained (Table 1, second row). We found overall discrimination accuracy of 85% when units maximally active to the object classes being discriminated were removed (Table 2, second row), compared with 85% when all responses were included (Simulation 1; Table 2, first row).

Table 2. 

Average Pairwise Category Discrimination Accuracies, from Simulations 1–3 (Column 1), Haxby et al. (2001) (Column 2), and Spiridon and Kanwisher (2002) (Column 3)

Region
Average Percent Correct Discrimination
Simulations
Haxby et al. (2001)
Spiridon and Kanwisher (2002)
All category-selective units 85.0 95.0 96.0 
Minus units maximally active to compared categories 85.2 93.8 – 
Units maximally active to: 
 Faces 81.0 83.0 – 
 Houses 83.4 92.7 – 
 Small objects 83.8 93.2 – 
Region
Average Percent Correct Discrimination
Simulations
Haxby et al. (2001)
Spiridon and Kanwisher (2002)
All category-selective units 85.0 95.0 96.0 
Minus units maximally active to compared categories 85.2 93.8 – 
Units maximally active to: 
 Faces 81.0 83.0 – 
 Houses 83.4 92.7 – 
 Small objects 83.8 93.2 – 

Scores are averages across all six object categories.

Again, this analysis was very similar to one performed by Spiridon and Kanwisher (2002). In that study, the authors reported that the overall category discrimination accuracy obtained using all visually active voxels was not substantially altered by the removal of face-selective or house-selective voxels from the patterns analyzed.

Simulation 3: Category Discrimination Using Only Maximally Active Voxels

In this simulation, we replicated a third result from Haxby et al. (2001): Voxels in regions that are maximally responsive to a particular object category contain sufficient information about other object categories to discriminate between those categories. In their analysis, Haxby et al. took sets of voxels that were maximally (and differentially) activated by one category: faces, houses, or a category defined as “small man-made objects” (i.e., chairs, bottles, scissors, and shoes). For each set of category-selective voxels, they determined the discriminability of all pairs of object categories, finding good overall performance in each region, ranging from 83% to 94%. We replicated this analysis using simulated activation patterns and similarly found good discrimination performance, ranging from 81% to 84% across the three regions (Table 1, rows 3–5; Table 2, rows 3–5), indicating that units in the object representation layer in the model carry information about the type of object being viewed even when they respond maximally to other categories.

Simulation 4: Category Discrimination Using Maximally Active Voxels, with Number of Units Held Constant

On the basis of the three simulations described, Haxby et al. concluded that many different object categories—including those with limited biological significance during our evolutionary history, such as scissors and shoes—are represented robustly in the ventral temporal cortex by distinct patterns of responses that are overlapping and distributed in nature. They suggested that the information representing each category was not restricted to a small region in which voxels responded maximally to that category but, instead, that submaximal responses outside that region are also an integral part of the category representation. Moreover, they argued that voxels responding maximally to one particular category (such as faces or houses) also contribute to the representation of other object categories, rather than possessing a specialized, category-specific function.

Spiridon and Kanwisher's (2002) study was very similar to that of Haxby et al. (2001), but contrary to Haxby et al., they interpreted their results in favor of cortical specialization for face and place processing. One important analysis that provided evidence in favor of specialization for faces involved a modified replication of the third Haxby et al. analysis simulated above. In this analysis, Spiridon and Kanwisher (2002) investigated whether cortical regions specialized for processing object categories other than faces and houses could be found if the category-selective region was not constrained to be a spatially contiguous set of voxels (finding a cluster of voxels maximally active to faces or houses is a typical constraint in localizing the FFA and the parahippocampal place area, but such clusters typically do not exist for other object categories). They tested for the existence of distributed selective regions by examining discrimination performance within the set of 30 voxels “most selective” for each category, choosing those voxels without heed to their location in the cortex. Their analysis was therefore very similar to Simulation 3, except that the size of the set of voxels was equated across category-selective regions. Spiridon and Kanwisher (2002) found that patterns across face-selective voxels produced better accuracy on face discriminations than on discriminations involving other stimulus classes and, critically, that this preferred-category advantage was not true for any other set of 30 voxels (e.g., in the shoe-selective cluster, discrimination of shoes was not superior to that of other categories). The authors took this as evidence for specialized neural processing for faces that does not exist for other man-made categories.

We replicated this analysis, choosing only the 30 most category-selective units in the object representation layer of the model for each category-selective region (see Methods). As seen in Table 3, we find a qualitatively similar pattern of results: In the face-selective units, face discriminations are more accurate than those of other categories, but this preferred-category advantage does not exist in any other category-selective region. The simulation of this finding provides a key insight for interpreting the original finding from the empirical fMRI data. In the model used to simulate the activation patterns, we assumed no anatomical module for faces in the network architecture, nor did we implement any specialized processing mechanism for faces that differed from the mechanism used to learn about and generate activation patterns for other object categories. Yet, the pattern of better face discrimination in the face-selective units emerged spontaneously. Postulation of a specialized processing mechanism for faces is therefore not necessary to explain this result in the empirical fMRI data. Assuming that brain patterns, like the activation patterns in our neurocomputational model, reflect the natural similarity space in which the stimuli viewed by human participants reside (Kriegeskorte et al., 2008; O'Toole et al., 2005), this pattern of results emerges simply on the basis of the inherent properties of the stimuli themselves, and the topographic manner in which the brain (and our model) represents these properties.

Table 3. 

Mean ± SEM Pairwise Category Discrimination Accuracy, Simulation 4


Percent Correct Discrimination
Faces
Houses
Chairs
Bottles
Scissors
Shoes
30 units most selective for: 
 Faces 85.4 ± 0.96 74.9 ± 0.62 79.0 ± 0.92 81.2 ± 0.64 72.0 ± 0.57 69.3 ± 0.61 
 Houses 81.9 ± 0.53 81.8 ± 0.74 82.9 ± 0.77 74.5 ± 0.41 75.1 ± 0.54 75.3 ± 0.59 
 Chairs 80.3 ± 0.73 84.9 ± 0.43 76.5 ± 0.84 76.7 ± 0.63 77.1 ± 0.68 74.9 ± 0.46 
 Bottles 82.3 ± 0.74 76.1 ± 0.66 76.7 ± 0.89 79.0 ± 0.52 78.9 ± 0.62 76.2 ± 0.67 
 Scissors 82.2 ± 0.46 81.3 ± 0.54 75.0 ± 0.91 83.6 ± 0.4 76.5 ± 0.53 80.3 ± 0.67 
 Shoes 80.7 ± 0.63 80.9 ± 0.58 78.7 ± 0.66 85.0 ± 0.46 84.9 ± 0.57 69.3 ± 0.56 

Percent Correct Discrimination
Faces
Houses
Chairs
Bottles
Scissors
Shoes
30 units most selective for: 
 Faces 85.4 ± 0.96 74.9 ± 0.62 79.0 ± 0.92 81.2 ± 0.64 72.0 ± 0.57 69.3 ± 0.61 
 Houses 81.9 ± 0.53 81.8 ± 0.74 82.9 ± 0.77 74.5 ± 0.41 75.1 ± 0.54 75.3 ± 0.59 
 Chairs 80.3 ± 0.73 84.9 ± 0.43 76.5 ± 0.84 76.7 ± 0.63 77.1 ± 0.68 74.9 ± 0.46 
 Bottles 82.3 ± 0.74 76.1 ± 0.66 76.7 ± 0.89 79.0 ± 0.52 78.9 ± 0.62 76.2 ± 0.67 
 Scissors 82.2 ± 0.46 81.3 ± 0.54 75.0 ± 0.91 83.6 ± 0.4 76.5 ± 0.53 80.3 ± 0.67 
 Shoes 80.7 ± 0.63 80.9 ± 0.58 78.7 ± 0.66 85.0 ± 0.46 84.9 ± 0.57 69.3 ± 0.56 

In each row, activation patterns across only the 30 most selective units for the indicated category were used in the analysis. The highest accuracy for each set of units (i.e., in each row) is underlined. Cells on the diagonal are indicated in bold, for clarity. If any set of category-selective units possesses a preferred-category discrimination advantage, the cell on the diagonal contains the highest discrimination accuracy in the row, and is therefore shown in bold and underlined. This is true only for face-selective units, as in Spiridon and Kanwisher (2002).

Exploratory Analyses of Simulation 4

To determine what properties of the object representations in the network allowed us to replicate the results from Spiridon and Kanwisher, we explored the model's representations by (1) examining the BCCs (the similarity of patterns in one category to patterns in other categories) and WCCs (the similarity of patterns elicited by objects in the same category) for each object category, (2) constructing dendrograms to depict the similarity of object activation patterns, and (3) visually examining plots of simulated activation patterns.

BCC and WCC

The accuracy of a category discrimination increases as the WCC of activation patterns increases or as the BCC of the activation patterns decreases (see Methods, Equation 3). For example, in a pairwise discrimination between faces and shoes, the higher the correlation between activation patterns caused by faces in the halves of the data (the WCC) or between activation patterns caused by shoes in two halves of the data, the better the discrimination of faces and shoes. In contrast, the higher the correlation between faces in one half of the data and shoes in the other half of the data (the BCC), the poorer the discrimination of faces and shoes. Figure 4 shows, for each subregion of the grid analyzed in Simulation 4 (i.e., the six sets of 30 units most selective for each category), the mean WCC and BCC values for activation patterns elicited by stimuli in the six categories. In Figure 4, for each set of 30 units, the WCC and BCC values (first and third rows) are shown directly above the discrimination accuracy (second and fourth rows) for each category of activation patterns. Categories with high WCCs and low BCCs yielded good discrimination scores, but the BCC value tended to have a greater influence on discrimination accuracy. For example, in face units, although houses, bottles, and scissors have slightly higher WCC values than face stimuli, none of these categories have a low value of BCC as faces do. Therefore, the superior discrimination of faces in the face units seems largely driven by the fact that activation patterns caused by faces are the most distinct from other categories' activation patterns. The powerful influence of the BCC is also observed for chair patterns across face units: Good discrimination accuracy is obtained despite low WCC values, presumably owing to the relatively low BCC values for chairs in these units. Thus, the model's account of the finding by Spiridon and Kanwisher—an advantage in face-selective voxels for the discrimination of faces over other categories, which does not exist for any other category—is that face patterns across the face units are maximally distinct from the patterns of other categories, whereas this is not true for preferred-category patterns in any other set of units.

Figure 4. 

WCC and BCC values for activation patterns in the object encoding layer of the model. For each set of 30 units used in Simulation 4—the units maximally selective for faces, houses, chairs, bottles, scissors, and shoes—the upper plot (row 1 or 3) shows the WCC and BCC values for each class of stimulus, and the lower plot (row 2 or 4) shows the discrimination accuracy for that stimulus class. Discrimination accuracy is based on WCC and BCC values, according to Equation 3. Higher WCC and lower BCC lead to better discrimination scores.

Figure 4. 

WCC and BCC values for activation patterns in the object encoding layer of the model. For each set of 30 units used in Simulation 4—the units maximally selective for faces, houses, chairs, bottles, scissors, and shoes—the upper plot (row 1 or 3) shows the WCC and BCC values for each class of stimulus, and the lower plot (row 2 or 4) shows the discrimination accuracy for that stimulus class. Discrimination accuracy is based on WCC and BCC values, according to Equation 3. Higher WCC and lower BCC lead to better discrimination scores.

Close modal

Dendrograms

To visualize how face representations across face-selective units are self-similar and distinct from other categories, we constructed dendrograms depicting the clustering (i.e., similarity) of activation patterns elicited by individual stimuli across the six sets of 30 category-selective units from Simulation 4 (Figure 5). The six separate dendrograms, one for each set of category-selective units, include patterns generated by a single, typical model simulation. In each dendrogram, eight example activation patterns from the 40 stimuli per category were randomly chosen for inclusion, because it was not possible to depict more than eight stimuli per category in a single plot.

Figure 5. 

Dendrograms showing the clustering of activation patterns elicited by individual stimuli across the six sets of 30 category-selective units defined in Simulation 4. Dendrogram leaves (x axis) are labeled and color-coded according to the category of the activation pattern (f = faces, h = houses, c = chairs, b = bottles, x = scissors, and s = shoes). In each dendrogram, the smallest possible cluster of patterns that contains all eight exemplars belonging to the preferred category is shown in color, with thickened lines. For all sets of units except face and bottle units, this cluster includes all 48 stimuli; that is, for all categories except faces and bottles, the patterns elicited by stimuli in a given category were not highly similar across the units that were maximally activated by that category. For bottle units, the cluster encompassing all eight bottles contains 34 of the 48 the stimuli, that is, there were 26 nonbottle stimuli that elicited patterns that were at least as similar to the bottles as some bottles were to each other, suggesting relatively poor clustering of bottle patterns in terms of similarity. For face units, the cluster encompassing all eight faces contains only 11 of 48 stimuli, suggesting a tight clustering of face patterns.

Figure 5. 

Dendrograms showing the clustering of activation patterns elicited by individual stimuli across the six sets of 30 category-selective units defined in Simulation 4. Dendrogram leaves (x axis) are labeled and color-coded according to the category of the activation pattern (f = faces, h = houses, c = chairs, b = bottles, x = scissors, and s = shoes). In each dendrogram, the smallest possible cluster of patterns that contains all eight exemplars belonging to the preferred category is shown in color, with thickened lines. For all sets of units except face and bottle units, this cluster includes all 48 stimuli; that is, for all categories except faces and bottles, the patterns elicited by stimuli in a given category were not highly similar across the units that were maximally activated by that category. For bottle units, the cluster encompassing all eight bottles contains 34 of the 48 the stimuli, that is, there were 26 nonbottle stimuli that elicited patterns that were at least as similar to the bottles as some bottles were to each other, suggesting relatively poor clustering of bottle patterns in terms of similarity. For face units, the cluster encompassing all eight faces contains only 11 of 48 stimuli, suggesting a tight clustering of face patterns.

Close modal

Dendrograms connect objects according to similarity in a hierarchical tree. The height of the inverted U connecting each pair of objects (or object groups) indicates the similarity of those objects. The dendrograms in Figure 5 show that, for all sets of units except face and bottle units, stimuli in the preferred category are no more similar to each other and no more distinct from other categories than stimuli in the nonpreferred categories. For activation patterns across the bottle units, bottle patterns are somewhat self-similar and distinct from other categories, in that all bottle patterns reside within a subcluster of the dendrogram that contains about two thirds of the patterns. For activation patterns across the face units, faces are even more self-similar and distinct: All eight face patterns fall within a subcluster that contains only 11 of the 48 patterns. The distinctiveness of a category can be quantified in terms of “cluster purity,” which we define as the proportion of leaves corresponding to preferred category patterns within the smallest cluster that contains all eight preferred category patterns in the dendrogram. For face units, the mean cluster purity was 0.68 (across all dendrograms generated for all 100 simulated networks). For house, chair, bottle, scissor, and shoe units, the mean cluster purities were 0.22, 0.18, 0.24, 0.18, and 0.2, respectively. The dendrograms in Figure 5 possess “purity” values representative of the mean cluster purity for each set of 30 category-selective units. In summary, only across face units did we consistently find that activation patterns elicited by stimuli from the preferred category were well separated into a relatively category-pure cluster.

Activation Patterns

We plotted the simulated activation patterns to visualize the properties of the stimulus representations (Figures 3, 6, and 7). In Figures 6 and 7, all plots are taken from a single simulation (network size = 40 × 40 units), which was also used to generate Figures 2 and 3; the regions of maximal response to faces and scissors in Figure 6 can therefore be compared with the map of category preferences in Figure 2. Whereas faces elicit remarkably consistent activation patterns across different individuals and different views, the activation patterns elicited by scissors vary considerably, leading to higher WCCs for faces than for scissors and other nonface categories. However, this effect is diluted in the MVPA results by the fact that, before calculating WCCs, patterns in each category are averaged into halves, removing much inter-individual and inter-view variability. It is therefore more instructive to examine plots of activation patterns that have been averaged across halves of the data.

Figure 6. 

Activation of units in the object encoding layer of the model, elicited by individual face (top row) and scissor (bottom row) stimuli. Each plot is generated by a unique exemplar pictured from a unique view. The black polygons delineate the regions of units that responded maximally to faces (top row) and scissors (bottom row). Activation patterns elicited by different scissor stimuli exhibit much more variability than activation patterns elicited by different face stimuli.

Figure 6. 

Activation of units in the object encoding layer of the model, elicited by individual face (top row) and scissor (bottom row) stimuli. Each plot is generated by a unique exemplar pictured from a unique view. The black polygons delineate the regions of units that responded maximally to faces (top row) and scissors (bottom row). Activation patterns elicited by different scissor stimuli exhibit much more variability than activation patterns elicited by different face stimuli.

Close modal
Figure 7. 

Mean activation of units in the object encoding layer of the model, in each half of the data, for the face, scissor, and shoe categories. Activation patterns were assigned to the halves of the data according to the “different views” protocol. Regions of the grid that are activated highly by faces tend not to be activated by shoes or scissors, and vice versa, whereas activation patterns elicited by shoes and scissors tend to include many units in common.

Figure 7. 

Mean activation of units in the object encoding layer of the model, in each half of the data, for the face, scissor, and shoe categories. Activation patterns were assigned to the halves of the data according to the “different views” protocol. Regions of the grid that are activated highly by faces tend not to be activated by shoes or scissors, and vice versa, whereas activation patterns elicited by shoes and scissors tend to include many units in common.

Close modal

Figure 7 shows mean activation values across patterns in the halves of the data, for the face, scissor, and shoe stimuli. Two important properties of the representations are evident. First, for shoes, even when averaged across half of the stimuli in the category, the two resulting average activation patterns are remarkably different, a fact likely to lead to low WCC values and poor discrimination accuracy. Second, regions of the grid that are activated by faces are in general not activated by scissors or shoes, and vice versa; in contrast, many regions of the grid activated by scissors are also activated by shoes. In other words, patterns caused by faces are quite distinct from patterns caused by other categories, eliciting high values only in units that do not contribute to the representations of other categories; this is not true for scissor and shoe patterns. This unique property of face representations in the model accounts for the high BCCs for faces. Moreover, we note that this is particularly true in the portion of the grid that is maximally activated by faces (the top-left corner in Figures 2 and 6), that is, the area from which the subset of 30 most-selective face units were drawn in Simulation 4.

We used a model of object processing in visual cortex to simulate MVPA results from two fMRI studies: Spiridon and Kanwisher (2002) and Haxby et al. (2001). The model employs a neurocomputationally plausible mechanism to govern the unsupervised development of visual object representations into a topographically organized map. We developed a method for analyzing the activation patterns elicited by objects from different categories across the topographic object-encoding layer of the model, in a manner analogous to the MVPA procedures of the fMRI studies. Replicating Haxby et al., we demonstrated that activation patterns in the model contain sufficient information to perform pairwise category discrimination under three conditions: using all object-selective units in the object-encoding layer, using only object-selective units that were not maximally activated by the categories taking part in the discrimination, and using subsets of object-selective units that were maximally activated by a single category or group of categories. Replicating Spiridon and Kanwisher (2002), we demonstrated that, for activation patterns across the 30 units most selective for faces, faces were better discriminated than any other category, and this preferred-category discrimination advantage did not exist for any other set of 30 units most selective for a particular object category.

The successful replication of these empirical results indicates that the object representations of the model share important properties (in particular, the similarity relationships that determine category discriminability) with object representations in the ventral temporal cortex measured with fMRI. Moreover, the method we devised for analyzing simulated activation patterns is a viable approach for predicting the discriminability of brain patterns. The approach is therefore suitable for testing the ability of theoretical assumptions concerning neural architecture and neural processing mechanisms to explain patterns of fMRI data.

Three of the empirical results we simulated have been interpreted as evidence for distributed object processing (Simulations 1–3), and the fourth as evidence for a specialized face-processing module (Simulation 4). In simulating key findings on both sides, the present computational study offers an explanation for the contradiction that seems to be posed by the empirical data. In the model, there was neither an anatomical module for the category of faces nor any specialized processing mechanism or distinct representational assumptions for faces. Rather, the MVPA results that arose in the model were driven by the inherent similarity properties of the face and object stimuli themselves, in combination with a neurocomputationally plausible learning rule that produced topography in the model representations mimicking that seen in visual cortex. The simulations therefore demonstrate that it is not necessary to posit specialized neural mechanisms to account for the finding by Spiridon and Kanwisher (2002) that there is a face-discrimination advantage across voxels maximally activated by faces, whereas there is no such preferred-category discrimination advantage across voxels maximally activated by other categories.

Examination of the WCC and BCC values for activation patterns in the different categories revealed that pairwise category discriminability in the model was most strongly influenced by the presence of low BCCs. Face patterns across face-selective units showed particularly low BCCs, meaning activation patterns caused by faces were the most distinct from other categories' activation patterns, likely making an important contribution to the result observed in Simulation 4. This account of the simulated data was corroborated by the hierarchical cluster analysis carried out on individual stimulus activation patterns from the model: Only across face units did we consistently find that activation patterns elicited by stimuli from the preferred category (faces) were well separated into a relatively category-pure cluster (Figure 5). Finally, in line with the notion that faces' distinctiveness as a stimulus class drove the result seen in Simulation 4, the plots of Figure 7 indicated that patterns caused by faces were quite distinct from patterns caused by other classes of stimuli, eliciting high values only in units that do not contribute to the representations of other categories.

We note that the extreme distinctiveness of face representations in the model also seems likely to be a property of the neural representations of faces. In the fMRI literature, category-selective voxels are often defined as those that respond more than twice as strongly to the preferred category than to other categories (Kanwisher, 2010; Spiridon & Kanwisher, 2002). What the present computational study shows is that the superior discriminability of faces in face-preferring regions of cortex might emerge simply on the basis of the raw stimulus properties themselves, provided that ventral temporal cortex representations reflect stimulus similarity space to some extent (an idea finding support in O'Toole et al., 2005, and Kriegeskorte et al., 2008).

We emphasize that our simulations cannot rule out a “face module” account of the fMRI data. However, a stimulus similarity explanation that does not appeal to cortical modules is a more parsimonious account, given that a specialized face-processing module is not necessary to explain the empirical results. Moreover, our results are in line with those of another computational study by Jiang et al. (2006), in which a shape-based account of object processing that assumes no face-specific processes can account for both behavioral findings (such as the face-inversion and “configural” effects) and a series of fMRI results indicating face selectivity in the FFA.

A key demonstration made by the present study is that these photographic images of everyday objects do not constitute a well-controlled stimulus set. Our simulation experiment deliberately eschews not only any specialized architecture or processing mechanism for faces but also many aspects of everyday learning and experience that are different for faces than for other classes of object (e.g., more time spent looking at faces or lifelong practice with subordinate-level, rather than category-level, discriminations). This choice was made to isolate and test the ability of the raw stimulus properties inherent in the images presented to human participants to account for the MVPA findings. Because of the simplicity of the model, some properties of the representations measured in ventral temporal cortex with fMRI are not reflected in our simulated activation patterns, for example, the finding that there are many more voxels selective for faces and houses than for other classes of object (e.g., see Figure 6, Spiridon & Kanwisher, 2002). It is remarkable that, despite the discrepancy between the object representations in our networks and those measured with fMRI for this seemingly important property, the MVPA results are in accord. Indeed, one would have expected that having fewer voxels for faces would make it more difficult to obtain this result. This demonstrates that the relative number of voxels selective for the various object categories is not a property of the representations that is necessary to account for the MVPA findings; stimulus properties alone can provide an explanation.

The stimuli used by both Haxby et al. (also used in this study) and Spiridon and Kanwisher were not standardized in terms of within- and between-category similarity; they simply reflected the inherent variability in similarity that exists across real-world object categories. Sets of images such as these cannot be used to test whether there are specialized neural processing mechanisms for faces that do not apply to other categories, because any appearance in the fMRI data of qualitative differences in the neural representations of faces might simply be attributable to inherent “properties” of faces as a class of visual stimulus (namely, that faces are in general highly similar to each other and very different from objects in other categories).

In summary, the novel computational technique that we present constitutes a valuable tool for the interpretation of fMRI data in terms of hypotheses concerning neurocognitive architecture and mechanisms. The present result offers a cautionary tale for the interpretation of MVPA results: It can be difficult to determine the consequences of a particular neurocognitive theory for patterns of brain activation in fMRI without explicit simulation of the hypothesized mechanisms. However, when the assumptions of a neurocognitive theory are instantiated in a concrete computational model and the consequences of those assumptions are simulated, it is possible to determine which inferences regarding neurocognitive mechanisms can and cannot be drawn from the empirical data.

Reprint requests should be sent to Rosemary A. Cowell, Psychology Department, Tobin Hall, University of Massachusetts, Amherst, MA 01003, or via e-mail: [email protected].

1. 

This is not to say that there may not be other data that can support this claim.

2. 

We also trained a set of networks on seven categories including cats and performed MVPA on those networks. The MVPA results for the six categories we report were not qualitatively (or substantially quantitatively) affected by the addition of cats. We did not include cats in the simulations we report because networks did not perform well behaviorally at categorizing cats, likely because of the greater variability in viewing angle from which images in this set were pictured, relative to the more restricted range of viewing angles within other stimulus categories.

3. 

We replicated all MVPA simulations reported here using a set of novel stimuli. The novel images had been used only as holdout images to test categorization performance during training; networks' weights had never been updated in response to the presentation of these stimuli. All qualitative aspects of the MVPA results we report for training stimuli were the same using these novel stimuli. We report the results from analyses with training stimuli because this afforded a larger stimulus set with which to perform MVPA (because many more items were required for the training set than for the holdout set).

4. 

All of the qualitative patterns of results that emerged in the MVPA results averaged over four network sizes were also evident in at least three of the four network sizes examined independently. The vast majority of the qualitative patterns we report were in fact evident in all four network sizes tested.

Aitkin
,
L. M.
,
Merzenich
,
M. M.
,
Irvine
,
D. R.
,
Clarey
,
J. C.
, &
Nelson
,
J. E.
(
1986
).
Frequency representation in auditory cortex of the common marmoset (Callithrix jacchus jacchus).
Journal of Comparative Neurology
,
252
,
175
185
.
Cohen
,
L.
,
Lehericy
,
S.
,
Chochon
,
F.
,
Lemer
,
C.
,
Rivaud
,
S.
, &
Dehaene
,
S.
(
2002
).
Language-specific tuning of visual cortex? Functional properties of the visual word form area.
Brain
,
125
,
1054
1069
.
Cowell
,
R. A.
,
Huber
,
D. E.
, &
Cottrell
,
G. W.
(
2009
).
Virtual brain reading: A connectionist approach to understanding fMRI.
Proceedings of the 31st Annual Conference of the Cognitive Science Society (pp. 212–217), Washington, DC
.
Cynader
,
M. S.
,
Swindale
,
N. V.
, &
Matsubara
,
J. A.
(
1987
).
Functional topography in cat area 18.
Journal of Neuroscience
,
7
,
1401
1413
.
Dailey
,
M. N.
, &
Cottrell
,
G. W.
(
1999
).
Organization of face and object recognition in modular neural network models.
Neural Networks
,
12
,
1053
1074
.
Daugman
,
J. G.
(
1985
).
Uncertainty relation for resolution in space, spatial frequency, and orientation optimized by two-dimensional visual cortical filters.
Journal of the Optical Society of America A—Optics Image Science and Vision
,
2
,
1160
1169
.
Downing
,
P. E.
,
Jiang
,
Y. H.
,
Shuman
,
M.
, &
Kanwisher
,
N.
(
2001
).
A cortical area selective for visual processing of the human body.
Science
,
293
,
2470
2473
.
Epstein
,
R.
, &
Kanwisher
,
N.
(
1998
).
A cortical representation of the local visual environment.
Nature
,
392
,
598
601
.
Hanson
,
S. J.
,
Matsuka
,
T.
, &
Haxby
,
J. V.
(
2004
).
Combinatorial codes in ventral temporal lobe for object recognition: Haxby (2001) revisited: Is there a “face” area?
Neuroimage
,
23
,
156
166
.
Haxby
,
J. V.
,
Gobbini
,
M. I.
,
Furey
,
M. L.
,
Ishai
,
A.
,
Schouten
,
J. L.
, &
Pietrini
,
P.
(
2001
).
Distributed and overlapping representations of faces and objects in ventral temporal cortex.
Science
,
293
,
2425
2430
.
Ishai
,
A.
,
Ungerleider
,
L. G.
,
Martin
,
A.
, &
Haxby
,
J. V.
(
2000
).
The representation of objects in the human occipital and temporal cortex.
Journal of Cognitive Neuroscience
,
12(Suppl. 2)
,
35
51
.
Ishai
,
A.
,
Ungerleider
,
L. G.
,
Martin
,
A.
,
Schouten
,
J. L.
, &
Haxby
,
J. V.
(
1999
).
Distributed representation of objects in the human ventral visual pathway.
Proceedings of the National Academy of Sciences, U.S.A.
,
96
,
9379
9384
.
Jiang
,
X.
,
Rosen
,
E.
,
Zeffiro
,
T.
,
Vanmeter
,
J.
,
Blanz
,
V.
, &
Riesenhuber
,
M.
(
2006
).
Evaluation of a shape-based model of human face discrimination using fMRI and behavioral techniques.
Neuron
,
50
,
159
172
.
Kanwisher
,
N.
(
2010
).
Functional specificity in the human brain: A window into the functional architecture of the mind.
Proceedings of the National Academy of Sciences, U.S.A.
,
107
,
11163
11170
.
Kanwisher
,
N.
,
McDermott
,
J.
, &
Chun
,
M. M.
(
1997
).
The fusiform face area: A module in human extrastriate cortex specialized for face perception.
Journal of Neuroscience
,
17
,
4302
4311
.
Kohonen
,
T.
(
1982
).
Self-organized formation of topologically correct feature maps.
Biological Cybernetics
,
43
,
59
69
.
Kohonen
,
T.
(
1984
).
Self-organization and associative memory.
Berlin
:
Springer-Verlag
.
Kriegeskorte
,
N.
,
Mur
,
M.
,
Ruff
,
D. A.
,
Kiani
,
R.
,
Bodurka
,
J.
,
Esteky
,
H.
,
et al
(
2008
).
Matching categorical object representations in inferior temporal cortex of man and monkey.
Neuron
,
60
,
1126
1141
.
Krubitzer
,
L. A.
, &
Calford
,
M. B.
(
1992
).
Five topographically organized fields in the somatosensory cortex of the flying fox: Microelectrode maps, myeloarchitecture, and cortical modules.
Journal of Comparative Neurology
,
317
,
1
30
.
Lemon
,
R. N.
(
1981
).
Variety of functional organization within the monkey motor cortex.
Journal of Physiology
,
311
,
521
540
.
Logothetis
,
N. K.
(
2008
).
What we can do and what we cannot do with fMRI.
Nature
,
453
,
869
878
.
McCarthy
,
G.
,
Puce
,
A.
,
Gore
,
J. C.
, &
Allison
,
T.
(
1997
).
Face-specific processing in the human fusiform gyrus.
Journal of Cognitive Neuroscience
,
9
,
605
610
.
O'Toole
,
A. J.
,
Jiang
,
F.
,
Abdi
,
H.
, &
Haxby
,
J. V.
(
2005
).
Partially distributed representations of objects and faces in ventral temporal cortex.
Journal of Cognitive Neuroscience
,
17
,
580
590
.
Petersen
,
S. E.
,
Fox
,
P. T.
,
Snyder
,
A. Z.
, &
Raichle
,
M. E.
(
1990
).
Activation of extrastriate and frontal cortical areas by visual words and word-like stimuli.
Science
,
249
,
1041
1044
.
Sanger
,
T. D.
(
1989
).
Optimal unsupervised learning in a single-layer linear feedforward neural network.
Neural Networks
,
2
,
459
473
.
Spiridon
,
M.
, &
Kanwisher
,
N.
(
2002
).
How distributed is visual category information in human occipito-temporal cortex? An fMRI study.
Neuron
,
35
,
1157
1165
.
Swindale
,
N. V.
,
Matsubara
,
J. A.
, &
Cynader
,
M. S.
(
1987
).
Surface organization of orientation and direction selectivity in cat area 18.
Journal of Neuroscience
,
7
,
1414
1427
.
Tong
,
M. H.
,
Joyce
,
C. A.
, &
Cottrell
,
G. W.
(
2008
).
Why is the fusiform face area recruited for novel categories of expertise? A neurocomputational investigation.
Brain Research
,
1202
,
14
24
.