Skip to main content
PLOS Computational Biology logoLink to PLOS Computational Biology
. 2024 Feb 26;20(2):e1011887. doi: 10.1371/journal.pcbi.1011887

Disentangled deep generative models reveal coding principles of the human face processing network

Paul Soulos 1, Leyla Isik 1,*
Editor: Jean Daunizeau2
PMCID: PMC10919870  PMID: 38408105

Abstract

Despite decades of research, much is still unknown about the computations carried out in the human face processing network. Recently, deep networks have been proposed as a computational account of human visual processing, but while they provide a good match to neural data throughout visual cortex, they lack interpretability. We introduce a method for interpreting brain activity using a new class of deep generative models, disentangled representation learning models, which learn a low-dimensional latent space that “disentangles” different semantically meaningful dimensions of faces, such as rotation, lighting, or hairstyle, in an unsupervised manner by enforcing statistical independence between dimensions. We find that the majority of our model’s learned latent dimensions are interpretable by human raters. Further, these latent dimensions serve as a good encoding model for human fMRI data. We next investigate the representation of different latent dimensions across face-selective voxels. We find that low- and high-level face features are represented in posterior and anterior face-selective regions, respectively, corroborating prior models of human face recognition. Interestingly, though, we find identity-relevant and irrelevant face features across the face processing network. Finally, we provide new insight into the few "entangled" (uninterpretable) dimensions in our model by showing that they match responses in the ventral stream and carry information about facial identity. Disentangled face encoding models provide an exciting alternative to standard “black box” deep learning approaches for modeling and interpreting human brain data.

Author summary

We use a class of interpretable deep neural network models, disentangled variational autoencoders (dVAEs), to analyze human fMRI data. We find that a dVAE learns human interpretable dimensions of faces, such as lighting, expression, and hairstyle, and provides as good a match to human fMRI data as matched, non-disentangled models. Our disentangled encoding approach allows us to map different disentangled features to ROI and voxel activity. A decoding analysis confirms that the model separates identity relevant and irrelevant information and reveals that the remaining entangled dimensions contain identity-relevant information. Together these results highlight the use of disentangled models for more interpretable fMRI encoding than standard deep learning models.

Introduction

Humans are highly skilled at recognizing faces despite the complex high dimensional space that face stimuli occupy and the many transformations they undergo. Some dimensions (such as 3D rotation and lighting) are constantly changing and thus irrelevant to recognizing a face, while others (such as facial features or skin tone) are generally stable and useful for recognizing an individual’s identity, and still others (such as hairstyle) can change but also offer important clues to identity. Face processing networks in the macaque and human brain have been thoroughly mapped [13] and many general coding principles have been identified, including separation of static vs. dynamic face representations [4,5] and increasing transformation invariance from posterior to anterior regions [6]. However, much is still unknown about the computations carried out across these regions, particularly in the human brain. Even fundamental information, such as how facial identity is represented, is still largely unknown [2]. This lack of understanding can be seen in the relatively poor decoding of face identity from fMRI data compared to other visual categories [7].

Recently, deep convolutional neural networks (DCNNs) trained on face recognition have been shown to learn effective face representations that provide a good match to human behavior [8], but such discriminatively trained models are difficult to interpret [9] and provide a poor match to human neural data [10]. Alternatively, deep generative models have been shown to provide a good match to human fMRI face processing data [11]. These models, however, transform faces into complex high dimensional latent spaces and thus suffer from the same lack of interpretability as standard DCNNs. Here we use a new class of deep generative models, disentangled representation learning models that isolate semantically meaningful factors of variation in individual latent dimensions, to understand the neural computations underlying human face processing.

Multiple disentangled representation learning models have been developed [1217], many of which are based on Variational Autoencoders (VAEs) [18]. These disentangled variational autoencoders (dVAEs) learn a latent space that “disentangles” different explanatory factors in the training distribution by enforcing statistical independence between latent dimensions during training [19]. Intriguingly, when applied to faces, dVAEs have been shown to learn latent dimensions that are not only statistically independent, but also isolate specific, interpretable face features.

dVAEs learn a latent representation that is compact and highly interpretable by humans, so we investigate complex face representations across the human brain using dVAEs in an encoding model framework. We find that representations in disentangled models match those found in the human face processing network at least as well or better than standard deep learning models without the disentanglement cost (including traditional VAEs and DCNNs). We then map the learned semantically meaningful dVAE dimensions to voxel responses and quantify their facial identity information, providing new insight into the models and the human face processing network.

Results

Disentangled generative models factor latent space into human-interpretable dimensions

We trained several dVAE models on the CelebA dataset [20], with the goal of selecting one as an encoding model of face-selective responses in the human brain. Like standard VAEs, these models have an encoder, which transforms an image into a lower-dimensional latent space via convolution, and a decoder, which aims to reconstruct the image from the latent representation (Fig 1A). The models were trained to minimize reconstruction error and had an additional training objective to maximize KL divergence between latent dimensions. Based on a hyperparameter search over previously published model architectures, number of latent dimensions, and model-specific disentanglement parameters to maximize disentanglement (see Methods M1), we selected FactorVAE [16] with 24 latent dimensions as our dVAE model.

Fig 1. Model overview and encoding procedure.

Fig 1

A) dVAE model overview. The model takes in a face image (left) and passes it through an encoder, consisting of several convolutional layers (left white trapezoid), to generate a latent vector (blue). Then a decoder (right white trapezoid) uses de-convolution to reconstruct the face image from the latent vector. Like a standard VAE, the dVAE is trained with a cost function to minimize reconstruction error. The dVAE has an additional term to maximize statistical independence (KL-divergence) between elements in the latent vector. B) Encoding procedure. A GLM is used to learn a linear mapping (beta weights) between the latent response to each training image and fMRI responses. At testing, a new test image is passed through the model to extract its 24-dimensional latent vector. This vector is then multiplied by the learned beta weights to generate a predicted voxel response. Because of license restriction, face images are representative of images in the CelebA dataset. Images from top to bottom are cropped from “2150881.png”, “woman-1867431_1280.jpg”, “1867175.png”, and “856125.png” from Pixabay.com and distributed under their content license.

After training the dVAE, the authors rated all dimensions by inspecting the faces generated by traversing values of a single latent dimension while keeping all others constant. These latent traversals were often highly interpretable, producing faces that seem to vary along a single dimension, such as facial expression or 3D rotation (Fig 2 and S1S2 Videos). Out of the 24 latent dimensions, the authors agreed on semantic labels for 16 (14 unanimously and two for a single rater, See Methods M2, Table 1), which included both identity-relevant (dimensions 8–12, 14–16) and irrelevant (dimensions 1–7, 13). These dimensions can also be separated roughly into lower-level visual dimensions that are not face-specific (1–7: lighting, image tone, background, rotation), and face-specific features (8–16), though we note this distinction is not entirely clear cut (see Discussion). The other 8 dimensions were considered entangled, containing multiple or uninterpretable transformations. In contrast to the dVAE, traversals from the standard VAE each contained several changing factors (S1S3 Videos).

Fig 2. Latent traversals for two dVAE dimensions.

Fig 2

Images generated by the dVAE when one latent dimension labeled as ‘smile’ (A) or ‘3D rotation’ (B) is varied from -1 to +1, and other dimensions are held constant. The primary change in the image corresponds to the labeled dimension, suggesting these dimensions are effectively disentangled and highly interpretable. Base images are from the CelebA dataset [20].

Table 1. Labels assigned to 24 dimensions by human annotators.

Annotators agreed on labels for 16 of 24 dimensions. Colors correspond to plots in Figs 56.

1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17–24
Lighting /Face width Image tone Background Background 3D rotation/ Lighting 3D rotation Elevation Hair part Hair Hair Hair Hairline Smile Skin tone Gender appearance Face width Entangled

We compared our dVAE to two control models. First, we used a standard entangled generative VAE matched in terms of training and hyper-parameters. Second, we used the penultimate layer of a popular DCNN, the discriminatively trained VGG-Face based on VGG16 [21,22]. To match model dimensions, we reduced the dimensionality of the VGG-Face representations to the first 24 principal components, which captured 70.7% of the variance. While the dVAE and VAE latent dimensions shared a similar geometry (CCA r = 0.92), the dVAE and VGG latent spaces were only moderately correlated (CCA r = 0.52), suggesting that discriminative versus generative training frameworks result in different face representations.

Disentangled models provide a good match to ventral face-selective regions

We used a publicly available fMRI dataset [11], where four subjects viewed roughly 8000 face images each a single time. Each subject also viewed 20 face test images between 40–60 times. This approach is in line with recent theoretical and empirical work highlighting the benefits of a small-n, condition-rich design [23,24]. Data were pre-processed and projected onto subjects’ individual cortical surfaces. We estimated a linear map between the latent representation of each model and the fMRI data via a generalized linear model (GLM) on the training data (Fig 1B). To predict fMRI responses to each held out test image, we extracted the latent representation for that test image from each model and multiplied them by the linear mapping learned in the GLM.

We evaluated encoding performance for three face-selective ROIs, the Fusiform Face Area (FFA), Occipital Face Area (OFA), and posterior Superior Temporal Sulcus (pSTS), as well as face-selective voxels across the whole brain, identified in a separate face-object localizer experiment (see Methods M4). Despite the additional disentanglement constraint, the dVAE model achieves similar encoding performance to the standard VAE and VGG in FFA and OFA (Fig 3 and S1 Table). At the group level, all models perform significantly above chance (p<0.001) in OFA and FFA. Additionally, both dVAE and VAE have significantly higher predictivity than VGG in the OFA and FFA at the group level (S1 Table). The models also performed similarly across all face-selective voxels in the brain (Figs 4, S1, and S2). None of the models provided consistently above chance accuracy in pSTS (Fig 3), perhaps due to the fact that all stimuli were static faces and lateral face regions have been shown to be selective for dynamic stimuli [4].

Fig 3. Encoding performance by ROI.

Fig 3

Average correlation between model prediction and true fMRI responses on held out test images for dVAE (blue), VAE (green), and VGG (red). Dots represent individual subject performance. Asterisks represent significant (p<0.05) results at the group level based on permutation test. Dashed line is average split half reliability across subjects on test image responses.

Fig 4. Whole brain encoding.

Fig 4

Encoding model performance in all face-selective voxels. Ventral view of two representative subjects for dVAE (left), VAE (center) and VGG-face (right).

Higher-level, identity-relevant dimensions are represented in more anterior face-selective regions

The main advantage of disentangled encoding models is the ability to examine how voxels respond to semantically meaningful dimensions. To do this, we performed preference mapping by predicting fMRI responses based on the dVAE latent vector and learned beta weights for each individual latent dimension. Preference mapping is similar to directly comparing the learned beta weights for each feature, but more robust since it is done on held out test data, and more interpretable since the outputs are bounded correlation values versus arbitrarily scaled beta weights [25]. High predictivity of a particular dimension in a particular brain region indicates that changes along that dimension predict changes in neural activity and does not necessarily mean that specific region codes for or is selective to that dimension.

We first performed preference mapping within each ROI (Fig 5). In the OFA, two dimensions, lighting/face width and image tone, were significantly predictive at the group level. These dimensions include lower-level visual changes, which are not face-specific, though we note the first dimension also contains information about face width, which may be predictive of identity (see Discussion). In FFA, one lower-level dimension (background) was significantly predictive at the group level. Additionally, two higher-level, face-specific dimensions, smile and skin tone, and one entangled dimension were also significantly predictive of FFA voxel responses at the group level. When comparing the two ROIs, OFA was significantly better predicted by lighting, image tone, and one entangled dimension, and FFA was significantly better predicted by two entangled dimensions. These FFA-dimensions included both identity-specific features like skin tone and changeable aspects of faces like expression. As with the full model performance in the STS, the performance from most individual dimensions is also worse than the other ROIs (S3 Fig).

Fig 5. Preference mapping results for each dimension and ROI.

Fig 5

Average feature predictivity across subjects is shown for each individual dimension. Dots represent individual subjects. Asterisks represents significance versus chance and > represents significantly greater response in one ROI versus the other (p < .05) at the group level. Dimensions are colored according to Table 1.

To understand how dimensions are represented across the brain, we can visualize their predictivity in a winner take all manner on the surface of the brain (S4 Fig). Similar to the ROI analysis, most posterior voxels were best predicted by image-level changes in background and lighting. More anterior regions, including FFA, and in some subjects, anterior temporal lobe (ATL) regions not included in our ROI analysis, also showed responses for face-specific dimensions like smile (light red) and identity-relevant dimensions like skin tone (light pink), hairstyle (light green, light yellow and light orange), and gender appearance (light purple). Some subjects also show anterior ventral voxels best predicted by visual features like background (dark orange). Interestingly, entangled dimensions (white) were predictive in face-selective voxels throughout the cortex.

Disentangled models isolate identity relevant face information

Another benefit of disentangled encoding models is the ability to study and group dimensions based on semantically meaningful attributes. One particularly important distinction for face processing is the separation of identity relevant factors (e.g., gender appearance, skin tone, and face shape) from identity irrelevant factors (e.g. lighting, viewpoint, and background). We decoded identity from our 20 test images using different subsets of dimensions: identity-relevant, identity-irrelevant, and entangled. Note that in our set, identity-relevant dimensions include all face-specific features identified above, with the exception of smile, which is not relevant to identity. Identity-relevant dimensions provided the highest identity decoding accuracy, almost equal to using all dimensions, whereas identity-irrelevant dimensions had the lowest, providing proof of concept that distinctions between our disentangled dimensions capture meaningful semantic information (Fig 6).

Fig 6. Facial identity decoding.

Fig 6

Pairwise decoding of face identity from different subsets of dimensions: 8 identity relevant dimensions, 8 identity irrelevant dimensions, 8 entangled dimensions, the combination of identity relevant with entangled dimensions, as well as all dimensions. Mean values shown as blue bars with individual subjects shown as dots. The dashed line at 50% indicates the level of chance performance.

The role of information contained in the remaining entangled dimensions of a disentangled model is an open question in AI, so we next sought to examine the extent of identity information in these dimensions. The entangled dimensions contained some identity information as illustrated by their above-chance decoding. However, entangled dimensions do not appear to capture information beyond the identity-relevant dimensions as shown by the similar decoding performances of identity-relevant features and the combination of identity-relevant and entangled features.

Discussion

We introduced a novel encoding framework for interpreting human fMRI data. Our method allows us to identify semantically meaningful dimensions in an unsupervised manner from large datasets. This disentanglement improves interpretability without a large degradation in encoding performance, as seen by the similar performance between the disentangled and standard VAEs. Our results also suggest that low- and high-level properties are represented in posterior versus anterior brain regions, consistent with prior data and models of face processing [2,2628]. Here we operationalized low versus high-level dimensions as those that are general visual changes (dimensions 1–7) versus face-specific changes (dimensions 8–16), though these are not necessarily identity-relevant. We note though that this distinction is not entirely clear cut as the 3D properties of faces are unique, and it is an open question whether the learned representations for changes in lighting or 3D rotation would generalize to other objects. In addition, properties like skin tone or face-width could also be considered low-level as they affect the overall pixel-level properties of the image. While we identified several identity-relevant dimensions in FFA, consistent with prior work [28,29], we also found sensitivity to several changeable aspects of faces, including expression, in FFA and other ventral face-selective voxels. These results add to mounting evidence [2] challenging the idea of a clear-cut distinction between identity and expression coding in ventral and lateral face regions [26,30,31]. We note, however, that these conclusions are largely based on analyses in OFA and FFA due to low reliability in lateral and anterior regions. An interesting question for future work is how disentangled model models match representations in the extended face processing network.

In addition to improving our understanding of the human brain, this work also yields new insights into representations learned in disentangled models. Through decoding analysis, we showed that the disentangled identity-relevant dimensions contain almost all the face identity information in the fMRI signals, providing novel support for disentanglement in these models. The nature of learned representations in models trained on naturalistic data is an open question in AI [32]. Our approach allows us to investigate the content contained in the remaining entangled dimensions of the dVAE. We found that these entangled dimensions are represented in face-selective regions, and showed for the first time that they contain identity-relevant information, providing new insight into their computational role.

Prior work has found that DCNNs trained for facial identity discrimination only capture a small amount of variance in human face selective regions [33] and do not replicate activity in the primate face patch hierarchy or human behavioral responses [34]. We see an advantage of our generatively trained encoding models versus the discriminatively trained DCNN particularly in the FFA, although this is not significant in all individual subjects (Fig 3). Interestingly, recent work [33,35] has shown that object-trained networks do a better job of matching human neural responses to faces than face-trained discriminative networks like VGG-Face tested here, though in general face-selective responses are not as well explained by DCNNs as object- and scene-selective regions. It is possible that this is due to richer training datasets available for objects than faces [36], which may lead to higher latent dimensionality [37] and improve model match to visual cortex [38]. To date no disentangled models have been successfully trained on such large and diverse datasets. For the fairest comparison, we focused on models trained only on face images. As disentangled models improve at capturing image variation at larger scales, future work can compare disentangled and non-disentangled models on richer, more varied datasets.

The original paper presenting this fMRI dataset also found good decoding performance across the brain with a generative VAE, achieving much higher decoding performance than the results presented here [11], as have other studies comparing generatively trained neural networks to visual brain responses [3941]. The original study focused on maximizing fMRI reconstruction and decoding with a high dimensional network (1024 dimensions vs. our 24). We chose our model to have the highest disentanglement which yielded the lowest dimensional network from all those tested in our hyperparameter search (see Methods M1), likely because enforcing statistical independence between latent dimensions via regularization during training becomes less effective as dimensionality increases. Thus our 24-dimensional network is less expressive than higher dimensional networks because it has a much smaller bottleneck for modeling the data distribution. However, the added interpretability afforded by disentanglement allows more fine-grained interpretability of the fMRI data not possible with standard models. Future work should investigate how to combine the interpretability benefits of disentangled models with the expressiveness of high dimensional networks.

Another recent approach has sought to learn disentangled latent representations in a supervised manner [34]. They learn a model which inverts a 3D face graphics program by supervising intermediate representations to match the primitives defined in the program (e.g. 3D shape, texture, and lighting). They find that this network matches primate face representations better than identity trained networks. Importantly, these intermediate representations are prespecified and need to be learned from labeled synthetic data. Many of these prespecified dimensions match those learned by our dVAE, providing further support for disentangled learning as a method to learn relevant latent dimensions in an unsupervised manner. On the other hand, the learned dimensions do not seem to represent fine-grained facial features, like the relative position and size of the eyes or nose, which prior work has found to be represented in the primate face patch system [42]. While it is difficult to say whether this is due to the specific model or dataset we used, it adds to mounting work [43,44] suggesting a more holistic face-based feature coding.

One recent prior study has investigated the correspondence between dVAEs and single neurons in macaque IT [43]. They find several IT neurons that show high one-to-one match with single units in their dVAE. They also demonstrate a high degree of disentanglement in the macaque neurons by showing a strong correlation between model disentanglement and alignment with IT neurons. It is worth noting that only a handful of neurons in the macaque data show high alignment with single disentangled dimensions. Perhaps unsurprisingly given the lower spatial resolution of fMRI, we do not see the same high disentanglement in our data as evidenced by the fact that each region is well predicted by multiple latent dimensions. Even at the voxel level, it may not be possible to see evidence for single disentangled dimensions. It remains an open question whether, and at what spatial scale, the primate face network is disentangled or shows exact correspondence to the dimensions learned by dVAEs across the visual system.

The content of the disentangled dimensions learned by our dVAE, and all other disentangled models, reflects the distribution of features in its training set. CelebA is a dataset of celebrity images which does not reflect the underlying distribution of faces that people see in daily life. In particular, CelebA faces tend to be young adults, white, and smiling at a camera. One example of how this can affect learned representations can be seen in the smile dimension, which is sometimes entangled with wearing sunglasses (S1 Video), likely reflecting a bias in CelebA that people wearing sunglasses tend to be smiling. More critically, the visual as well as racial and ethnic biases in the dataset likely impact the quality of the learned dimensions [45] and the model’s generalization to other datasets. Training models on a more ecologically valid dataset may improve encoding performance by better reflecting the statistics of real-world visual experience.

This work has important applications for cognitive neuroscientists to understand the relationship between semantic factors and neural activity using natural datasets without labels, in a scalable manner. Successfully scaling these models to larger, less constrained datasets is an important, ongoing research challenge. While our current results show some inter-subject variability, as the quality of models and fMRI data increase, our method can be used to identify new semantically meaningful data dimensions with higher precision. Disentangled models have been created for various visual domains including object and scene processing [4648] and can in theory be applied to any large scale visual dataset. While disentangled models are an active area of research in AI, there has been little investigation of their cognitive and neural plausibility. Our work sheds light on the role of entangled and disentangled dimensions in face representations in the brain and provides avenues for follow-up questions pertaining to their role in identity decoding. Understanding the neural coding of disentangled dimensions in the brain can help inspire novel data representations in AI systems.

Methods

M1. Neural Net architecture and training

We trained our VAE models using the TensorFlow DisentanglementLib package [49]. To identify the best disentangled model for our fMRI analyses, we performed a hyperparameter search over model architectures (including beta-VAE [17] and FactorVAE [16]), number of latent dimensions (24, 32, 48, and 64), and architecture-specific disentanglement parameters (beta-VAE β ϵ [1,2,4,6,8,16], FactorVAE γ ϵ [0, 10, 20, 30, 40, 50, 100] where γ = 0 is the same model as a beta-VAE with β = 1, a standard VAE). These hyperparameters were selected based on prior work [49]. For every hyperparameter combination, we performed 10 random initializations. This resulted in 240 FactorVAE models (we used the β = 1 beta-VAE as our γ = 0 FactorVAE instead of training new models) and 240 beta-VAE models. We used beta-VAE without disentanglement (beta = 1) for the standard, non-disentangled VAE models. After training, models were evaluated using the unsupervised disentanglement metric (UDR) [50]. We selected the model with the highest disentanglement score, a FactorVAE model with 24 latent dimensions and γ = 10, for subsequent encoding analyses. Of the dimension-matched standard VAE models, we selected the randomly initialized model with the highest disentanglement score as our baseline.

For our baseline discriminative model, we used VGG-Faces [22], a network that uses the VGG architecture [21] and is trained from scratch on 2.6 million face images to predict face identity. To facilitate model comparison, we take the representations at the final fully connected layer and use Principal Component Analysis to reduce the dimensionality to match that of the VAEs.

M2. Dimension annotation

After training and selecting our disentangled model, we passed 15 face images, not included in training, to the model. For each face image, we generated a set of “traversal images” by changing the value of a single latent dimension (e.g., Fig 2) from -2 to +2. The traversal images for each latent dimension were combined into an animated gif. The two authors labeled each dimension in each gif. We first consolidated the annotations for each annotator across images for each dimension by tallying the labels across the 15 face images (see S2 Table). We then selected labels where both annotators agreed on the majority of images for our final labels (Table 1). The annotators agreed on 14 out of the 16 labeled dimensions. For the two dimensions that the annotators did not agree on, one annotator assigned a majority label and the other did not. In these two cases (dimension 7 head elevation and dimension 11 hair), we assigned the majority label from one annotator. The remaining 8 dimensions were either not labeled or were not labeled consistently.

M3. fMRI data and preprocessing

We used publicly available fMRI data of four subjects from [11]. Subjects viewed around 8000 “training” face images each presented once, and 20 “test” face images presented between 40–60 times each. Face images were selected at random from the CelebA dataset and passed through a VAE-GAN. Each face was on the screen for 1s followed by a 2s ISI. The experiment was split over eight scan sessions. Subjects were also scanned on 8–10 separate face-object localizer runs to identify face-selective voxels. Data were collected on a Philips 3T ACHEIVA scanner. Subjects provided informed consent and all experiments were conducted in accordance with Comité de Protection des Personnes standards. For more details, refer to the original paper.

Data were pre-processed and projected onto subjects’ individual cortical surfaces using Freesurfer [51]. Preprocessing consisted of motion correcting each functional run, aligning it to each subject’s anatomical volume and then resampling to each subject’s high-density surface. After alignment, data were smoothed using a 5 mm FWHM Gaussian kernel. All individual analyses were performed on each subject’s native surface.

M4. ROI definition

Regions of interest were defined using a group-constrained subject-specific approach [52]. The regions we investigated were the right Fusiform Face Area (FFA), Occipital Face Area (OFA), and Superior Temporal Sulcus (STS). To define our regions of interest (ROIs), we used the published group parcels from [52].

We selected the top 10% of voxels in each parcel using a metric that combined both face-selectivity and reliability on the test data. We first calculated face-selectivity based on face-object localizer runs, and z-scored each subjects’ face>object p-values within each parcel to yield a selectivity score vs for each voxel. We next calculated the split-half reliability in our test data (Spearman r), and z-scored these values within each parcel to generate a reliability score vr. We then summed the normalized selectivity and reliability scores to yield our final selection metric (v = vs + vr). We restricted our ROI analyses to the right hemisphere because of more selective face responses and increased reliability in our test data. Across the subjects, the FFA had roughly 170 voxels, the OFA 110 voxels, and the STS 170 voxels. For our whole brain analyses, we computed the above metric (v = vs + vr) for each cortical voxel. We then selected all voxels that scored more than 1.5 standard deviations above the mean.

M5. Encoding model procedure

We estimated a linear map between the latent dimensions in our models and the fMRI data via a generalized linear model (GLM), following the procedure in the original study [11]. Since each training face image was shown only once, the latent values for that image (rather than the image itself) were included as weighted regressors to increase reliability of the learned beta weights. The latent values for each training face image, the test faces, and a general face “bias” term were all included as regressors, as well as well as nuisance regressors for linear drift removal and motion correction (x, y, z) per run.

To test the accuracy of the encoding model, we extracted the latent dimensions for each test image and multiplied this by the beta weights learned in the GLM and adding the above “bias” term to get a predicted voxel response to each test image (Fig 1). This produced an estimated brain response for each test image. We then compared the predicted fMRI response in each voxel activity to the true voxel activity across all test images using Spearman correlation.

M6. Preference mapping

To understand the contribution of each latent dimension to brain responses, we followed the same encoding model training procedure described above. In model testing, we then generated the voxel prediction using a single latent dimension value instead of all the latent dimension values and calculated the correlation between the single dimension’s predictions and ground truth. We calculated the average prediction for each latent variable within each ROI (Fig 5). For whole brain analyses, we performed preference mapping [25], assigning each voxel’s preference label as the dimension which yielded the highest prediction.

M7. Identity decoding

To understand the identity-relevant information in different latent dimensions we performed identity decoding of our test images. To decode identity, we took the same learned betas (W) from the encoding training procedure and multiplied the test fMRI data (y) by its pseudo-inverse (x^=W1*(yb)) where b is the face bias term. This generated a predicted set of latent dimensions for each test image. We correlated the predicted latent dimensions with the true test latent dimensions and one random foil to assess the pairwise accuracy of the decoding. If the correlation between the predicted latent dimensions and the true latent dimensions was larger than the correlation between the predicted latent dimensions and the foil, that indicated that the identity was correctly decoded. This pairwise comparison was repeated with every test image as the ground truth, and every other test image as the foil for a total of 20*19 = 380 samples per subject. The total accuracy was the number of correct identity decoding divided by 380. We repeated this for different subsets of latent dimensions: all those labeled as identity-relevant (including hair as it offers important cues to facial identity and prior work has shown sensitivity in face-selective voxels [42]), identity-irrelevant, and entangled dimensions.

M8. Statistical testing

As the underlying distribution of our data was unknown, we used non-parametric, resampling-based statistics. To evaluate whether each model achieved above chance performance, we generated null hypotheses by repeating the above prediction correlations with shuffled test image labels 1000 resample runs. We performed shuffling within subject, and then computed p-values for each individual as well as group-average prediction.

To compare models, we take the difference in prediction and compare this to a null distribution with shuffled model labels. We generated 1000 resample runs and calculate p-values for each two-tailed pairwise model comparison at the individual and group levels (S1 Table).

We followed the above procedure to assess the significance of our preference mapping results to evaluate whether each feature was significantly predictive in each ROI. To compare predictivity of OFA and FFA, we compared their difference in prediction to a shuffled baseline.

M9. Analysis code

The code for the analysis is available at https://github.com/psoulos/disentangle-faces.

Supporting information

S1 Table. Group and individual subject model versus model significance results.

P-value are in parentheses.

(XLSX)

pcbi.1011887.s001.xlsx (9.5KB, xlsx)
S2 Table. Tallies for labels assigned to each of fifteen individual images in the labeling experiments by two raters, and each raters’ consensus label.

(CSV)

S1 Video. Animated latent traversals for all 24 latent dVAE dimensions for one example rendered face.

Dimensions are varied from -2 to +2, with all other dimensions held constant. Images are model generated. Base image is from the CelebA dataset [20].

(GIF)

pcbi.1011887.s003.gif (1.6MB, gif)
S2 Video. Animated latent traversals for all 24 latent dVAE dimensions for a second example rendered face.

Dimensions are varied from -2 to +2, with all other dimensions held constant. Images are model generated. Base image is from the CelebA dataset [20].

(GIF)

pcbi.1011887.s004.gif (1.6MB, gif)
S3 Video. Animated latent traversals for all 24 latent VAE dimensions for the same exampled rendered face in S1 Video.

Dimensions are varied from -2 to + 2 with all other dimensions held constant. Images are model generated. Base image is from the CelebA dataset [20].

(GIF)

pcbi.1011887.s005.gif (2.2MB, gif)
S4 Video. Animated latent traversals for all 24 latent VAE dimensions for the same exampled rendered face in S2 Video.

Dimensions are varied from -2 to + 2 with all other dimensions held constant. Images are model generated. Base image is from the CelebA dataset [20].

(GIF)

pcbi.1011887.s006.gif (2.2MB, gif)
S1 Fig. Whole brain encoding.

Encoding model performance in all face-selective voxels. Ventral view of the remaining two subjects for dVAE (left), VAE (center) and VGG-face (right).

(EPS)

pcbi.1011887.s007.eps (4.1MB, eps)
S2 Fig. Whole brain correlations for all 4 lateral view.

(EPS)

pcbi.1011887.s008.eps (8.8MB, eps)
S3 Fig. Preference mapping for each dimension in the STS.

Average feature predictivity across subjects is shown for each individual dimension. Dots represent individual subjects. Asterisks represents significance (p < .05). Dimensions are colored according to Table 1.

(EPS)

pcbi.1011887.s009.eps (2.2MB, eps)
S4 Fig. Whole brain preference maps.

Each voxel is shaded based on the latent dimension that provides the highest predictivity.

(EPS)

pcbi.1011887.s010.eps (6.4MB, eps)

Acknowledgments

We thank Michael Bonner for helpful discussions on this work, Emalie McMahon and Raj Magesh for feedback on the manuscript.

Data Availability

All data used were publicly available and linked to the original paper: https://openneuro.org/datasets/ds001761. All analysis code is publicly available on our github repo https://github.com/psoulos/disentangle-faces.

Funding Statement

This work was supported with funds from The Clare Boothe Luce Program for Women in STEM (LI). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1.Tsao D. Y., Moeller S., and Freiwald W. A., “Comparing face patch systems in macaques and humans,” Proc. Natl. Acad. Sci. U. S. A., vol. 105, no. 49, pp. 19514–19519, Dec. 2008. doi: 10.1073/pnas.0809662105 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Duchaine B. and Yovel G., “A Revised Neural Framework for Face Processing,” Annu. Rev. Vis. Sci., vol. 1, no. 1, pp. 393–416, Nov. 2015. doi: 10.1146/annurev-vision-082114-035518 [DOI] [PubMed] [Google Scholar]
  • 3.Freiwald W., Duchaine B., and Yovel G., “Face Processing Systems: From Neurons to Real-World Social Perception,” vol. 39, pp. 325–346, Jul. 2016. doi: 10.1146/annurev-neuro-070815-013934 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Pitcher D., Dilks D. D., Saxe R. R., Triantafyllou C., and Kanwisher N., “Differential selectivity for dynamic versus static information in face-selective cortical regions,” Neuroimage, vol. 56, no. 4, pp. 2356–2363, Jun. 2011. doi: 10.1016/j.neuroimage.2011.03.067 [DOI] [PubMed] [Google Scholar]
  • 5.Polosecki P., Moeller S., Schweers N., Romanski L. M., Tsao D. Y., and Freiwald W. A., “Faces in Motion: Selectivity of Macaque and Human Face Processing Areas for Dynamic Stimuli,” J. Neurosci., vol. 33, no. 29, pp. 11768–11773, Jul. 2013. doi: 10.1523/JNEUROSCI.5402-11.2013 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Freiwald W. A. and Tsao D. Y., “Functional compartmentalization and viewpoint generalization within the macaque face-processing system.,” Science, vol. 330, no. 6005, pp. 845–51, Nov. 2010. doi: 10.1126/science.1194908 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Kriegeskorte N., Formisano E., Sorger B., and Goebel R., “Individual faces elicit distinct response patterns in human anterior temporal cortex,” Proc. Natl. Acad. Sci. U. S. A., vol. 104, no. 51, pp. 20600–20605, Dec. 2007. doi: 10.1073/pnas.0705654104 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.O’Toole A. J., Castillo C. D., Parde C. J., Hill M. Q., and Chellappa R., “Face Space Representations in Deep Convolutional Neural Networks,” Trends Cogn. Sci., vol. 22, no. 9, pp. 794–809, Sep. 2018. doi: 10.1016/j.tics.2018.06.006 [DOI] [PubMed] [Google Scholar]
  • 9.O’Toole A. J. and Castillo C. D., “Face Recognition by Humans and Machines: Three Fundamental Advances from Deep Learning,”, vol. 7, pp. 543–570, Sep. 2021. doi: 10.1146/annurev-vision-093019-111701 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Jiahui G. et al. , “Modeling naturalistic face processing in humans with deep convolutional neural networks,” Proc. Natl. Acad. Sci., vol. 120, no. 43, p. e2304085120, Oct. 2023. doi: 10.1073/pnas.2304085120 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.VanRullen R. and Reddy L., “Reconstructing faces from fMRI patterns using deep generative neural networks,” Commun. Biol., vol. 2, no. 1, p. 193, 2019. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Kulkarni T. D., Whitney W. F., Kohli P., and Tenenbaum J., “Deep Convolutional Inverse Graphics Network,” in Advances in {Neural} {Information} {Processing} {Systems}, 2015, vol. 28. [Google Scholar]
  • 13.Chen X., Duan Y., Houthooft R., Schulman J., Sutskever I., and Abbeel P., “Infogan: Interpretable representation learning by information maximizing generative adversarial nets,” Adv. Neural Inf. Process. Syst., vol. 29, 2016. [Google Scholar]
  • 14.Carlson T., Tovar D. A., Alink A., and Kriegeskorte N., “Representational dynamics of object vision: the first 1000 ms.,” J. Vis., vol. 13, no. 10, pp. 1-, Jan. 2013. doi: 10.1167/13.10.1 [DOI] [PubMed] [Google Scholar]
  • 15.Chen R. T. Q., Li X., Grosse R. B., and Duvenaud D. K., “Isolating sources of disentanglement in variational autoencoders,” Adv. Neural Inf. Process. Syst., vol. 31, 2018. [Google Scholar]
  • 16.Kim H. and Mnih A., “Disentangling by Factorising,” 2018. [Google Scholar]
  • 17.Higgins I. et al. , “beta-{VAE}: Learning Basic Visual Concepts with a Constrained Variational Framework,” in International Conference on Learning Representations, 2017. [Google Scholar]
  • 18.Kingma D. P. and Welling M., “Auto-Encoding Variational Bayes,” no. arXiv:1312.6114. May-2014. [Google Scholar]
  • 19.Bengio Y., Courville A., and Vincent P., “Representation learning: A review and new perspectives,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 35, no. 8, pp. 1798–1828, 2013. doi: 10.1109/TPAMI.2013.50 [DOI] [PubMed] [Google Scholar]
  • 20.Liu Z., Luo P., Wang X., and Tang X., “Deep Learning Face Attributes in the Wild.” pp. 3730–3738, 2015. [Google Scholar]
  • 21.Simonyan K. and Zisserman A., “Very Deep Convolutional Networks for Large-Scale Image Recognition,” Sep. 2014. [Google Scholar]
  • 22.Parkhi O., Vedaldi A., and Zisserman A., “Deep face recognition,” in BMVC 2015—Proceedings of the British Machine Vision Conference 2015, 2015, pp. 1–12. [Google Scholar]
  • 23.Naselaris T., Allen E., and Kay K., “Extensive sampling for complete models of individual brains,” Curr. Opin. Behav. Sci., vol. 40, pp. 45–51, Aug. 2021. [Google Scholar]
  • 24.Fedorenko E., Hsieh P.-J., Nieto-Castañón A., Whitfield-Gabrieli S., and Kanwisher N., “New method for fMRI investigations of language: defining ROIs functionally in individual subjects.,” J. Neurophysiol., vol. 104, no. 2, pp. 1177–94, Aug. 2010. doi: 10.1152/jn.00032.2010 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Lee Masson H. and Isik L., “Functional selectivity for social interaction perception in the human superior temporal sulcus during natural viewing,” Neuroimage, vol. 245, p. 118741, Dec. 2021. doi: 10.1016/j.neuroimage.2021.118741 [DOI] [PubMed] [Google Scholar]
  • 26.Haxby J. V., Hoffman E. A., and Gobbini M. I., “The distributed human neural system for face perception,” Trends Cogn. Sci., vol. 4, no. 6, pp. 223–233, Jun. 2000. doi: 10.1016/s1364-6613(00)01482-0 [DOI] [PubMed] [Google Scholar]
  • 27.Rossion B., Caldara R., Seghier M., Schuller A. M., Lazeyras F., and Mayer E., “A network of occipito-temporal face-sensitive areas besides the right middle fusiform gyrus is necessary for normal face processing,” Brain, vol. 126, no. 11, pp. 2381–2395, Nov. 2003. doi: 10.1093/brain/awg241 [DOI] [PubMed] [Google Scholar]
  • 28.Tsantani M., Kriegeskorte N., Storrs K., Williams A. L., McGettigan C., and Garrido L., “FFA and OFA Encode Distinct Types of Face Identity Information,” J. Neurosci., vol. 41, no. 9, pp. 1952–1969, Mar. 2021. doi: 10.1523/JNEUROSCI.1449-20.2020 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.Contreras J. M., Banaji M. R., and Mitchell J. P., “Multivoxel Patterns in Fusiform Face Area Differentiate Faces by Sex and Race,” PLoS One, vol. 8, no. 7, p. e69684, Jul. 2013. doi: 10.1371/journal.pone.0069684 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.Zhang H., Japee S., Nolan R., Chu C., Liu N., and Ungerleider L. G., “Face-selective regions differ in their ability to classify facial expressions,” Neuroimage, vol. 130, pp. 77–90, Apr. 2016. doi: 10.1016/j.neuroimage.2016.01.045 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31.Harris R. J., Young A. W., and Andrews T. J., “Brain regions involved in processing facial identity and expression are differentially selective for surface and edge information,” Neuroimage, vol. 97, pp. 217–223, Aug. 2014. doi: 10.1016/j.neuroimage.2014.04.032 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Zhou S., Zelikman E., Lu F., Ng A. Y., Carlsson G., and Ermon S., “Evaluating the Disentanglement of Deep Generative Models through Manifold Topology,” ICLR 2021 - 9th Int. Conf. Learn. Represent., Jun. 2020. [Google Scholar]
  • 33.Jiahui G. et al. , “Modeling naturalistic face processing in humans with deep convolutional neural networks,” bioRxiv, p. 2021.11.17.469009, Apr. 2022. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Yildirim I., Belledonne M., Freiwald W., and Tenenbaum J., “Efficient inverse graphics in biological face processing,” Sci. Adv., vol. 6, no. 10, p. eaax5979, Mar. 2020. doi: 10.1126/sciadv.aax5979 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35.Vinken K., Prince J. S., Konkle T., and Livingstone M. S., “The neural code for ‘face cells’ is not face-specific,” Sci. Adv., vol. 9, no. 35, p. eadg1736, Sep. 2023. doi: 10.1126/sciadv.adg1736 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36.Socher R., “ImageNet: A large-scale hierarchical image database,” in 2009 IEEE Conference on Computer Vision and Pattern Recognition, 2009, pp. 248–255. [Google Scholar]
  • 37.Elmoznino E. and Bonner M. F., “High-performing neural network models of visual cortex benefit from high latent dimensionality,” bioRxiv, p. 2022.07.13.499969, Jun. 2023. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 38.Conwell C., Prince J. S., Kay K. N., Alvarez G. A., and Konkle T., “What can 1.8 billion regressions tell us about the pressures shaping high-level visual representation in brains and machines?,” bioRxiv, p. 2022.03.28.485868, Jul. 2023. [Google Scholar]
  • 39.Han K. et al. , “Variational autoencoder: An unsupervised model for encoding and decoding fMRI activity in visual cortex,” Neuroimage, vol. 198, pp. 125–136, Sep. 2019. doi: 10.1016/j.neuroimage.2019.05.039 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40.Seeliger K., Güçlü U., Ambrogioni L., Güçlütürk Y., and van Gerven M. A. J., “Generative adversarial networks for reconstructing natural images from brain activity,” Neuroimage, vol. 181, pp. 775–785, Nov. 2018. doi: 10.1016/j.neuroimage.2018.07.043 [DOI] [PubMed] [Google Scholar]
  • 41.Güçlütürk Y., Güçlü U., Seeliger K., Bosch S., van Lier R., and van Gerven M. A., “Reconstructing perceived faces from brain activations with deep adversarial neural decoding,” Adv. Neural Inf. Process. Syst., vol. 30, 2017. [Google Scholar]
  • 42.Freiwald W. A., Tsao D. Y., and Livingstone M. S., “A face feature space in the macaque temporal lobe,” Nat. Neurosci., vol. 12, no. 9, pp. 1187–1196, Sep. 2009. doi: 10.1038/nn.2363 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43.Higgins I. et al. , “Unsupervised deep learning identifies semantic disentanglement in single inferotemporal face patch neurons,” Nat. Commun. 2021 121, vol. 12, no. 1, pp. 1–14, Nov. 2021. doi: 10.1038/s41467-021-26751-5 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 44.Chang L. and Tsao D. Y., “The Code for Facial Identity in the Primate Brain,” Cell, vol. 169, no. 6, pp. 1013–1028.e14, Jun. 2017. doi: 10.1016/j.cell.2017.05.011 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 45.Birhane A. and Guest O., “Towards decolonising computational sciences.” [Google Scholar]
  • 46.Lin C., Yuan Z., Zhao S., Sun P., Wang C., and Cai J., “Domain-Invariant Disentangled Network for Generalizable Object Detection.” pp. 8771–8780, 2021. [Google Scholar]
  • 47.Epstein D., Park T., Zhang R., Shechtman E., and Efros A. A., “BlobGAN: Spatially Disentangled Scene Representations,” pp. 616–635, 2022. [Google Scholar]
  • 48.Denton E. L. and vighnesh Birodkar, “Unsupervised Learning of Disentangled Representations from Video,” Adv. Neural Inf. Process. Syst., vol. 30, 2017. [Google Scholar]
  • 49.Locatello F. et al. , “Challenging common assumptions in the unsupervised learning of disentangled representations,” in international conference on machine learning, 2019, pp. 4114–4124. [Google Scholar]
  • 50.Duan S. et al. , “Unsupervised Model Selection for Variational Disentangled Representation Learning,” 2019. [Google Scholar]
  • 51.Fischl B., Sereno M. I., and Dale A. M., “Cortical Surface-Based Analysis,” Neuroimage, vol. 9, no. 2, pp. 195–207, Feb. 1999. [DOI] [PubMed] [Google Scholar]
  • 52.Julian J. B., Fedorenko E., Webster J., and Kanwisher N., “An algorithmic method for functionally defining regions of interest in the ventral visual pathway.,” Neuroimage, vol. 60, no. 4, pp. 2357–64, May 2012. doi: 10.1016/j.neuroimage.2012.02.055 [DOI] [PubMed] [Google Scholar]
PLoS Comput Biol. doi: 10.1371/journal.pcbi.1011887.r001

Decision Letter 0

Lyle J Graham, Jean Daunizeau

18 Jun 2023

Dear Dr. Isik,

Thank you very much for submitting your manuscript "Disentangled deep generative models reveal coding principles of the human face processing network" for consideration at PLOS Computational Biology.

Your manuscript was reviewed by members of the editorial board and by two independent reviewers. In light of the reviews (below this email), we would like to invite the resubmission of a significantly-revised version that takes into account the reviewers' comments.

We cannot make any decision about publication until we have seen the revised manuscript and your response to the reviewers' comments. Your revised manuscript is also likely to be sent to reviewers for further evaluation.

When you are ready to resubmit, please upload the following:

[1] A letter containing a detailed list of your responses to the review comments and a description of the changes you have made in the manuscript. Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out.

[2] Two versions of the revised manuscript: one with either highlights or tracked changes denoting where the text has been changed; the other a clean version (uploaded as the manuscript file).

Important additional instructions are given below your reviewer comments.

Please prepare and submit your revised manuscript within 60 days. If you anticipate any delay, please let us know the expected resubmission date by replying to this email. Please note that revised manuscripts received after the 60-day due date may require evaluation and peer review similar to newly submitted manuscripts.

Thank you again for your submission. We hope that our editorial process has been constructive so far, and we welcome your feedback at any time. Please don't hesitate to contact us if you have any questions or comments.

Sincerely,

Jean Daunizeau

Academic Editor

PLOS Computational Biology

Lyle Graham

Section Editor

PLOS Computational Biology

***********************

Reviewer's Responses to Questions

Comments to the Authors:

Please note here if the review is uploaded as an attachment.

Reviewer #1: The authors explored disentangled generative neural networks as a model for human face processing. Identifying a Factor VAE through model selection, the authors employed human raters to interpret the 24 model feature dimensions, finding 16 semantically meaningful. The authors then compared the disentangled model with two conventional models—a VAE without disentanglement and a VGG-based face classification network, showing the disentangled VAE (dVAE) to be a competitive encoding model of human fMRI activity. Furthermore, higher-level dVAE features better matched responses in more anterior face-selective regions. The dVAE also enabled identity decoding from fMRI activity, almost entirely through the eight identity-relevant disentangled features.

Developing high-performance yet interpretable models is an important research goal in visual neuroscience. Disentangled models are a promising direction showing emerging evidence to match neural representations uniquely well (e.g., Higgins et al., Nat. Commun. 2021; Whittington et al., ICLR 2023, arXiv:2210.01768). However, several concerns limit the potential impact of the manuscript.

A key promise of disentangled models is more interpretable features. The manuscript provides limited objective evidence for this supposition. Figure 1 shows examples of feature dimension traversals, which the reader must evaluate subjectively. Reasonable viewers may disagree as to how cleanly disentangled the features are. For instance, it looks to me that the “smile” dimension spuriously correlates with jaw width. Table 1 provides a consensus label for each feature; it would be more informative to report the original annotations given to individual images per rater. The raw data, perhaps quantified with a word embedding metric and evaluated for inter-rater consistency, would provide a more granular and rigorous measure of how interpretable the disentangled features are. The interpretability of the dVAE should ideally be compared to non-disentangled models (i.e., the VAE with beta = 1 and VGG-Face PCs) to establish the advantage of using a dVAE. The authors should also detail how they recruited the human raters. While it is not necessarily a problem for the raters to include the author(s), this should be reported.

The authors conclude that the disentangled features constitute a decent encoding model of human fMRI data. However, the encoding performance analysis (Fig. 3) likely unfairly represents conventional classification networks. The final fully connected layer of VGG-Face is considerably worse than its intermediate layers (Jiahui et al., biorxiv 2021). VGG-Face is a worse encoding model than ImageNet-trained VGG, which is, in turn, worse than AlexNet and CORnet (Chang et al., Curr. Biol. 2021), both widely used models. It thus remains unclear whether, or by how much, disentangled models suffer an encoding performance cost relative to conventional, less interpretable models for an (as-yet unquantified) gain in interpretability.

A third major conclusion of the manuscript is a posterior-anterior gradient of low- to high-level feature representations. This claim should be supported by a quantitative summary of the evidence (Fig. 4). While the overall conclusion likely holds, some presumably low-level features (e.g., “background”) are predictive in FFA, and other high-level features (e.g., “smile”) are also predictive in OFA. There is also considerable inter-subject variation. In quantifying the low-to-high level gradient, the authors would also help the reader by unambiguously defining what low- and high-level features signify. The manuscript seems to suggest high-level features as equivalent to identity-specific, “less changeable” features. However, this interpretation is confusing given that the authors claim no spatial segregation between identity-relevant and -irrelevant features. For another example of the present ambiguities, skin tone may be considered either a low-level visual attribute (because it affects luminance semi-globally) or a high-level one (in being identity-relevant).

Finally, although outside the scope of the current analyses, the authors may consider exploring the “alignment” between disentangled feature dimensions and fMRI responses (Higgins et al., 2021). Aside from interpretability, neuron-to-neuron alignment is perhaps the other main promise of disentangled models. This alignment is the main advantage over conventional models found by Higgins et al. (who did not quantify interpretability). While a negative result would be difficult to interpret given the limited spatial resolution of fMRI, a positive finding would add considerably to the study's impact.

Minor concerns:

I am unsure about the significance of analyzing the information in the “entangled” dimensions. A conventional model has only entangled features, which necessarily contain all the identity information available to the model. The entangled features of a dVAE might simply correspond to the entangled dimensions in conventional models. It seems more remarkable (and thus worth more emphasis) that the disentangled, identity-specific features in a dVAE (i.e., a third of all its dimensions) almost completely explain its ability to decode identity from fMRI.

The authors laudably performed an extensive model search. However, the optimal model corresponded to extreme values in the searched parameter ranges—the lowest gamma and fewest feature dimensions. Should this prompt the authors to consider model instances with even lower gamma and fewer dimensions? Moreover, although beta-VAE underperforms Factor VAE in the UDR selection metric, it would still be useful to analyze beta-VAE to facilitate comparison with prior studies.

In Fig. 3, the asterisks look very similar to the gray scatter dots.

On lines 4 and 11, pg. 6, the author used “significantly.” It is unclear whether this refers to statistical significance and, if so, what test was concerned.

Line 42, pg. 6: gender appearance should correspond to dark olive color, not green.

In the Introduction, the authors state that "where […] face identity is represented is still largely unknown" (line 36). The Discussion suggests that identity and expression coding are still understood to be clear-cut (lines 38-40). The authors should moderate both statements in light of the Duchaine & Yovel review the authors also cite. The review presents preponderant evidence that FFA contains both identity and expression information.

The authors used Z-scored p-values to define face-selective ROIs, but p-values do not measure effect size. The raw responses are more appropriate for defining a selectivity index, perhaps with a p-value mask.

Citation [14] (Higgins et al., “beta-VAE […]”) should be dated 2017, not 2022.

Reviewer #2: In the manuscript “Disentangled deep generative models reveal coding principles of the human face processing network”, Soulos and colleagues trained a disentangled deep generative model (dVAE) and used this model as an encoding model for human fMRI data to understand the neural representations of the semantic features of human faces. Using semantically meaningful dimensions from the deep learning models to address the problem of human face representations in the brain is an attractive approach that could integrate AI and neuroscience research. The manuscript is concise and well-written. However, I have reservations about the clarity and robustness of the current results.

Major concerns:

1. The authors trained their model solely on a set of images with limited diversity in face features (CelebA). There's a possibility that the model is biased towards certain dimensions or common image statistics specific to this image set. As numerous face image datasets are readily accessible now, I encourage the authors to broaden their training image set to retrain the model. More importantly, to validate their findings (especially with the neural data), they need to use another deep generative model trained with a different image set.

2. Many current disentangled dimensions are not face-relevant, including identity-relevant features such as hairstyles. Only a few dimensions (e.g., gender appearance, face width) are purely face dimensions. These disentangled dimensions also fail to separate fine features such as the location, distance, or size of the eyes and nose, which are crucial for face recognition and are represented hierarchically in the brain. Thus, the current dVAE provides a simple and coarse model of portrait images that is heavily influenced by prominent face-irrelevant features.

3. The authors assert that "Higher-level, identity-relevant dimensions are represented in more anterior face-selective regions." However, this is not clear to me. The dimensions are mixed, and it's challenging to determine which dimensions are “higher-level” ones. Additionally, this study seems to focus more on the three posterior regions (OFA, FFA, and pSTS), so I'm unclear which regions are considered “anterior” and how these anterior regions were defined. This part of the results needs a clear definition and rigorous analysis to compare the representations between “higher vs lower” level dimensions and “anterior vs posterior” face-selective regions to substantiate the conclusion.

4. This study primarily relies on neural data from four participants, and the inconsistent prediction performance among these participants raises concerns. There are significant inter-subject variations in the results, and with only four data points, it's challenging to gauge the robustness of the results and their generalizability across individuals. I have significant reservations about the small number of participants and strongly recommend the authors to include another dataset or more participants to demonstrate the reliability and generalizability of their findings.

5. I am concerned about the capability of human fMRI data to work with deep neural network models, given that fMRI data lacks the spatial resolution necessary to accurately map the features. The correlations shown in this study are also not particularly high, especially considering the strong effect from non-identity low-level features such as the background. The authors have briefly touched upon this in the discussion, but expanding on this point could help readers understand the limitations of this study.

Minor concerns:

1. The method of calculating accuracies in Figure 6 is unclear. From the methods section, it appears that the decoding was based on the correlations between the predicted and actual latent dimensions . However, it is not clear how the authors converted these correlation values into accuracy values. This process requires further explanation.

2. There are no error bars in any of the plots, possibly due to the large variability across individuals when the study only includes four participants. If more participants can be included, it is recommended that error bars are added to aid understanding of the bar plots.

**********

Have the authors made all data and (if applicable) computational code underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data and code underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data and code should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data or code —e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

Figure Files:

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email us at figures@plos.org.

Data Requirements:

Please note that, as a condition of publication, PLOS' data policy requires that you make available all data used to draw the conclusions outlined in your manuscript. Data must be deposited in an appropriate repository, included within the body of the manuscript, or uploaded as supporting information. This includes all numerical values that were used to generate graphs, histograms etc.. For an example in PLOS Biology see here: http://www.plosbiology.org/article/info%3Adoi%2F10.1371%2Fjournal.pbio.1001908#s5.

Reproducibility:

To enhance the reproducibility of your results, we recommend that you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. Additionally, PLOS ONE offers an option to publish peer-reviewed clinical study protocols. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols

PLoS Comput Biol. doi: 10.1371/journal.pcbi.1011887.r003

Decision Letter 1

Lyle J Graham, Jean Daunizeau

11 Oct 2023

Dear Dr. Isik,

Thank you very much for submitting your manuscript "Disentangled deep generative models reveal coding principles of the human face processing network" for consideration at PLOS Computational Biology. As with all papers reviewed by the journal, your manuscript was reviewed by members of the editorial board and by several independent reviewers. The reviewers appreciated the attention to an important topic. Based on the reviews, we are likely to accept this manuscript for publication, providing that you modify the manuscript according to the review recommendations.

Please prepare and submit your revised manuscript within 30 days. If you anticipate any delay, please let us know the expected resubmission date by replying to this email.

When you are ready to resubmit, please upload the following:

[1] A letter containing a detailed list of your responses to all review comments, and a description of the changes you have made in the manuscript. Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out

[2] Two versions of the revised manuscript: one with either highlights or tracked changes denoting where the text has been changed; the other a clean version (uploaded as the manuscript file).

Important additional instructions are given below your reviewer comments.

Thank you again for your submission to our journal. We hope that our editorial process has been constructive so far, and we welcome your feedback at any time. Please don't hesitate to contact us if you have any questions or comments.

Sincerely,

Jean Daunizeau

Academic Editor

PLOS Computational Biology

Lyle Graham

Section Editor

PLOS Computational Biology

***********************

A link appears below if there are any accompanying review attachments. If you believe any reviews to be missing, please contact ploscompbiol@plos.org immediately:

Reviewer's Responses to Questions

Comments to the Authors:

Please note here if the review is uploaded as an attachment.

Reviewer #1: I appreciate the revisions the authors made. The results are more straightforward to evaluate and better contextualized.

My main comments this time are to tighten the conclusions. The authors claim the following in the abstract:

1. The majority of learned latent dimensions in [the dVAE] are interpretable by human raters

2. These latent dimensions serve as a good encoding model for human fMRI data

3. [There is] a gradient from low- to high-level face feature representations along posterior to anterior face-selective regions

4. A decoding analysis confirms that the model separates identity-relevant and -irrelevant information

5. [There is] no spatial segregation between identity-relevant and -irrelevant face features

6. The few "entangled" (uninterpretable) dimensions

6a. match responses across the ventral stream

6b. carry significant information about facial identity

I think claims 1, 2, and 4 are now reasonably well supported. I appreciate the more direct description of interpretability ratings and the addition of Table S2. I agree that adding participants for rating is not essential to the paper's more interesting results on brain decoding.

Claim 3 is still unclear. Part of the issue is the wording—I think 'gradient' does not aptly describe comparisons between two ROIs. Moreover, while per-ROI statistics identify two face-specific (i.e., 'high-level') dimensions in FFA, between-ROI statistics only show statistically significant differences in an entangled feature in FFA, which does not directly support the claim of a gradient from low- to high-level features. (I do appreciate the addition of between-ROI statistics.) Another issue is that the evaluation of the voxel-wise results is purely qualitative, and I am unsure to what degree they support the claim of a gradient. Fig. S4 was hard for me to read. A different color map would help, giving high-level and low-level two families of colors (e.g., warm and cold). (The current color map puts features in groups of 2 that are irrelevant to the paper's claims, distracting, and require repeated references to the figure key.)

It is unclear what evidence directly supports claims 5 and 6a. Particularly for claim 5, the word segregation never appears in the main text. Are claims 5 and 6a based on the same evidence as claim 3 (i.e., Figs. 5 and S4)? Are both claims rigorously testable? I.e., what results, ideally quantitative ones, would support or reject the respective claims?

In claim 6b, it again helps to specify whether 'significant' refers to statistical significance. I found no statistical tests associated with Fig. 6. Is claim 6 really about 'above-chance' decoding? It helps to indicate chance in Fig. 6 (50% if I understand correctly).

I think suitably re-wording claims 3, 5, and 6 will not detract from the paper's significance and requires no additional analysis, although additional analysis may further strengthen claim 3.

Minor comments:

The Discussion explains well why the authors chose 24 latent dimensions (lines 10.31-10.40). A preview of this is due when this parameter was first introduced (line 4.31). The number of model latent dimensions is relevant. A different choice can potentially affect the conclusions about interpretability and the two classes of features (identity-relevant or not).

Why do the authors distinguish high- and low-level features vs. identity-relevant and -irrelevant features? The reason is implicit in some places (e.g., Fig. 6) but not in others, and juxtaposing the two categorization systems was confusing (e.g., in the abstract and on page 7). It would help the reader to explain why each analysis used either categorization and emphasize the subtle difference between the two since only one feature distinguishes them (dim 13. 'smile').

Line 4.39, 'agreed on': This phrasing is confusing. The Method (and reviewer response) is unambiguous—the authors agreed on 14 dimensions, and the other two dimensions were interpretable to one rater and conceded by the other.

Line 5.13, 'correlated': Do the latent dimensions correlate or have similar geometry? The word 'correlated' is confusing because it could mean individual features are correlated, which would contradict the result that dVAE features are more disentangled and interpretable.

In Fig. 5, it helps to annotate the dimensions showing significant differences between OFA and FFA.

Lines 7.37–7.40: The posterior voxels are not that clear. Where posteriorly can I see background (oranges) and image tone (pale blue)?

Lines 8.24-26, 'The role of information contained in the remaining entangled dimensions of a disentangled model is an open question in AI': I'm still unconvinced this is a significant question in AI. I must be less familiar with the literature than the authors. Thus, the authors can help by discussing or citing work that discusses why residual entangled dimensions are an interesting and important open question in AI.

Line 9.29, 'he combination': typo.

Lines 10.16–19: Given that the authors have conducted a direct (preliminary) analysis on alignment, it is well to mention it here. I agree with the authors about not including the relevant figures in the reviewer response as supplementary figures, but only because the plots show no interpretable differences, not because the analysis itself is distracting.

Reviewer #2: The authors have successfully addressed most of my questions and concerns. The only two points I would like to discuss with the authors are related to previous major concerns 3 and 5.

1. My understanding remains unclear regarding how the authors define the anterior and posterior regions. I am skeptical about relying solely on a few ventral regions to draw conclusions about the gradient. I would recommend that the authors discuss this limitation in the discussion section.

2. Research involving human fMRI data and deep neural network models appears to produce more inconsistent results when modeling faces compared to other object categories, such as scenes from natural datasets. I would be interested in the authors' perspective on this discrepancy.

**********

Have the authors made all data and (if applicable) computational code underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data and code underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data and code should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data or code —e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: No: Are the trained model weights tested in the paper shared? The GitHub repo has no README, and I didn't find a link to model weights after skimming the code. If the trained weights cannot be shared, the author should affirm that the shared code is sufficient for others to fully reproduce the study.

Reviewer #2: Yes

**********

PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

Figure Files:

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email us at figures@plos.org.

Data Requirements:

Please note that, as a condition of publication, PLOS' data policy requires that you make available all data used to draw the conclusions outlined in your manuscript. Data must be deposited in an appropriate repository, included within the body of the manuscript, or uploaded as supporting information. This includes all numerical values that were used to generate graphs, histograms etc.. For an example in PLOS Biology see here: http://www.plosbiology.org/article/info%3Adoi%2F10.1371%2Fjournal.pbio.1001908#s5.

Reproducibility:

To enhance the reproducibility of your results, we recommend that you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. Additionally, PLOS ONE offers an option to publish peer-reviewed clinical study protocols. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols

References:

Review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript.

If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice.

PLoS Comput Biol. doi: 10.1371/journal.pcbi.1011887.r005

Decision Letter 2

Lyle J Graham, Jean Daunizeau

2 Feb 2024

Dear Dr. Isik,

We are pleased to inform you that your manuscript 'Disentangled deep generative models reveal coding principles of the human face processing network' has been provisionally accepted for publication in PLOS Computational Biology.

Before your manuscript can be formally accepted you will need to complete some formatting changes, which you will receive in a follow up email. A member of our team will be in touch with a set of requests.

Please note that your manuscript will not be scheduled for publication until you have made the required changes, so a swift response is appreciated.

IMPORTANT: The editorial review process is now complete. PLOS will only permit corrections to spelling, formatting or significant scientific errors from this point onwards. Requests for major changes, or any which affect the scientific understanding of your work, will cause delays to the publication date of your manuscript.

Should you, your institution's press office or the journal office choose to press release your paper, you will automatically be opted out of early publication. We ask that you notify us now if you or your institution is planning to press release the article. All press must be co-ordinated with PLOS.

Thank you again for supporting Open Access publishing; we are looking forward to publishing your work in PLOS Computational Biology. 

Best regards,

Jean Daunizeau

Academic Editor

PLOS Computational Biology

Lyle Graham

Section Editor

PLOS Computational Biology

***********************************************************

Reviewer's Responses to Questions

Comments to the Authors:

Please note here if the review is uploaded as an attachment.

Reviewer #1: The authors have commendably addressed all my concerns. I heartily recommend publication. I only have some minor suggestions and do not need to see the manuscript again.

On page 2, lines 19–21, 27–29, 31: Are the authors presenting the notion that disentangled models learn semantically meaningful dimensions as a known fact or a hypothesis tested in the present study? I thought it was the latter. If so, lines 19–21 and 31 jump the gun a bit.

Page 5, lines 15–18: To my understanding, CCA identifies multiple canonical directions in descending order of the data correlation they capture. I’m guessing that the reported scalar r-values correspond to the top canonical direction per model pair. This may well be an implicit assumption in some literatures, but the authors can clarify the meaning.

Page 10, line 7, ‘model models’: typo.

Reviewer #2: The authors have addressed all my concerns, and I have no further questions. I recommend the article to be published.

**********

Have the authors made all data and (if applicable) computational code underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data and code underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data and code should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data or code —e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: Yes: Guo Jiahui

PLoS Comput Biol. doi: 10.1371/journal.pcbi.1011887.r006

Acceptance letter

Lyle J Graham, Jean Daunizeau

19 Feb 2024

PCOMPBIOL-D-23-00246R2

Disentangled deep generative models reveal coding principles of the human face processing network

Dear Dr Isik,

I am pleased to inform you that your manuscript has been formally accepted for publication in PLOS Computational Biology. Your manuscript is now with our production department and you will be notified of the publication date in due course.

The corresponding author will soon be receiving a typeset proof for review, to ensure errors have not been introduced during production. Please review the PDF proof of your manuscript carefully, as this is the last chance to correct any errors. Please note that major changes, or those which affect the scientific understanding of the work, will likely cause delays to the publication date of your manuscript.

Soon after your final files are uploaded, unless you have opted out, the early version of your manuscript will be published online. The date of the early version will be your article's publication date. The final article will be published to the same URL, and all versions of the paper will be accessible to readers.

Thank you again for supporting PLOS Computational Biology and open-access publishing. We are looking forward to publishing your work!

With kind regards,

Bernadett Koltai

PLOS Computational Biology | Carlyle House, Carlyle Road, Cambridge CB4 3DN | United Kingdom ploscompbiol@plos.org | Phone +44 (0) 1223-442824 | ploscompbiol.org | @PLOSCompBiol

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 Table. Group and individual subject model versus model significance results.

    P-value are in parentheses.

    (XLSX)

    pcbi.1011887.s001.xlsx (9.5KB, xlsx)
    S2 Table. Tallies for labels assigned to each of fifteen individual images in the labeling experiments by two raters, and each raters’ consensus label.

    (CSV)

    S1 Video. Animated latent traversals for all 24 latent dVAE dimensions for one example rendered face.

    Dimensions are varied from -2 to +2, with all other dimensions held constant. Images are model generated. Base image is from the CelebA dataset [20].

    (GIF)

    pcbi.1011887.s003.gif (1.6MB, gif)
    S2 Video. Animated latent traversals for all 24 latent dVAE dimensions for a second example rendered face.

    Dimensions are varied from -2 to +2, with all other dimensions held constant. Images are model generated. Base image is from the CelebA dataset [20].

    (GIF)

    pcbi.1011887.s004.gif (1.6MB, gif)
    S3 Video. Animated latent traversals for all 24 latent VAE dimensions for the same exampled rendered face in S1 Video.

    Dimensions are varied from -2 to + 2 with all other dimensions held constant. Images are model generated. Base image is from the CelebA dataset [20].

    (GIF)

    pcbi.1011887.s005.gif (2.2MB, gif)
    S4 Video. Animated latent traversals for all 24 latent VAE dimensions for the same exampled rendered face in S2 Video.

    Dimensions are varied from -2 to + 2 with all other dimensions held constant. Images are model generated. Base image is from the CelebA dataset [20].

    (GIF)

    pcbi.1011887.s006.gif (2.2MB, gif)
    S1 Fig. Whole brain encoding.

    Encoding model performance in all face-selective voxels. Ventral view of the remaining two subjects for dVAE (left), VAE (center) and VGG-face (right).

    (EPS)

    pcbi.1011887.s007.eps (4.1MB, eps)
    S2 Fig. Whole brain correlations for all 4 lateral view.

    (EPS)

    pcbi.1011887.s008.eps (8.8MB, eps)
    S3 Fig. Preference mapping for each dimension in the STS.

    Average feature predictivity across subjects is shown for each individual dimension. Dots represent individual subjects. Asterisks represents significance (p < .05). Dimensions are colored according to Table 1.

    (EPS)

    pcbi.1011887.s009.eps (2.2MB, eps)
    S4 Fig. Whole brain preference maps.

    Each voxel is shaded based on the latent dimension that provides the highest predictivity.

    (EPS)

    pcbi.1011887.s010.eps (6.4MB, eps)
    Attachment

    Submitted filename: DisentangledFaces_response.pdf

    pcbi.1011887.s011.pdf (833.7KB, pdf)
    Attachment

    Submitted filename: Review_Response.pdf

    pcbi.1011887.s012.pdf (115.8KB, pdf)

    Data Availability Statement

    All data used were publicly available and linked to the original paper: https://openneuro.org/datasets/ds001761. All analysis code is publicly available on our github repo https://github.com/psoulos/disentangle-faces.


    Articles from PLOS Computational Biology are provided here courtesy of PLOS

    RESOURCES