TY - JOUR T1 - Explaining face representation in the primate brain using different computational models JF - bioRxiv DO - 10.1101/2020.06.07.111930 SP - 2020.06.07.111930 AU - Le Chang AU - Bernhard Egger AU - Thomas Vetter AU - Doris Y. Tsao Y1 - 2021/01/01 UR - http://biorxiv.org/content/early/2021/03/30/2020.06.07.111930.abstract N2 - Understanding how the brain represents the identity of complex objects is a central challenge of visual neuroscience. The principles governing object processing have been extensively studied in the macaque face patch system, a sub-network of inferotemporal (IT) cortex specialized for face processing. A previous study reported that single face patch neurons encode axes of a generative model called the “active appearance” model, which transforms 50-d feature vectors separately representing facial shape and facial texture into facial images. However, a systematic investigation comparing this model to other computational models, especially convolutional neural network models that have shown success in explaining neural responses in the ventral visual stream, has been lacking. Here, we recorded responses of cells in the most anterior face patch AM to a large set of real face images and compared a large number of models for explaining neural responses. We found that the active appearance model better explained responses than any other model except CORnet-Z, a feedforward deep neural network trained on general object classification to classify non-face images, whose performance it tied on some face image sets and exceeded on others. Surprisingly, deep neural networks trained specifically on facial identification did not explain neural responses well. A major reason is that units in the network, unlike neurons, are less modulated by face-related factors unrelated to facial identification such as illumination.Competing Interest StatementThe authors have declared no competing interest. ER -