Abstract
Population receptive field (pRF) mapping is an influential neuroimaging technique used to estimate the region of visual space modulating the response of a neuronal population. While pRF mapping has advanced our understanding of visual cortical organization, evidence linking variation in pRF properties to behavioral performance remains limited. One of the most compelling pRF-to-behavior relationships has emerged from research into developmental prosopagnosia (DP). Individuals with DP have severe deficits in facial identity recognition sometimes linked to diminished holistic processing of faces. This perceptual deficit could be explained at the neural level by abnormally small pRFs in face-selective regions that restrict spatial integration of the face information. This hypothesis is supported by data from a small group of DPs but needs to be rigorously evaluated in a larger sample. Here, we measured pRF properties in 20 individuals with DP and 20 controls using a stimulus designed to robustly activate both low- and high-level visual areas. Consistent with previous studies, DPs exhibited weaker face selectivity in core ventral face-selective areas. Crucially, however, across the visual processing hierarchy – from early visual cortex, to intermediate visual areas, and face-selective areas – DPs and controls exhibited remarkable similarity in pRF properties, including pRF size, visual field coverage, and the distribution of pRF centers. Using a larger sample and the latest methods for mapping and modeling pRFs, these results challenge theories attributing DP to reduced spatial integration in face-selective regions. This underscores the need to explore alternative neural mechanisms of DP and to re-evaluate links between pRF properties and behavior more broadly.
Introduction
The receptive field (RF) is a fundamental concept in sensory neuroscience. In visual cortex, the RF of a sensory neuron refers to the region of visual space influencing the response of a neuron (Hubel and Wiesel, 1965). Nearby points in visual space are mapped to nearby points on the cortical surface forming a continuously varying map of the retinal input (i.e. a retinotopic map). Because neurons with similar RFs tend to be grouped together, human neuroimaging techniques have been developed for studying the aggregate RF properties of populations of neurons, termed the population RF or pRF (Dumoulin and Wandell, 2008; Harvey et al., 2013; Nasiotis et al., 2017; Winawer et al., 2013). Studies of pRFs measured within the volumetric units recorded by functional magnetic resonance imaging (fMRI) have revealed that a visuospatial framework structures much of the cortex – from early visual areas (Benson et al., 2018; Deyoe et al., 1996; Engel et al., 1997; Groen et al., 2022; Sereno et al., 1995) to high-level visual regions (Brewer et al., 2005; Finzi et al., 2021; Grill-Spector et al., 2018; Groen et al., 2022; Kay et al., 2015; Silson et al., 2015), and even other sensory modalities (Hagler and Sereno, 2006; Silver and Kastner, 2009; Steel et al., 2024; Szinte and Knapen, 2020; van Es et al., 2019).
Though a wealth of studies have produced many detailed descriptive accounts of pRF organization across a variety of brain regions, few studies to date have linked individual variation in pRF properties to variation in behavioral performance. One study found a correlation between individual variation in pRF sizes in early visual cortex and perceptual position discrimination thresholds (Song et al., 2015). Another study examining ventral face-selective regions found that pRFs mapped with upright face stimuli were larger and higher in the visual field than those mapped with inverted face stimuli (Poltoratski et al., 2021; but see Morsi et al., 2024), suggesting that large and foveally positioned pRFs are implicated in the common behavioral advantage for recognizing upright versus inverted faces.
The most striking finding linking variations in pRF properties to behavior, however, has emerged from the study of developmental prosopagnosia (DP; also known as congenital prosopagnosia). Individuals with DP have great difficulty recognizing facial identity despite having normal low-level vision, normal intelligence, and no history of brain damage (Avidan and Behrmann, 2021; Bate and Tree, 2017; Duchaine, 2011; McConachie, 1976; Susilo and Duchaine, 2013). Witthoft and colleagues (Witthoft et al., 2016) conducted retinotopic mapping in seven DPs and 15 controls and found that the DPs had pRFs that were far smaller and closer to the fovea in ventral face-selective regions. For instance, pRFs in a combined ventral face-selective ROI were approximately three times smaller in DPs compared to controls (control mean±SD = 3.4° ± 3.1°, DP mean±SD = 1.0° ± 0.37°). Similar differences were also seen in intermediate regions including hV4 and VO1, whereas no group differences were detected in early visual cortex. Furthermore, mean pRF sizes in the face-selective areas were positively correlated with performance on the Cambridge Face Memory Test (Duchaine and Nakayama, 2006).
These findings provide an appealing and tractable hypothesis about the abnormal nature of the computations underlying DP. A long history of perceptual research has characterized normal face processing as heavily dependent on the ability to form holistic representations consisting of facial features and their spatial relationships (Levine and Calvanio, 1989; Mckone and Yovel, 2009; Rossion, 2009; Tanaka and Farah, 1993; Yin, 1969; Young et al., 1987). Although the term holistic processing has been conceptualized in a variety of ways (Maurer et al., 2002), recent theoretical work has grounded it in an understanding of the receptive field characteristics of face-selective regions (Avidan and Behrmann, 2021; Grill-Spector et al., 2018; Witthoft et al., 2016). When a face is viewed from a typical distance of one meter, an early visual area, such as V1, would not enable holistic face processing, since a representative pRF from V1 only covers an area equivalent in size to the corner of the eye. In contrast, a pRF in a typical observer from a ventral face-selective area, such as one found on the medial fusiform (mFUS) gyrus, covers nearly the entire face, thus enabling the processing of multiple features in parallel as would be required for holistic processing. Therefore, a parsimonious account of the impairment in DP, and one supported by the data from Witthoft et al. (2016), is that the scope of the spatial integration window in face-selective regions is too small to permit simultaneous analysis of distal features, a phenomenon leading to the inability to get an overview of the face "as a whole in a single glance" (Levine and Calvanio, 1989). This interpretation dovetails with anecdotal reports from some DPs who have described relying on slow, compensatory strategies that depend on the piecemeal processing of individual face features (Duchaine, 2011).
The idea that reduced spatial integration underlies the deficit in DP has gained support (Avidan and Behrmann, 2021; Jiahui et al., 2018; Kaiser et al., 2019; Linka et al., 2022; Towler et al., 2018) and serves as one of the best examples linking pRF organization to high-level perception. However, several methodological limitations of the original study highlight the need for additional investigations into pRF characteristics in DP. First, in the years since the study was conducted, pRF results, especially in high-level regions, have been shown to depend critically on the type of carrier stimulus used during mapping. For instance, the ventral occipito-temporal reading circuitry showed less visual field coverage and higher variance explained for a word stimulus compared to a black-and- white checkerboard stimulus (Le et al., 2017). Across another set of studies, ventral face-selective regions were found to have lower/upper visual field biases when pRFs were mapped with scene images (Silson et al., 2015) but no bias when mapped with cartoon imagery (Finzi et al., 2021). Therefore, pRF mapping stimuli consisting of images of faces, objects, bodies, and scenes have been increasingly used in an effort to better match the preferences of a broad range of category-selective regions (Finzi et al., 2021; Kay et al., 2015; Silson et al., 2015). However, Witthoft et al. (2016), used a black-and-white checkerboard pattern as the mapping stimulus. While this decision is unlikely to have affected pRF estimates in EVC, results in higher-level areas are likely to have been impacted. In addition, the sample size used by Witthoft et al. (2016) is small by current standards (𝑁 = 7 DPs) and the findings have not been published in a peer-reviewed journal. Large-scale studies and meta-analyses in developmental dyslexia, a condition with many similarities to DP, indicate that neuroimaging studies with small samples are prone to false positives (Jednoróg et al., 2015; Ramus et al., 2018), echoing the concerns about reproducibility in the wider neuroscience community (Boekel et al., 2015; Button et al., 2013; Poldrack et al., 2017; Szucs and Ioannidis, 2017; Turner et al., 2018).
To deepen our understanding of the relationship between pRFs in category-selective areas and behavior as well as the role that spatial integration deficits might play in DP, we measured pRF properties in 20 DPs and 20 controls across early, intermediate, and category-selective areas. For mapping pRFs, we used stimuli containing vivid high-level images that effectively stimulate high-level category-selective regions as well as early visual areas. Moreover, we used the compressive spatial summation (CSS) model for estimating pRFs which is a more accurate model for high-level regions where responses to visual stimuli often sum in a subadditive rather than linear manner (Kay et al., 2013a,b). If DP stems from a deficit in spatial integration affecting face perception, we would expect to observe reduced pRF sizes and restricted visual field coverage in face-selective regions of DPs compared to neurotypical controls.
Results
To investigate spatial integration in developmental prosopagnosia (DP), we measured population receptive fields (pRFs) in a sample of 20 DPs and 20 controls across early, intermediate, and category-selective visual regions. During the pRF mapping experiment, participants fixated a central dot while variously-sized apertures (bars, wedges, and a disk) occupied portions of the visual field revealing colorful, high-level images flickering at a rate of 5 Hz (Supplemental Figure 5). To ensure stable fixation, participants performed a color change-detection task at fixation, and eye movements were recorded for quality control. The compressive spatial summation (CSS) model was used for estimating pRFs (Kay et al., 2013a,b), which produces estimates of pRF center position (expressed in either Cartesian or polar coordinates), pRF size (defined as 
Qualitatively, the pRF mapping procedure yielded the expected pattern of retinotopic polar angle and eccentricity maps in all participants – a hemifield representation along the calcarine sulcus flanked by mirror reversals of polar angle, accompanied by foveal representations near the occipital pole that gradually shift to more peripheral eccentricities anteriorly and medially. These maps served as the basis for defining V1, V2, V3, and hV4 with the dorsal and ventral arms of V1 through V3 combined. Delineation of hV4 followed the procedure outlined in Winawer and Witthoft (2017). Polar angle maps for each individual in the study are displayed in Supplemental Figures 6 and 7 as they serve to demonstrate that participants maintained fixation and pRF mapping worked appropriately.
Participants underwent functional localizer scans to define category selective regions including face-selective regions on the inferior occipital gyrus (occipital face area or OFA but also referred to as IOG-faces, Pitcher et al., 2011), posterior fusiform gyrus (pFUS, Weiner and Grill-spector, 2010) and mid-fusiform gyrus (mFUS, Weiner and Grill-spector, 2010). Additionally, an region on the parahippocampal gyrus selective for scenes (parahippocampal place area or PPA, Epstein and Kanwisher, 1998) was included to serve as a high-level control region not involved in the perception of faces.
Compared to controls, DPs had lower face-selectivity in all six face-selective ROIs, with face-selectivity defined as the percent signal change to blocks of faces minus the percent signal change to blocks of objects. These differences reached statistical significance in three areas (right and left OFA, and left pFUS; see Supplemental Table 3). Inspection of t-maps thresholded at t > 3.3 (uncorrected, p < 0.001) overlaid on the anatomical masks revealed that, despite slightly weaker face selectivity, DPs had detectable clusters of activity near the expected anatomical landmarks which is in keeping with previous findings (Avidan and Behrmann, 2009; Dinkelacker et al., 2011; Fur et al., 2011; Jiahui et al., 2018).
PRF parameters were fit individually to each voxel’s preprocessed time course. Voxels for which the variance explained by the pRF model (coefficient of determination, R2) was less than 20% were excluded from further analysis. ROIs from any participant with fewer than 10 voxels remaining after R2 thresholding were omitted from further analysis as the data was deemed insufficient to fully characterize the retinotopic response. The number of surviving ROIs was similar between DPs and controls (Figure 1C) with the exception of left mFUS, which was present in 18/20 controls versus 14/20 DPs.

PRF diagnostics and mean parameter values by group (control, DP) and hemisphere (left, right).
A) Mean coefficient of determination, R2, of all pRF fits included in ROIs B) Mean category selectivity of pRF selections from the functional localizer. Category selectivity was defined as the percent signal change for each ROI’s preferred category with respect to percent signal change for objects. Voxels were chosen by selecting the top 20% highest t values within anatomical masks. C) Counts of surviving ROIs after pRF fitting by group and hemisphere. An ROI was defined as retinotopically driven, and therefore included for analysis, if more than 10 voxels remained after R2 thresholding. D) The proportion of voxels that were retained after thresholding on R2 > .20. E) Mean pRF size, which was defined as 
To test for differences between groups (DPs, controls) by hemisphere (right, left) on various measures of pRF properties, a series of linear mixed-effects models (LMMs) were created. LMMs were selected for their ability to: accommodate missing data (e.g. absent ROIs), model the hierarchical/nested structure of the data (such as the fact that right/left hemisphere is nested within each functional ROI and ROIs were nested within subject (Etzel et al., 2011)) and provide robust inference without strict assumptions about sphericity or homogeneity of variances (Quené and Van Den Bergh, 2004). For each LMM, the random effects component included, at minimum, a random intercept for participant. The fixed effects portion of the model included group, hemisphere, and the interaction between group and hemisphere. In all models, the intercept was mapped to the control group in the right hemisphere. To evaluate the impact of each fixed effect on the dependent variable, likelihood ratio tests were conducted comparing models including the variable (or interaction) in question to a simpler model without it. All statistical analyses were performed in R using the lme4 Bates et al. (2015) and lmerTest Kuznetsova et al. (2017) packages.
Goodness of pRF model fits
We first compared the goodness-of-fit (R2) of the voxels following pRF fittings. Figure 1A shows that, consistent with previous studies (Finzi et al., 2021; Witthoft et al., 2016), a higher mean R2 was observed in the early visual cortex (V1, V2, and V3) and intermediate visual area hV4 than in face-selective areas (OFA, pFUS, mFUS) (Finzi et al., 2021; Witthoft et al., 2016). This pattern held for both controls (mean ± SE: EVC = 53.3±1.29; hV4 = 55.7±1.75; ventral face-selective regions = 41.6±0.81) and DPs (mean ± SE: EVC = 51.5±1.01; hV4 = 53.7±1.72; ventral face-selective regions = 42.6±1.24). Similar regional trends were found for the proportion of voxels within an ROI retained after thresholding R2 at 20% (see Figure 1D). Despite these regional differences in goodness-of-fit, the face-selective ROIs were still strongly retinotopically modulated by the pRF stimulus. Furthermore, across all ROIs examined, DPs and controls did not differ significantly in terms of R2 nor were there any significant interactions between group and hemisphere (LMMs on R2 with subject-specific random slopes nested within hemisphere, all p𝑠 > 0.01, see Supplementary Table 4 for all parameter estimates). The comparable goodness-of-fit values between DPs and controls and strong retinotopic modulation by the pRF stimulus indicate that the data quality was sufficient to support meaningful comparisons of pRF properties in these regions.
Eyetracking analysis
All participants were instructed on the importance of maintaining steady fixation throughout the entirety of the retinotopic mapping runs. Before beginning the experiment, participants were given ample time to practice fixating while inside the scanner, and eye tracking was performed whenever possible. The experimenters monitored real-time plots of eye positions during data collection and provided feedback in between runs, if needed. One control was eliminated from the study due to drowsiness and inability to maintain adequate central fixation. Another control had one of five retinotopic mapping runs dropped from analysis because the spread of eye positions in that run was unusually large. In total, eye tracking data was successfully recorded for all or some runs of the experiment in 14/20 controls and 14/20 DPs. An LMM on the area of the 95% probability contour of the distribution of eye positions (a summary of bivariate scatter) revealed no significant effect of group (𝜒2(1) = 0.40, p = 0.529), though there was a main effect of run (𝜒2(4) = 16.30, p = 0.003), with later runs having more dispersed eye positons. Despite this main effect of run, 95% probability contours were small across the majority of runs revealing that participants, regardless of group, were able to maintain central fixation throughout the retinotopic mapping runs. Heatmaps of eye position recordings for individual participants and runs are displayed in Supplemental Figures 8 and 9 and a group summary is displayed in Supplemental Figure 10.
DPs do not have restricted visual field coverage in face-selective or intermediate visual regions
Spatial representations within functional ROIs are built up from the responses of many pRFs that vary in their position and size within the visual field. To evaluate how pRFs collectively tile the visual field in DPs and controls, we measured the visual field coverage (VFC) within each ROI. To do this, pRFs were conceptualized as circles of radius 

Measures of visual field coverage (VFC).
VFC was computed as the proportion of pRFs covering each point in the visual field. A) Group averaged plots of VFC by ROI. Data from the right hemisphere has been reflected across the vertical meridian and combined with data from the left hemisphere. The dotted lines show the convex hull polygon at the 50% density threshold. B) Convex hull polygons, computed at the 50% density threshold, displayed for each individual subject by group and ROI. C) Cuts through the horizontal meridian showing normalized group coverage to the left and right of fixation (at 0 degrees) by hemisphere. D) Total area (in degrees of visual angle squared) for the convex hull polygons in panel B. Datapoints represent individual subjects. E) Proportion of the total convex hull area falling above the horizontal meridian. Values above 0.5 indicate an upper visual field bias and values below 0.5 indicate a lower visual field bias. F) Proportion of the convex hull area falling into the contralateral visual field.
To better quantify variation in VFC across individuals and groups, we computed the convex hull area of the VFC density, which is the area (in degrees of visual angle squared) of the smallest convex polygon wrapping around all points at a given density threshold. For the density threshold, we chose 50% of the normalized density as a robust middle ground. The convex hull area was measured separately for each participant and ROI. Figure 2B displays the convex hull polygons from each participant. In addition, a group-averaged convex hull polygon is superimposed on the visual field coverage density plots in Figure 2A as a dotted line.
In general, the VFC increased as one ascends from V1 through hV4 to the ventral face-selective regions, in accordance with previous studies (Benson et al., 2018; Finzi et al., 2021; Kay et al., 2015). An LMM with convex hull area as dependent measure and subject-specific random intercepts and fixed effects of ROI (V1, V2, V3, hV4, OFA, pFUS, mFUS, and PPA), group (DP, control), and hemisphere (right, left) revealed a strong main effect of ROI (𝜒2(6) = 160.71, p < 0.001), indicating that convex hull area was larger for later visual areas. However, no other main effects or interactions were significant (all p𝑠 > .05).
For all participants, we also computed the proportion of the total convex hull area that fell above the horizontal meridian (upper visual field bias) and the proportion that fell to the contralateral side of the vertical meridian (contralateral visual field bias). An LMM on the proportion of convex hull area above the horizontal meridian revealed a strong main effect of ROI (𝜒2(7) = 216.31, p < 0.001) but no other significant main effects (including group or hemisphere) and no significant interactions. The regions of V1, V2, and OFA exhibited a lower visual field bias (V1: t(39) = -6.56, p < 0.001; V2: t(39) = -3.53, p = 0.001; OFA: t(39) = -7.18, p < 0.001), and hV4 and PPA exhibited an upper visual field bias (hV4: t(39) = 2.94, p = 0.006; PPA: t(39) = 16.80, p < 0.001). Neither pFUS nor mFUS exhibited either an upper or lower visual field bias, which is consistent with several previous reports (Finzi et al., 2021; Kay et al., 2015, but see Silson et al., 2015). An LMM on the proportion of the convex hull area covering the contralateral visual field revealed a strong significant effect of ROI (𝜒2(7) = 132.72, p < 0.001) but no other significant main effects or interactions.
Together, these results indicate that, across the visual processing hierarchy, DPs and controls have highly similar visual field characteristics and, notably, DPs do not have reduced coverage of the visual field in any of the ROIs studied, including face-selective ones.
DPs have normal scaling between eccentricity and pRF size
Previous studies have found a strong linear relationship between pRF eccentricity and pRF size that spans early, intermediate, and ventral (category-selective) visual areas (Finzi et al., 2021; Groen et al., 2022; Kay et al., 2015). Critically, the slope of the line relating pRF eccentricity to pRF size increases as one ascends through the visual hierarchy. The previous results – highly similar mean pRF size between DPs and controls – could therefore have hidden atypical scaling of pRF size and eccentricity. Therefore, to compare how pRF sizes scale with eccentricity within DPs and controls, ROI-specific LMMs were created that modeled the linear relationship between pRF size and pRF eccentricity with subject-specific intercepts and slopes as random effects. Then, the effects of group, hemisphere, and their interaction were introduced as fixed effects. Parameter estimates from each ROI’s model are displayed in Table 1 and subject-specific slopes by group and ROI are displayed graphically in Figure 3. Since multiple models were constructed, a conservative alpha, 𝛼 < 0.01, was adopted to guide statistical interpretation. In none of the ROIs did group membership, hemisphere, or the interaction between the two significantly explain additional variance in relating pRF eccentricity to pRF size (all p𝑠 > 0.01). Hence, DPs and controls show highly similar relationships between pRF size and pRF eccentricity.

Lines of best fit relating pRF eccentricities (in degrees of visual angle) to pRF sizes by group and ROI.
Gray lines show lines of best fit for individual subjects and colored lines show the group average. Lines were constrained to the central 6 degrees of visual angle because data in far eccentricities were often sparse, especially in more anterior face-selective regions.

Fixed effect parameter estimates of pRF size by hemisphere (right, left) for the control and developmental prosopagnosic (DP) groups.
Separate models were created for each region of interest. Note that in each one the intercept has been mapped to the control group in the left hemisphere. Formula (R, lme4 package): size ∼ group + hemisphere + group:hemisphere + (1+eccentricity|subjectID) * p < 0.01; ** p < 0.001
DPs have normal distributions of pRF centers
Our results show that DPs and controls have very similar VFC. However, since VFC depends on both pRF sizes and pRF center locations – with larger pRFs covering more visual space and pRFs typically increasing in size with eccentricity – similar VFC patterns can arise even if the groups differ in the underlying spatial distribution of pRF centers. In other words, if one group had pRFs located more peripherally, the accompanying increase in pRF size could lead to comparable VFC between the groups despite underlying differences in where the pRFs are located. Therefore, in the next analysis, we directly compare groups on the distributions of pRF center locations.
To this end, we computed the area of the 95% probability contour encircling the bivariate distributions of 𝑥 and 𝑦 coordinates of pRF centers. Larger values indicate greater dispersion of pRF centers. In none of the ROIs (V1-hV4, OFA, pFUS, or mFUS) did the pRF dispersion differ significantly between DPs and controls (two sample t-tests, uncorrected, all p𝑠 > 0.05). The group-wise density of pRF center locations across all ROIs is represented in Figure 4A, and they show similar patterns between groups. Both groups also showed similar proportions of total pRFs occupying distinct eccentricity bins (Figure 4 B).

Spatial distributions of pRF center coordinates across ROIs and hemispheres.
A) Contours representing the group density distributions of pRF center coordinates by group and hemisphere. Green colors represent pRFs in the right hemisphere and blue colors represent pRFs in the left hemisphere. B) The proportion of pRFs falling into four distinct eccentricity bins. Datapoints represent individual participants. C) Measures of bivariate dispersion of pRF centers by group and ROI. Datapoints represent individual participants. Bivariate spread was measured by computing the area of the 95% probability contour encircling the distribution of pRF centers.
In summary, across multiple measures of retinotopic organization, DPs show no deviation from control participants. DPs and controls have highly similar coverage of the visual field, scaling between pRF eccentricity and size, and distribution of pRF centers.
Discussion
Population receptive field (pRF) mapping is a valuable technique in visual neuroscience that has provided an understanding of the location, size, and shape of aggregate receptive fields across the cortex (Groen et al., 2022; Wandell and Winawer, 2015), however few results have demonstrated strong relationships between individual variation in pRF properties and behavioral measures of perception. One of the most striking studies showing a link between pRF properties and behavior reported that a group of individuals with developmental prosopagnosia (DP) had exceptionally small and foveally-concentrated pRFs in intermediate visual areas and face-selective regions (Witthoft et al., 2016). Moreover, pRF size correlated with their face recognition performance. These findings presented an appealing neural explanation for the diminished holistic face processing sometimes observed in DPs (Avidan and Behrmann, 2021; Avidan et al., 2011; Degutis, 2012; Palermo et al., 2011).
To rigorously investigate spatial integration in DPs, we used fMRI to measure pRF properties in DPs and controls (𝑁 = 20 per group) using high-level mapping stimuli capable of robustly stimulating both early and category-selective visual regions. In contrast to Witthoft et al. (2016), our results provide strong empirical evidence that pRF properties are highly similar between DPs and controls. Both groups showed comparable: 1) visual field coverage (VFC) maps, providing a measure of how pRFs within an ROI collectively tile the visual field; 2) scaling between pRF eccentricity and pRF size, both within regions of interest (ROIs) and across the visual hierarchy; and 3) dispersion of pRF centers around the visual field. These three patterns of results held for early, intermediate, and ventral face-selective areas, as well as a category-selective control ROI not involved in face processing (the scene-selective parahippocampal place area or PPA; Epstein and Kanwisher (1998)). Similarly, we found no significant association between participants’ CFMT scores and mean pRF size in OFA, pFUS, or mFUS.
Methodological strengths of the current study
Our methods incorporated several features that ensure the validity and robustness of the findings. First, we employed a carefully thought-out, ecologically valid stimulus capable of significantly stimulating both early visual and category-selective regions. The stimulus featured vivid, high-contrast content including images of faces, objects, bodies, and places randomly placed on a background of pink noise. Although high-level regions have been mapped with low-level stimuli before (e.g., black- and-white checkerboard patterns), these types of stimuli do not drive category-selective regions as strongly, producing poorer goodness-of-fit (Le et al., 2017; Silson et al., 2015) and unstable pRF estimates.
Second, we acquired in-scanner eye-tracking data during the retinotopic mapping sessions whenever it was feasible. Eye movements can severely distort pRF measurements, usually by increasing estimated pRF size, so it was critical to compare fixation behavior in the DPs and controls. Participants with usable eye-tracking data demonstrated steady fixation, with no significant differences in gaze stability between DPs and controls. For participants without usable eye tracking data – due to either interference from thick prescription lenses or incompatibility of head anatomy and hardware setup – consistent fixation was confirmed by checking for clearly delineated polar angle and eccentricity maps in early visual cortex (Supplemental Figures 6 and 7).
One explanation for the comparable results for the DPs and controls is that the putative DPs may not have possessed a true face recognition deficit. However, we are confident that our selection process effectively categorized DPs for the following reasons. First, all DPs self-reported substantial face recognition difficulties affecting their daily lives, quantified by scores on the PI-20 questionnaire (Shah et al., 2015). Furthermore, objective inclusion criteria required DPs to score more than 1.8 standard deviations below the mean of a control sample on two independent tests of face recognition – the CFMT and a famous faces test – a stringent threshold consistent with or exceeding those used in numerous studies of DP (DeGutis et al., 2023). Furthermore, using not one but two tests for classification substantially minimizes the risk of false identification. Additional testing with the DPs was conducted to rule out low-level visual deficits or autistic-like traits as alternative causes for poor face performance. Univariate measures of face-selectivity provided another opportunity to confirm that our DP participants showed the neural abnormalities that have been found in previous fMRI studies of DP (Fur et al., 2011; Jiahui et al., 2018). Univariate functional localizer analyses from our pool of participants revealed that, at the group level, DPs compared to controls exhibited significantly lower face-selective activation in many core ventral face-selective regions (See Supplemental Table 3), replicating previous findings (Dinkelacker et al., 2011; Fur et al., 2011; Gerlach et al., 2019; Jiahui et al., 2018). Notably, limiting voxels to only those that showed strong retinotopic modulation, still revealed significant reductions in face-selectivity in DPs (Figure 1 B) - a novel finding that provides additional evidence that DPs were classified correctly.
Confidence in our pRF mapping methods was further supported by the fact that our results replicated many key pRF phenomenon from prior studies. Both groups showed significant increases in pRF size across the visual processing hierarchy (Finzi et al., 2021; Kay et al., 2015; Silson et al., 2015) and, within ROIs, there was a consistent increase in pRF size with pRF eccentricity (Benson et al., 2018; Kay et al., 2015). Regardless of group, the three ventral face-selective areas exhibited a higher proportion of foveally-positioned pRFs than early visual areas (Grill-Spector et al., 2018; Kay et al., 2015; Silson et al., 2015). Furthermore, these face-selective areas showed a relatively higher degree of subadditive spatial summation (larger static nonlinearity parameter in the CSS model; Kay et al., 2015, 2013a,b). Also, we replicated previously reported visual field biases – namely, a lower visual field bias in OFA (Gomez et al., 2018; Silson et al., 2015) and an upper visual field bias in PPA (Silson et al., 2015). Across both groups, qualitative inspection of polar angle maps near the calcarine sulcus in each hemisphere revealed the canonical retinotopic organization of smooth sweeps from the upper vertical meridian to the lower vertical meridian through the contralateral side of the visual field. Eccentricity maps likewise showed the expected progression from foveal pRFs near the occipital pole to increasingly more peripheral pRFs anteriorly and medially (Benson et al., 2018). Together, these multiple converging lines of evidence confirm that data quality was high and closely matched between groups.
How do our pRF results fit in with other behavioral data from DPs?
Our results directly challenge the view that DP stems from impaired spatial integration caused by abnormally small pRFs. While this view has garnered substantial interest (Avidan and Behrmann, 2021; Grill-Spector et al., 2018; Witthoft et al., 2016), some behavioral evidence is inconsistent with it. Behaviorally, the most common tasks used to evaluate holistic face processing are the composite (Young et al., 1987) and part-whole tasks (Tanaka and Farah, 1993). Comparisons between DPs and controls on these measures have yielded some mixed results: some report reduced holistic processing (Avidan et al., 2011; Degutis, 2012; Palermo et al., 2011), while others find it intact (Biotti et al., 2017; Le Grand et al., 2006; Susilo et al., 2010). This inconsistency highlights the heterogeneous nature of DP and complicates linking DP to a single underlying cause such as poor spatial integration.
Behavioral face studies that experimentally restrict spatial integration further challenge the notion that spatial integration deficits underlie DP. For instance, Tsantani et al. (2020) compared the ability of DPs and controls to identify famous faces under two viewing conditions: when faces were presented briefly in their entirety versus progressively revealed by a dynamic, narrow aperture. The dynamic aperture viewing condition allowed participants to inspect local features of the face but prevented distant regions from being processed simultaneously, thus simulating the effect of having a small window of integration. In this condition, control participants exhibited substantial decrements in performance likely because it hindered holistic face processing mechanisms. If the DPs’ deficits at face recognition stemmed from deficient sampling of the visual field by pRFs, one would expect the dynamic aperture viewing condition to elicit a smaller decrement in performance. However, contrary to this prediction, both groups showed comparable performance decrements. Similar results were found when DPs were tested with gaze-contingent displays restricting information to a small region around the fovea (Verfaillie et al., 2014), reinforcing the conclusion that spatial integration deficits alone do not underlie DP.
Lastly, DP is defined by deficits with facial identity processing. However, if DP is a deficit in spatial integration, individuals with DP should have difficulties with any aspect of face processing that involves holistic perception. Facial sex judgments (Baudouin and Humphreys, 2006) and facial expression recognition (Calder et al., 2000) both elicit composite effects that are similar in size to composite tasks involving facial identity recognition. Most DPs though perform as well as controls on face perception tasks involving judgments not involving identity such as sex classification (Chatterjee and Nakayama, 2012; Degutis, 2012; Marsh et al., 2019) and facial expression recognition (Bell et al., 2023; Palermo et al., 2011), so they do not appear to have broad deficits with spatial integration of facial information (Baudouin and Humphreys, 2006; Calder et al., 2000; Durand et al., 2007; Yokoyama et al., 2014).
What factors may have contributed to the different results for the present study and Witthoft et al. (2016)?
In light of the difference in findings between our study and that of Witthoft et al. (2016), several methodological differences between the two studies are worth considering. First, while Witthoft et al. (2016) mapped pRFs using black-and-white checkerboard patterns as the carrier image, we employed a more ecologically valid stimulus featuring colorful high-level images including faces, scenes, body parts, and objects. Although both types of stimuli would be expected to robustly stimulate early visual cortex, our stimulus better targets higher-level regions as reflected in noticeably higher goodness of fit, R2, in our data compared to that of Witthoft et al. (2016) (current study mean R2: OFA=.45, pFUS=.42, mFUS=.33; Witthoft et al. (2016) approximate mean R2: OFA=.21, pFUS=.15, mFUS=.07;). This boost agrees with prior within-subject studies demonstrating that more naturalistic and region-appropriate mapping stimuli outperform black-and-white checkerboard patterns in category-selective areas (Le et al., 2017; Silson et al., 2015).
More critically, the choice of stimulus can systematically alter pRF parameters themselves. For instance, Le et al. (2017) showed that visual field coverage (VFC) in ventral occipito-temporal cortex, a word form-selective region, shrinks toward the central visual field when mapped with word stimuli compared to checkerboard patterns. Thus, the results with DPs - reduced VFC with checker-boards but normal VFC with more naturalistic mapping stimuli - may reflect a stimulus-dependent effect that is elicited solely by low-level, non-preferred stimuli in face-selective regions. Nevertheless, our findings, demonstrating normal spatial integration in face-selective regions when using more appropriate, naturalistic stimuli offer a more definite and convincing proof of neural tuning in these higher-level regions.
To model pRFs, we used the compressive spatial summation (CSS; Kay et al. (2013a,b)) model as opposed to a linear pRF model as used in Witthoft et al. (2016). The main difference between these two models is that the CSS model applies a compressive static nonlinearity to account for subadditive spatial summation (situations where the sum of the individual responses to apertures forming complementary pairs is < 1). Subadditive spatial summation is present in EVC and grows more pronounced in higher-order regions, making this model more appropriate for studying spatial integration in face-selective regions.
It is also worth noting that Witthoft et al. (2016) mapped pRFs with stimuli spanning approximately 30° in diameter, compared to 14.7° in the present study, a substantial difference in field of view. For us, it was necessary to limit the field of view presented inside the scanner in order to collect high-quality eye-tracking data for verification of stable fixation, which we consider a critical step. Nevertheless, the field of view we used still greatly exceeds the typical visual angle subtended by a face at normal conversational distance (∼8°, McKone (2009)), so this design appears to provide a sufficiently wide enough visual field to detect any spatial integration deficits in DPs, if present.
Finally, there were differences in the aperture design used to map pRFs. Witthoft et al. (2016) employed blocked runs of rotating wedges and expanding rings (2 runs each, 4 runs total), whereas we used a sequence of apertures including bars (sweeping in 8 different directions), one rotating wedge, and a brief full-field stimulation. While the precise impact of aperture design on pRF estimates is not well understood (for one investigation see Alvarez et al. (2015)), our aperture design successfully replicated many established pRF properties and it remains unclear why this difference between the designs would have differentially affect DPs and controls. Moreover, our study col-lected more than twice as much data per participant (5 runs of 6.8 mins each versus 4 runs of 3.7 mins each), enhancing the precision and reliability of pRF parameter estimates, especially when combined with a more effective mapping stimulus.
Conclusion
Although we do agree that abnormally small pRFs and restricted VFC in face-selective regions could cause deficits to spatial integration mechanisms necessary for face recognition, our results – using rigorous pRF methods, a large sample, and region-appropriate mapping stimuli – indicate that spatial integration, as measured by pRFs, is normal in DP. In line with many previous studies, however, we did discover diminished face selectivity in DPs within the posterior ventral face areas that points to qualitative differences in how face-selective regions respond specifically to face stimuli. Face recognition deficits in DP may therefore arise from atypical neural tuning to facial information rather than from atypical spatial integration, shifting theoretical accounts of DP toward mechanisms involving representational content rather than receptive field properties.
Materials and methods
Participants
Twenty individuals with DP (15 females, 𝑀𝑎𝑔𝑒 = 38.18, 𝑆𝐷𝑎𝑔𝑒 = 11.13) and 20 typical controls (14 females, 𝑀𝑎𝑔𝑒 = 33.86, 𝑆𝐷𝑎𝑔𝑒 = 10.92) participated in the study. The mean age of the groups did not differ significantly (t(40.97) = 1.286, p = .206, 𝑑 = 0.398, 𝐶𝐼95% = -0.216, 1.006). Written informed consent was obtained from all participants in accordance with the Declaration of Helsinki and a protocol approved by the Dartmouth College Committee for the Protection of Human Subjects (#23282).
Diagnostic testing and participant selection
DP participants were recruited from our prosopagnosia database (www.faceblind.org), and interested participants first completed the Twenty-Item Prosopagnosia Index (PI20; Shah et al. (2015)). The PI20 responses from every DP participant enrolled in the study indicated substantial face recognition difficulties that affected their daily lives (M = 81.90/100, SD = 8.86).
Face recognition ability was assessed with two online tests: The Cambridge Face Memory Test (CFMT; Duchaine and Nakayama (2006)) and a Famous Faces Test (FFT; Kieseler and Duchaine (2023)). Individual scores and control summary statistics are presented in Table 2. The famous faces test was newly created and required participants to identify (via free response) the faces of 40 celebrities familiar to North Americans. Immediately after completing the test, participants were asked about their familiarity with each celebrity. Human raters judged if each response was sufficiently correct, and each participant’s performance was quantified as the number of correct responses out of the total number of celebrities they reported being familiar with. Potential DPs’ face recognition was deemed sufficiently impaired if they scored more than 1.8 standard deviations below the control sample means on both the CFMT and FFT. Scores on the CFMT were compared against data from 50 typical observers reported by (Duchaine and Nakayama, 2006), and scores on the FFT were compared against data from N typical observers collected online through www.testable.org.

Scores from each member of the prosopagnosic sample on diagnostic tests consisting of: The Hanover Early Visual Assessment (HEVA), The 20-Item Prosopagnosia Index (PI20), The Cambridge Face Memory Test (CFMT), The Cambridge Face Memory Test with Australian Faces (CFMT Aus), famous faces test (FFT), old/new faces test, and The Subthreshold Autism Trait Questionairre (SATQ).
Asterisks (*) indicates a given score meets the inclusion criterion set in advance of the study.
In cases where potential DP participants met the inclusion criteria for only one of the two face recognition tests, participants were given a third test. If a participant’s CFMT score was in the normal range, they were administered the CFMT-Australian (McKone et al., 2011); if their Famous Face Test score was in the normal range, they were tested with an old/new test involving female faces (Duchaine and Nakayama, 2005). The CFMT-Australian follows the same format and scoring as the original CFMT. The old/new test required memorization of ten target faces, followed by 50 test trials (20 targets plus 30 distractors). Performance on the old/new test was evaluated by calculating A prime scores. Participants were included in the study if they scored below the threshold (CFMT-Australian: 𝑍 < -1.8; Old/new faces: 𝑍 < -2) on either of the follow-up tests.
To ensure that the DPs’ impairments in face recognition reflected a high-level deficit rather than a deficit to early visual processing, each DP recruit was assessed with the Hanover Early Visual Assessment (Kieseler et al., 2022). Any potential recruit with a score below 1.8 SDs below the mean of a control sample (𝑁 = 117) was not given further consideration for the study. This left a final sample of DPs that scored very similarly to controls (Controls: mean±SE = 90.0±13.3; DPs: mean±SE = 95.6±7.8).
Because rates of face recognition impairments are elevated in Autism Spectrum condition (Griffin et al., 2021), participants were also screened for subthreshold autistic traits using the self-report Subthreshold Autism Trait Questionnaire (SATQ; Kanne et al. (2012)). In our analysis of the SATQ responses, we omitted 5 of the 24 questions that measured behavioral tendencies likely to be influenced by poor face recognition (“I enjoy social situations where I can meet new people and chat (i.e. parties, dances, sports, and games)”; “I seek out and approach others for social interactions”; “Others think I am strange or bizarre”; “I have some behaviors that others consider strange or odd”; “I make eye contact when talking with others”). All DPs were less than 2 SDs above the mean of a control sample (𝑁 = 133; Bell et al. (2023)).
Experiment design
Session 1: Functional localizer
To localize face-selective regions of interest (ROIs), a category localizer was run that presented six classes of stimuli: faces, bodies, objects, phase-scrambled objects, natural scenes, and words. Stimuli consisted of dynamic video clips featuring various visual elements in motion, all filmed against a uniform black backdrop to maximize contrast and visibility. Videos of the natural scenes filled the entire frame and displayed a slow forward camera movement equivalent to a normal walking pace from a first-person perspective. Video clips subtended ∼26° 𝑥 15° of visual angle in width and height. Participants viewed four runs, each of which lasted 8 minutes and 22 seconds. Within each run, stimuli were grouped into 14-second blocks containing five videos each. Blocks of each category were displayed four times in each run in a quasi-random order, twice in color and twice in grayscale. The trial order was the same for all participants. Throughout the experiment, participants were allowed to freely move their eyes and instructed to identify back-to-back repeats with a button press (1-back task). Each of the four runs presented unique stimuli.
Face-selective ROIs were created by analyzing voxels within masks centered on associated anatomical landmarks and selecting the 20% of the voxels within the mask with the highest t-values for the contrast comparing faces versus objects. This method was chosen for its ability to generate individually tailored functional ROIs, while at the same time ensuring equal voxel counts across participants as well as retention of data from participants whose activations might otherwise fall below fixed thresholds (Jiahui et al., 2018; Norman-Haignere et al., 2013). Additionally, an ROI on the parahippocampal gyrus selective more for scenes than objects (parahippocampal place area or PPA Epstein and Kanwisher, 1998) was included to serve as a category-selective control region not involved in face perception.
Session 2: pRF mapping
Each participant completed five runs of retinotopy to identify the portion of the visual field capable of eliciting a response from the population of neurons within a voxel (i.e., each voxel’s pRF). In each run, participants fixated a central dot (0.2° x 0.2°) while apertures of various shapes and sizes gradually swept across the visual field revealing colorful high-level images flickering at a rate of 5 Hz.
To elicit a strong BOLD response across a variety of category-selective areas, we used stimuli introduced by Benson et al. (2018). These stimuli were created by taking cutout images from Kriegeskorte et al. (2008) (containing faces, body parts, foods, objects, street signs, animals, and architectural elements, etc.) and placing them at random positions and scales on achromatic pink-noise backgrounds (1/𝑓 amplitude spectrum). As such, we refer to them as “image mashup”. Our early pilot studies demonstrated the image mashup stimuli were effective at driving responses in face-selective regions but not scene-selective regions, which were of tangential interest to our study of DP (see Jiahui et al. (2018)). Therefore, we interleaved the image mashup stimuli with photographs of natural landscapes, and pilot data showed this combined approach generated a strong response in both face-selective and scene-selective regions. The image mashup and natural landscape stimuli alternated at a rate of 5 Hz, so each 2 s aperture presentation included an equal number of presentations of each stimulus type. Apertures and images were prepared at a resolution of 1060 x 1060 pixels and were constrained to a central circular region of diameter 14.7°. All stimuli were presented against a medium gray background with a faint semi-transparent polar grid pattern superimposed to facilitate fixation.
Three different types of apertures were presented in every run – a full field disk, a sequence of bars, and a rotating wedge (a schematic diagram is included in Supplemental Figure 5). Apertures progressed the same way in every run: After 8 s of initial blank fixation, a disk-shaped aperture appeared that revealed stimuli within the entire circular extent of the visual field, lasting for a duration of 4 s. Following 6 s of rest, bar-shaped apertures traversed the visual field in sweeps of 18 discrete, evenly spaced steps, each lasting 2 s. Each bar presentation overlapped 50% of the area of the previous position. Eight sweeps occurred, each interleaved with 6 s of rest, in the following order for all runs: Top–Bottom, Bottom–Top, Right–Left, Left–Right, Upper Right–Bottom Left, Bottom Left–Upper Right, Upper Left–Bottom Right, Bottom Right–Upper Left. Following another 6 s of rest, a 90° wedge appeared in the upper left quadrant and made 16 discrete clockwise steps, 2 s each, sweeping a full 360° around the central circular visual field (22.5° turn per step). Each run ended with 17 s of final fixation.
Throughout all retinotopy scans, participants performed a color detection task at fixation in which they reported via a button press when the white fixation dot changed to red. The central dot changed color to either red, green, or blue semi-randomly at a rate of approximately 30 per minute.
MRI data acquisition
All functional and structural images were acquired using a 3 Tesla Siemens Prisma scanner at the Dartmouth Brain Imaging Center at Dartmouth College. A 32-channel receive-only phased array head coil was used for all data acquisition.
T1 images: For each participant, a high-resolution whole-brain anatomical volume was collected using a T1-weighted magnetization-prepared rapid acquisition gradient echo (MPRAGE) sequence (1 mm isovoxel resolution; 208 sagittal slices; TR = 2,300 ms; TE = 2.03 ms; flip angle = 9◦; FOV = 256 x 256 mm, bandwidth = 240 Hz/px).
Functional images: Two types of functional scans were acquired across two sessions, both using a T2*-weighted gradient-recalled echoplanar imaging multiband pulse sequence. Across all scans, slices were oriented approximately in plane with the calcarine sulcus to ensure coverage of most, and often all, of the temporal, parietal, and occipital lobes. At the beginning of each session, a pair of EPI images with phase encoding directions of opposite polarity in the anterior-to-posterior plane were collected for post-hoc correction of EPI spatial distortion. Session 1 consisted of localizer scans designed to identify category-selective regions of interest (nominal spatial resolution = 2 mm x 2 mm x 2mm; 69 oblique slices; TR = 2,000 ms; TE = 30 ms; flip angle = 79◦; matrix size 106 x 106; field of view = 212 mm x 212 mm; phase partial Fourier scheme of 6/8; bandwidth = 1,814 Hz/px; echo spacing = 0.66 ms; excite pulse duration = 8,200 microseconds; multiband acceleration factor = 3; phase encoding direction = AP). Session 2 comprised the event-related retinotopy scans and therefore used a scanning protocol that prioritized a faster TR to sample the hemodynamic response more finely (nominal spatial resolution = 2 mm x 2 mm x 2mm; 69 oblique slices; TR = 1,350 ms; TE = 33.60 ms; flip angle = 68◦; matrix size 106 x 106; field of view = 212 mm x 212 mm; phase partial Fourier scheme off; bandwidth = 2358 Hz/px; echo spacing = 0.54 ms; excite pulse duration = 5,140 microseconds; multiband acceleration factor = 4; phase encoding direction = AP)
Stimulus display and scanner peripherals
Stimuli were presented using a Panasonic PT-DW750BU 7000-Lumen WXGA DLP Projector and an SR Research in-bore back-projection screen positioned just inside the head end of the magnet. Use of an in-bore screen allowed for a stimulus that subtended a greater visual angle. Participants viewed the screen via a mirror mounted atop the head coil. The viewing distance was 127 mm from the eye to the mirror + 1,010 mm from the mirror to the screen = 1,137 mm total. The maximum extent of the image projected on the screen was 297 mm x 297 mm. This resulted in a maximum possible visual angle of 14.97 ° square.
A Lenovo Thinkpad T480s computer running Linux Ubuntu 20.04.6 (Focal Fossa) controlled the stimulus presentations and recorded button presses using Matlab R2022a and extensions from PsychToolbox version 3.0.18.
To verify accurate and steady eye fixation during the retinotopic mapping runs, eye movements were recorded using an SR Research Eyelink 1000 eyetracker mounted underneath the back projection screen. Eye tracking was performed for the right eye, and samples were obtained at 250 Hz using the Pupil-CR centroid mode with default thresholding. Black gaffer tape was applied to the eye cutouts of the headcoil in an effort to reduce unwanted reflections from the infrared illuminator. Before each run, a calibration, validation, and drift correction was performed, and data was discarded if the results of the calibration were poor. During data collection, the experimenters monitored real-time plots of eye position shown on the Eyelink host computer located inside the MRI control room. Experimenters intervened and restarted data collection if eye fixation was not maintained steadily enough throughout the run. The eyetracking data, however, was low-quality for some participants. Notably, achieving sufficient pupil contrast was more difficult for participants who required MR-compatible corrective lenses. In addition, the nose bridge on the head coil sometimes partially occluded the pupil for participants with narrow inter-pupillary distance. High-quality eye-tracking data was obtained from 14/20 controls and 14/20 DPs. Based on visual inspection of the eye gaze recordings, one control participant was discarded due to insufficiently stable eye fixation.
Behavioral responses from scanning sessions were collected using a Current Designs two-button fiber optic handheld response pad connected to a Current Designs 932 interface.
Pre-processing of MRI data
All DICOM images were converted to NIfTI format using dcm2niix version 1.0.2018.11.25 (Li et al., 2016). From each participant’s T1-weighted volume, cortical surface meshes of the pial/gray matter boundaries and gray matter/white matter boundaries were reconstructed using Freesurfer’s recon-all program (Fischl and Dale, 2000). Functional data from the retinotopy scans were temporally resampled with cubic interpolation to correct for differences in slice time acquisition and to simultaneously upsample the time resolution from 1,350 ms to 1,000 ms with a custom Matlab script.
Subsequent pre-processing was performed using AFNI version 23.3.12 Septimius Severus (Cox, 1996). AFNI’s afni_proc.py was used to create processing pipelines for both the localizer and retinotopy scans. Slice timing differences were corrected, and non-linear geometric distortion correction was applied using separately acquired volumes in which the phase-encoding direction was reversed (‘blip-up/blip-down’ strategy). EPI-to-anatomical alignment was carried out using the lpc+ZZ cost function while checking for possible left-right flips (Reynolds et al., 2023). Motion correction was applied, and timepoints were censored when the Euclidean norm (enorm) exceeded 0.3 mm or when the outlier fraction exceeded 5%. Spatial blurring of FWHM = 4 mm (twice the EPI voxel dimensions) was applied to the localizer data but omitted from preprocessing of the retinotopy data. Volumetric data from each scan was then projected onto the corresponding hemispheres of the high-density SUMA standard meshes (created from the Freesurfer output with AFNI function @SUMA_Make_Spec_FS). Finally, each time series was scaled to units of local BOLD percent signal change. All pre-processing results were carefully visually inspected to ensure quality control using the QC HTML reports generated by afni_proc.py (Reynolds et al., 2023).
pRF analysis
The population receptive field (pRF) was modeled using the compressive spatial summation (CSS) model (Kay et al., 2013a), which was fit independently to each voxel’s preprocessed timecourse. The CSS model fits a 2D isotropic Gaussian to each voxel of the form:

with parameters 𝑥′ and 𝑦′ describing the center coordinates of the pRF, and 𝜎, the standard deviation of its Gaussian profile. The Gaussian is normalized to the unit area to make the amplitude of the response more interpretable as a percent increase.
The input stimulus to the CSS model consists solely of the spatial extent (not content) of the flickering imagery revealed through the aperture or mask at each timepoint. Therefore, the stimuli for the purpose of model fitting were prepared by converting the aperture masks to binarized (contrast) images with values including 0 (no contrast) and 1 (full contrast). Masks were downsampled to 200 by 200 pixels and then concatenated across time. The predicted response at each timepoint was obtained by computing the overlap (via dot product) between each stimulus mask and the 2D Gaussian, then multiplying it by a non-negative scaling factor, 𝑔 for gain, and exponentiating the result by a parameter 𝑛 constituting a static power-law nonlinearity. Afterward, the resulting time series was convolved with a model of the hemodynamic response function. The predicted BOLD response at each timepoint, 𝐵(t), can therefore be expressed formally as:

where 𝑆(𝑥, 𝑦, t) is the contrast image at timepoint t, 𝐺 is the 2D isotropic Gaussian, and ℎ(t) is the hemodynamic response function used in SPM. The static power-law nonlinearity, 𝑛, is included because it captures the nonlinear spatial summation and position tolerance of pRFs, particularly in the later stages of the visual hierarchy, in a single parameter that is easy to interpret (Finzi et al., 2021; Kay et al., 2013a; Le et al., 2017). The parameter 𝑛 is typically observed to be less than 1 making it a compressive exponent.
After fitting the five pRF parameters (𝑥, 𝑦, 𝜎, 𝑔, 𝑛) in the CSS model, the Cartesian coordinates, (𝑥, 𝑦), of the pRF centers were converted to polar coordinates to produce maps of polar angle, 


Model fitting was performed using the Levenberg-Marquardt approach to nonlinear parameter estimation (MATLAB function lsqcurvefit() from the Optimization Toolbox). The initial seed for the solver was chosen by pre-computing model predictions for 4,632 grid points, with different combinations of 𝑥, 𝑦, 𝜎, and 𝑛, and picking the point that produces the highest correlation with the observed data. To avoid local minima, a two-stage refined fitting procedure was implemented. In the first stage, we optimized the 𝑥, 𝑦, 𝜎, and 𝑔 parameters with the 𝑛 parameter, controlling the compressive nonlinearity, fixed at one of three values (0.5, 0.25, or 0.125, chosen based on the prior grid search). In the second stage, the seed was updated to the solution found in the first stage, and then all five parameters were optimized simultaneously. Lower bounds of 0 and 0.1 were placed on the 𝜎 and 𝑛 parameters, respectively. The maximum allowed iterations was 500, and the step tolerance was 1𝑒 − 6. Data from all five runs was averaged together in time prior to fitting pRFs.
Eye tracking analysis
For participants for whom eye movements were successfully recorded (14/20 controls and 14/20 DPs), the time series of horizontal and vertical eye positions were obtained and parsed by the proprietary Eyelink software . All pre-processing steps were performed separately by run. First, time periods detected by the Eyelink software as blinks were removed along with an additional 300 msecs on either shoulder. Censored datapoints were then interpolated using a method based on discrete cosine transforms (Wang et al., 2012), and the resultant time series were low-pass zero-phase filtered using a second-order Butterworth filter with cutoff frequency of 15 Hz. Then, the data were detrended using a linear and quadratic polynomial to account for slow drifts in gaze position. Finally, the data were median-centered, the blink periods were recensored, and the time series were downsampled to 50 Hz. Units for all eye fixations were converted from pixel coordinates on the display screen to degrees of visual angle.
Fixation performance for each subject and run was summarized by computing the area of the 95% probability contour (in units of degrees of visual angle squared) encircling the distribution of eye positions.
Data Availability
The raw fMRI recordings from both the functional localizer and retinotopy experiments are available via OpenNeuro at https://openneuro.org/datasets/ds006472.
Code Availability
The code used to preprocess the raw fMRI timeseries and fit pRFs to the resultant timecourses is available via Figshare at https://doi.org/10.6084/m9.figshare.29582888.v1.
Supplementary Figures

Schematic diagram of retinotopic mapping runs.
Diagram is for illustration purposes only and not drawn to scale. All original faces have been replaced with AI-generated faces to preserve identifying information of original people. Runs began with a 4 second full-field exposure followed by eight sweeps of bar stimuli in eight different directions and ending with a single clockwise rotation of a wedge. Participants fixated a central dot throughout each run and indicated via button press whenever the dot changed color to red. The carrier images shown inside each aperture changed every 250 msecs (5 Hz), alternating between natural outdoor scenes and collages of various visual objects on backgrounds of pink noise (shown). See Methods for more details.

Polar angle plots displayed on inflated cortical surfaces for all control participants.

Polar angle plots displayed on inflated cortical surfaces for all DP participants.

Two dimensional density plots (heatmaps) of eye position recordings by subject (controls) and run.

Two dimensional density plots (heatmaps) of eye position recordings by subject (DPs) and run.

Eye fixation performance during the retinotopic mapping experiment for controls and DPs.
Eye fixation performance was summarized by computing the area of the 95% probability contour of eye position samples. Points are averages across all runs obtained. Diamonds are averages for each group (Controls, DPs).

Face-selectivity was defined as the percent signal change to blocks of face stimuli minus the percent signal change to blocks of object stimuli.
To quantitatively compare face-selectivity in DPs and controls, we used the variable window method (Norman-Haignere et al., 2013) whereby voxels were selected for the analysis by applying group-defined region of interest (ROI) masks and choosing the top 20% of the voxels with the highest t value for the faces-minus-objects contrast. Voxel selection was fully cross-validated in a leave-one-run-out fashion – voxels were selected based on three out of four runs and face selectivity was measured from the left out run. For each participant, the final measure of face-selectivity was the average across four cross-validated folds. For each ROI and hemisphere, separate Welche two sample t-tests were conducted comparing data from DPs to controls. Due to the number of tests conducted, a more conservative alpha threshold of p < 0.01 was used to establish statistical significance. * p < 0.01; ** p < 0.001

Fixed effect parameter estimates for models evaluating pRF model goodness-of-fit, R2, by hemisphere (right, left) and group (control, DP).
Separate models were created for each region of interest. In each model, the intercept has been mapped to the control group in the left hemisphere. Formula (R, lme4 package): R2 ∼ group + hemisphere + group:hemisphere + (1|subjectID/hemisphere) * p < 0.01; ** p < 0.001
Acknowledgements
We are truly grateful to all our research participants, whose dedication and enthusiasm was essential to this work. We extend our heartfelt thanks to Logan Dowdle for his generous support in developing and testing the fMRI pulse sequences used. We also deeply appreciate the thoughtful guidance and encouragement provided by Kalanit Grill-Spector, Edward Silson, and Ivan Alvarez throughout various stages of the project. This work was made possible by the generous support of the National Institutes of Health (NIH) under grants 1R01EY030613-01A1 to Bradley Duchaine and R01EY034118 to Kendrick Kay.
Additional information
Funding
NIH (1R01EY030613-01A1)
NIH (R01EY034118)
References
- Comparing different stimulus configurations for population receptive field mapping in human fMRIFrontiers in Human Neuroscience 9:1–16Google Scholar
- Functional MRI Reveals Compromised Neural Integrity of the Face Processing Network in Congenital ProsopagnosiaCurrent Biology 19:1146–1150Google Scholar
- Spatial Integration in Normal Face Processing and Its Breakdown in Congenital ProsopagnosiaAnnual Review of Vision Science, pages :1–21Google Scholar
- Impaired holistic processing in congenital prosopagnosiaNeuropsychologia 49:2541–2552Google Scholar
- The definition and diagnosis of developmental prosopagnosiaThe Quarterly Journal of Experimental Psychology 70:193–200Google Scholar
- Fitting Linear Mixed-Effects Models using lme4Journal of Statistical Software 67:1–48Google Scholar
- Configural information in gender categorisationPerception 35:531–540Google Scholar
- Dissociations between face identity and face expression processing in developmental prosopagnosiaCognition 238:105469Google Scholar
- The Human Connectome Project 7 Tesla retinotopy dataset: Description and population receptive field analysisJournal of Vision 18:1–22Google Scholar
- Normal composite face effects in developmental prosopagnosiaCortex 95:63–76Google Scholar
- A purely confirmatory replication study of structural brain-behavior correlationsCortex 66:115–133Google Scholar
- Visual field maps and stimulus selectivity in human ventral occipital cortexNature Neuroscience 8:1102–1109Google Scholar
- Power failure: Why small sample size undermines the reliability of neuroscienceNature Reviews Neuroscience 14:365–376Google Scholar
- Configural information in facial expression perceptionJournal of Experimental Psychology: Human Perception and Performance 26:527–551Google Scholar
- Normal facial age and gender perception in developmental prosopagnosiaCognitive Neuropsychology 29:482–502Google Scholar
- AFNI : Software for Analysis and Visualization of Functional Magnetic Resonance NeuroimagesComputets and Biomedical Research 29Google Scholar
- Face gender recognition in developmental prosopagnosia : Evidence for holistic processing and use of configural informationVisual Cognition 20Google Scholar
- What is the prevalence of developmental prosopagnosia? An empirical assessment of different diagnostic cutoffsCortex 161:51–64Google Scholar
- Mapping striate and extrastriate visual areas in human cerebral cortexProceedings of the National Academy of Sciences of the United States of America 93:2382–2386Google Scholar
- Congenital prosopagnosia: Multistage anatomical and functional deficits in face processing circuitryJournal of Neurology 258:770–782Google Scholar
- Developmental Prosopagnosia: Cognitive, Neural, and Developmental InvestigationsIn:
- Rhodes G.
- Haxby J
- Dissociations of face and object recognition in developmental prosopagnosiaJournal of Cognitive Neuroscience 17:249–261Google Scholar
- The Cambridge Face Memory Test: Results for neurologically intact individuals and an investigation of its validity using inverted face stimuli and prosopagnosic participantsNeuropsychologia 44:576–585Google Scholar
- Population receptive field estimates in human visual cortexNeuroImage 39:647–660Google Scholar
- The development of facial emotion recognition: The role of configural informationJournal of Experimental Child Psychology 97:14–27Google Scholar
- Retinotopic Organization in Human Visual Cortex and the Spatial Precision of Functional MRICerebral Cortex 7:181–192Google Scholar
- A cortical representation of the local visual environmentNature 392:598–601Google Scholar
- The impact of certain methodological choices on multivariate analysis of fMRI data with support vector machinesNeuroImage 54:1159–1167Google Scholar
- Differential spatial computations in ventral and lateral face-selective regions are scaffolded by structural connectionsNature Communications 12:1–14Google Scholar
- Measuring the thickness of the human cerebral cortex from magnetic resonance imagesPNAS 2000:11050–11055Google Scholar
- Fusiform gyrus face selectivity relates to individual differences in facial recognition abilityJournal of Cognitive Neuroscience 23:1723–1740Google Scholar
- Left hemisphere abnormalities in developmental prosopagnosia when looking at faces but not wordsBrain Communications 1Google Scholar
- Development differentially sculpts receptive fields across early and high-level human visual cortexNature Communications 9:1–12Google Scholar
- Supplemental Material for A Quantitative Meta-Analysis of Face Recognition Deficits in Autism: 40 Years of ResearchPsychological Bulletin 147:268–292Google Scholar
- The functional neuroanatomy of face perception: From brain measurements to deep neural networksInterface Focus 8Google Scholar
- Visuospatial coding as ubiquitous scaffolding for human cognitionTrends in Cognitive Sciences 26:81–96Google Scholar
- Spatial maps in frontal and prefrontal cortexNeuroImage 29:567–577Google Scholar
- Frequency specific spatial interactions in human electrocorticography: V1 alpha oscillations reflect surround suppressionNeuroImage 65:424–432Google Scholar
- Receptive Fields and Functional Architecture in Two Nonstriate Visual Areas (18 and 19) of the CatJournal of neurophysiology 28:229–289Google Scholar
- How reliable are gray matter disruptions in specific reading disability across multiple countries and languages? Insights from a large-scale voxel-based morphometry studyHuman Brain Mapping 36:1741–1754Google Scholar
- Developmental prosopagnosics have widespread selectivity reductions across category-selective visual cortexProceedings of the National Academy of Sciences of the United States of America 115:E6418–E6427Google Scholar
- Object Vision in a Structured WorldTrends Cogn Sci 23:672–685Google Scholar
- The Subthreshold Autism Trait Questionnaire (SATQ): Development of a brief self-report measure of subthreshold autism traitsJournal of Autism and Developmental Disorders 42:769–780Google Scholar
- Attention Reduces Spatial Uncertainty in Human Ventral Temporal Cortex Attention Reduces Spatial Uncertainty in Human Ventral Temporal CortexCurrent Biology 25:595–600Google Scholar
- Compressive spatial summation in human visual cortexJ Neurophysiol 94305:481–494Google Scholar
- A Two-Stage Cascade Model of BOLD Responses in Human Visual CortexPLoS Computational Biology 9Google Scholar
- HEVA-A new basic visual processing testJournal of Vision 22:4109–4109Google Scholar
- Persistent prosopagnosia following COVID-19Cortex 162:56–64Google Scholar
- Matching Categorical Object Representations in Inferior Temporal Cortex of Man and MonkeyNeuron 60:1126–1141Google Scholar
- LmerTest Package: Tests in Linear Mixed Effects ModelsJournal of Statistical Software 82:1–26Google Scholar
- The field of view available to the ventral occipito-temporal reading circuitryJournal of Vision 17:1–19Google Scholar
- What aspects of face processing are impaired in developmental prosopagnosia?Brain and Cognition 61:139–158Google Scholar
- Prosopagnosia: A defect in visual configural processingBrain and Cognition 10:149–170Google Scholar
- The first step for neuroimaging data analysis : DICOM to NIfTI conversionJournal of Neuroscience Methods 264:47–56Google Scholar
- Characteristic fixation biases in Super-RecognizersJournal of Vision 22Google Scholar
- The discrimination of facial sex in developmental prosopagnosiaScientific Reports 9:1–8Google Scholar
- The many faces of configural processingTrends in cognitive sciences 6:255–260Google Scholar
- Developmental Prosopagnosia. A Single Case ReportCortex 12:76–82Google Scholar
- Holistic processing for faces operates over a wide range of sizes but is strongest at identification rather than conversational distancesVision Research 49:268–283Google Scholar
- Face ethnicity and measurement reliability affect face recognition performance in developmental prosopagnosia: Evidence from the Cambridge face memory test-AustralianCognitive Neuropsychology 28:109–146Google Scholar
- Why does picture-plane inversion sometimes dissociate perception of features and spacing in faces, and sometimes not? toward a new theory of holistic processingPsychonomic Bulletin and Review 16:778–797Google Scholar
- Shared spatial selectivity in early visual cortex and face-selective brain regionsbioRxiv Google Scholar
- High-resolution retinotopic maps estimated with magnetoencephalographyNeuroImage 145Google Scholar
- Cortical pitch regions in humans respond primarily to resolved harmonics and are located in specific tonotopic regions of anterior auditory cortexJournal of Neuroscience 33:19451–19469Google Scholar
- Impaired holistic coding of facial expression and facial identity in congenital prosopagnosiaNeuropsychologia 49:1226–1235Google Scholar
- The role of the occipital face area in the cortical face perception networkExperimental Brain Research 209:481–493Google Scholar
- Scanning the horizon: Towards transparent and reproducible neuroimaging researchNature Reviews Neuroscience 18:115–126Google Scholar
- Holistic face recognition is an emergent phenomenon of spatial processing in face-selective regionsNature Communications 12:1–13Google Scholar
- On multi-level modeling of data from repeated measures designs: A tutorialSpeech Communication 43:103–121Google Scholar
- Neuroanatomy of developmental dyslexia: Pitfalls and promiseNeuroscience and Biobehavioral Reviews 84Google Scholar
- Quality control practices in FMRI analysis: Philosophy, methods and examples using AFNIFrontiers in Neuroscience 16:1–36Google Scholar
- Distinguishing the cause and consequence of face inversion: The perceptual field hypothesisActa Psychologica 132:300–312Google Scholar
- Borders of multiple visual areas in humans revealed by functional magnetic reson2ance imagingScience 268:889–893Google Scholar
- The 20-item prosopagnosia index (PI20): A self-report instrument for identifying developmental prosopagnosiaRoyal Society Open Science 2Google Scholar
- A Retinotopic Basis for the Division of High-Level Scene Processing between Lateral and Ventral Human Occipitotemporal CortexJ.Neurosci 35:11921–11935Google Scholar
- Topographic maps in human frontal and parietal cortexTrends in Cognitive Sciences 13:488–495Google Scholar
- Neural Population Tuning Links Visual Cortical Anatomy to Human Visual PerceptionNeuron 85:641–656Google Scholar
- A retinotopic code structures the interaction between perception and memory systemsNature Neuroscience 27:339–347Google Scholar
- Advances in developmental prosopagnosia researchCurrent Opinion in Neurobiology 23:423–429Google Scholar
- Face recognition impairments despite normal holistic processing and face space coding: Evidence from a case of developmental prosopagnosiaCognitive Neuropsychology 27:636–664Google Scholar
- Visual Organization of the Default NetworkCerebral Cortex 30:3518–3527Google Scholar
- Empirical assessment of published effect sizes and power in the recent cognitive neuroscience and psychology literaturePLoS Biology 15:1–18Google Scholar
- Parts and wholes in face recognitionThe Quarterly Journal of Experimental Psychology Section A 46:225–245Google Scholar
- Holistic face perception is impaired in developmental prosopagnosiaCortex 108:112–126Google Scholar
- Holistic processing of facial identity in developmental prosopagnosiaCortex 130:318–326Google Scholar
- Small sample sizes reduce the replicability of task-based fMRI studiesCommunications Biology 1Google Scholar
- Topographic Maps of Visual Space in the Human CerebellumCurrent Biology 29:1689–1694Google Scholar
- Impaired holistic and analytic face processing in cong prosop Evidence from eye contingent mask windowVisual Cognition 22:503–521Google Scholar
- Computational neuroimaging and population receptive fieldsTrends in Cognitive Sciences 19:349–357Google Scholar
- A three-dimensional gap filling method for large geophysical datasets: Application to global satellite soil moisture observationsEnvironmental Modelling and Software 30:139–142Google Scholar
- Sparsely-distributed organization of face and limb activations in human ventral temporal cortexNeuroImage 52:1559–1573Google Scholar
- Asynchronous broadband signals are the principal source of the bold response in human visual cortexCurrent Biology 23:1145–1153Google Scholar
- Identification of the ventral occipital visual field maps in the human brainF1000Research Google Scholar
- Reduced spatial integration in the ventral visual cortex underlies face recognition deficits in developmental prosopagnosiabioRxiv Google Scholar
- Looking at Upside-Down FacesJournal of Experimental Psychology 81:141–145Google Scholar
- A critical role of holistic processing in face gender perceptionFrontiers in Human Neuroscience 8:1–10Google Scholar
- Configural information in face perceptionPerception 16:747–759Google Scholar
Article and author information
Author information
Version history
- Preprint posted:
- Sent for peer review:
- Reviewed Preprint version 1:
Cite all versions
You can cite all versions using the DOI https://doi.org/10.7554/eLife.108657. This DOI represents all versions, and will always resolve to the latest one.
Copyright
© 2025, Stehr et al.
This article is distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use and redistribution provided that the original author and source are credited.
Metrics
- views
- 0
- downloads
- 0
- citations
- 0
Views, downloads and citations are aggregated across all versions of this paper published by eLife.