Graphic illustration of the horizontal smearing occurring when averaging multiple views of a face.

The tolerance of human face identity recognition to drastic appearance variations caused by varying lighting, viewpoint, facial expression, etc. has been proposed to emerge thanks to an averaging mechanism that would progressively whiten these accidental variations and preserve the stable identity cues to identity. Past illustrations (Burton, 2013) used varying lighting and expressions but moderate pose variations. Here we show that when pose varies from left to right profile, averaging results in horizontally smeared face cues suggesting that across encounters with a face, cues at most orientations except horizontal are whitened. As the observer learns the natural statistics of a person’s face, it seems plausible that they increasingly rely on horizontal cues for identity recognition (see Pachai et al., 2017 for supporting evidence). Images of two celebrities (George Clooney and Daniel Ratclife) were sampled from the internet and sorted into three view categories: frontal, left-, and right-averted. In order to illustrate the horizontal smearing due to view variations, the averages were made of 40% left-averted, 40% right-averted, and 20% frontal views. The luminance and RMS contrast of the averaged faces were set to a luminance of .5 and contrast of .4. Using this procedure, one can appreciate the emergence of the so-called bar code, namely the vertical arrangement of horizontally-oriented cues typical of the face at basic and individual levels of categorization (Dakin & Watt, 2009).

Stimulus conditions.

Columns. Each identity was viewed from seven different viewpoints ranging from +75° to −75° in steps of 25°. Rows. All images were filtered in the Fourier domain to preserve only a selective range of orientation, from 0° (vertical) to 157.5° in steps of 22.5°.

A. Sensitivity of human observers to facial identity (d’) as a function of the orientation filter (0° to 180° in 22.5° steps) and face viewpoint (yaw: +75° to −75° in 25° steps). Dots and error bars represent mean d’ values and 95% confidence intervals across participants. Solid lines and shaded areas indicate the mean posterior predictions and 95% credible intervals from the Gaussian Bayesian multilevel model. B. Population-level mean parameters of the Gaussian Bayesian Multilevel model, plotted with 95% credible intervals as a function of face viewpoint. The 95% credible intervals reflect the uncertainty of the model and can be interpreted as follows: there is a 95% chance that the value of the population parameter lies within this interval.

Posterior mean and 95% credible interval for each parameter of the Gaussian model, at each viewpoint.

Sensitivity (i.e., performance at the recognition task) of human and model observers and image energy across viewpoints and orientations.

Left column. 3D surf plots of the normalized energy/sensitivy across orientation and viewpoints. Middle column. Matrix representations of the normalized energy/sensitivity across orientation and viewpoints. Right column. Matrix representations of the Pearson correlation (non Fisher-z transformed) of the normalized orientation distributions of energy/sensitivity across viewpoints.

Mean and 95% confidence interval of the (Fisher Z-transformed) partial correlation coefficients between human and each model orientation d’ profiles while controlling for the variance in image energy and alternate model.

Fisher-z transformed Pearson partial correlation of the orientation sensitivity profiles between humans and each model, while controlling for the alternate model and image energy.

Error bars show the 95% confidence intervals. The fade grey line depicts the maximally achievable correlation for the separate viewpoint conditions in the human dataset (see Methods for details).

Difference of the human-model correlation coefficients between viewpoint-selective and view-tolerant model observers.

Positive t values indicate a stronger correlation with the viewpoint-selective model and negative t values stronger correlation with the view-tolerant model.