Attention modulates human visual responses to objects by tuning sharpening

  1. Narges Doostani  Is a corresponding author
  2. Gholam-Ali Hossein-Zadeh
  3. Radoslaw M Cichy
  4. Maryam Vaziri-Pashkam  Is a corresponding author
  1. School of Cognitive Sciences, Institute for Research in Fundamental Sciences (IPM), Iran
  2. Department of Education and Psychology, Freie Universität Berlin, Germany
  3. School of Electrical and Computer Engineering, College of Engineering, University of Tehran, Iran
  4. Department of Psychological and Brain Sciences, University of Delaware, United States
  5. Laboratory of Brain and Cognition, National Institute of Mental Health, United States
13 figures, 5 tables and 1 additional file

Figures

Figure 1 with 1 supplement
Stimuli, paradigm, and regions of interest.

(A) Top images represent the four categories used in the main experiment: body, car, house, and cat. The stimulus set consisted of 10 exemplars from each category (here: cats), with exemplars differing in pose and 3D-orientation. (B) The experimental design comprised 16 task conditions (12 paired, four isolated). The 4×4 matrix on the left illustrates the 12 paired conditions, with the to-be-attended category (outlined in orange for illustration purposes, not present in the experiment) on the y-axis and the to-be-ignored category on the x-axis. The right column illustrates the four isolated conditions. (C) Experimental paradigm. A paired block is depicted with superimposed body and house stimuli. In this example block, house stimuli were cued as targets, and the participant responded on the repetition of the exact same house in two consecutive trials, as marked here by the arrow. (D) Regions of interest for an example participant; the primary visual cortex (V1), the object-selective regions lateral occipital cortex (LO) and posterior fusiform gyrus (pFs), the body-selective region extrastriate body area (EBA), and the scene-selective region parahippocampal place area (PPA).

Figure 1—figure supplement 1
Behavioral performance of the participants during the recording.

B, Cr, H, and Ct represent the Body, Car, House, and Cat categories, respectively. Each x-axis label represents the two conditions with stimuli from the same two categories but with different attentional targets, with the dark gray bar illustrating the results related to the condition in which the first category was attended, and the light gray bar representing the results in the condition in which the second category was attended. For instance, the B-Cr bars represent the results related to the BodyatCar and BodyCarat conditions, respectively. Error bars represent standard errors of the mean. N=15 human participants. (A) Average detection rate for each category pair. (B) Average reaction time for each category pair.

Response vectors related to x and y stimuli in isolated and paired conditions.

(A) Example illustration of the isolated and paired responses in three-dimensional space. Vxat and Vyat denote the response vectors related to isolated x and isolated y conditions and Vxaty illustrates the response in the paired condition with attention directed to stimulus x. The paired response is projected on the plane defined by the two isolated responses Vxat and Vyat. This projection is illustrated by Pxaty. (B) Two-dimensional illustration of the plane defined by the two isolated response vectors Vxat and Vyat, along with the paired response vectors Vxat and Vyat and Pxaty as the projection of the paired response, Vxaty, on the plane. We calculated the weight of the two isolated responses in the paired response using multiple regression, with the weights of Vxat and Vyat shown as a1 and a2, respectively.

Attentional modulation by the response gain and tuning sharpening models.

We illustrate the models here for the example of a neuron with high selectivity for cat stimuli. Solid curves denote the response to unattended stimuli and dashed curves denote the response to attended stimuli. (A) According to the response gain model, the response of the neuron to attended stimuli is scaled by a constant attention factor. Therefore, the response of the cat-selective neuron to an attended stimulus is enhanced to the same degree for all stimuli. (B) According to the tuning sharpening model, the response modulation by attention depends on the neuron’s tuning for the attended stimulus. Therefore, for optimal and near-optimal stimuli such as cat and body stimuli the response is highly increased, while for non-optimal stimuli such as houses, the response is suppressed.

Figure 4 with 4 supplements
Average voxel response in the extastriate body area (EBA) for each pair of stimulus categories.

The x-axis labels represent the four conditions related to each category pair, Mat, MatL, MLat, Lat, with M and L denoting the presence of the more preferred and the less preferred category and the superscript at denoting the attended category. For instance, Mat refers to the condition in which the more preferred stimulus was presented in isolation (and automatically attended), and MLat refers to the paired condition in which the less preferred stimulus was attended to. Red arrows in each panel illustrate the observed change in response (univariate shift) caused by the shift of attention from the more preferred to the less preferred stimulus. Green arrows in panels B and C illustrate the difference in the response to isolated stimuli. Error bars represent standard errors of the mean. N=17 human participants.

Figure 4—figure supplement 1
Average voxel response in the primary visual cortex (V1) for each pair of stimulus categories.

The x-axis labels represent the four conditions related to each category pair, Mat, MatL, MLat, Lat, with M, and L denoting the presence of the more preferred and the less preferred category and the superscript at denoting the attended category. For instance, Mat refers to the condition in which the more preferred stimulus was presented in isolation (and automatically attended), and MLat refers to the paired condition in which the less preferred stimulus was attended to.

Figure 4—figure supplement 2
Average voxel response in the lateral occipital cortex (LO) for each pair of stimulus categories.

The x-axis labels represent the four conditions related to each category pair, Mat, MatL, MLat, Lat, with M, and L denoting the presence of the more preferred and the less preferred category and the superscript at denoting the attended category.

Figure 4—figure supplement 3
Average voxel response in the posterior fusiform gyrus (pFs) for each pair of stimulus categories.

The x-axis labels represent the four conditions related to each category pair, Mat, MatL, MLat, Lat, with M, and L denoting the presence of the more preferred and the less preferred category and the superscript at denoting the attended category.

Figure 4—figure supplement 4
Average voxel response in the parahippocampal place area (PPA) for each pair of stimulus categories.

The x-axis labels represent the four conditions related to each category pair, Mat, MatL, MLat, Lat, with M, and L denoting the presence of the more preferred and the less preferred category and the superscript at denoting the attended category.

Univariate shift versus category distance in each region of interest (ROI).

(A–E) The value of univariate shift versus category distance. MatL and MLat denote the two paired conditions with attention directed to the more preferred (M) or less preferred (L) stimulus, respectively. Mat and Lat represent the isolated conditions, respectively, with the more preferred or the less preferred stimulus presented in isolation. The blue, red, yellow, purple, green, and sky blue circles in each panel represent the values related to the Body-Car, Body-House, Body-Cat, Car-House, Car-Cat, and House-Cat pairs, respectively. The correlation coefficients in each ROI were calculated for single participants and the lines in the average plots are only shown for illustration purposes. (F) Correlation coefficient for the correlation between the univariate shift and category distance in each ROI. Asterisks indicate that the correlation coefficients are significantly positive (p<0.05). Error bars represent standard errors of the mean. N=17 human participants.

Figure 6 with 1 supplement
Weight shift versus category distance in each region of interest (ROI).

(A–E) Attentional weight shift versus category distance. The blue, red, yellow, purple, green, and sky blue circles in each panel represent the values related to the Body-Car, Body-House, Body-Cat, Car-House, Car-Cat, and House-Cat pairs, respectively. We calculated the correlation coefficients for single participants and the lines in these average plots are only for illustration purposes. (F) Correlation coefficient for the correlation between attentional weight shift and category distance in each ROI. Asterisks indicate that the correlations are significantly positive (p<0.05). Error bars represent standard errors of the mean. N=17 human participants.

Figure 6—figure supplement 1
Sum of weights in the multivariate analysis for each category pair in each region of interest (ROI), averaged across conditions with attention directed to each of the two categories of a pair and across participants.

For instance, the blue Body-Car bars represent the average results related to the BodyatCar and BodyCarat conditions, averaged across participants. Error bars represent standard errors of the mean. N=17 human participants.

Univariate shift as a function of category distance, as predicted by the two attentional mechanisms.

MatL and MLat denote the two paired conditions with attention directed to the more preferred (M) or the less preferred (L) stimulus, respectively. Mat and Lat represent the isolated conditions, respectively with the M or the L stimulus presented in isolation. Top panels represent predictions in a region with a strong preference for a specific category, and bottom panels illustrate predictions in an object-selective region. Each circle represents a pair of categories. (A–B) Predicted univariate shift based on the response gain model in a region with a strong preference for a specific category (A) and in an object-selective region (B). (C–D) Predicted univariate shift based on the tuning model in a region with strong preference for a specific category (C) and in an object-selective region (D).

Predicted weight shift as a function of category distance.

Weight shift for each pair is calculated using Equation 5. Category distance represents the difference in multi-voxel representation between responses to the two isolated stimuli, calculated by Equation 3. Top panels are related to predictions in a region with a strong preference for a specific category and the bottom panels illustrate predictions in an object-selective region. (A–B) Weight shift predicted by the response gain model in a region with a strong preference for a specific category (A) and in an object-selective region (B). (C–D) Weight shift predicted by the tuning model in a region with strong preference for a specific category (C) and in an object-selective region (D).

Appendix 1—figure 1
Based on the labeled line model, attention enhances the response of a neuron only when the attended stimulus is the neuron’s preferred stimulus.

We illustrate the models here for the example of a neuron with high selectivity for cat stimuli. Solid curves denote the response to unattended stimuli and dashed curves denote the response to attended stimuli.

Appendix 1—figure 2
Simulation results for the labeled line model.

(A–B) Predicted univariate shift based on the labeled line model in a region with a strong preference for a specific category (A) and in an object-selective region (B). (C–D) Weight shift predicted by the labeled line model in a region with a strong preference for a specific category (C) and in an object-selective region (D).

Author response image 1
Range of category distance in each ROI averaged across participants.
Author response image 2
Univariate shift versus category distance including single-subject data points in all ROIs.
Author response image 3
Weight shift versus category distance including single-subject data points in all ROIs.

Tables

Appendix 1—table 1
Percentage of voxels in each regions of interest (ROI) most responsive to each of the four stimulus categories, averaged across participants.
Preferred CategoryV1LOpFsEBAPPA
Body21%31%21%70%10%
Car19%18%17%4%12%
House33%32%40%3%68%
Cat27%19%22%23%10%
Appendix 1—table 2
Average voxel response (general linear model, GLM coefficients) to each category in each region of interest (ROI), averaged across participants.
Stimulus CategoryV1LOpFsEBAPPA
Body0.281.590.881.620.001
Car0.281.430.880.600.12
House0.471.511.030.480.55
Cat0.351.490.891.32–0.02
Appendix 1—table 3
Single-subject correlation coefficients for the correlation between the univariate shift and category distance in all regions of interest (ROIs).
V1LOpFsEBAPPA
–0.600.67–0.0730.720.18
0.160.380.0150.640.46
–0.530.520.890.600.42
–0.0620.780.930.870.69
–0.0980.0721.00.720.96
0.630.930.880.940.58
–0.310.52–0.110.880.89
0.290.180.620.670.63
–0.490.0940.630.810.80
0.210.570.310.810.87
–0.015–0.570.790.350.69
0.350.790.900.900.88
–0.470.058–0.0500.93–0.24
–0.047–0.210.910.280.86
0.660.700.910.640.45
–0.740.540.240.60–0.31
–0.0950.730.890.640.95
Appendix 1—table 4
Weights of each stimulus for each category pair in each region of interest (ROI), averaged across participants.
PairAttended StimulusV1LOpFsEBAPPA
Body-CarBody0.54×RBat+0.53×RCrat0.79×RBat+0.31×RCrat0.68×RBat+0.42×RCrat0.93×RBat+0.15×RCrat0.41×RBat+0.42×RCrat
Body-CarCar0.40×RBat+0.70×RCrat0.43×RBat+0.68×RCrat0.39×RBat+0.70×RCrat0.53×RBat+0.46×RCrat0.35×RBat+0.54×RCrat
Body-HouseBody0.43×RBat+0.65×RHat0.65×RBat+0.45×RHat0.65×RBat+0.47×RHat0.84×RBat+0.21×RHat0.42×RBat+0.61×RHat
Body-HouseHouse0.27×RBat+0.78×RHat0.3×RBat+0.74×RHat0.27×RBat+0.81×RHat0.37×RBat+0.57×RHat0.18×RBat+0.86×RHat
Body-CatBody0.63×RBat+0.55×RCtat0.86×RBat+0.35×RCtat0.81×RBat+0.45×RCtat0.90×RBat+0.30×RCtat0.61×RBat+0.32×RCtat
Body-CatCat0.59×RBat+0.55×RCtat0.69×RBat+0.48×RCtat0.67×RBat+0.55×RCtat0.67×RBat+0.48×RCtat0.55×RBat+0.37×RCtat
Car-HouseCar0.42×RCrat+0.65×RHat0.44×RCrat+0.70×RHat0.55×RCrat+0.60×RHat0.56×RCrat+0.59×RHat0.35×RCrat+0.72×RHat
Car-HouseHouse0.31×RCrat+0.73×RHat0.25×RCrat+0.85×RHat0.36×RCrat+0.78×RHat0.46×RCrat+0.65×RHat0.27×RCrat+0.93×RHat
Car-CatCar0.59×RCrat+0.52×RCtat0.68×RCrat+0.44×RCtat0.74×RCrat+0.40×RCtat0.47×RCrat+0.54×RCtat0.53×RCrat+0.40×RCtat
Car-CatCat0.45×RCrat+0.70×RCtat0.41×RCrat+0.75×RCtat0.55×RCrat+0.60×RCtat0.23×RCrat+0.92×RCtat0.46×RCrat+0.46×RCtat
House-CatHouse0.76×RHat+0.32×RCtat0.87×RHat+0.24×RCtat0.91×RHat+0.23×RCtat0.56×RHat+0.44×RCtat0.99×RHat+0.11×RCtat
House-CatCat0.61×RHat+0.47×RCtat0.47×RHat+0.65×RCtat0.53×RHat+0.62×RCtat0.17×RHat+0.90×RCtat0.57×RHat+0.38×RCtat
Appendix 1—table 5
Single-subject correlation coefficients for the correlation between weight shift and category distance in all regions of interests (ROIs).
V1LOpFsEBAPPA
–0.360.93–0.250.460.06
0.31–0.060.160.82–0.25
–0.290.350.800.48–0.42
–0.0380.700.750.570.38
0.720.600.900.58–0.77
0.530.690.760.510.59
–0.320.90–0.250.290.065
0.330.22–0.360.24–0.48
0.520.310.630.540.91
0.0190.960.800.610.71
–0.110.870.890.690.51
–0.270.940.810.630.50
–0.280.28–0.110.490.59
0.420.110.610.420.59
0.670.770.780.360.36
–0.960.780.480.43–0.24
–0.120.690.820.630.56

Additional files

Download links

A two-part list of links to download the article, or parts of the article, in various formats.

Downloads (link to download the article as PDF)

Open citations (links to open the citations from this article in various online reference manager services)

Cite this article (links to download the citations from this article in formats compatible with various reference manager tools)

  1. Narges Doostani
  2. Gholam-Ali Hossein-Zadeh
  3. Radoslaw M Cichy
  4. Maryam Vaziri-Pashkam
(2024)
Attention modulates human visual responses to objects by tuning sharpening
eLife 12:RP89836.
https://doi.org/10.7554/eLife.89836.3