Crossref Citations
This article has been cited by the following publications. This list is generated based on data provided by
Crossref.
O’Callaghan, Claire
Kveraga, Kestutis
Shine, James M.
Adams, Reginald B.
and
Bar, Moshe
2017.
Predictions penetrate perception: Converging insights from brain, behaviour and disorder.
Consciousness and Cognition,
Vol. 47,
Issue. ,
p.
63.
González-Casillas, Adrián
Parra, Luis
Martin, Luis
Avila-Contreras, Cynthia
Ramirez-Pedraza, Raymundo
Vargas, Natividad
del Valle-Padilla, Juan Luis
and
Ramos, Félix
2018.
Towards a model of visual recognition based on neurosciences.
Procedia Computer Science,
Vol. 145,
Issue. ,
p.
214.
González-Casillas, Adrían
Parra, Luis
Martin, Luis
Avila-Contreras, Cynthia
Ramirez-Pedraza, Raymundo
Vargas, Natividad
del Valle-Padilla, Juan Luis
and
Ramos, Félix
2018.
Towards a model of visual recognition based on neurosciences.
Biologically Inspired Cognitive Architectures,
Vol. 25,
Issue. ,
p.
119.
Matar, Elie
Phillips, Joseph R.
Martens, Kaylena A. Ehgoetz
Halliday, Glenda M.
and
Lewis, Simon J. G.
2019.
Impaired Color Discrimination—A Specific Marker of Hallucinations in Lewy Body Disorders.
Journal of Geriatric Psychiatry and Neurology,
Vol. 32,
Issue. 5,
p.
257.
Klever, Lena
Mamassian, Pascal
and
Billino, Jutta
2022.
Age-related differences in visual confidence are driven by individual differences in cognitive control capacities.
Scientific Reports,
Vol. 12,
Issue. 1,
Hadders-Algra, Mijna
2022.
Human face and gaze perception is highly context specific and involves bottom-up and top-down neural processing.
Neuroscience & Biobehavioral Reviews,
Vol. 132,
Issue. ,
p.
304.
Oswald, Flora
and
Adams, Reginald B.
2023.
Feminist Social Vision: Seeing Through the Lens of Marginalized Perceivers.
Personality and Social Psychology Review,
Vol. 27,
Issue. 3,
p.
332.
Quesque, François
and
Rossetti, Yves
2023.
Are TPJs conflict detectors or solvers?.
Physics of Life Reviews,
Vol. 44,
Issue. ,
p.
163.
Kveraga, Kestutis
2024.
Modeling Visual Aesthetics, Emotion, and Artistic Style.
p.
71.
Kent, Ernest W.
2024.
The Roles of Representation in Visual Perception.
Vol. 486,
Issue. ,
p.
305.
A neural mechanism for cognitive penetrability in visual perception
In their target article, Firestone & Scholl (F&S) readily dismiss the extensive presence of descending neural pathways (Angelucci et al. Reference Angelucci, Levitt, Walton, Hupe, Bullier and Lund2002; Bullier Reference Bullier2001), claiming they have no necessary implications for cognitive penetrability. Yet it is precisely this architecture of feedforward and feedback projections, beginning in the primary visual cortex, ascending through dorsal or ventral visual pathways, dominated respectively by magnocellular and parvocellular cells (Goodale & Milner Reference Goodale and Milner1992; Ungerleider & Mishkin Reference Ungerleider, Mishkin, Ingle, Goodale and Mansfield1982), and matched with reciprocal feedback connections (Felleman & Van Essen Reference Felleman and Van Essen1991; Salin & Bullier Reference Salin and Bullier1995), that provides the starting point for evidence in favour of top-down effects on visual perception.
Numerous studies capitalised on inherent differences in the speed and content of magnocellular (M) versus parvocellular (P) processing to reveal their role in top-down effects (Panichello et al. Reference Panichello, Cheung and Bar2012). Early work using functional magnetic resonance imaging (fMRI) suggested that formation of top-down expectations, signalled by a gradually increasing ventrotemporal lobe activity, facilitated the recognition of previously unseen objects (Bar et al. Reference Bar, Tootell, Schacter, Greve, Fischl, Mendola, Rosen and Dale2001). In subsequent studies using both intact line drawings and achromatic, low spatial frequency (LSF) stimuli (which preferentially recruit M pathways), early activity was evident in the orbitofrontal cortex (OFC) ~130 ms after stimulus presentation, well before object recognition–related activity peaks in the ventrotemporal cortex (Bar et al. Reference Bar, Kassam, Ghuman, Boshyan, Schmid, Dale, Hämäläinen, Marinkovic, Schacter and Rosen2006). An fMRI study using dynamic causal modelling later confirmed that M-biased stimuli specifically activated a pathway from the occipital cortex to OFC, which then initiated top-down feedback to the fusiform gyrus. This connectivity pattern was different from that evoked by stimuli activating the P pathway, where only feedforward flow increased between the occipital cortex and fusiform gyrus (Kveraga et al. Reference Kveraga, Boshyan and Bar2007a). OFC activity predicted recognition of M, but not P, stimuli, and resulted in faster recognition of M stimuli by ~100 ms. Another fMRI study showed that this OFC facilitation of object recognition was triggered for meaningful LSF images exclusively: Only meaningful images, but not meaningless images (from which predictions could not be generated), revealed increased functional connectivity between the lateral OFC and ventral visual pathway (Chaumon et al. Reference Chaumon, Kveraga, Barrett and Bar2013). We argue not only that these results demonstrate the importance of descending neural pathways, which F&S do not dispute (cf. sect. 2.2), but also that these recurrent connections penetrate bottom-up perception and facilitate perception via feedback of activated information from the OFC.
This top-down activity does not merely reflect recognition or “back-end” memory-based processes, as F&S suggest are commonly conflated with top-down effects. Instead, the rapid onset of OFC activation and subsequent coupling with visual cortical regions indicate that these top-down processes affect perception proper, which we suggest occurs in the form of predictions that constrain the ongoing perceptual process. These predictions categorise ambiguous visual input into a narrow set of most probable alternatives based on all available information. As a richly connected association region, receiving inputs from sensory, visceral and limbic modalities, the OFC is ideally situated to integrate crossmodal information and generate expectations based on previous experience that can be compared with incoming sensory input. Predictive information from the OFC is then back-propagated to inferior temporal regions and integrated with high spatial frequency information. Thus, by constraining the number of possible interpretations, the OFC provides a signal that guides continued, low-level visual processing, resulting in a refined visual percept that is identified faster (Trapp & Bar Reference Trapp and Bar2015).
Another aspect of this top-down guidance process that can penetrate bottom-up visual perceptual processing involves constraints imposed by the stimulus context. In the model that emerged from these data, “gist” information is extracted from LSFs in the visual input, and predictions are generated about the most probable interpretation of the input, given the current context (Bar Reference Bar2004). When bottom-up visual input is ambiguous, the same object can be perceived as a hair dryer or a drill, depending on whether it appears in a bathroom or a workshop context (e.g., Bar Reference Bar2004, Box 1). A network sensitive to contextual information, which also includes the parahippocampal, retrosplenial, and medial orbitofrontal cortices (Aminoff et al. Reference Aminoff, Gronau and Bar2007; Bar & Aminoff Reference Bar and Aminoff2003), has been implicated in computing this context signal. Crucially, this process is not simply influencing better guesswork. Using magnetoencephalography and phase synchrony analyses, these top-down contextual influences are shown to occur during the formation stages of a visual percept, extending all the way back to early visual cortex (Kveraga et al. Reference Kveraga, Ghuman, Kassam, Aminoff, Hämäläinen, Chaumon and Bar2011).
The emerging picture from this work suggests that ongoing visual perception is directly and rapidly influenced by previously learnt information about the world. This is undoubtedly a highly adaptive mechanism, promoting more efficient processing amidst the barrage of complex visual input that our brains receive. In the next section, we extend this model by incorporating ecologically valid examples of how top-down effects on visual perception facilitate complex human interactions, and the ramifications when the delicate balance between prediction and sensory input is lost in clinical disorders.
Cognitive penetrability in broader contexts – visual hallucinations and social vision
Top-down influences on visual perception are also observable in clinical disorders that manifest visual hallucinations, including schizophrenia, psychosis, and Parkinson's disease. Most strikingly, the perceptual content of visual hallucinations can be determined by autobiographical memories; familiar people or animals are a common theme (Barnes & David Reference Barnes and David2001). Frequency and severity of visual hallucinations is exacerbated by mood and physiological states (e.g., stress, depression, and fatigue), with mood also playing an important role in determining the content of hallucinations (e.g., when the image of a deceased spouse is perceived during a period of bereavement) (Waters et al. Reference Waters, Collerton, ffytche, Jardri, Pins, Dudley, Blom, Mosimann, Eperjesi and Ford2014). Such phenomenological enquiry into visual hallucinations suggests their content is influenced in a top-down manner, by stored memories and current emotional state. Anecdotal report is mirrored by experimental confirmation that the psychosis spectrum is associated with overreliance on prior knowledge, or predictive processing, when interpreting ambiguous visual stimuli (Teufel et al. Reference Teufel, Subramaniam, Dobler, Perez, Finnemann, Mehta, Goodyer and Fletcher2015). Together, these mechanisms are consistent with a framework in which top-down influences tend to dominate visual processing in hallucinations. Important for theories of cognitive penetrability, visual hallucinations typically involve a hallucinated object being perceived as embedded within the actual scenery, such that the hallucination is thoroughly integrated with sensory input (Macpherson Reference Macpherson, Zeimbekis and Raftopoulos2015). Existing neural frameworks for visual hallucinations account for an imbalance between bottom-up sensory information and top-down signals. These frameworks implicate overactivity in regions supplying top-down information during normal visual perception, including the medial temporal and prefrontal sites in the model outlined above. Abnormal activity in these regions, and in their connectivity with the visual cortex, plays a causative role in creating the hallucinatory percepts that effectively hijack visual perception (Shine et al. Reference Shine, O'Callaghan, Halliday and Lewis2014). Electrical stimulation studies targeting these regions independently confirm that abnormal activity in temporal lobe and midline areas is capable of generating complex visual hallucinations (Selimbeyoglu & Parvizi Reference Selimbeyoglu and Parvizi2010).
Visual information conveying social cues is some of the subtlest, yet richest, perceptual input we receive – consider the abundance of information delivered in a sidelong glance or a furrowed brow. Top-down influences allowing us to recognise patterns in our social environment and interpret them rapidly are a cornerstone of adaptive social behaviour (de Gelder & Tamietto Reference de Gelder, Tamietto, Adams, Ambady, Nakayama and Shimojo2011). Available evidence suggests that social visual processing leverages precisely the same neural mechanism described above for object and scene recognition. However, because of typically greater ambiguity in social cues, social vision must rely on top-down expectations to an even greater extent than object recognition. Social cues, including eye gaze, gender, culture, and race are found to directly influence the perception and neural response to facial emotion (Adams & Kleck Reference Adams and Kleck2005; Adams et al. Reference Adams, Gordon, Baird, Ambady and Kleck2003; Reference Adams, Hess and Kleck2015) and exert increasing influence with increasing ambiguity in the given expression (Graham & LaBar Reference Graham and LaBar2012). Critically, these effects are also modulated by individual differences such as trait anxiety and progesterone levels in perceptual tasks (Conway et al. Reference Conway, Jones, DeBruine, Welling, Smith, Perrett, Sharp and Al-Dujaili2007; Fox et al. Reference Fox, Mathews, Calder and Yiend2007) and in amygdala response to threat cues (Ewbank et al. Reference Ewbank, Fox and Calder2010). Dovetailing with the model outlined above, fusiform cortex activation is found to track closely with objective gradations between morphed male and female faces, whereas OFC responses track with categorical perceptions of face gender (Freeman et al. Reference Freeman, Rule, Adams and Ambady2010). As in object recognition, OFC may be categorising continuously varying social stimuli into a limited set of alternative interpretations. Social vision therefore provides an important example of cognitive penetrability in visual perception that utilises stored memories and innate templates to makes sense of perceptual input.
Conclusion
The convergent experimental and ecological evidence we have outlined suggests a visual processing system profoundly influenced by top-down effects. The model we describe fits with a “predictive brain” harnessing previous experience to hone sensory perception. In the face of evidence reviewed here, it seems difficult to categorically argue that cognitive penetrability in visual perception is yet to be convincingly demonstrated.