A series of experiments at EPFL provide conclusive evidence that the brain uses a single mechanism (supramodality) to estimate confidence in different senses such as audition, touch, or vision. The study is published in the Journal of Neuroscience.



Behavioral scientists and psychologists use the term “metacognition” to describe our ability to access, report and regulate our own mental states: “thinking about thinking”, “knowing about knowing” “being aware about being aware”, are all higher-order cognitive skills that fit this category.

Specifically, metacognition enables the brain to compute a degree of confidence when we perceive events from the external world, such as a sound, light, or touch. The accuracy of confidence estimates is crucial in daily life, for instance when hearing a baby crying, or smelling a gas leak. Confidence estimates also need to combine input from multiple senses simultaneously, for instance when buying a violin based on how it sounds, feels, and looks.

From a neuroscience point of view, the way metacognition operates in different senses, and for combination of senses is still a mystery: Does metacognition use the same rules for visual, auditory, or tactile stimuli, or does it use different components of each of sensory domains? The first of these two ideas – i.e. the “common rules” – is known as “supramodality” and it has proven controversial among neuroscientists.

Settling the matter

A series of experiments by Olaf Blanke’s lab at EPFL now provide evidence in favor of supramodality. The study, led by researcher Nathan Faivre, tested human volunteers using three different types of experimental techniques: behavioral psychophysics, computational modeling, and electrophysiological recordings.

The behavioral part of the study found that participants with high metacognitive performance for one sense (e.g. vision) were likely to perform well in other senses (e.g. audition or touch). “In other words,” explains Faivre, “those of us who are good at knowing what they see are also good at knowing what they hear and what they touch.”

The computational modeling indicated that the confidence estimates we build when seeing an image or hearing a sound can be efficiently compared to one another. This implies that they share the same format.

Finally, the electrophysiological recordings revealed similar characteristics when the volunteers reported confidence in their responses to audio or audiovisual stimuli. This suggests that visual and audiovisual metacognition is based on similar neural mechanisms.

“These results make a strong case in favor of the supramodality hypothesis,” says Faivre. “They show that there is a common currency for confidence in different sensory domains – in other words, that confidence in a signal is encoded with the same format in the brain no matter where the signal comes from. This gives metacognition a central status, whereby the monitoring of perceptual processes occurs through a common neural mechanism.”

The study is an important step towards a mechanistic understanding of human metacognition. It tells us something about how we perceive the world and become aware of our surroundings, and can potentially lead to ways of treating several neurological and psychiatric disorders where metacognition is impaired.

Contributing institutions

CNRS

Max Plank Institute for Human development

Humboldt-Universität zu Berlin

Bernstein Center for Computational Neuroscience Berlin

Universidad de Buenos Aires

University Medical Center Hamburg-Eppendorf

University Hospital Geneva

Funding

Bertarelli Foundation,

Swiss National Science Foundation

European Science Foundation

Reference

Nathan Faivre, Elisa Filevich, Guillermo Solovey, Simone Kuhn, Olaf Blanke. Behavioural, modeling, and electrophysiological evidence for domain-generality in human metacognition. Journal of Neuroscience 15 September 2017, 0322-17. DOI: 10.1523/JNEUROSCI.0322-17.2017