BA or DA? Decoding syllables to show the limits of AI

UNIGE researchers show the limits of the precision of decoding brain activity, via modern machine learning techniques, particularly in processing speech sounds.



The (traditional) univariate analysis of MEG (magnetoencephalography) activity recorded in human subjects when they perform a simple task such as « did you hear BA or DA? », shows that the decision is encoded very early and very focally in the human brain. On the contrary, new decoding techniques of neuronal activity by machine-learning suggest that decisional information is present all over the brain. Building on the temporal precision of the MEG technique, we show that machine-decoding algorithms capture patterns of neuronal activity that follow the decision and, hence, do not reveal how our brain uses information contained in these patterns. © UNIGE


For about the last ten years, researchers have been using artificial intelligence techniques called machine learning to decode human brain activity. Applied to neuroimaging data, these algorithms can reconstitute what we see, hear, and even what we think. For example, they show that words with similar meanings are grouped together in zones in different parts of our brain. However, by recording brain activity during a simple task—whether one hears BA or DA—neuroscientists from the University of Geneva (UNIGE), Switzerland, and the Ecole normale supérieure (ENS) in Paris now show that the brain does not necessarily use the regions of the brain identified by machine learning to perform a task. Above all, these regions reflect the mental associations related to this task. While machine learning is thus effective for decoding mental activity, it is not necessarily effective for understanding the specific information processing mechanisms in the brain. The results are available in the PNAS journal.


Modern neuroscientific data techniques have recently highlighted how the brain spatially organises the portrayal of word sounds, which researchers were able to precisely map by region of activity. UNIGE neuroscientists thus asked how these spatial maps are used by the brain itself when it performs specific tasks. “We have used all the available human neuroimagery techniques to try to answer this question”, says Anne-Lise Giraud, a professor at the Department of Basic Neurosciences of the UNIGE Faculty of Medicine.

A focal region for selecting information

UNIGE neuroscientists had about fifty people listen to a continuum of syllables ranging from BA to DA. The central phonemes were very ambiguous and it was difficult to distinguish between the two options. They then used a functional MRI and magnetoencephalography to see how the brain behaves when the acoustic stimulus is very clear, or, on the contrary, when it is ambiguous and requires an active mental representation of the phoneme and its interpretation by the brain. “We have observed that regardless of how difficult it is to classify the syllable that was heard, between BA and DA, the decision always engages a small region of the posterior superior temporal lobe”, notes Anne-Lise Giraud.

Neuroscientists then double-checked their results on a patient with an injury in the specific region of the posterior superior temporal lobe used to distinguish between BA and DA. “And indeed, although the patient did not appear to have symptoms, he was no longer able to distinguish between the BA and DA phonemes ... this confirms that this small region is important in processing this type of phoneme information”, adds Sophie Bouton, a researcher from Anne-Lise Giraud’s team.

The “false positives” of machine learning decoding

But is the information on the identity of the syllable just locally present, as the experiment of these Genevan scientists has shown, or is it present more generally in our brain, as suggested by the maps produced via machine learning? To answer this question, the neuroscientists reproduced the BA / DA task with people who have electrodes directly implanted in their brains for medical reasons. This technique can collect very focal neural activity. A univariate analysis made it possible to see which region of the brain was solicited during the task, electrode by electrode, contact by contact. Solely the contacts in the posterior superior temporal lobe were active, thus confirming the results of the Geneva study.
However, when a machine-learning algorithm was applied to all of the data, thus making a multivariate decoding of data possible, positive results were observed in the entire temporal lobe, and even beyond it. “Learning algorithms are intelligent but ignorant”, specifies Anne-Lise Giraud.

“They are very sensitive and use all of the information in the signals. However, they do not allow us to know whether this information was used to perform the task, or if it reflects the consequences of this task—in other words, spreading information in our brain”, continues Valérian Chambon, researcher at the Departement d’études cognitives at the ENS. The mapped regions outside of the posterior superior temporal lobe are thus false positives, in a way. These regions retain information on the decision that the subject makes (BA or DA), but aren’t solicited to perform this task.
This research offers a better understanding of how our brain portrays syllables and, by showing the limits of artificial intelligence in certain research contexts, fosters welcome reflection on how to interpret data produced by machine learning algorithms.

31 Jan 2018


Our experts

Talk to specialists from all disciplines

The Photo Library

Discover and download UNIGE images


All our press releases since 2012

Subscribe to press releases

Université de Genève

24 rue Général-Dufour

CH-1211 Genève 4

T. +41 22 379 77 96

F. +41 22 379 77 29