This is a recent theory on how our brain works: it is a predictive machine, which continuously the compares sensory information that we pick up (such as images, sounds and language) with internal predictions. "This theoretical idea is extremely popular in neuroscience, but the existing evidence for it is often restricted to artificial situations," says lead author Micha Heilbron of the Donders Institute (University) and the Max Planck Institute for Psycholinguistics. "I would really like to understand precisely how this works and test it in different situations."
Brain research into this phenomenon is usually done in an artificial setting, Heilbron reveals. Subjects are told to stare at dots for half an hour, or they listen to abstract noises that sound like 'bib bib boop'. "Studies of this kind do in fact reveal that our brain can make predictions, but not that this always happens in the complexity of everyday life as well. We are trying to take it out of the lab setting. We are studying the same type of phenomenon, how the brain deals with unexpected information, but then in natural situations that are much less predictable."
Hemingway and Holmes
The researchers analysed the brain activity of people listening to stories by Hemingway or about Sherlock Holmes. At the same time, they analysed the texts of the books using computer models (deep neural networks). By this means, they were able to calculate for each word how unpredictable it was.
For each word or sound, the brain makes detailed statistical expectations and turns out to be extremely sensitive to the degree of unpredictability: the brain response is stronger whenever a word is unexpected in the context. "That is not very surprising: after all, everyone knows that you can sometimes predict language. For example, you complete sentences yourself if someone starts to speak very slowly, stutters or is unable to think of a word. But what we have shown here is that this happens continuously. Our brain is constantly guessing at words; the predictive machinery is always turned on."
More than software
"In fact, our brain does something comparable to speech recognition software. Speech recognisers using artificial intelligence are also constantly making predictions and are allowing themselves to be guided by their expectations, just like the autocomplete function on your phone. Nevertheless, we observed a big difference: brains predict not only words, but make predictions on many different levels, from abstract meaning and grammar to specific sounds."
There is good reason for the ongoing interest from tech companies who would like to use new insights of this kind to build better language and image recognition software, for example. But these sorts of applications are not the main aim for Heilbron. "I would really like to understand how our predictive machinery works and when at a fundamental level. I’m now working with the same research setup, but for visual and auditive perceptions, like music."