Researchers develop an AI that can read your eyes to identify neurological disorders

A new tool developed at the Kavli Institute for Systems Neuroscience in Norway and described in an article in Nature Neuroscience [1], predicts gaze direction and eye movement directly from magnetic resonance imaging (MRI) scans. The goal is to make eye-tracking diagnostics a standard in brain imaging research and hospital clinics.

Whenever you explore an environment or search for something, you scan the scene using continuous rapid eye movements. Your eyes also make short stops to fixate on certain elements of the scene that you want more detailed information about. The way you move your eyes and the selection of details you fixate your gaze upon can be summed up as your viewing behaviour.

Scientists can infer a great deal from analysing eye movements. For instance, when you recall a memory, your eyes move in a pattern similar to how they did when you first experienced the event. Or, when you enter your kitchen hungry, your eyes will be drawn to other items and follow different patterns than when you enter the kitchen to do the dishes.


DeepMReye is an eye tracker that uses the MRI signal from the eyes. Eliminating the use of a camera promises to overcome many obstacles that have prevented eye-tracking from being implemented as a standard in the imaging field. It will also open up completely new research and diagnostic possibilities.

“Our viewing behaviour is an expression of our thoughts, memories and goals at any given moment in time,” said Matthias Nau. In collabo- ration with Markus Frey and Christian Doeller, he developed the DeepMReye tool at the Kavli Institute for Systems Neuroscience at the Norwegian University of Science and Technology (NTNU) in Trondheim. Viewing behaviour can even be a diagnostic for brain diseases.

“Brain diseases manifest themselves as characteristic eye movement patterns and disturbances in viewing behaviour,” Nau said. “Almost every cognitive or neural disorder, such as working memory deficits, amnesia, Parkinson’s disease and Alzheimer’s disease will affect your viewing behaviour.”

AI decodes your gaze pattern

The human eye really can’t extract signals from noise in big datasets. Artificial intelligence, on the other hand, is surprisingly efficient and well equipped for discerning relevant patterns in haystacks of data. So the researchers started training a deep learning network to do what they themselves could not: pay attention to the viewing behaviour of subjects in the MR brain scanner and recognize patterns that are shared across people.

A long time and many MRI datasets later, the model had figured out the complicated task of generalizing patterns of gaze across people. In a process involving convolutions that extract features of the data through dimensionality reduction, the artificial intelligence had learned how to use the vast pool of knowledge it had been fed, to extract and interpret the hidden meaning from any person’s viewing behaviour.

The AI’s ability to generalize knowledge to new individuals, combined with the fact that all the data the AI requires for analysis is already found in most MRI maps, brings new life and interest to old brain scanning data.

“Researchers can use this model to analyse retrospectively from 20 years of existing fMRI data, from many thousands of subjects, and ask completely new questions,” Nau said. For instance, researchers can ask how ongoing behaviour really influences so called “resting state networks” – currently one of the juicier topics in the brain imaging world.

DeepMReye can also be used for patient groups and in categories of research that the old camera-based eye tracking did not support. While camera eye tracking could only collect information from subjects when their eyes were open, the MR signal locates eyeball movement patterns even when the eyelid is closed. This makes the DeepMReye model relevant for sleep labs, for instance by studying eye movements to classify sleep stages.

Another example is clinical health screenings of congenitally blind people, who up till now have been excluded from eye tracking studies because the camera calibration process requires a seeing and focusing eye. DeepMReye does not discriminate between seeing and blind eyes when it comes to health screening.

Imaging research and diagnostics

There is a general agreement between scientists and clinicians that eye tracking is an important measure for investigating cognition and diseases. Combining precise measurements of eye movements with neuroimaging maps of whole-brain activity makes for a powerful tool. However, eye tracking has yet to become a standard in imaging research labs and the clinical routine of hospitals.

So why hasn’t eye tracking caught on? Nau has a clear explanation: “The existing camera eye-tracking procedure is quite expensive, it is tricky to use, and way too time consuming. We don’t have the time, the clinical routine, or the expertise to use these cameras for every single patient or research subject.”

This is a suboptimal situation well known amongst professionals. Currently, 90% of MRI studies in the top research journals published in the last two years did not track eye movement. Of the 10% that used eye tracking, 5% reported poor data quality, and the remaining 5% used the data in a way that didn’t really require a camera.

“This is a missed opportunity. But it is also alarming,” Nau said. “There are many problems and artefacts associated with fMRI brain activity studies that these researchers and clinicians are making themselves blind to, which could be cleaned up by including eye tracking into the procedure.”

Subsequent memory effect

Several studies are currently claiming that they have identified memory encoding and all sorts of higher-level cognitive processes, which at the end of the day may just be an eye-movement confound. One candidate for this category is the subsequent memory effect.

“If I show you two images and you later remember only one of them, I may also find that the activity in your hippocampus, which is your brain’s memory encoding hub, was higher for the image that you remembered than for the one that you had forgotten,” Nau said.

The theory of a subsequent memory effect suggested that the correlation between the two is directly related to memory encoding, and that the bump in activity is like a signature of memory. Thus, higher activity means better recall of that image later.

However, it turns out that the images you remember are also the ones that you looked at more. There is a direct correlation between eye movement, rise in local brain activity, and successful memory encoding. Your eyes made more fixations on that image. You invested way more energy in getting information from that image. And that is why you remember it later on. More fixations mean stronger brain activity. So, if you take into account the fact that people move their eyes more during encoding, then this subsequent memory effect disappears.

“The reason why we encode and remember some images and forget others, may simply be because we find some scenes more interesting, we are drawn to them, and we are willing to spend more time studying them to extract unique features of the scene. What we are beginning to realize, is that we can’t fully piece apart these memory effects from our viewing behaviour, they are inherently linked,” Nau said.

A free, open-source, plug and play tool

Crafting the camera-less eye tracking tool started as a weekend project between the students about three years ago.

“We wanted to build the package that we ourselves would like to have,” Nau said. “A user-friendly eye tracker that uses the MRI signal from the eyes, a model that could improve interpretability and robustness of neuroimaging research, and that would over- come the obstacles that had prevented eye-tracking from being implemented as a standard in the imaging field.”

“This project has been different from the average paper,” Nau said. “Usually you work on something for a few years, you write it up, you publish it, and then you close the book and move on to a new project. Whereas this paper will take on a whole new life once we publish it. Once the users start using DeepMReye, they will have questions, they will have requests, they will find bugs.”

The researchers have made the tool open-source and fully available from GitHub [2].

“You don’t need any money, any equipment, or any trained personnel to use the model. You don’t have to spend costly experimental time or medical consultation time for setting up and calibrating a camera. The model can be used post-hoc, after the patient is well at home,” Nau said.

The researchers also worked hard to make the tool as user friendly as possible, and created a website with user recommendations and a FAQ [3].

“Hopefully this will be plug-and-play for everybody,” Nau said.

Christian Doeller’s research group at Kavli Institute for Systems Neuroscience. First row from left: Lilith Sommer, Dörte Kuhrt, Bianca Somai, Annelene Gulden Dahl, Gøril Rolfseng Grøntvedt, Tobias Navarro Schroeder, Britt Veldman. Second row from left: Christian Doeller, Joshua Julian, Mona Gravert, Matthias Nau, Renil Mathew, Ignacio Polti, Markus Frey, Jacob Bellmund. Christian Doeller currently heads the Max Planck Institute for Human Cognitive and Brain Sciences (MPI CBS) in Leipzig. CREDIT: Kavli Institute for Systems Neuroscience 


1. Frey, M., Nau, M. & Doeller, C.F. Magnetic resonance-based eye tracking using deep neural networks. Nature Neuroscience (2021).

2. DeepMReye on Github

3. DeepMReye documentation