In a Bad Mood? Change the Channel

Share Article

Researcher uses ASL eye tracking technology to help explain higher levels of happiness in older adults.


Associate professor Derek Isaacowitz uses ASL mobile eye-tracking technology to identify exactly what a person is looking at and for how long. Northeastern file photo.

We exert con­trol over what we are exposed to, and that may be a really crit­ical way that we reg­u­late our emotion.

If a war scene, horror flick, or some other negative image appears on the television, older adults will tend to avert their eyes.

Derek Isaacowitz, associate professor in the College of Science’s Department of Psychology at Northeastern University, has studied this behavior for several years using Applied Science Laboratories (ASL) eye-tracking technology in a novel way to pinpoint exactly what a person is fixing his or her attention on and for how long. In the case of older adults, the results so far have revealed a pattern of viewing focused on positive stimuli that may serve emotion regulation - and may partially explain the higher levels of happiness that are also recorded among older adults.

But the current research setting is a bit too contrived, Isaacowitz said, and it may not be the best way to test the question of how older and younger adults differ in their strategies to control their emotions.

“In our everyday lives, when we’re regulating our emotion, it’s rarely the case that somebody forces you to look at something,” he said. “We exert control over what we are exposed to, and that may be a really critical way that we regulate our emotion.”

For instance, instead of simply looking away from the TV or computer screen, one could change the channel or navigate to a new Web page. In order to get around this, he has recently begun to simulate the real world experience in the lab by offering participants a variety of media to choose from, including multiple screens and websites.

“But there’s a big technological problem here,” Isaacowitz explained. “The environment is different for every person because every person chooses different things so there’s no way for the system to automatically process data so we can look at attention in a more fine-grained manner.”

With funding from a new grant from the National Institutes of Health, Isaacowitz’s team will address this problem by joining forces with researchers from the game design program who are well versed in exactly this sort of challenge.

Isaacowitz said Northeastern’s commitment to use-inspired, interdisciplinary research made this project possible. It’s here that he met Magy Seif El-Nasr, associate professor of game design, at his new faculty orientation and immediately recognized a great collaborative opportunity in his midst.

The researchers will first build a software platform to automatically detect what a person is looking at on a screen the moment he or she decides to look at or away from it. Once they have this platform, they will begin testing it with two separate questions: first, how do people use their choices and their attention together to regulate out of a bad mood? And second, how do they use them to maintain a good mood?

“We’re trying to understand in the environments that older people actually spend their time in, how do they regulate the emotional inputs of their environment,” said Isaacowitz. “We want to know how people use their choices differently by age, and also how these age patterns are influenced by the mood we are in when we start making the choices.” This will help them understand how adults of different ages use the tools in the environment to help regulate how they feel.

Professor Isaacowitz has been using ASL eye tracking technology for his research studies on emotion and aging since 2001. Most recently, the wireless ASL Mobile Eye-XG glasses have provided him with a lightweight tracker that could be used with adults of all ages, with and without eyeglasses, for psychology experiments. The lightweight glasses support two digital high resolution cameras, one that records the scene image and the other, the participant’s eye. The glasses provide highly accurate point of gaze information as his participants view their environment. To view a video of ASL Mobile Eye-XG in action please click here.

About Applied Science Laboratories (ASL)

Applied Science Laboratories (ASL) in Bedford, Massachusetts, is the recognized authority in the field of eye tracking technology. The Mobile Eye-XG and a comprehensive EYE-TRAC® suite of versatile solutions designed and manufactured by ASL are used worldwide in academic research, market research and industrial markets. ASL’s innovations include: desktop optics with head motion compensation, eye/head integration, long range optics for fMRI and revolutionary data analysis options. With dynamic, scalable and innovative solutions delivered with world-class customer support, ASL continues to be the complete eye tracking hardware, software and professional services provider. Visit

Northeastern University published original article at

Share article on social media or email:

View article via:

Pdf Print

Contact Author

Yvonne Vant Erve
781-275-4000 10
Email >
Visit website