This article has been reviewed according to Science X's editorial process and policies. Editors have highlighted the following attributes while ensuring the content's credibility:


peer-reviewed publication

trusted source


Unleashing the power of AI to track animal behavior

Unleashing the power of AI to track animal behavior
Human hand marked with GlowTrack fluorescent tags. Credit: Salk Institute

Movement offers a window into how the brain operates and controls the body. From clipboard-and-pen observation to modern artificial intelligence-based techniques, tracking human and animal movement has come a long way. Current cutting-edge methods utilize artificial intelligence to automatically track parts of the body as they move. However, training these models is still time-intensive and limited by the need for researchers to manually mark each body part hundreds to thousands of times.

Now, Associate Professor Eiman Azim and team have created GlowTrack, a non-invasive movement tracking method that uses markers to train . GlowTrack is robust, time-efficient, and —capable of tracking a single digit on a mouse's paw or hundreds of landmarks on a human hand.

The technique, published in Nature Communications on September 26, 2023, has applications spanning from biology to robotics to medicine and beyond.

"Over the last several years, there has been a revolution in tracking behavior as powerful artificial intelligence tools have been brought into the laboratory," says Azim, senior author and holder of the William Scandling Developmental Chair. "Our approach makes these tools more versatile, improving the ways we capture diverse movements in the laboratory. Better quantification of movement gives us better insight into how the brain controls behavior and could aid in the study of movement disorders like (ALS) and Parkinson's disease."

Current methods to capture often require researchers to manually and repeatedly mark body parts on a computer screen—a time-consuming process subject to and time constraints. Human annotation means that these methods can usually only be used in a narrow testing environment, since artificial intelligence models specialize to the limited amount of training data they receive. For example, if the light, orientation of the animal's body, camera angle, or any number of other factors were to change, the model would no longer recognize the tracked body part.

To address these limitations, the researchers used fluorescent dye to label parts of the animal or human body. With these "invisible" fluorescent dye markers, an enormous amount of visually diverse data can be created quickly and fed into the artificial intelligence models without the need for human annotation. Once fed this robust data, these models can be used to track movements across a much more diverse set of environments and at a resolution that would be far more difficult to achieve with manual human labeling.

This opens the door for easier comparison of movement data between studies, as different laboratories can use the same models to track body movement across a variety of situations. According to Azim, comparison and reproducibility of experiments are essential in the process of scientific discovery.

"Fluorescent dye markers were the perfect solution," says first author Daniel Butler, a Salk bioinformatics analyst. "Like the on a dollar bill that lights up only when you want it to, our fluorescent dye markers can be turned on and off in the blink of an eye, allowing us to generate a massive amount of training data."

In the future, the team is excited to support diverse applications of GlowTrack and pair its capabilities with other tracking tools that reconstruct movements in three dimensions, and with analysis approaches that can probe these vast movement datasets for patterns.

"Our approach can benefit a host of fields that need more sensitive, reliable, and comprehensive tools to capture and quantify movement," says Azim. "I am eager to see how other scientists and non-scientists adopt these methods, and what unique, unforeseen applications might arise."

Other authors include Alexander Keim and Shantanu Ray of Salk.

More information: Large-scale capture of hidden fluorescent labels for training generalizable markerless motion capture models, Nature Communications (2023). DOI: 10.1038/s41467-023-41565-3

Journal information: Nature Communications

Provided by Salk Institute

Citation: Unleashing the power of AI to track animal behavior (2023, September 26) retrieved 7 December 2023 from
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.

Explore further

3D deep neural network precisely reconstructs freely-behaving animal's movements


Feedback to editors