New algorithm improves the way computers interpret readings of the brain’s electrical signals

May 30, 2011 By Lee Swee Heng
Credit: 2010

Electroencephalography (EEG) records the electrical signals produced by the brain using an array of electrodes placed on the scalp. Computers use an algorithm called common spatial pattern (CSP) to translate these signals into commands for the control of various devices.

Haiping Lu at the A*STAR Institute for Infocomm Research and co-workers have now developed an improved version of CSP for classifying signals. The new algorithm will facilitate the development of advanced –computer interfaces that may one day enable paralyzed patients to control devices such as computers and robotic arms.

CSP distinguishes and interprets commands by estimating the variations between EEG signals, and its accuracy strongly depends on how many signals are provided. As a result, CSP may make an incorrect interpretation when the number of EEG signals is small.

The new CSP algorithm developed by Lu and his colleagues uses two parameters to regularize the estimation of EEG signal variations. One parameter lowers the variations of the estimates, while the other reduces the tendency of the CSP algorithm to bias the estimates towards values from only a small number of samples.

Together, these parameters significantly improve the accuracy of CSP for classifying EEG signals. The researchers optimized the new algorithm even further by aggregating a number of different regularizations.

They evaluated their algorithm in the third Brain–Computer Interface Competition, held in 2005. The competition enables researchers who are developing advanced brain–computer interface technologies to test their methods for processing and classifying EEG signals using publicly available data sets.

The algorithm developed by Lu’s group significantly out-performed four other groups in three sets of experiments with varying testing scenarios, and its superiority was particularly evident when the number of sample EEG signals was small.

Conventional algorithms require 20–90 EEG signals, but the algorithm of Lu and co-workers needs only ten. This significantly reduces the effort required to collect data for brain–computer interfaces, the memory requirements for EEG signal processing applications, and the processing time needed for processing the signals.

“This is a method to improve the accuracy of current brain–computer interfaces,” says Lu. “Our applies ensemble-based learning in the feature-extraction stage of an EEG-based brain–computer interface, which could be integrated with training data ensembles in the data pre-processing stage. It would be among many other improvements to be tested and used in existing brain–computer interface systems.”

Explore further: Computer scientists win a major grant to network mobile devices in the cloud

More information: Lu, H., et al. Regularized common spatial pattern with aggregation for EEG classification in small-sample setting. IEEE Transactions of Biomedical Engineering 57, 2936–2946 (2010)

Provided by Agency for Science, Technology and Research (A*STAR)

4.5 /5 (2 votes)

Related Stories

New way to detect epileptic seizures

Mar 23, 2011

Researchers at Concordia University have pioneered a computer-based method to detect epileptic seizures as they occur – a new technique that may open a window on the brain's electrical activity. Their paper, "A Novel ...

Using your mood to operate a computer game

May 28, 2009

( -- Brain Computer Interfaces measure electrical signals from the brain and convert them into data that can be used by a computer. You can move a cursor on your screen, for example, simply by ...

Towards zero training for brain-computer interfacing

Aug 13, 2008

While invasive electrode recordings in humans show long-term promise, non-invasive techniques can also provide effective brain-computer interfacing (BCI) and localization of motor activity in the brain for paralyzed patients ...

Recommended for you

Cutting the cloud computing carbon cost

Sep 12, 2014

Cloud computing involves displacing data storage and processing from the user's computer on to remote servers. It can provide users with more storage space and computing power that they can then access from anywhere in the ...

Teaching computers the nuances of human conversation

Sep 12, 2014

Computer scientists have successfully developed programs to recognize spoken language, as in automated phone systems that respond to voice prompts and voice-activated assistants like Apple's Siri.

Mapping the connections between diverse sets of data

Sep 12, 2014

What is a map? Most often, it's a visual tool used to demonstrate the relationship between multiple places in geographic space. They're useful because you can look at one and very quickly pick up on the general ...

User comments : 1

Adjust slider to filter visible comments by rank

Display comments: newest first

not rated yet May 30, 2011
"They evaluated their algorithm in the third BrainComputer Interface Competition, held in 2005."
No time machine was involved, despite what this sentence says.

Apparently, they evaluated their algorithm against the 3rd BCI dataset and got the reported, excellent results. Sadly, no trophy, though.