Machine-Learning Methods for Decoding Intentional Brain States
PDF WebBrain-computer interfaces (BCI) work by making the user perform a specific mental task, such as imagining moving body parts or performing some other covert mental activity, or attending to a particular stimulus out of an array of options, in order to encode their intention into a measurable brain signal. Signal-processing and machine-learning techniques are then used to decode the measured signal to identify the encoded mental state and hence extract the user‘s initial intention. The high-noise high-dimensional nature of brain-signals make robust decoding techniques a necessity. Generally, the approach has been to use relatively simple feature extraction techniques, such as template matching and band-power estimation, coupled to simple linear classifiers. This has led to a prevailing view among applied BCI researchers that (sophisticated) machine-learning is irrelevant since it doesn‘t matter what classifier you use once your features are extracted. Using examples from our own MEG and EEG experiments, I‘ll demonstrate how machine-learning principles can be applied in order to improve BCI performance, if they are formulated in a domain-specific way. The result is a type of data-driven analysis that is more than just classification, and can be used to find better feature extractors.
| Author(s): | Hill, NJ. |
| Links: | |
| Year: | 2010 |
| Month: | March |
| Day: | 30 |
| BibTeX Type: | Talk (talk) |
| Digital: | 0 |
| Electronic Archiving: | grant_archive |
| Event Name: | Symposium "Non-Invasive Brain Computer Interfaces: Current Developments and Applications" (BIOMAG 2010) |
| Event Place: | Dubrovnik, Croatia |
| Language: | en |
| Organization: | Max-Planck-Gesellschaft |
| School: | Biologische Kybernetik |
BibTeX
@talk{6430,
title = {Machine-Learning Methods for Decoding Intentional Brain States},
abstract = {Brain-computer interfaces (BCI) work by making the user perform a specific mental task, such as imagining moving body parts or performing some other covert mental activity, or attending to a particular stimulus out of an array of options, in order to encode their intention into a measurable brain signal. Signal-processing and machine-learning techniques are then used to decode the measured signal to identify the encoded mental state and hence extract the user‘s initial intention.
The high-noise high-dimensional nature of brain-signals make robust decoding techniques a necessity. Generally, the approach has been to use relatively simple feature extraction techniques, such as template matching and band-power estimation, coupled to simple linear classifiers. This has led to a prevailing view among applied BCI researchers that (sophisticated) machine-learning is irrelevant since it doesn‘t matter what classifier you use once your features are extracted.
Using examples from our own MEG and EEG experiments, I‘ll demonstrate how machine-learning principles can be applied in order to improve BCI performance, if they are formulated in a domain-specific way. The result is a type of data-driven analysis that is more than just classification, and can be used to find better feature extractors.},
organization = {Max-Planck-Gesellschaft},
school = {Biologische Kybernetik},
month = mar,
year = {2010},
author = {Hill, NJ.},
month_numeric = {3}
}