검색 전체 메뉴
PDF
맨 위로
OA 학술지
Brain Computer Interfacing: A Multi-Modal Perspective
  • 비영리 CC BY-NC
  • 비영리 CC BY-NC
ABSTRACT
Brain Computer Interfacing: A Multi-Modal Perspective
KEYWORD
Brain computer interfaces , Multi-modal , Subject-independent classification , EEG-NIRS
  • I. INTRODUCTION

    In recent brain computer interface (BCI) research, training and calibration times have been drastically reduced due to the use of machine learning and adaptive signal processing techniques [1] as well as novel dry electrodes [2-7]. Initial BCI systems were based on operant conditioning, and could easily require months of training on the subject side before it was possible to use them in an online feedback setting [8,9].

    Second-generation BCI systems required the recording of a brief calibration session, during which a subject assumes a fixed number of brain states such as that during movement imagination, after which subject-specific spatio-temporal filters [10] are inferred along with individualized classifiers [1]. The first steps to transfer a BCI user’s filters and classifiers between sessions have been studied [11], and a further online study confirmed that such a transfer is indeed possible without significant performance loss [12]. However, while this work focused on reusing data from previous sessions of expert users, more recent approaches designed subject-independent zerotraining BCI classifiers that enable both experienced and novice BCI subjects to use a BCI system immediately, without the need for recording calibration sessions [13-15]. Various of state-of-the-art learning methods (e.g., SVM, Lasso, etc.) have been applied in order to construct one-size-fits-all classifiers from a vast number of redundant features. The use of sparsifying techniques leads to features within the electroencephalogram (EEG) data that are predictive for future BCI users. The findings show a distribution of different alpha-band features in combination with a number of characteristic common spatial patterns (CSPs) that are highly predictive for most users. Clearly, these type of procedures may also be of use in scientific fields other than BCI, where complex characteristic features need to be homogenized into one overall inference model.

    To date, EEG is the most widely used technology in the context of BCI. Some of the main reasons are the relatively low costs and fast setup times partly attributable to the advent of dry-electrode and subject-independent classifiers. Another important reason is the high temporal resolution that EEG offers. However, EEG sensory motor rhythm (SMR)-based BCI still suffers from a number of problems. Unfortunately, not all subjects are able to alter their sensory motor rhythms, and thus, SMR-based BCIs do not work for all subjects. To this end, simultaneous recordings of near-infrared spectroscopy (NIRS) and EEG have not only been shown to increase BCI performance, but have also enabled some subjects to operate a BCI, who previously were not able to do so using solely EEG [16].

    While EEG has the highest temporal resolution of all neuroimaging methods, NIRS is dependent on changes of blood flow, as it measures the oxygenated and deoxygenated hemoglobin (HbO2 and HbR) in the superficial layers of the human cortex. The temporal resolution of NIRS is therefore orders of magnitudes lower, which substantially reduces the upper bound of information transfer rates (ITRs) for NIRS-based BCIs. Every neuroimaging method suffers from particular limitations. EEG has poor spatial resolution, while NIRS suffers from sluggishness of the underlying vascular response, which limits its temporal resolution. By employing a multi-modal neuroimaging approach, combining EEG and NIRS, it becomes possible to focus on their individual strengths and to partly overcome these limitations. In particular, we will review how extracting relevant NIRS features to support and complement high-speed EEG-based BCI and thus forming a hybrid BCI [17], allows for exploitation of the responsiveness of EEG (i.e., high ITR) as well as achieving enhanced and robust overall BCI performance by using information from the vascular response not contained within the EEG.

    II. SUBJECT-INDEPENDENT BCI CLASSIFICATION

    On the path of bringing BCI technology from the lab into practical use, it becomes indispensable to reduce the setup time. While dry electrodes provide a first step to eliminating the time needed for placing a cap, the need for recording calibration sessions has still remained. Employing a large body of high-quality experimental data accumulated over the years now enables the experimenter to choose, by means of machine learning technology, a very sparse set of voting classifiers, which perform as well as the standard, state-of-the-art subject-calibrated methods. L1-regularized regression in this case performs better than other methods such as majority voting (Table 1). The features that show best generalization do not exclusively come from the best-performing subjects, and in fact, some average performers were also selected (Fig. 1). Interestingly, some subjects with very high BCI performance were not selected at all, while others generalized well in the sense that their models were able to predict other subjects’ data. No single frequency band dominated with regard to classification accuracy (Fig. 1a). Therefore, diverse features must be selected for the regularization. Nevertheless, there is significant redundancy between classifiers in the ensemble. The approach of finding a sparse solution reduces the dimensionality of the chosen features significantly. For very able subjects, zero-training methods exhibit a slight performance decrease, though this will not prevent them from performing successfully in BCI. The sparsification of classifiers leads to insight into neurophysiological processes [15]. These processes identify relevant cortical locations and frequency bands of neuronal population activity that are in agreement with general neuroscientific knowledge. While these studies concentrated on zero-training classification and not brain activity interpretation, a much closer look is warranted. Movement imagination detection is not only determined by the cortical representation of the limb whose control is being imagined (in this case the arm), but also by differentially located cortical regions involved in movement planning (frontal), execution (fronto-parietal), and sensory feedback (occipito-parietal). Patterns relevant to BCI detection appear in all these areas, and while dominant discriminant frequencies are in the a range, higher frequencies appear in the ensemble, albeit in combination with less focused patterns. So, that which is found by the machine learning algorithm should be interpreted as representing the characteristic neurophysiological variation a large subject group, which in itself is a highly relevant topic.

    III. MULTI-MODAL RECORDINGS FOR BCI

    BCIs that solely rely on NIRS have been realized recently [18,19]. However, when looking at plain NIRS classification rates, it becomes apparent that NIRS cannot be seen as a viable alternative to EEG-based BCIs on its own. However, in a combination with EEG, we find that NIRS is capable of significantly enhancing event-related desynchronization (ERD)-based BCI performance. Not only does it increase performance for most subjects, but it also allows for meaningful classification rates for those who would otherwise not be able to operate a solely EEG-based BCI [16].

    Some subjects, who are able to operate SMR-based BCIs, experience a high level of non-stationarities in their BCI performance, as can be seen in Fig. 2. Possible solutions have been proposed, such as adaptive models [20,21], stationary subspace analysis [22], and a multi-modal approach [23]. The findings indicate that the performance fluctuations of EEG-based BCI control can be predicted by the preceding NIRS activity. These NIRS-based predictions are then employed to generate new, more robust EEG-based BCI classifiers, which enhance classification significantly, while at the same time minimize performance fluctuations and thus increase the general stability

    [Table 1.] Comparing subject-independent classifier result of various machine learning techniques to various baseline

    label

    Comparing subject-independent classifier result of various machine learning techniques to various baseline

    of BCI performance [23].

    Two blocks of real-time EEG-based, visual feedbackcontrolled motor imagery (50 trials per block per condition) were recorded for the estimation of the EEG classifier. The first 2 seconds of each trial began with a black fixation cross, which appeared at the center of the screen. Then, an arrow appeared as a visual cue pointing to the left or right, and the fixation cross started moving for 4 seconds, according to the classifier output. After 4 seconds, the cross disappeared and the screen remained blank for 10.5 ± 1.5 seconds. The online processing was based on the concept of coadaptive calibration [21].

    The user was given instantaneous EEG-based BCI feedback for the two blocks of motor imagery. During the first block of 100 trials, a subject-independent classifier was used, depending on band power estimates of Laplacian- filtered, motor-related EEG channels. For the second block subject-dependent spatial and temporal filters were estimated from the data of the first block and combined with some subject-independent features to form the classifier for the second block. During the online feedback, features were calculated every 40 ms with a sliding window of 750 ms. For further details on coadaptive calibration, we would like to refer the reader to a study by Vidaurre et al. [21].

    Once the EEG classifier is estimated, a feedback block with 300 trials and a relatively short inter-stimulus interval of 7 seconds is recorded, lasting a total of 35 minutes. As before, the first 2 seconds of each trial began with the appearance of a black fixation cross at the center of the screen. Then, a visual cue in the form of an arrow appeared to indicate the required class, and the fixation cross started moving for 4 seconds, according to the classifier output. After the 4 seconds, the cross disappeared

    and the screen remained blank for a short interval of 1 ± 0.5 seconds before the next trial began.

    The long inter-trial intervals in the first two blocks were chosen to evaluate the NIRS signals with respect to motor imagery [16]. Here, we only investigate the 300 trials of the fast feedback dataset. These trials are split into chronological blocks of 5 trials each, resulting in 60 blocks and the EEG-BCI performance of these smaller blocks are computed. For each block the EEG-based classifier output, out is multiplied by its true label

    image

    and summed over the 5 trials within the block, resulting in the performance y of the given block:

    image

    where i is the trial within block k. By calculating the performance this way, a continuous performance measure can be obtained, which is preferable to a mere 0?1 loss, since it is more accurate and also more suitable for the purpose of regressing NIRS features onto this measure, which will be explained. The overall performance is subtracted from each block-performance, such that a ‘time course’ of above- and below-average EEG-BCI performance is obtained.

    The NIRS signal is divided into multiple epochs, each with a width of 2 seconds, preceding each 5-trial block by 2, 4, 6, 8, and 10 seconds. Noisy channels are discarded, and the signal is transferred into the spectral domain. An L1-regularized regression optimization problem [24] is formulated (and implemented with CVX [25]) to identify spectral NIRS features predicting the EEG performance of the following 5 trials:

    image

    where X are the preceding NIRS features, β is the regressor, and λ is the regularization variable. Refer to Fig. 3 for a graphical explanation. The thin black boxes on the

    upper part represent trials, where left and right-hand motor imagery are cued. The optimization problem is repeated 60 times, and each time a different block is left out, resulting in a performance prediction for each block. Using the predictions as well as the actual performance, the correlation coefficient and its p-value are calculated. The corresponding p-values test the hypothesis of zero correlation. Since the method is repeated a number of times for various intervals, the p-values are Bonferroni corrected.

    Depending on the prediction of the NIRS, the EEG data is grouped into three categories: blocks with good performance, medium performance, and bad performance. All three groups have an even number of trials (100 trials per group). A validation scheme is setup, where one block (consisting of 5 trials) is left out as a test-set. Using the training, set we calculate four EEG classifiers: one for each group, defined by the performance prediction, and a fourth comprising all training data. The individual classifiers

    [Table 2.] Correlation coefficients (cc) and p-values (p) of predicted vs. actual electroencephalogram performances

    label

    Correlation coefficients (cc) and p-values (p) of predicted vs. actual electroencephalogram performances

    [Table 3.] Percentage classification loss of all individual subjects as well as their means

    label

    Percentage classification loss of all individual subjects as well as their means

    consist of a fixed broadband temporal filter (5thorder Butterworth digital filter with 5?30 Hz), a spatial filter (CSP [10]) and a linear classifier (LDA). In a second step, we train a meta-classifier, combining all four individual classifiers, based on the training set. The outputs of this meta-classifier are then compared to the true labels of the left-out trials, and its 0?1 loss is computed. This procedure is repeated for all blocks. As a baseline, we use the EEG classifier that was trained on all training data.

    Table 2 shows the results of the optimization problem defined by Equation (2). For 9 out of 14 subjects, the pvalues of the correlation coefficients calculated between the predicted and actual EEG performance are significant.

    Table 3 shows the classification results of all subjects, and their means. A paired t-test between the standard procedure of treating all training trials the same, as compared to a meta-classifier that combines four classifiers, based on the performance of blocks, results in a value of

    p = 0.013 (which is indicated by a ‘*’ in Table 3).

    To evaluate whether our method reduces the performance variability during the feedback session, we calculate the standard deviation over all 60 blocks for the standard method, where all trials are treated the same, as well as for the meta-classifier.

    Fig. 4 shows the results in the form of a scatter-plot. As can be seen, our proposed method reduces the variability of performance in 11 out of 14 subjects. In one subject, the variability is the same, and for two subjects, the standard procedure has lower performance fluctuations. A paired t-test reveals a significant relationship with p < 0.05.

    IV. DISCUSSION

    Multi-modal techniques can be useful in a number of ways. For the case of subject-independent decoding, we found that the outcome of a machine learning experiment can also be viewed as a compact quantitative description of the characteristic variability between individuals in the large subject group. Note that it is not the best subjects that characterize the variance necessary for a subjectindependent algorithm, but the spread over existing physiology is represented concisely.

    For the case of combining NIRS and EEG, it can be concluded that this novel approach of combining NIRS and EEG is a viable technique, suitable for SMR-based BCI, since it preserves the responsiveness of the EEG, while at the same time significantly enhances classification rates, and minimizes performance fluctuations.

참고문헌
  • 1. Dornhege G., Millan J. R., Hinterberger T., McFarland D., Muller K. R. 2007 Toward Brain-Computer Interfacing google
  • 2. Popescu F., Fazli S., Badower Y., Blankertz B., Muller K. R. 2007 “Single trial classification of motor imagination using 6 dry EEG electrodes” [PLoS ONE] Vol.2 P.e637 google cross ref
  • 3. Grozea C., Voinescu C. D., Fazli S. 2011 “Bristle-sensors: low-cost flexible passive dry EEG electrodes for neurofeedback and BCI applications” [Journal of Neural Engineering] Vol.8 P.25008 google cross ref
  • 4. Zander T. O., Lehne M., Ihme K., Jatzev S., Correia J., Kothe C., Picht B., Nijboer F. 2011 “A dry EEG-system for scientific research and brain-computer interfaces” [Frontiers in Neuroscience] Vol.5 P.53 google cross ref
  • 5. Chi Y. M., Wang Y. T., Wang Y., Maier C., Jung T. P., Cauwenberghs G. 2012 “Dry and noncontact EEG sensors for mobile brain-computer interfaces” [IEEE Transactions on Neural Systems and Rehabilitation Engineering] Vol.20 P.228-235 google cross ref
  • 6. Liao L. D., Chen C. Y., Wang I. J., Chen S. F., Li S. Y., Chen B.W., Chang J. Y., Lin C. T. 2012 “Gaming control using a wearable and wireless EEG-based brain-computer interface device with novel dry foam-based sensors” [Journal of Neuroengineering and Rehabilitation] Vol.9 P.5 google cross ref
  • 7. Guger C., Krausz G., Allison B. Z., Edlinger G. 2012 “Comparison of dry and gel based electrodes for P300 brain-computer interfaces” [Frontiers in Neuroscience] Vol.6 P.60 google
  • 8. Elbert T., Rockstroh B., Lutzenberger W., Birbaumer N. 1980 “Biofeedback of slow cortical potentials. I” [Electroencephalography and Clinical Neurophysiology] Vol.48 P.293-301 google cross ref
  • 9. Birbaumer N., Ghanayim N., Hinterberger T., Iversen I., Kotchoubey B., Kubler A., Perelmouter J., Taub E., Flor H. 1999 “A spelling device for the paralysed” [Nature] Vol.398 P.297-298 google cross ref
  • 10. Blankertz B., Tomioka R., Lemm S., Kawanabe M., Muller K.R. 2008 “Optimizing spatial filters for robust EEG singletrial analysis” [IEEE Signal Processing Magazine] Vol.25 P.41-56 google cross ref
  • 11. Krauledat M., Schroder M., Blankertz B., Muller K. R. 2007 “Reducing calibration time for brain-computer interfaces: a clustering approach,” in Advances in Neural Information Processing Systems 19, B. Scholkopf, J. Platt, and T. Hoffman, editors P.753-760 google
  • 12. Krauledat M., Tangermann M., Blankertz B., Muller K. R. 2008 “Towards zero training for brain-computer interfacing” [PLoS ONE] Vol.3 P.e2967 google cross ref
  • 13. Fazli S., Popescu F., Danoczy M., Blankertz B., Muller K. R., Grozea C. 2009 “Subject-independent mental state classification in single trials” [Neural Networks] Vol.22 P.1305-1312 google cross ref
  • 14. Fazli S., Grozea C., Danoczy M., Blankertz B., Popescu F., Muller K. R. 2009 “Subject independent EEG-based BCI decoding,” in Advances in Neural Information Processing Systems 22, Y. Bengio, D. Schuurmans, J. Lafferty, C. K. I. Williams, and A. Culotta, editors P.513-521 google
  • 15. Fazli S., Danoczy M., Schelldorfer J., Muller K. R. 2011 “(1)- penalized linear mixed-effects models for high dimensional data with application to BCI” [Neuroimage] Vol.56 P.2100-2108 google cross ref
  • 16. Fazli S., Mehnert J., Steinbrink J., Curio G., Villringer A., Muller K.R., Blankertz B. 2012 “Enhanced performance by a hybrid NIRS-EEG brain computer interface” [Neuroimage] Vol.59 P.519-529 google cross ref
  • 17. Pfurtscheller G., Allison B. Z., Brunner C., Bauernfeind G., Solis-Escalante T., Scherer R., Zander T. O., Mueller-Putz G., Neuper C., Birbaumer N. 2010 “The hybrid BCI” [Frontiers in Neuroscience] Vol.4 P.30 google
  • 18. Tsubone T., Muroga T., Wada Y. 2007 “Application to robot control using brain function measurement by near-infrared spectroscopy” [in Proceedings of the 29th Annual International Conference of the IEEE Engineering in Medicine and Biology Society] P.5342-5345 google
  • 19. Abdelnour A. F., Huppert T. 2009 “Real-time imaging of human brain function by near-infrared spectroscopy using an adaptive general linear model” [Neuroimage] Vol.46 P.133-143 google cross ref
  • 20. Krauledat M. 2008 “Analysis of nonstationarities in EEG signals for improving brain-computer interface performance” google
  • 21. Vidaurre C., Sannelli C., Muller K. R., Blankertz B. 2011 “Co-adaptive calibration to improve BCI efficiency” [Journal of Neural Engineering] Vol.8 P.25009 google cross ref
  • 22. von Bunau P., Meinecke F. C., Kiraly F., Muller K. R. 2009 “Finding stationary subspaces in multivariate time series” [Physical Review Letters] Vol.103 P.214101 google cross ref
  • 23. Fazli S., Mehnert J., Steinbrink J., Blankertz B. 2012 “Using NIRS as a predictor for EEG-based BCI performance” [in Proceedings of the 34th Annual International Conference of the IEEE Engineering in Medicine and Biology Society] P.4911-4914 google
  • 24. Tibshirani R. 1996 “Regression shrinkage and selection via the Lasso” [Journal of the Royal Statistical Society B] Vol.58 P.267-288 google
  • 25. Grant M., Boyd S. 2011 “CVX: MATLAB software for disciplined convex programming, version 1.21” google
OAK XML 통계
이미지 / 테이블
  • [ Table 1. ]  Comparing subject-independent classifier result of various machine learning techniques to various baseline
    Comparing subject-independent classifier result of various machine learning techniques to various baseline
  • [ Fig. 1. ]  (a) The used temporal filters and in color-code their contribution to the final L1-regularized regression classification (the scale is normalized from 0 to 1). Cleary μ-band temporal filters between 10-13 Hz are most predictive. (b) Number of features used vs. selfpredicted cross-validation. A high self-prediction can be seen to yield a large number of features that are predictable for the whole ensemble.
    (a) The used temporal filters and in color-code their contribution to the final L1-regularized regression classification (the scale is normalized from 0 to 1). Cleary μ-band temporal filters between 10-13 Hz are most predictive. (b) Number of features used vs. selfpredicted cross-validation. A high self-prediction can be seen to yield a large number of features that are predictable for the whole ensemble.
  • [ Fig. 2. ]  The electroencephalogram classification rate is generally non-stationary in time. We show the fluctuations in performance for the first 6 subjects. Each dot represents the classification performance of 5 trials. On the y-axis the classification accuracy is depicted in percent, and the x-axis represents the time-course of the experiment. The magenta color shows early stages of the experiment, where the classifier is not yet considered to be stable, the green indicates a later state of the experiment, where the classifier is kept constant. Horizontal lines show the average classification performance.
    The electroencephalogram classification rate is generally non-stationary in time. We show the fluctuations in performance for the first 6 subjects. Each dot represents the classification performance of 5 trials. On the y-axis the classification accuracy is depicted in percent, and the x-axis represents the time-course of the experiment. The magenta color shows early stages of the experiment, where the classifier is not yet considered to be stable, the green indicates a later state of the experiment, where the classifier is kept constant. Horizontal lines show the average classification performance.
  • [ Fig. 3. ]  The continuous electroencephalogram (EEG) during the fast feedback phase (top). The simultaneously recorded continuous oxygenated and deoxygenated hemoglobin (HbO2 and HbR) near-infrared spectroscopy (NIRS) chormophores. Note that random data is shown here for visualization purposes (bottom).
    The continuous electroencephalogram (EEG) during the fast feedback phase (top). The simultaneously recorded continuous oxygenated and deoxygenated hemoglobin (HbO2 and HbR) near-infrared spectroscopy (NIRS) chormophores. Note that random data is shown here for visualization purposes (bottom).
  • [ Table 2. ]  Correlation coefficients (cc) and p-values (p) of predicted vs. actual electroencephalogram performances
    Correlation coefficients (cc) and p-values (p) of predicted vs. actual electroencephalogram performances
  • [ Table 3. ]  Percentage classification loss of all individual subjects as well as their means
    Percentage classification loss of all individual subjects as well as their means
  • [ Fig. 4. ]  Scatter-plot of the standard deviation of performance over all blocks. Each dot represents a single subject. The percentage on the top left indicates for how many subjects the meta classifier has lower standard deviation. p indicated the significance of a paired t-test.
    Scatter-plot of the standard deviation of performance over all blocks. Each dot represents a single subject. The percentage on the top left indicates for how many subjects the meta classifier has lower standard deviation. p indicated the significance of a paired t-test.
(우)06579 서울시 서초구 반포대로 201(반포동)
Tel. 02-537-6389 | Fax. 02-590-0571 | 문의 : oak2014@korea.kr
Copyright(c) National Library of Korea. All rights reserved.