Comparative Analysis of Spectral Approac
Comparative Analysis of Spectral Approac
Abstract—The quantification of the spectral content of elec- aimed at replacing the impaired functionality of a human neuro-
troencephalogram (EEG) recordings has a substantial role in muscular system with computer-based analysis of the EEG and
clinical and scientific applications. It is of particular relevance establishing a communication channel independent of periph-
in the analysis of event-related brain oscillatory responses. This
work is focused on the identification and quantification of relevant eral nerves and muscles [4]. Most BCIs utilize EEG features that
frequency patterns in motor imagery (MI) related EEGs utilized are well-defined physiologically, which include oscillations in
for brain–computer interface (BCI) purposes. The main objective neuronal circuits or potentials evoked by particular stimuli [4].
of the paper is to perform comparative analysis of different ap- In this work, lateralized changes in the signal’s spectral content
proaches to spectral signal representation such as power spectral within (8–12 Hz) and (18–25 Hz) sensorimotor rhythms due
density (PSD) techniques, atomic decompositions, time-frequency
(t-f) energy distributions, continuous and discrete wavelet ap- to imagination of left- or right-hand movement are exploited for
proaches, from which band power features can be extracted and MI discrimination. The underlying brain phenomena is referred
used in the framework of MI classification. The emphasis is on to as event-related de-synchronization (ERD) and event-related
identifying discriminative properties of the feature sets repre- synchronization (ERS) [1], [3].
senting EEG trials recorded during imagination of either left– or The task of EEG spectral quantification is particularly
right-hand movement. Feature separability is quantified in the
offline study using the classification accuracy (CA) rate obtained challenging considering the complexity of the dynamics of
with linear and nonlinear classifiers. PSD approaches demonstrate nonstationary EEG [1]. It is required that the time variation of
the most consistent robustness and effectiveness in extracting the the relevant frequency components is accounted for. Statistical
distinctive spectral patterns for accurately discriminating between spectrum estimation methods, both parametric [5] and nonpara-
left and right MI induced EEGs. This observation is based on an metric [6], [7], have been extensively utilized for EEG analysis
analysis of data recorded from eleven subjects over two sessions
of BCI experiments. In addition, generalization capabilities of the in studies of MIs. Digital filtering, commonly applied in BCIs,
classifiers reflected in their intersession performance are discussed e.g., [8], represents a different concept of feature extraction.
in the paper. Instead of estimating spectral content of signal segments,
Index Terms—Alternative communication, brain–computer digital filters select their rhythmical components depending on
interface (BCI), electroencephalogram (EEG), spectral analysis, the frequency band of interest. Joint time-frequency (t-f) and
time-frequency (t-f) analysis, wavelet transforms. time-scale (t-s) techniques are considered to be particularly
well suited for analysis of EEGs as representative examples of
signals with time-varying spectral content [2]. Hence, quadratic
I. INTRODUCTION t-f distributions and wavelet-based methods have also been
HE electroencephalogram (EEG) is one of the most clin- exploited in BCI, e.g., [9], [10].
T ically and scientifically exploited signals recorded from
humans. Hence, its quantification plays a prominent role in brain
EEG feature extraction approaches are clearly dominated
by methods estimating the signal’s energy distributed in the
studies [1]. In particular, the examination of the frequency con- frequency, t-f or t-s domains. However, to the best of authors’
knowledge, a comparative analysis of a sizeable set of these
tent of EEG signals has been recognized as the most prepon-
derant approach to the problem of extracting knowledge of the methods applied to the same data in the BCI framework using
brain dynamics [1], [2] and thus it has been heavily exploited in the-state-of-the-art linear and nonlinear BCI classifiers has not
the study of the brain phenomena related to motor imagery (MI) been undertaken yet. Most comparative evaluations reported in
the literature are confined to a small number of techniques or only
[3]. The outcome of this study is crucial for successful devel-
opment of brain–computer interface (BCI) technology, which is one classifier [8], [11], [12]. Alternatively, studies involving an
examination of different feature types are concerned with the
problem of their selection from multichannel EEG recordings,
Manuscript received April 15, 2007; revised July 25, 2007; accepted Feb-
ruary 16, 2008. First published June 3, 2008; last published August 13, 2008
which does not provide much insight into the relation between
(projected). The work of P. Herman was supported by the Vice-Chancellor’s the nature of the given t-f signal representation and discrimi-
Research Scholarship at the University of Ulster. native properties of the features generated [7], [13]. Therefore,
The authors are with the Intelligent Systems Research Centre, School of Com- this paper is aimed at a more systematic analysis of different
puting and Intelligent Systems, University of Ulster, Derry, BT48 7JL, U.K.
(e-mail: [email protected]; [email protected]). approaches to quantifying the frequency content of the EEG trials
Digital Object Identifier 10.1109/TNSRE.2008.926694 recorded for BCI purposes. They are categorized into four groups:
1534-4320/$25.00 © 2008 IEEE
318 IEEE TRANSACTIONS ON NEURAL SYSTEMS AND REHABILITATION ENGINEERING, VOL. 16, NO. 4, AUGUST 2008
to subject [3], [8]. The resultant characteristic changes in the function of the signal or using Welch’s approach
EEG spectral content around (8–12 Hz) and central (18–25 , referred to as the modified periodogram [20],
Hz) ranges depend on the associated neural structure and the consisting in averaging multiple spectral estimates for shorter
locations where the neurons are active during MI. The experi- data segments and windowing. The average power in a given
ments showed that for the eleven subjects examined, there was frequency band was obtained from a rectangle approximation
ERD of the rhythm on the contralateral side and a slight ERS of the integral of the signal’s PSD. This led to the calculation
in the central rhythm on the ipsilateral hemisphere. This hemi- of in (2).
spheric asymmetry reflected in the EEGs is exploited to differ- 2) Atomic Decompositions: A t-f representation produced
entiate between the MIs. by methods falling into this category has the form of a linear
The quality of the spectral representations examined in this composition of elementary components referred to as atoms. In
paper is assessed in terms of distinctive properties of the fea- this work, two techniques were applied, the discrete short-time
tures obtained from windowed EEG signals. To this end, the FT (STFT) and the S transform. The major limitation of the
EEG trials of length were divided into windows along the STFT is a trade-off between time and frequency resolutions de-
time axis, depending on the settings of two parameters: window pending on the size of the window along the time axis
size, , and the amount of overlap, . The number of [1]. Here, a Gaussian window was applied. Since it is kept con-
windows could be calculated, as shown below stant throughout the signal analysis, the balance between time
and frequency localization capabilities should be determined
(1) beforehand.
The S transform combines the separate strengths of the STFT
Next, relevant portions of information corresponding to spectral and wavelet transform [21]. First, the signal is decomposed into
correlates of ERD/ERS phenomena were extracted separately t-f atoms and second, since the analysis window changes with
from each EEG window within a trial using the techniques ex- frequency it can provide satisfactory resolution in both domains
amined in this paper. The frequency components related to ERD of interest allowing for better time localization at higher fre-
and ERS were merged together to reduce the dimensionality of quency components. The discrete formula of the S transform is
the feature space (as explained in Sections II-B1–II-B5). Closer the following:
analysis performed at initial stages using separate correlates of
ERD and ERS revealed instability and variability of ERS pat- (3)
terns in frequency bands. Therefore, using each feature as a
separate dimension of the feature space was abandoned. As a where is a sampled signal (with samples) and
result, a feature vector representing a BCI trial was composed is a discrete form of the window . It is analogous
of elements given two recording to that of the STFT with the difference in the form of the
channels, Gaussian window , which is additionally parameterized with
the oscillations frequency
(2)
(4)
The most reactive frequency bands from which to extract fea-
tures for the given subjects were tuned by optimizing the CA The concepts of wavelet and STFT analysis in S transform are
evaluated with an LDA classifier in a five-fold cross-validation incorporated by separating the kernel into two parts: the slowly
(CV) procedure. varying frame and the oscillatory exponential element which
1) Spectral Estimation Methods: Spectral density methods chooses frequency bins being localized as in the STFT [21].
extract information from a signal as a stochastic process to Unlike the frame , the oscillatory component is not translated
describe the distribution of its power in the frequency domain. over the time axis, which allows for independent localization of
The PSD is defined as the Fourier transform (FT) of the signal’s the phase and the amplitude spectrum. The implementation of
autocorrelation function provided that the signal is stationary the decomposition methods was done using algorithms based on
in a wide sense [20]. In practice, a comprehensive statistical the fast FT (FFT) [2].
characteristic of the random signal is not available and its In order to determine the feature vector components, s in
spectral content can only be estimated from a sequence of time (2), the amplitudes of the atoms representing an EEG window
samples. Parametric approaches work on the assumption that were time averaged over its length, , and then the square
the underlying process can be described parametrically as an norm of these mean values within the frequency bands of
output of a linear system driven by white noise. In this work, interest (adjusted and ) was calculated (cf. Fig. 2).
the coefficients of AR model were estimated from segments 3) Quadratic Energy Distributions: Unlike the linear atomic
of EEG time series using the Yule–Walker algorithm [20] and decomposition that deals with elementary t-f components, the
the corresponding PSD was then derived from its frequency energy-based approach consists in distributing the energy of the
response. This parametric spectral estimation technique is signal in the joint t-f domain. The Wigner–Ville distribution
referred to throughout the paper as . (WVD) served as a starting point in this research. It belongs,
The two other methods allowed for nonparametric estima- like other transforms studied here, to the Cohen’s class of the
tion of the signal’s PSD from its periodogram [20]. It was quadratic t-f representations based on second-order moments of
either calculated as the FT of the windowed autocorrelation the signal, covariant by shifts in time and frequency [2], [22].
320 IEEE TRANSACTIONS ON NEURAL SYSTEMS AND REHABILITATION ENGINEERING, VOL. 16, NO. 4, AUGUST 2008
TABLE I TABLE II
LIST OF SPECTRAL TECHNIQUES USED IN FEATURE EXTRACTION LIST OF THE CLASSIFIERS APPLIED IN THE STUDY
One of the strengths of SVMs exploited in this work is the than those originally selected by the GAs, which confirmed the
fact that they are kernel machines, thus allowing for straight- desired optimality of the GA solutions in the given regions of
forward transformation into nonlinear classifiers [16]. A non- the SVM parameter space.
linear kernel function is implicitly utilized to map the original
input space to a higher dimensional feature space in which the
problem is more likely to be linearly separable. Two nonlinear III. RESULTS
kernels were employed—Gaussian and polynomial [16]. How- The process of parameter selection for each of the feature
ever, due to the poor performance of SVMs with the polyno- extraction methods was conducted first. The most reactive fre-
mial mapping when applied to the MI classification, only the quency bands, a setup of the windowing procedure ( ,
following Gaussian kernel function was used: ) and a set of adjustable parameters of the spectral techniques
listed in Table I were optimized using a ten-run five-fold CV
(8) procedure on Session I data with the average CA as a perfor-
mance measure. A preliminary study showed consistent trends
It is emphasized that the kernel is inhomogeneous if the of these parameter setups across different classifiers, which al-
components of the -dimensional vectors, lowed for further tuning to maximize the interclass feature sep-
and , have different scaling factors arability quantified using LDA only.
. Next, the overall separability of the EEG features was eval-
Selecting an optimal parameter set poses a considerable dif- uated on Session I data in terms of the discrimination error es-
ficulty but is crucial for effective utilization of SVMs. An SVM timate made using the inner–outer CV scheme (five-fold splits
classifier is controlled by multiple parameters such as the regu- with multiple runs) with all the classifiers under consideration.
larization constant and kernel parameters s, which need to To this end, Session I data was divided into five outer folds and
be carefully chosen to minimize the generalization error [27]. the classifiers’ parameters were optimized in the GA framework
In this work, it was estimated using a CV scheme and a theo- only on the training set, inner fold (again five-fold inner CV
retical radius-margin bound [30]. The search for an optimal set split), of each outer split. In addition, both the outer and inner
of the parameters was performed with a genetic algorithm (GA) five-fold splits were performed in multiple runs after random
due to its ability to efficiently solve optimization problems given data shuffling. The final overall CA rate was obtained as the
vast, multidimensional and highly irregular search spaces [31]. outer-test-fold average across multiple runs.
A roulette wheel selector along with a two-point crossover (with Analogously to the selection procedure used for spectral pa-
probability rate ) and binary mutation rameters, another ten-run five-fold CV procedure was utilized
operators were utilized. Through a smaller-scale pilot study em- (without inner–outer fold splits) within the framework of the
ploying micro-GAs, this setup was found to deliver satisfactory GA to identify the parameter setup of the SVM/RFD classi-
performance in terms of computational time and the quality of fiers for use in the second experiment—a single-pass test on
the resultant SVM classifier. Moreover, an elitist strategy was Session II data. This experiment facilitated verification of the
applied with a replacement rate of (for the population session-to-session generalization performance. It is crucial for
size of ). The GA search was halted after BCI systems since the dynamics of EEG recordings acquired
generations due to a considerable drop in the convergence rate. at distant times (different sessions) is known to vary consider-
Two variants of the objective function to be minimized were ably. These variations degrade the intersession performance of
used. The first was based solely on the average rate of the clas- BCIs [32]. A key question is to what extent the spectral methods
sification error evaluated using multiple-run (here, 10) five-fold applied in this work capture the most salient and persistent (ses-
CV whereas the second was defined as a weighted sum of this sion-to-session) characteristics of EEGs in the context of MI
CV error estimate and the theoretical upper bound on the gener- classification. Generalization capabilities of the classifiers play
alization error referred to as a radius-margin bound. The out- an important role in this regard also. The test of session-to-ses-
come of their application is further discussed in Section IV. sion performance was conducted using the unseen Session II
An extensive examination of the stochastic nature of the GA data. To this end, optimal configurations of the parameters iden-
optimization procedure facilitated the observation that ten ran- tified using CV on Session I data were kept unchanged. The clas-
domly initiated runs of the algorithm resulted in low variance sifiers were then trained on the features extracted from Session
average values of both objective functions applied in the exper- I and tested on those from Session II.
iments. Moreover, after ten runs of the GA search scheme it be- The results of this multisubject study evaluated using the
came clear that the regions of the solution space explored by inner–outer CV over Session I and the single-pass test on
the algorithm at final stages of the optimization process started Session II are presented in Fig. 3(a) and (b), respectively. The
to overlap to a considerable degree. Therefore, GA-based algo- CA rates were averaged over eleven subjects for every tech-
rithm was run ten times to identify the desirable parameter setup nique and classifier type. For SVMs, the classifiers performing
of an SVM classifier for each given feature set. In addition, gra- best on Session I in each category, linear and nonlinear, were
dient-based minimization of the estimate of the generalization considered as and . It was noted that the
error bound [30] was conducted for the resultant SVM designs Session I performance estimated with the inner–outer CV
following GA search to enhance local optimization capabilities scheme [Fig. 3(a)] showed similar trends to the mean CAs
of the algorithm. It did not lead however to finding better so- obtained from the ordinary ten-run five-fold CV during the
lutions in terms of the CV estimate of the generalization error selection of the classifier’s parameters (not reported here).
HERMAN et al.: COMPARATIVE ANALYSIS OF SPECTRAL APPROACHES TO FEATURE EXTRACTION 323
Fig. 3. (a) Mean CA rates obtained on Session I data using the inner–outer Fig. 4. Comparative analysis of the categories of frequency techniques in terms
CV scheme (five-fold splits with multiple runs) and averaged over 11 subjects. of the average CA rates obtained over 11 subjects with different classifiers: (a)
Vertical lines denote the intersubject standard deviations of the respective mean on session I data (inner–outer CV scheme) and (b) on session II data (single-pass
CA values. (b) Average CA rates obtained in the single-pass test on Session II test). Vertical lines denote the intersubject standard deviations of the respective
data for 11 subjects. Vertical lines denote the intersubject standard deviations of mean CA values.
the respective mean CA values.
TABLE III
Fig. 4(a) and (b) depicts the average classification perfor- LIST OF THE AVERAGE AND BEST CA RATES OBTAINED FOR EACH SUBJECT
mance associated with different categories of frequency tech- ACROSS ALL THE CLASSIFIERS IN SESSIONS I AND II. OPTIMAL FEATURE
TYPES CORRESPONDING TO THE BEST RESULTS ARE PRESENTED IN
niques: atomic decompositions, WV energy distributions, RIDs, COLUMNS “FEATURE” FOR BOTH SESSIONS
PSD estimators, t-s approaches, and AR modeling. The resultant
mean value for each category shown in the form of a bar was cal-
culated for every subject as the average CA rate obtained with
all the classifiers and the category-specific features. Again, the
mean and the standard deviation across the subjects are shown
for Session I and Session II. In addition, Table III lists the best
and average CA rates obtained for each subject across all the
classifiers. It illustrates intersubject differences in performance
on Session I and Session II data.
A comparative analysis of the performance of the classifiers
employed in this work, including all types of SVMs listed in
Table II, was also conducted. The corresponding CAs were av-
eraged over the spectral methods and their means obtained for
eleven subjects are presented along with the standard deviations
in Fig. 5 for Session I and Session II data.
IV. DISCUSSION
The results of this multisubject study were processed in-
dependently for two sessions of the data in the framework
of two-way analysis of variance (ANOVA) with repeated
measures [33]. This approach facilitated effective accounting factors were frequency technique (14 types) and classifier,
for rather high intersubject variance. The two within-subject represented by LDA, RFD, and two collective categories of
324 IEEE TRANSACTIONS ON NEURAL SYSTEMS AND REHABILITATION ENGINEERING, VOL. 16, NO. 4, AUGUST 2008
trial should serve then as the end point of the feature extraction able for BCI. The combination of different EEG datasets with a
window. At the same time, this approach is associated with a range of motor imagery dynamics from beginning to end of trial
delayed response of the BCI system to the changes in the spec- obtained during MI-based BCI experiments, undertaken by sub-
tral content of the ongoing EEG activity. The window size of up jects with varying levels of prior experience and capabilities in
to 2 s, used in this work, allows for balancing the time and the using EEG-based communication devices, enabled the produc-
frequency resolution of analysis of the spectral phenomena rele- tion of more general conclusions at the cost of relatively lower
vant to MI related EEG pattern discrimination. In consequence, average performance and higher intersubject variance. The ex-
the time delay of the causal feature extractor mentioned above tensive complementary online analysis, within the framework of
is considered to be manageable and acceptable in online BCI causal feature extraction with a moving window approach and
operation. It should be emphasized that the proposed transfor- continuous classification, is intended for further work.
mation enforcing causality still requires time efficient imple-
mentation of the spectral techniques. Sample-by-sample clas- ACKNOWLEDGMENT
sification in online mode within the framework discussed here The authors would like to thank the reviewers and the Asso-
implies that the spectral quantification of an EEG trial within ciate Editor for their constructive comments and suggestions.
a moving window has to be performed faster than the sam-
pling frequency, . Preliminary simulations conducted in the REFERENCES
ISRC laboratory setup (with Pentium IV 3 GHz, 512 MB RAM) [1] , E. Niedermeyer and F. Lopes da Silva, Eds., Electroencephalography:
Basic Principles, Clinical Applications and Related Fields, 4th ed.
demonstrated the feasibility of applying the spectral methods, Baltimore, MD: Williams & Wilkins, 1998.
including quadratic energy distributions with their highest com- [2] , M. Akay, Ed., Time Frequency and Wavelets in Biomedical Signal
putational complexity, in online EEG-based BCI systems with Processing, ser. Biomedical Engineering. Piscataway, NJ: IEEE
Press, 1997.
real-time constraints ( for moving windows of [3] G. Pfurtscheller, C. Neuper, D. Flotzinger, and M. Pregenzer,
equal or less than 2 s duration, which corresponds to the max- “EEG-based discrimination between imagination of right and left
hand movement,” Electroencephalogr. Clin. Neurophysiol., vol. 103,
imum of 256 data points processed within 1/128 s). pp. 642–651, 1997.
[4] J. R. Wolpaw, N. Birbaumer, D. J. McFarland, G. Pfurtscheller, and T.
V. CONCLUSION AND FUTURE WORK M. Vaughan, “Brain-computer interfaces for communication and con-
trol,” Clin. Neurophysiol., vol. 113, pp. 767–791, 2002.
This paper has investigated various approaches to quantifying [5] J. R. Wolpaw, D. J. McFarland, G. W. Neat, and C. A. Forneris, “An
the relevant spectral content of the EEG utilized for MI classi- eeg-based brain-computer interface for cursor control,” Electroen-
cephalogr. Clin. Neurophysiol., vol. 78, pp. 252–259, 1991.
fication. The separability of the EEG feature sets representing [6] J. del R. Millan, M. Franze, J. Mourino, F. Cincotti, and F. Babiloni,
entire BCI trials has been evaluated offline. Special attention “Relevant EEG features for the classification of spontaneous motor-
has been paid to session-to-session generalization properties of related tasks,” Biol. Cybern., vol. 86, pp. 89–95, 2002.
[7] E. Yom-Tov and G. F. Inbar, “Feature selection for the classification
the classifiers and the feature sets. of movements from single movement-related potentials,” IEEE Trans.
A detailed ranking of the most effective spectral methods in Neural Syst. Rehab. Eng., vol. 10, no. 3, pp. 170–177, Sep. 2002.
[8] G. Pfurtscheller et al., “Current trends in graz brain-computer interface
terms of the resultant CA rates is rather subject-specific. There (BCI) research,” IEEE Trans. Rehab. Eng., vol. 8, no. 2, pp. 216–219,
are a few cases of techniques achieving good performance in Jun. 2000.
one subject and poor performance in the other one. Similarly, [9] B. H. Yang, G. Z. Yan, R. G. Yan, and T. Wu, “Adaptive subject-based
feature extraction in brain-computer interfaces using wavelet packet
high CA rates in one session do not necessarily correspond with best basis decomposition,” Med. Eng. Phys., vol. 29, no. 1, pp. 48–53,
satisfactory results in the other session. Close observation of the 2007.
trends across all eleven subjects has demonstrated the superior [10] S. Lemm, C. Schafer, and G. Curio, “BCI competition 2003—data set
III: Probabilistic modeling of sensorimotor rhythms for classification
robustness of PSD approaches consistent with all the classifiers of imaginary hand movements,” IEEE Trans. Biomed. Eng., vol. 51, no.
used in one- and two-session classification tasks. The most in- 6, pp. 1077–1080, Jun. 2004.
[11] D. Garrett, D. A. Peterson, C. W. Anderson, and M. H. Thaut, “Com-
ferior session-to-session stability was produced by the parison of linear, nonlinear, and feature selection methods for EEG
approach. signal classification,” IEEE Trans. Neural Syst. Rehabil. Eng., vol. 11,
With regard to classification methods, the study has revealed no. 2, pp. 141–144, Jun. 2003.
[12] B. Obermaier, C. Guger, C. Neuper, and G. Pfurtscheller, “Hidden
the superiority of SVMs with Gaussian kernel markov models for online classification of single trial EEG data,” Pat-
in Session I experiments for the majority of the feature types. tern Recognit. Lett., vol. 22, pp. 1299–1309, 2001.
Clearly poorer performance is observed only for the [13] I. Kalatzis et al., “Design and implementation of an SVM-based com-
puter classification system for discriminating depressive patients from
approach. In session-to-session experiments, the selection of healthy controls using the P600 component of ERP signals,” Comput.
an optimal classifier that robustly maintains good performance Methods Programs Biomed., vol. 75, pp. 11–22, 2004.
[14] C. Bishop, Neural Networks For Pattern Recognition. Oxford, U.K.:
from one session to the other one is more subject and feature Oxford Univ. Press, 1995.
type dependent. Generally, the consistent effectiveness of linear [15] K.-R. Muller, M. Krauledat, G. Dornhege, G. Curio, and B. Blankertz,
approaches investigated in this work, - , - , “Machine learning techniques for brain-computer interfaces,” Biomedi-
zinische Technik, vol. 49, no. 1, pp. 11–22, 2004.
and RFD in particular, is emphasized. [16] J. Shawe-Taylor and N. Cristianini, Kernel Methods For Pattern Anal-
Although the concept of discrete classification performed at ysis. Cambridge, U.K.: Cambridge Univ. Press, 2004.
[17] N. F. Ince, S. Arica, and A. Tewfik, “Classification of single trial motor
the end of each trial is less practical in real-world BCI applica- imagery EEG recordings with subject adapted non-dyadic arbitrary
tions than a continuous approach, this thorough study still gives time-frequency tilings,” J. Neural Eng., vol. 3, pp. 235–244, 2006.
a valuable insight into the separability of MI induced EEG pat- [18] C. Neuper, G. R. Muller, A. Kubler, N. Birbaumer, and G. Pfurtscheller,
“Clinical application of an eeg-based brain-computer interface: A case
terns, in the same spirit as in [9], [17], and [18]. It also offers study in a patient with severe motor impairment,” Clin. Neurophysiol.,
some informed guidance as to which techniques are more suit- vol. 114, pp. 399–409, 2003.
326 IEEE TRANSACTIONS ON NEURAL SYSTEMS AND REHABILITATION ENGINEERING, VOL. 16, NO. 4, AUGUST 2008
[19] E. Haselsteiner and G. Pfurtscheller, “Using time-dependent neural Girijesh Prasad (M’98–SM’07) received the
networks for EEG classification,” IEEE Trans. Neural Syst. Rehabil. B.Tech. degree in electrical engineering, the M.Tech.
Eng., vol. 8, no. 4, pp. 457–462, Dec. 2000. degree in computer science and technology, and
[20] P. Stoica and R. Moses, Introductions to Spectral Analysis. Engle- the Ph.D. degree from Queen’s University Belfast,
wood Cliffs, NJ: Prentice-Hall, 1997. Belfast, U.K.
[21] R. G. Stockwell, L. Mansinha, and R. P. Lowe, “Localization of the He is a Senior Lecturer within the Faculty of
complex spectrum: The S transform,” IEEE Trans. Signal Process., vol. Computing and Engineering in the University of
44, no. 4, pp. 998–1001, Apr. 1996. Ulster, Ulster, U.K. His research interests include
[22] F. Auger, P. Flandrin, P. Goncalves, and O. Lemoine, Time-Frequency self-organizing hybrid intelligent systems, neural
toolbox CNRS Research Group, Rice University, 1996. computation, fuzzy logic, evolutionary algorithms,
[23] B. Boashash, , S. Haykin, Ed., “Time-frequency signal analysis,” in predictive modeling, and control with applications in
Advances in Spectrum Analysis and Array Processing 1. Englewood industrial and biological systems including brain–computer interface (BCI) and
Cliffs, NJ: Prentice-Hall, 1991, pp. 418–517. robotic systems. He has published over 70 papers in journals and conference
[24] S. Haykin, Adaptive Filter Theory, 2nd ed. Englewood Cliffs, NJ: proceedings.
Prentice-Hall, 1991.
Dr. Prasad is a Chartered Engineer and a member of the IET.
[25] M. T. Mitchell, Mach Learn. New York: McGraw-Hill, 1997.
[26] K.-R. Muller, C. W. Anderson, and G. E. Birch, “Linear and nonlinear
methods for brain-computer interfaces,” IEEE Trans. Neural. Syst. Re-
habil. Eng., vol. 11, no. 2, pp. 165–169, Jun. 2003.
[27] V. N. Vapnik, The Nature of Statitical Learning Theory. New York: Thomas Martin McGinnity (M’82) received the
Springer-Verlag, 1995. first class honours degree in physics and the Ph.D.
[28] S. Canu, Y. Grandvalet, V. Guigue, and A. Rakotomamonjy, degree from the University of Durham, Durham,
Perception Systèmes et Information, INSA de Rouen, SVM and U.K.
Kernel methods Matlab toolbox France, 2005 [Online]. Available: He is Professor of Intelligent Systems Engineering
http://www.asi.insa-rouen.fr/~arakotom/toolbox/index.html within the Faculty of Computing and Engineering,
[29] J. A. K. Suykens, T. Van Gestel, J. De Brabanter, B. De Moor, and the University of Ulster, Ulster, U.K. He has 28 years
J. Vandewalle, Least Squares Support Vector Machines. Singapore: experience in teaching and research in electronic and
World Scientific, 2002. computer engineering and is currently Director of the
[30] O. Chapelle, V. Vapnik, O. Bousqet, and S. Mukherjee, “Choosing mul- Intelligent Systems Research Centre. His current re-
tiple parameters for support vector machines,” Mach. Learn., vol. 46, search interests are focussed on computational intel-
no. 1, pp. 131–159, 2002. ligence, hardware and software implementations of biologically plausible ar-
[31] D. E. Goldberg , Genetic Algorithms For Search, Optimization, and tificial neural networks, brain computer interfacing, and intelligent systems in
Machine Learning. Reading, MA: Addison-Wesley, 1989. robotics.
[32] P. Shenoy, M. Krauledat, B. Blankertz, R. P. N. Rao, and K.-R. Müller, Prof. McGinnity is a Fellow of the IET and a Chartered Engineer.
“Towards adaptive classification for BCI,” J. Neural Eng., vol. 3, pp.
13–23, 2006.
[33] Y. Hochberg and A. C. Tamhane, Multiple Comparison Procedures.
New York: Wiley, 1987.
[34] Event-Related Dynamics of Brain Oscillations, C. Neuper and W. Damien Coyle (M’05) received the first class degree
Klimesch, Eds. Amsterdam, The Netherlands: Elsevier, 2006, vol. in computing and electronic engineering and the
159, Progress in Brain Research. Ph.D. degree in intelligent systems engineering from
the University of Ulster, Ulster, U.K., in 2002 and
2006, respectively.
He is currently a Lecturer and a member of the
Pawel Herman (M’04) received the M.Sc. Eng. de- Intelligent Systems Research Centre. His research
gree in applied computer science in medicine and en- interests include biosignal processing, bio-inspired
gineering from Wroclaw University of Technology, cognitive and adaptive systems, and brain–computer
Wroclaw, Poland, in 2002. He is currently working interface technology.
toward the Ph.D. degree in intelligent systems at the Dr. Coyle is a member of the IET and the Isaac
University of Ulster, Derry, U.K. Newton Institute for Mathematical Sciences. He is a member of the executive
He is a member of the Intelligent Systems Re- committees of the IEEE Computation Intelligence Society (CIS) and the IEEE
search Centre as a Research Associate in Brain- Engineering in Medicine and Biology Society (EMBS) and Vice Chair of the
Computer Interfaces. His research interests en- IEEE CIS GOLD subcommittee and the CIS Awards subcommittee.
compass biosignal analysis, machine learning,
computational intelligence methods with emphasis
on hybrid adaptive systems and type-2 fuzzy logic systems, and a range of
problems in cognitive science mainly in the context of neurofeedback treatment.