Abstract: A brain-computer interface (BCI) is a hardware and software communications system that permits cerebral activity alone to control computers or external devices. The immediate goal of BCI research is to provide communications capabilities to severely disabled people who are totally paralyzed or ‘locked in’ by neurological neuromuscular disorders, such as amyotrophic lateral sclerosis, brain stem stroke, or spinal cord injury. Here, we review the state-of-the-art of BCIs, looking at the different steps that form a standard BCI: signal acquisition, preprocessing or signal enhancement, feature extraction, classification and the control interface. We discuss their advantages, drawbacks, and latest advances, and we survey the numerous technologies reported in the scientific literature to design each step of a BCI. First, the review examines the neuroimaging modalities used in the signal acquisition step, each of which monitors a different functional brain activity such as electrical, magnetic or metabolic activity. Second, the review discusses different electrophysiological control signals that determine user intentions, which can be detected in brain activity. Third, the review includes some techniques used in the signal enhancement step to deal with the artifacts in the control signals and improve the performance. Fourth, the review studies some mathematic algorithms used in the feature extraction and classification steps which translate the information in the control signals into commands that operate a computer or other device. Finally, the review provides an overview of various BCI applications that control a range of devices. Keywords: brain-computer interface (BCI); electroencephalography (EEG); rehabilitation; artifact; neuroimaging; brain-machine interface; collaborative sensor system OPEN ACCESS Sensors 2012, 12 1212 1. Introduction A brain computer interface (BCI), also referred to as a brain machine interface (BMI), is a hardware and software communications system that enables humans to interact with their surroundings, without the involvement of peripheral nerves and muscles, by using control signals generated from electroencephalographic activity. BCI creates a new non-muscular channel for relaying a person’s intentions to external devices such as computers, speech synthesizers, assistive appliances, and neural prostheses. That is particularly attractive for individuals with severe motor disabilities. Such an interface would improve their quality of life and would, at the same time, reduce the cost of intensive care. A BCI is an artificial intelligence system that can recognize a certain set of patterns in brain signals following five consecutive stages: signal acquisition, preprocessing or signal enhancement, feature extraction, classification, and the control interface [1]. The signal acquisition stage captures the brain signals and may also perform noise reduction and artifact processing. The preprocessing stage prepares the signals in a suitable form for further processing. The feature extraction stage identifies discriminative information in the brain signals that have been recorded. Once measured, the signal is mapped onto a vector containing effective and discriminant features from the observed signals. The extraction of this interesting information is a very challenging task. Brain signals are mixed with other signals coming from a finite set of brain activities that overlap in both time and space. Moreover, the signal is not usually stationary and may also be distorted by artifacts such as electromyography (EMG) or electrooculography (EOG). The feature vector must also be of a low dimension, in order to reduce feature extraction stage complexity, but without relevant information loss. The classification stage classifies the signals taking the feature vectors into account. The choice of good discriminative features is therefore essential to achieve effective pattern recognition, in order to decipher the user’s intentions. Finally the control interface stage translates the classified signals into meaningful commands for any connected device, such as a wheelchair or a computer. BCI technology has traditionally been unattractive for serious scientific investigation. The idea of successfully deciphering thoughts or intentions by means of brain activity has often been rejected in the past as very strange and remote. Hence investigation in the field of brain activity has usually been limited to the analysis of neurological disorders in the clinic or to the exploration of brain functions in the laboratory. The BCI design was considered too complex, because of the limited resolution and reliability of information that was detectable in the brain and its high variability. Furthermore, BCI systems require real-time signal processing, and up until recently the requisite technology either did not exist or was extremely expensive [2]. However, this context has undergone radical change over the last two decades. BCI research, which was confined to only three groups 20 years ago and only six to eight groups 10 years ago, is now a flourishing field with more than 100 active research groups all over the World studying the topic [3]. The number of articles published regarding neural interface technology has increased exponentially over the past decade [4]. Successful studies on brain signal phenomena have lent further weight to these advances. The development of more and more inexpensive computer hardware and software has allowed more sophisticated online analysis. Likewise, the chances of using BCIs as auxiliary technology that might serve severely disabled people has increased social acceptance in this field and the need to accelerate its progress. Interest in this technology is now found outside of the laboratory or Sensors 2012, 12 1213 the clinic. Small specialized companies such as Emotiv [5] or Neurosky [6] have already developed some initial applications oriented towards the general public. Nevertheless, despite these advances, most BCI-based applications are still limited to the laboratory. Broader applicability of BCIs requires greater ease of use, which in turn means reducing time spent on preparation, training and calibration [7]. BCI research is a relatively young multidisciplinary field integrating researchers from neuroscience, physiology, psychology, engineering, computer science, rehabilitation, and other technical and health-care disciplines. As a result, in spite of some notable advances, a common language has yet to emerge, and existing BCI technologies vary, which makes their comparison difficult and, in consequence, slows down the research. The community of BCI researchers has therefore stressed the need to establish a general framework for BCI design [8]. Mason et al. [9], for example, proposed a new functional model for BCI systems and taxonomy design. This review of the state-of-the-art of BCI systems is arranged as follows: Section 2 discusses existing neuroimaging approaches to BCIs and Section 3 describes the most commonly found control signals in BCI systems. Section 4 briefly explains certain types of BCIs. Sections 5, 6 and 7, respectively, cover the different signal processing methods used for feature extraction, artifact reduction and feature classification. Section 8 provides an overview of BCI applications and, finally, the conclusions are drawn in Section 9. 2. Neuroimaging Approaches in BCIs BCIs use brain signals to gather information on user intentions. To that effect, BCIs rely on a recording stage that measures brain activity and translates the information into tractable electrical signals. Two types of brain activities may be monitored: (i) electrophysiological and (ii) hemodynamic. Electrophysiological activity is generated by electro-chemical transmitters exchanging information between the neurons. The neurons generate ionic currents which flow within and across neuronal assemblies. The large variety of current pathways can be simplified as a dipole conducting current from a source to a sink through the dendritic trunk. These intracellular currents are known as primary currents. Conservation of electric charges means that the primary currents are enclosed by extracellular current flows, which are known as secondary currents [10]. Electrophysiological activity is measured by electroencephalography, electrocorticography, magnetoencephalography, and electrical signal acquisition in single neurons. The hemodynamic response is a process in which the blood releases glucose to active neurons at a greater rate than in the area of inactive neurons. The glucose and oxygen delivered through the blood stream results in a surplus of oxyhemoglobin in the veins of the active area, and in a distinguishable change of the local ratio of oxyhemoglobin to deoxyhemoglobin [11]. These changes can be quantified by neuroimaging methods such as functional magnetic resonance and near infrared spectroscopy. These kinds of methods are categorized as indirect, because they measure the hemodynamic response, which, in contrast to electrophysiological activity, is not directly related to neuronal activity. Most current BCIs obtain the relevant information from the brain activity through electroencephalography. Electroencephalography is by far the most widely used neuroimaging modality, owing to its high temporal resolution, relative low cost, high portability, and few risks to the users. BCIs based on electroencephalography consist of a set of sensors that acquire electroencephalography signals Sensors 2012, 12 1214 from different brain areas. However, the quality of electroencephalography signals is affected by scalp, skull, and many other layers as well as background noise. Noise is key to electroencephalography and to other neuroimaging methods, insofar as it reduces the SNR and therefore the ability to extract meaningful information from the recorded signals. Non-invasive approaches have successfully been used by severely and partially paralyzed patients to reacquire basic forms of communication and to control neuroprostheses and wheelchairs [12–14]. Despite the outstanding utility of non-invasive approaches in BCI applications, motor recovery has been limited, because of the need for brain signals with a higher resolution. Invasive recording methods such as electrocorticography or intracortical neuron recording were introduced, in an effort to improve the quality of brain signals monitored by BCIs. Most researchers agree that movement restoration through prostheses with multiples degrees of freedom can only be achieved through invasive approaches [15]. It is unlikely that the power of non-invasive modalities will be enhanced in the near future. Accordingly, it would appear that invasive modalities are indispensable for accurate neuroprostheses control. Nevertheless, this issue is not yet entirely clear and some opinions disagree with this conjecture. Contrary to established opinion, Wolpaw [3] suggested that performance in multidimensional control may be independent of the recording method. Further refinements of recording and analysis techniques will probably increase the performance of both invasive and non-invasive modalities. However, the latest studies in neuroprostheses control appear to indicate that invasive modalities have inherent advantages in neuroprosthesis control applications [4]. Invasive modalities need to implant microelectrode arrays inside the skull that involves significant health risks, which restricts their use to experimental settings. Two invasive modalities can be found in BCI research: electrocorticography, which places electrodes on the surface of the cortex, either outside the dura mater (epidural electrocorticography) or under the dura mater (subdural electrocorticography), and intracortical neuron recording which implants electrodes inside the cortex. Several issues had to be addressed, before they become suitable for long-term applications. First, tissue acceptance of the microelectrode has to be addressed, for which reason proposals exist for electrodes with neurotropic mediums that promote neuronal growth to improve biocompatibility [16]. Perhaps, the future of nanotechnologies that might develop nano-detectors to be implanted inertly in the brain, may provide a definite solution to the problems of long-term invasive applications. Second, a link between the microelectrode and external hardware that uses wireless technology is needed to reduce the risks of infection. Wireless transmission of neuronal signals has already been tested in animals [17]. And third, continuous stress caused by plugging and unplugging the recording system may lead to tissue damage or system failure. Each neuroimaging modality is explained below. Firstly, electrophysiological methods such as electroencephalography, electrocorticography, magnetoencephalography, and electrical signal acquisition in single neurons will be discussed. Secondly, metabolic methods such as functional magnetic resonance and near infrared spectroscopy will be described. Finally, functional imaging modalities are listed in Table 1, along with information related to activity measured, temporal and spatial resolutions, safety, and portability. Sensors 2012, 12 1215 Table 1. Summary of neuroimaging methods. Neuroimaging method Activity measured Direct/ Indirect Measurement Temporal resolution Spatial resolution Risk Portability EEG Electrical Direct ~0.05 s ~10 mm Non-invasive Portable MEG Magnetic Direct ~0.05 s ~5 mm Non-invasive Non-portable ECoG Electrical Direct ~0.003 s ~1 mm Invasive Portable Intracortical neuron recording Electrical Direct ~0.003 s ~0.5 mm (LFP) ~0.1 mm (MUA) ~0.05 mm (SUA) Invasive Portable fMRI Metabolic Indirect ~1 s ~1 mm Non-invasive Non-portable NIRS Metabolic Indirect ~1 s ~5 mm Non-invasive Portable 2.1. Electroencephalography (EEG) EEG measures electric brain activity caused by the flow of electric currents during synaptic excitations of the dendrites in the neurons and is extremely sensitive to the effects of secondary currents [10]. EEG signals are easily recorded in a non-invasive manner through electrodes placed on the scalp, for which that reason it is by far the most widespread recording modality. However, it provides very poor quality signals as the signals have to cross the scalp, skull, and many other layers. This means that EEG signals in the electrodes are weak, hard to acquire and of poor quality. This technique is moreover severely affected by background noise generated either inside the brain or externally over the scalp. The EEG recording system consists of electrodes, amplifiers, A/D converter, and a recording device. The electrodes acquire the signal from the scalp, the amplifiers process the analog signal to enlarge the amplitude of the EEG signals so that the A/D converter can digitalize the signal in a more accurate way. Finally, the recording device, which may be a personal computer or similar, stores, and displays the data. The EEG signal is measured as the potential difference over time between signal or active electrode and reference electrode. An extra third electrode, known as the ground electrode, is used to measure the differential voltage between the active and the reference points. The minimal configuration for EEG measurement therefore consists of one active, one reference, and one ground electrode. Multi-channel configurations can comprise up to 128 or 256 active electrodes [18]. These electrodes are usually made of silver chloride (AgCl) [19]. Electrode-scalp contact impedance should be between 1 kΩ and 10 kΩ to record an accurate signal [20]. The electrode-tissue interface is not only resistive but also capacitive and it therefore behaves as a low pass filter. The impedance depends on several factors such as the interface layer, electrode surface area, and temperature [20]. EEG gel creates a conductive path between the skin and each electrode that reduces the impedance. Use of the gel is cumbersome, however, as continued maintenance is required to assure a relatively good quality signal. Electrodes that do not need to use of gels, called ‘dry’ electrodes, have been made with other materials such as titanium and stainless-steel [21]. These kinds of electrodes may be ‘dry’ active electrodes, which have preamplification circuits for dealing with very high electrode/skin interfacial Sensors 2012, 12 1216 impedances [21,22], or ‘dry’ passive electrodes, which have no active circuits, but are linked to EEG recording systems with ultra-high input impedance [23]. The amplitude of electrical bio-signals is in the order of microvolts. Consequently, the signal is very sensitive to electronic noise. External sources such power-lines may generate background noise and thermal, shot, flicker, and burst noises are generated by internal sources [24]. Design considerations should be addressed to reduce the effects of the noise, such as electromagnetic interference shielding or reduction for common mode signal, amongst others [20]. EEG comprises a set of signals which may be classified according to their frequency. Well-known frequency ranges have been defined according to distribution over the scalp or biological significance. These frequency bands are referred to as delta (δ), theta (θ), alpha (α), beta (β), and gamma (γ) from low to high, respectively. Relevant characteristics of these bands are detailed below. The delta band lies below 4 Hz, and the amplitude of delta signals detected in babies decreases as they age. Delta rhythms are usually only observed in adults in deep sleep state and are unusual in adults in an awake state. A large amount of delta activity in awake adults is abnormal and is related to neurological diseases [25]. Due to low frequency, it is easy to confuse delta waves with artifact signals, which are caused by the large muscles of the neck or jaw. Theta waves lie within the 4 to 7 Hz range. In a normal awake adult, only a small amount of theta frequencies can be recorded. A larger amount of theta frequencies can be seen in young children, older children, and adults in drowsy, meditative or sleep states [25]. Like delta waves, a large amount of theta activity in awake adults is related to neurological disease [25]. Theta band has been associated with meditative concentration [26,27] and a wide range of cognitive processes such as mental calculation [28], maze task demands [29], or conscious awareness [30]. Alpha rhythms are found over the occipital region in the brain [31]. These waves lie within the 8 to 12 Hz range. Their amplitude increases when the eyes close and the body relaxes and they attenuate when the eyes open and mental effort is made [32]. These rhythms primarily reflect visual processing in the occipital brain region and may also be related to the memory brain function [33]. There is also evidence that alpha activity may be associated with mental effort. Increasing mental effort causes a suppression of alpha activity, particularly from the frontal areas [34]. Consequently, these rhythms might be useful signals to measure mental effort. Mu rhythms may be found in the same range as alpha rhythms, although there are important physiological differences between both. In contrast to alpha rhythms, mu rhythms are strongly connected to motor activities and, in some cases, appear to correlate with beta rhythms [31,35]. Beta rhythms, within the 12 to 30 Hz range, are recorded in the frontal and central regions of the brain and are associated with motor activities. Beta rhythms are desynchronized during real movement or motor imagery [36]. Beta waves are characterized by their symmetrical distribution when there is no motor activity. However, in case of active movement, the beta waves attenuate, and their symmetrical distribution changes [36]. Gamma rhythms belong to the frequency range from 30 to 100 Hz. The presence of gamma waves in the brain activity of a healthy adult is related to certain motor functions or perceptions, among others [37]. Some experiments have revealed a relationship in normal humans between motor activities and gamma waves during maximal muscle contraction [38]. This gamma band coherence is replaced by a beta band coherence during weak contractions, suggesting a correlation between gamma or beta Sensors 2012, 12 1217 cortical oscillatory activity and force [39]. Also, several studies have provided evidence for the role of gamma activity in the perception of both visual and auditory stimuli [37,40–42]. Gamma rhythms are less commonly used in EEG-based BCI systems, because artifacts such as electromyography (EMG) or electrooculography (EOG) are likely to affect them [43]. Nevertheless, this range is attracting growing attention in BCI research because, compared to traditional beta and alpha signals, gamma activity may increase the information transfer rate and offer higher spatial specifity [44,45]. As explained above, EEG is recorded by electrodes. The electrodes placed over the scalp are commonly based on the International 10–20 system [46], which has been standardized by the American Electroencephalographic Society. The 10–20 system uses two reference points in the head to define the electrode location. One of these reference points is the nasion, located at the top of the nose at the same level as the eyes. The other reference point is the inion, which is found in the bony lump at the base of the skull. The transverse and median planes divide the skull from these two points. The electrode locations are determined by marking these planes at intervals of 10% and 20% (Figure 1). The letters in each location corresponds to specific brain regions in such a way that A represents the ear lobe, C the central region, Pg the nasopharyngeal, P the parietal, F the frontal, Fp the frontal polar, and O the occipital area. Figure 1. Electrode placement over scalp. 2.2. Magnetoencephalography (MEG) MEG is a non-invasive imaging technique that registers the brain’s magnetic activity by means of magnetic induction. MEG measures the intracellular currents flowing through dendrites which produce magnetic fields that are measurable outside of the head [47]. The neurophysiological processes that produce MEG signals are identical to those that produce EEG signals. Nevertheless, while EEG is extremely sensitive to secondary current sources, MEG is more sensitive to those of primary 20 % 20 % 20 % 20 % Inion Nasion 10 % 10 % 20 % 20 % 10 % 20 % 20 % 10 % Pg1 Pg2 Fp1 Fp2 F7 Fz F4 F8 T3 C3 Cz C4 T4 T5 P3 Pz P4 T6 A1 A2 O1 O2 F3 C3 T3 T5 P3 Pz Cz F7 F3 Fz Fp1 C3 A1 Pg1 Inion Nasion Sensors 2012, 12 1218 currents [10]. The advantage of MEG is that magnetic fields are less distorted by the skull and scalp than electric fields [48]. Magnetic fields are detected by superconducting quantum interferences devices, which are extremely sensitive to magnetic disturbances produced by neural activity [49]. The electronic equipment that measures magnetic brain activity is cooled to almost −273 degrees Celsius to facilitate sensor superconductivity. MEG requires effective shielding from electromagnetic interferences. The electronic equipment is installed inside a magnetically shielded room, which attenuates the effects of magnetic fields from external sources. MEG provides signals with higher spatiotemporal resolution than EEG, which reduces the training time needed to control a BCI and speeds up reliable communications [50]. MEG has also been successfully used to localize active regions inside the brain [51]. In spite of these advantageous features, MEG is not often used in BCI design because MEG technology is too bulky and expensive to become an acquisition modality suitable for everyday use. In 2005, Lal et al. [52] presented the first online MEG-based BCI. Although further studies have followed [53–57], MEG-based BCIs, as compared to EEG-based BCIs, are still at an early stage. 2.3. Electrocorticography (ECoG) ECoG is a technique that measures electrical activity in the cerebral cortex by means of electrodes placed directly on the surface of the brain. Compared to EEG, ECoG provides higher temporal and spatial resolution as well as higher amplitudes and a lower vulnerability to artifacts such as blinks and eye movement [58]. However, ECoG is an invasive recording modality which requires a craniotomy to implant an electrode grid, entailing significant health hazards. For that reason, the first studies on ECoG were with animals. Early studies involving animals evaluated the long-term stability of the signals from the brain that ECoG could acquire [59–62]. The results showed that subdural electrodes could provide stable signals over several months. Nevertheless, the long-term stability of the signals acquired by ECoG is currently unclear. More recent experiments with monkeys have shown that ECoG can perform at a high level for months without any drift in accuracy or recalibration [63]. The hand positions and arm joint angles could be successfully decoded during asynchronous movements. These studies have also developed minimally invasive protocols to implant the ECoG probes [64]. In humans, ECoG has been used for the analysis of alpha and beta waves [65] or gamma waves [66,67] produced during voluntary motor action. With regard to the use of ECoG in BCIs systems, Levine et al. [68] designed a BCI which classified motor actions on the basis of the identification of the event-related potentials (ERP) using ECoG. Leuthardt et al. [69] showed for the first time that an ECoG-based BCI could provide information to control a one-dimensional cursor, as this information is more precise and more quickly acquired than by EEG-based BCIs. Some years later, Schalk et al. [70] presented a more advanced ECoG-based BCI which allowed the user to control a two-dimensional cursor. The results of all these studies might make it more feasible for people with severe motor disabilities to use ECoG-based BCIs for their communication and control needs. Sensors 2012, 12 1219 2.4. Intracortical Neuron Recording Intracortical neuron recording is a neuroimaging technique that measures electrical activity inside the gray matter of the brain. It is an invasive recording modality that needs to implant microelectrode arrays inside the cortex to capture spike signals and local field potentials from neurons. Three signals can be obtained by intracortical neuron recording: single-unit activity (SUA), multi-unit activity (MUA), and local field potentials (LFPs) [47]. SUA is obtained by high-pass filtering (>300 Hz) of the signal of a single neuron. MUA is obtained in the same way, but the signals may come from multiple neurons. LFPs are extracted by low-pass filtering (100 bits/min) − Synchronous signal is necessary − User training required − More suitable for application with many options The typical VEP-based BCI application displays flashing stimuli, such as digits or letters, on a screen to induce SSVEPs while the user stares at one of the symbols. The user can move their gaze to the flashing digits or letters, in order to communicate with the computer [111]. The advantage of this type of control signal is that very little training is required. However, it presents the drawback that the user has to watch the screen and keep his eyes fixed on one point. This type of control signal can only be used for exogenous BCIs (see Section 0). Therefore, VEPs are not suitable for patients in advanced stages of Amyotrophic Lateral Sclerosis (ALS) or with uncontrollable eye or neck movements. Some independent SSVEP-based BCIs that are controlled by the attention of the user have been introduced to overcome this drawback [112,113]. SSVEP are usually elicited through light-emitting diodes (LEDs), cathode-ray tube (CRT) monitors, or liquid crystal display (LCD). LEDs outperform LCD or CRT stimulators but they need more complex hardware. LCD and CRT monitors make the target presentation easier than LED stimulators, because both systems can easily be connected to a PC. However, LED stimulators may be preferable for a multiple target BCI, because the refresh rate of an LCD or CRT monitor can limit the number of targets. LED stimulators offer more versatility because the flickering frequency and phase of each LED can be controlled independently by a programmable logic device [114]. The stimulation decision can be made on the basis of the number of choices that the BCI offers [115]. LCD screens are optimal for low complexity BCI (less than 10 choices), because they induce less eye-tiredness than CRT screens. For medium complexity BCI (10–20 choices), LCD or CRT screens are optimal. For high complexity BCI (more than 20 commands), LED are preferred. 3.2. Slow Cortical Potentials (SCPs) SCPs are slow voltage shifts in the EEG that last a second to several seconds. SCPs belong to the part of the EEG signals below 1 Hz [116]. SCPs are associated with changes in the level of cortical activity. Negative SCPs correlate with increased neuronal activity, whereas positive SCPs coincide with decreased activity in individual cells [116]. These brain signals can be self-regulated by both healthy users and paralyzed patients to control external devices by means of a BCI. SCP shifts can be used to move a cursor and select the targets presented on a computer screen [117]. Sensors 2012, 12 1224 People can be trained to generate voluntary SCP changes using a thought-translation device [117]. The thought-translation device is a tool used for self-regulation SCP training, which shows visual-auditory marks so that the user can learn to shift the SCP. The thought-translation device typically comprises a cursor on a screen in such a way that the vertical position of the cursor constantly reflects the amplitude of SCP shifts. Although most thought-translation devices show continuous feedback, it is possible to train SCP self-modulation in the absence of continuous feedback [118]. Success in SCP self-regulation training depends on numerous factors, such as the patient’s psychological and physical state, motivation, social context, or the trainer-patient relationship [117]. It is known that the learning capability of the user drastically affects SCP modulation training. Self-regulation training is therefore strongly recommended for patients at the early stage of a progressive disease [117]. Furthermore, initial SCP modulation skills have an effect on future performance following training [119]. Therefore, the value of SCPs as a suitable control signal for each patient can only be determined on the basis of initial trials. Other factors, such as sleep quality, pain, and mood also have an influence on self-regulation performance [117]. Their effects are not identical for all patients and further investigation is certainly needed to establish general rules on this matter. Self-regulation of SCPs has been tested extensively with patients suffering from ALS [120–122]. Typical accuracy rates achieved for SCP classification are acceptable and vary between 70 and 80 per cent, but the rates of information provided by SCP-based BCI are relatively low. Besides, longer training is required to use SCP-based BCI and it is likely that users will need continuous practice for several months. 3.3. P300 Evoked Potentials P300 evoked potentials are positive peaks in the EEG due to infrequent auditory, visual, or somatosensory stimuli. These endogenic P300 responses are elicited about 300 ms after attending to an oddball stimulus among several frequent stimuli [123,124]. Some studies have proven that the less probable the stimulus, the larger the amplitude of the response peak [125]. The use of P300-based BCIs does not require training. However, the performance may be reduced because the user gets used to the infrequent stimulus and consequently P300 amplitude is decreased [126]. A typical application of a BCI based on visual P300 evoked potentials comprises a matrix of letters, numbers, or other symbols or commands [123,127,128]. The rows or columns of this matrix are flashed at random while the EEG is monitored. The user gazes at the desired symbol and counts how many times the row or column containing the desired choice flashes. P300 is elicited only when the desired row or column flashes. Thus, the BCI uses this effect to determine the target symbol. Due to the low signal-to-noise ratio in EEG signals, the detection of target symbols from a single trial is very difficult. The rows or columns must be flashed several times for each choice. The epochs corresponding to each row or column are averaged over the trials, in order to improve their accuracy. However, these repetitions decrease the number of choices per minute, e.g., with 15 repetitions, only two characters are spelled per minute [123]. Although most of the applications based on P300 evoked potentials employ visual stimuli, auditory stimuli have been used for people with visual impairment [129]. P300-based BCIs provide a very low rate of information transmission because the classifier based on an average is too simple, and the accuracy of P300 potential detection is too low [130]. Sensors 2012, 12 1225 Consequently, too many trials are required to select a single symbol in the matrix. Accuracy of P300-based BCIs can be improved, while using a more complicated classifier than a simple average to ensure that the number of repetitions remain unaffected [130,131]. Other studies have proven that the detection accuracy of visual P300 evoked potentials also depends on the properties of the visual matrix such as the dimensions or colors of the symbols. Performance decreases when matrices with smaller symbols are used [132], and it is enhanced when a green and blue chromatic flicker matrix is used, rather than a gray and black one [133,134]. Information transmission rates provided by P300-based BCI can be also improved by considering the BCI as a noisy transmission system. BCI can therefore benefit from the use of error correcting codes [135]. However, optimizing the code solely according to the maximal minimum-Hamming-distance implies an increase in target frequency of target stimuli which might violate physiological constraints leading to difficulties in classifying the individual ERPs, due to overlap and refractory effects. Further, overlap and refractory effects are generally the main error source in these kinds of BCIs [136]. Some recent novel approaches have tried to reduce them, by superimposing the targets on a checkerboard [137] or by using alternative stimulus type methods based on motion [136]. The P300 response is not markedly affected by whether or not the subject gazes directly at the target, in contrast to the VEP response, which is larger when the target is foveated [138]. This distinction is important for clinical applications, because eye movements are often impaired or lost in the target population. Nevertheless, the performance of a P300-based BCI is substantially improved when subjects gaze at the desired item [138]. Therefore, the performance of the visual P300-based BCIs depends not only on the P300-evoked potential, but also on the VEP response that, in turn, strongly depends on eye-gaze direction. 3.4. Sensorimotor Rhythms (mu and beta rhythms) Sensorimotor rhythms comprise mu and beta rhythms, which are oscillations in the brain activity localized in the mu band (7–13 Hz), also known as the Rolandic band, and beta band (13–30 Hz), respectively. Both rhythms are associated in such a way that some beta rhythms are harmonic mu rhythms, although some beta rhythms may also be independent [139]. The amplitude of the sensorimotor rhythms varies when cerebral activity is related to any motor task although actual movement is not required to modulate the amplitude of sensorimotor rhythms [140,141]. Similar modulation patterns in the motor rhythms are produced as a result of mental rehearsal of a motor act without any overt motor output [140]. Sensorimotor rhythms have been used to control BCIs, because people can learn to generate these modulations voluntarily in the sensorimotor rhythms [36,142]. Sensorimotor rhythms can endure two kinds of amplitude modulations known as event-related desynchronization (ERD) and event-related synchronization (ERS) that are generated sensory stimulation, motor behavior, and mental imagery [36]. ERD involves an amplitude suppression of the rhythm and ERS implies amplitude enhancement. Figure 2 (left panel) shows the temporal behavior of ERD and ERS during a voluntary movement experiment which involves brisk finger lifting [36]. The mu band ERD starts 2.5 s before movement on-set, reaches the maximal ERD shortly after movement-onset, and recovers its original level within a few seconds. In contrast, the beta rhythm shows a short ERD during the movement initiation of movement, followed by ERS that reaches the Sensors 2012, 12 1226 maximum after movement execution. This ERS occurs while the mu rhythm is still attenuated. Figure 2 also shows the gamma oscillation (36–40 Hz), which is another rhythm related to motor tasks as well [36]. Gamma rhythms reveal an ERS shortly before movement-onset. Finally, the right panel of Figure 2 illustrates that simultaneous ERD and ERS are possible at different scalp locations [36]. Figure 2. Left panel: Superimposed band power time courses computed for three different frequency bands (10–12 Hz, 14–18 Hz, and 36–40 Hz) from EEG trials recorded from electrode position C3 during right index finger lifting. EEG data triggered with respect to movement-offset (vertical line at t = 0 s); Right panel: Examples of ongoing EEG recorded during right finger movement (adapted from [36]). Sensorimotor rhythms are related to motor imagery without any actual movement [141]. This makes it possible to use sensorimotor rhythms for the design of endogenous BCIs, which are more useful than exogenous BCIs. Nevertheless, self-control of sensorimotor rhythms is not easy, and most people have difficulties with motor imagery. People tend to imagine visual images of related real movements, which is not sufficiently useful for a BCI system, because the patterns of these sensorimotor rhythms differ from actual motor imagery. User training should emphasize kinesthetic experiences instead of visual representations of actions [143]. Motor imagery training is traditionally based on visual or auditory feedback [144]. This kind of training asks the users to perform a certain motor imagery task, and then the sensorimotor rhythms are extracted and classified by comparing them with a reference. Finally, visual or auditory feedback is provided to the participant according to the success of the result. This kind of training has been widely used although usually its effectiveness was not very high [145]. Hwang et al. [145] presented more effective motor imagery training based on a system that displayed real-time cortical activity as feedback, which allowed the users to watch their own cortical activity through a real-time monitoring system. Sensorimotor rhythms have been investigated extensively in BCI research. Well-known BCI systems such as Wadsworth [146], Berlin [147], or Graz [148] BCIs employ sensorimotor rhythms as control signals. The BCIs based on sensorimotor rhythms can operate in either synchronous or asynchronous mode. The latest advances in the field of BCIs based on sensorimotor rhythms have shown that it is possible to predict human voluntary movements before they occur based on the modulations in sensorimotor rhythms [149]. Furthermore, this prediction could be provided without the user making any movements at all. 250 200 150 100 50 0 -50 100-6 -5 -4 -3 -2 -1 0 1 2 3 C3 Gamma ERS 36-40 Hz Post-movement beta ERS 14-18 Hz Mu ERD 10-12 Hz E RD / E R S % Fz C3 C4 Cz P3 P4 Oz s -5 -4 -3 -2 -1 0 1 2 3 ERD ERS Sensors 2012, 12 1227 4. Types of BCIs The BCIs can be categorized into (i) exogenous or endogenous and (ii) synchronous (cue-paced) or asynchronous (self-paced). Types of BCI are listed in Tables 4 and 5, along with information related to brain signals that can be modulated to convey information as well as advantages and disadvantages. Also, BCIs can be classified into dependent and independent [2]. This distinction will not be detailed in this review because it is very similar to exogenous and endogenous distinction. Advantages and disadvantages in both taxonomies are analogous. Table 4. Main differences between exogenous and endogenous BCI. Approach Brain signals Advantages Disadvantages Exogenous BCI − SSVEP − P300 − Minimal training − Control signal set-up easily and quickly − High bit rate (60 bits/min) − Only one EEG channel required − Permanent attention to external stimuli − May cause tiredness in some users Endogenous BCI − SCPs − Sensorimotor rhythms − Independent of any stimulation − Can be operated at free will − Useful for users with sensory organs affected − Suitable for cursor control applications − Very time-consuming training (months or weeks) − Not all users are able to obtain control − Multichannel EEG recordings required for good performance − Lower bit rate (20–30 bits/min) Table 5. Main differences between synchronous and asynchronous BCIs. Approach Advantages Disadvantages Synchronous BCI − Simpler design and performance evaluation − The user can avoid generating artifacts since they can perform blinks and other eye movements when brain signals are not analyzed − Does not offer a more natural mode of interaction Asynchronous BCI − No requirement to wait for external cues − Offers a more natural mode of interaction − Much more complicate design − More difficult evaluation According to the nature of the signals used as input, BCI systems can be classified as either exogenous or endogenous. Exogenous BCI uses the neuron activity elicited in the brain by an external stimulus such as VEPs or auditory evoked potentials [150]. Exogenous systems do not require extensive training since their control signals, SSVEPs and P300, can be easily and quickly set-up. Besides, the signal controls can be realized with only one EEG channel and can achieve a high information transfer rate of up to 60 bits/min. On the other hand, endogenous BCI is based on self-regulation of brain rhythms and potentials without external stimuli [150]. Through neurofeedback training, the users learn to generate specific brain patterns which may be decoded by the BCI such as modulations in the sensorimotor rhythms [151] or the SCPs [117]. The advantage of an endogenous BCI is that the user can operate the BCI at free will and move a cursor to any point in a Sensors 2012, 12 1228 two-dimensional space, while an exogenous BCI may constrain the user to the choices presented. Also, endogenous BCI are especially useful for users with advanced stages of ALS or whose sensory organs are affected. Table 4 summarizes the differences between exogenous and endogenous BCIs. According to the input data processing modality, BCI systems can be classified as synchronous or asynchronous. Synchronous BCIs analyze brain signals during predefined time windows. Any brain signal outside the predefined window is ignored. Therefore, the user is only allowed to send commands during specific periods determined by the BCI system. For example, the standard Graz BCI [148] represents a synchronous BCI system. The advantage of a synchronous BCI system is that the onset of mental activity is known in advance and associated with a specific cue [152]. Moreover, the patients may also perform blinks and other eye movements, which would generate artifacts, if the BCI did not analyze the brain signals to avoid their misleading effects. This simplifies the design and evaluation of synchronous BCI. Asynchronous BCIs continuously analyze brain signals no matter when the user acts. They offer a more natural mode of human-machine interaction than synchronous BCI. However, asynchronous BCIs are more computation demanding and complex. Table 5 summarizes the differences between synchronous and asynchronous BCIs. 5. Features Extraction and Selection Different thinking activities result in different patterns of brain signals. BCI is seen as a pattern recognition system that classifies each pattern into a class according to its features. BCI extracts some features from brain signals that reflect similarities to a certain class as well as differences from the rest of the classes. The features are measured or derived from the properties of the signals which contain the discriminative information needed to distinguish their different types. The design of a suitable set of features is a challenging issue. The information of interest in brain signals is hidden in a highly noisy environment, and brain signals comprise a large number of simultaneous sources. A signal that may be of interest could be overlapped in time and space by multiple signals from different brain tasks. For that reason, in many cases, it is not enough to use simple methods such as a band pass filter to extract the desired band power. Brain signals can be measured through multiples channels. Not all information provided by the measured channels is generally relevant for understanding the underlying phenomena of interest. Dimension reduction techniques such as principal component analysis or independent component analysis can be applied to reduce the dimension of the original data, removing the irrelevant and redundant information. Computational costs are thereby reduced. Brain signals are inherently non-stationary. Time information about when a certain feature occurs should be obtained. Some approaches divide the signals into short segments and the parameters can be estimated from each segment. However, the segment length affects the accuracy of estimated features. FFT performs very poorly with short data segments [153]. Wavelet transform or adaptive autoregressive components are preferred to reveal the non-stationary time variations of brain signals. Also, a novel technique called stationary subspace analysis (SSA) has recently been introduced to deal with the non-stationarity of EEG signals [154]. SSA decomposes multivariate time series into stationary and non- stationary components. Sensors 2012, 12 1229 Multiples features can be extracted from several channels and from several time segments before being concatenated into a single feature vector. One of the major difficulties in BCI design is choosing relevant features from the vast number of possible features. High dimensional feature vectors are not desirable due to the “curse of dimensionality” in training classification algorithms (see next section). The feature selection may be attempted examining all possible subsets of the features. However, the number of possibilities grows exponentially, making an exhaustive search impractical for even a moderate number of features. Some more efficient optimization algorithms can be applied with the aim of minimizing the number of features while maximizing the classification performance. This section discusses methods to obtain the relevant characteristics of brain signals as well as feature selection methods. Firstly, dimensional reduction methods, such as principal component analysis or independent component analysis are explained. Secondly, time and/or frequency methods, such as matched filtering or wavelet transform, and parametric modeling, such as autoregressive component, are also surveyed. Thirdly, an explanation is given of the common spatial pattern algorithm. This method designs a preprocessing spatial filter, by means of spatial covariance from input data and signal whitening, that enhances the difference between classes before the feature extraction stage. And, finally, feature selection methods such as genetic algorithms or sequential selection are included. All these methods, including feature extraction and feature selection methods, are listed respectively in Tables 6 and 7 along with information on their properties and BCI applications. Table 6. Summary of feature extraction methods. Method Properties Applications Dimension reduction PCA − Linear transformation − Set of possibly correlated observations is transformed into a set of uncorrelated variables − Optimal representation of data in terms of minimal mean-square-error − No guarantees always a good classification − Valuable noise and dimension reduction method. PCA requires that artifacts are uncorrelated with the EEG signal [155,157,158] ICA − Splits a set of mixed signals into its sources − Mutual statistical independence of underlying sources is assumed − Powerful and robust tool for artifact removal. Artifacts are required to be independent from the EEG signal − May corrupt the power spectrum [160,161, 164–168] Space CSP − Spatial filter designed for 2-class problems. Multiclass extensions exist − Good result for synchronous BCIs. Less effective for asynchronous BCIs − Its performance is affected by the spatial resolution. Some electrode locations offer more discriminative information for some specific brain activities than others − Improved versions of CSP: WCSP, CSSP, CSSSP [183–187] Sensors 2012, 12 1230 Table 6. Cont. Timefrequency AR − Spectrum model − High frequency resolution for short time segments − Not suitable for non-stationary signals − Adaptive version of AR: MVAAR [170,172] MF − Detects a specific pattern on the basis of its matches with predetermined known signals or templates − Suitable for detection of waveforms with consistent temporal characteristics [151,173] CWT − Provides both frequency and temporal information − Suitable for non-stationary signals [179,180] DWT − Provides both frequency and temporal information − Suitable for non-stationary signals − Reduces the redundancy and complexity of CWT [181,182] Table 7. Summary of feature extraction methods. Method Properties Applications Features selection GA − High resource consumption − Possible premature convergence [188,189] SFS/SBS − Suboptimal methods [191,192] SFFS/SBFS − Modified versions of SFS/SBS methods − Based on plus l-take away r algorithm − Partially overcome the deficiencies of SFS/SBS [194–196] 5.1. Principal Component Analysis (PCA) PCA is a statistical features extraction method that uses a linear transformation to convert a set of observations possibly correlated into a set of uncorrelated variables called principal components. Linear transformation generates a set of components from the input data, sorted according to their variance in such a way that the first principal component has the highest possible variance. This variance allows PCA to separate the brain signal into different components. PCA projects the input data on a k-dimension eigenspace of k eigenvectors, which are calculated from the covariance matrix ∑ of the training data p = [p1 p2 … pn] [155]. pi is i-th d-dimension training sample, and n is the number of samples. The covariance matrix ∑ is computed as: ∑ ൌ ሺ െ ݉ሻሺ െ ݉ሻ௧ ୀଵ (1) where, ݉ ൌ ଵ ∑ ୀଵ is the mean vector of the training samples pi. The covariance matrix ∑ is a real and symmetric ݀ ݔ ݀ matrix, therefore ∑ has d different eigenvectors and eigenvalues. By means of the eigenvalues, it is possible to know which eigenvectors represent the most significant information contained in the dataset. The eigenvectors with the highest Sensors 2012, 12 1231 eigenvalue represent the principal components of the training dataset p. PCA selects that k, with k < d, eigenvectors having the largest eigenvalues. These selected eigenvectors serve to build a projection matrix A that will be used to extract the feature vector from the test data q. The k eigenvectors are sorted into columns in Matrix A, such that the first column of A corresponds to the largest eigenvalue. Finally, PCA computes the feature vector v from the data in matrix A, by projecting the test data q onto the new subspace, such that: ݒ ൌ ܣ௧ሺݍെ݉ሻ (2) where,݉ ൌ ଵ ∑ ୀଵ represents the mean vector of training samples pi. PCA is also a procedure to reduce the dimension of the feature. Since the number of columns is less than the number of eigenvectors, the dimension of the output projected data is less than the dimension of the input data. This decrease in dimensionality can reduce the complexity of the subsequent classifying step in a BCI system. PCA does not always guarantee a good classification since the best discriminating components may not figure among the largest principal components [156]. PCA reduces data dimension by seeking a new optimal representation of data in terms of minimal mean-square-error between the representation and the original data. It will not guarantee that the discriminative features are optimal for classification. Despite this shortcoming, it has been proven that PCA is a reliable noise reduction method. With regard to the applications of PCA in BCI systems, PCA has been used to identify the artifactual components in a reasonably successful way in EEG signals and to reconstruct the signals without the artifactual components [157,158]. Nevertheless, the artifacts must not be correlated with the EEG signal for PCA to function in this way. PCA has also been employed, in order to reduce feature space dimensionality [155]. 5.2. Independent Component Analysis (ICA) ICA is a statistical procedure that splits a set of mixed signals into its sources with no previous information on the nature of the signal. The only assumption involved in ICA is that the unknown underlying sources are mutually independent in statistical terms. ICA assumes that the observed EEG signal is a mixture of several independent source signals coming from multiple cognitive activities or artifacts. ICA therefore expresses the resulting EEG signal x(t) in relation to their sources s(t) as: ࢞ሺݐሻ ൌ ࢌሺ࢙ሺݐሻሻ ሺݐሻ (3) where, f is any unknown mixer function, and n(t) is an additive random noisy vector. The dimension of the input vector s(t) depends on the number of sources. The dimension of output vector x(t) is equal to the number of measured data channels. The number of sources is usually assumed to be less than or equal to the number of channels although more generalized ICA methods are possible [159]. The whole ICA problem consists in the calculation of the unmixing function by inverting f and obtaining an estimation of s(t), by mapping x(t) to the source space. To solve the problem, ICA can fall into two different models on the basis of f, which may be either a linear or nonlinear function. The nonlinear assumption is suitable in those cases where the linear model might be too simple to describe the observed data x(t). However, the nonlinear problem is usually too complex and generally intractable due to its high number of indeterminations. The assumption of a linear mixing function Sensors 2012, 12 1232 simplifies Equation (3). It is possible to rewrite it as a matrix multiplication where A is the mixing matrix. The Equation (4) gives the mathematical expression of the linear ICA model: ࢞ሺݐሻ ൌ ࢙ሺݐሻ ሺݐሻ (4) Although the approximation given by Equation (4) can be considered too simple, it works reasonably well in brain signal processing applications. Furthermore, it is possible to remove the noise term n(t) from Equation (4), by assuming that the observed data is noiseless or that the noise is too weak for consideration [160,161]. Finally, s(t) and A are obtained from x(t) by means of certain algorithms, such as Infomax [162] or further modification of the Infomax [163]. ICA has traditionally been used as a preprocessing tool before the feature extraction step, in order to remove ocular artifacts in BCI systems [164–166]. Although ICA has been proven to be a powerful and robust tool for artifact removal in signal analysis, some studies have indicated that artifact suppression may also corrupt the power spectrum of the underlying neural activity [167]. In addition, ICA requires that the artifacts are independent in relation to the EEG signal. It is also possible to find authors that have employed ICA as a classifier. ICA can be modified to classify EEG signals by fitting the generative ICA model to each task and employing Bayes’ rule to create the classifier [168]. 5.3. AutoRegressive Components (AR) AR spectral estimation is a method for modeling signals. AR models the EEG signal as the output random signal of a linear time invariant filter, where the input is white noise with a mean of zero and a certain variance of σ 2 . The aim of the AR procedure is to obtain the filter coefficients, since it is assumed that different thinking activities will produce different filter coefficients. The filter coefficients will be used as the features of the signal. AR assumes that the transfer function of the filter will only contain poles in the denominator. The number of poles in the denominator corresponds to the order of the autoregressive model. The assumption of an all-pole filter makes the filter coefficients computation easier because it is only necessary to solve linear equations. Mathematically, the AR model of order p describes the EEG signal y(t) as: ݕሺݐሻ ൌ ܽଵݕሺݐെ1ሻ ܽଶݕሺݐെ2ሻ ܽଷݕሺݐെ3ሻ ڮܽݕሺݐെሻ ݊ሺݐሻ (5) where, ai is the i-th filter coefficient, and n(t) is the noise. There are several methods that compute the filter coefficients such as the Yule-Walker, Burg, covariance, and forward-backward algorithms [169]. The resulting coefficients can be used to estimate the power spectrum of the EEG signal y(ω), such that: ݕሺ߱ሻ ൌ 1 ห1 െ ∑ ܽ ݁ ିఠ ୀଵ ห ଶ (6) where, ak are the estimated filter coefficients, and p is the AR model order, in other words, the number of poles. In the AR model, the determination of an appropriate order p for a given input signal is a trade-off issue. If the order is too low to model the input signal, the result will not faithfully represent the signal Sensors 2012, 12 1233 because the spectrum is too smooth. In contrast, if the order is too high, the spectrum may exhibit spurious peaks. AR spectral estimation is preferred to Fourier Transform, because of its superior resolution for short time segments [170]. Nevertheless, AR performs poorly when the signal is not stationary [171]. Due to the non-stationary nature of EEG signals, a multivariate adaptive AR (MVAAR) model has been proposed to design more effective on-line BCI systems. Jiang et al. [172] applied MVAAR for the classification of motor imagery, showing that MVAAR is a valuable adaptive method for feature extraction. The computation algorithm was very similar to the original AR model. In a BCI with m channels, the vector of m EEG values, at each point in time k, was represented as: ݕ Ԧൌ ሾݕ,ଵ ݕ,ଶ … ݕ,ሿ ் (7) As in the AR case, the MVAAR model was expressed as: (8Ԧ (ሬԦି ݊ݕܣ ڮ ଷԦିݕଷܣ ଶԦିݕଶܣ ଵԦିݕଵܣ ൌԦ ݕ where, ݊ሬԦ was the vector of white noise values, ܣଵ ܣ were the adaptive coefficients, and p was the model order. The Recursive Least Squares algorithm, a special variant of the Kalman Filter, were used to update coefficients ܣଵ ܣ at every point k. 5.4. Matched Filtering (MF) MF is a feature extraction method that attempts to detect a specific pattern on the basis of its matches with predetermined known signals or templates. The intention of the user is revealed by means of the correlation between the unknown EEG signals and the set of templates. Each template represents an intention of the user. A higher correlation would imply better matching between the template and the user’s intention. Each matched filter can simply be modeled as a sum of the harmonically related sinusoidal components [151]: ܨܯሺ݊ሻ ൌܽcos ൬ 2ߨ݂݇ி ݂௦ ൰ߔ݊ ே ୀଵ (9) where, n is the template sample number, fs is the sampling frequency, fF is the fundamental frequency of the rhythm template, N-1 is the number of harmonics, and at and ߔare the amplitude and phase of the individual harmonics, respectively. The model parameters at and ߔ can be obtained from the FFT spectrum [151]. MF has been proven especially effective for the detection of waveforms with consistent temporal characteristics. Krusienski et al. [151] used MF for the identification of user intentions through µrhythms and Brunner et al. [173] also used it for SSVEP feature extraction. 5.5. Wavelet Transform (WT) WT is a mathematical tool widely used for extracting information from many different kinds of data, such as audio or image data, among others. WT is particularly suitable when signals are not stationary, because it provides a flexible way of representing the time-frequency of a signal [174]. Sensors 2012, 12 1234 Wavelets are functions of varying frequency and limited duration that allow simultaneous study of the signal in both the time and the frequency domain [175], in contrast to other modalities of signal analysis such as Fourier transform (FT). FT provides only an analysis of the signal activity in the frequency domain. FT gives information about the frequency content, but it is not accompanied by information on when those frequencies occur. Short-term Fourier Transform (STFT) was proposed to overcome this shortcoming of the Fourier analysis. The STFT divides the signal into successive time windows and applies the FT in each epoch of the signal in time. In this approach, the design of window length is a trade-off because smaller windows lead to higher temporal resolution but also to lower frequency resolution at the same time. The WT overcomes this drawback by decomposing the signal in both the time and the frequency domain at multiple resolutions, by using a modulated window that is shifted along the signal at various scales. Continuous wavelet transform (CWT) is defined as the convolution of the signal ݔሺݐሻ with the wavelet function ߰௦,ఛሺݐሻ [175]: ݓሺݏ߬ ,ሻ ൌන ݔሺݐሻ߰௦,ఛ ݐ݀ሻݐሺ כ ஶ ିஶ (10) ݓሺݏ߬ ,ሻ is the wavelet coefficient that corresponds to the frequency associated with the scale s and the time τ of the wavelet function ߰௦,ఛሺݐሻ, and the symbol ‘*’ expresses the complex conjugation. The wavelet function ߰௦,ఛሺݐሻ is a dilated and shifted version of a mother wavelet ߰ሺݐሻ: ߰௦,ఛሺݐሻ ൌ 1 ݏ√ ߰ሺ ߬െݐ ݏ ሻ (11) A mother wavelet can take multiples shapes, but it always satisfies the next condition: න ߰ሺݐሻ݀ݐ ൌ 0 ஶ ିஶ (12) The CWT defined in the Equation (10) is actually a kind of template matching, similar to a matched filter in which the cross variance between the signal and a predefined waveform is calculated [151]. The advantage of the CWT over classic template matching methods arises from the special properties of the wavelet template. The wavelets are suitable for transient signal analysis, in which the spectral properties of the signal vary over time [176]. WT is a powerful tool for the decomposition of transient brain signals into their constituent parts, based on a combination of criteria such as frequency and temporal position. Signals of identical frequency ranges can be distinguished by means of the temporal position. Likewise, it is possible to separate temporally overlapping processes thanks to the different frequency content. The CWT introduces a lot of redundancy and complexity since it involves the analysis of a signal at a very high number of frequencies using multiple dilations and shifting of the mother wavelet. Discrete wavelet transform (DWT) was introduced to reduce this redundancy and complexity. The DWT translates and dilates the mother wavelet in certain discrete values only [177]. Farina et al. [178] showed a pattern recognition approach for the classification of single trial movement-related cortical potentials, where the feature space is built from coefficients of a discrete wavelet transformation. Although DWT is less redundant and less complex than CWT, CWT is still employed to extract Sensors 2012, 12 1235 features from P300 and SCP, because it can clarify subtle information that DWT is unable to extract [179]. The use of WT requires the selection of a mother wavelet. Many different mother wavelets can be found in BCI applications and the selection of any one depends on what types of features need to be extracted from the signal. The Mexican Hat wavelet is well localized in the time domain and is employed for the localization of ERP components in time [179]. The Morlet wavelet is well localized in the frequency domain and has been used for the analysis of gamma activity [180]. The bi-scale wavelet has been employed successfully for designing an asynchronous BCI based on detection of imaginary movement in the 1–4 Hz frequency range [181]. Also, the Daubechies wavelet, a very well-known mother wavelet, has been used for the classification of SCPs [182]. 5.6. Common Spatial Pattern (CSP) CSP is a feature extraction method that projects multichannel EEG signals into a subspace, where the differences between classes are highlighted and the similarities are minimized. It aims to make the subsequent classification much more effective, by designing a spatial filter that transforms the input data into output data with an optimal variance for the subsequent discrimination [183]. CSP has been designed for the analysis of multichannel data belonging to 2-class problems. Nevertheless, some extensions for multiclass BCIs have also been proposed [184]. CSP calculates the normalized spatial covariance C from the input data E, which represents the raw data of a single trial, by means of: ൌ ܥ Ԣܧܧ ᇱܧܧሺܽܿ݁ݎݐ ሻ (13) where, E is an ܰ ݔ ܶ matrix, in which T is the number of channels, i.e., recording electrodes, and N the number of samples per channel. The apostrophe Ԣ denotes the transpose operator, while trace(x) is the sum of the diagonal elements of x. Assuming CSP is used to classify two classes, e.g., left and right motor imagery, CSP calculates the spatial covariances ܥ ഥ and ܥҧ for each of the two classes by averaging the covariances over the successive training trials of each class over time. The composite spatial covariance ܥ is computed as: ܥ ൌ ܥ ഥ ܥҧ (14) Since ܥ is real and symmetric, it can be factored as ܥ ൌ ܷߣܷ ᇱ , where ܷ is the matrix of eigenvectors, and ߣ is the diagonal matrix of eigenvalues. By means of the whitening transform: ܲ ൌ ඥߣ ିଵܷ ᇱ (15) the variances are equalized in the space spanned by ܷ ᇱ and all eigenvalues of ܲܥҧ ܲԢ are equal to one. If ҧܥ and ܥҧ are transformed as: ܵ ൌ ܲܥҧ ܲԢ (16) ܵ ൌ ܲܥҧ ܲԢ (17) then, ܵ and ܵ will share common eigenvectors. If ܵ ൌ ߣܤܤԢ, then ܵ ൌ ߣܤܤԢ, and ߣ ߣ ൌ ܫ , where ܫ is the identity matrix. As a result of the sum of two corresponding eigenvalues being always Sensors 2012, 12 1236 one, the eigenvectors with the largest eigenvalues for ܵ correspond to the smallest eigenvalue for ܵ, and vice versa. This property is very useful for subsequent classification, because the variance of the signal is maximized for one class while minimized for the other class. Finally, the feature vector Z is obtained from the trial E as: ܼ ൌ ܹܧ) 18) where, ܹ ൌ ሺܤᇱ ܲሻԢ is the spatial filter matrix built by the CSP procedure. CSP increases the accuracy of synchronous BCIs where it is allowed to send signals only during certain predefined time periods. However, CSP does not offer the same improvement in asynchronous BCIs. This is mainly due to the nonstationary properties of EEG signals [185]. Also, the performance of CSP is affected by the spatial resolution, and it has been proven that some electrode locations offer more discriminative information for some specific brain activities than others. For these reasons, several methods improving the original CSP method have been proposed to increase the performance: Wavelet Common Spatial Pattern (WCSP) [185], Common Spatio-Spectral Pattern (CSSP) [186], and Common Sparse Spectral Spatial Pattern (CSSSP) [187]. 5.7. Genetic Algorithm (GA) GA is an optimization procedure to establish whether a certain set of features is the most efficient. GA has been used in very diverse fields to solve optimization problems. In BCI research, GA has been used as an automatic method to extract an optimal set of relevant features [188,189]. The baseline of the algorithm is a population of candidate solutions called individuals, creatures, or phenotypes which are encoded by strings named chromosomes or the genotype of the genome. These strings are coded either by binary information or no binary information. The standard steps of the GA can be explained briefly as follows (Figure 3). GA begins with an initial population which is randomly generated unless the algorithm has previous of the final solution. In the case of having initial information, the initial population may be directed towards areas where optimal solutions are more likely to reduce the number of iterations. The fitness of every individual population is evaluated. According to their fitness, some representatives of the population may be discarded to vacate space for newly generated individuals. Other individuals may be selected as parents in order to breed new individuals. Also, some individuals may be stochastically selected to keep diversity in the population preventing premature convergence. After the selection step, the individuals are crossed with each other. In the crossover step, mating is performed among the selected parents to generate one or more offspring. To keep a fixed population size, the number of offspring is usually the same as the number of discarded individuals. The parents’ genes are split into pieces and then combined to form new offspring. Following the crossover step, mutations are introduced to alter the population in order to avoid converging towards a local suboptimum solution before exploring the entire search space. As a result of the mutation, it is possible to discover areas that cannot be explored by crossover. Finally, the fitness of the new population is evaluated. When an acceptable solution is reached or the maximum number of generations has been produced, the algorithm is terminated. Otherwise, another iteration of the algorithm is produced. Sensors 2012, 12 1237 Figure 3. Genetic algorithm. 5.8. Sequential Selection Sequential selection is an optimization approach that aims at finding the optimal subset of features by adding or removing features sequentially. There are two algorithms that perform sequential selection: sequential forward selection and sequential backward selection. Sequential forward selection (SFS) [190] is a bottom up algorithm. Firstly, the best individual feature is found as the first feature in the subset. Next, for each subsequent step, the algorithm chooses the feature from the remaining set, which in combination with the previously selected features, yields the best subset of features. Finally, the algorithm finishes when the required number of features is reached. The shortcoming of this algorithm is that the superfluous features are not removed once other features are added. Sequential backward selection (SBS) [190], in contrast to SFS, is a top down process. The process starts with the entire set of features and removes step by step features in such a way that the error is as low as possible. This algorithm is also suboptimal, because it discards some features that may be helpful after discarding other features. SFS has been used with success in the field of BCIs [191,192]. Another refined method is introduced to partially overcome the aforementioned deficiencies. This method, known as plus l take away r method (l > r), adds l features, and remove r features that is not working well with other selected features. Sequential forward floating search (SFFS) or sequential backward floating search (SBFS) are based on the plus l-take away r method [193]. SFFS starts with a null feature set and, for each step, the r best features are included in the current feature set. In other words, r steps of SFS are performed. Next, the algorithm verifies the possibility that some feature may be excluded. Then, l worst features are eliminated from the set; in other words, l steps of SBS. SFFS increases and decreases the number of features until the desired number of features is reached. SBFS works analogously, but starting with the full feature set and performing the search until the desired dimension is reached, using SBS and SFS steps. In BCI research, SFFS has been used to reduce the dimensionality of the feature space to an appropriate size for the available training data [194–196]. Initial population Crossover Mutation Selection Evaluation Optimum population or maximum number of iterations? End Start Yes No Sensors 2012, 12 1238 6. Artifacts in BCIs Artifacts are undesirable signals that contaminate brain activity and are mostly of non-cerebral origin. Since the shape of neurological phenomenon is affected, artifacts may reduce the performance of BCI-based systems. Artifacts may be classified into two major categories: physiological artifacts and non-physiological or technical artifacts. Physiological artifacts are usually due to muscular, ocular and heart activity, known as electromyography (EMG), electrooculography (EOG), and electrocardiography (ECG) artifacts respectively [197]. EMG artifacts, which imply typically large disturbances in brain signals, come from electrical activity caused by muscle contractions, which occur when patients are talking, chewing or swallowing. EOG artifacts are produced by blinking and other eye movements. Blinking makes generally high-amplitude patterns over brain signals in contrast to eye movements which produce low-frequency patterns. These electrical patterns are due to the potential difference between the cornea and the retina, as their respective charges are positive and negative. For that reason, the electric field around the eye changes when this dipole moves. EOG artifacts mostly affect the frontal area, because they are approximately attenuated according to the square of the distance [198]. Finally, ECG artifacts, which reflect heart activity, introduce a rhythmic signal into brain activity [197]. Technical artifacts are mainly attributed to power-line noises or changes in electrode impedances, which can usually be avoided by proper filtering or shielding [197]. Therefore, the BCI community focuses principally on physiological artifacts, given that their reduction during brain activity acquisition is a much more challenging issue than non-physiological artifact handling. Several ways of handling physiological artifacts can be found in the literature. Artifacts may be avoided, rejected or removed from recordings of brain signals. Artifact avoidance involves asking patients to avoid blinking or moving their body during the experiments [199]. This approach to artifact handling is very simple, because it does not require any computation as brain signals are not assumed to have artifacts. However, this assumption is not always feasible given that some artifacts - involuntary heart beats, eye and bodily twitches- are not easily avoidable during data recording, especially in cases of strong neurological disorders [199]. Artifact rejection approaches suggest discarding the epochs contaminated by the artifacts. Manual artifact rejection is an option to remove artifacts in brain signals and an expert could identify and eliminate all artifact-contaminated epochs. The main disadvantage in using manual rejection is that it requires intensive human labor, so this approach is not suitable for on-line BCI systems. Nevertheless, this task can be performed automatically by EMG and EOG artifact detection. If EMG and EOG signals are monitored, the brain signal samples may be removed whenever ocular or muscular activity of the arms is detected [200]. Automatic rejection is an effective way of artifact handling, but it may fail when EOG amplitudes are too small. Besides, rejection methodology means that the user loses device control when artifact contaminated signals are discarded. Instead of rejecting samples, the artifact removal approach attempts to identify and remove artifacts while keeping the neurological phenomenon intact. Common methods for removing artifacts in EEG are linear filtering, linear combination and regression, BSS and PCA [197]; some of which were discussed in Section 0. Instead of avoided, rejected or removed artifacts from recordings of brain signals, some systems acquire and process artifacts to offer a communication path that either disabled or healthy people can Sensors 2012, 12 1239 use in many tasks and in different environments. This kind of system is not considered a BCI, because communication is not independent of peripheral nerves and muscles. EMG computer interface [201], human-computer interface (HCI) [202], EMG-based human-computer interface [203], EMG-Based Human-Machine Interface [204], EMG-based human-robot interface [205], muscle-computer interface (MuCI) [206], man-machine interface (MMI) [207], and biocontroller interface [208] are different terms used to name communication interfaces in the scientific literature that can employ artifact signals, among others. These systems usually have greater reliability than BCIs, but they cannot be used by severely disabled people with strong constraints in voluntary movements. 7. Classification Algorithms The aim of the classification step in a BCI system is recognition of a user’s intentions on the basis of a feature vector that characterizes the brain activity provided by the feature step. Either regression or classification algorithms can be used to achieve this goal, but using classification algorithms is currently the most popular approach [209]. Regression algorithms employ the features extracted from EEG signals as independent variables to predict user intentions. In contrast, classification algorithms use the features extracted as independent variables to define boundaries between the different targets in feature space. McFarland et al. [210] illustrated the differences between the two alternatives. For a two-target case, both the regression approach and the classification approach require the parameters of a single function to be determined. In a four-target case, assuming that the targets are distributed linearly, the regression approach still requires only a single function. In contrast, the classification approach requires the determination of three functions, one for each of the three boundaries between the four targets. Therefore, the classification approach might be more useful for two-target applications and the regression approach may be preferable for greater numbers of targets, when these targets can be ordered along one or more dimensions. Moreover, the regression approach is better for continuous feedback e.g., applications which involve continuous control of cursor movement. Figure 4 illustrates the differences between classification and regression approaches. Figure 4. Classification and regression approaches to BCI control of two-targets (adapted from [210]). The regression algorithms employ the features extracted from EEG signals as independent variables to predict user intentions. In contrast, the classification approach uses the features extracted as independent variables to define boundaries between the different targets in feature space. Classification Regression Sensors 2012, 12 1240 Classification algorithms can be developed via either offline, online or both kinds of sessions. The offline session involves the examination of data sets, such as BCI competitions data sets [211], which are collected from an adaptive or closed-loop system. The statistics of the data may be estimated from observations across entire sessions and long-term computations may be performed. The results can be reviewed by the analyst with the aim of fine-tuning the algorithms. Offline data analysis is valuable, but it does not address real-time issues. In contrast, online sessions provide a means of BCI system evaluation in a real-world environment. The data are processed in a causal manner and the algorithms are tested in an environment in which the users change over the time as a result of e.g., changes in motivation or fatigue. Although some researchers test new algorithms with only offline data, both offline simulation and online experiments are necessary for effective algorithm design in closed-loop systems. In other words, offline simulation and cross-validation can be valuable methods to develop and test new algorithms, but only online analysis can yield solid evidence of BCI system performance [137,212,213]. Classification algorithms have traditionally been calibrated by users through supervised learning using a labeled data set. It is assumed that the classifier is able to detect the patterns of the brain signal recorded in online sessions with feedback. However, this assumption results in a reduction in the performance of BCI systems, because the brain signals are inherently non-stationary. In this regard, Shenoy et al. [214] described two main sources of non-stationarity. On the one hand, the patterns observed in the experimental samples during calibration sessions may be different from those recorded during the online session. On the other hand, progressive mental training of the users or even changes in concentration, attentiveness, or motivation may affect the brain signals. Therefore, adaptive algorithms are essential for improving BCI accuracy. Adaptation to non-stationary signals is particularly necessary in asynchronous and non-invasive BCIs [215,216]. Apart from the fact that supervised learning is not optimal for non-stationary signals classification, large data sets and, thus, long initial calibration sessions are usually required to achieve acceptable accuracy. Semi-supervised learning has been suggested to reduce training time and to update the classifier in the online session on a continuous basis [217]. In semi-supervised learning, the classifier is initially trained using a small labeled data set, after which the classifier is updated with on-line test data. In a realistic BCI scenario, the signal associated with the subject’s intentions is not usually known and the labels are not available. Either unsupervised learning or reinforcement learning can be applied for BCI adaptation when the labeled data set is not available. Unsupervised methods attempt to find hidden structures in unlabeled data, in order to classify them. Some unsupervised methods rely on techniques for co-adaptive learning of user and machine [218,219] or covariate shift adaptation [220]. Reinforcement learning methods are based on the fact that distinguishing EEG potentials are elicited when a subject is aware of an erroneous decision. These potentials are used as learning signals to prevent that error from being repeated in the future [221]. The adaptation generally results in enhanced performance. Nevertheless, it is worth highlighting that inherent risks exist in an adaptive BCI. A BCI that learns too fast may confuse the user, because training will take place in a changing environment [222]. In addition, adaptive procedures can hide some relevant signal features. Accordingly, there is a tradeoff between highly sensitive adaptation and feature extraction. Sensors 2012, 12 1241 Classifiers also have to face two main problems related to the pattern recognition task: the curse of dimensionality and the bias-variance tradeoff. The curse of dimensionality means that the number of training data needed to offer good results increases exponentially with the dimensionality of the feature vector [223]. Unfortunately, the available training sets are usually small in BCI research, because training process takes a long time and is a tiring process for users. The bias-variance tradeoff represents the natural trend of the classifiers towards a high bias with low variance and vice versa. Stable classifiers are characterized by high bias with low variance, while unstable classifiers show high variance with low bias. To achieve the lowest classification error, bias and variance should be low simultaneously. A set of stabilization techniques such as the combination of classifiers or regularization can be used to reduce the variance. The design of the classification step involves the choice of one or several classification algorithms from many alternatives. Several classification algorithms have been proposed such as k-nearest neighbor classifiers, linear classifiers, support vector machines, and neural networks, among others. The general trend prefers simple algorithms to complex alternatives. Simple algorithms have an inherent advantage because their adaptation to the features of the brain signal is inherently simpler and more effective than for more complex algorithms. Nevertheless, simple algorithms, whenever outperformed in online and offline evaluations, should be replaced by more complex alternatives [213]. Table 8. Summary of classification methods. Approach Properties Applications Generative model Bayesian analysis − Assigns the observed feature vector to the labeled class to which it has the highest probability of belonging − Produces nonlinear decision boundaries − Not very popular in the BCI systems [245–248] Linear LDA − Simple classifier with acceptable accuracy − Low computation requirements − Fails in the presence of outliers or strong noise. Regularization required − Usually two class. Extended multiclass version exits. − Improved LDA versions: BLDA, FLDA [179,230,231, 233–235] SVM − Linear and non-linear (Gaussian) modalities − Binary or multiclass method − Maximizes the distance between the nearest training samples and the hyperplanes − Fails in the presence of outliers or strong noise. Regularization required − Speedy classifier [131,228,230, 237,239–244] Non-linear k-NNC − Uses metric distances between the test feature and their neighbors − Multiclass − Efficient with low dimensional feature vectors. Very sensitive to the dimensionality of the feature vectors [227–229] ANN − Very flexible classifier − Multiclass − Multiple architectures (PNN, Fuzzy ARTMAP ANN, FIRNN, PeGNC) [200,215, 249–256] Sensors 2012, 12 1242 Finally, certain inherent dangers of classification algorithm usage should be pointed out. Although classification algorithms have clearly helped to characterize task relevant brain states, several pitfalls may occur when these algorithms are used by non-experts. Bias and variance of the estimated error of the algorithms, and their overfitting are the main source of difficulties [224]. If a classifier is overfitted, then it will only be able to classify the training data or similar data. Overfitting can be avoided by restricting the complexity of the classification procedure [224]. Classification error is estimated by means of cross validation. Once a classification algorithm is trained, the algorithm is validated on a validation data set, which should be independent of the training data set. This procedure is usually repeated several times, using different partitions of the sample data. The resulting validation errors are averaged across multiple rounds. This approach presents some inherent dangers that must be prevented, because some elements of the partition may not be independent of each other or may not be identically distributed, among other reasons [224].Next, this section presents the properties of a set of classifiers, in order to make it easier to choose an appropriate classifier for a given type of BCI. All classifier methods are listed in Table 8, along with their main properties. 7.1. K-Nearest Neighbor Classifier (k-NNC) K-nearest neighbor classifiers (k-NNC) are based on the principle that the features corresponding to the different classes will usually form separate clusters in the feature space, while the close neighbors belong to the same class. This classifier takes k metric distances into account between the test samples features and those of the nearest classes, in order to classify a test feature vector. The metric distances are a measure of the similarities between the features of the test vector and the features of each class. The advantage of taking k neighbors into account in the classification is that error probability in the decision is decreased. Some training samples may be affected by noise and artifacts, which may seriously influence the classification results. If a decision involving several neighbors is made, then it is less likely that an error will occur, because the probability of several simultaneous erroneous datum is much lower [225]. Rather than only the closest sample, if several k closest classes are considered, then a voting scheme is required to decide between competing choices. Since there are no reasons to assume that the distributions of those neighbors are homogenous, it is clear to see that the k-NNC has to assign different ranks to the nearest neighbors, according to their distances from the test example. Therefore, k-NNC needs to define a weighting function, which varies with the distance in such a way that the output value decreases as the distance between the test feature vector and the neighbor increases. The function defined by Equation (19) [226] meets this requirement: ݓሺሻ ൌ ቐ ݀ሺሻ െ ݀ሺሻ ݀ሺሻ െ ݀ሺଵሻ ݂݅ ݀ሺሻ ് ݀ሺଵሻ 1 ݂݅ ݀ሺሻ ൌ ݀ሺଵሻ (19) where, ݀ሺሻ denotes the distance of the i-th nearest neighbor from a test example. That is, ݀ሺଵሻ corresponds to the nearest neighbor and ݀ሺሻ to the furthest. The decision rule of k-NNC assigns the unknown examples to the class with the greatest sum of weights among its k nearest neighbors. k-NNC is not very common in BCI research, because this classifier is very sensitive to the dimensionality of the feature vector [227]. Nevertheless, k-NNC has been proven to be efficient with Sensors 2012, 12 1243 low dimension feature vectors. Also, k-NNC has been tested in a multiclass environment [228] and applied to cursor movements on a vertical axis, when classifying SCPs [229]. 7.2. Linear Discriminant Analysis (LDA) LDA is a very simple classifier that provides acceptable accuracy without high computation requirements. LDA is very common in the BCI community and is a good choice for designing online BCI systems with a rapid response, but limited computational resources. LDA provides relatively acceptable accuracy and has been used successfully in numerous BCI systems, such as P300 speller [179], multiclass [230], or synchronous [231] BCIs. Nevertheless, it can lead to completely erroneous classifications in the presence of outliers or strong noise [232]. LDA is usually applied to classify patterns into two classes, although it is possible to extend the method to multiples classes [230]. For a two-class problem, LDA assumes that the two classes are linearly separable. According to this assumption, LDA defines a linear discrimination function which represents a hyperplane in the feature space in order to distinguish the classes. The class to which the feature vector belongs will depend on the side of the plane where the vector is found (Figure 5). In the case of an N-class problem (N > 2), several hyperplanes are used. The decision plane can be represented mathematically as: (20 (ݓ்࢞ݓ ൌ ሻ࢞ሺ݃ where, w is known as the weight vector, x is the input feature vector and w0 is a threshold. The input feature vector is assigned to one class or the other on the basis of the sign of ݃ሺ࢞ሻ. Figure 5. Linear classifier and margins. The decision boundary is the thick line. (adapted from [232]). There are many methods to compute w. For example, w may be calculated as [233]: ߑൌݓ ିଵሺμଶ െ μଵሻ (21) where, μi is the estimated mean of class i and ߑ ൌ ଵ ଶ (ߑଵ ߑଶሻ is the estimated common covariance matrix; the average of the two class empirical covariance matrices. The estimators of the covariance matrix and of the mean are calculated as: ൌ ߑ 1 ݊െ1ሺݔ െ μሻሺݔ െ μሻ் ୀଵ (22) Sensors 2012, 12 1244 μ ൌ 1 ݔ ݊ ୀଵ (23) where, ࢞ is a matrix containing n feature vectors ݔଵ, ݔଶ,…,ݔ א Թௗ. Figure 6. Eigenvalue spectrum of a given covariance matrix (bold line) and eigenvalue spectra of covariance matrices estimated from a finite number of samples (N = 50, 100, 200, 500). Note that accuracy increases as the number of trials increase (adapted from [233]). The estimation of the covariance defined in Equation (22) is unbiased and has good properties under usual conditions. Nevertheless, it may become imprecise in some cases where the dimensionality of the features is too high compared to the number of available trials. The estimated covariance matrix is different from the true covariance matrix, because the large eigenvalues of the original covariance matrix are over estimated and the small eigenvalues are under estimated (Figure 6). It leads to a systematic error which degrades LDA performance [233]. For this reason, a new procedure has been proposed to estimate the covariance, improving the standard estimator defined in the Equation (22). The new standard estimator of the covariance matrix is given by: ߑሺߛሻ ൌ ሺ1െߛሻߑ ܫߥߛ) 24) The γ value is referred to as a shrinkage parameter and is tunable between 0 and 1. ν is defined as ݎݐ݁ܿܽሺߑሻ/݀ with d being the dimensionality of the features space. The selection of a shrinkage parameter implies a trade-off and is estimated on the basis of the input data [234]. Some improved algorithms have been introduced based on LDA such as Fisher LDA (FLDA) and Bayesian LDA (BLDA) [235]. In the first example, performance was improved by projecting the data to a lower dimensional space, in order to achieve larger intervals between the projected classes and, simultaneously, to reduce the variability of the data in each class. However, FLDA does not work well real N=500 N=200 N=100 N=50 0 50 100 150 200 0 50 100 150 200 250 Index of eigenvalue Eig env ect or Sensors 2012, 12 1245 when the number of features becomes too large in relation to the number of training examples. This is known as the small sample size problem [235]. The second modification can be seen as an extension of FLDA. BLDA solves the small sample size problem by introducing a statistical method known as regularization. The regularization is estimated through Bayesian analysis of training data and is used to prevent overfitting of high dimensional and possibly noisy datasets. Overfitting means the classifier has lost generality and is therefore undesirable in a classifier. If a classifier is overfitted, then it is only able to classify the training data or similar data. In comparison to FLDA, the BLDA algorithm provides higher classification accuracy and bitrates, especially in those cases where the number of features is large [235]. Additionally, BLDA requires only slightly more computation time, which is a crucial requirement in real BCI systems. 7.3. Support Vector Machine (SVM) SVM is a classifier that, in a similar way to LDA classifiers, constructs a hyperplane or set of hyperplanes, in order to separate the feature vectors into several classes. However, in contrast to LDA, SVM selects the hyperplanes that maximize the margins, that is, the distance between the nearest training samples and the hyperplanes [236]. The basis of SVM is to map data into a high dimensional space and find a separating hyperplane with the maximal margin [237] according to Cover’s theorem on the separability of patterns [238]. Cover’s theorem states that a complex classification problem cast in a high-dimensional nonlinear space is more likely to be linearly separable than in a low-dimensional nonlinear space. Also, as for linear analysis classifier, an SVM uses regularization, in order to prevent the classifier from accommodating possibly noisy datasets. SVM has been used to classify feature vectors for binary [239,240] and multiclass problems [228,230]. It has also been successfully used in a large number of synchronous BCIs [131,230,240]. Such a classifier is regarded as a linear classifier, since it uses one or several hyperplanes. Nevertheless, it is also possible to create a SVM with non-linear decision boundary by means of a kernel function K(x, y). Non-linear SVM leads to a more flexible decision boundary in the data space, which may increase classification accuracy. The kernel that is usually used in the BCI field is the Gaussian or Radial Basis Function (RBF): ܭሺݔ ,ݕሻ ൌ exp ቆെԡݔെݕԡଶ 2ߪଶ ቇ (25) The Gaussian SVM has been applied in BCIs to classify P300 evoked potentials [241–243]. SVM has been widely used in BCI, because it is a simple classifier that performs well and is robust with regard to the curse of dimensionality, which means a large training set is not required for good results, even with very high dimensional feature vectors [228]. These advantages come at the expense of execution speed. Nevertheless, SVM is speedy enough for real-time BCIs [243,244]. 7.4. Bayesian Statistical Classifier Bayesian statistical classifier is a classifier which aims to assign, with the highest probability, an observed feature vector x from its class y. The Bayes’ rule is used to obtain the a posteriori probability ܲሺݕ|ݔሻ that a feature vector has of belonging to a given class. Assuming, for example, two classes L Sensors 2012, 12 1246 and R corresponding to imaginary left and right movements of the hand, the a posteriori probabilities of each class are computed using the Bayes’ rule as: ܲሺݕ|ݔሻ ൌ ܲሺݕሻܲሺݔ|ݕሻ ܲሺݔሻ ൌ ܲሺݕሻܲሺݔ|ݕሻ ܲሺݔ|ܮሻܲሺܮሻ ܲሺݔܴ|ሻܲሺܴሻ ൌ ܲሺݔ|ݕሻ ܲሺݔ|ܮሻ ܲሺݔܴ|ሻ (26) Typically, it is assumed that the a priori probabilities are equal, P(y) = P(L) = P(R) = 0.5, since it is supposed the user has no predilection for any movement. In order to calculate the probabilities ܲሺݕ|ݔሻ, it is usually supposed that a Gaussian statistical distribution applies to the features for each class, although it may also be assumed that the distribution is a weighted mixture of Gaussian distributions [245]: ݓൌ ሻݕ|ݔሺܲ ெ ୀଵ ܲሺݔܿ|ሻ (27) where, wi is the weight of each Gaussian prototype and M is the number of prototypes. Two ways are feasible to estimate the Gaussian prototypes mixture [245]. The first is to divide the feature space in several equally sized regions and calculate the mean and variance of the Gaussian prototypes in each area from training data. The set of Gaussian prototypes is equally weighted and the weights wi are equal to ଵ ெ . The second uses a Gaussian mixture models (GMM). The different weights wi and the mean, variance, and covariance matrices that define each Gaussian prototype, are calculated by the expectation maximization (EM) algorithm. EM algorithm is an iterative procedure which guarantees the maximum likelihood or maximum a posteriori (MAP) estimates of the parameters in the statistical model. Lui et al. [246] made GMM adaptive to significant changes in the statistical distribution of the data during long-term use. In these improvements, the initial mean, variance and covariance of each class is updated over time using a specific number of recent trials. Bayesian statistical classifiers are not very popular in the BCI community. Nevertheless, they have been used for classifying motor imagery [247] or visual P300 evoked potentials [248]. 7.5. Artificial Neural Network (ANN) ANNs are non-linear classifiers that have been used in many applications, in a wide variety of disciplines such as computer science, physics, and neuroscience. The idea of ANNs is inspired in how the brain processes the information. The purpose is to mimic brain activity that immediately solves certain problems, which a conventional computer program processes poorly. For example, ANNs are widely used in pattern recognition, because they are capable of learning from training data. The ability to learn from examples is one of most important properties of ANNs. Once trained, the ANNs are capable of recognizing a set of training data-related patterns. ANNs are therefore associated with BCI applications, since pattern recognition is performed to ascertain user intentions. An ANN comprises a set of nodes and connections that are modified during the training process. The ANN is fed on a set of training examples and the output is observed. If the output is incorrect, then the internal weights are modified by the training algorithm to minimize the difference between desired and actual output. This training continues until the network reaches a steady state, where no further Sensors 2012, 12 1247 significant improvement is achieved. In this state, not only should the ANN produce correct outputs for all examples of the training set, but also for inputs that were not encountered during training. From a mathematical point of view, ANNs define a mapping from an input space to an output space, that can be described as a vector-valued function ࢟ ൌ ࢌሺ࢞ሻ, where both x and y may be of any dimensionality. The mapping function f is a combination of mappings, which are individually performed by single nodes or neurons. Each neuron processes the information non-linearly and the resulting mapping is therefore non-linear. This property is important, especially in those cases where the physical mechanism that generates the input signal is non-lineal. One of the most well-known ANN structures is the multilayer perceptron (MLP) introduced by Rumelhart and McClelland in 1986. MLPs are very flexible classifiers that can classify any number of classes and adapt to numerous kinds of problems. In the field of BCIs, MLP have been applied to classify two [249], three [200], and five [250] different tasks, and to design synchronous [251] and asynchronous [215] BCIs. Moreover, MLP has been used for preprocessing EEG signals before the feature extraction step rather than the classification step, in order to improve the separability of EEG features [252]. Besides MLP, different types of ANN architecture have been used in the design of BCI systems such as Probabilistic Neural Networks (PNN) [253, 254], Fuzzy ARTMAP Neural Networks [255], Finite Impulse Response Neural Networks (FIRNN) [251] or Probability estimating Guarded Neural Classifiers (PeGNC) [256]. 8. BCI Applications BCIs offer their users new communication and control channels without any intervention of peripheral nerves and muscles. Hence, many researchers focus on building BCI applications, in the hope that this technology could be helpful for those with severe motor disabilities. Various BCI applications have very recently been developed thanks to significant advances in the field of EEG-based BCI. EEG signals are used by most BCI applications, because they offer an acceptable signal quality that combines low cost and easy-to-use equipment. Thanks to BCI applications, it is hoped that the quality of life of severely disabled people can be improved. Likewise, the attention given by caregivers will be less intensive, reducing its costs and making the life of relatives less onerous. Moreover, BCI applications potentially represent a powerful tool for revealing hidden information in the user’s brain that cannot be expressed. The main target populations for BCI applications fall into three classes. The first group includes Complete Locked-In State (CLIS) patients who have lost all motor control, because they may be at a terminal stage of ALS or suffer severe cerebral palsy. The second group comprises Locked-In State (LIS) patients who are almost completely paralyzed, but with residual voluntary movement, such as eye movement, eye blinks, or twitches with the lip. The third group of potential BCI users includes abled bodied people and those with substantial neuromuscular control, particularly speech and/or hand control. BCI have little to offer to the third group, because they can send the same information much more quickly and easily via other interfaces, rather than a BCI. Despite this, BCIs are increasingly used by healthy people in neuromarketing and video games as a tool to reveal affective information of Sensors 2012, 12 1248 the users, which cannot be so easily reported through conventional interfaces. Likewise, BCI can be used for some people that suffer from neurological disorders such as schizophrenia or depression. The level of impairment of the potential target population is related to the performance of a BCI system. Kübler et al. [257] reported a strong correlation between physical impairment and BCI performance. CLIS patients were unable to control a BCI. Voluntary brain regulation for communication was only possible in LIS patients. However, considering only LIS patients, this relationship between physical impairment and BCI performance disappeared. Figure 7 shows the relationship between BCI application areas and BCI information transfer rates and user capabilities. Figure 7. Relationship between BCI application areas, BCI information transfer rates and user capabilities. Horizontal axis: information transfer rate that would make the application controllable. Vertical axis: the degree of capability. It is currently unclear whether BCI technology will ever outperform other established technologies that include eye or muscle-based devices. Currently the latter devices tend to be easier to use and offer better benefit/cost ratios [258,259]. For example, the detection of eye movement is quicker, easier, and more accurate than the detection of ERP modulations. A spelling rate of 10 words per minute can be obtained with unimpaired eye movement, by means of an eyetracker [260]. In that regard, hybrid BCI systems have been proposed to improve performance. They are the combination of two different kinds of BCIs or the combination a BCI with other existing assistive technology [261]. Unless the performance of BCI systems improves considerably, BCI as assistive technology may only be especially attractive for severely disabled people, when other technologies are unsuitable. At present, LIS patients and those likely to develop CLIS constitute the principal candidates for BCI. Despite the low information transfer rates provided by BCI, the high grade of disability among LIS patients force them to use a BCI rather than more reliable conventional interfaces, such as muscle or eye-gaze based system. Eye-gaze control constraints in some LIS patients are an important issue, because they are obliged to use BCIs that does not depend on eye-gaze control [262,263]. Also, eye-gaze control constraints make some BCI applications more difficult, such as steering a wheelchair. Nowadays, there are a vast number of very different BCI applications, such as word processors, adapted web browsers, brain control of a wheelchair or neuroprostheses, and games, among others. However, most applications have solely been designed for training or demonstration purposes. Despite the most recent significant advances in BCI technology, there are still many challenges to employing BCI control for real-world tasks [264]: (i) the information transfer rate provided by BCIs is too low for user capabilities information transfer rate BCI Communication Environmental control Locomotion Neuroprosthesis Entertainment LIS patients CLIS patients Healthy users and non-severely disabled people Sensors 2012, 12 1249 natural interactive conversation, even for experienced subjects and well-tuned BCI systems; (ii) the high error rate further complicates the interaction; (iii) BCI systems cannot be used autonomously by disabled people, because BCI systems require assistants to apply electrodes or signal-receiving devices before the disabled person can communicate; (iv) a BCI user may be able to turn the BCI system off by means of brain activity as input, but usually cannot turn it back on again, which is termed the “Midas touch” problem; and (v) handling BCI applications demands a high cognitive load that can usually be achieved by users in quiet laboratory environment, but not in the real world. Nevertheless, despite all these challenging difficulties, the first steps on the path to long-term independent home use of BCIs have already been taken [12]. Before describing the practical usage of BCI applications, it is worth considering the distinction between BCIs and their applications [8]. As a tool that executes a specific function, particular BCI specifications correspond to the way it performs that function. These specifications can therefore be applied to wide variety of applications, even though the function remains unchanged. The important thing in BCI evaluation is its performance when executing its specific function. In contrast, applications are described in terms of the tools they employ and the purposes they serve. Therefore, BCI evaluation focuses on how well it performs its purpose. In other words, the term BCI refers to the system that records, analyses, and translates the input into commands and the term application denotes the environment in which the BCI estimated output commands are applied. Consequently, the evaluation procedures for BCI systems and their applications differ in each case. The following sub-sections briefly describe BCI applications, classified into five main areas: communication, motor restoration, environmental control, locomotion and entertainment. 8.1. Communication BCI applications for communication deal with severe communication disabilities resulting from neurological diseases. This kind of application probably represents the most pressing research in the field of BCI, because communication activity is essential for humans. Applications for communication purposes outline an operation that typically displays a virtual keyboard on screen, where the user selects a letter from the alphabet by means of a BCI. The distinguishing element in each approach is usually the BCI and the type of control signal. Voluntary control of SCPs may be used for letter selection. With extensive training, completely paralyzed patients are able to produce positive and negative changes in their SCP to drive the vertical movement of a cursor [117]. Based on this kind of control signal, Birbaumer et al. [265] developed a spelling device with an on-screen display, which used a cursor to select letters of the alphabet. Trials involving two patients at advanced stages of ALS showed that they achieved a rate of about 2 characters per minute when writing text messages. Other types of control signals, such as detection of eye blinks [266], which normally represent an artifact in EEG signals, or classification of three mental tasks [215], are also used to select the blocks or characters in a virtual keyboard. Both approaches are nearly the same apart from the control signal. In both cases, the virtual keyboard consisted of a total of 27 symbols, 26 English letters plus the space to separate words, organized in a three row by nine column matrix. Likewise, both applications were based on the same protocol of writing a single letter, which required three steps. Firstly, the whole keyboard was divided into three Sensors 2012, 12 1250 blocks, each with nine letters each. Then, the user could select a set of nine letters by producing a single, two or three eye blinks [266] or imagining one of three available tasks [215] depending on the case. After the first selection, the set of nine letters was distributed into three subsets, each with three letters, and once again the user again selected one of them. Finally, at the third level, the user chose a single letter amongst the three remaining symbols. The correct spelling rate of each speller was one character per minute using blinks [266] and 2.73 characters per minute for three mental tasks [215]. Obermaier et al. [267] also designed a letter spelling based on standard Graz-BCI which also included a virtual keyboard. The letter selection protocol is very similar to the approaches discussed above, except that the entire alphabet consisted of 32 letters and was divided into two halves at each step. In this case, the user chooses either subset of letters by EEG modulation through mental hand and leg motor imagery. The spelling rate achieved by three healthy users varied between 0.5 and 0.85 letters per minute. This is a lower rate than in previous cases, nevertheless, it appears easy to increase the number of letters spelled per minute just by expanding the number of classes to more than two. P300 event-related brain potentials are also very popular in BCI letter spelling applications. P300-based BCIs have been proven sufficiently suitable for ALS patients in the early and middle stages of the disease [268]. Besides, this kind of BCI is very handy because the P300 response occurs spontaneously and consequently does not require substantial training. Furthermore, recent progress with P300-based spellers have allowed the development of commercial applications available to the general public [269]. One of the best-known P300 spellers was designed by Farwell and Donchin in 1988 [123]. In this speller, the 26 letters of the alphabet, together with several other symbols and commands, are displayed on-screen in a 6 × 6 matrix (Figure 8) with randomly flashing rows and columns. Then, the user focuses attention on the screen and concentrates successively on the characters to be written, while the EEG response is monitored. Two P300 are elicited for each looked-for element on the matrix, when the desired row or column flashed, thereby allowing the system to identify the desired symbol. The results of the Farwell-Donchin speller trials involving 4 healthy people yielded an acceptable spelling rate of about 2 characters per minute. Figure 8. Original P300 speller. Matrix of symbols displayed on a screen computer which serves as the keyboard or prosthetic device (adapted from [123]). The Farwell-Donchin speller provides a relatively high rate and accuracy, but its precision can be improved by reducing perceptual errors in the Farwell-Donchin paradigm [270]. Perceptual error happens when a P300 response is elicited due to flashing rows or columns adjacent to the target symbol, an issue which is its major source of error. Hence, a new letter distribution was presented to Sensors 2012, 12 1251 overcome this problem (Figure 9) [270]. The idea is to have several regions flashing instead of using rows and columns. The characters are placed into a two-level distribution. At the first level, the characters are distributed into seven groups, each with seven characters, which are also flashing randomly. The group containing the target character is found by P300 detection. At the second level, the characters in the detected group are repositioned and the level one procedure is repeated, and so on until the target character is finally selected. Figure 9. The proposed region-based paradigm for the improved P300 speller: (a) The first level of intensification where each group contains up to seven characters; and (b) One region is expanded at the second level (adapted from [270]). (a) (b) Townsend et al. [137] presented a newly enhanced BCI based on a checkerboard paradigm instead of the standard row/column paradigm introduced by Farwell and Donchin. In this new approach, the standard matrix containing the targets was superimposed on a checkerboard. Trials with advanced ALS patients and healthy people showed a significantly higher mean accuracy for the checkerboard paradigm than for the row/column paradigm. Ahi et al. [271] also recently improved the Farwell-Donchin P300 speller by introducing a dictionary to decrease the number of misclassifications in the spelling. The dictionary was used for checking the candidate word proposed by the classifier of P300 responses. In case of misspelling, the dictionary gave a certain number of suggestions from which the system could select. Additionally, in order to reduce the probability of misspelling due to perceptual errors, the usual letter position in the matrix was changed according to the analysis of word similarities in the constructed dictionary. All previous P300 spellers are based on the recording of visual event-related brain potentials. However, there is no sense in using visual stimuli in cases of severely paralyzed patients with impaired vision or poor control over eye movements. In these cases, auditory stimulation is used in order to make P300 spellers suitable for this group of patients [129,272–274]. Other important applications of communication-related BCIs are Internet browsers adapted to users with severe disabilities because, over the last decade, the Internet has become a very important part of daily life. In this area, “Descartes” is one of the first EEG-controlled web browsers which can be operated by SCPs [275]. Its browser interface is based on arranging the links alphabetically in a dichotomous decision tree, where the user selects or rejects each item, producing positive or negative SCP shifts. “Descartes” presents the shortcoming that only a limited number of web pages can be A B C D F G E H I J K M N L V W X Y 1 2 Z 3 4 5 6 8 9 7 O P Q R T U S 0 ~ ! @ $ % # ^ * ( ) = + _ Sensors 2012, 12 1252 browsed, because the user receives a number of predefined links arranged in a tree at the start of the web surfing. Besides, graphical links cannot be chosen since the textual label is used to identify the link. A more advanced prototype, called “Nessi”, overcomes these shortcomings thanks to a better user interface [276]. Colored frames are placed around links or selectable items on the web page instead of arranging the links in a tree. More recently, evoked potentials are also used to enhance browser functionality. Mugler et al. [128] built an Internet browser with P300 control where the options are all presented as icons in an 8x8 matrix. Jinghai et al. [103] developed a browser based on VEPs. One of the advantages of ERPs is that they occur quickly and can lead to relatively high web surfing speeds. 8.2. Motor Restoration Spinal cord injury (SCI) or other neurological diseases with associated loss of sensory and motor functions dramatically decrease the patient’s quality of life and create life-long dependency on home care services. Motor restoration may alleviate their psychological and social suffering. Restoring movement, such as grasping, is feasible in quadriplegic patients through neuroprostheses guided by functional electrical stimulation (FES). FES compensates for the loss of voluntary functions by eliciting artificial muscle contractions. Electrical currents generate artificial action potential by depolarizing intact peripheral motor nerves that innervate the targeted muscle and cause a muscle contraction (see [277] for a review). EEG-based BCI can be used to generate a control signal for the operation of FES, because EEG signals are unaffected by electrical activation of upper extremity muscles [278]. Thanks to their merging of BCI and FES, Pfurtscheller et al. [279] developed an application where a tetraplegic patient, suffering from a traumatic spinal cord injury, was able to control paralyzed hands to grasp a cylinder. In that application, the patient generated beta oscillations in the EEG by foot movement imagery. Then, the BCI analyzed and classified the beta burst and the output signal was used to control the FES device that activated the extremity. Also, FES has been used for rehabilitation training after a stroke. Hu et al. [280] developed a combined FES-robot system which was continuously driven by the user’s residual electromyography on the affected side for wrist joint training after a stroke, in order to involve the user’s own neuromuscular effort during the training. FES has been proven to be an effective way to restore movement. Nevertheless, FES requires the use of residual movements, which are not possible in severely injured patients. For this reason, some groups have started to explore approaches that couple neuroprostheses and BCI without FES intervention. Pfurtscheller et al. [281] demonstrated that a tetraplegic patient, whose residual upper-limb muscle activity was restricted to the left biceps, due to an upper spinal cord injury, could effectively control a hand orthosis using changes in Rolandic oscillations, which were produced by motor imagery. A lengthy training period was required to use this application. However, the patient was finally able to open and close the hand orthosis almost without any errors. Some years later, the same group validated the coupling of EEG-based BCIs and an implanted neuroprosthesis giving further evidence that BCI is a feasible option for the control of a neuroprostheses [282]. In this study, BCI classified distinctive EEG-patterns that involved power decreases in certain specific frequency bands. These patterns were generated by the user from mental imagery of his paralyzed left hand in motion. More recently, ERPs are also used to provide motor restoration. Muller et al. [13] presented a novel neuroprosthetic device for the restoration of the grasp function for people spinal cord injuries. This Sensors 2012, 12 1253 neuroprosthetic device consisted of a dual-axis electrical hand prosthesis controlled by BCI based on four-class SSVEPs. Hence, it is possible to select only four movements according to the four LEDs flickering in different frequencies. The user’s gaze shifted between the different LEDs in order to select a movement. One light on the finger index flickering at 6 Hz and another light on the pinky finger flickering at 7 Hz served to turn the hand in supination or pronation. The two remaining lights on the wrist flickering at 8 Hz and 13 Hz represented the orders to open and close each hand. Within the field of BCI application in motor restoration, BCI systems have been also applied for movement reconstruction in patients with severe post-stroke motor disability. BCI training is hypothesized to provide feedback to sensorimotor cortex and, by doing so, movement is restored as cerebral pathways reorganize to link up motor commands with motor movements. Buch et al. [283] developed a BCI system that used MEG activity evoked by patient intent to move a completely paralyzed hand, in order to control grasping motions of a mechanical orthosis attached to the affected hand. Thanks to the hand prosthesis attached to the paralyzed hand and using visual feedback, the patient could learn to open a hand by increasing SMR over the injured hemisphere and to close the hand by decreasing it. MEG provides a much larger and more localized SMR response, which means that even a digit finger may be controlled [284]. MEG-based BCI is too expensive for widespread applications. For that reason, Broetz et al. [285] proposed a combination of MEG and EEG-based BCIs. Initially, the MEG-based BCI was used to boost rehabilitation training success. Later, the user continued rehabilitation training with an EEG-based BCI; a more affordable technology than MEG. Finally, the patient practiced physiotherapy training. The results of this study suggest that the combination of BCI training with goal-directed active physical therapy improves the motor abilities of chronic stroke patients. In similar experiments, Caria et al. [286] reaffirmed the success of a combination of BCI training and physiotherapy. This study encourages further research on the role of BCIs in brain plasticity and post-stroke recovery. 8.3. Environmental Control One of the main goals of BCI-based applications is to achieve maximum independence for the patient, despite any motor disability. People who suffer severe motor disabilities are often homebound and for this reason, environmental control applications focus on the control of domestic devices such as TV, lights or ambient temperatures. Apart from improving the quality of life of severely disabled people, assistive devices mean that the tasks of the caregiver are less intensive, costs are reduced, and the life of relatives is less onerous. Cincotti et al. [14] presented a pilot study dealing with the integration of BCI technology into the domestic environment. In this study, fourteen patients with severe motor disabilities, due to progressive neurodegenerative disorders, tested a device that provided environmental control through an interface designed to support different levels of motor capacities for each user. Typical peripherals such as keyboard, mouse or joystick were offered to allow the device control through upper limb residual motor abilities. Head trackers and microphones for voice recognition were also available in cases of people with impaired limbs but intact neck muscles and comprehensive speech. Lastly, in cases of totally disabled people, the system could be controlled by voluntary modulations of sensorimotor rhythms recorded by the EEG-based BCI. Thereby, the application offered the patient Sensors 2012, 12 1254 different access modalities that matched their gradual loss of motor abilities due to progressive neurodegenerative diseases. As output devices, the system allowed the use of a basic group of domestic appliances such as lights, TV and stereo sets, a motorized bed, an acoustic alarm, a front door opener, and a telephone, as well as wireless cameras to monitor the surrounding environment. Invasive techniques have also been proposed in environmental control applications. Hochberg et al. [287] implanted BrainGate sensors in the primary motor cortex of a tetraplegic patient to control a cursor. The initial trials yielded promising results, where the patient could handle e-mail applications or operate devices such as a television by imagining limb motions, even while conversing. 8.4. Locomotion BCI applications that allow disabled people to control a means of transportation represent an important field in their use. Thanks to these applications, people suffering from paraplegia or with other physical impairments can autonomously drive a wheelchair, making them more autonomous and improving their life quality. Portability is a necessity for these kinds of applications. Hence the use of a BCI based on EEG recording is enforced. EEG signals are typically very noisy and are highly variable, which means a relatively long time between commands that will be of high uncertainty. Therefore, the main challenge is to achieve sufficient accuracy in driving as well as reaching real-time control, in spite of the ultra-low information transfer rates provided by BCI. For this reason, some studies on BCIs proposed invasive techniques to record EEG signals, because they achieved more spatial resolution and reduced noise. Serruya et al. [288] experimented with monkeys implanting an electrode array in the motor cortex. These initial experiments showed that the monkeys were able to move a computer cursor to any position, opening up new human applications. However, the risks related to invasive BCIs lead research towards building non-invasive applications for human use. Some pilot experiments concerned with locomotion illustrate the feasibility of using EEG signals for continuous control of a mobile robot in an indoor environment with several rooms, corridors, and doorways [215,289]. The results of these experiments opened the possibility for physically disabled people to use a portable EEG-based BCI for controlling wheelchairs. To the best of our knowledge, in 2005, Tanaka et al. [127] presented the first application of wheelchair control using only EEG. In this study, the surrounding floor was divided into squares between which the user decided to move by imagining left or right-limb movements. Driven by user decisions, the wheelchair therefore moved from one square to another. Tests with six healthy subjects were quite encouraging and demonstrated the viability of wheelchairs control solely through the use of EEG signals. In wheelchair control by BCI-based systems, the usual problems are the infrequent control signal and the low information transfer rate and accuracy provided by a BCI. In that respect, some improvements have been presented over the past few years. Synchronous P300-based BCIs have been introduced in order to assure better accuracy. Likewise, to overcome the usual low bit rate in BCIs, the systems have been endowed with certain autonomy, decreasing the number of interactions required. Rebsamen et al. [290] designed a simplified wheelchair control by constraining the movements to guidepaths defined by the patient or a helper. These guidepaths were attached to a specific point in the environment and stored by the system. The user selected the destination through a P300-based BCI and the wheelchair automatically followed the path. The user only had to decide when the wheelchair Sensors 2012, 12 1255 would stop. For path guidance, the system steering the wheelchair had to be kept informed of its localization uninterruptedly. To that effect, the wheelchair relied on an odometer and a bar-code scanner to read bar-code patterns previously placed on the floor along the paths. Some years later, the system was improved to ensure safer control. Two faster BCIs based on P300 and the µ/β rhythm were employed, allowing the user to stop the wheelchair more quickly [291]. Both applications were tested with healthy people. The main disadvantage found in the preceding approaches is that the control assistance has little flexibility and is not capable of dealing with unknown and populated scenarios. Iturrate et al. [292] overcame this shortcoming by making the system create a dynamic reconstruction of the surrounding scenario. Other studies suggested that help should only be available in those cases where the user experienced more difficulties driving the wheelchair e.g., in a narrow corridor [216,293,294]. Three levels of assistance may be possible in the shared control: collision avoidance, obstacle avoidance and orientation recovery, which are only activated as required by the user [293]. Before executing the user’s steering commands, the share control evaluates the situation from the data provided by a set of laser scanners. Scanners inspect the environment and detect potential obstacles or walls. 8.5. Entertainment Entertainment-orientated BCI applications have typically had a lower priority in this field. Until now, research into BCI technology has usually focused on assistive applications, such as spelling devices, wheelchair control or neuroprostheses rather than applications with entertainment purposes. However, interest in entertainment applications has arisen over the recent years due to the significant advances in this technology. In fact, improvements in its performance have opened the way to extending BCI use to non-disabled people. BCIs create a new interaction modality which may turn video games into even more challenging and attractive experiences. Additionally, BCI may provide a way of accessing knowledge on the user’s experiences, thereby improving games through information from brain activity. BCIs can report when the gamer is bored, anxious or frustrated with the aim of using this knowledge for designing future games [295]. Figure 10. Pacman game. The gamer has to move through the maze to reach the exit in the right wall. The shortest path is marked with gray track marks, but the gamer can decide to run the rest of maze to receive additional credits (adapted from [296]). Nose: Direction of next move. Discrimination Line: Indication of competing laterality intentions S P il w f n g r d ( c O S S b L ti to im u c E S 1 a a Sensors 201 Entertain Pacman, Pon llustration, with the aim filled with r nose points game was d recording te External designed a s SSVER). T could be mo On the one SSVERs. A SSVER. Th because it re Lalor et al. ight-rope w o phase-rev mplemented useful tool f All previ context. Ho Emotiv [5] Spirit Moun 14 electrode an Applicati applications 2, 12 nment-orient ng and simi Figure 10 s m of giving red and gre to the direc developed, chniques fo evoked pot simple fligh This simulat oved to the hand, the c And on the e results of equired littl [299] prese walker in b versing che d the “Min for attention Figure 11 ous exampl wever, ther has already ntain Demo es, the so-ca ion Program is made ted BCI ap ilar games s shows a scre the gamer en color fro ction in wh in order t or complex c tentials have ht simulator tor was ver left or the r control com other hand, f the trials w e or no train ented the “ balance. Th eckerboard ndGame” ba n training, b 1. (a) Emotiv (a) les involved re are some y developed o Game, am alled EPOC mming Inter much sim pplications so that they eenshot of a enough tim om bottom hich the use o illustrate control task e been also controlled ry modest a right only. T mmand (righ , the select with able-bo ning since t MindBalan he applicati patterns. B ased on P3 ecause the P v EPOC neu d experimen e companie d a numero mong others C neurohead rface (API) mpler. The have adap y may be pla a Pacman ga me to perfor up as the p r intends to that it is ks [297]. used for ga by a BCI ba and only tw Two metho ht or left) w tion was ide odied partic the system c nce” game w ion was al By means o 00 events. P300 potent uroheadset ntal games t es preparing us set of B s. Furtherm dset (Figure ). Thanks to company pted very w ayed throug ame. The Pa rm a contro player’s inte o take the P possible to ame implem ased on Ste wo control a ods were tes was detecte entified tak cipants show capitalized where six h lso based o of other ki It was sug tials are als [5]; (b) Neu that have on g commerci BCI-based g more, the co e 11(a)) whi o this API, Neurosky well-known gh motor im acman mak ol command ention to tu Pacman. In o successfu mentations. eady State V actions were sted over th d, accordin king into ac wed that the on the natu healthy user on SSVEP inds of ER ggested that o attention m urosky Min (b) nly been fo ial BCI gam games, such ompany sell ich can be b the develop [6] also m n video gam magery [296 kes one step d. The Pacm urn rises an another stu ully apply Middendor Visual-Evok e possible. he airplane c ng to the str ccount the f e last one w ural occurrin rs were ask generated RPs, Finke t this game markers. ndwave [6]. ) und useful mes for fut h as Cortex ls a low co bought acco pment of th markets the 125 mes such a 6]. By way o each 1.5–2 man’s head nd the yellow udy, a pinba non-invasiv rf et al. [298 ked Respons The positio control trial rength of th frequency o was preferre ng response ked to keep in respons et al. [300 e would be in a researc ture market x Arcade an ost BCI wit ompanied b he BCI-base e Mindwav 56 as of 2 s is w all ve 8] se on ls. he of ed es. a se 0] a ch ts. nd th by ed ve Sensors 2012, 12 1257 neuroheadset (Figure 11(b)) with software applications that can respond to user brainwaves or mental states. Likewise, it provides a set of software tools for developers. Also, large software companies such as Microsoft have shown interest in BCI research, exploring the development of pilot novel applications that use BCIs [301]. 8.6. Other BCI Applications BCI systems have also been used in a broad variety of applications beyond the traditional areas of communication, motor restoration, environmental control, locomotion, and entertainment. The ability of BCI feedback to induce cortical plasticity may be the basis for medical applications. Users can acquire selective control over certain brain areas by means of neurofeedback, with the aim of inducing behavioral changes in the brain. Neurofeedback provided by a BCI system may improve cognitive performance [302,303], speech skills [304], affection [305], and pain management [306], and has been used in the treatment of mental disorders, such as epilepsy [307,308], attention deficit [309], schizophrenia [310], depression [311], alcohol dependence [312], or paedophilia [313]. On the other hand, brain signal recordings can be used in an assessment of brain functions to evaluate their status in health and disease [314]. The opportunity to examine brain signals can also be commercially exploited. Neuromarketing is a relatively young field of research that applies neuroscientific methods to marketing research. To date, few neuromarketing studies have been conducted, although some evidence has been found to suggest that neuroimaging could have a role in several areas of marketing [315–318]. Neuromarketing may provide a more efficient trade-off between costs and benefits. Product concepts could be tested by means of neuromarketing, removing those that are not promising at the start of the manufacturing process. This would lead to a more efficient distribution of sources, because only the more promising products would be developed [319]. In addition, neuromarketing may be a source of more accurate information on the underlying preferences of the users, rather than data from standard market research studies [319]. Neuroimaging may reveal hidden information on consumers’ true preferences that cannot be explicitly expressed. The brain’s response to advertisements could be measured and the effectiveness of advertising campaigns could therefore be quantified. Despite it being an emerging field, several companies such as Neurofocus [320], Neuroconsult [321], Neuro Insight [322] or EmSense [323], among others, currently offer neuromarketing services. It is also attracting increasing attention among researchers. The field has raised some ethical issues concerning this technology, in as much as it may be able to manipulate the brain and consumer behavior [324]. 9. Conclusions This article has reviewed the state-of-the-art of BCI systems, discussing fundamental aspects of BCI system design. The most significant goals that have driven BCI research over the last 20 years have been presented. It has been noted that many breakthroughs were achieved in BCI research. Different neuroimaging approaches have been successfully applied in BCI: (i) EEG, which provides acceptable quality signals with high portability and is by far the most usual modality in BCI; (ii) fMRI and MEG, which are proven and effective methods for localizing active regions inside the brain; (iii) NIRS, Sensors 2012, 12 1258 which is a very promising neuroimaging method in BCI; and (iv) invasive modalities, which have been presented as valuable methods to provide the high quality signals required in some multidimensional control applications e.g., neuroprostheses control. A wide variety of signal features and classification algorithms have been tested in the BCI design. Although BCI research is relatively young, many advances have been achieved in a little over two decades, because many of these methods are based on previous signal processing and pattern recognition research. Many studies have demonstrated the valuable accuracy of BCIs and provided acceptable information bit rate, despite the inherent major difficulties in brain signal processing. Accordingly, user training time has been significantly reduced, which has led to more widespread BCI applications in the daily life of disabled people, such as word processing, browsers, email, wheelchair control, simple environmental control or neuroprostheses among others. In spite of the recent important advances in the BCI field, some issues still need to be solved. First, the relative advantages and disadvantages of the different signal acquisition methods are still unclear. Their clarification will require further human and animal studies. Second, invasive methods need further investigation to deal with tissue damage, risk of infection, and long-term stability concerns. Electrodes that contain neurotropic mediums that promote neuronal growth and wireless transmission of neuronal signals recorded have already been proposed. Third, the electrophysiological and metabolic signals that are best able to encode user intent should be better identified and characterized. The majority of BCI studies have treated time, frequency, and spatial dimensions of brain signals independently. These signal dimension interdependencies may lead to significant improvement in BCI performance. Fourth, information bit rate provided by current BCIs is low for effective humanmachine interaction in some applications. Exogenous-based BCI may provide much higher throughput. Fifth, the unsupervised adaptation is a key challenge for BCI deployment outside the lab. Some moderately successful adaptive classification algorithms have already been proposed. And finally, most BCI applications are at the research stage and they are not ready to be introduced into people’s homes for continuous use in their daily life. In addition to their low information transfer rates and variable reliability, most current BCI systems are uncomfortable, because the electrodes need to be moistened, the software may require initiation, and the electrode contacts need continuous correction. An easy-to-use P300-based BCI with remote monitoring using a high-speed internet connection has already been proposed to reduce dependence on technical experts. The latest advances in BCI research suggest that innovative developments may be forthcoming in the near future. These achievements and the potential for new BCI applications have obviously given a significant boost to BCI research involving multidisciplinary scientists e.g., neuroscientists, engineers, mathematicians, and clinical rehabilitation specialists, among others. Interest in the BCI field is expected to increase and BCI design and development will in all probability continue to bring benefits to the daily lives of disabled people. Furthermore, recent commercial interest within certain companies suggests that BCI systems may find useful applications in the general population, and not just for people living with severe disabilities. In the near future, BCI systems may therefore become a new mode of human-machine interaction with levels of everyday use that are similar to other current interfaces. Sensors 2012, 12 1259 Acknowledgements This work was partially supported by the regional 2010 Research Project Plan of the Junta de Castilla y León, (Spain), under project VA034A10-2. References 1. Khalid, M.B.; Rao, N.I.; Rizwan-i-Haque, I.; Munir, S.; Tahir, F. Towards a Brain Computer Interface Using Wavelet Transform with Averaged and Time Segmented Adapted Wavelets. In Proceedings of the 2nd International Conference on Computer, Control and Communication (IC4’09) Karachi, Sindh, Pakistan, February 2009; pp. 1–4. 2. Wolpaw, J.R.; Birbaumer, N.; McFarland, D.J.; Pfurtscheller, G.; Vaughan, T.M. Brain-computer interfaces for communication and control. Clin. Neurophysiol. 2002, 113, 767–791. 3. Wolpaw, J.R. Brain–computer interfaces as new brain output pathways. J. Physiol. 2007, 579, 613–619. 4. Konrad, P.; Shanks, T. Implantable brain computer interface: Challenges to neurotechnology translation. Neurobiol. Dis. 2010, 38, 369–375. 5. Emotiv—Brain Computer Interface Technology. Available online: http://www.emotiv.com (accessed on 12 July 2011). 6. NeuroSky—Brainwave Sensors for Everybody. Available online: http://www.neurosky.com (accessed on 12 July 2011). 7. Blankertz, B.; Tangermann, M.; Vidaurre, C.; Fazli, S.; Sannelli, C.; Haufe, S.; Maeder, C.; Ramsey, L.; Sturm, I.; Curio, G.; Mueller, K.R. The Berlin Brain-Computer Interface: Non-Medical Uses of BCI Technology. Front. Neurosci. 2010, doi: 10.3389/fnins.2010.00198. 8. Wolpaw, J.R.; Birbaumer, N.; Heetderks, W.J.; McFarland, D.J.; Peckham, P.H.; Schalk, G.; Donchin, E.; Quatrano, L.A.; Robinson, C.J.; Vaughan, T.M. Brain-computer interface technology: A review of the first international meeting. IEEE Trans. Rehabil. Eng. 2000, 8, 164–173. 9. Mason, S.G.; Birch, G.E. A general framework for brain-computer interface design. IEEE Trans. Neural Syst. Rehabil. Eng. 2003, 11, 70–85. 10. Baillet, S.; Mosher, J.C.; Leahy, R.M. Electromagnetic brain mapping. IEEE Signal Process. Mag. 2001, 18, 14–30. 11. Laureys, S.; Boly, M.; Tononi, G.; Functional Neuroimaging. In The Neurology of Consciousness; Steven, L., Giulio, T., Eds.; Academic Press: New York, NY, USA, 2009; pp. 31–42. 12. Sellers, E.W.; Vaughan, T.M.; Wolpaw, J.R. A brain-computer interface for long-term independent home use. Amyotroph. Lateral Scler. 2010, 11, 449–455. 13. Muller-Putz, G.R.; Pfurtscheller, G. Control of an Electrical Prosthesis With an SSVEP-Based BCI. IEEE Trans. Biomed. Eng. 2008, 55, 361–364. 14. Cincotti, F.; Mattia, D.; Aloise, F.; Bufalari, S.; Schalk, G.; Oriolo, G.; Cherubini, A.; Marciani, M.G.; Babiloni, F. Non-invasive brain-computer interface system: Towards its application as assistive technology. Brain Res. Bull. 2008, 75, 796–803. Sensors 2012, 12 1260 15. Lebedev, M.A.; Nicolelis, M.A.L. Brain-machine interfaces: Past, present and future. Trends Neurosci. 2006, 29, 536–546. 16. Kennedy, P.R.; Bakay, R.A.E.; Moore, M.M.; Adams, K.; Goldwaithe, J. Direct control of a computer from the human central nervous system. IEEE Trans. Rehabil. Eng. 2000, 8, 198–202. 17. Bossetti, C.A.; Carmena, J.M.; Nicolelis, M.A.L.; Wolf, P.D. Transmission latencies in a telemetry-linked brain-machine interface. IEEE Trans. Biomed. Eng. 2004, 51, 919–924. 18. Teplan, M. Fundamentals of EEG measurement. Meas. Sci. Rev. 2002, 2, 1–11. 19. Sinclair, C.M.; Gasper, M.C.; Blum, A.S.; Basic Electronics in Clinical Neurophysiology. In The Clinical Neurophysiology Primer; Blum, A.S., Rutkove, S.B., Eds.; Humana Press Inc.: Totowa, NJ, USA, 2007; pp. 3–18. 20. Usakli, A.B. Improvement of EEG signal acquisition: An electrical aspect for state of the art of front end. Comput. Intell. Neurosci. 2010, 2010, 630649. 21. Fonseca, C.; Cunha, J.P.S.; Martins, R.E.; Ferreira, V.M.; de Sa, J.P.M.; Barbosa, M.A.; da Silva, A.M. A Novel Dry Active Electrode for EEG Recording. IEEE Trans. Biomed. Eng. 2007, 54, 162–165. 22. Taheri, B.A.; Knight, R.T.; Smith, R.L. A dry electrode for EEG recording. Electroencephalogr. Clin. Neurophysiol. 1994, 90, 376–383. 23. Gargiulo, G.; Calvo, R.A.; Bifulco, P.; Cesarelli, M.; Jin, C.; Mohamed, A.; van Schaik, A. A new EEG recording system for passive dry electrodes. Clin. Neurophysiol. 2010, 121, 686–693. 24. Leach, W.M., Jr. Fundamentals of low-noise analog circuit design. Proc. IEEE 1994, 82, 1515–1538. 25. Kübler, A.; Kotchoubey, B.; Kaiser, J.; Wolpaw, J.R.; Birbaumer, N. Brain-Computer Communication: Unlocking the Locked in; American Psychological Association: Washington, DC, USA, 2001. 26. Anand, B.K.; Chhina, G.S.; Singh, B. Some aspects of electroencephalographic studies in Yogis. Electroencephalogr. Clin. Neurophysiol. 1961, 13, 452–456. 27. Aftanas, L.I.; Golocheikine, S.A. Human anterior and frontal midline theta and lower alpha reflect emotionally positive state and internalized attention: High-resolution EEG investigation of meditation. Neurosci. Lett. 2001, 310, 57–60. 28. Fernández, T.; Harmony, T.; Rodríguez, M.; Bernal, J.; Silva, J.; Reyes, A.; Marosi, E. EEG activation patterns during the performance of tasks involving different components of mental calculation. Electroencephalogr. Clin. Neurophysiol. 1995, 94, 175–182. 29. Caplan, J.B.; Madsen, J.R.; Raghavachari, S.; Kahana, M.J. Distinct Patterns of Brain Oscillations Underlie Two Basic Parameters of Human Maze Learning. J. Neurophysiol. 2001, 86, 368–380. 30. Klimesch, W.; Doppelmayr, M.; Yonelinas, A.; Kroll, N.E.A.; Lazzara, M.; Röhm, D.; Gruber, W. Theta synchronization during episodic retrieval: Neural correlates of conscious awareness. Cogn. Brain. Res. 2001, 12, 33–38. 31. Jaime A.P. The functional significance of mu rhythms: Translating “seeing” and “hearing” into “doing”. Brain Res. Rev. 2005, 50, 57–68. 32. Black, A.H. The Operant Conditioning of Central Nervous System Electrical Activity. In Psychology of Learning and Motivation, Gordon, H.B., Ed.; Academic Press: New York, NY, USA, 1972; Volume 6, pp. 47–95. Sensors 2012, 12 1261 33. Klimesch, W. EEG-alpha rhythms and memory processes. Int. J. Psychophysiol. 1997, 26, 319–340. 34. Venables, L.; Fairclough, S. The influence of performance feedback on goal-setting and mental effort regulation. Motiv. Emotion 2009, 33, 63–74. 35. Pfurtscheller, G.; Brunner, C.; Schlögl, A.; Lopes da Silva, F.H. Mu rhythm (de)synchronization and EEG single-trial classification of different motor imagery tasks. Neuroimage 2006, 31, 153–159. 36. Pfurtscheller, G.; Neuper, C. Motor imagery and direct brain-computer communication. Proc. IEEE 2001, 89, 1123–1134. 37. Lee, K.-H.; Williams, L.M.; Breakspear, M.; Gordon, E. Synchronous Gamma activity: A review and contribution to an integrative neuroscience model of schizophrenia. Brain Res. Rev. 2003, 41, 57–78. 38. Brown, P.; Salenius, S.; Rothwell, J.C.; Hari, R. Cortical correlate of the piper rhythm in humans. J. Neurophysiol. 1998, 80, 2911–2917. 39. Mima, T.; Simpkins, N.; Oluwatimilehin, T.; Hallett, M. Force level modulates human cortical oscillatory activities. Neurosci. Lett. 1999, 275, 77–80. 40. Lutzenberger, W.; Pulvermüller, F.; Elbert, T.; Birbaumer, N. Visual stimulation alters local 40-Hz responses in humans: an EEG-study. Neurosci. Lett. 1995, 183, 39–42. 41. Müller, M.M.; Keil, A.; Gruber, T.; Elbert, T. Processing of affective pictures modulates right-hemispheric gamma band EEG activity. Clin. Neurophysiol. 1999, 110, 1913–1920. 42. Müller, M.M.; Bosch, J.; Elbert, T.; Kreiter, A.; Sosa, M.V.; Sosa, P.V.; Rockstroh, B. Visually induced gamma-band responses in human electroencephalographic activity—a link to animal studies. Exp. Brain Res. 1996, 112, 96–102. 43. Zhang, L.; He, W.; He, C.; Wang, P. Improving Mental task classification by adding high frequency band information. J. Med. Syst. 2010, 34, 51–60. 44. Darvas, F.; Scherer, R.; Ojemann, J.G.; Rao, R.P.; Miller, K.J.; Sorensen, L.B. High gamma mapping using EEG. Neuroimage 2010, 49, 930–938. 45. Miller, K.J.; Leuthardt, E.C.; Schalk, G.; Rao, R.P.N.; Anderson, N.R.; Moran, D.W.; Miller, J.W.; Ojemann, J.G. Spectral changes in cortical surface potentials during motor movement. J. Neurosci. 2007, 27, 2424–2432. 46. Jasper, H.H. The ten-twenty electrode system of the International Federation. Electroencephalogr. Clin. Neurophysiol. 1958, 10, 371–375. 47. Waldert, S.; Pistohl, T.; Braun, C.; Ball, T.; Aertsen, A.; Mehring, C. A review on directional information in neural signals for brain-machine interfaces. J. Phisiol. Paris 2009, 103, 244–254. 48. Salmelin, R.; Hámáaláinen, M.; Kajola, M.; Hari, R. Functional segregation of movement-related rhythmic activity in the human brain. Neuroimage 1995, 2, 237–243. 49. Babiloni, C.; Pizzella, V.; Gratta, C.D.; Ferretti, A.; Romani, G.L.; Fundamentals of Electroencefalography, Magnetoencefalography, and Functional Magnetic Resonance Imaging. In Brain Machine Interfaces for Space Applications; Luca, R., Dario, I., Leopold, S., Eds.; Academic Press: New York, NY, USA, 2009; Volume 86, pp. 67–80. 50. Mellinger, J.; Schalk, G.; Braun, C.; Preissl, H.; Rosenstiel, W.; Birbaumer, N.; Kübler, A. An MEG-based brain-computer interface (BCI). NeuroImage 2007, 36, 581–593. Sensors 2012, 12 1262 51. Wang, W.; Sudre, G.P.; Xu, Y.; Kass, R.E.; Collinger, J.L.; Degenhart, A.D.; Bagic, A.I.; Weber, D.J. Decoding and cortical source localization for intended movement direction with MEG. J. Neurophysiol. 2010, 104, 2451–2461. 52. Lal, T.N.; Schröder, M.; Hill, N.J.; Preissl, H.; Hinterberger, T.; Mellinger, J.; Bogdan, M.; Rosenstiel, W.; Hofmann, T.; Birbaumer, N.; Schölkopf, B. A Brain Computer Interface with Online Feedback Based on Magnetoencephalography. In Proceedings of the 22nd International Conference on Machine Learning (ICML’ 05), Bonn, Germany, 7–11 August 2005; pp. 465–472. 53. Kauhanen, L.; Nykopp, T.; Lehtonen, J.; Jylanki, P.; Heikkonen, J.; Rantanen, P.; Alaranta, H.; Sams, M. EEG and MEG brain-computer interface for tetraplegic patients. IEEE Trans. Neural Syst. Rehabil. Eng. 2006, 14, 190–193. 54. Georgopoulos, A.; Langheim, F.; Leuthold, A.; Merkle, A. Magnetoencephalographic signals predict movement trajectory in space. Exp. Brain Res. 2005, 167, 132–135. 55. Mellinger, J.; Schalk, G.; Braun, C.; Preissl, H.; Rosenstiel, W.; Birbaumer, N.; Kübler, A. An MEG-based brain-computer interface (BCI). Neuroimage 2007, 36, 581–593. 56. Jinyin, Z.; Sudre, G.; Xin, L.; Wei, W.; Weber, D.J.; Bagic, A. Clustering linear discriminant analysis for MEG-Based brain computer interfaces. IEEE Trans. Neural Syst. Rehabil. Eng. 2011, 19, 221–231. 57. Sabra, N.I.; Wahed, M.A. The Use of MEG-Based Brain Computer Interface for Classification of Wrist Movements in Four Different Directions. In Proceedings of the 28th National Radio Science Conference (NRSC’ 11), Cairo, Egypt, April 2011; pp. 1–7. 58. Ball, T.; Kern, M.; Mutschler, I.; Aertsen, A.; Schulze-Bonhage, A. Signal quality of simultaneously recorded invasive and non-invasive EEG. Neuroimage 2009, 46, 708–716. 59. Loeb, G.E.; Walker, A.E.; Uematsu, S.; Konigsmark, B.W. Histological reaction to various conductive and dielectric films chronically implanted in the subdural space. J. Biomed. Mater. Res. 1977, 11, 195–210. 60. Bullara, L.A.; Agnew, W.F.; Yuen, T.G.H.; Jacques, S.; Pudenz, R.H. Evaluation of electrode array material for neural prostheses. Neurosurgery 1979, 5, 681–686. 61. Yuen, T.G.H.; Agnew, W.F.; Bullara, L.A. Tissue response to potential neuroprosthetic materials implanted subdurally. Biomaterials 1987, 8, 138–141. 62. Margalit, E.; Weiland, J.D.; Clatterbuck, R.E.; Fujii, G.Y.; Maia, M.; Tameesh, M.; Torres, G.; D'Anna, S.A.; Desai, S.; Piyathaisere, D.V.; et al. Visual and electrical evoked response recorded from subdural electrodes implanted above the visual cortex in normal dogs under two methods of anesthesia. J. Neurosci. Methods 2003, 123, 129–137. 63. Chao, Z.C.; Nagasaka, Y.; Fujii, N. Long-term asynchronous decoding of arm motion using electrocorticographic signals in monkeys. Front. Neuroeng. 2010, 3, doi: 10.3389/fneng.2010.00003. 64. Matsuo, T.; Kawasaki, K.; Osada, T.; Sawahata, H.; Suzuki, T.; Shibata, M.; Miyakawa, N.; Nakahara, K.; Iijima, A.; Sato, N.; et al. Intrasulcal electrocorticography in macaque monkeys with minimally invasive neurosurgical protocols. Front. Syst. Neurosci. 2011, doi: 10.3389/ fnsys.2011.00034. 65. Crone, N.E.; Miglioretti, D.L.; Gordon, B.; Sieracki, J.M.; Wilson, M.T.; Uematsu, S.; Lesser, R.P. Functional mapping of human sensorimotor cortex with electrocorticographic spectral analysis. I. Alpha and beta event-related desynchronization. Brain 1998, 121, 2271–2299. Sensors 2012, 12 1263 66. Crone, N.E.; Miglioretti, D.L.; Gordon, B.; Lesser, R.P. Functional mapping of human sensorimotor cortex with electrocorticographic spectral analysis. II. Event-related synchronization in the gamma band. Brain 1998, 121, 2301–2315. 67. Miller, K.J.; denNijs, M.; Shenoy, P.; Miller, J.W.; Rao, R.P.N.; Ojemann, J.G. Real-time functional brain mapping using electrocorticography. NeuroImage 2007, 37, 504–507. 68. Levine, S.P.; Huggins, J.E.; BeMent, S.L.; Kushwaha, R.K.; Schuh, L.A.; Passaro, E.A.; Rohde, M.M.; Ross, D.A. Identification of electrocorticogram patterns as the basis for a direct brain interface. J. Clin. Neurophysiol. 1999, 16, 439–447. 69. Leuthardt, E.C.; Schalk, G.; Wolpaw, J.R.; Ojemann, J.G; Moran, D.W. A brain-computer interface using electrocorticographic signals in humans. J. Neural Eng. 2004, doi: 10.1088/1741- 2560/1/2/001. 70. Schalk, G.; Schalk, G.; Kubánek, J.; Miller, K.J.; Anderson, N.R.; Leuthardt, E.C.; Ojemann, J.G.; Limbrick, D.; Moran, D.; Gerhardt, L.A.; et al. Decoding two-dimensional movement trajectories using electrocorticographic signals in humans. J. Neural Eng. 2007, doi: 10.1088/1741-2560/4/3/012. 71. Polikov, V.S.; Tresco, P.A.; Reichert, W.M. Response of brain tissue to chronically implanted neural electrodes. J. Neurosci. Methods 2005, 148, 1–18. 72. Maynard, E.M.; Nordhausen, C.T.; Normann, R.A. The Utah Intracortical Electrode Array: A recording structure for potential brain-computer interfaces. Electroencephalogr. Clin. Neurophysiol. 1997, 102, 228–239. 73. Lauer, R.T.; Peckham, P.H.; Kilgore, K.L.; Heetderks, W.J. Applications of cortical signals to neuroprosthetic control: A critical review. IEEE Trans. Rehabil. Eng. 2000, 8, 205–208. 74. Georgopoulos, A.; Schwartz, A.; Kettner, R. Neuronal population coding of movement direction. Science 1986, 233, 1416–1419. 75. Schwartz, A.B. Motor cortical activity during drawing movements: population representation during sinusoid tracing. J. Neurophysiol. 1993, 70, 28–36. 76. Wessberg, J.; Stambaugh, C.R.; Kralik, J.D.; Beck, P.D.; Laubach, M.; Chapin, J.K.; Kim, J.; Biggs, S.J.; Srinivasan, M.A.; Nicolelis, M.A.L. Real-time prediction of hand trajectory by ensembles of cortical neurons in primates. Nature 2000, 408, 361–365. 77. Taylor, D.M.; Tillery, S.I.H.; Schwartz, A.B. Direct cortical control of 3D neuroprosthetic devices. Science 2002, 296, 1829–1832. 78. Jackson, A.; Mavoori, J.; Fetz, E.E. Correlations between the same motor cortex cells and arm muscles during a trained task, free behavior, and natural sleep in the macaque monkey. J. Neurophysiol. 2007, 97, 360–374. 79. Velliste, M.; Perel, S.; Spalding, M.C.; Whitford, A.S.; Schwartz, A.B. Cortical control of a prosthetic arm for self-feeding. Nature 2008, 453, 1098–1101. 80. Vargas-Irwin, C.E.; Shakhnarovich, G.; Yadollahpour, P.; Mislow, J.M.K.; Black, M.J.; Donoghue, J.P. Decoding complete reach and grasp actions from local primary motor cortex populations. J. Neurosci. 2010, 30, 9659–9669. 81. Carpaneto, J.; Umiltà, M.A.; Fogassi, L.; Murata, A.; Gallese, V.; Micera, S.; Raos, V. Decoding the activity of grasping neurons recorded from the ventral premotor area F5 of the macaque monkey. Neuroscience 2011, 188, 80–94. Sensors 2012, 12 1264 82. Kennedy, P.R.; Kirby, M.T.; Moore, M.M.; King, B.; Mallory, A. Computer control using human intracortical local field potentials. IEEE Trans. Neural Syst. Rehabil. Eng. 2004, 12, 339–344. 83. deCharms, R.C.; Christoff, K.; Glover, G.H.; Pauly, J.M.; Whitfield, S.; Gabrieli, J.D.E. Learned regulation of spatially localized brain activation using real-time fMRI. Neuroimage 2004, 21, 436–443. 84. Weiskopf, N.; Mathiak, K.; Bock, S.W.; Scharnowski, F.; Veit, R.; Grodd, W.; Goebel, R.; Birbaumer, N. Principles of a brain-computer interface (BCI) based on real-time functional magnetic resonance imaging (fMRI). IEEE Trans. Biomed. Eng. 2004, 51, 966–970. 85. Lee, J.; Ryu, J.; Jolesz, F.A.; Cho, Z.H.; Yoo, S.S. Brain-machine interface via real-time fMRI: Preliminary study on thought-controlled robotic arm. Neurosci. Lett. 2009, 450, 1–6. 86. Logothetis, N.K.; Pauls, J.; Augath, M.; Trinath, T.; Oeltermann, A. Neurophysiological investigation of the basis of the fMRI signal. Nature 2001, 412, 150–157. 87. Weiskopf, N.; Sitaram, R.; Josephs, O.; Veit, R.; Scharnowski, F.; Goebel, R.; Birbaumer, N.; Deichmann, R.; Mathiak, K. Real-time functional magnetic resonance imaging: Methods and applications. Magn. Reson. Imaging 2007, 25, 989–1003. 88. Christopher deCharms, R. Applications of real-time fMRI. Nat. Rev. Neurosci. 2008, 9, 720–729. 89. Moench, T.; Hollmann, M.; Grzeschik, R.; Mueller, C.; Luetzkendorf, R.; Baecke, S.; Luchtmann, M.; Wagegg, D.; Bernarding, J. Real-Time Classification of Activated Brain Areas for fMRI-Based Human-Brain-Interfaces; Society of Photo-Optical Instrumentation Engineers: Belligham, WA, USA, 2008. 90. Ward, B.D.; Mazaheri, Y. Information transfer rate in fMRI experiments measured using mutual information theory. J. Neurosci. Methods 2008, 167, 22–30. 91. Coyle, S.M.; Ward, T.E.; Markham, C.M. Brain-computer interface using a simplified functional near-infrared spectroscopy system. J. Neural Eng. 2007, doi: 10.1088/1741-2560/4/3/007. 92. Taga, G.; Homae, F.; Watanabe, H. Effects of source-detector distance of near infrared spectroscopy on the measurement of the cortical hemodynamic response in infants. NeuroImage 2007, 38, 452–460. 93. Kennan, R.P.; Horovitz, S.G.; Maki, A.; Yamashita, Y.; Koizumi, H.; Gore, J.C. Simultaneous recording of event-related auditory oddball response using transcranial near infrared optical topography and surface EEG. Neuroimage 2002, 16, 587–592. 94. Izzetoglu, M.; Devaraj, A.; Bunce, S.; Onaral, B. Motion artifact cancellation in NIR spectroscopy using Wiener filtering. IEEE Trans. Biomed. Eng. 2005, 52, 934–938. 95. Strangman, G.; Boas, D.A.; Sutton, J.P. Non-invasive neuroimaging using near-infrared light. Biol. Psychiatry 2002, 52, 679–693. 96. Cui, X.; Bray, S.; Reiss, A.L. Functional near infrared spectroscopy (NIRS) signal improvement based on negative correlation between oxygenated and deoxygenated hemoglobin dynamics. Neuroimage 2010, 49, 3039–3046. 97. Shirley, C.; Ward, T.; Markham, C.; McDarby, G. On the suitability of near-infrared (NIR) systems for next-generation brain-computer interfaces. Physiol. Meas. 2004, 25, 815. 98. Power, S.D.; Kushki, A.; Chau, T. Towards a system-paced near-infrared spectroscopy brain-computer interface: Differentiating prefrontal activity due to mental arithmetic and mental singing from the no-control state. J. Neural Eng. 2011, 8, 066004. Sensors 2012, 12 1265 99. Villringer, A.; Planck, J.; Hock, C.; Schleinkofer, L.; Dirnagl, U. Near infrared spectroscopy (NIRS): A new tool to study hemodynamic changes during activation of brain function in human adults. Neurosci. Lett. 1993, 154, 101–104. 100. Sitaram, R.; Zhang, H.; Guan, C.; Thulasidas, M.; Hoshi, Y.; Ishikawa, A.; Shimizu, K.; Birbaumer, N. Temporal classification of multichannel near-infrared spectroscopy signals of motor imagery for developing a brain–computer interface. Neuroimage 2007, 34, 1416–1427. 101. Regan, D. Human Brain Electrophysiology: Evoked Potentials and Evoked Magnetic Fields in Science and Medicine; Elsevier: New York, NY, USA, 1989. 102. Yijun, W.; Ruiping, W.; Xiaorong, G.; Bo, H.; Shangkai, G. A practical VEP-based brain-computer interface. IEEE Trans. Neural Syst. Rehabil. Eng. 2006, 14, 234–240. 103. Jinghai, Y.; Derong, J.; Jianfeng, H. Design and Application of Brain-Computer Interface Web Browser Based on VEP. In Proceedings of the International Conference on Future BioMedical Information Engineering (FBIE’09), Sanya, China, 13–14 December 2009; pp. 77–80. 104. Xiaorong, G.; Dingfeng, X.; Ming, C.; Shangkai, G. A BCI-based environmental controller for the motion-disabled. IEEE Trans. Neural Syst. Rehabil. Eng. 2003, 11, 137–140. 105. Odom, J.V.; Bach, M.; Barber, C.; Brigell, M.; Marmor, M.F.; Tormene, A.P.; Holder, G.E. Visual evoked potentials standard (2004). Doc. Ophthalmol. 2004, 108, 115–123. 106. Zhu, D.; Bieger, J.; Garcia Molina, G.; Aarts, R.M. A survey of stimulation methods used in SSVEP-Based BCIs. Comput. Intell. Neurosci. 2010, doi: 10.1155/2010/702357. 107. Galloway, N. Human brain electrophysiology: Evoked potentials and evoked magnetic fields in science and medicine. Br. J. Ophthalmol. 1990, 74, 255. 108. Perlstein, W.M.; Cole, M.A.; Larson, M.; Kelly, K.; Seignourel, P.; Keil, A. Steady-state visual evoked potentials reveal frontally-mediated working memory activity in humans. Neurosci. Lett. 2003, 342, 191–195. 109. Gray, M.; Kemp, A.H.; Silberstein, R.B.; Nathan, P.J. Cortical neurophysiology of anticipatory anxiety: An investigation utilizing steady state probe topography (SSPT). Neuroimage 2003, 20, 975–986. 110. Guangyu, B.; Xiaorong, G.; Yijun, W.; Bo, H.; Shangkai, G. VEP-based brain-computer interfaces: Time, frequency, and code modulations [Research Frontier]. IEEE Comput. Intell. Mag. 2009, 4, 22–26. 111. Lee, P.; Hsieh, J.; Wu, C.; Shyu, K.; Wu, Y. Brain computer interface using flash onset and offset visual evoked potentials. Clin. Neurophysiol. 2008, 119, 605–616. 112. Allison, B.Z.; McFarland, D.J.; Schalk, G.; Zheng, S.D.; Jackson, M.M.; Wolpaw, J.R. Towards an independent brain-computer interface using steady state visual evoked potentials. Clin. Neurophysiol. 2008, 119, 399–408. 113. Dan, Z.; Alexander, M.; Xiaorong, G.; Bo, H.; Andreas, K.E.; Shangkai, G. An independent brain-computer interface using covert non-spatial visual selective attention. J. Neural Eng. 2010, 7, 016010. 114. Yijun, W.; Xiaorong, G.; Bo, H.; Chuan, J.; Shangkai, G. Brain-computer interfaces based on visual evoked potentials. IEEE Eng. Med. Biol. Mag. 2008, 27, 64–71. 115. Wu, Z.; Lai, Y.; Xia, Y.; Wu, D.; Yao, D. Stimulator selection in SSVEP-based BCI. Med. Eng. Phys. 2008, 30, 1079–1088. Sensors 2012, 12 1266 116. Birbaumer, N.; Elbert, T.; Canavan, A.G.; Rockstroh, B. Slow potentials of the cerebral cortex and behavior. Physiol. Rev. 1990, 70, 1–41. 117. Hinterberger, T.; Schmidt, S.; Neumann, N.; Mellinger, J.; Blankertz, B.; Curio, G.; Birbaumer, N. Brain-computer communication and slow cortical potentials. IEEE Trans. Biomed. Eng. 2004, 51, 1011–1018. 118. Kaiser, J. Self-initiation of EEG-based communication in paralyzed patients. Clin. Neurophysiol. 2001, 112, 551–554. 119. Neumann, N.; Birbaumer, N. Predictors of successful self control during brain-computer communication. J. Neurol. Neurosurg. Psychiatry 2003, 74, 1117–1121. 120. Kübler, A.; Kotchoubey, B.; Hinterberger, T.; Ghanayim, N.; Perelmouter, J.; Schauer, M.; Fritsch, C.; Taub, E.; Birbaumer, N. The thought translation device: A neurophysiological approach to communication in total motor paralysis. Exp. Brain Res. 1999, 124, 223–232. 121. Birbaumer, N.; Hinterberger, T.; Kubler, A.; Neumann, N. The thought-translation device (TTD): Neurobehavioral mechanisms and clinical outcome. IEEE Trans. Neural Syst. Rehabil. Eng. 2003, 11, 120–123. 122. Iversen, I.H.; Ghanayim, N.; Kübler, A.; Neumann, N.; Birbaumer, N.; Kaiser, J. A brain-computer interface tool to assess cognitive functions in completely paralyzed patients with amyotrophic lateral sclerosis. Clin. Neurophysiol. 2008, 119, 2214–2223. 123. Farwell, L.A.; Donchin, E. Talking off the top of your head: Toward a mental prosthesis utilizing event-related brain potentials. Electroencephalogr. Clin. Neurophysiol. 1988, 70, 510–523. 124. Donchin, E.; Smith, D.B.D. The contingent negative variation and the late positive wave of the average evoked potential. Electroencephalogr. Clin. Neurophysiol. 1970, 29, 201–203. 125. Polich, J.; Ellerson, P.C.; Cohen, J. P300, stimulus intensity, modality, and probability. Int. J. Psychophysiol. 1996, 23, 55–62. 126. Ravden, D.; Polich, J. On P300 measurement stability: Habituation, intra-trial block variation, and ultradian rhythms. Biol. Psychol. 1999, 51, 59–76. 127. Tanaka, K.; Matsunaga, K.; Wang, H.O. Electroencephalogram-based control of an electric wheelchair. IEEE Trans. Robot. 2005, 21, 762–766. 128. Mugler, E.M.; Ruf, C.A.; Halder, S.; Bensch, M.; Kubler, A. Design and implementation of a P300-based brain-computer interface for controlling an internet browser. IEEE Trans. Neural Syst. Rehabil. Eng. 2010, 18, 599–609. 129. Furdea, A.; Halder, S.; Krusienski, D.J.; Bross, D.; Nijboer, F.; Birbaumer, N.; Kübler, A. An auditory oddball (P300) spelling system for brain-computer interfaces. Psychophysiology 2009, 46, 617–625. 130. Rivet, B.; Souloumiac, A.; Attina, V.; Gibert, G. xDAWN algorithm to enhance evoked potentials: Application to brain-computer interface. IEEE Trans. Biomed. Eng. 2009, 56, 2035–2043. 131. Rakotomamonjy, A.; Guigue, V. BCI Competition III: Dataset II-Ensemble of SVMs for BCI P300 Speller. IEEE Trans. Biomed. Eng. 2008, 55, 1147–1154. 132. Salvaris, M.; Sepulveda, F. Visual modifications on the P300 speller BCI paradigm. J. Neural Eng. 2009, 6, 046011. Sensors 2012, 12 1267 133. Takano, K.; Komatsu, T.; Hata, N.; Nakajima, Y.; Kansaku, K. Visual stimuli for the P300 brain-computer interface: A comparison of white/gray and green/blue flicker matrices. Clin. Neurophysiol. 2009, 120, 1562–1566. 134. Ikegami, S.; Takano, K.; Saeki, N.; Kansaku, K. Operation of a P300-based brain-computer interface by individuals with cervical spinal cord injury. Clin. Neurophysiol. 2011, 122, 991–996. 135. Hill, J.; Farquhar, J.; Martens, S.; Bießmann, F.; Schölkopf, B. Effects of stimulus type and of error-correcting code design on BCI speller performance. In Proceedings of the Twenty-Second Annual Conference on Neural Information Processing Systems (NIPS 2008), Vancouver, Canada, 8–11 December 2008. 136. Martens, S.M.M.; Hill, N.J.; Farquhar, J.; Schölkopf, B. Overlap and refractory effects in a brain-computer interface speller based on the visual P300 event-related potential. J. Neural Eng. 2009, 6, 026003. 137. Townsend, G.; LaPallo, B.K.; Boulay, C.B.; Krusienski, D.J.; Frye, G.E.; Hauser, C.K.; Schwartz, N.E.; Vaughan, T.M.; Wolpaw, J.R.; Sellers, E.W. A novel P300-based brain-computer interface stimulus presentation paradigm: Moving beyond rows and columns. Clin. Neurophysiol. 2010, 121, 1109–1120. 138. Brunner, P.; Joshi, S.; Briskin, S.; Wolpaw, J.R.; Bischof, H.; Schalk, G. Does the ‘P300’ speller depend on eye gaze? J. Neural Eng. 2010, 7, 056013. 139. Pfurtscheller, G.; Lopes da Silva, F.H. Event-related EEG/MEG synchronization and desynchronization: Basic principles. Clin. Neurophysiol. 1999, 110, 1842–1857. 140. Jeannerod, M. Mental imagery in the motor context. Neuropsychologia 1995, 33, 1419–1432. 141. Pfurtscheller, G.; Neuper, C.; Flotzinger, D.; Pregenzer, M. EEG-based discrimination between imagination of right and left hand movement. Electroencephalogr. Clin. Neurophysiol. 1997, 103, 642–651. 142. Blankertz, B.; Sannelli, C.; Halder, S.; Hammer, E.M.; Kübler, A.; Müller, K.R.; Curio, G.; Dickhaus, T. Neurophysiological predictor of SMR-based BCI performance. Neuroimage 2010, 51, 1303–1309. 143. Neuper, C.; Scherer, R.; Reiner, M.; Pfurtscheller, G. Imagery of motor actions: Differential effects of kinesthetic and visual-motor mode of imagery in single-trial EEG. Cogn. Brain. Res. 2005, 25, 668–677. 144. Nijboer, F.; Furdea, A.; Gunst, I.; Mellinger, J.; McFarland, D.J.; Birbaumer, N.; Kübler, A. An auditory brain-computer interface (BCI). J. Neurosci. Methods 2008, 167, 43–50. 145. Hwang, H.; Kwon, K.; Im, C. Neurofeedback-based motor imagery training for brain-computer interface (BCI). J. Neurosci. Methods 2009, 179, 150–156. 146. Wolpaw, J.R.; McFarland, D.J.; Vaughan, T.M. Brain-computer interface research at the Wadsworth Center. IEEE Trans. Rehabil. Eng. 2000, 8, 222–226. 147. Blankertz, B.; Losch, F.; Krauledat, M.; Dornhege, G.; Curio, G.; Muller, K.R. The berlin brain-computer interface: Accurate performance from first-session in BCI-Naïve subjects. IEEE Trans. Biomed. Eng. 2008, 55, 2452–2462. 148. Pfurtscheller, G.; Neuper, C.; Muller, G.R.; Obermaier, B.; Krausz, G.; Schlogl, A.; Scherer, R.; Graimann, B.; Keinrath, C.; Skliris, D.; et al. Graz-BCI: State of the art and clinical applications. IEEE Trans. Neural Sys. Rehabil. Eng. 2003, 11, 1–4. Sensors 2012, 12 1268 149. Bai, O.; Rathi, V.; Lin, P.; Huang, D.; Battapady, H.; Fei, D.; Schneider, L.; Houdayer, E.; Chen, X.; Hallett, M. Prediction of human voluntary movement before it occurs. Clin. Neurophysiol. 2011, 122, 364–372. 150. Kleber, B.; Birbaumer, N. Direct brain communication: neuroelectric and metabolic approaches at Tübingen. Cogn. Process. 2005, 6, 65–74. 151. Krusienski, D.J.; Schalk, G.; McFarland, D.J.; Wolpaw, J.R. A mu-rhythm matched filter for continuous control of a brain-computer interface. IEEE Trans. Biomed. Eng. 2007, 54, 273–280. 152. Tsui, C.; Gan, J.; Asynchronous BCI Control of a Robot Simulator with Supervised Online Training. In Intelligent Data Engineering and Automated Learning—IDEAL 2007; Yin, H., Tino, P., Corchado, E., Byrne, W., Yao, X., Eds.; Springer: Berlin, Germany, 2007; Volume 4881, pp. 125–134. 153. Ghafar, R.; Hussain, A.; Samad, S.A.; Tahir, N.M. Comparison of FFT and AR Techniques for Scalp EEG Analysis. In Proceedings of the 4th Kuala Lumpur International Conference on Biomedical Engineering 2008, Kuala Lumpur, Malaysia, June 2008; Abu Osman, N.A., Ibrahim, F., Wan Abas, W.A.B., Abdul Rahman, H.S., Ting, H.-N., Magjarevic, R., Eds.; Springer Berlin Heidelberg: Berlin, Germany, 2008; Volume 21, pp. 158–161. 154. von Bünau, P.; Meinecke, F.C.; Király, F.C.; Müller, K.R. Finding stationary subspaces in multivariate time series. Phys. Rev. Lett. 2009, 103, 214101. 155. Lin, C.J.; Hsieh, M.H. Classification of mental task from EEG data using neural networks based on particle swarm optimization. Neurocomputing 2009, 72, 1121–1130. 156. McFarland, D.J.; Anderson, C.W.; Muller, K.R.; Schlogl, A.; Krusienski, D.J. BCI meeting 2005-workshop on BCI signal processing: Feature extraction and translation. IEEE Trans. Neural Syst. Rehabil. Eng. 2006, 14, 135–138. 157. Lins, O.G.; Picton, T.W.; Berg, P.; Scherg, M. Ocular artifacts in recording EEGs and event-related potentials II: Source dipoles and source components. Brain Topogr. 1993, 6, 65–78. 158. Boye, A.T.; Kristiansen, U.Q.; Billinger, M.; do Nascimento, O.F.; Farina, D. Identification of movement-related cortical potentials with optimized spatial filtering and principal component analysis. Biomed. Signal Process. Control 2008, 3, 300–304. 159. Te-Won, L.; Lewicki, M.S.; Girolami, M.; Sejnowski, T.J. Blind source separation of more sources than mixtures using overcomplete representations. IEEE Signal Process. Lett. 1999, 6, 87–90. 160. Kun, L.; Sankar, R.; Arbel, Y.; Donchin, E. Single trial independent component analysis for P300 BCI system. In Proceedings of the 31th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBCS’09), Minneapolis, MN, USA, September 2009; pp. 4035–4038. 161. Castellanos, N.P.; Makarov, V.A. Recovering EEG brain signals: Artifact suppression with wavelet enhanced independent component analysis. J. Neurosci. Methods 2006, 158, 300–312. 162. Bell, A.; Sejnowski, T. An information-maximization approach to blind separation and blind deconvolution. Neural Comput. 1995, 7, 1129–1159. Sensors 2012, 12 1269 163. Lee, T.; Girolami, M.; Sejnowski, T.J. Independent component analysis using an extended infomax algorithm for mixed subgaussian and supergaussian sources. Neural Comput. 1999, 11, 417–441. 164. Gao, J.; Yang, Y.; Lin, P.; Wang, P.; Zheng, C. Automatic removal of eye-movement and blink artifacts from EEG signals. Brain Topogr. 2010, 23, 105–114. 165. Flexer, A.; Bauer, H.; Pripfl, J.; Dorffner, G. Using ICA for removal of ocular artifacts in EEG recorded from blind subjects. Neural Netw. 2005, 18, 998–1005. 166. Jung, T.P.; Makeig, S.; Westerfield, M.; Townsend, J.; Courchesne, E.; Sejnowski, T.J. Removal of eye activity artifacts from visual event-related potentials in normal and clinical subjects. Clin. Neurophysiol. 2000, 111, 1745–1758. 167. Wallstrom, G.L.; Kass, R.E.; Miller, A.; Cohn, J.F.; Fox, N.A. Automatic correction of ocular artifacts in the EEG: a comparison of regression-based and component-based methods. Int. J. Psychophysiol. 2004, 53, 105–119. 168. Chiappa, S.; Barber, D. EEG classification using generative independent component analysis. Neurocomputing 2006, 69, 769–777. 169. Kay, S.M.; Marple, S.L., Jr. Spectrum analysis—A modern perspective. Proc. IEEE 1981, 69, 1380-1419. 170. Krusienski, D.J.; McFarland, D.J.; Wolpaw, J.R. An Evaluation of Autoregressive Spectral Estimation Model Order for Brain-Computer Interface Applications. In Proceedings of the 28th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBS’06), New York, NY, USA, September 2006; pp. 1323–1326. 171. Florian, G.; Pfurtscheller, G. Dynamic spectral analysis of event-related EEG data. Electroencephalogr. Clin. Neurophysiol. 1995, 95, 393–396. 172. Jiang, W.; Guizhi, X.; Lei, W.; Huiyuan, Z. Feature Extraction of Brain-Computer Interface Based on Improved Multivariate Adaptive Autoregressive Models. In Proceedings of the 3rd International Conference on Biomedical Engineering and Informatics (BMEI’10), Yantai, China, October 2010; pp. 895–898. 173. Brunner, C.; Allison, B.Z.; Krusienski, D.J.; Kaiser, V.; Müller-Putz, G.R.; Pfurtscheller, G.; Neuper, C. Improved signal processing approaches in an offline simulation of a hybrid brain-computer interface. J. Neurosci. Methods 2010, 188, 165–173. 174. Ghanbari, A.A.; Kousarrizi, M.R.N.; Teshnehlab, M.; Aliyari, M. Wavelet and Hilbert transform-based Brain Computer Interface. In Proceedings of the International Conference on Advances in Computational Tools for Engineering Applications (ACTEA’09), Beirut, Lebanon, July 2009; pp. 438–442. 175. Samar, V.J.; Bopardikar, A.; Rao, R.; Swartz, K. Wavelet analysis of neuroelectric waveforms: A conceptual tutorial. Brain Lang 1999, 66, 7–60. 176. Demiralp, T.; Yordanova, J.; Kolev, V.; Ademoglu, A.; Devrim, M.; Samar, V.J. Time-frequency analysis of single-sweep event-related potentials by means of fast wavelet transform. Brain Lang 1999, 66, 129–145. 177. Hubbard, B.B. The World According to Wavelets: The Story of a Mathematical Technique in the Making; A.K. Peters: Natick, MA, USA, 1998. Sensors 2012, 12 1270 178. Farina, D.; do Nascimento, O.F.; Lucas, M.F.; Doncarli, C. Optimization of wavelets for classification of movement-related cortical potentials generated by variation of force-related parameters. J. Neurosci. Methods 2007, 162, 357–363. 179. Bostanov, V. BCI competition 2003-data sets Ib and IIb: Feature extraction from event-related brain potentials with the continuous wavelet transform and the t-value scalogram. IEEE Trans. Biomed. Eng. 2004, 51, 1057–1061. 180. Senkowski, D.; Herrmann, C.S. Effects of task difficulty on evoked gamma activity and ERPs in a visual discrimination task. Clin. Neurophysiol. 2002, 113, 1742–1753. 181. Mason, S.G.; Birch, G.E. A brain-controlled switch for asynchronous control applications. IEEE Trans. Biomed. Eng. 2000, 47, 1297–1307. 182. Hinterberger, T.; Kübler, A.; Kaiser, J.; Neumann, N.; Birbaumer, N. A brain-computer interface (BCI) for the locked-in: Comparison of different EEG classifications for the thought translation device. Clin. Neurophysiol. 2003, 114, 416–425. 183. Ramoser, H.; Muller-Gerking, J.; Pfurtscheller, G. Optimal spatial filtering of single trial EEG during imagined hand movement. IEEE Trans. Rehabil. Eng. 2000, 8, 441–446. 184. Grosse-Wentrup, M.; Buss, M. Multiclass common spatial patterns and information theoretic feature extraction. IEEE Trans. Biomed. Eng. 2008, 55, 1991–2000. 185. Mousavi, E.A.; Maller, J.J.; Fitzgerald, P.B.; Lithgow, B.J. Wavelet common spatial pattern in asynchronous offline brain computer interfaces. Biomed. Signal Process. Control 2011, 6, 121–128. 186. Lemm, S.; Blankertz, B.; Curio, G.; Muller, K.R. Spatio-spectral filters for improving the classification of single trial EEG. IEEE Trans. Biomed. Eng. 2005, 52, 1541–1548. 187. Dornhege, G.; Blankertz, B.; Krauledat, M.; Losch, F.; Curio, G.; Muller, K.R. Combined optimization of spatial and temporal filters for improving brain-computer interfacing. IEEE Trans. Biomed. Eng. 2006, 53, 2274–2281. 188. Dal Seno, B.; Matteucci, M.; Mainardi, L. A Genetic Algorithm for Automatic Feature Extraction in P300 Detection. In Proceedings of the IEEE International Joint Conference on Neural Networks (IJCNN’08), Hong Kong, China, June 2008; pp. 3145–3152. 189. Fatourechi, M.; Birch, G.E.; Ward, R.K. Application of a hybrid wavelet feature selection method in the design of a self-paced brain interface system. J. NeuroEng. Rehabil. 2007, 4, 1–13. 190. Ciaccio, E.; Dunn, S.; Akay, M. Biosignal pattern recognition and interpretation systems. 2. Methods for feature extraction and selection. IEEE Eng. Med. Biol. Mag. 1993, 12, 106–113. 191. Dias, N.; Kamrunnahar, M.; Mendes, P.; Schiff, S.; Correia, J. Feature selection on movement imagery discrimination and attention detection. Med. Biol. Eng. Comput. 2010, 48, 331–341. 192. Lakany, H.; Conway, B.A. Understanding intention of movement from electroencephalograms. Expert. Syst. 2007, 24, 295–304. 193. Pudil, P.; Novovičová, J.; Kittler, J. Floating search methods in feature selection. Pattern Recogn. Lett. 1994, 15, 1119–1125. 194. Townsend, G.; Graimann, B.; Pfurtscheller, G. A comparison of common spatial patterns with complex band power features in a four-class BCI experiment. IEEE Trans. Biomed. Eng. 2006, 53, 642–651. Sensors 2012, 12 1271 195. Hasan, B.A.S.; Gan, J.Q. Unsupervised Adaptive GMM for BCI. In Proceedings of the 4th International IEEE/EMBS Conference on Neural Engineering (NER’09), Antalya, Turkey, May 2009; pp. 295–298. 196. Neuper, C.; Scherer, R.; Wriessnegger, S.; Pfurtscheller, G. Motor imagery and action observation: Modulation of sensorimotor brain rhythms during mental control of a brain-computer interface. Clin. Neurophysiol. 2009, 120, 239–247. 197. Fatourechi, M.; Bashashati, A.; Ward, R.K.; Birch, G.E. EMG and EOG artifacts in brain computer interface systems: A survey. Clin. Neurophysiol. 2007, 118, 480–494. 198. Croft, R.J.; Barry, R.J. Removal of ocular artifact from the EEG: A review. Neurophysiol. Clin. 2000, 30, 5–19. 199. Vigário, R.N. Extraction of ocular artefacts from EEG using independent component analysis. Electroencephalogr. Clin. Neurophysiol. 1997, 103, 395–404. 200. del R Millan, J.; Mourino, J.; Franze, M.; Cincotti, F.; Varsta, M.; Heikkonen, J.; Babiloni, F. A local neural classifier for the recognition of EEG patterns associated to mental tasks. IEEE Trans. Neural Netw. 2002, 13, 678–686. 201. Changmok, C.; Micera, S.; Carpaneto, J.; Jung, K. Development and quantitative performance evaluation of a noninvasive EMG computer interface. IEEE Trans. Biomed. Eng. 2009, 56, 188–191. 202. Nojd, N.; Hannula, M.; Hyttinen, J. Electrode Position Optimization for Facial EMG Measurements for Human-Computer Interface. In Proceedings of the 2nd International Conference on Pervasive Computing Technologies for Healthcare (PCTH’11), Dublin, Ireland, February 2008; pp. 319–322. 203. Inhyuk, M.; Myoungjoon, L.; Museong, M. A Novel EMG-Based Human-Computer Interface for Persons with Disability. In Proceedings of the Proceedings of the IEEE International Conference on Mechatronics (ICM’04), Tunis, Tunisia, June 2004; pp. 519–524. 204. Gomez-Gil, J.; San-Jose-Gonzalez, I.; Nicolas-Alonso, L.F.; Alonso-Garcia, S. Steering a tractor by means of an EMG-based human-machine interface. Sensors 2011, 11, 7110–7126. 205. Nan, B.; Okamoto, M.; Tsuji, T. A hybrid motion classification approach for EMG-based human-robot interfaces using bayesian and neural networks. IEEE Trans. Robot. 2009, 25, 502–511. 206. Zhang, X.; Chen, X.; Lantz, V.; Yang, J.-H.; Wang, K.-Q. Exploration on the Feasibility of Building Muscle-Computer Interfaces Using Neck and Shoulder Motions. In Proceedings of the Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC’09), Minneapolis, MN, USA, 3–6 September 2009; pp. 7018–7021. 207. De la Rosa, R.; Alonso, A.; Carrera, A.; Durán, R.; Fernández, P. Man-machine interface system for neuromuscular training and evaluation based on EMG and MMG signals. 2010, 10, 11100–11125. 208. Eckhouse, R.H.; Maulucci, R.A. A multimedia system for augmented sensory assessment and treatment of motor disabilities. Telemat. Inf. 1997, 14, 67–82. 209. Lotte, F.; Congedo, M.; Lécuyer, A.; Lamarche, F.;Arnaldi, B. A review of classification algorithms for EEG-based brain-computer interfaces. J. Neural Eng. 2007, 4, R1. 210. McFarland, D.J.; Wolpaw, J.R. Sensorimotor rhythm-based brain-computer interface (BCI): Feature selection by regression improves performance. IEEE Trans. Neural Syst. Rehabil. Eng. 2005, 13, 372–379. Sensors 2012, 12 1272 211. Blankertz, B.; Muller, K.R.; Krusienski, D.J.; Schalk, G.; Wolpaw, J.R.; Schlogl, A.; Pfurtscheller, G.; Millan, J.R.; Schroder, M.; Birbaumer, N. The BCI competition III: Validating alternative approaches to actual BCI problems. IEEE Trans. Neural Syst. Rehabil. Eng. 2006, 14, 153–159. 212. McFarland, D.J.; Krusienski, D.J.; Wolpaw, J.R.; Brain-Computer Interface Signal Processing at the Wadsworth Center: mu and Sensorimotor Beta Rhythms. In Progress in Brain Research; Christa, N., Wolfgang, K., Eds.; Elsevier: New York, NY, USA, 2006; Volume 159, pp. 411–419. 213. Daly, J.J.; Wolpaw, J.R. Brain-computer interfaces in neurological rehabilitation. Lancet Neurol. 2008, 7, 1032–1043. 214. Shenoy, P.; Krauledat, M.; Blankertz, B.; Rao, R.P.N.; Müller, K.-R. Towards adaptive classification for BCI. J. Neural Eng. 2006, 3, R13. 215. Millan, J.R.; Mourino, J. Asynchronous BCI and local neural classifiers: an overview of the adaptive brain interface project. IEEE Trans. Neural Syst. Rehabil. Eng. 2003, 11, 159–161. 216. Galán, F.; Nuttin, M.; Lew, E.; Ferrez, P.W.; Vanacker, G.; Philips, J.; Millán, J.D.R. A brain-actuated wheelchair: Asynchronous and non-invasive brain-computer interfaces for continuous control of robots. Clin. Neurophysiol. 2008, 119, 2159–2169. 217. Li, Y.; Guan, C.; Li, H.; Chin, Z. A self-training semi-supervised SVM algorithm and its application in an EEG-based brain computer interface speller system. Pattern Recogn. Lett. 2008, 29, 1285–1294. 218. Vidaurre, C.; Sannelli, C.; Müller, K.-R.; Blankertz, B. Machine-learning-based coadaptive calibration for brain-computer interfaces. Neural Comput. 2010, 23, 791–816. 219. Shijian, L.; Cuntai, G.; Haihong, Z. Unsupervised brain computer interface based on intersubject information and online adaptation. IEEE Trans. Neural Syst. Rehabil. Eng. 2009, 17, 135–145. 220. Yan, L.; Kambara, H.; Koike, Y.; Sugiyama, M. Application of covariate shift adaptation techniques in brain-computer interfaces. IEEE Trans. Biomed. Eng. 2010, 57, 1318–1324. 221. Ferrez, P.W.; del R. Millan, J. Error-related EEG potentials generated during simulated brain-computer interaction. IEEE Trans. Biomed. Eng. 2008, 55, 923–929. 222. Millan, J.R. On the Need for On-Line Learning in Brain-Computer Interfaces. In Proceedings of the IEEE International Joint Conference on Neural Networks, (IJCNN’04), Budapest, Hungary, July 2004; pp. 2877-2882. 223. Jain, A.K.; Duin, R.P.W.; Jianchang, M. Statistical pattern recognition: A review. IEEE Trans. Pattern Anal. 2000, 22, 4–37. 224. Lemm, S.; Blankertz, B.; Dickhaus, T.; Müller, K.-R. Introduction to machine learning for brain imaging. NeuroImage 2011, 56, 387–399. 225. Denoeux, T. A k-nearest neighbor classification rule based on Dempster-Shafer theory. IEEE Trans. Syst. Man. Cyb. 1995, 25, 804–813. 226. Dudani, S.A. The distance-weighted k-nearest-neighbor rule. IEEE Trans. Syst. Man. Cyb. 1976, SMC-6, 325–327. 227. Borisoff, J.F.; Mason, S.G.; Bashashati, A.; Birch, G.E. Brain-computer interface design for asynchronous control applications: Improvements to the LF-ASD asynchronous brain switch. IEEE Trans. Biomed. Eng. 2004, 51, 985–992. Sensors 2012, 12 1273 228. Schlögl, A.; Lee, F.; Bischof, H.; Pfurtscheller, G. Characterization of four-class motor imagery EEG data for the BCI-competition 2005. J. Neural Eng. 2005, 2, L14. 229. Kayikcioglu, T.; Aydemir, O. A polynomial fitting and k-NN based approach for improving classification of motor imagery BCI data. Pattern Recogn. Lett. 2010, 31, 1207–1215. 230. Garrett, D.; Peterson, D.A.; Anderson, C.W.; Thaut, M.H. Comparison of linear, nonlinear, and feature selection methods for EEG signal classification. IEEE Trans. Neural Syst. Rehabil. Eng. 2003, 11, 141–144. 231. Scherer, R.; Muller, G.R.; Neuper, C.; Graimann, B.; Pfurtscheller, G. An asynchronously controlled EEG-based virtual keyboard: Improvement of the spelling rate. IEEE Trans. Biomed. Eng. 2004, 51, 979–984. 232. Muller, K.R.; Anderson, C.W.; Birch, G.E. Linear and nonlinear methods for brain-computer interfaces. IEEE Trans. Neural Syst. Rehabil. Eng. 2003, 11, 165–169. 233. Blankertz, B.; Lemm, S.; Treder, M.; Haufe, S.; Müller, K.R. Single-trial analysis and classification of ERP components—A tutorial. Neuroimage 2011, 56, 814–825. 234. Vidaurre, C.; Krämer, N.; Blankertz, B.; Schlögl, A. Time domain parameters as a feature for EEG-based brain-computer interfaces. Neural Netw. 2009, 22, 1313–-1319. 235. Hoffmann, U.; Vesin, J.M.; Ebrahimi, T.; Diserens, K. An efficient P300-based brain-computer interface for disabled subjects. J. Neurosci. Methods 2008, 167, 115–125. 236. Burges, C.J.C. A tutorial on support vector machines for pattern recognition. Data Min. Knowl. Discov. 1998, 2, 121–167. 237. Xiang, L.; Dezhong, Y.; Wu, D.; Chaoyi, L. Combining spatial filters for the classification of single-trial EEG in a finger movement task. IEEE Trans. Biomed. Eng. 2007, 54, 821–831. 238. Cover, T.M. Geometrical and statistical properties of systems of linear inequalities with applications in pattern recognition. IEEE Trans. Electron. Comput. 1965, EC-14, 326–334. 239. Garcia, G.N.; Ebrahimi, T.; Vesin, J.M. Support Vector EEG Classification in the Fourier and Time-Frequency Correlation Domains. In Proceedings of the First International IEEE EMBS Conference on Neural Engineering (NER’03), Capri Island, Italy, March 2003; pp. 591–594. 240. Blankertz, B.; Curio, G.; Müller, K. Classifying Single Trial EEG: Towards Brain Computer Interfacing. In Proceedings of the Advances in Neural Information Processing Systems (ANPS’02), Vancouver, BC, Canada, December 2002; pp. 157–164. 241. Haihong, Z.; Cuntai, G.; Chuanchu, W. Asynchronous P300-based brain--computer interfaces: A computational approach with statistical models. IEEE Trans. Biomed. Eng. 2008, 55, 1754–1763. 242. Kaper, M.; Meinicke, P.; Grossekathoefer, U.; Lingner, T.; Ritter, H. BCI competition 2003-data set IIb: Support vector machines for the P300 speller paradigm. IEEE Trans. Biomed. Eng. 2004, 51, 1073–1076. 243. Meinicke, P.; Kaper, M.; Hoppe, F.; Heumann, M.; Ritter, H. Improving Transfer Rates in Brain Computer Interfacing: A Case Study. In Proceedings of the Advances in Neural Inf. Proc. Systems (NIPS’02), Vancouver, BC, Canada, December 2002; pp. 1107–1114. 244. Thulasidas, M.; Cuntai, G.; Jiankang, W. Robust classification of EEG signal for brain-computer interface. IEEE Trans. Neural Sys. Rehabil. Eng. 2006, 14, 24–29. Sensors 2012, 12 1274 245. Ruiting, Y.; Gray, D.A.; Ng, B.W.; Mingyi, H. Comparative Analysis of Signal Processing in Brain Computer Interface. In Proceedings of the 4th IEEE Conference on Industrial Electronics and Applications (ICIEA’09), Xi'an, China, May 2009; pp. 580-585. 246. Liu, G.; Huang, G.; Meng, J.; Zhang, D.; Zhu, X. Improved GMM with parameter initialization for unsupervised adaptation of Brain-Computer interface. Int. J. Numer. Meth. BioMed. Eng. 2010, 26, 681–691. 247. Lemm, S.; Schafer, C.; Curio, G. BCI competition 2003-data set III: Probabilistic modeling of sensorimotor mu rhythms for classification of imaginary hand movements. IEEE Trans. Biomed. Eng. 2004, 51, 1077–1080. 248. Pires, G.; Castelo-Branco, M.; Nunes, U. Visual P300-Based BCI to Steer a Wheelchair: A Bayesian Approach. In Proceedings of the 30th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBS’08), Vancouver, Canada, August 2008; pp. 658–661. 249. Palaniappan, R. Brain Computer Interface Design Using Band Powers Extracted During Mental Tasks. In Proceedings of the 2nd International IEEE EMBS Conference on Neural Engineering (NER’05), Arlington, VA, USA, March 2005; pp. 321–324. 250. Nakayaman, K.; Inagaki, K. A Brain Computer Interface Based on Neural Network with Efficient Pre-Processing. In Proceedings of the International Symposium on Intelligent Signal Processing and Communications (ISPACS’06), Yonago, Japan, December 2006; pp. 673–676. 251. Haselsteiner, E.; Pfurtscheller, G. Using time-dependent neural networks for EEG classification. IEEE Trans. Rehabil. Eng. 2000, 8, 457–463. 252. Coyle, D.; McGinnity, T.M.; Prasad, G. Improving the separability of multiple EEG features for a BCI by neural-time-series-prediction-preprocessing. Biomed. Signal Process. Control 2010, 5, 196–204. 253. Hazrati, M.K.; Erfanian, A. An online EEG-based brain-computer interface for controlling hand grasp using an adaptive probabilistic neural network. Med. Eng. Phys. 2010, 32, 730–739. 254. Ting, W.; Guo-zheng, Y.; Bang-hua, Y.; Hong, S. EEG feature extraction based on wavelet packet decomposition for brain computer interface. Measurement 2008, 41, 618–625. 255. Palaniappan, R.; Paramesran, R.; Nishida, S.; Saiwaki, N. A new brain-computer interface design using fuzzy ARTMAP. IEEE Trans. Neural Sys. Rehabil. Eng. 2002, 10, 140–148. 256. Felzer, T.; Freisieben, B. Analyzing EEG signals using the probability estimating guarded neural classifier. IEEE Trans. Neural Syst. Rehabil. Eng. 2003, 11, 361–371. 257. Kübler, A.; Birbaumer, N. Brain-computer interfaces and communication in paralysis: Extinction of goal directed thinking in completely paralysed patients? Clin. Neurophysiol. 2008, 119, 2658–2666. 258. Betke, M.; Gips, J.; Fleming, P. The Camera Mouse: Visual tracking of body features to provide computer access for people with severe disabilities. IEEE Trans. Neural Sys. Rehabil. Eng. 2002, 10, 1–10. 259. Miller, L.A.; Stubblefield, K.A.; Lipschutz, R.D.; Lock, B.A.; Kuiken, T.A. Improved myoelectric prosthesis control using targeted reinnervation surgery: A case series. IEEE Trans. Neural Sys. Rehabil. Eng. 2008, 16, 46–50. Sensors 2012, 12 1275 260. Treder, M.S.; Blankertz, B. (C) overt attention and visual speller design in an ERP-based brain-computer interface. Behav. Brain Funct. 2010, 6, 28. 261. Pfurtscheller, G.; Allison, B.Z.; Bauernfeind, G.n.; Brunner, C.; Solis Escalante, T.; Scherer, R.; Zander, T.O.; Mueller-Putz, G.; Neuper, C.; Birbaumer, N. The hybrid BCI. Front. Neurosci. 2010, doi: 10.1186/1744-9081-6-28. 262. Liu, Y.; Zhou, Z.; Hu, D. Gaze independent brain-computer speller with covert visual search tasks. Clin. Neurophysiol. 2011, 122, 1127–1136. 263. Treder, M.S.; Schmidt, N.M.; Blankertz, B. Gaze-independent brain-computer interfaces based on covert attention and feature attention. J. Neural Eng. 2011, 8, 066003. 264. Moore, M.M. Real-world applications for brain-computer interface technology. IEEE Trans. Neural Syst. Rehabil. Eng. 2003, 11, 162–165. 265. Birbaumer, N.; Ghanayim, N.; Hinterberger, T.; Iversen, I.; Kotchoubey, B.; Kubler, A.; Perelmouter, J.; Taub, E.; Flor, H. A spelling device for the paralysed. Nature 1999, 398, 297–298. 266. Chambayil, B.; Singla, R.; Jha, R. Virtual Keyboard BCI Using Eye Blinks in EEG. In Proceedings of the 6th IEEE International Conference on Wireless and Mobile Computing, Networking and Communications (WiMob’10), Niagara Falls, ON, Canada, October 2010; pp. 466–470. 267. Obermaier, B.; Muller, G.R.; Pfurtscheller, G. “Virtual keyboard” controlled by spontaneous EEG activity. IEEE Trans. Neural Sys. Rehabil. Eng. 2003, 11, 422–426. 268. Silvoni, S.; Volpato, C.; Cavinato, M.; Marchetti, M.; Priftis, K.; Merico, A.; Tonin, P.; Koutsikos, K.; Beverina, F.; Piccione, F. P300-based brain-computer interface communication: evaluation and follow-up in amyotrophic lateral sclerosis. Front. Neurosci. 2009, 3, 60. 269. Intendix. Available online: http://www.intendix.com/ (accessed on 12 July 2011). 270. Fazel-Rezai, R.; Abhari, K. A region-based P300 speller for brain-computer interface. Can. J. Elect. Comput. E. 2009, 34, 81–85. 271. Ahi, S.T.; Kambara, H.; Koike, Y. A dictionary-driven P300 speller with a modified interface. IEEE Trans. Neural Syst. Rehabil. Eng. 2011, 19, 6–14. 272. Halder, S.; Rea, M.; Andreoni, R.; Nijboer, F.; Hammer, E.M.; Kleih, S.C.; Birbaumer, N.; Kübler, A. An auditory oddball brain-computer interface for binary choices. Clin. Neurophysiol. 2010, 121, 516–523. 273. Klobassa, D.S.; Vaughan, T.M.; Brunner, P.; Schwartz, N.E.; Wolpaw, J.R.; Neuper, C.; Sellers, E.W. Toward a high-throughput auditory P300-based brain-computer interface. Clin. Neurophysiol. 2009, 120, 1252–1261. 274. Kübler, A.; Furdea, A.; Halder, S.; Hammer, E.M.; Nijboer, F.; Kotchoubey, B. A brain-computer interface controlled auditory event-related potential (P300) spelling system for Locked-in patients. Ann. N.Y. Acad. Sci. 2009, 1157, 90–100. 275. Karim, A.A.; Hinterberger, T.; Richter, J.; Mellinger, J.; Neumann, N.; Flor, H.; Kübler, A.; Birbaumer, N. Neural internet: Web surfing with brain potentials for the completely paralyzed. Neurorehabil. Neural Repair 2006, 20, 508–515. 276. Bensch, M.; Karim, A.A.; Mellinger, J.; Hinterberger, T.; Tangermann, M.; Bogdan, M.; Rosenstiel, W.; Birbaumer, N. Nessi: An EEG-controlled web browser for severely paralyzed patients. Comput. Intell. Neurosci. 2007, 2007, 71863. Sensors 2012, 12 1276 277. Braz, G.P.; Russold, M.; Davis, G.M. Functional electrical stimulation control of standing and stepping after spinal cord injury: A review of technical characteristics. Neuromodulation: Technol. Neural Interface 2009, 12, 180–190. 278. Lauer, R.T.; Peckham, P.H.; Kilgore, K.L. EEG-based control of a hand grasp neuroprosthesis. NeuroReport 1999, 10, 1767–1771. 279. Pfurtscheller, G.; Müller, G.R.; Pfurtscheller, J.; Gerner, H.J.; Rupp, R. ‘Thought’-control of functional electrical stimulation to restore hand grasp in a patient with tetraplegia. Neurosci. Lett. 2003, 351, 33–36. 280. Hu, X.L.; Tong, K.Y.; Li, R.; Chen, M.; Xue, J.J.; Ho, S.K.; Chen, P.N. Post-stroke wrist rehabilitation assisted with an intention-driven functional electrical stimulation (FES)-robot system. In Proceedings of the IEEE International Conference on Rehabilitation Robotics (ICORR’11), Zurich, Switzerland, July 2011; pp. 1–6. 281. Pfurtscheller, G.; Guger, C.; Müller, G.; Krausz, G.; Neuper, C. Brain oscillations control hand orthosis in a tetraplegic. Neurosci. Lett. 2000, 292, 211–214. 282. Müller-Putz, G.R.; Scherer, R.; Pfurtscheller, G.; Rupp, R. EEG-based neuroprosthesis control: A step towards clinical practice. Neurosci. Lett. 2005, 382, 169–174. 283. Buch, E.; Weber, C.; Cohen, L.G.; Braun, C.; Dimyan, M.A.; Ard, T.; Mellinger, J.; Caria, A.; Soekadar, S.; Fourkas, A.; Birbaumer, N. Think to move: A neuromagnetic brain-computer interface (BCI) system for chronic stroke. Stroke 2008, 39, 910–917. 284. Braun, C.; Schweizer, R.; Elbert, T.; Birbaumer, N.; Taub, E. Differential activation in somatosensory cortex for different discrimination tasks. J. Neurosci. 2000, 20, 446–450. 285. Broetz, D.; Braun, C.; Weber, C.; Soekadar, S.R.; Caria, A.; Birbaumer, N. Combination of brain-computer interface training and goal-directed physical therapy in chronic stroke: A case report. Neurorehabil. Neural Repair 2010, 24, 674–679. 286. Caria, A.; Weber, C.; Brötz, D.; Ramos, A.; Ticini, L.F.; Gharabaghi, A.; Braun, C.; Birbaumer, N. Chronic stroke recovery after combined BCI training and physiotherapy: A case report. Psychophysiology 2011, 48, 578–582. 287. Hochberg, L.R.; Serruya, M.D.; Friehs, G.M.; Mukand, J.A.; Saleh, M.; Caplan, A.H.; Branner, A.; Chen, D.; Penn, R.D.; Donoghue, J.P. Neuronal ensemble control of prosthetic devices by a human with tetraplegia. Nature 2006, 442, 164–171. 288. Serruya, M.D.; Hatsopoulos, N.G.; Paninski, L.; Fellows, M.R.; Donoghue, J.P. Brain-machine interface: Instant neural control of a movement signal. Nature 2002, 416, 141–142. 289. Millan, J.R.; Renkens, F.; Mourino, J.; Gerstner, W. Noninvasive brain-actuated control of a mobile robot by human EEG. IEEE Trans. Biomed. Eng. 2004, 51, 1026–1033. 290. Rebsamen, B.; Burdet, E.; Cuntai, G.; Chee Leong, T.; Qiang, Z.; Ang, M.; Laugier, C. Controlling a Wheelchair Using a BCI with Low Information Transfer Rate. In Proceedings of the IEEE 10th International Conference on Rehabilitation Robotics (ICORR’07), Noordwijk, The Netherlands, June 2007; pp. 1003–1008. 291. Rebsamen, B.; Cuntai, G.; Haihong, Z.; Chuanchu, W.; Cheeleong, T.; Ang, M.H.; Burdet, E. A brain controlled wheelchair to navigate in familiar environments. IEEE Trans. Neural Syst. Rehabil. Eng. 2010, 18, 590–598. Sensors 2012, 12 1277 292. Iturrate, I.; Antelis, J.M.; Kubler, A.; Minguez, J. A Noninvasive brain-actuated wheelchair based on a P300 neurophysiological protocol and automated navigation. IEEE Trans. Robot. 2009, 25, 614–627. 293. Philips, J.; del R. Millan, J.; Vanacker, G.; Lew, E.; Galan, F.; Ferrez, P.W.; Van Brussel, H.; Nuttin, M. Adaptive Shared Control of a Brain-Actuated Simulated Wheelchair. In Proceedings of the IEEE 10th International Conference on Rehabilitation Robotics (ICORR’07), Noordwijk, The Netherlands, June 2007; pp. 408–414. 294. Vanacker, G.; del R. Millán, J.; Lew, E. Context-based filtering for assisted brain-actuated wheelchair driving. Comput. Intell. Neurosci. 2007, 2007, 25130. 295. Fun of gaming: Measuring the human experience of media enjoyment. Available online: http://fuga.aalto.fi/ (accessed on 12 July 2011). 296. Roman, K.; Benjamin, B.; Gabriel, C.; Klaus-Robert, M. The Berlin Brain-Computer Interface (BBCI)—Towards a new communication channel for online control in gaming applications. Multimed. Tools Appl. 2007, 33, 73–90. 297. Tangermann, M.; Krauledat, M.; Grzeska, K.; Sagebaum, M.; Blankertz, B.; Vidaurre, C.; Müller, K. Playing Pinball with non-invasive BCI. Proc. NIPS 2008, 1641–1648. 298. Middendorf, M.; McMillan, G.; Calhoun, G.; Jones, K.S. Brain-computer interfaces based on the steady-state visual-evoked response. IEEE Trans. Rehabil. Eng. 2000, 8, 211–214. 299. Lalor, E.C.; Kelly, S.P.; Finucane, C. Steady-state VEP-based brain-computer interface control in an immersive 3D gaming environment. J. Appl. Signal Proc. 2005, 2005, 3156–3164. 300. Finke, A.; Lenhardt, A.; Ritter, H. The MindGame: A P300-based brain-computer interface game. Neural Netw. 2009, 22, 1329–1333. 301. Microsoft Research—Computational User Experiences: Brain-Computer Interfaces. Available online: http://research.microsoft.com/en-us/um/redmond/groups/cue/bci/ (accessed on 12 July 2011). 302. Angelakis, E.; Stathopoulou, S.; Frymiare, J.; Green, D.; Lubar, J.; Kounios, J. EEG neurofeedback: A brief overview and an example of peak alpha frequency training for cognitive enhancement in the elderly. Clin. Neuropsychol. 2007, 21, 110–129. 303. Hanslmayr, S.; Sauseng, P.; Doppelmayr, M.; Schabus, M.; Klimesch, W. Increasing individual upper alpha power by neurofeedback improves cognitive performance in human subjects. Appl. Psychophysiol. Biofeedback 2005, 30, 1–10. 304. Rota, G.; Sitaram, R.; Veit, R.; Erb, M.; Weiskopf, N.; Dogil, G.; Birbaumer, N. Self-regulation of regional cortical activity using real-time fMRI: The right inferior frontal gyrus and linguistic processing. Hum. Brain Mapp. 2009, 30, 1605–1614. 305. Sitaram, R.; Lee, S.; Ruiz, S.; Rana, M.; Veit, R.; Birbaumer, N. Real-time support vector classification and feedback of multiple emotional brain states. Neuroimage 2011, 56, 753–765. 306. deCharms, R.C.; Maeda, F.; Glover, G.H.; Ludlow, D.; Pauly, J.M.; Soneji, D.; Gabrieli, J.D.E.; Mackey, S.C. Control over brain activation and pain learned by using real-time functional MRI. Proc. Natl. Acad. Sci. USA 2005, 102, 18626–18631. 307. Walker, J.E.; Kozlowski, G.P. Neurofeedback treatment of epilepsy. Child Adolesc. Psychiatr. Clin. N. Am. 2005, 14, 163–176. Sensors 2012, 12 1278 308. Sterman, M.; Egner, T. Foundation and practice of neurofeedback for the treatment of epilepsy. Appl. Psychophysiol. Biofeedback 2006, 31, 21–35. 309. Strehl, U.; Leins, U.; Goth, G.; Klinger, C.; Hinterberger, T.; Birbaumer, N. Self-regulation of slow cortical potentials: A new treatment for children with attention-deficit/hyperactivity disorder. Pediatrics 2006, 118, 1530–1540. 310. Schneider, F.; Rockstroh, B.; Heimann, H.; Lutzenberger, W.; Mattes, R.; Elbert, T.; Birbaumer, N.; Bartels, M. Self-regulation of slow cortical potentials in psychiatric patients: Schizophrenia. Appl. Psychophysiol. Biofeedback 1992, 17, 277–292. 311. Schneider, F.; Heimann, H.; Mattes, R.; Lutzenberger, W.; Birbaumer, N. Self-regulation of slow cortical potentials in psychiatric patients: Depression. Appl. Psychophysiol. Biofeedback 1992, 17, 203–214. 312. Schneider, F.; Elbert, T.; Heimann, H.; Welker, A.; Stetter, F.; Mattes, R.; Birbaumer, N.; Mann, K. Self-regulation of slow cortical potentials in psychiatric patients: Alcohol dependency. Appl. Psychophysiol. Biofeedback 1993, 18, 23–32. 313. Renaud, P.; Joyal, C.; Stoleru, S.; Goyette, M.; Weiskopf, N.; Birbaumer, N. Real-Time Functional Magnetic Imaging—Brain-Computer Interface and Virtual Reality: Promising Tools for the Treatment of Pedophilia. In Progress in Brain Research; Green, A.M., Chapman, C.E., Kalaska, J.F., Lepore, F., Eds.; Elsevier: New York, NY, USA, 2011; Volume 192, pp. 263–272. 314. Georgopoulos, A.P.; Karageorgiou, E.; Leuthold, A.C.; Lewis, S.M.; Lynch, J.K.; Alonso, A.A.; Aslam, Z.; Carpenter, A.F.; Georgopoulos, A.; Hemmy, L.S.; et al. Synchronous neural interactions assessed by magnetoencephalography: A functional biomarker for brain disorders. J. Neural Eng. 2007, doi: 10.1088/1741-2560/4/4/001. 315. Cook, I.A.; Warren, C.; Pajot, S.K.; Schairer, D.; Leuchter, A.F. Regional Brain activation with advertising images. J. Neurosci. Psychol. Econ. 2011, 4, 147–160. 316. Ambler, T.; Ioannides, A.; Rose, S. Brands on the brain: Neuro-images of advertising. Bus. Strateg. Rev. 2000, 11, 17–30. 317. Vecchiato, G.; De Vico Fallani, F.; Astolfi, L.; Toppi, J.; Cincotti, F.; Mattia, D.; Salinari, S.; Babiloni, F. The issue of multiple univariate comparisons in the context of neuroelectric brain mapping: An application in a neuromarketing experiment. J. Neurosci. Methods 2010, 191, 283–289. 318. Ambler, T.; Braeutigam, S.; Stins, J.; Rose, S.; Swithenby, S. Salience and choice: Neural correlates of shopping decisions. Psychol. Market 2004, 21, 247–261. 319. Ariely, D.; Berns, G.S. Neuromarketing: The hope and hype of neuroimaging in business. Nat. Rev. Neurosci. 2010, 11, 284–292. 320. Neuromarketing, Neuroscientific Consumer Testing. NeuroFocus. Available online: http://www.neurofocus.com/ (accessed on 28 December 2011). 321. Neuroconsult. Available online: http://www.neuroconsult.at/ (accessed on 8 October 2011). 322. Neuro Insight. Available online: http://www.neuro-insight.com/ (accessed on 8 October 2011). 323. EmSense: Quantitative Biosensory Metrics. Available online: http://www.emsense.com/ (accessed on 8 October 2011). Sensors 2012, 12 1279 324. Fisher, C.E.; Chin, L.; Klitzman, R. Defining neuromarketing: Practices and professional challenges. Harv. Rev. Psychiatry 2010, 18, 230–237. © 2012 by the authors; licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution license (http://creativecommons.org/licenses/by/3.0/). View publication stats