Brain Computer Interfaces

Slides:



Advertisements
Similar presentations
Abstract Electrical activity in the cortex can be recorded by surface electrodes. Electro Encephalography (EEG) machine records potential difference between.
Advertisements

TMS-evoked EEG responses in symptomatic and recovered patients with mild traumatic brain injury Jussi Tallus 1, Pantelis Lioumis 2, Heikki Hämäläinen 3,
Electroencephalography (EEG)
Non-Invasive BCI.
IntroductionMethods Participants  7 adults with severe motor impairment.  9 adults with no motor impairment.  Each participant was asked to utilize.
EEG An Introduction Aamir Saeed Malik Neuro-Signal Processing Group Universiti Teknologi Petronas Malaysia.
Brain-computer interfaces: classifying imaginary movements and effects of tDCS Iulia Comşa MRes Computational Neuroscience and Cognitive Robotics Supervisors:
1 1 MPI for Biological Cybernetics 2 Stanford University 3 Werner Reichardt Centre for Integrative Neuroscience Eberhard Karls University Tuebingen Epidural.
Sankalp Patel, Michael Kang, James Kelly.  The EEG (electroencephalogram) uses highly conductive silver electrodes coated with silver- chloride and gold.
Billy Vermillion. EEG  Electroencephalography A test to measure the electrical activity of the brain. ○ Brain cells communicate by producing tiny electrical.
BCI Systems Brendan Allison, Ph.D. Institute for Automation University of Bremen 6 November, 2008.
Sleep and Dreaming Methodology PAGE 48. EEG  electroencephalogram.
EEG. These are neurons. Your brain has hundreds of billions of them! Diagram of a neuron.A group of real neurons.
COMSATS Institute of Information Technology,Sahiwal.
Papavasileiou-1 CSE 5810 Brain Computer Interface in BMI Ioannis Papavasileiou Computer Science & Engineering Department The University of Connecticut.
NeuroPhone : Brain-Mobile Phone Interface using a wireless EEG Headset Ilho nam.
Brain-Computer Interfaces for Communication in Paralysis: A Clinical Experimental Approach By Adil Mehmood Khan.
Review, Hollan & Gorodnitsky Adrienne Moore, by
Human Brain the most complex living structure on the universe وَ فِیۡۤ اَنۡفُسِكُمْ اَفَلَا تُبْصِرُوۡنَ -- سُوۡرَۃُ الذّٰرِیٰتِ Dr. Abdel Ilah Alshbatat.
Da’Janel Roberts Matthew Morgan Jonathan James
Brain-Computer-Interface for Stress Detection Hassan Farooq, Ilona Wong Supervisor: Steve Mann Administrator: Cristiana Amza Section 8 Collecting Brainwave.
Shyamanta M. Hazarika, Adity Saikia, Simanta Bordoloi, Ujjal Sharma And Nayantara Kotoky Department Of Computer Sc. & Engineering, Tezpur University Tezpur,
Graz-Brain-Computer Interface: State of Research By Hyun Sang Suh.
Functional Brain Signal Processing: EEG & fMRI Lesson 4
Operant Conditioning of Cortical Activity E Fetz, 1969.
Virtual Reality in Brain- Computer Interface Research F. Lee 1, R. Scherer 2, H. Bischof 1, G. Pfurtscheller 2 1) Institute for Computer Graphics and Vision.
 A direct communication pathway between the brain and an external device.  Directed at assisting, augmenting, or repairing human cognitive or sensory-motor.
EEG-Based Communication and Control: Short-Term Role Feedback Present by: Yu Yuan-Chu Dennis J. Mcfarland, Lynn M. McCane, and J. R. Wolpaw.
Electromyography (EMG)
Electrocorticography-Based Brain Computer Interface – The Seattle Experience.
Analysis of Movement Related EEG Signal by Time Dependent Fractal Dimension and Neural Network for Brain Computer Interface NI NI SOE (D3) Fractal and.
Motivation Increase bandwidth of BCI. Reduce training time Use non invasive technique.
Electroencephalography Collecting and Analyzing Data: 101
Six seconds of data, recorded 5 minutes apart. Electroencephalography First recording of electrical fields of animals, Caton (1875); humans, Berger.
Memory, Brain (Chapter 8.1) Vengertsev Dmitry. Agenda Goals: 1) provide general overview of what we do know about underlying biological processes and.
1 Psychology 304: Brain and Behaviour Lecture 2. 2 Research Methods 1.What techniques do biological psychologists use to assess the structure and function.
Dr. Ali Saad modified from Dr. Carlos Davila Southe. metho univ 1 EEG Brain signal measurement and analysis 414BMT Dr Ali Saad, College of Applied medical.
HKIN 473 Recording Motor Units. Recording Electrical Signals Muscle fiber sarcolemma action potential is very small ~ 1 millivolt. Therefore, to be able.
BRAIN GATE TECHNOLOGY.. Brain gate is a brain implant system developed by the bio-tech company Cyberkinetics in 2003 in conjunction with the Department.
EEG Biofeedback Background Report Adrian Smith, gte198f Daniel Shinn, gte539f Ken Grove, gte262f ECE Group N1 January 31, 2002.
MIND CONTROLLED ROBOT BY ADITHYA KUMAR EIGHTH GRADE.
1 Psychology 304: Brain and Behaviour Lecture 4. 2 Research Methods and The Structure of the Nervous System 2. What are the primary divisions of the nervous.
Brain-Computer Interfaces
Biofeedback and Stress Management John Harrison – Leader Albert Kwansa – Communicator Eric Lee – BSAC Brenton Nelson – BWIG Dr. Daniel Muller – Client.
BMI DEVELOPMENT SYSTEMS -SRIKANTH.B. INTRODUCTION The core of this paper is that to operate machines from a remote area. In the given BMI DEVELOPMENT.
Introduction to Brain Computer Interface (BCI) Systems
Brain Machine Interface. EEGs Neurons is like a battery; when active, it’s voltage changes Free running EEGs vs ERP (event related potential)
SEMINAR on ‘BRAIN COMPUTER INTERFACE’ Submitted by: JYOTI DOSAYA
ELECTROENCEPHALOGRAPHY (EEG)
Brain Computer Interfaces
SCHEMATIC DIAGRAM OF AN EEG MACHINE
Brain Computer Interface
Bongjae Choi, Sungho Jo Presented by: Yanrong Wo
Brain operated wheelchair
[Ran Manor and Amir B.Geva] Yehu Sapir Outlines Review
The Thought Translation device
Brain Electrophysiological Signal Processing: Postprocessing
Electroencephalogram (EEG)
When to engage in interaction – and how
Fabien LOTTE, Cuntai GUAN Brain-Computer Interfaces laboratory
Major Project Presentation Phase - I
Mind Control of Systems
Playback control using mind
TRENDS IN CNS DEVELOPMENT
Brain Computer Interfaces
Ways of Studying the Brain
Ways of Studying the Brain
Sleep stages Awake Stage 1 Stage 2 Stage 3 Stage 4
The Berlin Brain-Computer Interface: Machine Learning-Based Detection of User Specific Brain States Umar Farooq.
A Low-Cost EEG System-Based Hybrid Brain-Computer Interface for Humanoid Robot Navigation and Recognition Bongjae Choi, Sungho Jo Presented by Megan Fillion.
Presentation transcript:

Brain Computer Interfaces or Krang’s Body

What is an EEG? An electroencephalogram is a measure of the brain's voltage fluctuations as detected from scalp electrodes. It is an approximation of the cumulative electrical activity of neurons.

What is it good for? Neurofeedback Brain Computer Interfaces treating ADHD guiding meditation Brain Computer Interfaces People with little muscle control (i.e. not enough control for EMG or gaze tracking) People with ALS, spinal injuries High Precision Low bandwidth (bit rate)

EEG Background 1875 - Richard Caton discovered electrical properties of exposed cerebral hemispheres of rabbits and monkeys. 1924 - German Psychiatrist Hans Berger discovered alpha waves in humans and invented the term “electroencephalogram” 1950s - Walter Grey Walter developed “EEG topography” - mapping electrical activity of the brain.

Physical Mechanisms EEGs require electrodes attached to the scalp with sticky gel Require physical connection to the machine At present perhaps the most cumbersome factor is the need for scalp electrodes, which require an electrolyte gel for electrical conductivity, and as little hair as possible. Users with normal hair have to deal with electrode prep before use and hair cleaning after use. http://www.icad.org/websiteV2.0/Conferences/ICAD2004/concert_call.htm The scalp electrodes may always be the limiting factor in resolution of a EEG- computer interface. There is probably much electrical activity concomitant with thought patterns and sensory images in the brain, but the fine resolution of this activity is not detectable with surface electrodes. Another difficulty, is that the EP systems are quite slow. The EP must be derived by signal averaging, that is, multiple repetitions of the evoked response must be accumulated in order to see the EP signal above the noise. In the case of Dr. Sutter's system, 1.5 seconds is required to discriminate the selection of a particular letter from the alphabet array. The continuous EEG interface systems have faster switch functions because the change in alpha or mu wave amplitude can be detected more quickly.

Electrode Placement Standard “10-20 System” Spaced apart 10-20% Letter for region F - Frontal Lobe T - Temporal Lobe C - Center O - Occipital Lobe Number for exact position Odd numbers - left Even numbers - right The listener wore headphones to hear the music, and a cap with EEG sensors on it to record neural activity. The 26 sensor electrodes were arranged according to the 10-20 standard for EEG placement. faculty.washington.edu/chudler/1020.html. The sensors are labelled by proximity over a regions of the brain (F=Front, T=Temporal, C=Central, P-Parietal, O=Occipital) followed by either a 'z' for the midline, or a number that increases as it moves further from the midline. Odd numbers (1,3,5) are on the left hemisphere and even numbers (2,4,6) on the right e.g. T4 is on the right temporal lobe, above the right ear. An additional 10 sensors were used to record heart-rate, skin conductance, eye movements, breathing and other data. The sensors were recorded as interleaved channels of signed 32 bit integers at a rate of 500 samples per second. The channels were separated into individually named files and converted to ascii format for simplicity of loading on different systems. http://www.icad.org/websiteV2.0/Conferences/ICAD2004/concert_call.htm http://faculty.washington.edu/chudler/1020.html

Electrode Placement A more detailed view:

Brain “Features” User must be able to control the output: use a feature of the continuous EEG output that the user can reliably modify (waves), or evoke an EEG response with an external stimulus (evoked potential)

Continuous Brain Waves Generally grouped by frequency: (amplitudes are about 100µV max) Type Frequency Location Use Delta <4 Hz everywhere occur during sleep, coma Theta 4-7 Hz temporal and parietal correlated with emotional stress (frustration & disappointment) Alpha 8-12 Hz occipital and parietal reduce amplitude with sensory stimulation or mental imagery Beta 12-36 Hz parietal and frontal can increase amplitude during intense mental activity Mu 9-11 Hz frontal (motor cortex) diminishes with movement or intention of movement Lambda sharp, jagged occipital correlated with visual attention Vertex higher incidence in patients with epilepsy or encephalopathy The continuous or resting rhythms of the brain, "brain waves", are categorized by frequency bands. Different brain wave frequencies correspond to behavioral and attentional states of the brain, and a traditional classification system has long been used to characterize these different EEG rhythms: Alpha waves are between 8 and 13 Hz with amplitude in the range of 25 - 100 µV. They appear mainly from the occipital and parietal brain regions and demonstrate reduced amplitude with afferent stimulation, especially light, and also with intentional visual imagery or mental effort. Beta activity normally occurs in the range of 14 to 30 Hz, and can reach 50 Hz during intense mental activity. Beta arises mainly from the parietal and frontal areas and is associated with the normal alert mental state. Theta waves occur in the 4 to 7 Hz range and arise from the temporal and parietal regions in children, but also occur in adults in response to emotional stress, especially frustration or disappointment. Delta activity is inclusive of all brain waves below 3.5 Hz. Delta occurs in deep sleep, during infancy, and in patients with severe organic brain disease Mu waves, also known the comb or wicket rhythm, appears in bursts at 9 - 11 Hz. This activity appears to be associated with the motor cortex and is diminished with movement or the intention to move. Lambda waves are large electropositive sharp or saw-toothed waves that appear mainly from the occipital region and are associated with visual attention. Vertex waves are electronegative waves of 100 µV amplitude which appear in normal individuals, especially children, in the absence of overt stimulation. These waves have been observed to have a higher incidence in patients with epilepsy or other encephalopathy.

Brain Waves Transformations wave-form averaging over several trials auto-adjustment with a known signal Fourier transforms to detect relative amplitude at different frequencies

Alpha and Beta Waves Studied since 1920s Found in Parietal and Frontal Cortex Relaxed - Alpha has high amplitude Excited - Beta has high amplitude So, Relaxed -> Excited means Alpha -> Beta Alpha waves can also be volitionally manipulated. Alpha activity appears with closing the eyes or defocussed attention. Also, alpha is suppressed by light or normal attentive activity. Thus, most people can learn to produce bursts or "epochs" of alpha activity, and then return to normal beta activity. This behavioral "switch" between beta and alpha activity can be used as the mental command for a brain wave controller. When the signal processor detects the alpha epoch by using an FFT to detect the change in the fundamental frequency of the brain rhythm, an instruction is sent to control an output device.

Mu Waves Studied since 1930s Found in Motor Cortex Amplitude suppressed by Physical Movements, or intent to move physically (Wolpaw, et al 1991) trained subjects to control the mu rhythm by visualizing motor tasks to move a cursor up and down (1D) The mu wave has been studied since the 1930s and came to be referred to as the "wicket rhythm" since the rounded waves on the EEG record resembled a croquet wicket. In a study in the 1950s, Gian Emilio Chatrian and colleagues showed that the amplitude of this wave could be suppressed by physical movements, and later studies showed that simply the intent to move or certain other efforts requiring visual or mental activity would also suppress the amplitude of the mu wave. In Wolpaw and MacFarlands' lab, subjects can learn to control the amplitude of this waveform by trial and error when visualizing various motor activities, such smiling, chewing, or swallowing. For different subjects, different images enhance or suppress the voltage of the mu waveform. Upon detection of the voltage change in the mu wave, the system sends output code to drive a cursor up or down on a computer screen. Thus, with a certain amount of feedback training, users can learn to move the cursor with the appropriate mental effort. The researchers hope that this system will eventually provide a communications link for profoundly disabled individuals.

Mu Waves

Mu and Beta Waves (Wolpaw and McFarland 2004) used a linear combination of Mu and Beta waves to control a 2D cursor. Weights were learned from the users in real time. Cursor moved every 50ms (20 Hz) 92% “hit rate” in average 1.9 sec

Mu and Beta Waves Movie!

Mu and Beta Waves How do you handle more complex tasks? Finite Automata, such as this from (Millán et al, 2004)

P300 (Evoked Potentials) occurs in response to a significant but low-probability event 300 milliseconds after the onset of the target stimulus found in 1965 by (Sutton et al., 1965; Walter, 1965) focus specific

P300 Experiments (Farwell and Donchin 1988) 95% accuracy at 1 character per 26s

P300 (Evoked Potentials) (Polikoff, et al 1995) allowed users to control a cursor by flashing control points in 4 different directions Each sample took 4 seconds Threw out samples masked by muscle movements (such as blinks)

(Polikoff, et al 1995) Results 50% accuracy at ~1/4 Hz 80% accuracy at ~1/30 Hz

VEP - Visual Evoked Potential Detects changes in the visual cortex Similar in use to P300 Close to the scalp

Model Generalization (time) EEG models so far haven’t adjusted to fit the changing nature of the user. (Curran et al 2004) have proposed using Adaptive Filtering algorithms to deal with this.

Model Generalization (users) Many manual adjustments still must be made for each person (such as EEG placement) Currently, users have to adapt to the system rather than the system adapting to the users. Current techniques learn a separate model for each user.

Model Generalization (users) (Müller 2004) applied typical machine learning techniques to reduce the need for training data. Support Vector Machines (SVM) and Regularized Linear Discriminant Analysis (RLDA) This is only the beginning of applying machine learning to BCIs!

BCI Examples - Communication Farwell and Donchin (1988) allowed the user to select a command by looking for P300 signals when the desired command flashed

BCI Examples - Prostheses (Wolpaw and McFarland 2004) allowed a user to move a cursor around a 2 dimensional screen (Millán, et al. 2004) allowed a user to move a robot around the room.

BCI Examples - Music 1987 - Lusted and Knapp demonstrated an EEG controlling a music synthesizer in real time. Atau Tanaka (Stanford Center for Computer Research in Music and Acoustics) uses it in performances to switch synthesizer functions while generating sound using EMG. In 1987, the authors (Lusted and Knapp) demonstrated an EEG controller which was configured to switch settings of a music synthesizer. Music was chosen for the controller's output because sound provided a good demonstration the real-time capabilities of this technology. By wearing a headband that positioned electrodes on the back of the head to detect the occipital alpha activity, users controlled a switch that responded to the transitions between beta and alpha epochs. More recently, composer Atau Tanaka of the Stanford Center for Computer Research in Music and Acoustics uses this EEG controller in his performance pieces to switch certain synthesizer functions while generating sounds using EMG signals. Another recent application for the EEG-alpha interface is being used as a controller for visual keyboard software. In Brazil, Roberto Santini is using a Biomuse system configured to provide him with the EEG switch, since he is immobilized with advanced ALS (amyotrophic lateral sclerosis) and cannot make use of his eye movements to use the EOG controller. With the EEG controller interfaced to the mouse port of his personal computer, Roberto can select letters from the visual keyboard on the screen. The selection process is somewhat laborious because each choice is binary. The word processing software allows him to zoom in on a given letter by dividing the screen in half. Thus, starting from the full keyboard, as many as 6 steps may be required to move down the branching pattern in order to select a desired letter. Roberto now writes complete letters and is pleased that he can again communicate with others. Currently, the authors and a few other researchers, notably a group headed by Alkira Hiraiwa at the NTT Laboratories in Japan, are continuing development in EEG controllers by using pattern recognition algorithms in an attempt to detect signature patterns of EEG activity which correspond to volitional behaviors. The eventual aim is to develop a vocabulary of EEG signals that are recognizable by the computer. The process of pattern recognition is similar to that used for EMG gesture recognition. In this case, the "gesture" is a thought pattern or type of visualization. For instance, attempts have been made to train a neural network to recognize subvocalized letters, where subjects think a particular letter as though about to speak it, and over many repetitions, train the neural net to recognize a brain wave pattern that occurs with this behavior. This is a promising technique, but the training period is laborious in order to obtain a high percentage of accuracy in matching letters with brain wave patterns. As mentioned earlier, another approach to development of an EEG-computer interface involves the use of an evoked potential (EP) paradigm. Evoked potentials are produced by activating a sensory pathway with a particular type of stimulus, such as a flash of light or a noise burst, and then recording a characteristic waveform from the brain at a particular time interval after the stimulus presentation. Since the characteristic evoked waveform appears at a specific time after the stimulus, researchers can discriminate between the EP and the noise because they know its temporal location in the post-stimulus EEG recording. Other electrical activity which occurs before and after the EP latency window can be ignored. Eric Sutter at the Smith-Kettlewell Institute in San Francisco has developed a visual EP controller system for physically handicapped users. The user can select words or phrases from a matrix of flashing squares on a computer screen. The flashing square upon which the user is fixating his or her gaze produces a characteristic EP from a particular portion of the visual cortex, and since the amplitude of the EP produced from the foveal portion (point of maximal accuity) of the retina is much larger than the response form surrounding retinal areas, the computer can discriminate which word square the user is watching at any given time. Dr. Sutter has implanted electrodes under the scalp to improve the quality of the EEG signal in these patients. Also, this eliminates the need to put on scalp electrodes for each test session since the patients simply "plug in" their transdermal connection to interface with the computer.

In Review… Brain Computer Interfaces Allow those with poor muscle control to communicate and control physical devices High Precision (can be used reliably) Requires somewhat invasive sensors Requires extensive training (poor generalization) Low bandwidth (today 24 bits/minute, or at most 5 characters/minute)

Future Work Improving physical methods for gathering EEGs Improving generalization Improving knowledge of how to interpret waves (not just the “new phrenology”)

References http://www.cs.man.ac.uk/aig/staff/toby/research/bci/richard.seabrook.brain.computer.interface.txt http://www.icad.org/websiteV2.0/Conferences/ICAD2004/concert_call.htm http://faculty.washington.edu/chudler/1020.html http://www.biocontrol.com/eeg.html http://www.asel.udel.edu/speech/Spch_proc/eeg.html Toward a P300-based Computer Interface James B. Polikoff, H. Timothy Bunnell, & Winslow J. Borkowski Jr. Applied Science and Engineering Laboratories Alfred I. Dupont Institute Various papers from PASCAL 2004 Original Paper on Evoked Potential: https://access.web.cmu.edu/http://www.jstor.org/cgi-bin/jstor/viewitem/00368075/ap003886/00a00500/0?searchUrl=http%3a//www.jstor.org/search/Results%3fQuery%3dEvoked-Potential%2bPotential%2bCorrelates%2bof%2bStimulus%2bUncertainty%26hp%3d25%26so%3dnull%26si%3d1%26mo%3dbs&frame=noframe&dpi=3&currentResult=00368075%2bap003886%2b00a00500%2b0%2c07&userID=80020b32@cmu.edu/01cce4403532f102af429e95&backcontext=page

Invasive BCIs Have traditionally provided much finer control than non-invasive EEGs (no longer true?) May have ethical/practical issues (Chapin et al. 1999) trained rats to control a “robot arm” to fetch water (Wessberg et al. 2000) allowed primates to accurately control a robot arm in 3 dimensions in real time.