

Abrain–computer interface (BCI), sometimes called abrain–machine interface (BMI), is a direct communication link between thebrain's electrical activity and an external device, most commonly a computer or robotic limb. BCIs are often directed at researching,mapping, assisting,augmenting, or repairing humancognitive orsensory-motor functions.[1] They are often conceptualized as ahuman–machine interface that skips the intermediary of moving body parts (e.g.hands orfeet). BCI implementations range from non-invasive (EEG,MEG,MRI) and partially invasive (ECoG and endovascular) to invasive (microelectrode array), based on how physically close electrodes are to brain tissue.[2]
Research on BCIs began in the 1970s by Jacques Vidal at theUniversity of California, Los Angeles (UCLA) under a grant from theNational Science Foundation, followed by a contract from theDefense Advanced Research Projects Agency (DARPA).[3][4] Vidal's 1973 paper introduced the expressionbrain–computer interface into scientific literature.
Due to thecortical plasticity of the brain, signals from implantedprostheses can, after adaptation, be handled by the brain like natural sensor or effector channels.[5] Following years of animal experimentation, the firstneuroprosthetic devices were implanted in humans in the mid-1990s.
The history of brain-computer interfaces (BCIs) starts withHans Berger's discovery of the brain's electrical activity and the development ofelectroencephalography (EEG). In 1924 Berger was the first to record human brain activity utilizing EEG. Berger was able to identifyoscillatory activity, such as thealpha wave (8–13 Hz), by analyzing EEG traces.[citation needed]
Berger's first recording device was rudimentary. He insertedsilver wires under the scalps of his patients. These were later replaced by silver foils attached to the patient's head by rubber bandages. Berger connected these sensors to aLippmann capillary electrometer, with disappointing results. However, more sophisticated measuring devices, such as theSiemens double-coil recordinggalvanometer, which displayedvoltages as small as 10−4 volt, led to success.[citation needed]
Berger analyzed the interrelation of alternations in his EEG wave diagrams withbrain diseases. EEGs permitted completely new possibilities for brain research.[citation needed]
Although the term had not yet been coined, one of the earliest examples of a working brain-machine interface was the pieceMusic for Solo Performer (1965) by American composerAlvin Lucier. The piece makes use of EEG andanalog signal processing hardware (filters, amplifiers, and a mixing board) to stimulate acoustic percussion instruments. Performing the piece requires producingalpha waves and thereby "playing" the various instruments via loudspeakers that are placed near or directly on the instruments.[6]
Jacques Vidal coined the term "BCI" and produced the first peer-reviewed publications on this topic.[3][4] He is widely recognized as the inventor of BCIs.[7][8][9] A review pointed out that Vidal's 1973 paper stated the "BCI challenge"[10] of controlling external objects using EEG signals, and especially use ofContingent Negative Variation (CNV) potential as a challenge for BCI control. Vidal's 1977 experiment was the first application of BCI after his 1973 BCI challenge. It was a noninvasive EEG (actuallyVisual Evoked Potentials (VEP)) control of a cursor-like graphical object on a computer screen. The demonstration was movement in a maze.[11]
1988 was the first demonstration of noninvasive EEG control of a physical object, a robot. The experiment demonstrated EEG control of multiple start-stop-restart cycles of movement, along an arbitrary trajectory defined by a line drawn on a floor. The line-following behavior was the default robot behavior, utilizing autonomous intelligence and an autonomous energy source.[12][13][14][15]
In 1990, a report was given on a closed loop, bidirectional, adaptive BCI controlling a computer buzzer by an anticipatory brain potential, the Contingent Negative Variation (CNV) potential.[16][17] The experiment described how an expectation state of the brain, manifested by CNV, used a feedback loop to control the S2 buzzer in the S1-S2-CNV paradigm. The resulting cognitive wave representing the expectation learning in the brain was termed Electroexpectogram (EXG). The CNV brain potential was part of Vidal's 1973 challenge.[citation needed]
Studies in the 2010s suggested neural stimulation's potential to restore functional connectivity and associated behaviors through modulation of molecular mechanisms.[18][19] This opened the door for the concept that BCI technologies may be able to restore function.[citation needed]
Beginning in 2013,DARPA funded BCI technology through the BRAIN initiative, which supported work out of teams includingUniversity of Pittsburgh Medical Center,[20] Paradromics,[21] Brown,[22] and Synchron.[23]
Neuroprosthetics is an area ofneuroscience concerned with neural prostheses, that is, using artificial devices to replace the function of impaired nervous systems and brain-related problems, or of sensory or other organs (bladder, diaphragm, etc.). As of December 2010,cochlear implants had been implanted as neuroprosthetic devices in some 736,900 people worldwide.[24] Other neuroprosthetic devices aim to restore vision, includingretinal implants. The first neuroprosthetic device, however, was thepacemaker.[citation needed]
The terms are sometimes used interchangeably. Neuroprosthetics and BCIs seek to achieve the same aims, such as restoring sight, hearing, movement, ability to communicate, and evencognitive function.[1] Both use similar experimental methods and surgical techniques.[citation needed]
Several laboratories have managed to read signals from monkey and ratcerebral cortices to operate BCIs to produce movement. Monkeys have movedcomputer cursors and commanded robotic arms to perform simple tasks simply by thinking about the task and seeing the results, without motor output.[25] In May 2008 photographs that showed a monkey at theUniversity of Pittsburgh Medical Center operating a robotic arm by thinking were published in multiple studies.[26] Sheep have also been used to evaluate BCI technology including Synchron's Stentrode.[citation needed]
In 2020,Elon Musk'sNeuralink was successfully implanted in a pig.[27] In 2021, Musk announced that the company had successfully enabled a monkey to play video games using Neuralink's device.[28]

In 1969operant conditioning studies by Fetz et al. at the Regional Primate Research Center and Department of Physiology and Biophysics,University of Washington School of Medicine showed that monkeys could learn to control the deflection of abiofeedback arm with neural activity.[29] Similar work in the 1970s established that monkeys could learn to control the firing rates of individual and multiple neurons in the primarymotor cortex if they were rewarded accordingly.[30]
Algorithms to reconstruct movements frommotor cortexneurons, which control movement, date back to the 1970s. In the 1980s, Georgopoulos atJohns Hopkins University found a mathematical relationship between the electrical responses of single motor cortex neurons inrhesus macaque monkeys and the direction in which they moved their arms. He also found that dispersed groups of neurons, in different areas of the monkey's brains, collectively controlled motor commands. He was able to record the firings of neurons in only one area at a time, due to equipment limitations.[31]
Several groups have been able to capture complex brain motor cortex signals by recording fromneural ensembles (groups of neurons) and using these to control external devices.[citation needed]
Phillip Kennedy (Neural Signals founder (1987) and colleagues built the first intracortical brain–computer interface by implanting neurotrophic-coneelectrodes into monkeys.[citation needed]

In 1999,Yang Dan et al. atUniversity of California, Berkeley decoded neuronal firings to reproduce images from cats. The team used an array of electrodes embedded in thethalamus (which integrates the brain's sensory input). Researchers targeted 177 brain cells in the thalamuslateral geniculate nucleus area, which decodes signals from theretina. Neuron firings were recorded from watching eight short movies. Using mathematical filters, the researchers decoded the signals to reconstruct recognizable scenes and moving objects.[32]
Duke University professorMiguel Nicolelis advocates using multiple electrodes spread over a greater area of the brain to obtain neuronal signals.[citation needed]
After initial studies in rats during the 1990s, Nicolelis and colleagues developed BCIs that decoded brain activity inowl monkeys and used the devices to reproduce monkey movements in robotic arms. Monkeys' advanced reaching and grasping abilities and hand manipulation skills, made them good test subjects.[citation needed]
By 2000, the group succeeded in building a BCI that reproduced owl monkey movements while the monkey operated ajoystick or reached for food.[33] The BCI operated in real time and could remotely control a separate robot. But the monkeys received no feedback (open-loop BCI).[citation needed]

Later experiments onrhesus monkeys includedfeedback and reproduced monkey reaching and grasping movements in a robot arm. Their deeply cleft and furrowed brains made them better models for humanneurophysiology than owl monkeys. The monkeys were trained to reach and grasp objects on a computer screen by manipulating a joystick while corresponding movements by a robot arm were hidden.[34][35] The monkeys were later shown the robot and learned to control it by viewing its movements. The BCI used velocity predictions to control reaching movements and simultaneously predictedgripping force.[citation needed]
In 2011 O'Doherty and colleagues showed a BCI with sensory feedback with rhesus monkeys. The monkey controlled the position of an avatar arm while receiving sensory feedback through directintracortical stimulation (ICMS) in the arm representation area of thesensory cortex.[36]
Other laboratories that have developed BCIs and algorithms that decode neuron signals includeJohn Donoghue at theCarney Institute for Brain Science atBrown University, Andrew Schwartz at theUniversity of Pittsburgh, andRichard Andersen atCaltech. These researchers produced working BCIs using recorded signals from far fewer neurons than Nicolelis (15–30 neurons versus 50–200 neurons).[citation needed]
The Carney Institute reported training rhesus monkeys to use a BCI to track visual targets on a computer screen (closed-loop BCI) with or without a joystick.[37] The group created a BCI for three-dimensional tracking in virtual reality and reproduced BCI control in a robotic arm.[38] The same group demonstrated that a monkey could feed itself pieces of fruit and marshmallows using a robotic arm controlled by the animal's brain signals.[39][40][41]
Andersen's group used recordings ofpremovement activity from theposterior parietal cortex, including signals created when experimental animals anticipated receiving a reward.[42]
In addition to predictingkinematic andkinetic parameters of limb movements, BCIs that predictelectromyographic or electrical activity of the muscles of primates are in process.[43] Such BCIs could restore mobility in paralyzed limbs by electrically stimulating muscles.[citation needed]
Nicolelis and colleagues demonstrated that large neural ensembles can predict arm position. This work allowed BCIs to read arm movement intentions and translate them into actuator movements. Carmena and colleagues[34] programmed a BCI that allowed a monkey to control reaching and grasping movements by a robotic arm. Lebedev and colleagues argued that brain networks reorganize to create a new representation of the robotic appendage in addition to the representation of the animal's own limbs.[35]
In 2019, researchers from the University of San Francisco, California, initiated a brain-computer interface (BCI) study that had the potential to aid patients with speech impairment resulting from neurological disorders. Their BCI utilized high-densityelectrocorticography to capture neural activity from a patient's brain and employeddeep learning to synthesize speech.[44][45] In 2021, those researchers reported the potential of a BCI to decode words and sentences in ananarthric patient who had been unable to speak for over 15 years.[46][47]
The biggest impediment to BCI technology is the lack of a sensor modality that provides safe, accurate and robust access to brain signals. The use of a better sensor expands the range of communication functions that can be provided using a BCI.[citation needed]
Development and implementation of a BCI system is complex and time-consuming. In response to this problem, Gerwin Schalk has been developingBCI2000, a general-purpose system for BCI research, since 2000.[48]
A new 'wireless' approach useslight-gated ion channels such aschannelrhodopsin to control the activity of genetically defined subsets of neuronsin vivo. In the context of a simple learning task, illumination oftransfected cells in thesomatosensory cortex influenced decision-making in mice.[49]
BCIs led to a deeper understanding of neural networks and thecentral nervous system. Research has reported that despite neuroscientists' inclination to believe that neurons have the most effect when working together, single neurons can be conditioned through the use of BCIs to fire in a pattern that allows primates to control motor outputs. BCIs led to development of the single neuron insufficiency principle that states that even with a well-tuned firing rate, single neurons can only carry limited information and therefore the highest level of accuracy is achieved by recording ensemble firings. Other principles discovered with BCIs include the neuronal multitasking principle, the neuronal mass principle, the neural degeneracy principle, and the plasticity principle.[50]
BCIs are proposed to be applied by users without disabilities. Passive BCIs allow for assessing and interpreting changes in the user state duringHuman–computer interaction (HCI). In a secondary, implicit control loop, the system adapts to its user, improving itsusability.[51]
BCI systems can potentially be used to encode signals from the periphery. These sensory BCI devices enable real-time, behaviorally-relevant decisions based upon closed-loop neural stimulation.[52]
Invasive BCI requires surgery to implant electrodes under the scalp for accessing brain signals. The main advantage is to increase accuracy. Downsides include side effects from the surgery, including scar tissue that can obstruct brain signals, or the body potentially rejecting the implanted electrodes.[53]
Invasive BCI research has targeted repairing damaged sight and providing new functionality for people with paralysis. Invasive BCIs are implanted directly into thegrey matter of the brain during neurosurgery. Because they lie in the grey matter, invasive devices produce the highest quality signals of BCI devices but are prone toscar-tissue build-up, causing the signal to weaken, or disappear, as the body reacts to the foreign object.[54]
Invision science, directbrain implants have been used to treat non-congenital (acquired) blindness. One of the first scientists to produce a working brain interface to restore sight was private researcherWilliam Dobelle. Dobelle's first prototype was implanted into "Jerry", a man blinded in adulthood, in 1978. A single-array BCI containing 68 electrodes was implanted onto Jerry'svisual cortex and succeeded in producingphosphenes, the sensation of seeing light. The system included cameras mounted on glasses to send signals to the implant. Initially, the implant allowed Jerry to see shades of grey in a limited field of vision at a low frame-rate. This also required him to be hooked up to amainframe computer, but shrinking electronics and faster computers made his artificial eye more portable and now enable him to perform simple tasks unassisted.[55]
In 2002, Jens Naumann, also blinded in adulthood, became the first in a series of 16 paying patients to receive Dobelle's second generation implant, one of the earliest commercial uses of BCIs. The second generation device used a more sophisticated implant enabling better mapping of phosphenes into coherent vision. Phosphenes are spread out across the visual field in what researchers call "the starry-night effect". Immediately after his implant, Jens was able to use his imperfectly restored vision todrive an automobile slowly around the parking area of the research institute.[56] Dobelle died in 2004 before his processes and developments were documented, leaving no one to continue his work.[57] Subsequently, Naumann and the other patients in the program began having problems with their vision, and eventually lost their "sight" again.[58][59]
BCIs focusing on motor neuroprosthetics aim to restore movement in individuals with paralysis or provide devices to assist them, such as interfaces with computers or robot arms.
Kennedy and Bakay were first to install a human brain implant that produced signals of high enough quality to simulate movement. Their patient, Johnny Ray (1944–2002), developed 'locked-in syndrome' after a brain-stemstroke in 1997. Ray's implant was installed in 1998 and he lived long enough to start working with the implant, eventually learning to control a computer cursor; he died in 2002 of abrain aneurysm.[60]
TetraplegicMatt Nagle became the first person to control an artificial hand using a BCI in 2005 as part of the first nine-month human trial ofCyberkinetics'sBrainGate chip-implant. Implanted in Nagle's rightprecentral gyrus (area of the motor cortex for arm movement), the 96-electrode implant allowed Nagle to control a robotic arm by thinking about moving his hand as well as a computer cursor, lights and TV.[61] One year later, Jonathan Wolpaw received theAltran Foundation for Innovation prize for developing a Brain Computer Interface with electrodes located on the surface of the skull, instead of directly in the brain.[62]
Research teams led by the BrainGate group and another atUniversity of Pittsburgh Medical Center, both in collaborations with theUnited States Department of Veterans Affairs (VA), demonstrated control of prosthetic limbs with many degrees of freedom using direct connections to arrays of neurons in the motor cortex of tetraplegia patients.[63][64]
In May 2021, a Stanford University team reported a successful proof-of-concept test that enabled a quadraplegic participant to produce English sentences at about 86 characters per minute and 18 words per minute. The participant imagined moving his hand to write letters, and the system performed handwriting recognition on electrical signals detected in the motor cortex, utilizingHidden Markov models andrecurrent neural networks.[65][66]Since researchers fromUCSF initiated a brain-computer interface (BCI) study, numerous reports have been made. In 2021, they reported that a paralyzed and withanarthria man was able to communicate fifteen words per minute using an implanted device that examined nerve cells controlling the muscles of the vocal tract.[67][68] In addition in 2022 it was announced that their implant could also be used to spell out words and entire sentences without speaking aloud. The first bilingual speech neuroprosthesis was reported to have been developed by the same team at the University of San Francisco, in 2024.[69][70][71] in 2025, in the beginning of the year, an article was published. The UCSF researchers reported that a man was able to control a robotic arm just by thinking.
In a review article, authors wondered whether human information transfer rates can surpass that of language with BCIs. Language research has reported that information transfer rates are relatively constant across many languages. This may reflect the brain's information processing limit. Alternatively, this limit may be intrinsic to language itself, as a modality for information transfer.[72]
In 2023 two studies used BCIs with recurrent neural network to decode speech at a record rate of 62 words per minute and 78 words per minute.[73][74][75]
There exist a number of technical challenges to recording brain activity with invasive BCIs. Advances inCMOS technology are pushing and enabling integrated, invasive BCI designs with smaller size, lower power requirements, and higher signal acquisition capabilities.[76] Invasive BCIs involve electrodes that penetrate brain tissue in an attempt to recordaction potential signals (also known as spikes) from individual, or small groups of, neurons near the electrode. The interface between a recording electrode and the electrolytic solution surrounding neurons has been modelled using theHodgkin-Huxley model.[77][78]
Electronic limitations to invasive BCIs have been an active area of research in recent decades. Whileintracellular recordings of neurons reveal action potential voltages on the scale of hundreds of millivolts, chronic invasive BCIs rely on recording extracellular voltages which typically are three orders of magnitude smaller, existing at hundreds of microvolts.[79] Further adding to the challenge of detecting signals on the scale of microvolts is the fact that the electrode-tissue interface has a highcapacitance at small voltages. Due to the nature of these small signals, for BCI systems that incorporate functionality onto an integrated circuit, each electrode requires its ownamplifier andADC, which convert analog extracellular voltages into digital signals.[79] Because a typical neuron action potential lasts for one millisecond, BCIs measuring spikes must have sampling rates ranging from 300 Hz to 5 kHz. Yet another concern is that invasive BCIs must be low-power, so as to dissipate less heat to surrounding tissue; at the most basic level more power is traditionally needed to optimizesignal-to-noise ratio.[78] Optimal battery design is an active area of research in BCIs.[80]

Challenges existing in the area ofmaterial science are central to the design of invasive BCIs. Variations in signal quality over time have been commonly observed with implantable microelectrodes.[81] Optimal material and mechanical characteristics for long term signal stability in invasive BCIs has been an active area of research.[82] It has been proposed that the formation ofglial scarring, secondary to damage at the electrode-tissue interface, is likely responsible for electrode failure and reduced recording performance.[83] Research has suggested thatblood-brain barrier leakage, either at the time of insertion or over time, may be responsible for the inflammatory and glial reaction to chronic microelectrodes implanted in the brain.[83][84] As a result, flexible[85][86][87] and tissue-like designs[88][89] have been researched and developed to minimizeforeign-body reaction by means of matching theYoung's modulus of the electrode closer to that of brain tissue.[88]
Partially invasive BCI devices are implanted inside the skull but rest outside the brain rather than within the grey matter. They produce higher resolution signals than non-invasive BCIs where the bone tissue of the cranium deflects and deforms signals and have a lower risk of forming scar-tissue in the brain than fully invasive BCIs. Preclinical demonstration of intracortical BCIs from the stroke perilesional cortex has been conducted.[90]
A systematic review published in 2020 detailed multiple clinical and non-clinical studies investigating the feasibility of endovascular BCIs.[91]
In 2010, researchers affiliated with University of Melbourne began developing a BCI that could be inserted via the vascular system. Australian neurologistThomas Oxley conceived the idea for this BCI, called Stentrode, earning funding fromDARPA. Preclinical studies evaluated the technology in sheep.[2]
Stentrode is a monolithicstent electrode array designed to be delivered via an intravenous catheter under image-guidance to thesuperior sagittal sinus, in the region which lies adjacent to themotor cortex.[92] This proximity enables Stentrode to measure neural activity. The procedure is most similar to how venous sinus stents are placed for the treatment ofidiopathic intracranial hypertension.[93] Stentrode communicates neural activity to a battery-less telemetry unit implanted in the chest, which communicates wirelessly with an external telemetry unit capable of power and data transfer. While an endovascular BCI benefits from avoiding acraniotomy for insertion, risks such asclotting andvenous thrombosis exist.[citation needed]
Human trials with Stentrode were underway as of 2021.[92] In November 2020, two participants withamyotrophic lateral sclerosis were able to wirelessly control an operating system to text, email, shop, and bank using direct thought using Stentrode,[94] marking the first time a brain-computer interface was implanted via the patient's blood vessels, eliminating the need for brain surgery. In January 2023, researchers reported no serious adverse events during the first year for all four patients, who could use it to operate computers.[95][96]
Electrocorticography (ECoG) measures brain electrical activity from beneath the skull in a way similar to non-invasive electroencephalography, using electrodes embedded in a thin plastic pad placed above the cortex, beneath thedura mater.[97] ECoG technologies were first trialled in humans in 2004 by Eric Leuthardt and Daniel Moran fromWashington University in St. Louis. In a later trial, the researchers enabled a teenage boy to playSpace Invaders.[98] This research indicates that control is rapid, requires minimal training, balancing signal fidelity and level of invasiveness.[note 1]
Signals can be either subdural or epidural, but are not taken from within the brainparenchyma. Patients are required to have invasive monitoring for localization and resection of an epileptogenic focus.[citation needed]
ECoG offers higher spatial resolution, better signal-to-noise ratio, wider frequency range, and less training requirements than scalp-recorded EEG, and at the same time has lower technical difficulty, lower clinical risk, and may have superior long-term stability than intracortical single-neuron recording.[100] This feature profile and evidence of the high level of control with minimal training requirements shows potential for real world application for people with motor disabilities.[101][102]
Edward Chang and Joseph Makin fromUCSF reported that ECoG signals could be used to decode speech from epilepsy patients implanted with high-density ECoG arrays over the peri-Sylvian cortices.[103][104] They reported word error rates of 3% (a marked improvement from prior efforts) utilizing an encoder-decoderneural network, which translated ECoG data into one of fifty sentences composed of 250 unique words.[citation needed]
In 2014, a BCI usingfunctional near-infrared spectroscopy for "locked-in" patients withamyotrophic lateral sclerosis (ALS) was able to restore basic ability to communicate.[105]

After Vidal stated the BCI challenge, the initial reports on non-invasive approaches included control of a cursor in 2D using VEP,[106] control of a buzzer using CNV,[107] control of a physical object, a robot, using a brain rhythm (alpha),[108] control of a text written on a screen using P300.[109][10]
In the early days of BCI research, another substantial barrier to using EEG was that extensive training was required. For example, in experiments beginning in the mid-1990s, Niels Birbaumer at theUniversity of Tübingen inGermany trained paralysed people to self-regulate the slow cortical potentials in their EEG to such an extent that these signals could be used as a binary signal to control a computer cursor. (Birbaumer had earlier trainedepileptics to prevent impending fits by controlling this low voltage wave.) The experiment trained ten patients to move a computer cursor. The process was slow, requiring more than an hour for patients to write 100 characters with the cursor, while training often took months. The slow cortical potential approach has fallen away in favor of approaches that require little or no training, are faster and more accurate, and work for a greater proportion of users.[110]
Another research parameter is the type ofoscillatory activity that is measured. Gert Pfurtscheller founded the BCI Lab 1991 and conducted the first online BCI based on oscillatory features and classifiers. Together with Birbaumer and Jonathan Wolpaw atNew York State University they focused on developing technology that would allow users to choose the brain signals they found easiest to operate a BCI, includingmu andbeta rhythms.[citation needed]
A further parameter is the method of feedback used as shown in studies ofP300 signals. Patterns of P300 waves are generated involuntarily (stimulus-feedback) when people see something they recognize and may allow BCIs to decode categories of thoughts without training.[citation needed]
A 2005 study reported EEG emulation of digital control circuits, using a CNV flip-flop.[111] A 2009 study reported noninvasive EEG control of a robotic arm using a CNV flip-flop.[112] A 2011 study reported control of two robotic arms solvingTower of Hanoi task with three disks using a CNV flip-flop.[113] A 2015 study described EEG-emulation of aSchmitt trigger, flip-flop,demultiplexer, andmodem.[114]
Advances byBin He and his team atUniversity of Minnesota suggest the potential of EEG-based brain-computer interfaces to accomplish tasks close to invasive brain-computer interfaces. Using advanced functional neuroimaging including BOLD functionalMRI andEEG source imaging, They identified the co-variation and co-localization ofelectrophysiological andhemodynamic signals.[115] Refined by a neuroimaging approach and a training protocol, They fashioned a non-invasive EEG based brain-computer interface to control the flight of a virtual helicopter in 3-dimensional space, based upon motor imagination.[116] In June 2013 they announced a technique to guide a remote-control helicopter through an obstacle course.[117] They also solved the EEGinverse problem and then used the resulting virtual EEG for BCI tasks. Well-controlled studies suggested the merits of such a source analysis-based BCI.[118]
A 2014 study reported that severely motor-impaired patients could communicate faster and more reliably with non-invasive EEG BCI than with muscle-based communication channels.[119]
A 2019 study reported that the application of evolutionary algorithms could improve EEG mental state classification with a non-invasiveMuse device, enabling classification of data acquired by a consumer-grade sensing device.[120]
In a 2021 systematic review ofrandomized controlled trials using BCI for post-stroke upper-limb rehabilitation, EEG-based BCI was reported to have efficacy in improving upper-limb motor function compared to control therapies. More specifically, BCI studies that utilized band power features,motor imagery, andfunctional electrical stimulation were reported to be more effective than alternatives.[121] Another 2021 systematic review focused on post-stroke robot-assisted EEG-based BCI for hand rehabilitation. Improvement in motor assessment scores was observed in three of eleven studies.[122]
In the early 1990s Babak Taheri, atUniversity of California, Davis demonstrated the first single and multichannel dry active electrode arrays.[123] The arrayed electrode was demonstrated to perform well compared tosilver/silver chloride electrodes. The device consisted of four sensor sites with integrated electronics to reduce noise byimpedance matching. The advantages of such electrodes are:[citation needed]
The active electrode array is an integrated system containing an array of capacitive sensors with local integrated circuitry packaged with batteries to power the circuitry. This level of integration was required to achieve the result.[citation needed]
The electrode was tested on a test bench and on human subjects in four modalities, namely:[citation needed]
Performance compared favorably with that of standard wet electrodes in terms of skin preparation, no gel requirements (dry), and higher signal-to-noise ratio.[124]
In 1999 Hunter Peckham and others atCase Western Reserve University used a 64-electrode EEG skullcap to return limited hand movements to aquadriplegic. As he concentrated on simple but opposite concepts like up and down. A basic pattern was identified in his beta-rhythm EEG output and used to control a switch: Above average activity was interpreted as on, below average off. The signals were also used to drive nerve controllers embedded in his hands, restoring some movement.[125]
In 2009, the NCTU Brain-Computer-Interface-headband was announced. Those researchers also engineered silicon-basedmicroelectro-mechanical system (MEMS)dry electrodes designed for application to non-hairy body sites. These electrodes were secured to the headband'sDAQ board with snap-on electrode holders. The signal processing module measuredalpha activity and transferred it overBluetooth to a phone that assessed the patients' alertness and cognitive capacity. When the subject became drowsy, the phone sent arousing feedback to the operator to rouse them.[126]
In 2011, researchers reported a cellular based BCI that could cause a phone to ring. The wearable system was composed of a four channel bio-signal acquisition/amplificationmodule, a communication module, and a Bluetooth phone. The electrodes were placed to pick up steady state visual evoked potentials (SSVEPs).[127] SSVEPs are electrical responses to flickering visual stimuli with repetition rates over 6 Hz[127] that are best found in the parietal and occipital scalp regions of the visual cortex.[128][129][130] It was reported that all study participants were able to initiate the phone call with minimal practice in natural environments.[131]
The scientists reported that a single channelfast Fourier transform (FFT) and multiple channel systemcanonical correlation analysis (CCA) algorithm can support mobile BCIs.[127][132] The CCA algorithm has been applied in experiments investigating BCIs with claimed high accuracy and speed.[133] Cellular BCI technology can reportedly be translated for other applications, such as picking up sensorimotormu/beta rhythms to function as a motor-imagery based BCI.[127]
In 2013, comparative tests performed onAndroid cell phone, tablet, and computer based BCIs, analyzed the powerspectrum density of resultant EEG SSVEPs. The stated goals of this study were to "increase the practicability, portability, and ubiquity of an SSVEP-based BCI, for daily use". It was reported that the stimulation frequency on all mediums was accurate, although the phone's signal was not stable. The amplitudes of the SSVEPs for the laptop and tablet were reported to be larger than those of the cell phone. These two qualitative characterizations were suggested as indicators of the feasibility of using a mobile stimulus BCI.[132]
One of the difficulties with EEG readings is susceptibility to motion artifacts.[134] In most research projects, the participants were asked to sit still in a laboratory setting, reducing head and eye movements as much as possible. However, since these initiatives were intended to create a mobile device for daily use,[132] the technology had to be tested in motion. In 2013, researchers tested mobile EEG-based BCI technology, measuring SSVEPs from participants as they walked on a treadmill. Reported results were that as speed increased, SSVEP detectability using CCA decreased.Independent component analysis (ICA) had been shown to be efficient in separating EEG signals from noise.[135] The researchers stated that CCA data with and without ICA processing were similar. They concluded that CCA demonstrated robustness to motion artifacts.[129] EEG-based BCI applications offer low spatial resolution. Possible solutions include: EEG source connectivity based ongraph theory, EEG pattern recognition based on Topomap and EEG-fMRI fusion.[citation needed]
Non-invasive BCIs have been applied to prosthetic upper and lower extremity devices in people with paralysis. For example, Gert Pfurtscheller ofGraz University of Technology and colleagues demonstrated a BCI-controlledfunctional electrical stimulation system to restore upper extremity movements in a person with tetraplegia due tospinal cord injury.[136] Between 2012 and 2013, researchers atUniversity of California, Irvine demonstrated for the first time that BCI technology can restore brain-controlled walking afterspinal cord injury. In theirstudy, a person withparaplegia operated a BCI-robotic gaitorthosis to regain basic ambulation.[137][138] In 2009 independent researcher Alex Blainey used theEmotiv EPOC to control a 5 axis robot arm.[139] He made several demonstrations of mind controlled wheelchairs andhome automation.[citation needed]

Magnetoencephalography (MEG) andfunctional magnetic resonance imaging (fMRI) have both been used as non-invasive BCIs.[140] In a widely reported experiment, fMRI allowed two users to playPong in real-time by altering theirhaemodynamic response or brain blood flow throughbiofeedback.[141]
fMRI measurements of haemodynamic responses in real time have also been used to control robot arms with a seven-second delay between thought and movement.[142]
In 2008 research developed in the Advanced Telecommunications Research (ATR)Computational Neuroscience Laboratories inKyoto, Japan, allowed researchers to reconstruct images from brain signals at aresolution of 10x10pixels.[143]
A 2011 study reported second-by-second reconstruction of videos watched by the study's subjects, from fMRI data.[144] This was achieved by creating a statistical model relating videos to brain activity. This model was then used to look up 100 one-second video segments, in a database of 18 million seconds of randomYouTube videos, matching visual patterns to brain activity recorded when subjects watched a video. These 100 one-second video extracts were then combined into a mash-up image that resembled the video.[145][146][147]
Motor imagery involves imagining the movement of body parts, activating thesensorimotor cortex, which modulates sensorimotor oscillations in the EEG. This can be detected by the BCI and used to infer user intent. Motor imagery typically requires training to acquire acceptable control. Training sessions typically consume hours over several days. Regardless of the duration of the training session, users are unable to master the control scheme. This results in very slow pace of the gameplay.[148] Machine learning methods were used to compute a subject-specific model for detecting motor imagery performance. The top performing algorithm from BCI Competition IV in 2022[149] dataset 2 for motor imagery was the Filter Bank Common Spatial Pattern, developed by Ang et al. fromA*STAR,Singapore.[150]
Biofeedback can be used to monitor a subject's mental relaxation. In some cases, biofeedback does not match EEG, while parameters such aselectromyography (EMG),galvanic skin resistance (GSR), andheart rate variability (HRV) can do so. Many biofeedback systems treat disorders such asattention deficit hyperactivity disorder (ADHD), sleep problems in children, teeth grinding, and chronic pain. EEG biofeedback systems typically monitor four brainwave bands (theta: 4–7 Hz, alpha:8–12 Hz, SMR: 12–15 Hz, beta: 15–18 Hz) and challenge the subject to control them. Passive BCI uses BCI to enrich human–machine interaction with information on the user's mental state, for example, simulations that detect when users intend to push brakes during emergency vehicle braking.[51] Game developers using passive BCIs understand that through repetition of game levels the user's cognitive state adapts. During the first play of a given level, the player reacts differently than during subsequent plays: for example, the user is less surprised by an event that they expect.[148]
A VEP is an electrical potential recorded after a subject is presented with a visual stimuli. The types of VEPs include SSVEPs and P300 potential.[citation needed]
Steady-state visually evoked potentials (SSVEPs) use potentials generated by exciting theretina, using visual stimuli modulated at certain frequencies. SSVEP stimuli are often formed from alternating checkerboard patterns and at times use flashing images. The frequency of the phase reversal of the stimulus used can be distinguished by EEG; this makes detection of SSVEP stimuli relatively easy. SSVEP is used within many BCI systems. This is due to several factors. The signal elicited is measurable in as large a population as the transient VEP and blink movement. Electrocardiographic artefacts do not affect the frequencies monitored. The SSVEP signal is robust; the topographic organization of the primary visual cortex is such that a broader area obtains afferents from the visual field's central or fovial region. SSVEP comes with problems. As SSVEPs use flashing stimuli to infer user intent, the user must gaze at one of the flashing or iterating symbols in order to interact with the system. It is, therefore, likely that the symbols become irritating and uncomfortable during longer play sessions.[citation needed]
Another type of VEP is theP300 potential. This potential is a positive peak in the EEG that occurs roughly 300 ms after the appearance of a target stimulus (a stimulus for which the user is waiting or seeking) oroddball stimuli. P300 amplitude decreases as the target stimuli and the ignored stimuli grow more similar. P300 is thought to be related to a higher level attention process or an orienting response. Using P300 requires fewer training sessions. The first application to use it was the P300 matrix. Within this system, a subject chooses a letter from a 6 by 6 grid of letters and numbers. The rows and columns of the grid flashed sequentially and every time the selected "choice letter" was illuminated the user's P300 was (potentially) elicited. However, the communication process, at approximately 17 characters per minute, was slow. P300 offers a discrete selection rather than continuous control. The advantage of P300 within games is that the player does not have to learn how to use a new control system, requiring only short training instances to learn gameplay mechanics and the basic BCI paradigm.[148]
Human-computer interaction can exploit other recording modalities, such aselectrooculography and eye-tracking. These modalities do not record brain activity and therefore do not qualify as BCIs.[151]
In 1989, a study reported control of a mobile robot by eye movement using electrooculography signals. A mobile robot was driven to a goal point using five EOG commands, interpreted as forward, backward, left, right, and stop.[152]
A 2016 article described a new non-EEG-based HCI that required novisual fixation, or ability to move the eyes.[153] The interface is based on covertinterest; directing attention to a chosen letter on a virtual keyboard, without the need to look directly at the letter. Each letter has its own (background) circle which micro-oscillates in brightness differently from the others. Letter selection is based on best fit between unintentional pupil-size oscillation and the background circle's brightness oscillation pattern. Accuracy is additionally improved by the user's mental rehearsal of the words 'bright' and 'dark' in synchrony with the brightness transitions of the letter's circle.[citation needed]
In the 1960s a researcher after training used EEG to createMorse code using alpha waves.[154] On 27 February 2013Miguel Nicolelis's group atDuke University and IINN-ELS connected the brains of two rats, allowing them to share information, inthe first-ever direct brain-to-brain interface.[155][156][157]
Gerwin Schalk reported that ECoG signals can discriminate vowels and consonants embedded in spoken and imagined words, shedding light on the mechanisms associated with their production and could provide a basis for brain-based communication using imagined speech.[102][158]
In 2002Kevin Warwick had an array of 100 electrodes fired into his nervous system in order to link his nervous system to the Internet. Warwick carried out a series of experiments. Electrodes were implanted into his wife's nervous system, allowing them to conduct the first direct electronic communication experiment between the nervous systems of two humans.[159][160][161][162]
Other researchers achieved brain-to-brain communication between participants at a distance using non-invasive technology attached to the participants' scalps. The words were encoded in binary streams by the cognitive motor input of the person sending the information. Pseudo-random bits of the information carried encoded words "hola" ("hi" in Spanish) and "ciao" ("goodbye" in Italian) and were transmitted mind-to-mind.[163]

Researchers have built devices to interface with neural cells and entire neural networksin vitro. Experiments on cultured neural tissue focused on building problem-solving networks, constructing basic computers and manipulating robotic devices. Research into techniques for stimulating and recording individual neurons grown on semiconductor chips is neuroelectronics orneurochips.[164]
Development of the first neurochip was claimed by a Caltech team led by Jerome Pine and Michael Maher in 1997.[165] The Caltech chip had room for 16 neurons.[citation needed]
In 2003 a team led by Theodore Berger, at theUniversity of Southern California, worked on a neurochip designed to function as an artificial or prosthetichippocampus. The neurochip was designed for rat brains. The hippocampus was chosen because it is thought to be the most structured and most studied part of the brain. Its function is to encode experiences for storage as long-term memories elsewhere in the brain.[166]
In 2004 Thomas DeMarse at theUniversity of Florida used a culture of 25,000 neurons taken from a rat's brain to fly aF-22 fighter jetaircraft simulator. After collection, the cortical neurons were cultured in apetri dish and reconnected themselves to form a living neural network. The cells were arranged over a grid of 60 electrodes and used to control thepitch andyaw functions of the simulator. The study's focus was on understanding how the human brain performs and learns computational tasks at a cellular level.[167]
Concerns center on the safety and long-term effects on users. These include obtaininginformed consent from individuals with communication difficulties, the impact on patients' and families' quality of life, health-related side effects, misuse of therapeutic applications, safety risks, and the non-reversible nature of some BCI-induced changes. Additionally, questions arise about access to maintenance, repair, and spare parts, particularly in the event of a company's bankruptcy.[168]
The legal and social aspects of BCIs complicate mainstream adoption. Concerns include issues of accountability and responsibility, such as claims that BCI influence overrides free will and control over actions, inaccurate translation of cognitive intentions, personality changes resulting from deep-brain stimulation, and the blurring of the line between human and machine.[169] Other concerns involve the use of BCIs in advanced interrogation techniques, unauthorized access ("brain hacking"),[170] social stratification through selective enhancement, privacy issues related to mind-reading, tracking and "tagging" systems, and the potential for mind, movement, and emotion control.[171]
In their current form, most BCIs are more akin to corrective therapies that engage few of such ethical issues. Bioethics is well-equipped to address the challenges posed by BCI technologies, with Clausen suggesting in 2009 that "BCIs pose ethical challenges, but these are conceptually similar to those that bioethicists have addressed for other realms of therapy."[172] Haselager and colleagues highlighted the importance of managing expectations and value.[173]
The evolution of BCIs mirrors that of pharmaceutical science, which began as a means to address impairments and now enhances focus and reduces the need for sleep. As BCIs progress from therapies to enhancements, the BCI community is working to create consensus on ethical guidelines for research, development, and dissemination.[174][175]
Various companies are developing inexpensive BCIs for research and entertainment. Toys such as the NeuroSky and Mattel MindFlex have seen some commercial success.

A consortium of 12 European partners completed a roadmap to support the European Commission in their funding decisions for theHorizon 2020 framework program. The project was funded by the European Commission. It started in November 2013 and published a roadmap in April 2015.[192] A 2015 publication describes this project, as well as the Brain-Computer Interface Society.[193] It reviewed work within this project that further defined BCIs and applications, explored recent trends, discussed ethical issues, and evaluated directions for new BCIs.[citation needed]
Other recent publications too have explored future BCI directions for new groups of disabled users.[7][194]
Some people have adisorder of consciousness (DOC). This state is defined to include people in a coma and those in avegetative state (VS) orminimally conscious state (MCS). BCI research seeks to address DOC. A key initial goal is to identify patients who can perform basic cognitive tasks, which would change their diagnosis, and allow them to make important decisions (such as whether to seek therapy, where to live, and their views on end-of-life decisions regarding them). Patients incorrectly diagnosed may die as a result of end-of-life decisions made by others. The prospect of using BCI to communicate with such patients is a tantalizing prospect.[195][196]
Many such patients cannot use BCIs based on vision. Hence, tools must rely on auditory and/or vibrotactile stimuli. Patients may wear headphones and/or vibrotactile stimulators placed on responsive body parts. Another challenge is that patients may be able to communicate only at unpredictable intervals. Home devices can allow communications when the patient is ready.[citation needed]
Automated tools can ask questions that patients can easily answer, such as "Is your father named George?" or "Were you born in the USA?" Automated instructions inform patients how to convey yes or no, for example by focusing their attention on stimuli on the right vs. left wrist. This focused attention produces reliable changes inEEG patterns that can help determine whether the patient is able to communicate.[197][198][199]
People may lose some of their ability to move due to many causes, such as stroke or injury. Research in recent years has demonstrated the utility of EEG-based BCI systems in aiding motor recovery and neurorehabilitation in patients who have had a stroke.[200][201][202][203] Several groups have explored systems and methods for motor recovery that include BCIs.[204][205][206][207] In this approach, a BCI measures motor activity while the patient imagines or attempts movements as directed by a therapist. The BCI may provide two benefits: (1) if the BCI indicates that a patient is not imagining a movement correctly (non-compliance), then the BCI could inform the patient and therapist; and (2) rewarding feedback such as functional stimulation or the movement of a virtual avatar also depends on the patient's correct movement imagery.[citation needed]
So far, BCIs for motor recovery have relied on the EEG to measure the patient's motor imagery. However, studies have also used fMRI to study different changes in the brain as persons undergo BCI-based stroke rehab training.[208][209][210] Imaging studies combined with EEG-based BCI systems hold promise for investigating neuroplasticity during motor recovery post-stroke.[210] Future systems might include the fMRI and other measures for real-time control, such as functional near-infrared, probably in tandem with EEGs. Non-invasive brain stimulation has also been explored in combination with BCIs for motor recovery.[211] In 2016, scientists out of theUniversity of Melbourne published preclinical proof-of-concept data related to a potential brain-computer interface technology platform being developed for patients with paralysis to facilitate control of external devices such as robotic limbs, computers and exoskeletons by translating brain activity.[212][213][214]
In 2014, some 400,000 people underwentbrain mapping during neurosurgery. This procedure is often required for people who do not respond tomedication.[215] During this procedure, electrodes are placed on the brain to precisely identify the locations of structures and functional areas. Patients may be awake during neurosurgery and asked to perform tasks, such as moving fingers or repeating words. This is necessary so that surgeons can remove the desired tissue while sparing other regions. Removing too much brain tissue can cause permanent damage, while removing too little can mandate additional neurosurgery.[citation needed]
Researchers explored ways to improve neurosurgical mapping. This work focuses largely on high gamma activity, which is difficult to detect non-invasively. Results improved methods for identifying key functional areas.[216]
Flexible electronics arepolymers or other flexible materials (e.g.silk,[217]pentacene,PDMS,Parylene,polyimide[218]) printed withcircuitry; the flexibility allows the electronics to bend. Thefabrication techniques used to create these devices resembles those used to createintegrated circuits andmicroelectromechanical systems (MEMS).[citation needed]
Flexible neural interfaces may minimize brain tissue trauma related to mechanical mismatch between electrode and tissue.[219]
Neural dust is millimeter-sized devices operated aswirelessly powered nerve sensors that were proposed in a 2011 paper from theUniversity of California, Berkeley Wireless Research Center.[220][221] In one model,local field potentials could be distinguished fromaction potential "spikes", which would offer greatly diversified data vs conventional techniques.[220]
ECoG- Based BCI has advantage in signal and durability that are absolutely necessary for clinical application
Justin Williams, a biomedical engineer at the university, has already transformed the ECoG implant into a micro device that can be installed with a minimum of fuss. It has been tested in animals for a long period of time – the micro ECoG stays in place and doesn't seem to negatively affect the immune system.