Auditory Evoked Potential (AEP) Based Brain-Computer Interface (BCI) Technology: A Short Review

Author(s):  
Md Nahidul Islam ◽  
Norizam Sulaiman ◽  
Bifta Sama Bari ◽  
Mamunur Rashid ◽  
Mahfuzah Mustafa
2021 ◽  
pp. 1-13
Author(s):  
Hamidreza Maymandi ◽  
Jorge Luis Perez Benitez ◽  
F. Gallegos-Funes ◽  
J. A. Perez Benitez

Author(s):  
Yao Li ◽  
T. Kesavadas

Abstract One of the expectations for the next generation of industrial robots is to work collaboratively with humans as robotic co-workers. Robotic co-workers must be able to communicate with human collaborators intelligently and seamlessly. However, industrial robots in prevalence are not good at understanding human intentions and decisions. We demonstrate a steady-state visual evoked potential (SSVEP)-based brain-computer interface (BCI) which can directly deliver human cognition to robots through a headset. The BCI is applied to a part-picking robot and sends decisions to the robot while operators visually inspecting the quality of parts. The BCI is verified through a human subject study. In the study, a camera by the side of the conveyor takes photos of each part and presents it to the operator automatically. When the operator looks at the photo, the electroencephalography (EEG) is collected through BCI. The inspection decision is extracted through SSVEPs in EEG. When a defective part is identified by the operator, the signal is communicated to the robot which locates the defective part through a second camera and removes it from the conveyor. The robot can grasp various part with our grasp planning algorithm (2FRG). We have developed a CNN-CCA model for SSVEP extraction. The model is trained on a dataset collected in our offline experiment. Our approach outperforms the existing CCA, CCA-SVM, and PSD-SVM models. The CNN-CCA is further validated in an online experiment that achieves 93% accuracy in identifying and removing a defective part.


Sign in / Sign up

Export Citation Format

Share Document