The Emotiv™ electroencephalogram (EEG) headset is a non-invasive brain-computer interface with 14 EEG channels that was used to record neural responses to stimuli in 5 male and 5 female participants. The stimuli included instructions (presented in both visual or auditory format successively) of 'right' or 'left.' Participants raised their right or left arm in response to the instruction given. The subjects were stimulated and their data was recorded using a scenario generated in OpenViBE™. Specifically, the Graz Motor BCI Stimulator algorithm was used to generate the number and frequency of auditory and visual stimulations. In addition, using EEGlab™ in conjunction with Matlab®, the electrodes most important during visual and auditory stimulation were identified. This aided in the development of processing real-time EEG signals from the Emotiv™ EEG headset. Filtering and feature extraction were accomplished via the SIMULINK® signal processing toolbox. An Arduino™ microcontroller acted as a connection between the Mecha Te Hand™ and the Emotiv™ EEG headset.