TY - GEN
T1 - Brain-machine interface control of a robot arm using actor-critic rainforcement learning
AU - Pohlmeyer, Eric A.
AU - Mahmoudi, Babak
AU - Geng, Shijia
AU - Prins, Noeline
AU - Sanchez, Justin C.
PY - 2012
Y1 - 2012
N2 - Here we demonstrate how a marmoset monkey can use a reinforcement learning (RL) Brain-Machine Interface (BMI) to effectively control the movements of a robot arm for a reaching task. In this work, an actor-critic RL algorithm used neural ensemble activity in the monkey's motor cortext to control the robot movements during a two-target decision task. This novel approach to decoding offers unique advantages for BMI control applications. Compared to supervised learning decoding methods, the actor-critic RL algorithm does not require an explicit set of training data to create a static control model, but rather it incrementally adapts the model parameters according to its current performance, in this case requiring only a very basic feedback signal. We show how this algorithm achieved high performance when mapping the monkey's neural states (94%) to robot actions, and only needed to experience a few trials before obtaining accurate real-time control of the robot arm. Since RL methods responsively adapt and adjust their parameters, they can provide a method to create BMIs that are robust against perturbations caused by changes in either the neural input space or the output actions they generate under different task requirements or goals.
AB - Here we demonstrate how a marmoset monkey can use a reinforcement learning (RL) Brain-Machine Interface (BMI) to effectively control the movements of a robot arm for a reaching task. In this work, an actor-critic RL algorithm used neural ensemble activity in the monkey's motor cortext to control the robot movements during a two-target decision task. This novel approach to decoding offers unique advantages for BMI control applications. Compared to supervised learning decoding methods, the actor-critic RL algorithm does not require an explicit set of training data to create a static control model, but rather it incrementally adapts the model parameters according to its current performance, in this case requiring only a very basic feedback signal. We show how this algorithm achieved high performance when mapping the monkey's neural states (94%) to robot actions, and only needed to experience a few trials before obtaining accurate real-time control of the robot arm. Since RL methods responsively adapt and adjust their parameters, they can provide a method to create BMIs that are robust against perturbations caused by changes in either the neural input space or the output actions they generate under different task requirements or goals.
UR - http://www.scopus.com/inward/record.url?scp=84880955826&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=84880955826&partnerID=8YFLogxK
U2 - 10.1109/EMBC.2012.6346870
DO - 10.1109/EMBC.2012.6346870
M3 - Conference contribution
C2 - 23366831
AN - SCOPUS:84880955826
SN - 9781424441198
T3 - Proceedings of the Annual International Conference of the IEEE Engineering in Medicine and Biology Society, EMBS
SP - 4108
EP - 4111
BT - 2012 Annual International Conference of the IEEE Engineering in Medicine and Biology Society, EMBC 2012
T2 - 34th Annual International Conference of the IEEE Engineering in Medicine and Biology Society, EMBS 2012
Y2 - 28 August 2012 through 1 September 2012
ER -