Academic Journal
Hidden Brain State-Based Internal Evaluation Using Kernel Inverse Reinforcement Learning in Brain-Machine Interfaces
العنوان: | Hidden Brain State-Based Internal Evaluation Using Kernel Inverse Reinforcement Learning in Brain-Machine Interfaces |
---|---|
المؤلفون: | Jieyuan Tan, Xiang Zhang, Shenghui Wu, Zhiwei Song, Yiwen Wang |
المصدر: | IEEE Transactions on Neural Systems and Rehabilitation Engineering, Vol 32, Pp 4219-4229 (2024) |
بيانات النشر: | IEEE, 2024. |
سنة النشر: | 2024 |
المجموعة: | LCC:Medical technology LCC:Therapeutics. Pharmacology |
مصطلحات موضوعية: | Brain-machine interface (BMI), inverse reinforcement learning, medial prefrontal cortex, internal evaluation, Medical technology, R855-855.5, Therapeutics. Pharmacology, RM1-950 |
الوصف: | Reinforcement learning (RL)-based brain machine interfaces (BMIs) assist paralyzed people in controlling neural prostheses without the need for real limb movement as supervised signals. The design of reward signal significantly impacts the learning efficiency of the RL-based decoders. Existing reward designs in the RL-based BMI framework rely on external rewards or manually labeled internal rewards, unable to accurately extract subjects’ internal evaluation. In this paper, we propose a hidden brain state-based kernel inverse reinforcement learning (HBS-KIRL) method to accurately infer the subject-specific internal evaluation from neural activity during the BMI task. The state-space model is applied to project the neural state into low-dimensional hidden brain state space, which greatly reduces the exploration dimension. Then the kernel method is applied to speed up the convergence of policy, reward, and Q-value networks in reproducing kernel Hilbert space (RKHS). We tested our proposed algorithm on the data collected from the medial prefrontal cortex (mPFC) of rats when they were performing a two-lever-discrimination task. We assessed the state-value estimation performance of our proposed method and compared it with naïve IRL and PCA-based IRL. To validate that the extracted internal evaluation could contribute to the decoder training, we compared the decoding performance of decoders trained by different reward models, including manually designed reward, naïve IRL, PCA-IRL, and our proposed HBS-KIRL. The results show that the HBS-KIRL method can give a stable and accurate estimation of state-value distribution with respect to behavior. Compared with other methods, the decoder guided by HBS-KIRL achieves consistent and better decoding performance over days. This study reveals the potential of applying the IRL method to better extract subject-specific evaluation and improve the BMI decoding performance. |
نوع الوثيقة: | article |
وصف الملف: | electronic resource |
اللغة: | English |
تدمد: | 1534-4320 1558-0210 |
Relation: | https://ieeexplore.ieee.org/document/10759843/; https://doaj.org/toc/1534-4320; https://doaj.org/toc/1558-0210 |
DOI: | 10.1109/TNSRE.2024.3503713 |
URL الوصول: | https://doaj.org/article/3e49e6bbe0544e09aec320491c34a027 |
رقم الانضمام: | edsdoj.3e49e6bbe0544e09aec320491c34a027 |
قاعدة البيانات: | Directory of Open Access Journals |
تدمد: | 15344320 15580210 |
---|---|
DOI: | 10.1109/TNSRE.2024.3503713 |