Reinforcement Learning from Implicit Neural Feedback for Human-Aligned Robot Control

Reading time: 5 minute
...

๐Ÿ“ Original Info

  • Title: Reinforcement Learning from Implicit Neural Feedback for Human-Aligned Robot Control
  • ArXiv ID: 2512.00050
  • Date: 2025-11-18
  • Authors: Suzie Kim

๐Ÿ“ Abstract

Conventional reinforcement learning (RL) approaches often struggle to learn effective policies under sparse reward conditions, necessitating the manual design of complex, task-specific reward functions. To address this limitation, reinforcement learning from human feedback (RLHF) has emerged as a promising strategy that complements hand-crafted rewards with human-derived evaluation signals. However, most existing RLHF methods depend on explicit feedback mechanisms such as button presses or preference labels, which disrupt the natural interaction process and impose a substantial cognitive load on the user. We propose a novel reinforcement learning from implicit human feedback (RLIHF) framework that utilizes non-invasive electroencephalography (EEG) signals, specifically error-related potentials (ErrPs), to provide continuous, implicit feedback without requiring explicit user intervention. The proposed method adopts a pre-trained decoder to transform raw EEG signals into probabilistic reward components, enabling effective policy learning even in the presence of sparse external rewards. We evaluate our approach in a simulation environment built on the MuJoCo physics engine, using a Kinova Gen2 robotic arm to perform a complex pick-and-place task that requires avoiding obstacles while manipulating target objects. The results show that agents trained with decoded EEG feedback achieve performance comparable to those trained with dense, manually designed rewards. These findings validate the potential of using implicit neural feedback for scalable and human-aligned reinforcement learning in interactive robotics.

๐Ÿ’ก Deep Analysis

Deep Dive into Reinforcement Learning from Implicit Neural Feedback for Human-Aligned Robot Control.

Conventional reinforcement learning (RL) approaches often struggle to learn effective policies under sparse reward conditions, necessitating the manual design of complex, task-specific reward functions. To address this limitation, reinforcement learning from human feedback (RLHF) has emerged as a promising strategy that complements hand-crafted rewards with human-derived evaluation signals. However, most existing RLHF methods depend on explicit feedback mechanisms such as button presses or preference labels, which disrupt the natural interaction process and impose a substantial cognitive load on the user. We propose a novel reinforcement learning from implicit human feedback (RLIHF) framework that utilizes non-invasive electroencephalography (EEG) signals, specifically error-related potentials (ErrPs), to provide continuous, implicit feedback without requiring explicit user intervention. The proposed method adopts a pre-trained decoder to transform raw EEG signals into probabilistic re

๐Ÿ“„ Full Content

Masterโ€™s Thesis Reinforcement Learning from Implicit Neural Feedback for Human-Aligned Robot Control Suzie Kim Department of Artificial Intelligence Graduate School Korea University February 2025 arXiv:2512.00050v1 [cs.RO] 18 Nov 2025 Reinforcement Learning from Implicit Neural Feedback for Human-Aligned Robot Control by Suzie Kim under the supervision of Professor Seong-Whan Lee A thesis submitted in partial fulfillment of the requirements for the degree of Master of Science Department of Artificial Intelligence Graduate School Korea University October 2025 The thesis of Suzie Kim has been approved by the thesis committee in partial fulfillment of the requirements for the degree of Master of Science December 2025 Committee Chair: Seong-Whan Lee Committee Member: Won-Zoo Chung Committee Member: Tae-Eui Kam Reinforcement Learning from Implicit Neural Feedback for Human-Aligned Robot Control by Suzie Kim Department of Artificial Intelligence under the supervision of Professor Seong-Whan Lee Abstract Conventional reinforcement learning (RL) approaches often struggle to learn effective policies under sparse reward conditions, necessitating the man- ual design of complex, task-specific reward functions. To address this limi- tation, reinforcement learning from human feedback (RLHF) has emerged as a promising strategy that complements hand-crafted rewards with human- derived evaluation signals. However, most existing RLHF methods depend on explicit feedback mechanisms such as button presses or preference labels, which disrupt the natural interaction process and impose a substantial cog- nitive load on the user. We propose a novel reinforcement learning from implicit human feedback (RLIHF) framework that utilizes non-invasive elec- troencephalography (EEG) signals, specifically error-related potentials (Er- rPs), to provide continuous, implicit feedback without requiring explicit user intervention. The proposed method adopts a pre-trained decoder to trans- i form raw EEG signals into probabilistic reward components, enabling effective policy learning even in the presence of sparse external rewards. We evaluate our approach in a simulation environment built on the MuJoCo physics en- gine, using a Kinova Gen2 robotic arm to perform a complex pick-and-place task that requires avoiding obstacles while manipulating target objects. The results show that agents trained with decoded EEG feedback achieve perfor- mance comparable to those trained with dense, manually designed rewards. These findings validate the potential of using implicit neural feedback for scalable and human-aligned reinforcement learning in interactive robotics. Keywords: human-robot interaction, brain-computer interface, electroen- cephalography, error-related potential, reinforcement learning from human feedback ii ์ธ๊ฐ„์˜๋„๊ธฐ๋ฐ˜๋กœ๋ด‡์ œ์–ด๋ฅผ์œ„ํ•œ ์•”๋ฌต์ ์‹ ๊ฒฝํ”ผ๋“œ๋ฐฑ๊ธฐ๋ฐ˜๊ฐ•ํ™”ํ•™์Šต ๊น€์ˆ˜์ง€ ์ธ๊ณต์ง€๋Šฅํ•™๊ณผ ์ง€๋„๊ต์ˆ˜: ์ด์„ฑํ™˜ ์ดˆ๋ก ๊ธฐ์กด์˜๊ฐ•ํ™”ํ•™์Šต(Reinforcement Learning, RL) ์ ‘๊ทผ๋ฒ•์€ํฌ์†Œ๋ณด์ƒ(sparse reward) ํ™˜๊ฒฝ์—์„œํšจ๊ณผ์ ์ธ์ •์ฑ…(policy)์„ํ•™์Šตํ•˜๋Š”๋ฐ์–ด๋ ค์›€์„๊ฒช์œผ๋ฉฐ, ์ด ๋กœ์ธํ•ด๋ณต์žกํ•˜๊ณ ๊ณผ์—…๋ณ„๋กœํŠนํ™”๋œ๋ณด์ƒํ•จ์ˆ˜๋ฅผ์ˆ˜๋™์œผ๋กœ์„ค๊ณ„ํ•ด์•ผํ•˜๋Š”ํ•œ ๊ณ„๊ฐ€์กด์žฌํ•œ๋‹ค. ์ด๋Ÿฌํ•œ๋ฌธ์ œ๋ฅผํ•ด๊ฒฐํ•˜๊ธฐ์œ„ํ•ด, ์ธ๊ฐ„ํ”ผ๋“œ๋ฐฑ๊ธฐ๋ฐ˜๊ฐ•ํ™”ํ•™์Šต (Reinforcement Learning from Human Feedback, RLHF) ์ด์ฃผ์–ด์ง„๋ณด์ƒ์— ๋”ํ•ด์ธ๊ฐ„์˜ํ‰๊ฐ€์‹ ํ˜ธ(human-derived evaluation signals) ๋ฅผ๋ณด์กฐ์ ์œผ๋กœํ™œ์šฉ ํ•˜๋Š”์œ ๋งํ•œ์ „๋žต์œผ๋กœ์ฃผ๋ชฉ๋ฐ›๊ณ ์žˆ๋‹ค. ๊ทธ๋Ÿฌ๋‚˜๋Œ€๋ถ€๋ถ„์˜๊ธฐ์กดRLHF ๊ธฐ๋ฒ•์€๋ฒ„ ํŠผ์ž…๋ ฅ์ด๋‚˜์„ ํ˜ธ๋„๋ผ๋ฒจ๊ณผ๊ฐ™์€๋ช…์‹œ์ (explicit) ํ”ผ๋“œ๋ฐฑ๋ฉ”์ปค๋‹ˆ์ฆ˜์—์˜์กดํ•˜๊ณ  ์žˆ์œผ๋ฉฐ, ์ด๋Š”์ƒํ˜ธ์ž‘์šฉ์˜์ž์—ฐ์Šค๋Ÿฌ์›€์„์ €ํ•ดํ•˜๊ณ ์‚ฌ์šฉ์ž์—๊ฒŒ์ƒ๋‹นํ•œ์ธ์ง€์  ๋ถ€๋‹ด(cognitive load)์„์ดˆ๋ž˜ํ•œ๋‹ค๋Š”ํ•œ๊ณ„๊ฐ€์žˆ๋‹ค. ๋ณธ์—ฐ๊ตฌ์—์„œ๋Š”์ด๋Ÿฌํ•œํ•œ๊ณ„๋ฅผ๊ทน๋ณตํ•˜๊ธฐ์œ„ํ•ด, ๋น„์นจ์Šต์ (Non-invasive) ๋‡ŒํŒŒ (Electroencephalography, EEG) ์‹ ํ˜ธ, ํŠนํžˆ์˜ค๋ฅ˜๊ด€๋ จ์ „์œ„(Error-related Po- tentials, ErrPs) ๋ฅผํ™œ์šฉํ•˜์—ฌ์‚ฌ์šฉ์ž์˜๋ช…์‹œ์ ๊ฐœ์ž…์—†์ด๋„์ง€์†์ ์ด๊ณ ์•”๋ฌต์  ์ธํ”ผ๋“œ๋ฐฑ์„์ œ๊ณตํ• ์ˆ˜์žˆ๋Š”์•”๋ฌต์ ์ธ๊ฐ„ํ”ผ๋“œ๋ฐฑ๊ธฐ๋ฐ˜๊ฐ•ํ™”ํ•™์Šต(Reinforcement iii Learning from Implicit Human Feedback, RLIHF) ํ”„๋ ˆ์ž„์›Œํฌ๋ฅผ์ œ์•ˆํ•œ๋‹ค. ์ œ์•ˆ๋œ๋ฐฉ๋ฒ•์€์‚ฌ์ „ํ•™์Šต๋œ(Pre-trained) ๋””์ฝ”๋”๋ฅผ์ด์šฉํ•ดEEG ์›์‹œ์‹ ํ˜ธ๋ฅผ ํ™•๋ฅ ์ ๋ณด์ƒ์„ฑ๋ถ„(probabilistic reward components)์œผ๋กœ๋ณ€ํ™˜ํ•จ์œผ๋กœ์จ, ์™ธ๋ถ€ ๋ณด์ƒ์ดํฌ์†Œํ•œ์ƒํ™ฉ์—์„œ๋„ํšจ๊ณผ์ ์ธ์ •์ฑ…ํ•™์Šต์ด๊ฐ€๋Šฅํ•˜๋„๋กํ•œ๋‹ค. ์ œ์•ˆ๋œ์ ‘๊ทผ๋ฒ•์˜์œ ํšจ์„ฑ์„๊ฒ€์ฆํ•˜๊ธฐ์œ„ํ•ด, MuJoCo ๋ฌผ๋ฆฌ์—”์ง„(physics en- gine)์„๊ธฐ๋ฐ˜์œผ๋กœํ•œ์‹œ๋ฎฌ๋ ˆ์ด์…˜ํ™˜๊ฒฝ์—์„œKinova Gen2 ๋กœ๋ด‡๋งค๋‹ˆํ“ฐ๋ ˆ์ดํ„ฐ ๋ฅผ์‚ฌ์šฉํ•˜์—ฌ๋ณต์žกํ•œํ”ฝ์•คํ”Œ๋ ˆ์ด์Šค(pick-and-place) ๊ณผ์—…์„์ˆ˜ํ–‰ํ•˜์˜€๋‹ค. ํ•ด๋‹น ๊ณผ์—…์€๋ชฉํ‘œ๋ฌผ์ฒด๋ฅผ์กฐ์ž‘ํ•˜๋ฉด์„œ์žฅ์• ๋ฌผ์„ํšŒํ”ผํ•ด์•ผํ•˜๋Š”๋ณตํ•ฉ์ ์กฐ์ž‘ํ™˜๊ฒฝ์„ ํฌํ•จํ•œ๋‹ค. ์‹คํ—˜๊ฒฐ๊ณผ, EEG ํ”ผ๋“œ๋ฐฑ์œผ๋กœํ•™์Šต๋œ์—์ด์ „ํŠธ๋Š”์กฐ๋ฐ€ํ•˜๊ณ ์ˆ˜๋™์„ค ๊ณ„๋œ๋ณด์ƒ(dense manual rewards) ์œผ๋กœํ•™์Šต๋œ๋ชจ๋ธ๊ณผ์œ ์‚ฌํ•œ์ˆ˜์ค€์˜์„ฑ๋Šฅ์„ ๋‹ฌ์„ฑํ•˜์˜€๋‹ค. ์ด๋Ÿฌํ•œ๊ฒฐ๊ณผ๋Š”์•”๋ฌต์ ์‹ ๊ฒฝํ”ผ๋“œ๋ฐฑ(implicit neural feedback)์„ํ™œ ์šฉํ•œ๊ฐ•ํ™”ํ•™์Šต์ด๋Œ€๊ทœ๋ชจํ™•์žฅ์„ฑ(scalability) ๊ณผ์ธ๊ฐ„์ ์‘ํ˜•(human-adaptive) ๋กœ๋ด‡ํ•™์Šต์„๊ตฌํ˜„ํ• ์ˆ˜์žˆ๋Š”์ž ์žฌ๋ ฅ์„์ง€๋‹˜์„์ž…์ฆํ•œ๋‹ค. ์ฃผ์ œ์–ด: ์ธ๊ฐ„-๋กœ๋ด‡์ƒํ˜ธ์ž‘์šฉ, ๋‡Œ-์ปดํ“จํ„ฐ์ธํ„ฐํŽ˜์ด์Šค, ๋‡ŒํŒŒ, ์˜ค๋ฅ˜๊ด€๋ จ์ „์œ„, ์ธ๊ฐ„ํ”ผ๋“œ๋ฐฑ๊ธฐ๋ฐ˜๊ฐ•ํ™”ํ•™์Šต iv Preface This dissertation is submitted for the degree of Master of Science in Ar- tificial Intelligence at Korea University. The research described herein was conducted under the supervision of Professor Seong-Whan Lee in the Depart- ment of Artificial Intelligence, Korea University. Part of this work has been submitted to the IEEE International Conference on Systems, Man, and Cy- bernetics. I was the lead investigator for the projects where I was responsible for all major areas of concept formation, data collection and analysis, as well as the majority of manuscript composition. Neither this, nor any substan- tially similar dissertation has been or is being submitted for any other degree, diploma, or other qualification at any other university v Acknowledgement This research was supported by the Institute of Information & Communications Technology Planning & Evaluation (IITP) grant, funded by the Korea

…(Full text truncated)…

๐Ÿ“ธ Image Gallery

figure1.png figure1.webp figure2.png figure2.webp figure3.png figure3.webp figure5.png figure5.webp figure6.png figure6.webp scan_signatured.png scan_signatured.webp

Reference

This content is AI-processed based on ArXiv data.

Start searching

Enter keywords to search articles

โ†‘โ†“
โ†ต
ESC
โŒ˜K Shortcut