๐ Original Info
- Title: Reinforcement Learning from Implicit Neural Feedback for Human-Aligned Robot Control
- ArXiv ID: 2512.00050
- Date: 2025-11-18
- Authors: Suzie Kim
๐ Abstract
Conventional reinforcement learning (RL) approaches often struggle to learn effective policies under sparse reward conditions, necessitating the manual design of complex, task-specific reward functions. To address this limitation, reinforcement learning from human feedback (RLHF) has emerged as a promising strategy that complements hand-crafted rewards with human-derived evaluation signals. However, most existing RLHF methods depend on explicit feedback mechanisms such as button presses or preference labels, which disrupt the natural interaction process and impose a substantial cognitive load on the user. We propose a novel reinforcement learning from implicit human feedback (RLIHF) framework that utilizes non-invasive electroencephalography (EEG) signals, specifically error-related potentials (ErrPs), to provide continuous, implicit feedback without requiring explicit user intervention. The proposed method adopts a pre-trained decoder to transform raw EEG signals into probabilistic reward components, enabling effective policy learning even in the presence of sparse external rewards. We evaluate our approach in a simulation environment built on the MuJoCo physics engine, using a Kinova Gen2 robotic arm to perform a complex pick-and-place task that requires avoiding obstacles while manipulating target objects. The results show that agents trained with decoded EEG feedback achieve performance comparable to those trained with dense, manually designed rewards. These findings validate the potential of using implicit neural feedback for scalable and human-aligned reinforcement learning in interactive robotics.
๐ก Deep Analysis
Deep Dive into Reinforcement Learning from Implicit Neural Feedback for Human-Aligned Robot Control.
Conventional reinforcement learning (RL) approaches often struggle to learn effective policies under sparse reward conditions, necessitating the manual design of complex, task-specific reward functions. To address this limitation, reinforcement learning from human feedback (RLHF) has emerged as a promising strategy that complements hand-crafted rewards with human-derived evaluation signals. However, most existing RLHF methods depend on explicit feedback mechanisms such as button presses or preference labels, which disrupt the natural interaction process and impose a substantial cognitive load on the user. We propose a novel reinforcement learning from implicit human feedback (RLIHF) framework that utilizes non-invasive electroencephalography (EEG) signals, specifically error-related potentials (ErrPs), to provide continuous, implicit feedback without requiring explicit user intervention. The proposed method adopts a pre-trained decoder to transform raw EEG signals into probabilistic re
๐ Full Content
Masterโs Thesis
Reinforcement Learning from
Implicit Neural Feedback for
Human-Aligned Robot Control
Suzie Kim
Department of Artificial Intelligence
Graduate School
Korea University
February 2025
arXiv:2512.00050v1 [cs.RO] 18 Nov 2025
Reinforcement Learning from
Implicit Neural Feedback for
Human-Aligned Robot Control
by
Suzie Kim
under the supervision of Professor Seong-Whan Lee
A thesis submitted in partial fulfillment of the
requirements for the degree of Master of Science
Department of Artificial Intelligence
Graduate School
Korea University
October 2025
The thesis of Suzie Kim has been approved
by the thesis committee in partial fulfillment of
the requirements for the degree of
Master of Science
December 2025
Committee Chair: Seong-Whan Lee
Committee Member: Won-Zoo Chung
Committee Member: Tae-Eui Kam
Reinforcement Learning from
Implicit Neural Feedback for
Human-Aligned Robot Control
by Suzie Kim
Department of Artificial Intelligence
under the supervision of Professor Seong-Whan Lee
Abstract
Conventional reinforcement learning (RL) approaches often struggle to
learn effective policies under sparse reward conditions, necessitating the man-
ual design of complex, task-specific reward functions. To address this limi-
tation, reinforcement learning from human feedback (RLHF) has emerged as
a promising strategy that complements hand-crafted rewards with human-
derived evaluation signals. However, most existing RLHF methods depend
on explicit feedback mechanisms such as button presses or preference labels,
which disrupt the natural interaction process and impose a substantial cog-
nitive load on the user.
We propose a novel reinforcement learning from
implicit human feedback (RLIHF) framework that utilizes non-invasive elec-
troencephalography (EEG) signals, specifically error-related potentials (Er-
rPs), to provide continuous, implicit feedback without requiring explicit user
intervention. The proposed method adopts a pre-trained decoder to trans-
i
form raw EEG signals into probabilistic reward components, enabling effective
policy learning even in the presence of sparse external rewards. We evaluate
our approach in a simulation environment built on the MuJoCo physics en-
gine, using a Kinova Gen2 robotic arm to perform a complex pick-and-place
task that requires avoiding obstacles while manipulating target objects. The
results show that agents trained with decoded EEG feedback achieve perfor-
mance comparable to those trained with dense, manually designed rewards.
These findings validate the potential of using implicit neural feedback for
scalable and human-aligned reinforcement learning in interactive robotics.
Keywords: human-robot interaction, brain-computer interface, electroen-
cephalography, error-related potential, reinforcement learning from human
feedback
ii
์ธ๊ฐ์๋๊ธฐ๋ฐ๋ก๋ด์ ์ด๋ฅผ์ํ
์๋ฌต์ ์ ๊ฒฝํผ๋๋ฐฑ๊ธฐ๋ฐ๊ฐํํ์ต
๊น์์ง
์ธ๊ณต์ง๋ฅํ๊ณผ
์ง๋๊ต์: ์ด์ฑํ
์ด๋ก
๊ธฐ์กด์๊ฐํํ์ต(Reinforcement Learning, RL) ์ ๊ทผ๋ฒ์ํฌ์๋ณด์(sparse
reward) ํ๊ฒฝ์์ํจ๊ณผ์ ์ธ์ ์ฑ
(policy)์ํ์ตํ๋๋ฐ์ด๋ ค์์๊ฒช์ผ๋ฉฐ, ์ด
๋ก์ธํด๋ณต์กํ๊ณ ๊ณผ์
๋ณ๋กํนํ๋๋ณด์ํจ์๋ฅผ์๋์ผ๋ก์ค๊ณํด์ผํ๋ํ
๊ณ๊ฐ์กด์ฌํ๋ค. ์ด๋ฌํ๋ฌธ์ ๋ฅผํด๊ฒฐํ๊ธฐ์ํด, ์ธ๊ฐํผ๋๋ฐฑ๊ธฐ๋ฐ๊ฐํํ์ต
(Reinforcement Learning from Human Feedback, RLHF) ์ด์ฃผ์ด์ง๋ณด์์
๋ํด์ธ๊ฐ์ํ๊ฐ์ ํธ(human-derived evaluation signals) ๋ฅผ๋ณด์กฐ์ ์ผ๋กํ์ฉ
ํ๋์ ๋งํ์ ๋ต์ผ๋ก์ฃผ๋ชฉ๋ฐ๊ณ ์๋ค. ๊ทธ๋ฌ๋๋๋ถ๋ถ์๊ธฐ์กดRLHF ๊ธฐ๋ฒ์๋ฒ
ํผ์
๋ ฅ์ด๋์ ํธ๋๋ผ๋ฒจ๊ณผ๊ฐ์๋ช
์์ (explicit) ํผ๋๋ฐฑ๋ฉ์ปค๋์ฆ์์์กดํ๊ณ
์์ผ๋ฉฐ, ์ด๋์ํธ์์ฉ์์์ฐ์ค๋ฌ์์์ ํดํ๊ณ ์ฌ์ฉ์์๊ฒ์๋นํ์ธ์ง์
๋ถ๋ด(cognitive load)์์ด๋ํ๋ค๋ํ๊ณ๊ฐ์๋ค.
๋ณธ์ฐ๊ตฌ์์๋์ด๋ฌํํ๊ณ๋ฅผ๊ทน๋ณตํ๊ธฐ์ํด, ๋น์นจ์ต์ (Non-invasive) ๋ํ
(Electroencephalography, EEG) ์ ํธ, ํนํ์ค๋ฅ๊ด๋ จ์ ์(Error-related Po-
tentials, ErrPs) ๋ฅผํ์ฉํ์ฌ์ฌ์ฉ์์๋ช
์์ ๊ฐ์
์์ด๋์ง์์ ์ด๊ณ ์๋ฌต์
์ธํผ๋๋ฐฑ์์ ๊ณตํ ์์๋์๋ฌต์ ์ธ๊ฐํผ๋๋ฐฑ๊ธฐ๋ฐ๊ฐํํ์ต(Reinforcement
iii
Learning from Implicit Human Feedback, RLIHF) ํ๋ ์์ํฌ๋ฅผ์ ์ํ๋ค.
์ ์๋๋ฐฉ๋ฒ์์ฌ์ ํ์ต๋(Pre-trained) ๋์ฝ๋๋ฅผ์ด์ฉํดEEG ์์์ ํธ๋ฅผ
ํ๋ฅ ์ ๋ณด์์ฑ๋ถ(probabilistic reward components)์ผ๋ก๋ณํํจ์ผ๋ก์จ, ์ธ๋ถ
๋ณด์์ดํฌ์ํ์ํฉ์์๋ํจ๊ณผ์ ์ธ์ ์ฑ
ํ์ต์ด๊ฐ๋ฅํ๋๋กํ๋ค.
์ ์๋์ ๊ทผ๋ฒ์์ ํจ์ฑ์๊ฒ์ฆํ๊ธฐ์ํด, MuJoCo ๋ฌผ๋ฆฌ์์ง(physics en-
gine)์๊ธฐ๋ฐ์ผ๋กํ์๋ฎฌ๋ ์ด์
ํ๊ฒฝ์์Kinova Gen2 ๋ก๋ด๋งค๋ํฐ๋ ์ดํฐ
๋ฅผ์ฌ์ฉํ์ฌ๋ณต์กํํฝ์คํ๋ ์ด์ค(pick-and-place) ๊ณผ์
์์ํํ์๋ค. ํด๋น
๊ณผ์
์๋ชฉํ๋ฌผ์ฒด๋ฅผ์กฐ์ํ๋ฉด์์ฅ์ ๋ฌผ์ํํผํด์ผํ๋๋ณตํฉ์ ์กฐ์ํ๊ฒฝ์
ํฌํจํ๋ค. ์คํ๊ฒฐ๊ณผ, EEG ํผ๋๋ฐฑ์ผ๋กํ์ต๋์์ด์ ํธ๋์กฐ๋ฐํ๊ณ ์๋์ค
๊ณ๋๋ณด์(dense manual rewards) ์ผ๋กํ์ต๋๋ชจ๋ธ๊ณผ์ ์ฌํ์์ค์์ฑ๋ฅ์
๋ฌ์ฑํ์๋ค. ์ด๋ฌํ๊ฒฐ๊ณผ๋์๋ฌต์ ์ ๊ฒฝํผ๋๋ฐฑ(implicit neural feedback)์ํ
์ฉํ๊ฐํํ์ต์ด๋๊ท๋ชจํ์ฅ์ฑ(scalability) ๊ณผ์ธ๊ฐ์ ์ํ(human-adaptive)
๋ก๋ดํ์ต์๊ตฌํํ ์์๋์ ์ฌ๋ ฅ์์ง๋์์
์ฆํ๋ค.
์ฃผ์ ์ด: ์ธ๊ฐ-๋ก๋ด์ํธ์์ฉ, ๋-์ปดํจํฐ์ธํฐํ์ด์ค, ๋ํ, ์ค๋ฅ๊ด๋ จ์ ์,
์ธ๊ฐํผ๋๋ฐฑ๊ธฐ๋ฐ๊ฐํํ์ต
iv
Preface
This dissertation is submitted for the degree of Master of Science in Ar-
tificial Intelligence at Korea University. The research described herein was
conducted under the supervision of Professor Seong-Whan Lee in the Depart-
ment of Artificial Intelligence, Korea University. Part of this work has been
submitted to the IEEE International Conference on Systems, Man, and Cy-
bernetics. I was the lead investigator for the projects where I was responsible
for all major areas of concept formation, data collection and analysis, as well
as the majority of manuscript composition. Neither this, nor any substan-
tially similar dissertation has been or is being submitted for any other degree,
diploma, or other qualification at any other university
v
Acknowledgement
This research was supported by the Institute of Information & Communications
Technology Planning & Evaluation (IITP) grant, funded by the Korea
…(Full text truncated)…
๐ธ Image Gallery
Reference
This content is AI-processed based on ArXiv data.