Neuro-Cognitive Reward Modeling for Human-Centered Autonomous Vehicle Control

IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2026 Main Track

Abstract

Recent advancements in computer vision have accelerated the development of autonomous driving. Despite these advancements, training machines to drive in a way that aligns with human expectations remains a significant challenge. Humans possess sophisticated cognitive systems capable of rapidly interpreting scene information and making decisions in complex environments. In this work, we propose an electroencephalography (EEG)-guided decision-making framework to incorporate human cognitive insights into reinforcement learning for autonomous driving. EEG signals were collected from 20 participants in a realistic driving simulator and event-related potentials (ERP) were analyzed in response to sudden environmental changes. We train a neural network to predict ERP strength directly from visual scene information and integrate this predicted cognitive signal into the reward function of a reinforcement learning policy. Experimental results show that the proposed neuro-cognitive reward modeling framework improves collision avoidance and driving safety. These results demonstrate the potential of leveraging human neuro-cognitive feedback to enhance human-centered autonomous driving systems.

Method and Contributions

Framework
Overview of the proposed neuro-cognitive reward modeling framework.

This study leverages the advantages of event-related potentials (ERPs), specifically their reliable millisecond-level representation of natural brain responses. It extends classical ERP analysis by exploring the potential of using ERP signals to train reinforcement learning (RL) models for autonomous intelligent vehicle (AIV) tasks.

We collect EEG data from 20 participants while they actively drive in a realistic VR-based driving simulator. By analyzing their ERP responses to sudden environmental changes, we observe a correlation between ERP amplitude and driver reaction time. Based on this neuro-cognitive insight, we design a human cognitive reward model that predicts ERP strength directly from scene images. The predicted ERP signal is then incorporated into the reward function of a reinforcement learning agent.

Unlike traditional RL with EEG feedback, which requires EEG signals during inference, our approach distills human cognitive responses into a vision-based model. As a result, the trained RL agent can leverage human cognitive feedback without requiring EEG signals during deployment.

Main Contributions

  • Neuro-cognitive driving dataset. A VR-based dataset with EEG, eye-tracking, and driving behavior.
  • Human cognitive reward model. Predicts EEG-derived signals from visual input.
  • Neuro-cognitive RL framework. Integrates cognitive reward for improved safety and performance.

Results

Qualitative result
Machine attention visualization across time steps.
Quantitative result
Summary statistics for driving performance.

Data Collection Setup