生成式人工智能驱动的EEG视觉重建脑机接口

A Review of Brain-Computer Interfaces Leveraging Generative AI for EEG Visual Reconstruction

  • 摘要: 生成式人工智能(GenAI)正重塑脑电图(EEG)视觉重建脑机接口(BCI)的研究范式。该文系统综述了该交叉领域的前沿进展,其核心贡献在于:从模型演进的视角,揭示了EEG视觉重建从判别式分类向生成式重构的范式迁移——在编码-解码架构下,任务形态已实现从静态图像到动态视频、三维模型的多维跃迁,生成模型亦历经GAN、VAE向扩散模型的迭代演进,各架构在语义一致性、结构保真度与训练稳定性上形成鲜明权衡。该文进一步梳理了公开数据集的演进脉络:从早期ImageNet-EEG的小样本验证,到THINGS-EEG、Alljoined等百万试次级大规模数据,再到SEED-DV、EEG-3D等时序化、多模态新基准,数据规模呈指数级增长,但跨被试一致性、生成式任务适配性等瓶颈仍制约领域发展。未来突破需在神经生理合理编码、视觉先验解耦与标准化评估体系构建三方面协同发力,推动EEG视觉重建从实验室原型走向真实场景,为解码大脑视觉表征机制提供可计算的建模新路径。

     

    Abstract: Generative models are reshaping the research paradigm of visual reconstruction via Electroencephalogram (EEG) in Brain-Computer Interface (BCI) research. This article provides a systematic review of the cutting-edge advancements in this interdisciplinary field. Its primary contribution lies in revealing, from the perspective of model evolution, a paradigm shift in EEG-based visual reconstruction from discriminative classification towards generative reconstruction. Under the prevailing encoder-decoder architecture, the task framework has evolved from static images to encompass dynamic videos and three-dimensional models. Concurrently, generative models themselves have undergone iterative advancements from Generative Adversarial Networks (GANs) and Variational Autoencoders (VAEs) to Diffusion Models, with each architecture presents distinct trade-offs in semantic consistency, structural fidelity, and training stability. Furthermore, this artick delineates the evolution of public datasets: from small-sample verification with early datasets like ImageNet-EEG, to large-scale datasets such as THINGS-EEG and Alljoined containing millions of trials, and further to newer benchmarks like SEED-DV and EEG-3D that feature temporal sequences and multimodality. While the scale of data has grown exponentially, bottlenecks including cross-subject consistency and suitability for generative tasks continue to challenge the field. Future breakthroughs necessitate concerted efforts in three key areas: the rational encoding of neurophysiological principles, the disentanglement of visual priors, and the establishment of standardized evaluation frameworks. Such advancements are crucial to transition EEG-based visual reconstruction from laboratory prototypes to real-world applications, thereby offering a novel computational modeling pathway for decoding the brain's visual representation mechanisms.

     

/

返回文章
返回