Multi-dimensional Neural Decoding with Orthogonal Representations for Brain-Computer Interfaces
DOI:
https://doi.org/10.1609/aaai.v40i3.37191Abstract
Current brain-computer interfaces primarily decode single motor variables, limiting natural control requiring simultaneous multi-dimensional extraction. We introduce Multi-dimensional Neural Decoding (MND), a task that simultaneously extracts multiple motor variables (direction, position, velocity, acceleration) from single neural population recordings. MND faces two key challenges: cross-task interference when decoding correlated motor dimensions from shared cortical representations, and generalization issues across sessions, subjects, and paradigms. To address these challenges, we propose OrthoSchema, a multi-task framework inspired by cortical orthogonal subspace organization and cognitive schema reuse. OrthoSchema enforces representation orthogonality to eliminate cross-task interference and employs selective feature reuse transfer for few-shot cross-session, subject and paradigm adaptation. Experiments on macaque motor cortex datasets demonstrate that OrthoSchema significantly improves decoding accuracy in cross-session, subject and paradigm generalization tasks, with larger performance improvements when fine-tuning samples are limited. Ablation studies confirm the synergistic effects of all components are crucial, with OrthoSchema effectively modeling cross-task features and capturing session relationships for robust transfer. Our results provide new insights into scalable and robust neural decoding for real-world BCI applications.Published
2026-03-14
How to Cite
Tian, K., Zhao, S., Zhang, Y., & Yu, S. (2026). Multi-dimensional Neural Decoding with Orthogonal Representations for Brain-Computer Interfaces. Proceedings of the AAAI Conference on Artificial Intelligence, 40(3), 2092–2100. https://doi.org/10.1609/aaai.v40i3.37191
Issue
Section
AAAI Technical Track on Cognitive Modeling & Cognitive Systems