TCANet is a novel end-to-end model designed for motor imagery EEG signal decoding, enhancing the capabilities of existing frameworks like CTNet and MSCFormer. It employs a combination of multi-scale CNN, temporal convolutional networks, and multi-head self-attention to effectively capture spatiotemporal dependencies, achieving high classification accuracies on BCI IV-2a and IV-2b datasets. The model demonstrates competitive performance in both subject-dependent and subject-independent settings, indicating its potential for advancing brain-computer interface systems.
The Low-to-high Multi-Level Transformer (LMLT) introduces a novel approach for image super-resolution that reduces the complexity and inference time associated with existing Vision Transformer models. By employing attention mechanisms with varying feature sizes and integrating results from lower heads into higher heads, LMLT effectively captures both local and global information, mitigating issues related to window boundaries in self-attention. Experimental results indicate that LMLT outperforms state-of-the-art methods while significantly reducing GPU memory usage.