With the development of deep neural networks (DNN), building visual decoding models based on functional magnetic resonance imaging (fMRI) to simulate the visual system of the human brain and studying visual mechanisms have become a research hotspot. Although existing visual decoding models built using DNNs have achieved a certain accuracy, most models ignore the differences between different voxels. Among them, the BRNN-based category decoding model uses the bidirectional long short term memory (LSTM) network to simulate the visual bidirectional information flow, which improves the decoding accuracy, but it uses the voxels of each brain area as an overall input model. Therefore, we embed the channel attention module, the Squeeze-and-Excitation Networks (SENet), into the LSTM network to construct an LSTM-SENet vision that introduces an attention mechanism The decoding model allows the model to learn by itself and assign different weights to each voxel, focusing on important voxels, thereby improving the classification accuracy of natural images. The experimental results show that our method improves the accuracy of (three-level) category decoding than other methods, and the results further verify the effectiveness of building a visual decoding model based on the visual mechanism.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
INSTITUTIONAL Select your institution to access the SPIE Digital Library.
PERSONAL Sign in with your SPIE account to access your personal subscriptions or to use specific features such as save to my library, sign up for alerts, save searches, etc.