Journal of Systems Engineering and Electronics ›› 2025, Vol. 36 ›› Issue (6): 1389-1397.doi: 10.23919/JSEE.2025.000051

• ELECTRONICS TECHNOLOGY •    

The brief self-attention module for lightweight convolution neural networks

Jie YAN1(), Yingmei WEI1(), Yuxiang XIE1,*(), Quanzhi GONG1(), Shiwei ZOU1(), Xidao LUAN2()   

  1. 1 Laboratory for Big Data and Decision, College of Systems Engineering, National University of Defense Technology, Changsha 410000, China
    2 College of Computer Science and Engineering, Changsha University, Changsha 410000, China
  • Received:2024-08-26 Accepted:2024-08-26 Online:2025-12-18 Published:2026-01-07
  • Contact: Yuxiang XIE E-mail:yanjie@nudt.edu.cn;weiyingmei@nudt.edu.cn;xyx89@163.com;charles_g27@qq.com;1530531454@qq.com;xidaoluan@ccsu.cn
  • About author:
    YAN Jie was born in 1999. She received her B.S. degree in cost engineering from Qingdao University of Technology, China, in 2016. She is pursing her Ph.D. degree with the College of Systems Engineering, National University of Defense Technology, China. Her research interests include multi-modal semantic understanding, image captioning, and image classification. E-mail: yanjie@nudt.edu.cn

    WEI Yingmei was born in 1972. She received her Ph.D. degree in computer science and technology from National University of Defense Technology, China, in 2000, where she is a professor with the College of Systems Engineering. Her research interests include virtual reality, information visualization, and visual analysis techniques. E-mail: weiyingmei@nudt.edu.cn

    XIE Yuxiang was born in 1976. She received her B.S., M.S. and Ph.D. degrees in management science and engineering from National University of Defense Technology in 1998, 2001 and 2004 respectively, all in the College of Systems Engineering. She is a professor in the College of Systems Engineering, National University of Defense Technology. Her research interests include computer vision, image and video analysis, classification and retrieval. E-mail: yxxie@nudt.edu.cn

    GONG Quanzhi was born in 1998. He received his B.S. and M.S. degrees in control science and engineering from National University of Defense Technology, China, in 2020 and 2022. His research interests include action recognition and fine-grained image classification. E-mail: charles_g27@qq.com

    ZOU Shiwei was born in 2001. She received her B.S. degree in target engineering from National University of Defense Technology, China, in 2023, where she is pursing her M.S. degree with the College of Systems Engineering, National University of Defense Technology, China. Her current research interests include multi-modal semantic understanding, remote sensing image captioning, and image classification. E-mail: zsw0915@nudt.edu.cn

    LUAN Xidao was born in 1976. He received his B.S. degree in applied mathematics in 1998, M.S. and Ph.D. degrees in management science and engineering in 2005, 2009 respectively, all from National University of Defense Technology. He is a professor with the College of Computer Science and Engineering, Changsha University. His research interests include computer vision, image and video analysis, classification and retrieval. E-mail: xidaoluan@ccsu.cn

Abstract:

Lightweight convolutional neural networks (CNNs) have simple structures but struggle to comprehensively and accurately extract important semantic information from images. While attention mechanisms can enhance CNNs by learning distinctive representations, most existing spatial and hybrid attention methods focus on local regions with extensive parameters, making them unsuitable for lightweight CNNs. In this paper, we propose a self-attention mechanism tailored for lightweight networks, namely the brief self-attention module (BSAM). BSAM consists of the brief spatial attention (BSA) and advanced channel attention blocks. Unlike conventional self-attention methods with many parameters, our BSA block improves the performance of lightweight networks by effectively learning global semantic representations. Moreover, BSAM can be seamlessly integrated into lightweight CNNs for end-to-end training, maintaining the network’s lightweight and mobile characteristics. We validate the effectiveness of the proposed method on image classification tasks using the Food-101, Caltech-256, and Mini-ImageNet datasets.

Key words: self-attention, lightweight neural network, deep learning