PCSA: ENHANCING CNN PERFORMANCE WITH PYRAMID CHANNEL AND SPATIAL ATTENTION

PCSA: Enhancing CNN Performance With Pyramid Channel and Spatial Attention

PCSA: Enhancing CNN Performance With Pyramid Channel and Spatial Attention

Blog Article

Recent studies have demonstrated that the attention mechanism can effectively enhance the effectiveness of deep convolutional neural networks.In this paper, we propose a “Pyramid channel and spatial attention” (PCSA), which consists of reconstructing the features after pyramidal multiscale convolution by extracting spatial weights and channel weights.This dual weight extraction process helps to merge the multiscale information more accurately and enhances the vibrating table for chocolate model’s focus on the complex locations of image objects.

As a plug-and-play module, PCSA can be easily added to various backbone networks to enhance the modeling effect.We apply the PCSA module to two kinds of backbone read more networks, VGG and ResNet, and the improved models are named: VGG-PCSA and PCSANet, respectively.Experimental results show that on the CIFAR-10, CIFAR-100, and NaSC-TG2 datasets, our model has a significant performance improvement over the backbone networks while keeping the number of parameters low and performs better than most of the state-of-the-art channel attention methods.

In addition, we visualize feature maps and class activation diagrams to explain the better performance of PCSA.

Report this page