咨询与建议

看过本文的还看了

相关文献

该作者的其他文献

文献详情 >Performer: A High-Performance ... 收藏

Performer: A High-Performance Global-Local Model-Augmented with Dual Network Interaction Mechanism

作     者:Tan, Dayu Hao, Rui Hua, Linfeng Xu, Qi Su, Yansen Zheng, Chunhou Zhong, Weimin 

作者机构:Anhui University Key Laboratory of Intelligent Computing and Signal Processing Ministry of Education Institutes of Physical Science and Information Technology Hefei230601 China East China University of Science and Technology Key Laboratory of Smart Manufacturing in Energy Chemical Process Ministry of Education Shanghai200237 China Dalian University of Technology School of Computer Science and Technology Dalian116024 China 

出 版 物:《IEEE Transactions on Cognitive and Developmental Systems》 (IEEE Trans. Cogn. Dev. Syst.)

年 卷 期:2024年

核心收录:

学科分类:1205[管理学-图书情报与档案管理] 0808[工学-电气工程] 08[工学] 

主  题:Semantic Segmentation 

摘      要:In deep learning, Convolutional Neural Networks (CNNs) focus on local information through convolutional kernels, while transformers attend to global information using self-attention mechanisms. The union of these distinct approaches enables a more comprehensive extraction of image features. However, the feature map dimensions of CNN and Transformer differ, leading to dimension mismatch issues when combining these architectures. Additionally, the parameter size of the hybrid model integrating both architectures remains large, making it difficult to train. To further augmenting the interpretation of complex image patterns, we present Performer, a dual-network architecture that seamlessly combines CNNs and transformers, resulting in a novel and efficient representation learning model. In the Performer model, we innovate by devising a unique interaction methodology for CNN and transformer architectures to enhance the image feature extraction capabilities mutually. To counteract issue of dimensionality mismatch, we also introduce a refined transformer block, a advancement over the transformer block of ViT. To validate the effectiveness of Performer, we conduct extensive experiments on both classification and segmentation tasks. Performer achieve an accuracy of 83.37% on the ImageNet-200 dataset. For semantic segmentation, Performer excels on the CamVid and Hippocampus datasets. On CamVid, our model achieves a mean Intersection over Union (mIoU) of 63.27% and Pixel Accuracy of 92.11%, demonstrating superior performance in capturing fine details and handling complex scenes effectively. The code is available at https://***/hlfthh/ Performer. © 2016 IEEE.

读者评论 与其他读者分享你的观点

用户名:未登录
我的评分