咨询与建议

看过本文的还看了

相关文献

该作者的其他文献

文献详情 >Epipolar Attention Field Trans... 收藏
arXiv

Epipolar Attention Field Transformers for Bird’s Eye View Semantic Segmentation

作     者:Witte, Christian Behley, Jens Stachniss, Cyrill Raaijmakers, Marvin 

作者机构:CARIAD SE Germany Center for Robotics University of Bonn Germany Lamarr Institute for Machine Learning and Artificial Intelligence Germany 

出 版 物:《arXiv》 (arXiv)

年 卷 期:2024年

核心收录:

主  题:Semantic Segmentation 

摘      要:Spatial understanding of the semantics of the surroundings is a key capability needed by autonomous cars to enable safe driving decisions. Recently, purely vision-based solutions have gained increasing research interest. In particular, approaches extracting a bird’s eye view (BEV) from multiple cameras have demonstrated great performance for spatial understanding. This paper addresses the dependency on learned positional encodings to correlate image and BEV feature map elements for transformer-based methods. We propose leveraging epipolar geometric constraints to model the relationship between cameras and the BEV by Epipolar Attention Fields. They are incorporated into the attention mechanism as a novel attribution term, serving as an alternative to learned positional encodings. Experiments show that our method EAFormer outperforms previous BEV approaches by 2% mIoU for map semantic segmentation and exhibits superior generalization capabilities compared to implicitly learning the camera configuration. Copyright © 2024, The Authors. All rights reserved.

读者评论 与其他读者分享你的观点

用户名:未登录
我的评分