咨询与建议

看过本文的还看了

相关文献

该作者的其他文献

文献详情 >EdgeOAR: Real-time Online Acti... 收藏
arXiv

EdgeOAR: Real-time Online Action Recognition On Edge Devices

作     者:Luo, Wei Zhang, Deyu Tang, Ying Wu, Fan Zhang, Yaoxue 

作者机构:School of Computer Science and Engineering Central South University Changsha410083 China Big Data Institute Central South University Changsha410083 China Department of Computer Science and Technology Tsinghua University Beijing100084 China 

出 版 物:《arXiv》 (arXiv)

年 卷 期:2024年

核心收录:

主  题:Spatio temporal data 

摘      要:This paper addresses the challenges of Online Action Recognition (OAR), a framework that involves instantaneous analysis and classification of behaviors in video streams. OAR must operate under stringent latency constraints, making it an indispensable component for real-time feedback for edge computing. Existing methods, which typically rely on the processing of entire video clips, fall short in scenarios requiring immediate recognition. To address this, we designed EdgeOAR, a novel framework specifically designed for OAR on edge devices. EdgeOAR includes the Early Exit-oriented Task-specific Feature Enhancement Module (TFEM), which comprises lightweight submodules to optimize features in both temporal and spatial dimensions. We design an iterative training method to enable TFEM learning features from the beginning of the video. Additionally, EdgeOAR includes an Inverse Information Entropy (IIE) and Modality Consistency (MC)-driven fusion module to fuse features and make better exit decisions. This design overcomes the two main challenges: robust modeling of spatio-temporal action representations with limited initial frames in online video streams and balancing accuracy and efficiency on resource-constrained edge devices. Experiments show that on the UCF-101 dataset, our method EdgeOAR reduces latency by 99.23% and energy consumption by 99.28% compared to state-of-the-art (SOTA) method. And achieves an adequate accuracy on edge devices. © 2024, CC BY.

读者评论 与其他读者分享你的观点

用户名:未登录
我的评分