• Complex
  • Title
  • Keyword
  • Abstract
  • Scholars
  • Journal
  • ISSN
  • Conference
搜索

Author:

Xie, B. (Xie, B..) | Deng, Y. (Deng, Y..) | Shao, Z. (Shao, Z..) | Xu, Q. (Xu, Q..) | Li, Y. (Li, Y..)

Indexed by:

EI Scopus SCIE

Abstract:

Event cameras are neuromorphic vision sensors that record a scene as sparse and asynchronous event streams. Most event-based methods project events into dense frames and process them using conventional vision models, resulting in high computational complexity. A recent trend is to develop point-based networks that achieve efficient event processing by learning sparse representations. However, existing works may lack robust local information aggregators and effective feature interaction operations, thus limiting their modeling capabilities. To this end, we propose an attention-aware model named Event Voxel Set Transformer (EVSTr) for efficient spatiotemporal representation learning on event streams. It first converts the event stream into voxel sets and then hierarchically aggregates voxel features to obtain robust representations. The core of EVSTr is an event voxel transformer encoder that consists of two well-designed components, including the Multi-Scale Neighbor Embedding Layer (MNEL) for local information aggregation and the Voxel Self-Attention Layer (VSAL) for global feature interaction. Enabling the network to incorporate a long-range temporal structure, we introduce a segment modeling strategy (S2TM) to learn motion patterns from a sequence of segmented voxel sets. The proposed model is evaluated on two recognition tasks, including object classification and action recognition. To provide a convincing model evaluation, we present a new event-based action recognition dataset (NeuroHAR) recorded in challenging scenarios. Comprehensive experiments show that EVSTr achieves state-of-the-art performance while maintaining low model complexity. IEEE

Keyword:

neuromorphic vision object classification Feature extraction Streams Computational modeling Event camera Spatiotemporal phenomena Transformers attention mechanism Task analysis action recognition Cameras

Author Community:

  • [ 1 ] [Xie B.]Department of Mechanical Engineering, City University of Hong Kong, Hong Kong, SAR, China
  • [ 2 ] [Deng Y.]College of Computer Science, Beijing University of Technology, Beijing, China
  • [ 3 ] [Shao Z.]College of Information Science and Engineering, Hunan Normal University, Changsha, China
  • [ 4 ] [Xu Q.]Department of Electromechanical Engineering, University of Macau, Macao, SAR, China
  • [ 5 ] [Li Y.]Department of Mechanical Engineering, City University of Hong Kong, Hong Kong, SAR, China

Reprint Author's Address:

Email:

Show more details

Related Keywords:

Related Article:

Source :

IEEE Transactions on Circuits and Systems for Video Technology

ISSN: 1051-8215

Year: 2024

Page: 1-1

8 . 4 0 0

JCR@2022

Cited Count:

WoS CC Cited Count: 36

SCOPUS Cited Count: 1

ESI Highly Cited Papers on the List: 0 Unfold All

WanFang Cited Count:

Chinese Cited Count:

30 Days PV: 16

Affiliated Colleges:

Online/Total:574/10591618
Address:BJUT Library(100 Pingleyuan,Chaoyang District,Beijing 100124, China Post Code:100124) Contact Us:010-67392185
Copyright:BJUT Library Technical Support:Beijing Aegean Software Co., Ltd.