MG-GCT: A Motion-Guided Graph Convolutional Transformer for Traffic Gesture Recognition

For autonomous driving systems, it is crucial to recognize the actions and gestures of traffic conductors and cyclists on the road to ensure safety. However, traffic gesture recognition is more challenging than action recognition in general scenarios due to the differences in action posture and sample composition between traffic gesture datasets and general action datasets. Therefore, general action recognition methods cannot identify traffic gestures well. To overcome these problems, the authors propose a novel motion-guided graph convolutional transformer (MG-GCT) for traffic gesture recognition. Firstly, they proposed a two-stream network to fully utilize joint data and motion data for action recognition. Secondly, the authors designed and implemented a motion-guided module between two streams, which leverages the powerful spatial representation ability of the motion data to guide the learning of the joint data stream in the spatial dimension. Thirdly, they implemented a temporal transformer network to process the temporal features of the skeleton. Finally, the authors conducted extensive experiments on two public datasets and one dataset presented by themselves to demonstrate the effectiveness of their network in traffic gesture recognition, which has a significant advantage over the state-of-the-art methods.

Language

  • English

Media Info

Subject/Index Terms

Filing Info

  • Accession Number: 01942202
  • Record Type: Publication
  • Files: TRIS
  • Created Date: Jan 13 2025 10:24AM