Long-distance gesture detection based on deep learning for 3D spatial interaction

2021 
With the technological development of stereoscopic display, an immersive 3D space with large size can be reconstructed more and more easily, and a 3D spatial interaction method with high-efficiency become more and more urgent. Gesture interaction, as the most natural and efficient way of human-computer interaction, can convey information very quickly and efficiently. However, the effective interaction distance of most existing gesture interaction methods is less than one meter, and can not meet the demand of the long distance 3D spatial interaction. In this paper, an efficient network named Gesture YOLO for long-distance gesture detection is proposed to achieve the small gesture object detection with improved accuracy. There are two modules in our Gesture YOLO, one is the Dual CSPDarknet53-tiny Backbone module for fusing person features and gesture features, and the other is the Progressive Multi-Scale Feature Fusion module for enhancing output features. The experimental results on our test set show that our Gesture YOLO can achieve higher gesture detection accuracy than the YOLOv4-tiny at distances ranging from 2m to 5m, and can mitigate the significant drop in gesture detection accuracy when the distance increases.
    • Correction
    • Source
    • Cite
    • Save
    • Machine Reading By IdeaReader
    0
    References
    0
    Citations
    NaN
    KQI
    []