http://scholars.ntou.edu.tw/handle/123456789/21404
標題: | Multi-fusion feature pyramid for real-time hand detection | 作者: | Chang, Chuan-Wang Santra, Santanu Hsieh, Jun-Wei Hendri, Pirdiansyah Lin, Chi-Fang |
關鍵字: | Object detection;Hand detection;Embedded system;Human;YOLOV4 | 公開日期: | 3-三月-2022 | 出版社: | SPRINGER | 來源出版物: | MULTIMEDIA TOOLS AND APPLICATIONS | 摘要: | Real-time HI (Human Interface) systems need accurate and efficient hand detection models to meet the limited resources in budget, dimension, memory, computing, and electric power. The detection task is also important for other applications such as homecare systems, fine-grained action recognition, movie interpretation, and even for understanding dance gestures. In recent years, object detection has become a less challenging task with the latest deep CNN-based state-of-the-art models, i.e., RCNN, SSD, and YOLO. However, these models cannot achieve desired efficiency and accuracy on HI-based embedded devices due to their complex time-consuming architecture. Another critical issue in hand detection is that small hands (<30 x 30 pixels) are still challenging for all the above methods. We proposed a shallow model named Multi-fusion Feature Pyramid for real-time hand detection to deal with the above problems. Experimental results on the Oxford hand dataset combined with the skin dataset show that the proposed method outperforms other SoTA methods in terms of accuracy, efficiency, and real-time speed. The COCO dataset is also used to compare with other state-of-the-art method and shows the highest efficiency and accuracy with the proposed CFPN model. Thus we conclude that the proposed model is useful for real-life small hand detection on embedded devices. |
URI: | http://scholars.ntou.edu.tw/handle/123456789/21404 | ISSN: | 1380-7501 | DOI: | 10.1007/s11042-021-11897-7 |
顯示於: | 資訊工程學系 |
在 IR 系統中的文件,除了特別指名其著作權條款之外,均受到著作權保護,並且保留所有的權利。