Reference Class-Based Improvement of Object Detection Accuracy

Raegeun Park (1), Jaechoon Jo (2)
(1) Department of Smart Information and Telecommunication Engineering, Sangmyung University, Cheonan, 31066, Republic of Korea
(2) Division of Computer Engineering, Hanshin University, Osan,18101, Korea, Republic of Korea
Fulltext View | Download
How to cite (IJASEIT) :
Park, Raegeun, and Jaechoon Jo. “Reference Class-Based Improvement of Object Detection Accuracy”. International Journal on Advanced Science, Engineering and Information Technology, vol. 10, no. 4, Aug. 2020, pp. 1526-35, doi:10.18517/ijaseit.10.4.12792.
To date, the Frames Per Second (FPS) and accuracy of object detection based on deep learning have made rapid progress. However, the accuracy is limited by issues such as false positive (FP) cases. FP cases can trigger malfunctions in applications requiring high accuracy, such as in autonomous vehicles, where it is essential to ensure driver safety when malfunctions occur. To reduce the occurrences of FP cases, we conducted an experiment to derive the association by separately detecting a highly relevant element called a reference class, in addition to the target class to be detected. To measure the association, we obtained the integrated association by first finding the associations between the bounding boxes of the target and reference classes. Then we generated a reference class-based model by applying the integrated association to a trained model. The reference class-based model achieved approximately 15% higher accuracy than the trained model at iteration 1,000. Besides, the proposed model reduced the FP cases to approximately half of the 18.964% in the conventional method; the FP reduction through an increase in iteration was only 11.008%. The reference class can be applied in various fields, such as security and autonomous vehicle technology. It can be used to reduce the FP cases and improve the accuracy performance limits in object detection. Furthermore, it is possible to reduce the cost of reinforcing the training dataset and using high-performance hardware, and the time cost of increasing training numbers.

A. Krizhevsky, I. Sutskever and G. E. Hinton, “Imagenet classification with deep convolutional neural networks,” In Advances in neural information processing systems, pp. 1097-1105, 2012

E. Karami, S. Prasad and M. Shehata, “Image matching using SIFT, SURF, BRIEF and ORB: performance comparison for distorted images,” arXiv preprint arXiv:1710.02726, Oct. 2017

J. Redmon and A. Farhadi, “YOLO9000: Better, Faster, Stronger,” The IEEE conference on computer vision and pattern recognition, pp. 7263-7271, 2017

J. Redmon and A. Farhadi, “YOLOv3: An Incremental Improvement,” arXiv:1804.02767, 2018

YOLO. (2020), YOLO:Real-Time Object Detection. [Online]. Available: https://pjreddie.com/darknet/yolo/

T. Y. Lin, M. Maire, S. Belongie, J. Hays, P. Perona, D. Ramanan, P. Dollí¡r and C. L. Zitnick, “Microsoft COCO: Common Objects in Context”, European conference on computer vision. Springer, pp. 740-755, 2014

Y. Long, Y. Gong, Z. Xiao and Q. Liu, “Accurate object localization in remote sensing images based on convolutional neural networks,” IEEE Transactions on Geoscience and Remote Sensing, 55(5), pp. 2486-2498, Jan. 2017

X. Wang, A. Shrivastava and A. Gupta, “A-fast-rcnn: Hard positive generation via adversary for object detection,” the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2606-2615, 2017

B. Singh and L. S. Davis, “An analysis of scale invariance in object detection snip,” the IEEE conference on computer vision and pattern recognition, pp. 3578-3587, 2018

Z. Cai and N. Vasconcelos, “Cascade r-cnn: Delving into high quality object detection,” the IEEE conference on computer vision and pattern recognition, pp. 6154-6162, 2018

Y. Chen, W. Li, C. Sakaridis, D. Dai and L. V. Gool, “Domain adaptive faster r-cnn for object detection in the wild,” the IEEE conference on computer vision and pattern recognition, pp. 3339-3348, 2018

J. Jeong, H. Park and N. Kwak, “Enhancement of SSD by concatenating feature maps for object detection,” arXiv preprint arXiv:1705.09587, May. 2017

X. Sun, P. Wu and S. C. H. Hoi, “Face detection using deep learning: An improved faster RCNN approach,” Neurocomputing, 299, 42-50, 2018

X. Zhu, Y. Wang, J. Dai, L. Yuan and Y. Wei, “Flow-guided feature aggregation for video object detection,” the IEEE International Conference on Computer Vision, pp. 408-417, 2017

T. Y. Lin, P. Goyal, R. Girshick, K. He and P. Dollí¡r, “Focal loss for dense object detection,” the IEEE international conference on computer vision, pp. 2980-2988, 2017

T. Y. Lin and S. Maji, “Improved bilinear pooling with cnns,” arXiv preprint arXiv:1707.06772, 2017

L. Tychsen-Smith and L. Petersson, “Improving object localization with fitness nms and bounded iou loss,” the IEEE conference on computer vision and pattern recognition, pp. 6877-6885, 2018

X. Wang, X. Hua, F. Xiao, Y. Li, X. Hu and P. Sun, “Multi-object detection in traffic scenes based on improved SSD,” Electronics, 7(11), 302, 2018

G. Bertasius, L. Torresani and J. Shi, “Object detection in video with spatiotemporal sampling networks,” the European Conference on Computer Vision (ECCV), pp. 331-346, 2018

J. Li, X. Liang, Y. Wei, T. Xu, J. Feng and S. Yan, “Perceptual generative adversarial networks for small object detection,” the IEEE conference on computer vision and pattern recognition, pp. 1222-1230, 2017

T. Kong, F. Sun, A. Yao, H. Liu, M. Lu and Y. Chen, “Ron: Reverse connection with objectness prior networks for object detection,” the IEEE conference on computer vision and pattern recognition, pp. 5936-5944, 2017

Y. Li, Y. Chen, N. Wang and Z. Zhang, “Scale-aware trident networks for object detection,” the IEEE International Conference on Computer Vision, pp. 6054-6063, 2019

P. Zhou, B. Ni, C. Geng, J. Hu and Y. Xu, “Scale-transferrable object detection,” the IEEE conference on computer vision and pattern recognition, pp. 528-537, 2018

Z. Zhang, S. Qiao, C. Xie, W. Shen, B. Wang and A. L. Yuille, “Single-shot object detection with enriched semantics,” the IEEE Conference on Computer Vision and Pattern Recognition, pp. 5813-5821, 2018

S. Zhang, L. Wen, X. Bian, Z. Lei and S. Z. Li, “Single-shot refinement neural network for object detection,” the IEEE conference on computer vision and pattern recognition, pp. 4203-4212, 2018

N. Bodla, B. Singh, R. Chellappa and L. S. Davis, “Soft-NMS--improving object detection with one line of code,” the IEEE international conference on computer vision, pp. 5561-5569, 2017

K. Kang, H. Li, J. Yan, X. Zeng, B. Yang, T. Xiao, C. Zhang, Z. Wang, R. Wang, X. Wang and W. Ouyang, “T-cnn: Tubelets with convolutional neural networks for object detection from videos,” IEEE Transactions on Circuits and Systems for Video Technology, 28(10), pp. 2896-2907, 2017

R. G. Park, H. J. Yun, E. G. Han, S. W. Kang, J. H. Park, E. J. Lee, D. H. Jeon, K. Y. Jung, S. B. Cho and T. K. Kang, “A Study on Countermeasures for ADAS Malfunction Based on YOLO”, Korean Institute of Electrical Engineers(KIEE) Conference, pp.188-190, Nov. 2019

T. Karras, S. Laine and T. Aila, “A style-based generator architecture for generative adversarial networks,” The IEEE Conference on Computer Vision and Pattern Recognition, pp. 4401-4410, 2019

N. Dalal and B. Triggs, “Histograms of oriented gradients for human detection,” The IEEE computer society conference on computer vision and pattern recognition, pp. 886-893, Jun. 2005

M. Everingham, A. Zisserman, C. K. I. Williams, L. V. Gool, M. Allan, C. M. Bishop, O. Chapelle, N. Dalal, T. Deselaers, G. Dork´o, S. Duffner, J. Eichhorn, J. D. R. Farquhar, M. Fritz, C. Garcia, T. Griffiths, F. Jurie, D. Keysers, M. Koskela, J. Laaksonen, D. Larlus and B. Leibe, “The 2005 PASCAL Visual Object Classes Challenge”, In Machine Learning Challenges Workshop, pp. 117-176, Apr. 2005

V. S. Rotenberg, “Moravec’s paradox: consideration in the context of two brain hemisphere functions,” Activitas Nervosa Superior, 55(3), pp. 108-111, 2013

(2020) Joint Base Langley-Eustis website. [Online]. Available: https://www.jble.af.mil/News/Photos/igphoto/2000197216/

Authors who publish with this journal agree to the following terms:

    1. Authors retain copyright and grant the journal right of first publication with the work simultaneously licensed under a Creative Commons Attribution License that allows others to share the work with an acknowledgement of the work's authorship and initial publication in this journal.
    2. Authors are able to enter into separate, additional contractual arrangements for the non-exclusive distribution of the journal's published version of the work (e.g., post it to an institutional repository or publish it in a book), with an acknowledgement of its initial publication in this journal.
    3. Authors are permitted and encouraged to post their work online (e.g., in institutional repositories or on their website) prior to and during the submission process, as it can lead to productive exchanges, as well as earlier and greater citation of published work (See The Effect of Open Access).