Volume 4,Issue 1
Analysis and Research on Students' Classroom Behavior Data Based on Object Detection
Analyzing and studying students' classroom behavior is crucial for enhancing both students' abilities and teachers' instructional methods. This topic has been a significant focus within the educational community. Recently, machine vision and object detection technologies have been extensively applied across various domains, yielding notable outcomes. Consequently, this paper introduces a method for modeling and analyzing classroom behavior data using an object detection neural network. Experimental results indicate that this approach can effectively facilitate the development of students' abilities and the improvement of teaching practices.
[1]Buniyamin N, Mat U, Arshad P M, 2015, Educational data mining for prediction and classification of engineering students achievement. //2015 IEEE 7th International Conference on Engineering Education (ICEED). IEEE: 49-53.
[2]Lu G, Xie K, Liu Q, 2021, Automated Annotation of Classroom Behaviours with AI Engine. Open Education Research, 27(06): 97-107.
[3]Le C Y, Bengio Y, Hinton G, 2015, Deep learning. nature, 521(7553): 436-444.
[4]Treichler D G, 1967, Are you missing the boat in training aids. Film and AV Communication, 1: 14-16.
[5] Lin T Y, Goyal P, Girshick R, et al., 2017, Focal loss for dense object detection, in: Proceedings of the IEEE international conference on computer vision, 2980–2988.
[6] Redmon J, Divvala S, Girshick R, et al., 2016, You only look once: Unified, real-time object detection, in: Proceedings of the IEEE conference on computer vision and pattern recognition, 779–788.
[7] Liu W, Anguelov D, Erhan D, et al., 2016, Ssd: Single shot multibox detector, in: Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, October 11–14, Proceedings, Part I 14, Springer, 21–37.
[8] Ren S, He K, Girshick R, et al., 2017, Faster r-cnn: Towards real-time object detection with region proposal networks, IEEE transactions on pattern analysis and machine intelligence, 39(6): 1137-1149.
[9] Girshick R, 2015, Fast r-cnn, in: Proceedings of the IEEE international conference on computer vision, 1440–1448.
[10]He K, Zhang X, Ren S, et al., 2015, Spatial pyramid pooling in deep convolutional networks for visual recognition, IEEE transactions on pattern analysis and machine intelligence, 37(9): 1904-1916.
[11]Tian Z, Shen C, Chen H, et al., 2022, Fcos: A simple and strong anchor-free object detector, IEEE Transactions on Pattern Analysis and Machine Intelligence, 44(4): 1922-1933.
[12] Zhou X, Wang D, Kr¨ahenb¨uhl P, 2019, Objects as points, arXiv preprint arXiv: 1904.07850.
[13] Duan K, Bai S, Xie L, et al., 2019, Centernet: Keypoint triplets for object detection, in: Proceedings of the IEEE/CVF international conference on computer vision: 6569-6578.
[14] Law H, Deng J, 2018, Cornernet: Detecting objects as paired keypoints, in: Proceedings of the European conference on computer vision (ECCV): 734–750.
[15] Carion N, Massa F, Synnaeve G, et al., 2020, End-to-end object detection with transformers, in: European conference on computer vision, Springer: 213–229.
[16] Liu S, Li F, Zhang H, et al., 2022, DAB-DETR: Dynamic anchor boxes are better queries for DETR, in: International Conference on Learning Representations. https: //openreview.net/forum?id=oMI9PjOb9Jl.
[17] Li F, Zhang H, Liu S, et al., 2022, Dn-detr: Accelerate detr training by introducing query denoising, in: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition: 13619–13627.
[18] Roh B, Shin J, Shin W, et al., 2021, Sparse detr: Efficient end-to-end object detection with learnable sparsity, arXiv preprint arXiv: 2111.14330.
[19] Chen Q, Chen X, Zeng G, et al., 2022, Group detr: Fast training convergence with decoupled one-to-many label assignment, arXiv preprint arXiv: 2207.2207.13085.
[20]Jocher G, Qiu J, Chaurasia A, 2023, Ultralytics YOLO (Version 8.0.0) [Computer software]. https://github.com/ultralytics/ultralytics.
[21]Woo S, Hwang S, Kweon I S, 2018, Stairnet: Top-down semantic aggregation for accurate one-shot detection. In: 2018 IEEE Winter Conference on Applications of Computer Vision (WACV), IEEE: 1093-1102.
[22]Liu Z, Gao G, Sun L, 2020, Ipg-net: Image pyramid guidance network for small object detection. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops: 1026-1027.
[23]Gong Y, Yu X, Ding Y, 2021, Effective fusion factor in fpn for tiny object detection. In: Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision: 1160-1168.
[24]Hong M, Li S, Yang Y, 2022, Sspnet: Scale selection pyramid network for tiny person detection from uav images. IEEE geoscience and remote sensing letters, 19: 1-5.
[25]Li S, Huang D, Wang Y, 2019, Learning spatial fusion for single-shot object detection. arXiv preprint arXiv:1911.09516.
[26]Wang C Y, Liao H Y M, Wu Y H, 2020, CSPNet: A new backbone that can enhance learning capability of CNN//Proceedings of the IEEE/CVF conference on computer vision and pattern recognition workshops. 390-391.
[27]Yu J, Jiang Y, Wang Z, 2016, Unitbox: An advanced object detection network//Proceedings of the 24th ACM international conference on Multimedia. 516-520.
[28]Wang C Y, Yeh I H, Mark L H Y, 2025, Yolov9: Learning what you want to learn using programmable gradient information//European Conference on Computer Vision. Springer, Cham, 1-21.
[29]Wang A, Chen H, Liu L, 2024, Yolov10: Real-time end-to-end object detection. arxiv preprint arxiv: 2405.14458.
[30]Khanam R, Hussain M, 2024, Yolov11: An overview of the key architectural enhancements. arxiv preprint arxiv:2410.17725.