Deep Learning for Detecting Abandoned Dogs
DOI:
https://doi.org/10.9781/ijimai.2026.2222Keywords:
Activity Recognition, Algorithms, Anomaly Detection, Deep Learning, Edge Computing, Machine Learning, Machine VisionAbstract
This research paper presents a methodology consisting of an algorithm and a workflow for finding abandoned dogs in natural surroundings. We propose a temporal-contextual methodology for identifying abandoned pets in public areas. This involves employing a temporal rule alongside context-sensitive object identification, whereby dog bounding boxes are deliberately expanded to encompass proximate visual indicators suggestive of abandonment. The proposed approach uses object detection techniques, trajectory analysis, and image segmentation to quickly differentiate between abandoned and owned dogs. The research addresses key challenges, such as data scarcity and the complexity of distinguishing between abandoned and accompanied dogs. To address the issue of sufficient and adequate training corpus to identify an abandoned dog, the research article employs single-channel image augmentation methods that improve model recall and precision by 4%. Several object detection algorithms were evaluated, and our findings indicate that single-stage detectors like YOLOv8 achieved a better trade-off between classification performance and speed for detecting abandoned dogs compared to multi-stage detectors like Faster R-CNN, reaching a mean average precision up to 86% and an inference time of 0.3 ms per frame. This research study contributes to animal welfare, biodiversity conservation, and public safety by providing a scalable solution for monitoring abandoned animals in diverse environments. The findings demonstrate the impact of object detection techniques on improving the generalization of deep learning models for real-world applications.
Downloads
References
[1] W. Hu, T. Tan, L. Wang, S. Maybank, “A survey on visual surveillance of object motion and behaviors,” IEEE Transactions on Systems, Man, and Cybernetics, Part C (Applications and Reviews), vol. 34, no. 3, pp. 334–352, 2004, doi: https://doi.org/10.1109/TSMCC.2004.829274.
[2] J. Redmon, A. Farhadi, “Yolov3: An incremental improvement,” arXiv preprint arXiv:1804.02767, 2018, doi: https://doi.org/10.48550/arXiv.1804.02767.
[3] J. Hughes, D. W. Macdonald, “A review of the interactions between freeroaming domestic dogs and wildlife,” Biological Conservation, vol. 157, pp. 341–351, 2013, doi: https://doi.org/10.1016/j.biocon.2012.07.005.
[4] World Health Organization, “Rabies,” 2023. Accessed: May 28, 2025, [Online]. Available: https://www.who.int/news-room/factsheets/detail/rabies.
[5] K. Hampson, L. Coudeville, T. Lembo, M. Sambo, A. Kieffer, M. Attlan, J. Barrat, J. D. Blanton, D. J. Briggs, S. Cleaveland, P. Costa, C. M. Freuling, E. Hiby, L. Knopf, F. Leanes, F.-X. Meslin, A. Metlin, M. E. Miranda, T. Müller, L. H. Nel, S. Recuenco, C. E. Rupprecht, C. Schumacher, L. Taylor, M. A. N. Vigilato, J. Zinsstag, J. Dushoff, “Estimating the global burden of endemic canine rabies,” PLoS Neglected Tropical Diseases, vol. 9, no. 4, p. e0003709, 2015, doi: https://doi.org/10.1371/journal.pntd.0003709.
[6] G. Jocher, “Yolov5 by ultralytics,” 2020. doi: https://doi.org/10.5281/zenodo.3908559.
[7] C. Li, L. Li, H. Jiang, K. Weng, Y. Geng, L. Li, Z. Ke, Q. Li, M. Cheng, W. Nie, Y. Li, B. Zhang, Y. Liang, L. Zhou, X. Xu, X. Chu, X. Wei, X. Wei, “Yolov6: A single-stage object detection framework for industrial applications,” arXiv preprint arXiv:2209.02976, 2022, doi: https://doi.org/10.48550/arXiv.2209.02976.
[8] R. Girshick, “Fast r-cnn,” in Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 2015, pp. 1440–1448.
[9] S. Ren, K. He, R. Girshick, J. Sun, “Faster rcnn: Towards real-time object detection with region proposal networks,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 39, no. 6, pp. 1137–1149, 2017, doi: https://doi.org/10.1109/TPAMI.2016.2577031.
[10] R. Girshick, I. Radosavovic, G. Gkioxari, P. Dollár, K. He, “Detectron,” 2018. Accessed: May 28, 2025, [Online]. Available: https://github.com/facebookresearch/detectron.
[11] Y. Wu, A. Kirillov, F. Massa, W.-Y. Lo, R. Girshick, “Detectron2,” 2019. Accessed: May 28, 2025, [Online]. Available: https://github.com/facebookresearch/detectron2.
[12] L. Ferrando, G. Gera, C. Regazzoni, “Classification of unattended and stolen objects in videosurveillance systems,” in Proceedings of the 2006 IEEE International Conference on Video and Signal Based Surveillance, Sydney, Australia, 2006, pp. 21– 21.
[13] F. I. Bashir, A. A. Khokhar, D. Schonfeld, “Realtime motion trajectorybased indexing and retrieval of video sequences,” IEEE Transactions on Multimedia, vol. 9, no. 1, pp. 58–65, 2007, doi: https://doi.org/10.1109/TMM.2006.886346.
[14] L. Taylor, G. Nitschke, “Improving deep learning using generic data augmentation,” arXiv preprint arXiv:1708.06020, 2017, doi: https://doi.org/10.48550/arXiv.1708.06020.
[15] C. Shorten, T. M. Khoshgoftaar, “A survey on image data augmentation for deep learning,” Journal of Big Data, vol. 6, no. 1, p. 60, 2019, doi: https://doi.org/10.1186/s40537-019-0197-0.
[16] CIMNE, “PIKSEL: Portal for the Integration of Knowledge for Sustainable Ecosystems and Land Management,” 2024. Accessed: May 28, 2025, [Online]. Available: https://pikselweb.cimne.com/.
[17] B. Benjdira, T. Khursheed, A. Koubaa, A. Ammar, K. Ouni, “Car detection using unmanned aerial vehicles: Comparison between faster r-cnn and yolov3,” arXiv preprint arXiv:1812.10968, 2018, doi: https://doi.org/10.48550/arXiv.1812.10968.
[18] E. D. Cubuk, B. Zoph, D. Mane, V. Vasudevan, Q. V. Le, “Autoaugment: Learning augmentation policies from data,” arXiv preprint arXiv:1805.09501, 2019, doi: https://doi.org/10.48550/arXiv.1805.09501.
[19] S. Lim, I. Kim, T. Kim, C. Kim, S. Kim, “Fast autoaugment,” in Proceedings of the 33rd International Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 2019.
[20] G. Gibbins, G. Newsam, M. Brooks, “Detecting suspicious background changes in video surveillance of busy scenes,” in Proceedings of the Third IEEE Workshop on Applications of Computer Vision (WACV’96), Sarasota, FL, USA, 1996, pp. 22–26.
[21] A. H. Freedman, I. Gronau, R. M. Schweizer, D. Ortega-Del Vecchyo, E. Han, P. M. Silva, et al., “Genome sequencing highlights the dynamic early history of dogs,” PLoS Genetics, vol. 10, no. 1, p. e1004016, 2014, doi: https://doi.org/10.1371/journal.pgen.1004016.
[22] M. E. Gompper, “The dog–human–wildlife interface: Assessing the scope of the problem,” in Free-Ranging Dogs and Wildlife Conservation, Oxford University Press, 2013.
[23] S. Barnard, C. Ippoliti, D. Di Flaviano, A. De Ruvo, S. Messori, A. Giovannini, P. Dalla Villa, “Smartphone and gps technology for freeroaming dog population surveillance - a methodological study,” Veterinaria Italiana, vol. 51, no. 3, pp. 165–172, 2015, doi: https://doi.org/10.12834/VetIt.233.2163.3.
[24] P. Howell, At Home and Astray: The Domestic Dog in Victorian Britain. Charlottesville, VA, USA: University of Virginia Press, 2015.
[25] O. Akinwehinmi, P. Arnau, F. Solsona, A. Priegue, J. Jimenez, “Paws for concern: Spotting strays and abandoned pets in natural spaces,” in 2023 IEEE International Smart Cities Conference (ISC2), Bucharest, Romania, 2023, pp. 1–7.
[26] J. Ruiz-Chavez, J. Salvador-Meneses, C. MejíaAstudillo, S. DiazQuilachamin, “Analysis of dogs’ abandonment problem using georeferenced multiagent systems,” in From Bioinspired Systems and Biomedical Applications to Machine Learning, Cham, Switzerland: Springer International Publishing, 2019, pp. 297–306.
[27] B. Graham, “Fractional max-pooling,” arXiv preprint arXiv:1412.6071, 2015, doi: https://doi.org/10.48550/arXiv.1412.6071.
[28] Ruman, “Yolo data augmentation explained: Turbocharge your object detection model,” 2023. Accessed: May 28, 2025, [Online]. Available: https://www.rumn.ai/blog/yolodata-augmentation-explained.
[29] J. Deng, W. Dong, R. Socher, L.-J. Li, K. Li, F.F. Li, “Imagenet: A large-scale hierarchical image database,” in Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Miami, FL, USA, 2009, pp. 248–255.
[30] O. Akinwehinmi, P. Arnau, F. Solsona, “Improving abandoned pet detection in natural spaces through adaptive single-coloured image augmentation techniques,” in Proceedings of the 15th Congress of the International Colour Association (AIC 2023), Chiang Rai, Thailand, 2023.
[31] T. DeVries, G. W. Taylor, “Improved regularization of convolutional neural networks with cutout,” arXiv preprint arXiv:1708.04552, 2017, doi: https://doi.org/10.48550/arXiv.1708.04552.
[32] H. Zhang, M. Cisse, Y. N. Dauphin, D. Lopez-Paz, “Mixup: Beyond empirical risk minimization,” arXiv preprint arXiv:1710.09412, 2018, doi: https://doi.org/10.48550/arXiv.1710.09412.
[33] Y. Guo, Y. Mao, R. Zhang, “Mixup as locally linear out-of-manifold regularization,” in Proceedings of the AAAI Conference on Artificial Intelligence, vol. 33, Honolulu, HI, USA, 2019, pp. 3714–3722.
[34] S. Yun, D. Han, S. Chun, S. J. Oh, Y. Yoo, J. Choe, “Cutmix: Regularization strategy to train strong classifiers with localizable features,” in Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, South Korea, 2019, pp. 6022–6031.
[35] A. Bruno, E. Ardizzone, S. Vitabile, M. Midiri, “A novel solution based on scale invariant feature transform descriptors and deep learning for the detection of suspicious regions in mammogram images,” Journal of Medical Signals and Sensors, vol. 10, no. 3, pp. 158–173, 2020, doi: https://doi.org/10.4103/jmss.JMSS_31_19.
[36] B. Recht, R. Roelofs, L. Schmidt, V. Shankar, “Do cifar-10 classifiers generalize to cifar10?,” arXiv preprint arXiv:1806.00451, 2018, doi: https://doi.org/10.48550/arXiv.1806.00451.
[37] A. Jurio, M. Pagola, “ M. Galar, C. LopezMolina, D. Paternain, A comparison study of different color spaces in clustering based image segmentation,” in Information Processing and Management of Uncertainty in Knowledge-Based Systems. Applications, Berlin, Heidelberg: Springer Berlin Heidelberg, 2010, pp. 532–541.
[38] Y. Bengio, F. Bastien, A. Bergeron, N. Boulanger–Lewandowski, T. Breuel, Y. Chherawala, M. Cisse, M. Côté, D. Erhan, J. Eustache, X. Glorot, X. Muller, S. Pannetier Lebeuf, R. Pascanu, S. Rifai, F. Savard, G.
Sicard, “Deep learners benefit more from out-of-distribution examples,” in Proceedings of the Fourteenth International Conference on Artificial Intelligence and Statistics (AISTATS), vol. 15 of Proceedings of Machine Learning Research, Fort Lauderdale, FL, USA, 2011, pp. 164–172.
[39] E. Azizi, L. Zaman, “Deep learning pet identification using face and body,” Information, vol. 14, no. 5, 2023, doi: https://doi.org/10.3390/info14050278.
[40] P. Kasnesis, V. Doulgerakis, D. Uzunidis, D. G. Kogias, S. I. Funcia, M. B. González, C. Giannousis, C. Z. Patrikakis, “Deep learning empowered wearable-based behavior recognition for search and rescue dogs,” Sensors, vol. 22, no. 3, 2022, doi: https://doi.org/10.3390/s22030993.
[41] Y. Sangve, Y. Firke, S. Shinde, S. Patil, P. Shinde, P. Mitake, “A comprehensive survey on real-time animal (dog) detection system using artificial intelligence methods,” in Proceedings of the 4th International Conference on Artificial Intelligence and Smart Energy, Cham, Switzerland: Springer Nature Switzerland, 2024, pp. 260–275.
[42] O. Akinwehinmi, “Abandoned dogs detection dataset,” 2025. Accessed: Jun. 3, 2025, [Online]. Available: https://www.kaggle. com/datasets/oluwakemiakinwehinmi/ abandoned-dogs-detection.
[43] O. Akinwehinmi, “Abandoned dog detection - code for reproduction,” 2025. Accessed: Jun. 3, 2025, [Online]. Available: https://github.com/DrKem/Abandoned-DogDetection/blob/main/Codes
[44] B. D. Lucas, T. Kanade, “An iterative image registration technique with an application to stereo vision,” in Proceedings of the 7th International Joint Conference on Artificial Intelligence (IJCAI’81), San Francisco, CA, USA, 1981, pp. 674–679.
[45] B. K. Horn, B. G. Schunck, “Determining optical flow,” Artificial Intelligence, vol. 17, no. 1, pp. 185– 203, 1981, doi: https://doi.org/10.1016/00043702(81)90024-2.
[46] A. Bewley, Z. Ge, L. Ott, F. Ramos, B. Upcroft, “Simple online and realtime tracking,” in Proceedings of the 2016 IEEE International Conference on Image Processing (ICIP), Phoenix, AZ, USA, 2016, pp. 3464–3468.
[47] N. Wojke, A. Bewley, D. Paulus, “Simple online and realtime tracking with a deep association metric,” in Proceedings of the 2017 IEEE International Conference on Image Processing (ICIP), Beijing, China, 2017,pp. 3645–3649.
[48] K. Chatfield, A. Zisserman, “Visor: Towards onthe-fly large-scale object category retrieval,” in Computer Vision – ACCV 2012, Berlin, Heidelberg: Springer Berlin Heidelberg, 2013, pp. 432–446.
[49] M. Everingham, L. Van Gool, C. K. I. Williams, J. Winn, A. Zisserman, “The pascal visual object classes (voc) challenge,” International Journal of Computer Vision, vol. 88, no. 2, pp. 303–338, 2010, doi: https://doi.org/10.1007/s11263-009-0275-4.
Downloads
Published
-
Abstract49
-
PDF36






