|تعداد مشاهده مقاله||2,245,300|
|تعداد دریافت فایل اصل مقاله||1,598,086|
|Journal of Electrical and Computer Engineering Innovations (JECEI)|
|مقاله 14، دوره 10، شماره 1، فروردین 2022، صفحه 163-174 اصل مقاله (1.43 M)|
|نوع مقاله: Original Research Paper|
|شناسه دیجیتال (DOI): 10.22061/jecei.2021.7871.443|
|A. Bosaghzadeh ؛ M. Shabani؛ R. Ebrahimpour|
|Artificial Intelligence Department, Faculty of Computer Engineering, Shahid Rajaee Teacher Training University, Tehran, Iran|
|تاریخ دریافت: 14 اردیبهشت 1400، تاریخ بازنگری: 14 تیر 1400، تاریخ پذیرش: 16 شهریور 1400|
|Background and Objectives: Visual attention is a high order cognitive process of human brain which defines where a human observer attends. Dynamic computational visual attention models are modeled on the behavior of the human brain and can predict what areas a human will pay attention to when viewing a scene such as a video. However, several types of computational models have been proposed to provide a better understanding of saliency maps in static and dynamic environments, most of these models are used for specific scenes. In this paper, we propose a model that can generate saliency maps in a variety of dynamic environments with complex scenes.|
Methods: We used a deep learner as a mediating network to combine basic saliency maps with appropriate weighting. Each of these basic saliency maps covers an important feature of human visual attention, and ultimately the final saliency map is very similar to human visual behavior.
Results: The proposed model is run on two datasets and the generated saliency maps are evaluated by different criteria such as ROC, CC, NSS, SIM and KLdiv. The results show that the proposed model has a good performance compared to other similar models.
Conclusion: The proposed model consists of three main parts, including basic saliency maps, gating network, and combinator. This model was implemented on the ETMD dataset and the resulting saliency maps (visual attention areas) were compared with some other models in this field by evaluation criteria and their results were evaluated. The results obtained from the proposed model are acceptable and based on the accepted evaluation criteria in this area, it performs better than similar models.
|Visual Attention؛ Dynamic Visual Attention؛ Bottom-up Attention؛ Visual Saliency؛ Human Eye Fixation|
 S.Treue, “Neural correlates of attention in primate visual cortex,” Trends Neurosci., 24(5): 295-300, 2001.
 L. Itti, C. Koch, E. Niebur, “A model of saliency-based visual attention for rapid scene analysis,” IEEE Trans. Pattern Anal. Mach. Intell., 20(11): 1254-1259, 1998.
 J.K. Tsotsos, S.M. Culhane, W.Y.K. Wai, Y. Lai, N. Davis, F. Nuflo, “Modeling visual attention via selective tuning,” Artif. Intell., 78: 507-545, 1995.
 A. Borji, L. Itti, "State-of-the-art in visual attention modeling," IEEE Trans. Pattern Anal. Mach. Intell., 35(1): 185-207, 2013.
 K. Koch, J. McLean, R. Segev, M.A. Freed, M.J. Berry, V.Balasubramanian, P. Sterling, “How much the eye tells the brain,” Curr. Biol., 16(14): 1428-34, 2006.
 L. Itti, “Models of bottom-up and top-down visual attention,” Ph.D. thesis, California Inst. of Technology, 2000.
 H.Xiaodi, L. Zhang, "Saliency detection: A spectral residual approach." in Proc. 2007 IEEE Conference on Computer Vision and Pattern Recognition: 1-8, 2007.
 B.W. Tatler, “The central fixation bias in scene viewing: selecting an optimal viewing position independently of motor bases and image feature distributions,” J. Vis., 14: 1-17, 2007.
 L. Zhang, M.H. Tong, T.K. Marks, H. Shan, G.W. Cottrell,“SUN: A bayesian framework for saliency using natural statistics,” J. Vis., 8(32): 1-20, 2008.
 N.D.B. Bruce, J.K. Tsotsos, “Saliency based on information maximization,” in Proc. Advances in Neural Information Processing Systems, 2005.
 J. Harel, C. Koch, P. Perona, “Graph-based visual saliency,” in Proc. Advances in Neural Information Processing Systems, 19: 545-552, 2007.
 C. Siagian, L. Itti, “Rapid biologically-inspired scene classification using features shared with visual attention,” IEEE Trans. Pattern Anal. Mach. Intell., 29(2): 300-312, 2007.
 G. Li, Y. Yizhou, "Visual saliency based on multiscale deep features," in Proc. the IEEE Conference on Computer Vision and Pattern Recognition: 5455-5463, 2015.
 R. Anna, P. Napieralski, “The visual attention saliency map for movie retrospection,” Open Phys. 16(1): 188-192, 2018.
 W. Hongfa, X. Zhou, Y. Sun, J. Zhang, Ch. Yan. “Deep fusion based video saliency detection,” J. Visual Commun. Image Represent., 62: 279-285, 2019.
 W.Wang, J. Shen, F. Porikli, “Saliency-aware geodesic video object segmentation,” IEEE CVPR: 3395-3402, 2015.
 W. Wang, J. Shen, R. Yang, F. Porikli, “Saliency-aware Video Object Segmentation,” IEEE Trans. Pattern Anal. Mach. Intell., 40(1): 20-33, 2018.
 S. Meijun, Z. Zhou, D. Zhang, Z. Wang. “Hybrid convolutional neural networks and optical flow for video visual attention prediction,” Multimed. Tool. Appl. 77(22): 29231-29244, 2018.
 K. Petros, P. Maragos. “SUSiNet: See, Understand and Summarize it,” in Proc. the IEEE Conference on Computer Vision and Pattern Recognition Workshops, 2019.
 S. Kolkur, et al. “Human skin detection using RGB, HSV and YCbCr color models”, arXiv preprint arXiv:1708.02694, 2017.
 K. Petros, A. Katsamanis, P. Maragos. “Predicting eyes’ fixations in movie videos: Visual Saliency experiments on a new eye-tracking database," in Proc. International Conference on Engineering Psychology and Cognitive Ergonomics: 183-194, Springer, Cham, 2014.
 Z. Bylinskii, J. Tilke, O. Aude, T. Antonio, D. Frédo, “What do different evaluation metrics tell us about saliency models?,” IEEE Trans. Pattern Anal. Mach. Intell., 41(3) :740-757, 2018.
 A. Mohammadi Anbaran; P. Torkzadeh; R. Ebrahimpour; N. Bagheri. "Fast and efficient hardware implementation of 2D gabor filter for a biologically-inspired visual processing algorithm," J. Electr. Comput. Eng. Innovations, 9(1): 93-102, 2021.
 S. Mohseni; G. Ardeshir; N. Zarei. "Facial expression recognition based on anatomical structure of human face," J. Electr. Comput. Eng. Innovations, 2(2): 77-83, 2014.
 M.R. Pishgoo; M.R. N. Avanaki, R. Ebrahimpour, "The application of multi-layer artificial neural networks in speckle reduction (Methodology)," J. Electr. Comput. Eng. Innovations), 2(1): 37-42, 2014.
 J. Khosravi; M. Shams Esfandabadi; R. Ebrahimpour, "Image registration based on sum of square difference cost function," J. Electr. Comput. Eng. Innovations, 6(2): 273-281, 2018.
 R. Veale, H. Yoshida, “How is visual salience computed in the brain? Insights from behaviour, neurobiology and modelling.” Phil. Trans. R. Soc. B 372, 2017.
 J. Li, Y. Tian, T. Huang, W. Gao, “Probabilistic multi-task learning for visual saliency estimation in video,” Int’l J. Comput. Vision, 90: 150-165, 2010.
 A. Oliva, A. Torralba, M.S. Castelhano, J.M. Henderson, “Top-down control of visual attention in object detection,” in Proc. Int’lConf. Image Processing: 253-256, 2003.
تعداد مشاهده مقاله: 306
تعداد دریافت فایل اصل مقاله: 170