Saliency detection in deep learning era: Trends of development

State University of Aerospace Instrumentation (SUAI)
Publication Type:
Journal Article
Informatsionno-Upravliaiushchie Sistemy, 2019, (3), pp. 10-36
Issue Date:
Filename Description Size
ILL Article 2 to Scott M_SALIENCY DETECTION IN DEEP LEARNING ERA ... 9.6.20.pdfPublished version2.44 MB
Adobe PDF
Full metadata record
© 2019 Saint Petersburg State University of Aerospace Instrumentation. All rights reserved. Introduction: Saliency detection is a fundamental task of computer vision. Its ultimate aim is to localize the objects of interest that grab human visual attention with respect to the rest of the image. A great variety of saliency models based on different approaches was developed since 1990s. In recent years, the saliency detection has become one of actively studied topic in the theory of Convolutional Neural Network (CNN). Many original decisions using CNNs were proposed for salient object detection and, even, event detection. Purpose: A detailed survey of saliency detection methods in deep learning era allows to understand the current possibilities of CNN approach for visual analysis conducted by the human eyes' tracking and digital image processing. Results: A survey reflects the recent advances in saliency detection using CNNs. Different models available in literature, such as static and dynamic 2D CNNs for salient object detection and 3D CNNs for salient event detection are discussed in the chronological order. It is worth noting that automatic salient event detection in durable videos became possible using the recently appeared 3D CNN combining with 2D CNN for salient audio detection. Also in this article, we have presented a short description of public image and video datasets with annotated salient objects or events, as well as the often used metrics for the results' evaluation. Practical relevance: This survey is considered as a contribution in the study of rapidly developed deep learning methods with respect to the saliency detection in the images and videos.
Please use this identifier to cite or link to this item: