Dr. NAWAZ MehmoodYan, HongHongYan2024-09-072024-09-072020IEEE Transactions on Multimedia, 2020, vol. 23, pp. 2902-2916.1520-92101941-0077http://hdl.handle.net/20.500.11861/10464Most existing saliency methods measure fore- ground saliency by using the contrast of a foreground region to its local context, or boundary priors and spatial compactness. These methods are not powerful enough to extract a precise salient region from noisy and cluttered backgrounds. To evaluate the contrast of salient and background regions effectively, we consider high-level features from both supervised and unsupervised methods. We propose an affinity-based robust background subtraction technique and maximum attention map using a pre-trained convolution neural network. This affinity-based technique uses pixel similarities to propagate the values of salient pixels among foreground and background regions and their union. The salient pixel value controls the foreground and background information by using multiple pixel affinities. The maximum attention map is derived from the convolution neural network using features of the Pooling and Relu layers. This method can detect salient regions from images that have noisy and cluttered backgrounds. Our experimental results demonstrate the effectiveness of the proposed approach on six different saliency data sets and benchmarks and show that it improves the quality of detection beyond current saliency detection methods.enBackground SubtractionDeep FeaturesSaliency DetectionNeural NetworkConvolutional Neural NetworkPooling LayerHigh-level FeaturesBackground RegionsAttention MapPre-trained Convolutional Neural NetworkBackground ClutterSalient RegionsReLU LayerForeground RegionsSubtraction TechniqueDeep Neural NetworkImage SegmentationObject DetectionMean Absolute ErrorTop-Down ApproachSaliency MapSalient ObjectDense ReconstructionForeground PixelsTechnical ResultsSimilar PixelsMultiple ObjectsBottom-Up ApproachPR CurveBackground PixelsSaliency detection using deep features and affinity-based robust background subtractionPeer Reviewed Journal Article10.1109/TMM.2020.3019688