Modeling Object Dissimilarity for Deep Saliency Prediction
- URL: http://arxiv.org/abs/2104.03864v1
- Date: Thu, 8 Apr 2021 16:10:37 GMT
- Title: Modeling Object Dissimilarity for Deep Saliency Prediction
- Authors: Bahar Aydemir, Deblina Bhattacharjee, Seungryong Kim, Tong Zhang,
Mathieu Salzmann and Sabine S\"usstrunk
- Abstract summary: We introduce a detection-guided saliency prediction network that explicitly models the differences between multiple objects.
Our approach is general, allowing us to fuse our object dissimilarities with features extracted by any deep saliency prediction network.
- Score: 86.14710352178967
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Saliency prediction has made great strides over the past two decades, with
current techniques modeling low-level information, such as color, intensity and
size contrasts, and high-level one, such as attention and gaze direction for
entire objects. Despite this, these methods fail to account for the
dissimilarity between objects, which humans naturally do. In this paper, we
introduce a detection-guided saliency prediction network that explicitly models
the differences between multiple objects, such as their appearance and size
dissimilarities. Our approach is general, allowing us to fuse our object
dissimilarities with features extracted by any deep saliency prediction
network. As evidenced by our experiments, this consistently boosts the accuracy
of the baseline networks, enabling us to outperform the state-of-the-art models
on three saliency benchmarks, namely SALICON, MIT300 and CAT2000.
Related papers
- Human-Object Interaction Detection Collaborated with Large Relation-driven Diffusion Models [65.82564074712836]
We introduce DIFfusionHOI, a new HOI detector shedding light on text-to-image diffusion models.
We first devise an inversion-based strategy to learn the expression of relation patterns between humans and objects in embedding space.
These learned relation embeddings then serve as textual prompts, to steer diffusion models generate images that depict specific interactions.
arXiv Detail & Related papers (2024-10-26T12:00:33Z) - Stanceformer: Target-Aware Transformer for Stance Detection [59.69858080492586]
Stance Detection involves discerning the stance expressed in a text towards a specific subject or target.
Prior works have relied on existing transformer models that lack the capability to prioritize targets effectively.
We introduce Stanceformer, a target-aware transformer model that incorporates enhanced attention towards the targets during both training and inference.
arXiv Detail & Related papers (2024-10-09T17:24:28Z) - Adversarial Robustness Assessment of NeuroEvolution Approaches [1.237556184089774]
We evaluate the robustness of models found by two NeuroEvolution approaches on the CIFAR-10 image classification task.
Our results show that when the evolved models are attacked with iterative methods, their accuracy usually drops to, or close to, zero.
Some of these techniques can exacerbate the perturbations added to the original inputs, potentially harming robustness.
arXiv Detail & Related papers (2022-07-12T10:40:19Z) - Self-attention Presents Low-dimensional Knowledge Graph Embeddings for
Link Prediction [6.789370732159177]
Self-attention is the key to applying query-dependant projections to entities and relations.
Our model achieves favorably comparable or better performance than our three best recent state-of-the-art competitors.
arXiv Detail & Related papers (2021-12-20T16:11:01Z) - How Well Do Sparse Imagenet Models Transfer? [75.98123173154605]
Transfer learning is a classic paradigm by which models pretrained on large "upstream" datasets are adapted to yield good results on "downstream" datasets.
In this work, we perform an in-depth investigation of this phenomenon in the context of convolutional neural networks (CNNs) trained on the ImageNet dataset.
We show that sparse models can match or even outperform the transfer performance of dense models, even at high sparsities.
arXiv Detail & Related papers (2021-11-26T11:58:51Z) - Bayesian deep learning of affordances from RGB images [5.939410304994348]
We present a deep learning method to predict the affordances available in the environment directly from RGB images.
Based on previous work on socially accepted affordances, our model is based on a multiscale CNN that combines local and global information from the object and the full image.
Our results show a marginal better performance of deep ensembles, compared to MC-dropout on the Brier score and the Expected Error.
arXiv Detail & Related papers (2021-09-27T07:39:47Z) - Recursive Contour Saliency Blending Network for Accurate Salient Object
Detection [0.0]
In this work, we designed a network for better edge quality in salient object detection.
We proposed a contour-saliency blending module to exchange information between contour and saliency.
Our model is lightweight and fast, with only 27.9 million parameters and real-time inferencing at 31 FPS.
arXiv Detail & Related papers (2021-05-28T14:19:54Z) - Contemplating real-world object classification [53.10151901863263]
We reanalyze the ObjectNet dataset recently proposed by Barbu et al. containing objects in daily life situations.
We find that applying deep models to the isolated objects, rather than the entire scene as is done in the original paper, results in around 20-30% performance improvement.
arXiv Detail & Related papers (2021-03-08T23:29:59Z) - Firearm Detection via Convolutional Neural Networks: Comparing a
Semantic Segmentation Model Against End-to-End Solutions [68.8204255655161]
Threat detection of weapons and aggressive behavior from live video can be used for rapid detection and prevention of potentially deadly incidents.
One way for achieving this is through the use of artificial intelligence and, in particular, machine learning for image analysis.
We compare a traditional monolithic end-to-end deep learning model and a previously proposed model based on an ensemble of simpler neural networks detecting fire-weapons via semantic segmentation.
arXiv Detail & Related papers (2020-12-17T15:19:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.