Out of Sight, Out of Mind: A Source-View-Wise Feature Aggregation for
Multi-View Image-Based Rendering
- URL: http://arxiv.org/abs/2206.04906v1
- Date: Fri, 10 Jun 2022 07:06:05 GMT
- Title: Out of Sight, Out of Mind: A Source-View-Wise Feature Aggregation for
Multi-View Image-Based Rendering
- Authors: Geonho Cha, Chaehun Shin, Sungroh Yoon, Dongyoon Wee
- Abstract summary: We propose a source-view-wise feature aggregation method, which facilitates us to find out the consensus in a robust way.
We validate the proposed method on various benchmark datasets, including synthetic and real image scenes.
- Score: 26.866141260616793
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: To estimate the volume density and color of a 3D point in the multi-view
image-based rendering, a common approach is to inspect the consensus existence
among the given source image features, which is one of the informative cues for
the estimation procedure. To this end, most of the previous methods utilize
equally-weighted aggregation features. However, this could make it hard to
check the consensus existence when some outliers, which frequently occur by
occlusions, are included in the source image feature set. In this paper, we
propose a novel source-view-wise feature aggregation method, which facilitates
us to find out the consensus in a robust way by leveraging local structures in
the feature set. We first calculate the source-view-wise distance distribution
for each source feature for the proposed aggregation. After that, the distance
distribution is converted to several similarity distributions with the proposed
learnable similarity mapping functions. Finally, for each element in the
feature set, the aggregation features are extracted by calculating the weighted
means and variances, where the weights are derived from the similarity
distributions. In experiments, we validate the proposed method on various
benchmark datasets, including synthetic and real image scenes. The experimental
results demonstrate that incorporating the proposed features improves the
performance by a large margin, resulting in the state-of-the-art performance.
Related papers
- Learning Spatial Similarity Distribution for Few-shot Object Counting [17.28147599627954]
Few-shot object counting aims to count the number of objects in a query image that belong to the same class as the given exemplar images.
Existing methods compute the similarity between the query image and exemplars in the 2D spatial domain and perform regression to obtain the counting number.
We propose a network learning Spatial Similarity Distribution (SSD) for few-shot object counting, which preserves the spatial structure of exemplar features.
arXiv Detail & Related papers (2024-05-20T04:15:59Z) - Rethinking Polyp Segmentation from an Out-of-Distribution Perspective [37.1338930936671]
We leverage the ability of masked autoencoders -- self-supervised vision transformers trained on a reconstruction task -- to learn in-distribution representations.
We perform out-of-distribution reconstruction and inference, with feature space standardisation to align the latent distribution of the diverse abnormal samples with the statistics of the healthy samples.
Experimental results on six benchmarks show that our model has excellent segmentation performance and generalises across datasets.
arXiv Detail & Related papers (2023-06-13T14:13:16Z) - Explicit Correspondence Matching for Generalizable Neural Radiance
Fields [49.49773108695526]
We present a new NeRF method that is able to generalize to new unseen scenarios and perform novel view synthesis with as few as two source views.
The explicit correspondence matching is quantified with the cosine similarity between image features sampled at the 2D projections of a 3D point on different views.
Our method achieves state-of-the-art results on different evaluation settings, with the experiments showing a strong correlation between our learned cosine feature similarity and volume density.
arXiv Detail & Related papers (2023-04-24T17:46:01Z) - Feature-based Image Matching for Identifying Individual K\=ak\=a [0.0]
This report investigates an unsupervised, feature-based image matching pipeline for the novel application of identifying individual k=ak=a.
Applying with a similarity network for clustering, this addresses a weakness of current supervised approaches to identifying individual birds.
We conclude that feature-based image matching could be used with a similarity network to provide a viable alternative to existing supervised approaches.
arXiv Detail & Related papers (2023-01-17T03:43:19Z) - Multiscale Analysis for Improving Texture Classification [62.226224120400026]
This paper employs the Gaussian-Laplacian pyramid to treat different spatial frequency bands of a texture separately.
We aggregate features extracted from gray and color texture images using bio-inspired texture descriptors, information-theoretic measures, gray-level co-occurrence matrix features, and Haralick statistical features into a single feature vector.
arXiv Detail & Related papers (2022-04-21T01:32:22Z) - LEAD: Self-Supervised Landmark Estimation by Aligning Distributions of
Feature Similarity [49.84167231111667]
Existing works in self-supervised landmark detection are based on learning dense (pixel-level) feature representations from an image.
We introduce an approach to enhance the learning of dense equivariant representations in a self-supervised fashion.
We show that having such a prior in the feature extractor helps in landmark detection, even under drastically limited number of annotations.
arXiv Detail & Related papers (2022-04-06T17:48:18Z) - Contextual Similarity Aggregation with Self-attention for Visual
Re-ranking [96.55393026011811]
We propose a visual re-ranking method by contextual similarity aggregation with self-attention.
We conduct comprehensive experiments on four benchmark datasets to demonstrate the generality and effectiveness of our proposed visual re-ranking method.
arXiv Detail & Related papers (2021-10-26T06:20:31Z) - Out-of-distribution Generalization via Partial Feature Decorrelation [72.96261704851683]
We present a novel Partial Feature Decorrelation Learning (PFDL) algorithm, which jointly optimize a feature decomposition network and the target image classification model.
The experiments on real-world datasets demonstrate that our method can improve the backbone model's accuracy on OOD image classification datasets.
arXiv Detail & Related papers (2020-07-30T05:48:48Z) - Object-Centric Multi-View Aggregation [86.94544275235454]
We present an approach for aggregating a sparse set of views of an object in order to compute a semi-implicit 3D representation in the form of a volumetric feature grid.
Key to our approach is an object-centric canonical 3D coordinate system into which views can be lifted, without explicit camera pose estimation.
We show that computing a symmetry-aware mapping from pixels to the canonical coordinate system allows us to better propagate information to unseen regions.
arXiv Detail & Related papers (2020-07-20T17:38:31Z) - Multi-Person Pose Estimation with Enhanced Feature Aggregation and
Selection [33.15192824888279]
We propose a novel Enhanced Feature Aggregation and Selection network (EFASNet) for multi-person 2D human pose estimation.
Our method can well handle crowded, cluttered and occluded scenes.
Comprehensive experiments demonstrate that the proposed approach outperforms the state-of-the-art methods.
arXiv Detail & Related papers (2020-03-20T08:33:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.