A Visual Active Search Framework for Geospatial Exploration
- URL: http://arxiv.org/abs/2211.15788v3
- Date: Sun, 29 Oct 2023 20:24:10 GMT
- Title: A Visual Active Search Framework for Geospatial Exploration
- Authors: Anindya Sarkar, Michael Lanier, Scott Alfeld, Jiarui Feng, Roman
Garnett, Nathan Jacobs, Yevgeniy Vorobeychik
- Abstract summary: Many problems can be viewed as forms of geospatial search aided by aerial imagery.
We model this class of problems in a visual active search (VAS) framework, which has three key inputs.
We propose a reinforcement learning approach for VAS that learns a meta-search policy from a collection of fully annotated search tasks.
- Score: 36.31732056074638
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Many problems can be viewed as forms of geospatial search aided by aerial
imagery, with examples ranging from detecting poaching activity to human
trafficking. We model this class of problems in a visual active search (VAS)
framework, which has three key inputs: (1) an image of the entire search area,
which is subdivided into regions, (2) a local search function, which determines
whether a previously unseen object class is present in a given region, and (3)
a fixed search budget, which limits the number of times the local search
function can be evaluated. The goal is to maximize the number of objects found
within the search budget. We propose a reinforcement learning approach for VAS
that learns a meta-search policy from a collection of fully annotated search
tasks. This meta-search policy is then used to dynamically search for a novel
target-object class, leveraging the outcome of any previous queries to
determine where to query next. Through extensive experiments on several
large-scale satellite imagery datasets, we show that the proposed approach
significantly outperforms several strong baselines. We also propose novel
domain adaptation techniques that improve the policy at decision time when
there is a significant domain gap with the training data. Code is publicly
available.
Related papers
- Local Feature Matching Using Deep Learning: A Survey [19.322545965903608]
Local feature matching enjoys wide-ranging applications in the realm of computer vision, encompassing domains such as image retrieval, 3D reconstruction, and object recognition.
In recent years, the introduction of deep learning models has sparked widespread exploration into local feature matching techniques.
The paper also explores the practical application of local feature matching in diverse domains such as Structure from Motion, Remote Sensing Image Registration, and Medical Image Registration.
arXiv Detail & Related papers (2024-01-31T04:32:41Z) - A Partially Supervised Reinforcement Learning Framework for Visual
Active Search [36.966522001393734]
Visual active search (VAS) has been proposed as a modeling framework in which visual cues are used to guide exploration.
We propose an approach that combines the strength of both DRL and conventional active search by decomposing the search policy into a prediction module.
We develop a novel meta-learning approach for jointly learning the resulting combined policy that can make effective use of supervised information obtained both at training and decision time.
arXiv Detail & Related papers (2023-10-15T00:29:35Z) - FORB: A Flat Object Retrieval Benchmark for Universal Image Embedding [7.272083488859574]
We introduce a new dataset for benchmarking visual search methods on flat images with diverse patterns.
Our flat object retrieval benchmark (FORB) supplements the commonly adopted 3D object domain.
It serves as a testbed for assessing the image embedding quality on out-of-distribution domains.
arXiv Detail & Related papers (2023-09-28T08:41:51Z) - RF-Next: Efficient Receptive Field Search for Convolutional Neural
Networks [86.6139619721343]
We propose to find better receptive field combinations through a global-to-local search scheme.
Our search scheme exploits both global search to find the coarse combinations and local search to get the refined receptive field combinations.
Our RF-Next models, plugging receptive field search to various models, boost the performance on many tasks.
arXiv Detail & Related papers (2022-06-14T06:56:26Z) - CrossBeam: Learning to Search in Bottom-Up Program Synthesis [51.37514793318815]
We propose training a neural model to learn a hands-on search policy for bottom-up synthesis.
Our approach, called CrossBeam, uses the neural model to choose how to combine previously-explored programs into new programs.
We observe that CrossBeam learns to search efficiently, exploring much smaller portions of the program space compared to the state-of-the-art.
arXiv Detail & Related papers (2022-03-20T04:41:05Z) - Global-Local Context Network for Person Search [125.51080862575326]
Person search aims to jointly localize and identify a query person from natural, uncropped images.
We exploit rich context information globally and locally surrounding the target person, which we refer to scene and group context, respectively.
We propose a unified global-local context network (GLCNet) with the intuitive aim of feature enhancement.
arXiv Detail & Related papers (2021-12-05T07:38:53Z) - Exposing Query Identification for Search Transparency [69.06545074617685]
We explore the feasibility of approximate exposing query identification (EQI) as a retrieval task by reversing the role of queries and documents in two classes of search systems.
We derive an evaluation metric to measure the quality of a ranking of exposing queries, as well as conducting an empirical analysis focusing on various practical aspects of approximate EQI.
arXiv Detail & Related papers (2021-10-14T20:19:27Z) - Nonmyopic Multifidelity Active Search [15.689830609697685]
We propose a model of multifidelity active search, as well as a novel, computationally efficient policy for this setting.
We evaluate the performance of our solution on real-world datasets and demonstrate significantly better performance than natural benchmarks.
arXiv Detail & Related papers (2021-06-11T12:55:51Z) - Addressing Visual Search in Open and Closed Set Settings [8.928169373673777]
We present a method for predicting pixel-level objectness from a low resolution gist image.
We then use to select regions for performing object detection locally at high resolution.
Second, we propose a novel strategy for open-set visual search that seeks to find all instances of a target class which may be previously unseen.
arXiv Detail & Related papers (2020-12-11T17:21:28Z) - Tasks Integrated Networks: Joint Detection and Retrieval for Image
Search [99.49021025124405]
In many real-world searching scenarios (e.g., video surveillance), the objects are seldom accurately detected or annotated.
We first introduce an end-to-end Integrated Net (I-Net), which has three merits.
We further propose an improved I-Net, called DC-I-Net, which makes two new contributions.
arXiv Detail & Related papers (2020-09-03T03:57:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.