COVINS-G: A Generic Back-end for Collaborative Visual-Inertial SLAM
- URL: http://arxiv.org/abs/2301.07147v3
- Date: Fri, 5 May 2023 08:17:00 GMT
- Title: COVINS-G: A Generic Back-end for Collaborative Visual-Inertial SLAM
- Authors: Manthan Patel, Marco Karrer, Philipp B\"anninger and Margarita Chli
- Abstract summary: Collaborative SLAM is at the core of perception in multi-robot systems.
CoVINS-G is a generalized back-end building upon the COVINS framework.
We show on-par accuracy with state-of-the-art multi-session and collaborative SLAM systems.
- Score: 13.190581566723917
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Collaborative SLAM is at the core of perception in multi-robot systems as it
enables the co-localization of the team of robots in a common reference frame,
which is of vital importance for any coordination amongst them. The paradigm of
a centralized architecture is well established, with the robots (i.e. agents)
running Visual-Inertial Odometry (VIO) onboard while communicating relevant
data, such as e.g. Keyframes (KFs), to a central back-end (i.e. server), which
then merges and optimizes the joint maps of the agents. While these frameworks
have proven to be successful, their capability and performance are highly
dependent on the choice of the VIO front-end, thus limiting their flexibility.
In this work, we present COVINS-G, a generalized back-end building upon the
COVINS framework, enabling the compatibility of the server-back-end with any
arbitrary VIO front-end, including, for example, off-the-shelf cameras with
odometry capabilities, such as the Realsense T265. The COVINS-G back-end
deploys a multi-camera relative pose estimation algorithm for computing the
loop-closure constraints allowing the system to work purely on 2D image data.
In the experimental evaluation, we show on-par accuracy with state-of-the-art
multi-session and collaborative SLAM systems, while demonstrating the
flexibility and generality of our approach by employing different front-ends
onboard collaborating agents within the same mission. The COVINS-G codebase
along with a generalized front-end wrapper to allow any existing VIO front-end
to be readily used in combination with the proposed collaborative back-end is
open-sourced. Video: https://youtu.be/FoJfXCfaYDw
Related papers
- Unveiling the Backbone-Optimizer Coupling Bias in Visual Representation Learning [54.956037293979506]
This paper delves into the interplay between vision backbones and vision backbones and their inter-dependent phenomenon termed textittextbfbackbonetextbfoptimizer textbfcoupling textbfbias (BOCB)
We observe that canonical CNNs, such as VGG and ResNet, exhibit a marked co-dependency with SGD families, while recent architectures like ViTs and ConvNeXt share a tight coupling with the adaptive learning rate ones.
arXiv Detail & Related papers (2024-10-08T21:14:23Z) - IFTR: An Instance-Level Fusion Transformer for Visual Collaborative Perception [9.117534139771738]
Multi-agent collaborative perception has emerged as a widely recognized technology in the field of autonomous driving.
Current collaborative perception predominantly relies on LiDAR point clouds, with significantly less attention given to methods using camera images.
This work proposes an instance-level fusion transformer for visual collaborative perception.
arXiv Detail & Related papers (2024-07-13T11:38:15Z) - Self-Localized Collaborative Perception [49.86110931859302]
We propose$mathttCoBEVGlue$, a novel self-localized collaborative perception system.
$mathttCoBEVGlue$ is a novel spatial alignment module, which provides the relative poses between agents.
$mathttCoBEVGlue$ achieves state-of-the-art detection performance under arbitrary localization noises and attacks.
arXiv Detail & Related papers (2024-06-18T15:26:54Z) - What Makes Good Collaborative Views? Contrastive Mutual Information Maximization for Multi-Agent Perception [52.41695608928129]
Multi-agent perception (MAP) allows autonomous systems to understand complex environments by interpreting data from multiple sources.
This paper investigates intermediate collaboration for MAP with a specific focus on exploring "good" properties of collaborative view.
We propose a novel framework named CMiMC for intermediate collaboration.
arXiv Detail & Related papers (2024-03-15T07:18:55Z) - APGL4SR: A Generic Framework with Adaptive and Personalized Global
Collaborative Information in Sequential Recommendation [86.29366168836141]
We propose a graph-driven framework, named Adaptive and Personalized Graph Learning for Sequential Recommendation (APGL4SR)
APGL4SR incorporates adaptive and personalized global collaborative information into sequential recommendation systems.
As a generic framework, APGL4SR can outperform other baselines with significant margins.
arXiv Detail & Related papers (2023-11-06T01:33:24Z) - Collaborative Multi-Agent Video Fast-Forwarding [30.843484383185473]
We develop two collaborative multi-agent video fast-forwarding frameworks in distributed and centralized settings.
In these frameworks, each individual agent can selectively process or skip video frames at adjustable paces based on multiple strategies.
We show that compared with other approaches in the literature, our frameworks achieve better coverage of important frames, while significantly reducing the number of frames processed at each agent.
arXiv Detail & Related papers (2023-05-27T20:12:19Z) - Plug-and-Play Regulators for Image-Text Matching [76.28522712930668]
Exploiting fine-grained correspondence and visual-semantic alignments has shown great potential in image-text matching.
We develop two simple but quite effective regulators which efficiently encode the message output to automatically contextualize and aggregate cross-modal representations.
Experiments on MSCOCO and Flickr30K datasets validate that they can bring an impressive and consistent R@1 gain on multiple models.
arXiv Detail & Related papers (2023-03-23T15:42:05Z) - COVINS: Visual-Inertial SLAM for Centralized Collaboration [11.65456841016608]
Collaborative SLAM enables a group of agents to simultaneously co-localize and jointly map an environment.
This article presents COVINS, a novel collaborative SLAM system, that enables multi-agent, scalable SLAM in large environments.
arXiv Detail & Related papers (2021-08-12T13:50:44Z) - A Flexible Framework for Designing Trainable Priors with Adaptive
Smoothing and Game Encoding [57.1077544780653]
We introduce a general framework for designing and training neural network layers whose forward passes can be interpreted as solving non-smooth convex optimization problems.
We focus on convex games, solved by local agents represented by the nodes of a graph and interacting through regularization functions.
This approach is appealing for solving imaging problems, as it allows the use of classical image priors within deep models that are trainable end to end.
arXiv Detail & Related papers (2020-06-26T08:34:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.