OpenBot: Turning Smartphones into Robots
- URL: http://arxiv.org/abs/2008.10631v2
- Date: Wed, 10 Mar 2021 19:08:00 GMT
- Title: OpenBot: Turning Smartphones into Robots
- Authors: Matthias M\"uller, Vladlen Koltun
- Abstract summary: Current robots are either expensive or make significant compromises on sensory richness, computational power, and communication capabilities.
We propose to leverage smartphones to equip robots with extensive sensor suites, powerful computational abilities, state-of-the-art communication channels, and access to a thriving software ecosystem.
We design a small electric vehicle that costs $50 and serves as a robot body for standard Android smartphones.
- Score: 95.94432031144716
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Current robots are either expensive or make significant compromises on
sensory richness, computational power, and communication capabilities. We
propose to leverage smartphones to equip robots with extensive sensor suites,
powerful computational abilities, state-of-the-art communication channels, and
access to a thriving software ecosystem. We design a small electric vehicle
that costs $50 and serves as a robot body for standard Android smartphones. We
develop a software stack that allows smartphones to use this body for mobile
operation and demonstrate that the system is sufficiently powerful to support
advanced robotics workloads such as person following and real-time autonomous
navigation in unstructured environments. Controlled experiments demonstrate
that the presented approach is robust across different smartphones and robot
bodies. A video of our work is available at
https://www.youtube.com/watch?v=qc8hFLyWDOM
Related papers
- $π_0$: A Vision-Language-Action Flow Model for General Robot Control [77.32743739202543]
We propose a novel flow matching architecture built on top of a pre-trained vision-language model (VLM) to inherit Internet-scale semantic knowledge.
We evaluate our model in terms of its ability to perform tasks in zero shot after pre-training, follow language instructions from people, and its ability to acquire new skills via fine-tuning.
arXiv Detail & Related papers (2024-10-31T17:22:30Z) - Unifying 3D Representation and Control of Diverse Robots with a Single Camera [48.279199537720714]
We introduce Neural Jacobian Fields, an architecture that autonomously learns to model and control robots from vision alone.
Our approach achieves accurate closed-loop control and recovers the causal dynamic structure of each robot.
arXiv Detail & Related papers (2024-07-11T17:55:49Z) - Open-TeleVision: Teleoperation with Immersive Active Visual Feedback [17.505318269362512]
Open-TeleVision allows operators to actively perceive the robot's surroundings in a stereoscopic manner.
The system mirrors the operator's arm and hand movements on the robot, creating an immersive experience.
We validate the effectiveness of our system by collecting data and training imitation learning policies on four long-horizon, precise tasks.
arXiv Detail & Related papers (2024-07-01T17:55:35Z) - HumanoidBench: Simulated Humanoid Benchmark for Whole-Body Locomotion and Manipulation [50.616995671367704]
We present a high-dimensional, simulated robot learning benchmark, HumanoidBench, featuring a humanoid robot equipped with dexterous hands.
Our findings reveal that state-of-the-art reinforcement learning algorithms struggle with most tasks, whereas a hierarchical learning approach achieves superior performance when supported by robust low-level policies.
arXiv Detail & Related papers (2024-03-15T17:45:44Z) - Amplifying robotics capacities with a human touch: An immersive
low-latency panoramic remote system [16.97496024217201]
"Avatar" system is an immersive low-latency panoramic human-robot interaction platform.
Under favorable network conditions, we achieved a low-latency high-definition panoramic visual experience with a delay of 357ms.
The system enables remote control over vast physical distances, spanning campuses, provinces, countries, and even continents.
arXiv Detail & Related papers (2024-01-07T06:55:41Z) - Giving Robots a Hand: Learning Generalizable Manipulation with
Eye-in-Hand Human Video Demonstrations [66.47064743686953]
Eye-in-hand cameras have shown promise in enabling greater sample efficiency and generalization in vision-based robotic manipulation.
Videos of humans performing tasks, on the other hand, are much cheaper to collect since they eliminate the need for expertise in robotic teleoperation.
In this work, we augment narrow robotic imitation datasets with broad unlabeled human video demonstrations to greatly enhance the generalization of eye-in-hand visuomotor policies.
arXiv Detail & Related papers (2023-07-12T07:04:53Z) - AnyTeleop: A General Vision-Based Dexterous Robot Arm-Hand Teleoperation System [51.48191418148764]
Vision-based teleoperation can endow robots with human-level intelligence to interact with the environment.
Current vision-based teleoperation systems are designed and engineered towards a particular robot model and deploy environment.
We propose AnyTeleop, a unified and general teleoperation system to support multiple different arms, hands, realities, and camera configurations within a single system.
arXiv Detail & Related papers (2023-07-10T14:11:07Z) - A Perspective on Robotic Telepresence and Teleoperation using Cognition:
Are we there yet? [0.0]
With the Artificial Intelligence (AI) revolution already being started, we can see a wide range of robotic applications being realized.
These technologies find significant application in health care, education, surveillance, disaster recovery, and corporate/government sectors.
But question still remains about their maturity, security and safety levels.
arXiv Detail & Related papers (2022-03-06T13:10:00Z) - Robotic Telekinesis: Learning a Robotic Hand Imitator by Watching Humans
on Youtube [24.530131506065164]
We build a system that enables any human to control a robot hand and arm, simply by demonstrating motions with their own hand.
The robot observes the human operator via a single RGB camera and imitates their actions in real-time.
We leverage this data to train a system that understands human hands and retargets a human video stream into a robot hand-arm trajectory that is smooth, swift, safe, and semantically similar to the guiding demonstration.
arXiv Detail & Related papers (2022-02-21T18:59:59Z) - Design and Development of Autonomous Delivery Robot [0.16863755729554888]
We present an autonomous mobile robot platform that delivers the package within the VNIT campus without any human intercommunication.
The entire pipeline of an autonomous robot working in outdoor environments is explained in this thesis.
arXiv Detail & Related papers (2021-03-16T17:57:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.