An Event-Based Perception Pipeline for a Table Tennis Robot
- URL: http://arxiv.org/abs/2502.00749v1
- Date: Sun, 02 Feb 2025 10:56:37 GMT
- Title: An Event-Based Perception Pipeline for a Table Tennis Robot
- Authors: Andreas Ziegler, Thomas Gossard, Arren Glover, Andreas Zell,
- Abstract summary: We present the first real-time perception pipeline for a table tennis robot that uses only event-based cameras.
We show that compared to a frame-based pipeline, event-based perception pipelines have an update rate which is an order of magnitude higher.
- Score: 12.101426862186072
- License:
- Abstract: Table tennis robots gained traction over the last years and have become a popular research challenge for control and perception algorithms. Fast and accurate ball detection is crucial for enabling a robotic arm to rally the ball back successfully. So far, most table tennis robots use conventional, frame-based cameras for the perception pipeline. However, frame-based cameras suffer from motion blur if the frame rate is not high enough for fast-moving objects. Event-based cameras, on the other hand, do not have this drawback since pixels report changes in intensity asynchronously and independently, leading to an event stream with a temporal resolution on the order of us. To the best of our knowledge, we present the first real-time perception pipeline for a table tennis robot that uses only event-based cameras. We show that compared to a frame-based pipeline, event-based perception pipelines have an update rate which is an order of magnitude higher. This is beneficial for the estimation and prediction of the ball's position, velocity, and spin, resulting in lower mean errors and uncertainties. These improvements are an advantage for the robot control, which has to be fast, given the short time a table tennis ball is flying until the robot has to hit back.
Related papers
- Investigating Event-Based Cameras for Video Frame Interpolation in Sports [59.755469098797406]
We present a first investigation of event-based Video Frame Interpolation (VFI) models for generating sports slow-motion videos.
Particularly, we design and implement a bi-camera recording setup, including an RGB and an event-based camera to capture sports videos, to temporally align and spatially register both cameras.
Our experimental validation demonstrates that TimeLens, an off-the-shelf event-based VFI model, can effectively generate slow-motion footage for sports videos.
arXiv Detail & Related papers (2024-07-02T15:39:08Z) - Table tennis ball spin estimation with an event camera [11.735290341808064]
In table tennis, the combination of high velocity and spin renders traditional low frame rate cameras inadequate.
We present the first method for table tennis spin estimation using an event camera.
We achieve a spin magnitude mean error of $10.7 pm 17.3$ rps and a spin axis mean error of $32.9 pm 38.2deg$ in real time for a flying ball.
arXiv Detail & Related papers (2024-04-15T15:36:38Z) - A multi-modal table tennis robot system [12.590158763556186]
We present an improved table tennis robot system with high accuracy vision detection and fast robot reaction.
Based on previous work, our system contains a KUKA robot arm with 6 DOF, with four frame-based cameras and two additional event-based cameras.
arXiv Detail & Related papers (2023-10-29T16:35:29Z) - EventTransAct: A video transformer-based framework for Event-camera
based action recognition [52.537021302246664]
Event cameras offer new opportunities compared to standard action recognition in RGB videos.
In this study, we employ a computationally efficient model, namely the video transformer network (VTN), which initially acquires spatial embeddings per event-frame.
In order to better adopt the VTN for the sparse and fine-grained nature of event data, we design Event-Contrastive Loss ($mathcalL_EC$) and event-specific augmentations.
arXiv Detail & Related papers (2023-08-25T23:51:07Z) - EV-Catcher: High-Speed Object Catching Using Low-latency Event-based
Neural Networks [107.62975594230687]
We demonstrate an application where event cameras excel: accurately estimating the impact location of fast-moving objects.
We introduce a lightweight event representation called Binary Event History Image (BEHI) to encode event data at low latency.
We show that the system is capable of achieving a success rate of 81% in catching balls targeted at different locations, with a velocity of up to 13 m/s even on compute-constrained embedded platforms.
arXiv Detail & Related papers (2023-04-14T15:23:28Z) - EvConv: Fast CNN Inference on Event Camera Inputs For High-Speed Robot
Perception [1.3869227429939426]
Event cameras capture visual information with a high temporal resolution and a wide dynamic range.
Current convolutional neural network inference on event camera streams cannot currently perform real-time inference at the high speeds at which event cameras operate.
This paper presents EvConv, a new approach to enable fast inference on CNNs for inputs from event cameras.
arXiv Detail & Related papers (2023-03-08T15:47:13Z) - Fast Trajectory End-Point Prediction with Event Cameras for Reactive
Robot Control [4.110120522045467]
In this paper, we propose to exploit the low latency, motion-driven sampling, and data compression properties of event cameras to overcome these issues.
As a use-case, we use a Panda robotic arm to intercept a ball bouncing on a table.
We train the network in simulation to speed up the dataset acquisition and then fine-tune the models on real trajectories.
arXiv Detail & Related papers (2023-02-27T14:14:52Z) - Neural Scene Representation for Locomotion on Structured Terrain [56.48607865960868]
We propose a learning-based method to reconstruct the local terrain for a mobile robot traversing urban environments.
Using a stream of depth measurements from the onboard cameras and the robot's trajectory, the estimates the topography in the robot's vicinity.
We propose a 3D reconstruction model that faithfully reconstructs the scene, despite the noisy measurements and large amounts of missing data coming from the blind spots of the camera arrangement.
arXiv Detail & Related papers (2022-06-16T10:45:17Z) - TimeReplayer: Unlocking the Potential of Event Cameras for Video
Interpolation [78.99283105497489]
Event camera is a new device to enable video at the presence of arbitrarily complex motion.
This paper proposes a novel TimeReplayer algorithm to interpolate videos captured by commodity cameras with events.
arXiv Detail & Related papers (2022-03-25T18:57:42Z) - Event Camera Based Real-Time Detection and Tracking of Indoor Ground
Robots [2.471139321417215]
This paper presents a real-time method to detect and track multiple mobile ground robots using event cameras.
The method uses density-based spatial clustering of applications with noise (DBSCAN) to detect the robots and a single k-dimensional (k-d) tree to accurately keep track of them as they move in an indoor arena.
arXiv Detail & Related papers (2021-02-23T19:50:17Z) - EventHands: Real-Time Neural 3D Hand Reconstruction from an Event Stream [80.15360180192175]
3D hand pose estimation from monocular videos is a long-standing and challenging problem.
We address it for the first time using a single event camera, i.e., an asynchronous vision sensor reacting on brightness changes.
Our approach has characteristics previously not demonstrated with a single RGB or depth camera.
arXiv Detail & Related papers (2020-12-11T16:45:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.