Contact Area Detector using Cross View Projection Consistency for
COVID-19 Projects
- URL: http://arxiv.org/abs/2008.07712v1
- Date: Tue, 18 Aug 2020 02:57:26 GMT
- Title: Contact Area Detector using Cross View Projection Consistency for
COVID-19 Projects
- Authors: Pan Zhang, Wilfredo Torres Calderon, Bokyung Lee, Alex Tessier, Jacky
Bibliowicz, Liviu Calin, Michael Lee
- Abstract summary: We show that the contact between an object and a static surface can be identified by projecting the object onto the static surface through two different viewpoints.
This simple method can be easily adapted to real-life applications.
- Score: 7.539495357219132
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The ability to determine what parts of objects and surfaces people touch as
they go about their daily lives would be useful in understanding how the
COVID-19 virus spreads. To determine whether a person has touched an object or
surface using visual data, images, or videos, is a hard problem. Computer
vision 3D reconstruction approaches project objects and the human body from the
2D image domain to 3D and perform 3D space intersection directly. However, this
solution would not meet the accuracy requirement in applications due to
projection error. Another standard approach is to train a neural network to
infer touch actions from the collected visual data. This strategy would require
significant amounts of training data to generalize over scale and viewpoint
variations. A different approach to this problem is to identify whether a
person has touched a defined object. In this work, we show that the solution to
this problem can be straightforward. Specifically, we show that the contact
between an object and a static surface can be identified by projecting the
object onto the static surface through two different viewpoints and analyzing
their 2D intersection. The object contacts the surface when the projected
points are close to each other; we call this cross view projection consistency.
Instead of doing 3D scene reconstruction or transfer learning from deep
networks, a mapping from the surface in the two camera views to the surface
space is the only requirement. For planar space, this mapping is the Homography
transformation. This simple method can be easily adapted to real-life
applications. In this paper, we apply our method to do office occupancy
detection for studying the COVID-19 transmission pattern from an office desk in
a meeting room using the contact information.
Related papers
- Decaf: Monocular Deformation Capture for Face and Hand Interactions [77.75726740605748]
This paper introduces the first method that allows tracking human hands interacting with human faces in 3D from single monocular RGB videos.
We model hands as articulated objects inducing non-rigid face deformations during an active interaction.
Our method relies on a new hand-face motion and interaction capture dataset with realistic face deformations acquired with a markerless multi-view camera system.
arXiv Detail & Related papers (2023-09-28T17:59:51Z) - Learning Explicit Contact for Implicit Reconstruction of Hand-held
Objects from Monocular Images [59.49985837246644]
We show how to model contacts in an explicit way to benefit the implicit reconstruction of hand-held objects.
In the first part, we propose a new subtask of directly estimating 3D hand-object contacts from a single image.
In the second part, we introduce a novel method to diffuse estimated contact states from the hand mesh surface to nearby 3D space.
arXiv Detail & Related papers (2023-05-31T17:59:26Z) - NeurOCS: Neural NOCS Supervision for Monocular 3D Object Localization [80.3424839706698]
We present NeurOCS, a framework that uses instance masks 3D boxes as input to learn 3D object shapes by means of differentiable rendering.
Our approach rests on insights in learning a category-level shape prior directly from real driving scenes.
We make critical design choices to learn object coordinates more effectively from an object-centric view.
arXiv Detail & Related papers (2023-05-28T16:18:41Z) - Grounding 3D Object Affordance from 2D Interactions in Images [128.6316708679246]
Grounding 3D object affordance seeks to locate objects' ''action possibilities'' regions in the 3D space.
Humans possess the ability to perceive object affordances in the physical world through demonstration images or videos.
We devise an Interaction-driven 3D Affordance Grounding Network (IAG), which aligns the region feature of objects from different sources.
arXiv Detail & Related papers (2023-03-18T15:37:35Z) - Neural Correspondence Field for Object Pose Estimation [67.96767010122633]
We propose a method for estimating the 6DoF pose of a rigid object with an available 3D model from a single RGB image.
Unlike classical correspondence-based methods which predict 3D object coordinates at pixels of the input image, the proposed method predicts 3D object coordinates at 3D query points sampled in the camera frustum.
arXiv Detail & Related papers (2022-07-30T01:48:23Z) - Tac2Pose: Tactile Object Pose Estimation from the First Touch [6.321662423735226]
We present Tac2Pose, an object-specific approach to tactile pose estimation from the first touch for known objects.
We simulate the contact shapes that a dense set of object poses would produce on the sensor.
We obtain contact shapes from the sensor with an object-agnostic calibration step that maps RGB tactile observations to binary contact shapes.
arXiv Detail & Related papers (2022-04-25T14:43:48Z) - Sparse Pose Trajectory Completion [87.31270669154452]
We propose a method to learn, even using a dataset where objects appear only in sparsely sampled views.
This is achieved with a cross-modal pose trajectory transfer mechanism.
Our method is evaluated on the Pix3D and ShapeNet datasets.
arXiv Detail & Related papers (2021-05-01T00:07:21Z) - 3D Object Detection and Pose Estimation of Unseen Objects in Color
Images with Local Surface Embeddings [35.769234123059086]
We present an approach for detecting and estimating the 3D poses of objects in images that requires only an untextured CAD model.
Our approach combines Deep Learning and 3D geometry: It relies on an embedding of local 3D geometry to match the CAD models to the input images.
We show that we can use Mask-RCNN in a class-agnostic way to detect the new objects without retraining and thus drastically limit the number of possible correspondences.
arXiv Detail & Related papers (2020-10-08T15:57:06Z) - HOPE-Net: A Graph-based Model for Hand-Object Pose Estimation [7.559220068352681]
We propose a lightweight model called HOPE-Net which jointly estimates hand and object pose in 2D and 3D in real-time.
Our network uses a cascade of two adaptive graph convolutional neural networks, one to estimate 2D coordinates of the hand joints and object corners, followed by another to convert 2D coordinates to 3D.
arXiv Detail & Related papers (2020-03-31T19:01:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.