HandNeRF: Learning to Reconstruct Hand-Object Interaction Scene from a Single RGB Image

AmazUtah_NLP at SemEval-2024 Task 9: A MultiChoice Question Answering System for Commonsense Defying Reasoning


View a PDF of the paper titled HandNeRF: Learning to Reconstruct Hand-Object Interaction Scene from a Single RGB Image, by Hongsuk Choi and 4 other authors

View PDF
HTML (experimental)

Abstract:This paper presents a method to learn hand-object interaction prior for reconstructing a 3D hand-object scene from a single RGB image. The inference as well as training-data generation for 3D hand-object scene reconstruction is challenging due to the depth ambiguity of a single image and occlusions by the hand and object. We turn this challenge into an opportunity by utilizing the hand shape to constrain the possible relative configuration of the hand and object geometry. We design a generalizable implicit function, HandNeRF, that explicitly encodes the correlation of the 3D hand shape features and 2D object features to predict the hand and object scene geometry. With experiments on real-world datasets, we show that HandNeRF is able to reconstruct hand-object scenes of novel grasp configurations more accurately than comparable methods. Moreover, we demonstrate that object reconstruction from HandNeRF ensures more accurate execution of downstream tasks, such as grasping and motion planning for robotic hand-over and manipulation. Homepage: this https URL

Submission history

From: Hongsuk Choi [view email]
[v1]
Thu, 14 Sep 2023 17:42:08 UTC (2,871 KB)
[v2]
Thu, 12 Oct 2023 18:46:09 UTC (4,222 KB)
[v3]
Tue, 17 Oct 2023 18:11:28 UTC (4,222 KB)
[v4]
Sun, 11 Feb 2024 22:52:38 UTC (4,333 KB)
[v5]
Tue, 10 Sep 2024 19:20:13 UTC (4,313 KB)



Source link
lol

By stp2y

Leave a Reply

Your email address will not be published. Required fields are marked *

No widgets found. Go to Widget page and add the widget in Offcanvas Sidebar Widget Area.