Show simple item record

RGB2Hands: Real-Time Tracking of 3D Hand Interactions from Monocular RGB Video

dc.contributor.authorWANG, JIAYI
dc.contributor.authorMUELLER, FRANZISKA
dc.contributor.authorBERNARD, FLORIAN
dc.contributor.authorSORLI, SUZANNE
dc.contributor.authorSOTNYCHENKO, OLEKSANDR
dc.contributor.authorQIAN, NENG
dc.contributor.authorOTADUY, MIGUEL A.
dc.contributor.authorCASAS, DAN
dc.contributor.authorTHEOBALT, CHRISTIAN
dc.date.accessioned2021-04-19T11:04:23Z
dc.date.available2021-04-19T11:04:23Z
dc.date.issued2020
dc.identifier.citationACM Trans. Graph., Vol. 39, No. 6, Article 218. Publication date: December 2020es
dc.identifier.issn1557-7368
dc.identifier.urihttp://hdl.handle.net/10115/17670
dc.descriptionTouchDesign (M1792)es
dc.description© 2020 Copyright held by the owner/author(s). Publication rights licensed to ACM. This is the author’s version of the work. It is posted here for your personal use. Not for redistribution. The definitive Version of Record was published in ACM Transactions on Graphics, https://doi.org/10.1145/3414685.3417852.
dc.description.abstractTracking and reconstructing the 3D pose and geometry of two hands in interaction is a challenging problem that has a high relevance for several human-computer interaction applications, including AR/VR, robotics, or sign language recognition. Existing works are either limited to simpler tracking settings (e.g., considering only a single hand or two spatially separated hands), or rely on less ubiquitous sensors, such as depth cameras. In contrast, in this work we present the first real-time method for motion capture of skeletal pose and 3D surface geometry of hands from a single RGB camera that explicitly considers close interactions. In order to address the inherent depth ambiguities in RGB data, we propose a novel multi-task CNN that regresses multiple complementary pieces of information, including segmentation, dense matchings to a 3D hand model, and 2D keypoint positions, together with newly proposed intra-hand relative depth and inter-hand distance maps. These predictions are subsequently used in a generative model fitting framework in order to estimate pose and shape parameters of a 3D hand model for both hands. We experimentally verify the individual components of our RGB two-hand tracking and 3D reconstruction pipeline through an extensive ablation study. Moreover, we demonstrate that our approach offers previously unseen two-hand tracking performance from RGB, and quantitatively and qualitatively outperforms existing RGB-based methods that were not explicitly designed for two-hand interactions. Moreover, our method even performs on-par with depth-based real-time methods.es
dc.language.isoenges
dc.publisherAssociation for Computing Machinery (ACM)es
dc.rightsAttribution-NonCommercial-NoDerivatives 4.0 Internacional*
dc.rights.urihttp://creativecommons.org/licenses/by-nc-nd/4.0/*
dc.subjecthand trackinges
dc.subjecthand pose estimationes
dc.subjecthand reconstructiones
dc.subjecttwo handses
dc.subjectmonocular RGBes
dc.subjectRGB videoes
dc.subjectcomputing methodologieses
dc.subjectComputer visiones
dc.subjectNeural networkses
dc.titleRGB2Hands: Real-Time Tracking of 3D Hand Interactions from Monocular RGB Videoes
dc.typeinfo:eu-repo/semantics/articlees
dc.identifier.doi10.1145/3414685.3417852es
dc.rights.accessRightsinfo:eu-repo/semantics/openAccesses
dc.relation.projectIDTouchDesign (M1792)es


Files in this item

This item appears in the following Collection(s)

Show simple item record

Attribution-NonCommercial-NoDerivatives 4.0 InternacionalExcept where otherwise noted, this item's license is described as Attribution-NonCommercial-NoDerivatives 4.0 Internacional