Back to Search
Start Over
Online generative model personalization for hand tracking
- Source :
- ACM Transactions on Graphics. 36:1-11
- Publication Year :
- 2017
- Publisher :
- Association for Computing Machinery (ACM), 2017.
-
Abstract
- We present a new algorithm for real-time hand tracking on commodity depth-sensing devices. Our method does not require a user-specific calibration session, but rather learns the geometry as the user performs live in front of the camera, thus enabling seamless virtual interaction at the consumer level. The key novelty in our approach is an online optimization algorithm that jointly estimates pose and shape in each frame, and determines the uncertainty in such estimates. This knowledge allows the algorithm to integrate per-frame estimates over time, and build a personalized geometric model of the captured user. Our approach can easily be integrated in state-of-the-art continuous generative motion tracking software. We provide a detailed evaluation that shows how our approach achieves accurate motion tracking for real-time applications, while significantly simplifying the workflow of accurate hand performance capture. We also provide quantitative evaluation datasets at http://gfx.uvic.ca/datasets/handy
- Subjects :
- Computer science
business.industry
Frame (networking)
articulated registration
020207 software engineering
real-time calibration
02 engineering and technology
Machine learning
computer.software_genre
generative tracking
Computer Graphics and Computer-Aided Design
Motion capture
Personalization
Generative model
Software
Match moving
motion capture
0202 electrical engineering, electronic engineering, information engineering
020201 artificial intelligence & image processing
Computer vision
Artificial intelligence
real-time hand tracking
Geometric modeling
business
computer
Subjects
Details
- ISSN :
- 15577368 and 07300301
- Volume :
- 36
- Database :
- OpenAIRE
- Journal :
- ACM Transactions on Graphics
- Accession number :
- edsair.doi.dedup.....4c002f08048c8b831a82210aea386479
- Full Text :
- https://doi.org/10.1145/3130800.3130830