Back to Search
Start Over
HapTable: An Interactive Tabletop Providing Online Haptic Feedback for Touch Gestures
- Source :
- IEEE Transactions on Visualization and Computer Graphics, IEEE Transactions on Visualization and Computer Graphics, Institute of Electrical and Electronics Engineers, 2019, 25 (9), pp.2749-2762. ⟨10.1109/TVCG.2018.2855154⟩
- Publication Year :
- 2021
-
Abstract
- We present HapTable; a multi-modal interactive tabletop that allows users to interact with digital images and objects through natural touch gestures, and receive visual and haptic feedback accordingly. In our system, hand pose is registered by an infrared camera and hand gestures are classified using a Support Vector Machine (SVM) classifier. To display a rich set of haptic effects for both static and dynamic gestures, we integrated electromechanical and electrostatic actuation techniques effectively on tabletop surface of HapTable, which is a surface capacitive touch screen. We attached four piezo patches to the edges of tabletop to display vibrotactile feedback for static gestures. For this purpose, the vibration response of the tabletop, in the form of frequency response functions (FRFs), was obtained by a laser Doppler vibrometer for 84 grid points on its surface. Using these FRFs, it is possible to display localized vibrotactile feedback on the surface for static gestures. For dynamic gestures, we utilize the electrostatic actuation technique to modulate the frictional forces between finger skin and tabletop surface by applying voltage to its conductive layer. To our knowledge, this hybrid haptic technology is one of a kind and has not been implemented or tested on a tabletop. It opens up new avenues for gesture-based haptic interaction not only on tabletop surfaces but also on touch surfaces used in mobile devices with potential applications in data visualization, user interfaces, games, entertainment, and education. Here, we present two examples of such applications, one for static and one for dynamic gestures, along with detailed user studies. In the first one, user detects the direction of a virtual flow, such as that of wind or water, by putting their hand on the tabletop surface and feeling a vibrotactile stimulus traveling underneath it. In the second example, user rotates a virtual knob on the tabletop surface to select an item from a menu while feeling the knob's detents and resistance to rotation in the form of frictional haptic feedback.<br />Scientific and Technological Research Council of Turkey (TÜBİTAK)
- Subjects :
- FOS: Computer and information sciences
vibrotactile haptic feedback
Computer science
InformationSystems_INFORMATIONINTERFACESANDPRESENTATION(e.g.,HCI)
Computer Vision and Pattern Recognition (cs.CV)
Computer Science - Human-Computer Interaction
Computer Science - Computer Vision and Pattern Recognition
02 engineering and technology
Human-Computer Interaction (cs.HC)
Data visualization
Computer Science - Graphics
Electrostatic actuation
human-computer interaction
0202 electrical engineering, electronic engineering, information engineering
Computer vision
[INFO.INFO-HC]Computer Science [cs]/Human-Computer Interaction [cs.HC]
Haptic technology
ComputingMethodologies_COMPUTERGRAPHICS
Haptic interaction
business.industry
gesture recognition
020207 software engineering
haptic interfaces
Computer Graphics and Computer-Aided Design
Graphics (cs.GR)
Visualization
Multimedia (cs.MM)
[SPI.TRON]Engineering Sciences [physics]/Electronics
Signal Processing
multimodal systems
Computer Vision and Pattern Recognition
Artificial intelligence
User interface
business
Mobile device
Laser Doppler vibrometer
Software
Computer Science - Multimedia
Gesture recognition
Haptic interfaces
Human-computer interaction
Multimodal systems
Vibrotactile haptic feedback
Gesture
Subjects
Details
- Language :
- English
- ISSN :
- 10772626
- Database :
- OpenAIRE
- Journal :
- IEEE Transactions on Visualization and Computer Graphics, IEEE Transactions on Visualization and Computer Graphics, Institute of Electrical and Electronics Engineers, 2019, 25 (9), pp.2749-2762. ⟨10.1109/TVCG.2018.2855154⟩
- Accession number :
- edsair.doi.dedup.....143221d40f71a5d40ef052849a03b76d
- Full Text :
- https://doi.org/10.1109/TVCG.2018.2855154⟩