loading page

Evaluating Visual-Spatiotemporal Coregistration of a Physics-based Virtual Reality Haptic Interface
  • Suman Chowdhury ,
  • Syed T. Mubarrat ,
  • Antonio Fernandes
Suman Chowdhury
Author Profile
Syed T. Mubarrat
Author Profile
Antonio Fernandes
Author Profile

Abstract

This study aimed to evaluate the visual-spatiotemporal co-registration of the real and virtual objects’ movement dynamics in a low-cost physics-based virtual reality (VR) system that provides real cutaneous and kinesthetic haptic feedback of the objects instead of computer-generated haptic feedback. Twelve healthy participants performed three human-robot collaborative (HRC) sequential pick-and-place lifting tasks while both motion capture and VR systems respectively traced the movement kinematics of the real and virtual objects. We used an iterative closest point algorithm to transform the 3D spatial point clouds of the VR system into the motion capture system. We employed a novel algorithm and principal component analysis to respectively calculate visual and spatiotemporal co-registration precisions between virtual and real objects. Results showed a high correlation (r > 0.96) between real and virtual objects’ movement dynamics and linear and angular co-registration errors of less than 5 cm and 8°, respectively. The trend also revealed a low temporal registration error of <12 ms and was only found along the vertical axis. The visual registration data indicated that the real cutaneous and kinesthetic haptics provided by the physical objects in the virtual environment enhanced proprioception and visuomotor functions of the users.
20 Feb 2024Submitted to TechRxiv
22 Feb 2024Published in TechRxiv
2024Published in IEEE Access volume 12 on pages 57017-57032. 10.1109/ACCESS.2024.3391186