Jialin Chen
Comparing a Graphical User Interface, Hand Gestures and Controller in Virtual Reality for Robot Teleoperation
Chen, Jialin; Moemeni, Armaghan; Caleb-Solly, Praminda
Authors
ARMAGHAN MOEMENI ARMAGHAN.MOEMENI@NOTTINGHAM.AC.UK
Assistant Professor
Professor PRAMINDA CALEB-SOLLY Praminda.Caleb-Solly@nottingham.ac.uk
Professor of Embodied Intelligence
Abstract
Robot teleoperation is being explored in a number of application areas, where combining human adaptive intelligence and high precision of robots can provide access to dangerous or inaccessible places, or augment human dexterity. Using virtual reality (VR) is one way to enable robot teleoperation, where additional information can be augmented in the display to make the remote control easier. In this paper, we present a robot teleoperation system, developed and deployed on a JAKA Minicobo robotic arm, to compare the user experience and performance for three different control methods. These include a VR controller, VR gestures, and a traditional graphical user interface (GUI). Each study participant was asked to conduct the experiment twice using all three methods, during which the time, the total spatial distance of the robot end-effector movement, and the frequency of errors in a teleoperation task were recorded for quantitative analysis. All participants were also asked to complete a questionnaire based on the NASA task load index. The results show that overall the VR gestures method enables users to complete the task faster than the VR controller, and using the traditional GUI is generally slower. While the quantitative results do not show statistically significant differences, both of the VR methods place greater perceived physical and mental demands on the users, in comparison to the GUI, although when asked which method they preferred, only three of the sixteen participants preferred the VR gestures method. Given that a number of applications use VR, this study indicates that if the task is not time critical, then a traditional GUI might be more suitable for reducing perceived mental and physical load, and controller-free VR interaction might not always be desirable.
Citation
Chen, J., Moemeni, A., & Caleb-Solly, P. (2023). Comparing a Graphical User Interface, Hand Gestures and Controller in Virtual Reality for Robot Teleoperation. In HRI '23: Companion of the 2023 ACM/IEEE International Conference on Human-Robot Interaction (644-648). https://doi.org/10.1145/3568294.3580165
Presentation Conference Type | Conference Paper (Published) |
---|---|
Conference Name | Companion of the 2023 ACM/IEEE International Conference on Human-Robot Interaction |
Start Date | Mar 13, 2023 |
End Date | Mar 16, 2023 |
Acceptance Date | Jan 11, 2023 |
Online Publication Date | Mar 13, 2023 |
Publication Date | Mar 13, 2023 |
Deposit Date | Mar 14, 2023 |
Publisher | Association for Computing Machinery (ACM) |
Pages | 644-648 |
Book Title | HRI '23: Companion of the 2023 ACM/IEEE International Conference on Human-Robot Interaction |
ISBN | 9781450399708 |
DOI | https://doi.org/10.1145/3568294.3580165 |
Public URL | https://nottingham-repository.worktribe.com/output/18237289 |
Publisher URL | https://dl.acm.org/doi/10.1145/3568294.3580165 |
You might also like
Trust in Robot Benchmarking and Benchmarking for Trustworthy Robots
(2024)
Book Chapter
Empowering future care workforces: scoping human capabilities to leverage assistive robotics
(2023)
Presentation / Conference Contribution
Using a Knowledge Café approach as a public engagement activity for raising awareness of data protection issues in robotics for health and social care
(2023)
Presentation / Conference Contribution
Digital twins for human-assistive robot teams in ambient assisted living
(2023)
Presentation / Conference Contribution
Robustness of Deep Learning Methods for Occluded Object Detection - A Study Introducing a Novel Occlusion Dataset
(2023)
Presentation / Conference Contribution
Downloadable Citations
About Repository@Nottingham
Administrator e-mail: discovery-access-systems@nottingham.ac.uk
This application uses the following open-source libraries:
SheetJS Community Edition
Apache License Version 2.0 (http://www.apache.org/licenses/)
PDF.js
Apache License Version 2.0 (http://www.apache.org/licenses/)
Font Awesome
SIL OFL 1.1 (http://scripts.sil.org/OFL)
MIT License (http://opensource.org/licenses/mit-license.html)
CC BY 3.0 ( http://creativecommons.org/licenses/by/3.0/)
Powered by Worktribe © 2024
Advanced Search