Ultra-Range Gesture Recognition using a web-camera in Human–Robot Interaction

Eran Bamani, Eden Nissinman, Inbar Meir, Lisa Koenigsberg, Avishai Sintov*

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

2 Scopus citations

Abstract

Hand gestures play a significant role in human interactions where non-verbal intentions, thoughts and commands are conveyed. In Human–Robot Interaction (HRI), hand gestures offer a similar and efficient medium for conveying clear and rapid directives to a robotic agent. However, state-of-the-art vision-based methods for gesture recognition have been shown to be effective only up to a user-camera distance of seven meters. Such a short distance range limits practical HRI with, for example, service robots, search and rescue robots and drones. In this work, we address the Ultra-Range Gesture Recognition (URGR) problem by aiming for a recognition distance of up to 25 m and in the context of HRI. We propose the URGR framework, a novel deep-learning, using solely a simple RGB camera. Gesture inference is based on a single image. First, a novel super-resolution model termed High-Quality Network (HQ-Net) uses a set of self-attention and convolutional layers to enhance the low-resolution image of the user. Then, we propose a novel URGR classifier termed Graph Vision Transformer (GViT ) which takes the enhanced image as input. GViT combines the benefits of a Graph Convolutional Network (GCN) and a modified Vision Transformer (ViT). Evaluation of the proposed framework over diverse test data yields a high recognition rate of 98.1%. The framework has also exhibited superior performance compared to human recognition in ultra-range distances. With the framework, we analyze and demonstrate the performance of an autonomous quadruped robot directed by human gestures in complex ultra-range indoor and outdoor environments, acquiring 96% recognition rate on average.

Original languageEnglish
Article number108443
JournalEngineering Applications of Artificial Intelligence
Volume132
DOIs
StatePublished - Jun 2024

Keywords

  • Graph Convolutional Network
  • Human–Robot Interaction
  • Ultra-Range Gesture Recognition
  • Vision Transformer

Fingerprint

Dive into the research topics of 'Ultra-Range Gesture Recognition using a web-camera in Human–Robot Interaction'. Together they form a unique fingerprint.

Cite this