Your browser doesn't support javascript.
loading
An Underwater Human-Robot Interaction Using a Visual-Textual Model for Autonomous Underwater Vehicles.
Zhang, Yongji; Jiang, Yu; Qi, Hong; Zhao, Minghao; Wang, Yuehang; Wang, Kai; Wei, Fenglin.
Afiliación
  • Zhang Y; College of Computer Science and Technology, Jilin University, Changchun 130012, China.
  • Jiang Y; College of Computer Science and Technology, Jilin University, Changchun 130012, China.
  • Qi H; State Key Lab of Symbolic Computation and Knowledge Engineering of Ministry of Education, Jilin University, Changchun 130012, China.
  • Zhao M; College of Computer Science and Technology, Jilin University, Changchun 130012, China.
  • Wang Y; State Key Lab of Symbolic Computation and Knowledge Engineering of Ministry of Education, Jilin University, Changchun 130012, China.
  • Wang K; College of Computer Science and Technology, Jilin University, Changchun 130012, China.
  • Wei F; College of Computer Science and Technology, Jilin University, Changchun 130012, China.
Sensors (Basel) ; 23(1)2022 Dec 24.
Article en En | MEDLINE | ID: mdl-36616794
ABSTRACT
The marine environment presents a unique set of challenges for human-robot interaction. Communicating with gestures is a common way for interacting between the diver and autonomous underwater vehicles (AUVs). However, underwater gesture recognition is a challenging visual task for AUVs due to light refraction and wavelength color attenuation issues. Current gesture recognition methods classify the whole image directly or locate the hand position first and then classify the hand features. Among these purely visual approaches, textual information is largely ignored. This paper proposes a visual-textual model for underwater hand gesture recognition (VT-UHGR). The VT-UHGR model encodes the underwater diver's image as visual features, the category text as textual features, and generates visual-textual features through multimodal interactions. We guide AUVs to use image-text matching for learning and inference. The proposed method achieves better performance than most existing purely visual methods on the dataset CADDY, demonstrating the effectiveness of using textual patterns for underwater gesture recognition.
Asunto(s)
Palabras clave

Texto completo: 1 Bases de datos: MEDLINE Asunto principal: Algoritmos / Robótica Límite: Humans Idioma: En Revista: Sensors (Basel) Año: 2022 Tipo del documento: Article País de afiliación: China

Texto completo: 1 Bases de datos: MEDLINE Asunto principal: Algoritmos / Robótica Límite: Humans Idioma: En Revista: Sensors (Basel) Año: 2022 Tipo del documento: Article País de afiliación: China