Your browser doesn't support javascript.
loading
A data-efficient and easy-to-use lip language interface based on wearable motion capture and speech movement reconstruction.
Liu, Shiqiang; Fawden, Terry; Zhu, Rong; Malliaras, George G; Bance, Manohar.
Afiliação
  • Liu S; State Key Laboratory of Precision Measurement Technology and Instrument, Department of Precision Instrument, Tsinghua University, Beijing 100084, China.
  • Fawden T; Department of Clinical Neurosciences, University of Cambridge, Cambridge CB23EB, UK.
  • Zhu R; State Key Laboratory of Precision Measurement Technology and Instrument, Department of Precision Instrument, Tsinghua University, Beijing 100084, China.
  • Malliaras GG; Electrical Engineering Division, Department of Engineering, University of Cambridge, Cambridge CB3 0FA, UK.
  • Bance M; Department of Clinical Neurosciences, University of Cambridge, Cambridge CB23EB, UK.
Sci Adv ; 10(26): eado9576, 2024 Jun 28.
Article em En | MEDLINE | ID: mdl-38924408
ABSTRACT
Lip language recognition urgently needs wearable and easy-to-use interfaces for interference-free and high-fidelity lip-reading acquisition and to develop accompanying data-efficient decoder-modeling methods. Existing solutions suffer from unreliable lip reading, are data hungry, and exhibit poor generalization. Here, we propose a wearable lip language decoding technology that enables interference-free and high-fidelity acquisition of lip movements and data-efficient recognition of fluent lip language based on wearable motion capture and continuous lip speech movement reconstruction. The method allows us to artificially generate any wanted continuous speech datasets from a very limited corpus of word samples from users. By using these artificial datasets to train the decoder, we achieve an average accuracy of 92.0% across individuals (n = 7) for actual continuous and fluent lip speech recognition for 93 English sentences, even observing no training burn on users because all training datasets are artificially generated. Our method greatly minimizes users' training/learning load and presents a data-efficient and easy-to-use paradigm for lip language recognition.
Assuntos

Texto completo: 1 Base de dados: MEDLINE Assunto principal: Fala / Dispositivos Eletrônicos Vestíveis Limite: Adult / Female / Humans / Male Idioma: En Ano de publicação: 2024 Tipo de documento: Article

Texto completo: 1 Base de dados: MEDLINE Assunto principal: Fala / Dispositivos Eletrônicos Vestíveis Limite: Adult / Female / Humans / Male Idioma: En Ano de publicação: 2024 Tipo de documento: Article