Your browser doesn't support javascript.
loading
Using Computer Vision to Annotate Video-Recoded Direct Observation of Physical Behavior.
Keadle, Sarah K; Eglowski, Skylar; Ylarregui, Katie; Strath, Scott J; Martinez, Julian; Dekhtyar, Alex; Kagan, Vadim.
Afiliação
  • Keadle SK; Department of Kinesiology and Public Health, California Polytechnic State University, San Luis Obispo, CA 93407, USA.
  • Eglowski S; Sentimetrix Inc., Bethesda, MD 20814, USA.
  • Ylarregui K; Department of Kinesiology and Public Health, California Polytechnic State University, San Luis Obispo, CA 93407, USA.
  • Strath SJ; College of Public Health, University of Wisconsin, Milwaukee, WI 53205, USA.
  • Martinez J; College of Public Health, University of Wisconsin, Milwaukee, WI 53205, USA.
  • Dekhtyar A; Department of Computer Science and Software Engineering, California Polytechnic State University, San Luis Obispo, CA 93407, USA.
  • Kagan V; Sentimetrix Inc., Bethesda, MD 20814, USA.
Sensors (Basel) ; 24(7)2024 Apr 08.
Article em En | MEDLINE | ID: mdl-38610576
ABSTRACT
Direct observation is a ground-truth measure for physical behavior, but the high cost limits widespread use. The purpose of this study was to develop and test machine learning methods to recognize aspects of physical behavior and location from videos of human movement Adults (N = 26, aged 18-59 y) were recorded in their natural environment for two, 2- to 3-h sessions. Trained research assistants annotated videos using commercially available software including the following taxonomies (1) sedentary versus non-sedentary (two classes); (2) activity type (four classes sedentary, walking, running, and mixed movement); and (3) activity intensity (four classes sedentary, light, moderate, and vigorous). Four machine learning approaches were trained and evaluated for each taxonomy. Models were trained on 80% of the videos, validated on 10%, and final accuracy is reported on the remaining 10% of the videos not used in training. Overall accuracy was as follows 87.4% for Taxonomy 1, 63.1% for Taxonomy 2, and 68.6% for Taxonomy 3. This study shows it is possible to use computer vision to annotate aspects of physical behavior, speeding up the time and reducing labor required for direct observation. Future research should test these machine learning models on larger, independent datasets and take advantage of analysis of video fragments, rather than individual still images.
Assuntos
Palavras-chave

Texto completo: 1 Bases de dados: MEDLINE Assunto principal: Trabalho de Parto / Computadores Limite: Adult / Female / Humans / Pregnancy Idioma: En Revista: Sensors (Basel) Ano de publicação: 2024 Tipo de documento: Article País de afiliação: Estados Unidos

Texto completo: 1 Bases de dados: MEDLINE Assunto principal: Trabalho de Parto / Computadores Limite: Adult / Female / Humans / Pregnancy Idioma: En Revista: Sensors (Basel) Ano de publicação: 2024 Tipo de documento: Article País de afiliação: Estados Unidos