Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Results 1 - 20 de 2.613
Filter
1.
Front Vet Sci ; 11: 1436795, 2024.
Article in English | MEDLINE | ID: mdl-39086767

ABSTRACT

Facial expressions are essential for communication and emotional expression across species. Despite the improvements brought by tools like the Horse Grimace Scale (HGS) in pain recognition in horses, their reliance on human identification of characteristic traits presents drawbacks such as subjectivity, training requirements, costs, and potential bias. Despite these challenges, the development of facial expression pain scales for animals has been making strides. To address these limitations, Automated Pain Recognition (APR) powered by Artificial Intelligence (AI) offers a promising advancement. Notably, computer vision and machine learning have revolutionized our approach to identifying and addressing pain in non-verbal patients, including animals, with profound implications for both veterinary medicine and animal welfare. By leveraging the capabilities of AI algorithms, we can construct sophisticated models capable of analyzing diverse data inputs, encompassing not only facial expressions but also body language, vocalizations, and physiological signals, to provide precise and objective evaluations of an animal's pain levels. While the advancement of APR holds great promise for improving animal welfare by enabling better pain management, it also brings forth the need to overcome data limitations, ensure ethical practices, and develop robust ground truth measures. This narrative review aimed to provide a comprehensive overview, tracing the journey from the initial application of facial expression recognition for the development of pain scales in animals to the recent application, evolution, and limitations of APR, thereby contributing to understanding this rapidly evolving field.

2.
Ann Biomed Eng ; 2024 Aug 03.
Article in English | MEDLINE | ID: mdl-39097542

ABSTRACT

PURPOSE: Estimating loading of the knee joint may be helpful in managing degenerative joint diseases. Contemporary methods to estimate loading involve calculating knee joint contact forces using musculoskeletal modeling and simulation from motion capture (MOCAP) data, which must be collected in a specialized environment and analyzed by a trained expert. To make the estimation of knee joint loading more accessible, simple input predictors should be used for predicting knee joint loading using artificial neural networks. METHODS: We trained feedforward artificial neural networks (ANNs) to predict knee joint loading peaks from the mass, height, age, sex, walking speed, and knee flexion angle (KFA) of subjects using their existing MOCAP data. We also collected an independent MOCAP dataset while recording walking with a video camera (VC) and inertial measurement units (IMUs). We quantified the prediction accuracy of the ANNs using walking speed and KFA estimates from (1) MOCAP data, (2) VC data, and (3) IMU data separately (i.e., we quantified three sets of prediction accuracy metrics). RESULTS: Using portable modalities, we achieved prediction accuracies between 0.13 and 0.37 root mean square error normalized to the mean of the musculoskeletal analysis-based reference values. The correlation between the predicted and reference loading peaks varied between 0.65 and 0.91. This was comparable to the prediction accuracies obtained when obtaining predictors from motion capture data. DISCUSSION: The prediction results show that both VCs and IMUs can be used to estimate predictors that can be used in estimating knee joint loading outside the motion laboratory. Future studies should investigate the usability of the methods in an out-of-laboratory setting.

3.
BMC Oral Health ; 24(1): 772, 2024 Jul 10.
Article in English | MEDLINE | ID: mdl-38987714

ABSTRACT

Integrating artificial intelligence (AI) into medical and dental applications can be challenging due to clinicians' distrust of computer predictions and the potential risks associated with erroneous outputs. We introduce the idea of using AI to trigger second opinions in cases where there is a disagreement between the clinician and the algorithm. By keeping the AI prediction hidden throughout the diagnostic process, we minimize the risks associated with distrust and erroneous predictions, relying solely on human predictions. The experiment involved 3 experienced dentists, 25 dental students, and 290 patients treated for advanced caries across 6 centers. We developed an AI model to predict pulp status following advanced caries treatment. Clinicians were asked to perform the same prediction without the assistance of the AI model. The second opinion framework was tested in a 1000-trial simulation. The average F1-score of the clinicians increased significantly from 0.586 to 0.645.


Subject(s)
Artificial Intelligence , Dental Caries , Humans , Dental Caries/therapy , Referral and Consultation , Patient Care Planning , Algorithms
4.
Ophthalmol Sci ; 4(5): 100540, 2024.
Article in English | MEDLINE | ID: mdl-39051045

ABSTRACT

Objective: An enlarged cup-to-disc ratio (CDR) is a hallmark of glaucomatous optic neuropathy. Manual assessment of the CDR may be less accurate and more time-consuming than automated methods. Here, we sought to develop and validate a deep learning-based algorithm to automatically determine the CDR from fundus images. Design: Algorithm development for estimating CDR using fundus data from a population-based observational study. Participants: A total of 181 768 fundus images from the United Kingdom Biobank (UKBB), Drishti_GS, and EyePACS. Methods: FastAI and PyTorch libraries were used to train a convolutional neural network-based model on fundus images from the UKBB. Models were constructed to determine image gradability (classification analysis) as well as to estimate CDR (regression analysis). The best-performing model was then validated for use in glaucoma screening using a multiethnic dataset from EyePACS and Drishti_GS. Main Outcome Measures: The area under the receiver operating characteristic curve and coefficient of determination. Results: Our gradability model vgg19_batch normalization (bn) achieved an accuracy of 97.13% on a validation set of 16 045 images, with 99.26% precision and area under the receiver operating characteristic curve of 96.56%. Using regression analysis, our best-performing model (trained on the vgg19_bn architecture) attained a coefficient of determination of 0.8514 (95% confidence interval [CI]: 0.8459-0.8568), while the mean squared error was 0.0050 (95% CI: 0.0048-0.0051) and mean absolute error was 0.0551 (95% CI: 0.0543-0.0559) on a validation set of 12 183 images for determining CDR. The regression point was converted into classification metrics using a tolerance of 0.2 for 20 classes; the classification metrics achieved an accuracy of 99.20%. The EyePACS dataset (98 172 healthy, 3270 glaucoma) was then used to externally validate the model for glaucoma classification, with an accuracy, sensitivity, and specificity of 82.49%, 72.02%, and 82.83%, respectively. Conclusions: Our models were precise in determining image gradability and estimating CDR. Although our artificial intelligence-derived CDR estimates achieve high accuracy, the CDR threshold for glaucoma screening will vary depending on other clinical parameters. Financial Disclosures: Proprietary or commercial disclosure may be found in the Footnotes and Disclosures at the end of this article.

5.
Maturitas ; : 108065, 2024 Jul 14.
Article in English | MEDLINE | ID: mdl-39054223
6.
Article in English | MEDLINE | ID: mdl-38978825

ABSTRACT

Background: The American Optometric Association defines computer vision syndrome (CVS), also known as digital eye strain, as "a group of eye- and vision-related problems that result from prolonged computer, tablet, e-reader and cell phone use". We aimed to create a well-structured, valid, and reliable questionnaire to determine the prevalence of CVS, and to analyze the visual, ocular surface, and extraocular sequelae of CVS using a novel and smart self-assessment questionnaire. Methods: This multicenter, observational, cross-sectional, descriptive, survey-based, online study included 6853 complete online responses of medical students from 15 universities. All participants responded to the updated, online, fourth version of the CVS questionnaire (CVS-F4), which has high validity and reliability. CVS was diagnosed according to five basic diagnostic criteria (5DC) derived from the CVS-F4. Respondents who fulfilled the 5DC were considered CVS cases. The 5DC were then converted into a novel five-question self-assessment questionnaire designated as the CVS-Smart. Results: Of 10 000 invited medical students, 8006 responded to the CVS-F4 survey (80% response rate), while 6853 of the 8006 respondents provided complete online responses (85.6% completion rate). The overall CVS prevalence was 58.78% (n = 4028) among the study respondents; CVS prevalence was higher among women (65.87%) than among men (48.06%). Within the CVS group, the most common visual, ocular surface, and extraocular complaints were eye strain, dry eye, and neck/shoulder/back pain in 74.50% (n = 3001), 58.27% (n = 2347), and 80.52% (n = 3244) of CVS cases, respectively. Notably, 75.92% (3058/4028) of CVS cases were involved in the Mandated Computer System Use Program. Multivariate logistic regression analysis revealed that the two most statistically significant diagnostic criteria of the 5DC were ≥2 symptoms/attacks per month over the last 12 months (odds ratio [OR] = 204177.2; P <0.0001) and symptoms/attacks associated with screen use (OR = 16047.34; P <0.0001). The CVS-Smart demonstrated a Cronbach's alpha reliability coefficient of 0.860, Guttman split-half coefficient of 0.805, with perfect content and construct validity. A CVS-Smart score of 7-10 points indicated the presence of CVS. Conclusions: The visual, ocular surface, and extraocular diagnostic criteria for CVS constituted the basic components of CVS-Smart. CVS-Smart is a novel, valid, reliable, subjective instrument for determining CVS diagnosis and prevalence and may provide a tool for rapid periodic assessment and prognostication. Individuals with positive CVS-Smart results should consider modifying their lifestyles and screen styles and seeking the help of ophthalmologists and/or optometrists. Higher institutional authorities should consider revising the Mandated Computer System Use Program to avoid the long-term consequences of CVS among university students. Further research must compare CVS-Smart with other available metrics for CVS, such as the CVS questionnaire, to determine its test-retest reliability and to justify its widespread use.

7.
Article in English | MEDLINE | ID: mdl-38978826

ABSTRACT

Background: Vascular endothelial growth factor (VEGF) is the primary substance involved in retinal barrier breach. VEGF overexpression may cause diabetic macular edema (DME). Laser photocoagulation of the macula is the standard treatment for DME; however, recently, intravitreal anti-VEGF injections have surpassed laser treatment. Our aim was to evaluate the efficacy of intravitreal injections of aflibercept or ranibizumab for managing treatment-naive DME. Methods: This single-center, retrospective, interventional, comparative study included eyes with visual impairment due to treatment-naive DME that underwent intravitreal injection of either aflibercept 2 mg/0.05 mL or ranibizumab 0.5 mg/0.05 mL at Al-Azhar University Hospitals, Egypt between March 2023 and January 2024. Demographic data and full ophthalmological examination results at baseline and 1, 3, and 6 months post-injection were collected, including the best-corrected distance visual acuity (BCDVA) in logarithm of the minimum angle of resolution (logMAR) notation, slit-lamp biomicroscopy, dilated fundoscopy, and central subfield thickness (CST) measured using spectral-domain optical coherence tomography. Results: Overall, the 96 eyes of 96 patients with a median (interquartile range [IQR]) age of 57 (10) (range: 20-74) years and a male-to-female ratio of 1:2.7 were allocated to one of two groups with comparable age, sex, diabetes mellitus duration, and presence of other comorbidities (all P >0.05). There was no statistically significant difference in baseline diabetic retinopathy status or DME type between groups (both P >0.05). In both groups, the median (IQR) BCDVA significantly improved from 0.7 (0.8) logMAR at baseline to 0.4 (0.1) logMAR at 6 months post-injection (both P = 0.001), with no statistically significant difference between groups at all follow-up visits (all P >0.05). The median (IQR) CST significantly decreased in the aflibercept group from 347 (166) µm at baseline to 180 (233) µm at 6 months post-injection, and it decreased in the ranibizumab group from 360 (180) µm at baseline to 190 (224) µm at 6 months post-injection (both P = 0.001), with no statistically significant differences between groups at all follow-up visits (all P >0.05). No serious adverse effects were documented in either group. Conclusions: Ranibizumab and aflibercept were equally effective in achieving the desired anatomical and functional results in patients with treatment-naïve DME in short-term follow-up without significant differences in injection counts between both drugs. Larger prospective, randomized, double-blinded trials with longer follow-up periods are needed to confirm our preliminary results.

8.
Article in English | MEDLINE | ID: mdl-38978827

ABSTRACT

Background: Diabetic retinopathy (DR), a sight-threatening ocular complication of diabetes mellitus, is one of the main causes of blindness in the working-age population. Dyslipidemia is a potential risk factor for the development or worsening of DR, with conflicting evidence in epidemiological studies. Fenofibrate, an antihyperlipidemic agent, has lipid-modifying and pleiotropic (non-lipid) effects that may lessen the incidence of microvascular events. Methods: Relevant studies were identified through a PubMed/MEDLINE search spanning the last 20 years, using the broad term "diabetic retinopathy" and specific terms "fenofibrate" and "dyslipidemia". References cited in these studies were further examined to compile this mini-review. These pivotal investigations underwent meticulous scrutiny and synthesis, focusing on methodological approaches and clinical outcomes. Furthermore, we provided the main findings of the seminal studies in a table to enhance comprehension and comparison. Results: Growing evidence indicates that fenofibrate treatment slows DR advancement owing to its possible protective effects on the blood-retinal barrier. The protective attributes of fenofibrate against DR progression and development can be broadly classified into two categories: lipid-modifying effects and non-lipid-related (pleiotropic) effects. The lipid-modifying effect is mediated through peroxisome proliferator-activated receptor-α activation, while the pleiotropic effects involve the reduction in serum levels of C-reactive protein, fibrinogen, and pro-inflammatory markers, and improvement in flow-mediated dilatation. In patients with DR, the lipid-modifying effects of fenofibrate primarily involve a reduction in lipoprotein-associated phospholipase A2 levels and the upregulation of apolipoprotein A1 levels. These changes contribute to the anti-inflammatory and anti-angiogenic effects of fenofibrate. Fenofibrate elicits a diverse array of pleiotropic effects, including anti-apoptotic, antioxidant, anti-inflammatory, and anti-angiogenic properties, along with the indirect consequences of these effects. Two randomized controlled trials-the Fenofibrate Intervention and Event Lowering in Diabetes and Action to Control Cardiovascular Risk in Diabetes studies-noted that fenofibrate treatment protected against DR progression, independent of serum lipid levels. Conclusions: Fenofibrate, an oral antihyperlipidemic agent that is effective in decreasing DR progression, may reduce the number of patients who develop vision-threatening complications and require invasive treatment. Despite its proven protection against DR progression, fenofibrate treatment has not yet gained wide clinical acceptance in DR management. Ongoing and future clinical trials may clarify the role of fenofibrate treatment in DR management.

9.
Article in English | MEDLINE | ID: mdl-38981809

ABSTRACT

This article discusses the role of computer vision in otolaryngology, particularly through endoscopy and surgery. It covers recent applications of artificial intelligence (AI) in nonradiologic imaging within otolaryngology, noting the benefits and challenges, such as improving diagnostic accuracy and optimizing therapeutic outcomes, while also pointing out the necessity for enhanced data curation and standardized research methodologies to advance clinical applications. Technical aspects are also covered, providing a detailed view of the progression from manual feature extraction to more complex AI models, including convolutional neural networks and vision transformers and their potential application in clinical settings.

10.
Data Brief ; 55: 110614, 2024 Aug.
Article in English | MEDLINE | ID: mdl-39015254

ABSTRACT

Battery technology is increasingly important for global electrification efforts. However, batteries are highly sensitive to small manufacturing variations that can induce reliability or safety issues. An important technology for battery quality control is computed tomography (CT) scanning, which is widely used for non-destructive 3D inspection across a variety of clinical and industrial applications. Historically, however, the utility of CT scanning for high-volume manufacturing has been limited by its low throughput as well as the difficulty of handling its large file sizes. In this work, we present a dataset of over one thousand CT scans of as-produced commercially available batteries. The dataset spans various chemistries (lithium-ion and sodium-ion) as well as various battery form factors (cylindrical, pouch, and prismatic). We evaluate seven different battery types in total. The manufacturing variability and the presence of battery defects can be observed via this dataset. This dataset may be of interest to scientists and engineers working on battery technology, computer vision, or both.

11.
Sensors (Basel) ; 24(13)2024 Jun 21.
Article in English | MEDLINE | ID: mdl-39000823

ABSTRACT

Unmanned aerial vehicle (UAV)-based object detection methods are widely used in traffic detection due to their high flexibility and extensive coverage. In recent years, with the increasing complexity of the urban road environment, UAV object detection algorithms based on deep learning have gradually become a research hotspot. However, how to further improve algorithmic efficiency in response to the numerous and rapidly changing road elements, and thus achieve high-speed and accurate road object detection, remains a challenging issue. Given this context, this paper proposes the high-efficiency multi-object detection algorithm for UAVs (HeMoDU). HeMoDU reconstructs a state-of-the-art, deep-learning-based object detection model and optimizes several aspects to improve computational efficiency and detection accuracy. To validate the performance of HeMoDU in urban road environments, this paper uses the public urban road datasets VisDrone2019 and UA-DETRAC for evaluation. The experimental results show that the HeMoDU model effectively improves the speed and accuracy of UAV object detection.

12.
Sensors (Basel) ; 24(13)2024 Jun 25.
Article in English | MEDLINE | ID: mdl-39000900

ABSTRACT

In recent years, the technological landscape has undergone a profound metamorphosis catalyzed by the widespread integration of drones across diverse sectors. Essential to the drone manufacturing process is comprehensive testing, typically conducted in controlled laboratory settings to uphold safety and privacy standards. However, a formidable challenge emerges due to the inherent limitations of GPS signals within indoor environments, posing a threat to the accuracy of drone positioning. This limitation not only jeopardizes testing validity but also introduces instability and inaccuracies, compromising the assessment of drone performance. Given the pivotal role of precise GPS-derived data in drone autopilots, addressing this indoor-based GPS constraint is imperative to ensure the reliability and resilience of unmanned aerial vehicles (UAVs). This paper delves into the implementation of an Indoor Positioning System (IPS) leveraging computer vision. The proposed system endeavors to detect and localize UAVs within indoor environments through an enhanced vision-based triangulation approach. A comparative analysis with alternative positioning methodologies is undertaken to ascertain the efficacy of the proposed system. The results obtained showcase the efficiency and precision of the designed system in detecting and localizing various types of UAVs, underscoring its potential to advance the field of indoor drone navigation and testing.

13.
Sensors (Basel) ; 24(13)2024 Jun 26.
Article in English | MEDLINE | ID: mdl-39000914

ABSTRACT

The acquisition of the body temperature of animals kept in captivity in biology laboratories is crucial for several studies in the field of animal biology. Traditionally, the acquisition process was carried out manually, which does not guarantee much accuracy or consistency in the acquired data and was painful for the animal. The process was then switched to a semi-manual process using a thermal camera, but it still involved manually clicking on each part of the animal's body every 20 s of the video to obtain temperature values, making it a time-consuming, non-automatic, and difficult process. This project aims to automate this acquisition process through the automatic recognition of parts of a lizard's body, reading the temperature in these parts based on a video taken with two cameras simultaneously: an RGB camera and a thermal camera. The first camera detects the location of the lizard's various body parts using artificial intelligence techniques, and the second camera allows reading of the respective temperature of each part. Due to the lack of lizard datasets, either in the biology laboratory or online, a dataset had to be created from scratch, containing the identification of the lizard and six of its body parts. YOLOv5 was used to detect the lizard and its body parts in RGB images, achieving a precision of 90.00% and a recall of 98.80%. After initial calibration, the RGB and thermal camera images are properly localised, making it possible to know the lizard's position, even when the lizard is at the same temperature as its surrounding environment, through a coordinate conversion from the RGB image to the thermal image. The thermal image has a colour temperature scale with the respective maximum and minimum temperature values, which is used to read each pixel of the thermal image, thus allowing the correct temperature to be read in each part of the lizard.


Subject(s)
Artificial Intelligence , Body Temperature , Lizards , Animals , Lizards/physiology , Body Temperature/physiology , Video Recording/methods , Image Processing, Computer-Assisted/methods
14.
Sensors (Basel) ; 24(13)2024 Jul 04.
Article in English | MEDLINE | ID: mdl-39001127

ABSTRACT

Compressive sensing (CS) is recognized for its adeptness at compressing signals, making it a pivotal technology in the context of sensor data acquisition. With the proliferation of image data in Internet of Things (IoT) systems, CS is expected to reduce the transmission cost of signals captured by various sensor devices. However, the quality of CS-reconstructed signals inevitably degrades as the sampling rate decreases, which poses a challenge in terms of the inference accuracy in downstream computer vision (CV) tasks. This limitation imposes an obstacle to the real-world application of existing CS techniques, especially for reducing transmission costs in sensor-rich environments. In response to this challenge, this paper contributes a CV-oriented adaptive CS framework based on saliency detection to the field of sensing technology that enables sensor systems to intelligently prioritize and transmit the most relevant data. Unlike existing CS techniques, the proposal prioritizes the accuracy of reconstructed images for CV purposes, not only for visual quality. The primary objective of this proposal is to enhance the preservation of information critical for CV tasks while optimizing the utilization of sensor data. This work conducts experiments on various realistic scenario datasets collected by real sensor devices. Experimental results demonstrate superior performance compared to existing CS sampling techniques across the STL10, Intel, and Imagenette datasets for classification and KITTI for object detection. Compared with the baseline uniform sampling technique, the average classification accuracy shows a maximum improvement of 26.23%, 11.69%, and 18.25%, respectively, at specific sampling rates. In addition, even at very low sampling rates, the proposal is demonstrated to be robust in terms of classification and detection as compared to state-of-the-art CS techniques. This ensures essential information for CV tasks is retained, improving the efficacy of sensor-based data acquisition systems.

15.
Sensors (Basel) ; 24(13)2024 Jul 05.
Article in English | MEDLINE | ID: mdl-39001152

ABSTRACT

The search for structural and microstructural defects using simple human vision is associated with significant errors in determining voids, large pores, and violations of the integrity and compactness of particle packing in the micro- and macrostructure of concrete. Computer vision methods, in particular convolutional neural networks, have proven to be reliable tools for the automatic detection of defects during visual inspection of building structures. The study's objective is to create and compare computer vision algorithms that use convolutional neural networks to identify and analyze damaged sections in concrete samples from different structures. Networks of the following architectures were selected for operation: U-Net, LinkNet, and PSPNet. The analyzed images are photos of concrete samples obtained by laboratory tests to assess the quality in terms of the defection of the integrity and compactness of the structure. During the implementation process, changes in quality metrics such as macro-averaged precision, recall, and F1-score, as well as IoU (Jaccard coefficient) and accuracy, were monitored. The best metrics were demonstrated by the U-Net model, supplemented by the cellular automaton algorithm: precision = 0.91, recall = 0.90, F1 = 0.91, IoU = 0.84, and accuracy = 0.90. The developed segmentation algorithms are universal and show a high quality in highlighting areas of interest under any shooting conditions and different volumes of defective zones, regardless of their localization. The automatization of the process of calculating the damage area and a recommendation in the "critical/uncritical" format can be used to assess the condition of concrete of various types of structures, adjust the formulation, and change the technological parameters of production.

16.
Neurosurg Rev ; 47(1): 327, 2024 Jul 15.
Article in English | MEDLINE | ID: mdl-39004685

ABSTRACT

With the current artificial intelligence (AI) boom, new innovative and accessible applications requiring minimal computer science expertise have been developed for discipline specific and mainstream purposes. Apple Intelligence, a new AI model developed by Apple, aims to enhance user experiences with new functionalities across many of its product offerings. Although designed for the everyday user, many of these advances have potential applications in neurosurgery. These include functionalities for writing, image generation, and upgraded integrations to the voice command assistant Siri. Future integrations may also include other Apple products such as the vision pro for preoperative and intraoperative applications. Considering the popularity of Apple products, particularly the iPhone, it is important to appraise this new technology and how it can be leveraged to enhance patient care, improve neurosurgical education, and facilitate more efficiency for the neurosurgeon.


Subject(s)
Artificial Intelligence , Neurosurgery , Neurosurgical Procedures , Humans , Neurosurgical Procedures/methods
17.
Front Transplant ; 3: 1305468, 2024.
Article in English | MEDLINE | ID: mdl-38993786

ABSTRACT

Two common obstacles limiting the performance of data-driven algorithms in digital histopathology classification tasks are the lack of expert annotations and the narrow diversity of datasets. Multi-instance learning (MIL) can address the former challenge for the analysis of whole slide images (WSI), but performance is often inferior to full supervision. We show that the inclusion of weak annotations can significantly enhance the effectiveness of MIL while keeping the approach scalable. An analysis framework was developed to process periodic acid-Schiff (PAS) and Sirius Red (SR) slides of renal biopsies. The workflow segments tissues into coarse tissue classes. Handcrafted and deep features were extracted from these tissues and combined using a soft attention model to predict several slide-level labels: delayed graft function (DGF), acute tubular injury (ATI), and Remuzzi grade components. A tissue segmentation quality metric was also developed to reduce the adverse impact of poorly segmented instances. The soft attention model was trained using 5-fold cross-validation on a mixed dataset and tested on the QUOD dataset containing n = 373 PAS and n = 195 SR biopsies. The average ROC-AUC over different prediction tasks was found to be 0.598 ± 0.011 , significantly higher than using only ResNet50 ( 0.545 ± 0.012 ), only handcrafted features ( 0.542 ± 0.011 ), and the baseline ( 0.532 ± 0.012 ) of state-of-the-art performance. In conjunction with soft attention, weighting tissues by segmentation quality has led to further improvement ( A U C = 0.618 ± 0.010 ) . Using an intuitive visualisation scheme, we show that our approach may also be used to support clinical decision making as it allows pinpointing individual tissues relevant to the predictions.

18.
Artif Intell Med ; 154: 102923, 2024 Jun 27.
Article in English | MEDLINE | ID: mdl-38970987

ABSTRACT

Computerized cognitive training (CCT) is a scalable, well-tolerated intervention that has promise for slowing cognitive decline. The effectiveness of CCT is often affected by a lack of effective engagement. Mental fatigue is a the primary factor for compromising effective engagement in CCT, particularly in older adults at risk for dementia. There is a need for scalable, automated measures that can constantly monitor and reliably detect mental fatigue during CCT. Here, we develop and validate a novel Recurrent Video Transformer (RVT) method for monitoring real-time mental fatigue in older adults with mild cognitive impairment using their video-recorded facial gestures during CCT. The RVT model achieved the highest balanced accuracy (79.58%) and precision (0.82) compared to the prior models for binary and multi-class classification of mental fatigue. We also validated our model by significantly relating to reaction time across CCT tasks (Waldχ2=5.16,p=0.023). By leveraging dynamic temporal information, the RVT model demonstrates the potential to accurately measure real-time mental fatigue, laying the foundation for future CCT research aiming to enhance effective engagement by timely prevention of mental fatigue.

19.
Artif Organs ; 2024 Jul 17.
Article in English | MEDLINE | ID: mdl-39016696

ABSTRACT

BACKGROUND: EXCOR Pediatric is one of the most commonly used ventricular assist devices (VAD) for small children; it requires visual inspection of the diaphragm movement to assess its operating status. Although this visual inspection can only be performed by trained medical professionals, it can also be attempted by the recent advances in computer vision technology. METHODS: Movement of the diaphragm in the operating EXCOR VAD was recorded as movies and annotated frame-by-frame in three classes according to the state of the diaphragm: "fill," "mid," and "empty." Three models, MobileNetV3, EfficientNetV2, and MobileViT, were trained using the frames, and their performance was evaluated based on the accuracy and area under the receiver operating characteristic curve (AROC). RESULTS: A total of 152 movies were available from two participants. Only the 10 mL pumps were used. Ninety-eight movies were used for annotation and frame extraction, and 7949 frames per class were included in the analysis. The macro-average accuracies of the three models were 0.88, 0.91, and 0.93, and the AROC were 0.99, 0.98, and 0.99 for MobileNetV3, EfficientNetV2, and MobileViT, respectively. CONCLUSION: Image recognition models based on lightweight deep neural networks could detect the diaphragm state of EXCOR VAD with sufficient accuracy, although there were limited variations in the dataset. This suggests the potential of computer vision for the automated monitoring of the EXCOR diaphragm position.

20.
Animals (Basel) ; 14(13)2024 Jun 21.
Article in English | MEDLINE | ID: mdl-38997962

ABSTRACT

Aquaculture requires precise non-invasive methods for biomass estimation. This research validates a novel computer vision methodology that uses a signature function-based feature extraction algorithm combining statistical morphological analysis of the size and shape of fish and machine learning to improve the accuracy of biomass estimation in fishponds and is specifically applied to tilapia (Oreochromis niloticus). These features that are automatically extracted from images are put to the test against previously manually extracted features by comparing the results when applied to three common machine learning methods under two different lighting conditions. The dataset for this analysis encompasses 129 tilapia samples. The results give promising outcomes since the multilayer perceptron model shows robust performance, consistently demonstrating superior accuracy across different features and lighting conditions. The interpretable nature of the model, rooted in the statistical features of the signature function, could provide insights into the morphological and allometric changes at different developmental stages. A comparative analysis against existing literature underscores the competitiveness of the proposed methodology, pointing to advancements in precision, interpretability, and species versatility. This research contributes significantly to the field, accelerating the quest for non-invasive fish biometrics that can be generalized across various aquaculture species in different stages of development. In combination with detection, tracking, and posture recognition, deep learning methodologies such as the one provided in the latest studies could generate a powerful method for real-time fish morphology development, biomass estimation, and welfare monitoring, which are crucial for the effective management of fish farms.

SELECTION OF CITATIONS
SEARCH DETAIL