Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Results 1 - 20 de 3.539
Filter
1.
Spectrochim Acta A Mol Biomol Spectrosc ; 324: 125010, 2025 Jan 05.
Article in English | MEDLINE | ID: mdl-39216368

ABSTRACT

Lithium, a rare metal of strategic importance, has garnered heightened global attention. This investigation delves into the laboratory visible-near infrared and short-wavelength infrared reflectance (VNIR-SWIR 350 nm-2500 nm) spectral properties of lithium-rich rocks and stream sediments, aiming to elucidate their quantitative relationship with lithium concentration. This research seeks to pave new avenues and furnish innovative technical solutions for probing sedimentary lithium reserves. Conducted in the Tuanjie Peak region of Western Kunlun, Xinjiang, China, this study analyzed 614 stream sediments and 222 rock specimens. Initial steps included laboratory VNIR-SWIR spectral reflectance measurements and lithium quantification. Following the preprocessing of spectral data via Savitzky-Golay (SG) smoothing and continuum removal (CR), the absorption positions (Pos2210nm, Pos1910nm) and depths (Depth2210, Depth1910) in the rock spectra, as well as the Illite Spectral Maturity (ISM) of the rock samples, were extracted. Employing both the Successive Projections Algorithm (SPA) and genetic algorithm (GA), wavelengths indicative of lithium content were identified. Integrating the lithium-sensitive wavelengths identified by these feature selection methods, A quantitative predictive regression model for lithium content in rock and stream sediments was developed using partial least squares regression (PLSR), support vector regression (SVR), and convolutional neural network (CNN). Spectral analysis indicated that lithium is predominantly found in montmorillonite and illite, with its content positively correlating with the spectral maturity of illite and closely related to Al-OH absorption depth (Depth2210) and clay content. The SPA algorithm was more effective than GA in extracting lithium-sensitive bands. The optimal regression model for quantitative prediction of lithium content in rock samples was SG-SPA-CNN, with a correlation coefficient prediction (Rp) of 0.924 and root-mean-square error prediction (RMSEP) of 0.112. The optimal model for the prediction of lithium content in stream sediment was SG-SPA-CNN, with an Rp and RMSEP of 0.881 and 0.296, respectively. The higher prediction accuracy for lithium content in rocks compared to sediments indicates that rocks are a more suitable medium for predicting lithium content. Compared to the PLSR and SVR models, the CNN model performs better in both sample types. Despite the limitations, this study highlights the effectiveness of hyperspectral technology in exploring the potential of clay-type lithium resources in the Tuanjie Peak area, offering new perspectives and approaches for further exploration.

2.
Med Sci (Basel) ; 12(3)2024 Sep 20.
Article in English | MEDLINE | ID: mdl-39311162

ABSTRACT

Osteoporosis, a skeletal disorder, is expected to affect 60% of women aged over 50 years. Dual-energy X-ray absorptiometry (DXA) scans, the current gold standard, are typically used post-fracture, highlighting the need for early detection tools. Panoramic radiographs (PRs), common in annual dental evaluations, have been explored for osteoporosis detection using deep learning, but methodological flaws have cast doubt on otherwise optimistic results. This study aims to develop a robust artificial intelligence (AI) application for accurate osteoporosis identification in PRs, contributing to early and reliable diagnostics. A total of 250 PRs from three groups (A: osteoporosis group, B: non-osteoporosis group matching A in age and gender, C: non-osteoporosis group differing from A in age and gender) were cropped to the mental foramen region. A pretrained convolutional neural network (CNN) classifier was used for training, testing, and validation with a random split of the dataset into subsets (A vs. B, A vs. C). Detection accuracy and area under the curve (AUC) were calculated. The method achieved an F1 score of 0.74 and an AUC of 0.8401 (A vs. B). For young patients (A vs. C), it performed with 98% accuracy and an AUC of 0.9812. This study presents a proof-of-concept algorithm, demonstrating the potential of deep learning to identify osteoporosis in dental radiographs. It also highlights the importance of methodological rigor, as not all optimistic results are credible.


Subject(s)
Artificial Intelligence , Deep Learning , Osteoporosis , Radiography, Panoramic , Humans , Osteoporosis/diagnostic imaging , Female , Middle Aged , Aged , Male , Neural Networks, Computer , Absorptiometry, Photon , Mandible/diagnostic imaging
3.
PeerJ Comput Sci ; 10: e2281, 2024.
Article in English | MEDLINE | ID: mdl-39314680

ABSTRACT

The evolution of social intelligence has led to the adoption of intelligent accounting practices in enterprises. To enhance the efficiency of enterprise accounting operations and improve the capabilities of accountants, we propose an intelligent accounting optimization approach that integrates meta-heuristic algorithms with convolutional neural networks (CNN). First, we enhance the CNN framework by incorporating document and voucher information into accounting audits, creating a multi-modal feature extraction mechanism. Utilizing these multi-modal accounting features, we then introduce a method for assessing accounting quality, which objectively evaluates financial performance. Finally, we propose an optimization technique based on meta-heuristic principles, combining genetic algorithms with annealing models to improve the accounting system. Experimental results validate our approach, demonstrating an accuracy of 0.943 and a mean average precision (mAP) score of 0.812. This method provides technological support for refining accounting audit mechanisms.

4.
PeerJ Comput Sci ; 10: e2052, 2024.
Article in English | MEDLINE | ID: mdl-39314724

ABSTRACT

Most natural disasters result from geodynamic events such as landslides and slope collapse. These failures cause catastrophes that directly impact the environment and cause financial and human losses. Visual inspection is the primary method for detecting failures in geotechnical structures, but on-site visits can be risky due to unstable soil. In addition, the body design and hostile and remote installation conditions make monitoring these structures inviable. When a fast and secure evaluation is required, analysis by computational methods becomes feasible. In this study, a convolutional neural network (CNN) approach to computer vision is applied to identify defects in the surface of geotechnical structures aided by unmanned aerial vehicle (UAV) and mobile devices, aiming to reduce the reliance on human-led on-site inspections. However, studies in computer vision algorithms still need to be explored in this field due to particularities of geotechnical engineering, such as limited public datasets and redundant images. Thus, this study obtained images of surface failure indicators from slopes near a Brazilian national road, assisted by UAV and mobile devices. We then proposed a custom CNN and low complexity model architecture to build a binary classifier image-aided to detect faults in geotechnical surfaces. The model achieved a satisfactory average accuracy rate of 94.26%. An AUC metric score of 0.99 from the receiver operator characteristic (ROC) curve and matrix confusion with a testing dataset show satisfactory results. The results suggest that the capability of the model to distinguish between the classes 'damage' and 'intact' is excellent. It enables the identification of failure indicators. Early failure indicator detection on the surface of slopes can facilitate proper maintenance and alarms and prevent disasters, as the integrity of the soil directly affects the structures built around and above it.

5.
Heliyon ; 10(18): e37167, 2024 Sep 30.
Article in English | MEDLINE | ID: mdl-39315129

ABSTRACT

With the advancement of technology in recent years, the application of artificial intelligence in real life has become more extensive. Graphic recognition is a hot spot in the current research of related technologies. It involves machines extracting key information from pictures and combining it with natural language processing for in-depth understanding. Existing methods still have obvious deficiencies in fine-grained recognition and deep understanding of contextual context. Addressing these issues to achieve high-quality image-text recognition is crucial for various application scenarios, such as accessibility technologies, content creation, and virtual assistants. To tackle this challenge, a novel approach is proposed that combines the Mask R-CNN, DCGAN, and ALBERT models. Specifically, the Mask R-CNN specializes in high-precision image recognition and segmentation, the DCGAN captures and generates nuanced features from images, and the ALBERT model is responsible for deep natural language processing and semantic understanding of this visual information. Experimental results clearly validate the superiority of this method. Compared to traditional image-text recognition techniques, the recognition accuracy is improved from 85.3% to 92.5%, and performance in contextual and situational understanding is enhanced. The advancement of this technology has far-reaching implications for research in machine vision and natural language processing and open new possibilities for practical applications.

6.
Curr Med Imaging ; 2024 Sep 18.
Article in English | MEDLINE | ID: mdl-39297463

ABSTRACT

BACKGROUND: Brain tumours represent a diagnostic challenge, especially in the imaging area, where the differentiation of normal and pathologic tissues should be precise. The use of up-to-date machine learning techniques would be of great help in terms of brain tumor identification accuracy from MRI data. Objective This research paper aims to check the efficiency of a federated learning method that joins two classifiers, such as convolutional neural networks (CNNs) and random forests (R.F.F.), with dual U-Net segmentation for federated learning. This procedure benefits the image identification task on preprocessed MRI scan pictures that have already been categorized. METHODS: In addition to using a variety of datasets, federated learning was utilized to train the CNN-RF model while taking data privacy into account. The processed MRI images with Median, Gaussian, and Wiener filters are used to filter out the noise level and make the feature extraction process easy and efficient. The surgical part used a dual U-Net layout, and the performance assessment was based on precision, recall, F1-score, and accuracy. RESULTS: The model achieved excellent classification performance on local datasets as CRPs were high, from 91.28% to 95.52% for macro, micro, and weighted averages. Throughout the process of federated averaging, the collective model outperformed by reaching 97% accuracy compared to those of 99%, which were subjected to different clients. The correctness of how data is used helps the federated averaging method convert individual model insights into a consistent global model while keeping all personal data private. CONCLUSION: The combined structure of the federated learning framework, CNN-RF hybrid model, and dual U-Net segmentation is a robust and privacypreserving approach for identifying MRI images from brain tumors. The results of the present study exhibited that the technique is promising in improving the quality of brain tumor categorization and provides a pathway for practical utilization in clinical settings.

7.
BMC Med Imaging ; 24(1): 253, 2024 Sep 20.
Article in English | MEDLINE | ID: mdl-39304839

ABSTRACT

BACKGROUND: Breast cancer is one of the leading diseases worldwide. According to estimates by the National Breast Cancer Foundation, over 42,000 women are expected to die from this disease in 2024. OBJECTIVE: The prognosis of breast cancer depends on the early detection of breast micronodules and the ability to distinguish benign from malignant lesions. Ultrasonography is a crucial radiological imaging technique for diagnosing the illness because it allows for biopsy and lesion characterization. The user's level of experience and knowledge is vital since ultrasonographic diagnosis relies on the practitioner's expertise. Furthermore, computer-aided technologies significantly contribute by potentially reducing the workload of radiologists and enhancing their expertise, especially when combined with a large patient volume in a hospital setting. METHOD: This work describes the development of a hybrid CNN system for diagnosing benign and malignant breast cancer lesions. The models InceptionV3 and MobileNetV2 serve as the foundation for the hybrid framework. Features from these models are extracted and concatenated individually, resulting in a larger feature set. Finally, various classifiers are applied for the classification task. RESULTS: The model achieved the best results using the softmax classifier, with an accuracy of over 95%. CONCLUSION: Computer-aided diagnosis greatly assists radiologists and reduces their workload. Therefore, this research can serve as a foundation for other researchers to build clinical solutions.


Subject(s)
Breast Neoplasms , Ultrasonography, Mammary , Humans , Female , Breast Neoplasms/diagnostic imaging , Ultrasonography, Mammary/methods , Neural Networks, Computer , Image Interpretation, Computer-Assisted/methods , Diagnosis, Computer-Assisted/methods
9.
Heliyon ; 10(18): e37743, 2024 Sep 30.
Article in English | MEDLINE | ID: mdl-39309774

ABSTRACT

An early identification and subsequent management of cerebral small vessel disease (cSVD) grade 1 can delay progression into grades II and III. Machine learning algorithms have shown considerable promise in medical image interpretation automation. An experimental cross-sectional study aimed to develop an automated computer-aided diagnostic system based on AI (artificial intelligence) tools to detect grade 1-cSVD with improved accuracy. Patients with Fazekas grade 1 cSVD on Non-Contrast Magnetic Resonance Imaging (MRI) Brain of age >40 years of both genders were included. The dataset was pre-processed to be fed into a 3D convolutional neural network (CNN) model. A 3D stack with the shape (120, 128, 128, 1) containing axial slices from the brain magnetic resonance image was created. The model was created from scratch and contained four convolutional and three fully connected (FC) layers. The dataset was preprocessed by making a 3D stack, and normalizing, resizing, and completing the stack was performed. A 3D-CNN model architecture was designed to train and test preprocessed images. We achieved an accuracy of 93.12 % when 2D axial slices were used. When the 2D slices of a patient were stacked to form a 3D image, an accuracy of 85.71 % was achieved on the test set. Overall, the 3D-CNN model performed very well on the test set. The earliest and the most accurate diagnosis from computational imaging methods can help reduce the huge burden of cSVD and its associated morbidity in the form of vascular dementia.

10.
Heliyon ; 10(18): e36694, 2024 Sep 30.
Article in English | MEDLINE | ID: mdl-39309932

ABSTRACT

The agricultural sector in Bangladesh is a cornerstone of the nation's economy, with key crops such as rice, corn, wheat, potato, and tomato playing vital roles. However, these crops are highly vulnerable to various leaf diseases, which pose significant threats to crop yields and food security if not promptly addressed. Consequently, there is an urgent need for an automated system that can accurately identify and categorize leaf diseases, enabling early intervention and management. This study explores the efficacy of the latest state-of-the-art object detection model, YOLOv8 (You Only Look Once), in surpassing previous models for the automated detection and categorization of leaf diseases in these five major crops. By leveraging modern computer vision techniques, the goal is to enhance the efficiency of disease detection and management. A dataset comprising 19 classes, each with 150 images, totaling 2850 images, was meticulously curated and annotated for training and evaluation. The YOLOv8 framework, known for its capability to detect multiple objects simultaneously, was employed to train a deep neural network. The system's performance was evaluated using standard metrics such as mean Average Precision (mAP) and F1 score. The findings demonstrate that the YOLOv8 framework successfully identifies leaf diseases, achieving a high mAP of 98% and an F1 score of 97%. These results underscore the significant potential of this approach to enhance crop disease management, thereby improving food security and promoting agricultural sustainability in Bangladesh.

11.
Yi Chuan ; 46(9): 701-715, 2024 Sep.
Article in English | MEDLINE | ID: mdl-39275870

ABSTRACT

With the rapid growth of data driven by high-throughput sequencing technologies, genomics has entered an era characterized by big data, which presents significant challenges for traditional bioinformatics methods in handling complex data patterns. At this critical juncture of technological progress, deep learning-an advanced artificial intelligence technology-offers powerful capabilities for data analysis and pattern recognition, revitalizing genomic research. In this review, we focus on four major deep learning models: Convolutional Neural Network(CNN), Recurrent Neural Network(RNN), Long Short-Term Memory(LSTM), and Generative Adversarial Network(GAN). We outline their core principles and provide a comprehensive review of their applications in DNA, RNA, and protein research over the past five years. Additionally, we also explore the use of deep learning in livestock genomics, highlighting its potential benefits and challenges in genetic trait analysis, disease prevention, and genetic enhancement. By delivering a thorough analysis, we aim to enhance precision and efficiency in genomic research through deep learning and offer a framework for developing and applying livestock genomic strategies, thereby advancing precision livestock farming and genetic breeding technologies.


Subject(s)
Deep Learning , Genomics , Genomics/methods , Animals , Neural Networks, Computer , Livestock/genetics , Humans , Computational Biology/methods
12.
Animals (Basel) ; 14(17)2024 Aug 27.
Article in English | MEDLINE | ID: mdl-39272273

ABSTRACT

Ovine pulmonary adenocarcinoma (OPA) is a contagious lung tumour caused by the Jaagsiekte Sheep Retrovirus (JSRV). Histopathological diagnosis is the gold standard for OPA diagnosis. However, interpretation of traditional pathology images is complex and operator dependent. The mask regional convolutional neural network (Mask R-CNN) has emerged as a valuable tool in pathological diagnosis. This study utilized 54 typical OPA whole slide images (WSI) to extract 7167 typical lesion images containing OPA to construct a Common Objects in Context (COCO) dataset for OPA pathological images. The dataset was categorized into training and test sets (8:2 ratio) for model training and validation. Mean average specificity (mASp) and average sensitivity (ASe) were used to evaluate model performance. Six WSI-level pathological images (three OPA and three non-OPA images), not included in the dataset, were used for anti-peeking model validation. A random selection of 500 images, not included in the dataset establishment, was used to compare the performance of the model with assessment by pathologists. Accuracy, sensitivity, specificity, and concordance rate were evaluated. The model achieved a mASp of 0.573 and an ASe of 0.745, demonstrating effective lesion detection and alignment with expert annotation. In Anti-Peeking verification, the model showed good performance in locating OPA lesions and distinguished OPA from non-OPA pathological images. In the random 500-image diagnosis, the model achieved 92.8% accuracy, 100% sensitivity, and 88% specificity. The agreement rates between junior and senior pathologists were 100% and 96.5%, respectively. In conclusion, the Mask R-CNN-based OPA diagnostic model developed for OPA facilitates rapid and accurate diagnosis in practical applications.

13.
Heliyon ; 10(16): e36411, 2024 Aug 30.
Article in English | MEDLINE | ID: mdl-39253213

ABSTRACT

This study introduces a groundbreaking method to enhance the accuracy and reliability of emotion recognition systems by combining electrocardiogram (ECG) with electroencephalogram (EEG) data, using an eye-tracking gated strategy. Initially, we propose a technique to filter out irrelevant portions of emotional data by employing pupil diameter metrics from eye-tracking data. Subsequently, we introduce an innovative approach for estimating effective connectivity to capture the dynamic interaction between the brain and the heart during emotional states of happiness and sadness. Granger causality (GC) is estimated and utilized to optimize input for a highly effective pre-trained convolutional neural network (CNN), specifically ResNet-18. To assess this methodology, we employed EEG and ECG data from the publicly available MAHNOB-HCI database, using a 5-fold cross-validation approach. Our method achieved an impressive average accuracy and area under the curve (AUC) of 91.00 % and 0.97, respectively, for GC-EEG-ECG images processed with ResNet-18. Comparative analysis with state-of-the-art studies clearly shows that augmenting ECG with EEG and refining data with an eye-tracking strategy significantly enhances emotion recognition performance across various emotions.

14.
Neural Netw ; 180: 106673, 2024 Aug 30.
Article in English | MEDLINE | ID: mdl-39260009

ABSTRACT

Image harmonization seeks to transfer the illumination distribution of the background to that of the foreground within a composite image. Existing methods lack the ability of establishing global-local pixel illumination dependencies between foreground and background of composite images, which is indispensable for sharp and color-consistent harmonized image generation. To overcome this challenge, we design a novel Simple Hybrid CNN-Transformer Network (SHT-Net), which is formulated into an efficient symmetrical hierarchical architecture. It is composed of two newly designed light-weight Transformer blocks. Firstly, the scale-aware gated block is designed to capture multi-scale features through different heads and expand the receptive fields, which facilitates to generate images with fine-grained details. Secondly, we introduce a simple parallel attention block, which integrates the window-based self-attention and gated channel attention in parallel, resulting in simultaneously global-local pixel illumination relationship modeling capability. Besides, we propose an efficient simple feed forward network to filter out less informative features and allow the features to contribute to generating photo-realistic harmonized results passing through. Extensive experiments on image harmonization benchmarks indicate that our method achieve promising quantitative and qualitative results. The code and pre-trained models are available at https://github.com/guanguanboy/SHT-Net.

15.
J Insur Med ; 51(2): 64-76, 2024 Jul 01.
Article in English | MEDLINE | ID: mdl-39266002

ABSTRACT

Recent artificial intelligence (AI) advancements in cardiovascular medicine offer potential enhancements in diagnosis, prediction, treatment, and outcomes. This article aims to provide a basic understanding of AI enabled ECG technology. Specific conditions and findings will be discussed, followed by reviewing associated terminology and methodology. In the appendix, definitions of AUC versus accuracy are explained. The application of deep learning models enables detecting diseases from normal electrocardiograms at accuracy not previously achieved by technology or human experts. Results with AI enabled ECG are encouraging as they considerably exceeded current screening models for specific conditions (i.e., atrial fibrillation, left ventricular dysfunction, aortic stenosis, and hypertrophic cardiomyopathy). This could potentially lead to a revitalization of the utilization of the ECG in the insurance domain. While we are embracing the findings with this rapidly evolving technology, but cautious optimism is still necessary at this point.


Subject(s)
Artificial Intelligence , Electrocardiography , Humans , Electrocardiography/methods , Deep Learning , Atrial Fibrillation/diagnosis
16.
Biol Methods Protoc ; 9(1): bpae063, 2024.
Article in English | MEDLINE | ID: mdl-39258158

ABSTRACT

Deep learning applications in taxonomic classification for animals and plants from images have become popular, while those for microorganisms are still lagging behind. Our study investigated the potential of deep learning for the taxonomic classification of hundreds of filamentous fungi from colony images, which is typically a task that requires specialized knowledge. We isolated soil fungi, annotated their taxonomy using standard molecular barcode techniques, and took images of the fungal colonies grown in petri dishes (n = 606). We applied a convolutional neural network with multiple training approaches and model architectures to deal with some common issues in ecological datasets: small amounts of data, class imbalance, and hierarchically structured grouping. Model performance was overall low, mainly due to the relatively small dataset, class imbalance, and the high morphological plasticity exhibited by fungal colonies. However, our approach indicates that morphological features like color, patchiness, and colony extension rate could be used for the recognition of fungal colonies at higher taxonomic ranks (i.e. phylum, class, and order). Model explanation implies that image recognition characters appear at different positions within the colony (e.g. outer or inner hyphae) depending on the taxonomic resolution. Our study suggests the potential of deep learning applications for a better understanding of the taxonomy and ecology of filamentous fungi amenable to axenic culturing. Meanwhile, our study also highlights some technical challenges in deep learning image analysis in ecology, highlighting that the domain of applicability of these methods needs to be carefully considered.

17.
Comput Biol Med ; 182: 109092, 2024 Sep 09.
Article in English | MEDLINE | ID: mdl-39255658

ABSTRACT

BACKGROUND: Detecting Attention-Deficit/Hyperactivity Disorder (ADHD) in children is crucial for timely intervention and personalized treatment. OBJECTIVE: This study aims to utilize deep learning techniques to analyze brain maps derived from Power Spectral Density (PSD) of Electroencephalography (EEG) signals in pediatric subjects for ADHD detection. METHODS: We employed a Siamese-based Convolutional Neural Network (CNN) to analyze EEG-based brain maps. Gradient-weighted class activation mapping (Grad-CAM) was used as an explainable AI (XAI) visualization method to identify significant features. RESULTS: The CNN model achieved a high classification accuracy of 99.17 %. Grad-CAM analysis revealed that PSD features from the theta band of the frontal and occipital lobes are effective discriminators for distinguishing children with ADHD from healthy controls. CONCLUSION: This study demonstrates the effectiveness of deep learning in ADHD detection and highlights the importance of regional PSD metrics in accurate classification. By utilizing Grad-CAM, we elucidate the discriminative power of specific brain regions and frequency bands, thereby enhancing the understanding of ADHD neurobiology for improved diagnostic precision in pediatric populations.

18.
Sci Rep ; 14(1): 21643, 2024 09 16.
Article in English | MEDLINE | ID: mdl-39284813

ABSTRACT

The main bottleneck in training a robust tumor segmentation algorithm for non-small cell lung cancer (NSCLC) on H&E is generating sufficient ground truth annotations. Various approaches for generating tumor labels to train a tumor segmentation model was explored. A large dataset of low-cost low-accuracy panCK-based annotations was used to pre-train the model and determine the minimum required size of the expensive but highly accurate pathologist annotations dataset. PanCK pre-training was compared to foundation models and various architectures were explored for model backbone. Proper study design and sample procurement for training a generalizable model that captured variations in NSCLC H&E was studied. H&E imaging was performed on 112 samples (three centers, two scanner types, different staining and imaging protocols). Attention U-Net architecture was trained using the large panCK-based annotations dataset (68 samples, total area 10,326 [mm2]) followed by fine-tuning using a small pathologist annotations dataset (80 samples, total area 246 [mm2]). This approach resulted in mean intersection over union (mIoU) of 82% [77 87]. Using panCK pretraining provided better performance compared to foundation models and allowed for 70% reduction in pathologist annotations with no drop in performance. Study design ensured model generalizability over variations on H&E where performance was consistent across centers, scanners, and subtypes.


Subject(s)
Carcinoma, Non-Small-Cell Lung , Deep Learning , Lung Neoplasms , Pathologists , Humans , Lung Neoplasms/pathology , Carcinoma, Non-Small-Cell Lung/pathology , Image Processing, Computer-Assisted/methods , Algorithms
19.
Sci Rep ; 14(1): 21525, 2024 09 14.
Article in English | MEDLINE | ID: mdl-39277634

ABSTRACT

Manual identification of tomato leaf diseases is a time-consuming and laborious process that may lead to inaccurate results without professional assistance. Therefore, an automated, early, and precise leaf disease recognition system is essential for farmers to ensure the quality and quantity of tomato production by providing timely interventions to mitigate disease spread. In this study, we have proposed seven robust Bayesian optimized deep hybrid learning models leveraging the synergy between deep learning and machine learning for the automated classification of ten types of tomato leaves (nine diseased and one healthy). We customized the popular Convolutional Neural Network (CNN) algorithm for automatic feature extraction due to its ability to capture spatial hierarchies of features directly from raw data and classical machine learning techniques [Random Forest (RF), XGBoost, GaussianNB (GNB), Support Vector Machines (SVM), Multinomial Logistic Regression (MLR), K-Nearest Neighbor (KNN)], and stacking for classifications. Additionally, the study incorported a Boruta feature filtering layer to capture the statistically significant features. The standard, research-oriented PlantVillage dataset was used for the performance testing, which facilitates benchmarking against prior research and enables meaningful comparisons of classification performance across different approaches. We utilized a variety of statistical classification metrics to demonstrate the robustness of our models. Using the CNN-Stacking model, this study achieved the highest classification performance among the seven hybrid models. On an unseen dataset, this model achieved average precision, recall, f1-score, mcc, and accuracy values of 98.527%, 98.533%, 98.527%, 98.525%, and 98.268%, respectively. Our study requires only 0.174 s of testing time to correctly identify noisy, blurry, and transformed images. This indicates our approach's time efficiency and generalizability in images captured under challenging lighting conditions and with complex backgrounds. Based on the comparative analysis, our approach is superior and computationally inexpensive compared to the existing studies. This work will aid in developing a smartphone app to offer farmers a real-time disease diagnosis tool and management strategies.


Subject(s)
Bayes Theorem , Deep Learning , Plant Diseases , Plant Leaves , Solanum lycopersicum , Algorithms , Support Vector Machine , Neural Networks, Computer , Machine Learning
20.
JMIR Form Res ; 8: e57335, 2024 Sep 03.
Article in English | MEDLINE | ID: mdl-39226096

ABSTRACT

BACKGROUND: Artificial intelligence (AI) models are being increasingly studied for the detection of variations and pathologies in different imaging modalities. Nasal septal deviation (NSD) is an important anatomical structure with clinical implications. However, AI-based radiographic detection of NSD has not yet been studied. OBJECTIVE: This research aimed to develop and evaluate a real-time model that can detect probable NSD using cone beam computed tomography (CBCT) images. METHODS: Coronal section images were obtained from 204 full-volume CBCT scans. The scans were classified as normal and deviated by 2 maxillofacial radiologists. The images were then used to train and test the AI model. Mask region-based convolutional neural networks (Mask R-CNNs) comprising 3 different backbones-ResNet50, ResNet101, and MobileNet-were used to detect deviated nasal septum in 204 CBCT images. To further improve the detection, an image preprocessing technique (contrast enhancement [CEH]) was added. RESULTS: The best-performing model-CEH-ResNet101-achieved a mean average precision of 0.911, with an area under the curve of 0.921. CONCLUSIONS: The performance of the model shows that the model is capable of detecting nasal septal deviation. Future research in this field should focus on additional preprocessing of images and detection of NSD based on multiple planes using 3D images.


Subject(s)
Cone-Beam Computed Tomography , Nasal Septum , Neural Networks, Computer , Proof of Concept Study , Humans , Cone-Beam Computed Tomography/methods , Nasal Septum/diagnostic imaging , Female , Male , Adult , Middle Aged
SELECTION OF CITATIONS
SEARCH DETAIL