Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Results 1 - 20 de 162
Filter
1.
J Imaging ; 10(8)2024 Aug 13.
Article in English | MEDLINE | ID: mdl-39194986

ABSTRACT

Currently, existing deep learning methods exhibit many limitations in multi-target detection, such as low accuracy and high rates of false detection and missed detections. This paper proposes an improved Faster R-CNN algorithm, aiming to enhance the algorithm's capability in detecting multi-scale targets. This algorithm has three improvements based on Faster R-CNN. Firstly, the new algorithm uses the ResNet101 network for feature extraction of the detection image, which achieves stronger feature extraction capabilities. Secondly, the new algorithm integrates Online Hard Example Mining (OHEM), Soft non-maximum suppression (Soft-NMS), and Distance Intersection Over Union (DIOU) modules, which improves the positive and negative sample imbalance and the problem of small targets being easily missed during model training. Finally, the Region Proposal Network (RPN) is simplified to achieve a faster detection speed and a lower miss rate. The multi-scale training (MST) strategy is also used to train the improved Faster R-CNN to achieve a balance between detection accuracy and efficiency. Compared to the other detection models, the improved Faster R-CNN demonstrates significant advantages in terms of mAP@0.5, F1-score, and Log average miss rate (LAMR). The model proposed in this paper provides valuable insights and inspiration for many fields, such as smart agriculture, medical diagnosis, and face recognition.

2.
Data Brief ; 55: 110720, 2024 Aug.
Article in English | MEDLINE | ID: mdl-39100779

ABSTRACT

Accurate inspection of rebars in Reinforced Concrete (RC) structures is essential and requires careful counting. Deep learning algorithms utilizing object detection can facilitate this process through Unmanned Aerial Vehicle (UAV) imagery. However, their effectiveness depends on the availability of large, diverse, and well-labelled datasets. This article details the creation of a dataset specifically for counting rebars using deep learning-based object detection methods. The dataset comprises 874 raw images, divided into three subsets: 524 images for training (60 %), 175 for validation (20 %), and 175 for testing (20 %). To enhance the training data, we applied eight augmentation techniques-brightness, contrast, perspective, rotation, scale, shearing, translation, and blurring-exclusively to the training subset. This resulted in nine distinct datasets: one for each augmentation technique and one combining all techniques in augmentation sets. Expert annotators labelled the dataset in VOC XML format. While this research focuses on rebar counting, the raw dataset can be adapted for other tasks, such as estimating rebar diameter or classifying rebar shapes, by providing the necessary annotations.

3.
J Imaging ; 10(7)2024 Jul 05.
Article in English | MEDLINE | ID: mdl-39057733

ABSTRACT

The domain of object detection was revolutionized with the introduction of Convolutional Neural Networks (CNNs) in the field of computer vision. This article aims to explore the architectural intricacies, methodological differences, and performance characteristics of three CNN-based object detection algorithms, namely Faster Region-Based Convolutional Network (R-CNN), You Only Look Once v3 (YOLO), and Single Shot MultiBox Detector (SSD) in the specific domain application of vehicle detection. The findings of this study indicate that the SSD object detection algorithm outperforms the other approaches in terms of both performance and processing speeds. The Faster R-CNN approach detected objects in images with an average speed of 5.1 s, achieving a mean average precision of 0.76 and an average loss of 0.467. YOLO v3 detected objects with an average speed of 1.16 s, achieving a mean average precision of 0.81 with an average loss of 1.183. In contrast, SSD detected objects with an average speed of 0.5 s, exhibiting the highest mean average precision of 0.92 despite having a higher average loss of 2.625. Notably, all three object detectors achieved an accuracy exceeding 99%.

4.
PeerJ Comput Sci ; 10: e2033, 2024.
Article in English | MEDLINE | ID: mdl-38855240

ABSTRACT

This research conducts a comparative analysis of Faster R-CNN and YOLOv8 for real-time detection of fishing vessels and fish in maritime surveillance. The study underscores the significance of this investigation in advancing fisheries monitoring and object detection using deep learning. With a clear focus on comparing the performance of Faster R-CNN and YOLOv8, the research aims to elucidate their effectiveness in real-time detection, emphasizing the relevance of such capabilities in fisheries management. By conducting a thorough literature review, the study establishes the current state-of-the-art in object detection, particularly within the context of fisheries monitoring, while discussing existing methods, challenges, and limitations. The findings of this study not only shed light on the superiority of YOLOv8 in precise detection but also highlight its potential impact on maritime surveillance and the protection of marine resources.

5.
BMC Med Imaging ; 24(1): 152, 2024 Jun 18.
Article in English | MEDLINE | ID: mdl-38890604

ABSTRACT

BACKGROUND: Leishmaniasis is a vector-born neglected parasitic disease belonging to the genus Leishmania. Out of the 30 Leishmania species, 21 species cause human infection that affect the skin and the internal organs. Around, 700,000 to 1,000,000 of the newly infected cases and 26,000 to 65,000 deaths are reported worldwide annually. The disease exhibits three clinical presentations, namely, the cutaneous, muco-cutaneous and visceral Leishmaniasis which affects the skin, mucosal membrane and the internal organs, respectively. The relapsing behavior of the disease limits its diagnosis and treatment efficiency. The common diagnostic approaches follow subjective, error-prone, repetitive processes. Despite, an ever pressing need for an accurate detection of Leishmaniasis, the research conducted so far is scarce. In this regard, the main aim of the current research is to develop an artificial intelligence based detection tool for the Leishmaniasis from the Geimsa-stained microscopic images using deep learning method. METHODS: Stained microscopic images were acquired locally and labeled by experts. The images were augmented using different methods to prevent overfitting and improve the generalizability of the system. Fine-tuned Faster RCNN, SSD, and YOLOV5 models were used for object detection. Mean average precision (MAP), precision, and Recall were calculated to evaluate and compare the performance of the models. RESULTS: The fine-tuned YOLOV5 outperformed the other models such as Faster RCNN and SSD, with the MAP scores, of 73%, 54% and 57%, respectively. CONCLUSION: The currently developed YOLOV5 model can be tested in the clinics to assist the laboratorists in diagnosing Leishmaniasis from the microscopic images. Particularly, in low-resourced healthcare facilities, with fewer qualified medical professionals or hematologists, our AI support system can assist in reducing the diagnosing time, workload, and misdiagnosis. Furthermore, the dataset collected by us will be shared with other researchers who seek to improve upon the detection system of the parasite. The current model detects the parasites even in the presence of the monocyte cells, but sometimes, the accuracy decreases due to the differences in the sizes of the parasite cells alongside the blood cells. The incorporation of cascaded networks in future and the quantification of the parasite load, shall overcome the limitations of the currently developed system.


Subject(s)
Azure Stains , Deep Learning , Microscopy , Humans , Microscopy/methods , Leishmaniasis/diagnostic imaging , Leishmaniasis/parasitology , Leishmania/isolation & purification
6.
Heliyon ; 10(10): e31233, 2024 May 30.
Article in English | MEDLINE | ID: mdl-38803938

ABSTRACT

With the development of Computer Vision, we can effectively and accurately identify trees, fruit or object images. But to build a high-performance image dataset for tree identification problems in Agriculture is a challenge. Realizing that Vietnam is a country with strong agriculture with many tropical fruits grown widely such as Dragon fruit, Mangosteen, Mango, Orange, Lychee, Longan … We chose the Dragon Fruit tree for the data set. of my proposed images, all images will be collected using the close-up method, including tasks such as taking photos of Dragon Fruit trees from many angles and in different conditions (weather, temperature, light, …). In this article, we want to improve the data quality of the collected images so we have applied image processing techniques such as noise filtering (using Gaussian filter), image quality enhancement (image rotation), flip the image, zoom out, zoom in, etc.). From the collected Dragon Fruit tree data set, we will propose to use the Faster R-CNN model for this data set to build a tree and Dragon Fruit identification system.

7.
Sci Rep ; 14(1): 10357, 2024 May 06.
Article in English | MEDLINE | ID: mdl-38710753

ABSTRACT

With constant growth of civilization and modernization of cities all across the world since past few centuries smart traffic management of vehicles is one of the most sorted after problem by research community. Smart traffic management basically involves segmentation of vehicles, estimation of traffic density and tracking of vehicles. The vehicle segmentation from videos helps realization of niche applications such as monitoring of speed and estimation of traffic. When occlusions, background with clutters and traffic with density variations, this problem becomes more intractable in nature. Keeping this motivation in this research work, we investigate Faster R-CNN based deep learning method towards segmentation of vehicles. This problem is addressed in four steps viz minimization with adaptive background model, Faster R-CNN based subnet operation, Faster R-CNN initial refinement and result optimization with extended topological active nets. The computational framework uses adaptive background modeling. It also addresses shadow and illumination issues. Higher segmentation accuracy is achieved through topological active net deformable models. The topological and extended topological active nets help to achieve stated deformations. Mesh deformation is achieved with minimization of energy. The segmentation accuracy is improved with modified version of extended topological active net. The experimental results demonstrate superiority of this framework with respect to other methods.

8.
Plant Methods ; 20(1): 63, 2024 May 06.
Article in English | MEDLINE | ID: mdl-38711143

ABSTRACT

BACKGROUND: The detection of internal defects in seeds via non-destructive imaging techniques is a topic of high interest to optimize the quality of seed lots. In this context, X-ray imaging is especially suited. Recent studies have shown the feasibility of defect detection via deep learning models in 3D tomography images. We demonstrate the possibility of performing such deep learning-based analysis on 2D X-ray radiography for a faster yet robust method via the X-Robustifier pipeline proposed in this article. RESULTS: 2D X-ray images of both defective and defect-free seeds were acquired. A deep learning model based on state-of-the-art object detection neural networks is proposed. Specific data augmentation techniques are introduced to compensate for the low ratio of defects and increase the robustness to variation of the physical parameters of the X-ray imaging systems. The seed defects were accurately detected (F1-score >90%), surpassing human performance in computation time and error rates. The robustness of these models against the principal distortions commonly found in actual agro-industrial conditions is demonstrated, in particular, the robustness to physical noise, dimensionality reduction and the presence of seed coating. CONCLUSION: This work provides a full pipeline to automatically detect common defects in seeds via 2D X-ray imaging. The method is illustrated on sugar beet and faba bean and could be efficiently extended to other species via the proposed generic X-ray data processing approach (X-Robustifier). Beyond a simple proof of feasibility, this constitutes important results toward the effective use in the routine of deep learning-based automatic detection of seed defects.

9.
J Imaging ; 10(5)2024 Apr 26.
Article in English | MEDLINE | ID: mdl-38786556

ABSTRACT

The world's most significant yield by production quantity is sugarcane. It is the primary source for sugar, ethanol, chipboards, paper, barrages, and confectionery. Many people are affiliated with sugarcane production and their products around the globe. The sugarcane industries make an agreement with farmers before the tillering phase of plants. Industries are keen on knowing the sugarcane field's pre-harvest estimation for planning their production and purchases. The proposed research contribution is twofold: by publishing our newly developed dataset, we also present a methodology to estimate the number of sugarcane plants in the tillering phase. The dataset has been obtained from sugarcane fields in the fall season. In this work, a modified architecture of Faster R-CNN with feature extraction using VGG-16 with Inception-v3 modules and sigmoid threshold function has been proposed for the detection and classification of sugarcane plants. Significantly promising results with 82.10% accuracy have been obtained with the proposed architecture, showing the viability of the developed methodology.

10.
Bioengineering (Basel) ; 11(5)2024 May 19.
Article in English | MEDLINE | ID: mdl-38790377

ABSTRACT

A deep convolution network that expands on the architecture of the faster R-CNN network is proposed. The expansion includes adapting unsupervised classification with multiple backbone networks to improve the Region Proposal Network in order to improve accuracy and sensitivity in detecting minute changes in images. The efficiency of the proposed architecture is investigated by applying it to the detection of cancerous lung tumors in CT (computed tomography) images. This investigation used a total of 888 images from the LUNA16 dataset, which contains CT images of both cancerous and non-cancerous tumors of various sizes. These images are divided into 80% and 20%, which are used for training and testing, respectively. The result of the investigation through the experiment is that the proposed deep-learning architecture could achieve an accuracy rate of 95.32%, a precision rate of 94.63%, a specificity of 94.84%, and a high sensitivity of 96.23% using the LUNA16 images. The result shows an improvement compared to a reported accuracy of 93.6% from a previous study using the same dataset.

11.
Sensors (Basel) ; 24(7)2024 Apr 03.
Article in English | MEDLINE | ID: mdl-38610494

ABSTRACT

Accurately and effectively detecting the growth position and contour size of apple fruits is crucial for achieving intelligent picking and yield predictions. Thus, an effective fruit edge detection algorithm is necessary. In this study, a fusion edge detection model (RED) based on a convolutional neural network and rough sets was proposed. The Faster-RCNN was used to segment multiple apple images into a single apple image for edge detection, greatly reducing the surrounding noise of the target. Moreover, the K-means clustering algorithm was used to segment the target of a single apple image for further noise reduction. Considering the influence of illumination, complex backgrounds and dense occlusions, rough set was applied to obtain the edge image of the target for the upper and lower approximation images, and the results were compared with those of relevant algorithms in this field. The experimental results showed that the RED model in this paper had high accuracy and robustness, and its detection accuracy and stability were significantly improved compared to those of traditional operators, especially under the influence of illumination and complex backgrounds. The RED model is expected to provide a promising basis for intelligent fruit picking and yield prediction.

12.
Skin Res Technol ; 30(4): e13698, 2024 Apr.
Article in English | MEDLINE | ID: mdl-38634154

ABSTRACT

BACKGROUND: Dermoscopy is a common method of scalp psoriasis diagnosis, and several artificial intelligence techniques have been used to assist dermoscopy in the diagnosis of nail fungus disease, the most commonly used being the convolutional neural network algorithm; however, convolutional neural networks are only the most basic algorithm, and the use of object detection algorithms to assist dermoscopy in the diagnosis of scalp psoriasis has not been reported. OBJECTIVES: Establishment of a dermoscopic modality diagnostic framework for scalp psoriasis based on object detection technology and image enhancement to improve diagnostic efficiency and accuracy. METHODS: We analyzed the dermoscopic patterns of scalp psoriasis diagnosed at 72nd Group army hospital of PLA from January 1, 2020 to December 31, 2021, and selected scalp seborrheic dermatitis as a control group. Based on dermoscopic images and major dermoscopic patterns of scalp psoriasis and scalp seborrheic dermatitis, we investigated a multi-network fusion object detection framework based on the object detection technique Faster R-CNN and the image enhancement technique contrast limited adaptive histogram equalization (CLAHE), for assisting in the diagnosis of scalp psoriasis and scalp seborrheic dermatitis, as well as to differentiate the major dermoscopic patterns of the two diseases. The diagnostic performance of the multi-network fusion object detection framework was compared with that between dermatologists. RESULTS: A total of 1876 dermoscopic images were collected, including 1218 for scalp psoriasis versus 658 for scalp seborrheic dermatitis. Based on these images, training and testing are performed using a multi-network fusion object detection framework. The results showed that the test accuracy, specificity, sensitivity, and Youden index for the diagnosis of scalp psoriasis was: 91.0%, 89.5%, 91.0%, and 0.805, and for the main dermoscopic patterns of scalp psoriasis and scalp seborrheic dermatitis, the diagnostic results were: 89.9%, 97.7%, 89.9%, and 0.876. Comparing the diagnostic results with those of five dermatologists, the fusion framework performs better than the dermatologists' diagnoses. CONCLUSIONS: Studies have shown some differences in dermoscopic patterns between scalp psoriasis and scalp seborrheic dermatitis. The proposed multi-network fusion object detection framework has higher diagnostic performance for scalp psoriasis than for dermatologists.


Subject(s)
Dermatitis, Seborrheic , Psoriasis , Skin Neoplasms , Humans , Scalp , Artificial Intelligence , Neural Networks, Computer , Dermoscopy/methods , Skin Neoplasms/diagnosis
13.
Sensors (Basel) ; 24(8)2024 Apr 10.
Article in English | MEDLINE | ID: mdl-38676049

ABSTRACT

Long-term, automated fish detection provides invaluable data for deep-sea aquaculture, which is crucial for safe and efficient seawater aquafarming. In this paper, we used an infrared camera installed on a deep-sea truss-structure net cage to collect fish images, which were subsequently labeled to establish a fish dataset. Comparison experiments with our dataset based on Faster R-CNN as the basic objection detection framework were conducted to explore how different backbone networks and network improvement modules influenced fish detection performances. Furthermore, we also experimented with the effects of different learning rates, feature extraction layers, and data augmentation strategies. Our results showed that Faster R-CNN with the EfficientNetB0 backbone and FPN module was the most competitive fish detection network for our dataset, since it took a significantly shorter detection time while maintaining a high AP50 value of 0.85, compared to the best AP50 value of 0.86 being achieved by the combination of VGG16 with all improvement modules plus data augmentation. Overall, this work has verified the effectiveness of deep learning-based object detection methods and provided insights into subsequent network improvements.


Subject(s)
Aquaculture , Deep Learning , Fishes , Animals , Aquaculture/methods , Infrared Rays , Image Processing, Computer-Assisted/methods , Neural Networks, Computer
14.
Sensors (Basel) ; 24(8)2024 Apr 21.
Article in English | MEDLINE | ID: mdl-38676267

ABSTRACT

The rapid increase in the number of vehicles has led to increasing traffic congestion, traffic accidents, and motor vehicle crime rates. The management of various parking lots has also become increasingly challenging. Vehicle-type recognition technology can reduce the workload of humans in vehicle management operations. Therefore, the application of image technology for vehicle-type recognition is of great significance for integrated traffic management. In this paper, an improved faster region with convolutional neural network features (Faster R-CNN) model was proposed for vehicle-type recognition. Firstly, the output features of different convolution layers were combined to improve the recognition accuracy. Then, the average precision (AP) of the recognition model was improved through the contextual features of the original image and the object bounding box optimization strategy. Finally, the comparison experiment used the vehicle image dataset of three vehicle types, including cars, sports utility vehicles (SUVs), and vans. The experimental results show that the improved recognition model can effectively identify vehicle types in the images. The AP of the three vehicle types is 83.2%, 79.2%, and 78.4%, respectively, and the mean average precision (mAP) is 1.7% higher than that of the traditional Faster R-CNN model.

15.
Nan Fang Yi Ke Da Xue Xue Bao ; 44(2): 344-353, 2024 Feb 20.
Article in Chinese | MEDLINE | ID: mdl-38501420

ABSTRACT

OBJECTIVE: To propose a method for mitigate the impact of anomaly points (such as dust, bubbles, scratches on the chip surface, and minor indentations) in images on the results of digital droplet PCR (ddPCR) detection to achieve high-throughput, stable, and accurate detection. METHODS: We propose a Filter Faster R-CNN ddPCR detection model, which employs Faster R-CNN to generate droplet prediction boxes followed by removing the anomalies within the positive droplet prediction boxes using an outlier filtering module (Filter). Using a plasmid carrying a norovirus fragment as the template, we established a ddPCR dataset for model training (2462 instances, 78.56%) and testing (672 instances, 21.44%). Ablation experiments were performed to test the effectiveness of 3 filtering branches of the Filter for anomaly removal on the validation dataset. Comparative experiments with other ddPCR droplet detection models and absolute quantification experiments of ddPCR were conducted to assess the performance of the Filter Faster R-CNN model. RESULTS: In low-dust and dusty environments, the Filter Faster R-CNN model achieved detection accuracies of 98.23% and 88.35% for positive droplets, respectively, with composite F1 scores reaching 99.15% and 99.14%, obviously superior to the other models. The introduction of the filtering module significantly enhanced the positive accuracy of the model in dusty environments. In the absolute quantification experiments, a regression line was plotted using the results from commercial flow cytometry equipment as the standard concentration. The results show a regression line slope of 1.0005, an intercept of -0.025, and a determination coefficient of 0.9997, indicating high consistency between the two results. CONCLUSION: The ddPCR detection technique using the Filter Faster R-CNN model provides a robust detection method for ddPCR under various environmental conditions.


Subject(s)
Dust , Polymerase Chain Reaction/methods
16.
Cancers (Basel) ; 16(3)2024 Feb 02.
Article in English | MEDLINE | ID: mdl-38339394

ABSTRACT

Performing a mitosis count (MC) is the diagnostic task of histologically grading canine Soft Tissue Sarcoma (cSTS). However, mitosis count is subject to inter- and intra-observer variability. Deep learning models can offer a standardisation in the process of MC used to histologically grade canine Soft Tissue Sarcomas. Subsequently, the focus of this study was mitosis detection in canine Perivascular Wall Tumours (cPWTs). Generating mitosis annotations is a long and arduous process open to inter-observer variability. Therefore, by keeping pathologists in the loop, a two-step annotation process was performed where a pre-trained Faster R-CNN model was trained on initial annotations provided by veterinary pathologists. The pathologists reviewed the output false positive mitosis candidates and determined whether these were overlooked candidates, thus updating the dataset. Faster R-CNN was then trained on this updated dataset. An optimal decision threshold was applied to maximise the F1-score predetermined using the validation set and produced our best F1-score of 0.75, which is competitive with the state of the art in the canine mitosis domain.

17.
Biomed Phys Eng Express ; 10(2)2024 Feb 15.
Article in English | MEDLINE | ID: mdl-38357907

ABSTRACT

The assessment of mitotic activity is an integral part of the comprehensive evaluation of breast cancer pathology. Understanding the level of tumor dissemination is essential for assessing the severity of the malignancy and guiding appropriate treatment strategies. A pathologist must manually perform the intricate and time-consuming task of counting mitoses by examining biopsy slices stained with Hematoxylin and Eosin (H&E) under a microscope. Mitotic cells can be challenging to distinguish in H&E-stained sections due to limited available datasets and similarities among mitotic and non-mitotic cells. Computer-assisted mitosis detection approaches have simplified the whole procedure by selecting, detecting, and labeling mitotic cells. Traditional detection strategies rely on image processing techniques that apply custom criteria to distinguish between different aspects of an image. Additionally, the automatic feature extraction from histopathology images that exhibit mitosis using neural networks.Additionally, the possibility of automatically extracting features from histopathological images using deep neural networks was investigated. This study examines mitosis detection as an object detection problem using multiple neural networks. From a medical standpoint, mitosis at the tissue level was also investigated utilising pre-trained Faster R-CNN and raw image data. Experiments were done on the MITOS-ATYPIA- 14 dataset and TUPAC16 dataset, and the results were compared to those of other methods described in the literature.


Subject(s)
Breast Neoplasms , Mitosis , Humans , Female , Neural Networks, Computer , Breast Neoplasms/diagnostic imaging , Image Processing, Computer-Assisted/methods
18.
Heliyon ; 10(3): e25367, 2024 Feb 15.
Article in English | MEDLINE | ID: mdl-38327447

ABSTRACT

Water quality can be negatively affected by the presence of some toxic phytoplankton species, whose toxins are difficult to remove by conventional purification systems. This creates the need for periodic analyses, which are nowadays manually performed by experts. These labor-intensive processes are affected by subjectivity and expertise, causing unreliability. Some automatic systems have been proposed to address these limitations. However, most of them are based on classical image processing pipelines with not easily scalable designs. In this context, deep learning techniques are more adequate for the detection and recognition of phytoplankton specimens in multi-specimen microscopy images, as they integrate both tasks in a single end-to-end trainable module that is able to automatize the adaption to such a complex domain. In this work, we explore the use of two different object detectors: Faster R-CNN and RetinaNet, from the one-stage and two-stage paradigms respectively. We use a dataset composed of multi-specimen microscopy images captured using a systematic protocol. This allows the use of widely available optical microscopes, also avoiding manual adjustments on a per-specimen basis, which would require expert knowledge. We have made our dataset publicly available to improve the reproducibility and to foment the development of new alternatives in the field. The selected Faster R-CNN methodology reaches maximum recall levels of 95.35%, 84.69%, and 79.81%, and precisions of 94.68%, 89.30% and 82.61%, for W. naegeliana, A. spiroides, and D. sociale, respectively. The system is able to adapt to the dataset problems and improves the results overall with respect to the reference state-of-the-art work. In addition, the proposed system improves the automation and abstraction from the domain and simplifies the workflow and adjustment.

19.
Article in English | MEDLINE | ID: mdl-38278999

ABSTRACT

Smart, secure, and environmentally friendly smart cities are all the rage in urban planning. Several technologies, including the Internet of Things (IoT) and edge computing, are used to develop smart cities. Early and accurate fire detection in a Smart city is always desirable and motivates the research community to create a more efficient model. Deep learning models are widely used for fire detection in existing research, but they encounter several issues in typical climate environments, such as foggy and normal. The proposed model lends itself to IoT applications for authentic fire surveillance because of its minimal configuration load. A hybrid Local Binary Pattern Convolutional Neural Network (LBP-CNN) and YOLO-V5 model-based fire detection model for smart cities in the foggy scenario is presented in this research. Additionally, we recommend a two-part technique for extracting features to be applied to YOLO throughout this article. Using a transfer learning technique, the first portion of the proposed approach for extracting features retrieves standard features. The section part is for retrieval of additional valuable information related to the current activity using the LBP (Local Binary Pattern) protective layer and classifications layers. This research utilizes an online Kaggle fire and smoke dataset with 13950 normal and foggy images. The proposed hybrid model is premised on a two-cascaded YOLO model. In the initial cascade, smoke and fire are detected in the normal surrounding region, and the second cascade fire is detected with density in a foggy environment. In experimental analysis, the proposed model achieved a fire and smoke detection precision rate of 96.25% for a normal setting, 93.2% for a foggy environment, and a combined detection average precision rate of 94.59%. The proposed hybrid system outperformed existing models in terms of better precision and density detection for fire and smoke.

20.
IEEE J Transl Eng Health Med ; 12: 119-128, 2024.
Article in English | MEDLINE | ID: mdl-38088993

ABSTRACT

The objective of this study was to develop an interpretable system that could detect specific lung features in neonates. A challenging aspect of this work was that normal lungs showed the same visual features (as that of Pneumothorax (PTX)). M-mode is typically necessary to differentiate between the two cases, but its generation in clinics is time-consuming and requires expertise for interpretation, which remains limited. Therefore, our system automates M-mode generation by extracting Regions of Interest (ROIs) without human in the loop. Object detection models such as faster Region Based Convolutional Neural Network (fRCNN) and RetinaNet models were employed to detect seven common Lung Ultrasound (LUS) features. fRCNN predictions were then stored and further used to generate M-modes. Beyond static feature extraction, we used a Hough transform based statistical method to detect "lung sliding" in these M-modes. Results showed that fRCNN achieved a greater mean Average Precision (mAP) of 86.57% (Intersection-over-Union (IoU) = 0.2) than RetinaNet, which only displayed a mAP of 61.15%. The calculated accuracy for the generated RoIs was 97.59% for Normal videos and 96.37% for PTX videos. Using this system, we successfully classified 5 PTX and 6 Normal video cases with 100% accuracy. Automating the process of detecting seven prominent LUS features addresses the time-consuming manual evaluation of Lung ultrasound in a fast paced environment. Clinical impact: Our research work provides a significant clinical impact as it provides a more accurate and efficient method for diagnosing lung diseases in neonates.


Subject(s)
Pneumonia , Pneumothorax , Humans , Infant, Newborn , Lung/diagnostic imaging , Neural Networks, Computer , Pneumothorax/diagnostic imaging , Thorax
SELECTION OF CITATIONS
SEARCH DETAIL