Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 1.587
Filtrar
1.
J Exp Bot ; 2024 Oct 04.
Artigo em Inglês | MEDLINE | ID: mdl-39363775

RESUMO

Artificial intelligence and machine learning (AI/ML) can be used to automatically analyze large image datasets. One valuable application of this approach is estimation of plant trait data contained within images. Here we review 39 papers that describe the development and/or application of such models for estimation of stomatal traits from epidermal micrographs. In doing so, we hope to provide plant biologists with a foundational understanding of AI/ML and summarize the current capabilities and limitations of published tools. While most models show human-level performance for stomatal density (SD) quantification at superhuman speed, they are often likely to be limited in how broadly they can be applied across phenotypic diversity associated with genetic, environmental or developmental variation. Other models can make predictions across greater phenotypic diversity and/or additional stomatal/epidermal traits, but require significantly greater time investment to generate ground-truth data. We discuss the challenges and opportunities presented by AI/ML-enabled computer vision analysis, and make recommendations for future work to advance accelerated stomatal phenotyping.

2.
Environ Monit Assess ; 196(11): 1044, 2024 Oct 11.
Artigo em Inglês | MEDLINE | ID: mdl-39392511

RESUMO

Tea leaf blight (TLB) is a common disease of tea plants and is widely distributed in tea gardens. Although the use of unmanned aerial vehicle (UAV) remote sensing can help to achieve a wider scale for TLB detection, the blurring of UAV images, overlapping of tea leaves, and small size of TLB spots pose significant challenges to the task of detection. This study proposes a method of detecting TLB in UAV remote sensing images by integrating super-resolution (SR) and detection networks. We use an SR network called SERB-Swin2sr to reconstruct the detailed features of UAV images and solve the problem of detail loss caused by the blurring in UAV images. In SERB-Swin2sr, a squeeze-and-excitation ResNet block (SERB) is introduced to enhance the models' ability to extract the target details in the images, and the convolution stem replaces the convolution block in order to increase the convergence rate and stability of the network. A detection network called SDDA-YOLO is applied to achieve precise detection of TLB in UAV remote sensing images. In SDDA-YOLO, a shuffle dual-dimensional attention (SDDA) module is introduced to enhance the feature fusion capability of the network, and an Xsmall-scale detection layer is used to enhance the detection ability of small lesions. Experimental results show that the proposed method is superior to current detection methods. Compared with a baseline YOLOv8 model, the precision, mAP@0.5, and mAP@0.5:0.95 of the proposed method are improved by 4.2%, 1.6%, and 1.8%, and the size of our model is only 4.6 MB.


Assuntos
Camellia sinensis , Doenças das Plantas , Folhas de Planta , Tecnologia de Sensoriamento Remoto , Dispositivos Aéreos não Tripulados , Folhas de Planta/microbiologia , Doenças das Plantas/microbiologia , Camellia sinensis/microbiologia , Monitoramento Ambiental/métodos
3.
Artigo em Inglês | MEDLINE | ID: mdl-39379641

RESUMO

PURPOSE: Dysphagia is the inability or difficulty to swallow normally. Standard procedures for diagnosing the exact disease are, among others, X-ray videofluoroscopy, manometry and impedance examinations, usually performed consecutively. In order to gain more insights, ongoing research is aiming to collect these different modalities at the same time, with the goal to present them in a joint visualization. One idea to create a combined view is the projection of the manometry and impedance values onto the right location in the X-ray images. This requires to identify the exact sensor locations in the images. METHODS: This work gives an overview of the challenges associated with the sensor detection task and proposes a robust approach to detect the sensors in X-ray image sequences, ultimately allowing to project the manometry and impedance values onto the right location in the images. RESULTS: The developed sensor detection approach is evaluated on a total of 14 sequences from different patients, achieving a F1-score of 86.36%. To demonstrate the robustness of the approach, another study is performed by adding different levels of noise to the images, with the performance of our sensor detection method only slightly decreasing in these scenarios. This robust sensor detection provides the basis to accurately project manometry and impedance values onto the images, allowing to create a multimodal visualization of the swallow process. The resulting visualizations are evaluated qualitatively by domain experts, indicating a great benefit of this proposed fused visualization approach. CONCLUSION: Using our preprocessing and sensor detection method, we show that the sensor detection task can be successfully approached with high accuracy. This allows to create a novel, multimodal visualization of esophageal motility, helping to provide more insights into swallow disorders of patients.

4.
J Environ Manage ; 370: 122742, 2024 Oct 08.
Artigo em Inglês | MEDLINE | ID: mdl-39383749

RESUMO

Sorting out plastic waste (PW) from municipal solid waste (MSW) by material type is crucial for reutilization and pollution reduction. However, current automatic separation methods are costly and inefficient, necessitating an advanced sorting process to ensure high feedstock purity. This study introduces a Swin Transformer-based model for effectively detecting PW in real-world MSW streams, leveraging both morphological and material properties. And, a dataset comprising 3560 optical images and infrared spectra data was created to support this task. This vision-based system can localize and classify PW into five categories: polypropylene (PP), polyethylene (PE), polyethylene terephthalate (PET), polyvinyl chloride (PVC), and polystyrene (PS). Performance evaluations reveal an accuracy rate of 99.75% and a mean Average Precision (mAP50) exceeding 91%. Compared to popular convolutional neural network (CNN)-based models, this well-trained Swin Transformer-based model offers enhanced convenience and performance in five-category PW detection task, maintaining a mAP50 over 80% in the real-life deployment. The model's effectiveness is further supported by visualization of detection results on MSW streams and principal component analysis of classification scores. These results demonstrate the system's significant effectiveness in both lab-scale and real-life conditions, aligning with global regulations and strategies that promote innovative technologies for plastic recycling, thereby contributing to the development of a sustainable circular economy.

5.
Sci Prog ; 107(4): 368504241280765, 2024.
Artigo em Inglês | MEDLINE | ID: mdl-39360473

RESUMO

As a pivotal task within computer vision, object detection finds application across a diverse spectrum of industrial scenarios. The advent of deep learning technologies has significantly elevated the accuracy of object detectors designed for general-purpose applications. Nevertheless, in contrast to conventional terrestrial environments, remote sensing object detection scenarios pose formidable challenges, including intricate and diverse backgrounds, fluctuating object scales, and pronounced interference from background noise, rendering remote sensing object detection an enduringly demanding task. In addition, despite the superior detection performance of deep learning-based object detection networks compared to traditional counterparts, their substantial parameter and computational demands curtail their feasibility for deployment on mobile devices equipped with low-power processors. In response to the aforementioned challenges, this paper introduces an enhanced lightweight remote sensing object detection network, denoted as YOLO-Faster, built upon the foundation of YOLOv5. Firstly, the lightweight design and inference speed of the object detection network is augmented by incorporating the lightweight network as the foundational network within YOLOv5, satisfying the demand for real-time detection on mobile devices. Moreover, to tackle the issue of detecting objects of different scales in large and complex backgrounds, an adaptive multiscale feature fusion network is introduced, which dynamically adjusts the large receptive field to capture dependencies among objects of different scales, enabling better modeling of object detection scenarios in remote sensing scenes. At last, the robustness of the object detection network under background noise is enhanced through incorporating a decoupled detection head that separates the classification and regression processes of the detection network. The results obtained from the public remote sensing object detection dataset DOTA show that the proposed method has a mean average precision of 71.4% and a detection speed of 38 frames per second.

6.
Data Brief ; 57: 110941, 2024 Dec.
Artigo em Inglês | MEDLINE | ID: mdl-39351130

RESUMO

This CIDACC dataset was created to determine the cell population of Chlorella vulgaris microalga during cultivation. Chlorella vulgaris has diverse applications, including use as food supplement, biofuel production, and pollutant removal. High resolution images were collected using a microscope and annotated, focusing on computer vision and machine learning models creation for automatic Chlorella cell detection, counting, size and geometry estimation. The dataset comprises 628 images, organized into hierarchical folders for easy access. Detailed segmentation masks and bounding boxes were generated using external tools enhancing the dataset's utility. The dataset's efficacy was demonstrated through preliminary experiments using deep learning architecture such as object detection and localization algorithms, as well as image segmentation algorithms, achieving high precision and accuracy. This dataset is a valuable tool for advancing computer vision applications in microalgae research and other related fields. The dataset is particularly challenging due to its dynamic nature and the complex correlations it presents across various application domains, including cell analysis in medical research. Its intricacies not only push the boundaries of current computer vision algorithms but also offer significant potential for advancements in diverse fields such as biomedical imaging, environmental monitoring, and biotechnological innovations.

7.
Mar Pollut Bull ; 209(Pt A): 117030, 2024 Oct 10.
Artigo em Inglês | MEDLINE | ID: mdl-39393229

RESUMO

Underwater litter is widely spread across aquatic environments such as lakes, rivers, and oceans, significantly impacting natural ecosystems. Current automated monitoring technologies for detecting this litter face limitations in survey efficiency, cost, and environmental conditions, highlighting the need for efficient, consumer-grade technologies for automatic detection. This research introduces the Aerial-Aquatic Speedy Scanner (AASS) combined with Super-Resolution Reconstruction (SRR) and an enhanced YOLOv8 detection network. The AASS system boosts data acquisition efficiency over traditional methods, capturing high-resolution images that accurately identify and categorize underwater waste. The SRR technique enhances image quality by mitigating common issues like motion blur and low resolution, thereby improving the YOLOv8 model's detection capabilities. Specifically, the RCAN model achieved the highest mean average precision (mAP) of 78.6 % for object detection accuracy on reconstructed underwater litter among the tested SR models. With a magnification factor of 4, the SR test set shows an improved mAP compared to the Bicubic test set. These results demonstrate the effectiveness of the proposed method in detecting underwater litter.

8.
Acta Crystallogr D Struct Biol ; 80(Pt 10): 744-764, 2024 Oct 01.
Artigo em Inglês | MEDLINE | ID: mdl-39361357

RESUMO

A group of three deep-learning tools, referred to collectively as CHiMP (Crystal Hits in My Plate), were created for analysis of micrographs of protein crystallization experiments at the Diamond Light Source (DLS) synchrotron, UK. The first tool, a classification network, assigns images into categories relating to experimental outcomes. The other two tools are networks that perform both object detection and instance segmentation, resulting in masks of individual crystals in the first case and masks of crystallization droplets in addition to crystals in the second case, allowing the positions and sizes of these entities to be recorded. The creation of these tools used transfer learning, where weights from a pre-trained deep-learning network were used as a starting point and repurposed by further training on a relatively small set of data. Two of the tools are now integrated at the VMXi macromolecular crystallography beamline at DLS, where they have the potential to absolve the need for any user input, both for monitoring crystallization experiments and for triggering in situ data collections. The third is being integrated into the XChem fragment-based drug-discovery screening platform, also at DLS, to allow the automatic targeting of acoustic compound dispensing into crystallization droplets.


Assuntos
Cristalização , Aprendizado Profundo , Cristalização/métodos , Cristalografia por Raios X/métodos , Proteínas/química , Processamento de Imagem Assistida por Computador/métodos , Síncrotrons , Automação , Software
9.
Rev Cardiovasc Med ; 25(9): 335, 2024 Sep.
Artigo em Inglês | MEDLINE | ID: mdl-39355611

RESUMO

Background: Congenital heart diseases (CHDs), particularly atrial and ventricular septal defects, pose significant health risks and common challenges in detection via echocardiography. Doctors often employ the cardiac structural information during the diagnostic process. However, prior CHD research has not determined the influence of including cardiac structural information during the labeling process and the application of data augmentation techniques. Methods: This study utilizes advanced artificial intelligence (AI)-driven object detection frameworks, specifically You Look Only Once (YOLO)v5, YOLOv7, and YOLOv9, to assess the impact of including cardiac structural information and data augmentation techniques on the identification of septal defects in echocardiographic images. Results: The experimental results reveal that different labeling strategies substantially affect the performance of the detection models. Notably, adjustments in bounding box dimensions and the inclusion of cardiac structural details in the annotations are key factors influencing the accuracy of the model. The application of deep learning techniques in echocardiography enhances the precision of detecting septal heart defects. Conclusions: This study confirms that careful annotation of imaging data is crucial for optimizing the performance of object detection algorithms in medical imaging. These findings suggest potential pathways for refining AI applications in diagnostic cardiology studies.

10.
J Robot Surg ; 18(1): 339, 2024 Sep 12.
Artigo em Inglês | MEDLINE | ID: mdl-39261441

RESUMO

Worldwide, healthcare systems are struggling to tackle a nursing shortage. As per the data released by the American Hospital Association, the healthcare workforce could face a loss of around 5 lakh nurses by the end of the year. Consequently, it could result in those deficiencies, which will be 1.1 million instead of 0.6 million. The current nursing scenario in India as per the Indian Nursing Council (INC), which is a board under the ministry and is responsible for legally confirming and maintaining universal standardized training for nursing, is that the 1.96 nurses out of every 1000 Indians that are there are much behind the World Health Organization's (WHO) recommended figure of 3 nurses per 1000. To mitigate the nurse shortage, a collaborative robotic system was designed that can assist with surgical procedures with a collaborative robot acting as a scrub nurse for cataract surgery (CRASCS) represented in Fig. 1. Accordingly, the model has been built to empower a customized 3d printed 5-Degree of freedom robotic arm by tracking the phase of surgery in real-time and automatically supplying the clinician with the ideal equipment that is needed for the particular phase of surgery. The system is supported with one more model which can identify where the surgical equipment is located within the arm range. The system is also supported with voice commands which help in picking up the right surgical equipment in the middle of any phase of the surgery. In this way, the system could be able to potentially handle the shortage of surgical nurses around the world and benefit humanity.


Assuntos
Extração de Catarata , Procedimentos Cirúrgicos Robóticos , Humanos , Extração de Catarata/métodos , Procedimentos Cirúrgicos Robóticos/métodos , Índia , Impressão Tridimensional
11.
Front Plant Sci ; 15: 1451018, 2024.
Artigo em Inglês | MEDLINE | ID: mdl-39239201

RESUMO

Introduction: Efficiently and precisely identifying tomatoes amidst intricate surroundings is essential for advancing the automation of tomato harvesting. Current object detection algorithms are slow and have low recognition accuracy for occluded and small tomatoes. Methods: To enhance the detection of tomatoes in complex environments, a lightweight greenhouse tomato object detection model named S-YOLO is proposed, based on YOLOv8s with several key improvements: (1) A lightweight GSConv_SlimNeck structure tailored for YOLOv8s was innovatively constructed, significantly reducing model parameters to optimize the model neck for lightweight model acquisition. (2) An improved version of the α-SimSPPF structure was designed, effectively enhancing the detection accuracy of tomatoes. (3) An enhanced version of the ß-SIoU algorithm was proposed to optimize the training process and improve the accuracy of overlapping tomato recognition. (4) The SE attention module is integrated to enable the model to capture more representative greenhouse tomato features, thereby enhancing detection accuracy. Results: Experimental results demonstrate that the enhanced S-YOLO model significantly improves detection accuracy, achieves lightweight model design, and exhibits fast detection speeds. Experimental results demonstrate that the S-YOLO model significantly enhances detection accuracy, achieving 96.60% accuracy, 92.46% average precision (mAP), and a detection speed of 74.05 FPS, which are improvements of 5.25%, 2.1%, and 3.49 FPS respectively over the original model. With model parameters at only 9.11M, the S-YOLO outperforms models such as CenterNet, YOLOv3, YOLOv4, YOLOv5m, YOLOv7, and YOLOv8s, effectively addressing the low recognition accuracy of occluded and small tomatoes. Discussion: The lightweight characteristics of the S-YOLO model make it suitable for the visual system of tomato-picking robots, providing technical support for robot target recognition and harvesting operations in facility environments based on mobile edge computing.

12.
Water Res ; 266: 122405, 2024 Sep 11.
Artigo em Inglês | MEDLINE | ID: mdl-39265217

RESUMO

Researchers and practitioners have extensively utilized supervised Deep Learning methods to quantify floating litter in rivers and canals. These methods require the availability of large amount of labeled data for training. The labeling work is expensive and laborious, resulting in small open datasets available in the field compared to the comprehensive datasets for computer vision, e.g., ImageNet. Fine-tuning models pre-trained on these larger datasets helps improve litter detection performances and reduces data requirements. Yet, the effectiveness of using features learned from generic datasets is limited in large-scale monitoring, where automated detection must adapt across different locations, environmental conditions, and sensor settings. To address this issue, we propose a two-stage semi-supervised learning method to detect floating litter based on the Swapping Assignments between multiple Views of the same image (SwAV). SwAV is a self-supervised learning approach that learns the underlying feature representation from unlabeled data. In the first stage, we used SwAV to pre-train a ResNet50 backbone architecture on about 100k unlabeled images. In the second stage, we added new layers to the pre-trained ResNet50 to create a Faster R-CNN architecture, and fine-tuned it with a limited number of labeled images (≈1.8k images with 2.6k annotated litter items). We developed and validated our semi-supervised floating litter detection methodology for images collected in canals and waterways of Delft (the Netherlands) and Jakarta (Indonesia). We tested for out-of-domain generalization performances in a zero-shot fashion using additional data from Ho Chi Minh City (Vietnam), Amsterdam and Groningen (the Netherlands). We benchmarked our results against the same Faster R-CNN architecture trained via supervised learning alone by fine-tuning ImageNet pre-trained weights. The findings indicate that the semi-supervised learning method matches or surpasses the supervised learning benchmark when tested on new images from the same training locations. We measured better performances when little data (≈200 images with about 300 annotated litter items) is available for fine-tuning and with respect to reducing false positive predictions. More importantly, the proposed approach demonstrates clear superiority for generalization on the unseen locations, with improvements in average precision of up to 12.7%. We attribute this superior performance to the more effective high-level feature extraction from SwAV pre-training from relevant unlabeled images. Our findings highlight a promising direction to leverage semi-supervised learning for developing foundational models, which have revolutionized artificial intelligence applications in most fields. By scaling our proposed approach with more data and compute, we can make significant strides in monitoring to address the global challenge of litter pollution in water bodies.

13.
Comput Biol Med ; 182: 109153, 2024 Sep 16.
Artigo em Inglês | MEDLINE | ID: mdl-39288557

RESUMO

OBJECTIVES: Cracked tooth syndrome (CTS) is one of the major causes of tooth loss, presents the problem of early microcrack symptoms that are difficult to distinguish. This paper aims to investigate the practicality and feasibility of an improved object detection algorithm for automatically detecting cracks in dental optical images. METHODS: A total of 286 teeth were obtained from Sun Yat-sen University and Guangdong University of Technology, and simulated cracks were generated using thermal expansion and contraction. Over 3000 images of cracked teeth were collected, including 360 real clinical images. To make the model more lightweight and better suited for deployment on embedded devices, this paper improves the YOLOv8 model for detecting tooth cracks through model pruning and backbone replacement. Additionally, the impact of image enhancement modules and coordinate attention modules on optimizing our model was analyzed. RESULTS: Through experimental validation, we conclude that that model pruning reduction maintains performance better than replacing a lightweight backbone network on a tooth crack detection task. This approach achieved a reduction in parameters and GFLOPs by 16.8 % and 24.3 %, respectively, with minimal impact on performance. These results affirm the effectiveness of the proposed method in identifying and labeling tooth fractures. In addition, this paper demonstrated that the impact of image enhancement modules and coordinate attention mechanisms on YOLOv8's performance in the task of tooth crack detection was minimal. CONCLUSIONS: An improved object detection algorithm has been proposed to reduce model parameters. This lightweight model is easier to deploy and holds potential for assisting dentists in identifying cracks on tooth surfaces.

14.
PeerJ Comput Sci ; 10: e2224, 2024.
Artigo em Inglês | MEDLINE | ID: mdl-39314678

RESUMO

Surface defect inspection methods have proven effective in addressing casting quality control tasks. However, traditional inspection methods often struggle to achieve high-precision detection of surface defects in castings with similar characteristics and minor scales. The study introduces DES-YOLO, a novel real-time method for detecting castings' surface defects. In the DES-YOLO model, we incorporate the DSC-Darknet backbone network and global attention mechanism (GAM) module to enhance the identification of defect target features. These additions are essential for overcoming the challenge posed by the high similarity among defect characteristics, such as shrinkage holes and slag holes, which can result in decreased detection accuracy. An enhanced pyramid pooling module is also introduced to improve feature representation for small defective parts through multi-layer pooling. We integrate Slim-Neck and SIoU bounding box regression loss functions for real-time detection in actual production scenarios. These functions reduce memory overhead and enable real-time detection of surface defects in castings. Experimental findings demonstrate that the DES-YOLO model achieves a mean average precision (mAP) of 92.6% on the CSD-DET dataset and a single-image inference speed of 3.9 milliseconds. The proposed method proves capable of swiftly and accurately accomplishing real-time detection of surface defects in castings.

15.
PeerJ Comput Sci ; 10: e2313, 2024.
Artigo em Inglês | MEDLINE | ID: mdl-39314705

RESUMO

To address issues such as misdetection and omission due to low light, image defocus, and worker occlusion in coal-rock image recognition, a new method called YOLOv8-Coal, based on YOLOv8, is introduced to enhance recognition accuracy and processing speed. The Deformable Convolution Network version 3 enhances object feature extraction by adjusting sampling positions with offsets and aligning them closely with the object's shape. The Polarized Self-Attention module in the feature fusion network emphasizes crucial features and suppresses unnecessary information to minimize irrelevant factors. Additionally, the lightweight C2fGhost module combines the strengths of GhostNet and the C2f module, further decreasing model parameters and computational load. The empirical findings indicate that YOLOv8-Coal has achieved substantial enhancements in all metrics on the coal rock image dataset. More precisely, the values for AP50, AP50:95, and AR50:95 were improved to 77.7%, 62.8%, and 75.0% respectively. In addition, optimal localization recall precision (oLRP) were decreased to 45.6%. In addition, the model parameters were decreased to 2.59M and the FLOPs were reduced to 6.9G. Finally, the size of the model weight file is a mere 5.2 MB. The enhanced algorithm's advantage is further demonstrated when compared to other commonly used algorithms.

16.
PeerJ Comput Sci ; 10: e2260, 2024.
Artigo em Inglês | MEDLINE | ID: mdl-39314711

RESUMO

Point clouds are highly regarded in the field of 3D object detection for their superior geometric properties and versatility. However, object occlusion and defects in scanning equipment frequently result in sparse and missing data within point clouds, adversely affecting the final prediction. Recognizing the synergistic potential between the rich semantic information present in images and the geometric data in point clouds for scene representation, we introduce a two-stage fusion framework (TSFF) for 3D object detection. To address the issue of corrupted geometric information in point clouds caused by object occlusion, we augment point features with image features, thereby enhancing the reference factor of the point cloud during the voting bias phase. Furthermore, we implement a constrained fusion module to selectively sample voting points using a 2D bounding box, integrating valuable image features while reducing the impact of background points in sparse scenes. Our methodology was evaluated on the SUNRGB-D dataset, where it achieved a 3.6 mean average percent (mAP) improvement in the mAP@0.25 evaluation criterion over the baseline. In comparison to other great 3D object detection methods, our method had excellent performance in the detection of some objects.

17.
PeerJ Comput Sci ; 10: e2233, 2024.
Artigo em Inglês | MEDLINE | ID: mdl-39314728

RESUMO

With the rapid increase in vehicle numbers, efficient traffic management has become a critical challenge for society. Traditional methods of vehicle detection and classification often struggle with the diverse characteristics of vehicles, such as varying shapes, colors, edges, shadows, and textures. To address this, we proposed an innovative ensemble method that combines two state-of-the-art deep learning models i.e., EfficientDet and YOLOv8. The proposed work leverages data from the Forward-Looking Infrared (FLIR) dataset, which provides both thermal and RGB images. To enhance the model performance and to address the class imbalances, we applied several data augmentation techniques. Experimental results demonstrate that the proposed ensemble model achieves a mean average precision (mAP) of 95.5% on thermal images, outperforming the individual performances of EfficientDet and YOLOv8, which achieved mAPs of 92.6% and 89.4% respectively. Additionally, the ensemble model attained an average recall (AR) of 0.93 and an optimal localization recall precision (oLRP) of 0.08 on thermal images. For RGB images, the ensemble model achieved mAP of 93.1%, AR of 0.91, and oLRP of 0.10, consistently surpassing the performance of its constituent models. These findings highlight the effectiveness of proposed ensemble approach in improving vehicle detection and classification. The integration of thermal imaging further enhances detection capabilities under various lighting conditions, making the system robust for real-world applications in intelligent traffic management.

18.
Open Res Eur ; 4: 101, 2024.
Artigo em Inglês | MEDLINE | ID: mdl-39309190

RESUMO

In recent years, deep learning has gained popularity for its ability to solve complex classification tasks. It provides increasingly better results thanks to the development of more accurate models, the availability of huge volumes of data and the improved computational capabilities of modern computers. However, these improvements in performance also bring efficiency problems, related to the storage of datasets and models, and to the waste of energy and time involved in both the training and inference processes. In this context, data reduction can help reduce energy consumption when training a deep learning model. In this paper, we present up to eight different methods to reduce the size of a tabular training dataset, and we develop a Python package to apply them. We also introduce a representativeness metric based on topology to measure the similarity between the reduced datasets and the full training dataset. Additionally, we develop a methodology to apply these data reduction methods to image datasets for object detection tasks. Finally, we experimentally compare how these data reduction methods affect the representativeness of the reduced dataset, the energy consumption and the predictive performance of the model.

19.
Data Brief ; 57: 110887, 2024 Dec.
Artigo em Inglês | MEDLINE | ID: mdl-39290432

RESUMO

This article describes a dataset comprising 16,426 real-world urban photographs, capturing vehicles, cyclists, motorbikes, and pedestrians across Morning, Evening, and Night scenes. The dataset is valuable for machine learning tasks in traffic analysis, urban planning, and public safety. It enables the development and validation of algorithms for pedestrian detection, traffic flow analysis, and infrastructure optimization. Our main goal is to assist academics, urban planners, and decision-makers in creating sophisticated models for pedestrian safety, traffic control, and accident avoidance. This dataset is a useful resource for training and verifying algorithms targeted at boosting real-time traffic monitoring systems, optimizing urban infrastructure, and raising overall road safety because of its high variability and significant volume. This dataset represents a major advancement for smart city projects and the creation of intelligent transportation systems.

20.
J Environ Manage ; 369: 122246, 2024 Oct.
Artigo em Inglês | MEDLINE | ID: mdl-39241598

RESUMO

Seagrass meadows are an essential part of the Great Barrier Reef ecosystem, providing various benefits such as filtering nutrients and sediment, serving as a nursery for fish and shellfish, and capturing atmospheric carbon as blue carbon. Understanding the phenotypic plasticity of seagrasses and their ability to acclimate their morphology in response to environ-mental stressors is crucial. Investigating these morphological changes can provide valuable insights into ecosystem health and inform conservation strategies aimed at mitigating seagrass decline. Measuring seagrass growth by measuring morphological parameters such as the length and width of leaves, rhizomes, and roots is essential. The manual process of measuring morphological parameters of seagrass can be time-consuming, inaccurate and costly, so researchers are exploring machine-learning techniques to automate the process. To automate this process, researchers have developed a machine learning model that utilizes image processing and artificial intelligence to measure morphological parameters from digital imagery. The study uses a deep learning model called YOLO-v6 to classify three distinct seagrass object types and determine their dimensions. The results suggest that the proposed model is highly effective, with an average recall of 97.5%, an average precision of 83.7%, and an average f1 score of 90.1%. The model code has been made publicly available on GitHub (https://github.com/sajalhalder/AI-ASMM).


Assuntos
Inteligência Artificial , Aprendizado de Máquina , Ecossistema , Alismatales/anatomia & histologia , Alismatales/crescimento & desenvolvimento
SELEÇÃO DE REFERÊNCIAS
DETALHE DA PESQUISA
...