Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 28
Filtrar
1.
Brief Bioinform ; 25(3)2024 Mar 27.
Artigo em Inglês | MEDLINE | ID: mdl-38561979

RESUMO

Peptide binding to major histocompatibility complex (MHC) proteins plays a critical role in T-cell recognition and the specificity of the immune response. Experimental validation such peptides is extremely resource-intensive. As a result, accurate computational prediction of binding peptides is highly important, particularly in the context of cancer immunotherapy applications, such as the identification of neoantigens. In recent years, there is a significant need to continually improve the existing prediction methods to meet the demands of this field. We developed ConvNeXt-MHC, a method for predicting MHC-I-peptide binding affinity. It introduces a degenerate encoding approach to enhance well-established panspecific methods and integrates transfer learning and semi-supervised learning methods into the cutting-edge deep learning framework ConvNeXt. Comprehensive benchmark results demonstrate that ConvNeXt-MHC outperforms state-of-the-art methods in terms of accuracy. We expect that ConvNeXt-MHC will help us foster new discoveries in the field of immunoinformatics in the distant future. We constructed a user-friendly website at http://www.combio-lezhang.online/predict/, where users can access our data and application.


Assuntos
Peptídeos , Peptídeos/metabolismo , Ligação Proteica
2.
Sensors (Basel) ; 24(13)2024 Jun 23.
Artigo em Inglês | MEDLINE | ID: mdl-39000856

RESUMO

Fire is a significant security threat that can lead to casualties, property damage, and environmental damage. Despite the availability of object-detection algorithms, challenges persist in detecting fires, smoke, and humans. These challenges include poor performance in detecting small fires and smoke, as well as a high computational cost, which limits deployments. In this paper, we propose an end-to-end object detector for fire, smoke, and human detection based on Deformable DETR (DEtection TRansformer) called FSH-DETR. To effectively process multi-scale fire and smoke features, we propose a novel Mixed Encoder, which integrates SSFI (Separate Single-scale Feature Interaction Module) and CCFM (CNN-based Cross-scale Feature Fusion Module) for multi-scale fire, smoke, and human feature fusion. Furthermore, we enhance the convergence speed of FSH-DETR by incorporating a bounding box loss function called PIoUv2 (Powerful Intersection of Union), which improves the precision of fire, smoke, and human detection. Extensive experiments on the public dataset demonstrate that the proposed method surpasses state-of-the-art methods in terms of the mAP (mean Average Precision), with mAP and mAP50 reaching 66.7% and 84.2%, respectively.

3.
Sheng Wu Yi Xue Gong Cheng Xue Za Zhi ; 41(3): 544-551, 2024 Jun 25.
Artigo em Chinês | MEDLINE | ID: mdl-38932541

RESUMO

Skin cancer is a significant public health issue, and computer-aided diagnosis technology can effectively alleviate this burden. Accurate identification of skin lesion types is crucial when employing computer-aided diagnosis. This study proposes a multi-level attention cascaded fusion model based on Swin-T and ConvNeXt. It employed hierarchical Swin-T and ConvNeXt to extract global and local features, respectively, and introduced residual channel attention and spatial attention modules for further feature extraction. Multi-level attention mechanisms were utilized to process multi-scale global and local features. To address the problem of shallow features being lost due to their distance from the classifier, a hierarchical inverted residual fusion module was proposed to dynamically adjust the extracted feature information. Balanced sampling strategies and focal loss were employed to tackle the issue of imbalanced categories of skin lesions. Experimental testing on the ISIC2018 and ISIC2019 datasets yielded accuracy, precision, recall, and F1-Score of 96.01%, 93.67%, 92.65%, and 93.11%, respectively, and 92.79%, 91.52%, 88.90%, and 90.15%, respectively. Compared to Swin-T, the proposed method achieved an accuracy improvement of 3.60% and 1.66%, and compared to ConvNeXt, it achieved an accuracy improvement of 2.87% and 3.45%. The experiments demonstrate that the proposed method accurately classifies skin lesion images, providing a new solution for skin cancer diagnosis.


Assuntos
Algoritmos , Diagnóstico por Computador , Neoplasias Cutâneas , Humanos , Neoplasias Cutâneas/patologia , Neoplasias Cutâneas/diagnóstico por imagem , Neoplasias Cutâneas/classificação , Diagnóstico por Computador/métodos , Pele/patologia , Interpretação de Imagem Assistida por Computador/métodos
4.
Sensors (Basel) ; 23(11)2023 Jun 05.
Artigo em Inglês | MEDLINE | ID: mdl-37300061

RESUMO

This article introduces a novel framework for diagnosing faults in rolling bearings. The framework combines digital twin data, transfer learning theory, and an enhanced ConvNext deep learning network model. Its purpose is to address the challenges posed by the limited actual fault data density and inadequate result accuracy in existing research on the detection of rolling bearing faults in rotating mechanical equipment. To begin with, the operational rolling bearing is represented in the digital realm through the utilization of a digital twin model. The simulation data produced by this twin model replace traditional experimental data, effectively creating a substantial volume of well-balanced simulated datasets. Next, improvements are made to the ConvNext network by incorporating an unparameterized attention module called the Similarity Attention Module (SimAM) and an efficient channel attention feature referred to as the Efficient Channel Attention Network (ECA). These enhancements serve to augment the network's capability for extracting features. Subsequently, the enhanced network model is trained using the source domain dataset. Simultaneously, the trained model is transferred to the target domain bearing using transfer learning techniques. This transfer learning process enables the accurate fault diagnosis of the main bearing to be achieved. Finally, the proposed method's feasibility is validated, and a comparative analysis is conducted in comparison with similar approaches. The comparative study demonstrates that the proposed method effectively addresses the issue of low mechanical equipment fault data density, leading to improved accuracy in fault detection and classification, along with a certain level of robustness.

5.
Sensors (Basel) ; 23(6)2023 Mar 22.
Artigo em Inglês | MEDLINE | ID: mdl-36992039

RESUMO

Along with society's development, transportation has become a key factor in human daily life, increasing the number of vehicles on the streets. Consequently, the task of finding free parking slots in metropolitan areas can be dramatically challenging, increasing the chance of getting involved in an accident and the carbon footprint, and negatively affecting the driver's health. Therefore, technological resources to deal with parking management and real-time monitoring have become key players in this scenario to speed up the parking process in urban areas. This work proposes a new computer-vision-based system that detects vacant parking spaces in challenging situations using color imagery processed by a novel deep-learning algorithm. This is based on a multi-branch output neural network that maximizes the contextual image information to infer the occupancy of every parking space. Every output infers the occupancy of a specific parking slot using all the input image information, unlike existing approaches, which only use a neighborhood around every slot. This allows it to be very robust to changing illumination conditions, different camera perspectives, and mutual occlusions between parked cars. An extensive evaluation has been performed using several public datasets, proving that the proposed system outperforms existing approaches.

6.
Sensors (Basel) ; 23(18)2023 Sep 19.
Artigo em Inglês | MEDLINE | ID: mdl-37766026

RESUMO

Historically, individuals with hearing impairments have faced neglect, lacking the necessary tools to facilitate effective communication. However, advancements in modern technology have paved the way for the development of various tools and software aimed at improving the quality of life for hearing-disabled individuals. This research paper presents a comprehensive study employing five distinct deep learning models to recognize hand gestures for the American Sign Language (ASL) alphabet. The primary objective of this study was to leverage contemporary technology to bridge the communication gap between hearing-impaired individuals and individuals with no hearing impairment. The models utilized in this research include AlexNet, ConvNeXt, EfficientNet, ResNet-50, and VisionTransformer were trained and tested using an extensive dataset comprising over 87,000 images of the ASL alphabet hand gestures. Numerous experiments were conducted, involving modifications to the architectural design parameters of the models to obtain maximum recognition accuracy. The experimental results of our study revealed that ResNet-50 achieved an exceptional accuracy rate of 99.98%, the highest among all models. EfficientNet attained an accuracy rate of 99.95%, ConvNeXt achieved 99.51% accuracy, AlexNet attained 99.50% accuracy, while VisionTransformer yielded the lowest accuracy of 88.59%.


Assuntos
Aprendizado Profundo , Língua de Sinais , Humanos , Estados Unidos , Qualidade de Vida , Gestos , Tecnologia
7.
Sensors (Basel) ; 23(17)2023 Sep 03.
Artigo em Inglês | MEDLINE | ID: mdl-37688096

RESUMO

The advancements in ship detection technology using convolutional neural networks (CNNs) regarding synthetic aperture radar (SAR) images have been significant. Yet, there are still some limitations in the existing detection algorithms. First, the backbones cannot generate high-quality multiscale feature maps. Second, there is a lack of suitable attention mechanisms to suppress false alarms. Third, the current feature intensification algorithms are unable to effectively enhance the shallow feature's semantic information, which hinders the detection of small ships. Fourth, top-level feature maps have rich semantic information; however, as a result of the reduction of channels, the semantic information is weakened. These four problems lead to poor performance in SAR ship detection and recognition. To address the mentioned issues, we put forward a new approach that has the following characteristics. First, we use Convnext as the backbone to generate high-quality multiscale feature maps. Second, to suppress false alarms, the multi-pooling channel attention (MPCA) is designed to generate a corresponding weight for each channel, suppressing redundant feature maps, and further optimizing the feature maps generated by Convnext. Third, a feature intensification pyramid network (FIPN) is specifically designed to intensify the feature maps, especially the shallow feature maps. Fourth, a top-level feature intensification (TLFI) is also proposed to compensate for semantic information loss within the top-level feature maps by utilizing semantic information from different spaces. The experimental dataset employed is the SAR Ship Detection Dataset (SSDD), and the experimental findings display that our approach exhibits superiority compared to other advanced approaches. The overall Average Precision (AP) reaches up to 95.6% on the SSDD, which improves the accuracy by at least 1.7% compared to the current excellent methods.

8.
Sensors (Basel) ; 22(10)2022 May 23.
Artigo em Inglês | MEDLINE | ID: mdl-35632344

RESUMO

Three-dimensional object detection in the point cloud can provide more accurate object data for autonomous driving. In this paper, we propose a method named MA-MFFC that uses an attention mechanism and a multi-scale feature fusion network with ConvNeXt module to improve the accuracy of object detection. The multi-attention (MA) module contains point-channel attention and voxel attention, which are used in voxelization and 3D backbone. By considering the point-wise and channel-wise, the attention mechanism enhances the information of key points in voxels, suppresses background point clouds in voxelization, and improves the robustness of the network. The voxel attention module is used in the 3D backbone to obtain more robust and discriminative voxel features. The MFFC module contains the multi-scale feature fusion network and the ConvNeXt module; the multi-scale feature fusion network can extract rich feature information and improve the detection accuracy, and the convolutional layer is replaced with the ConvNeXt module to enhance the feature extraction capability of the network. The experimental results show that the average accuracy is 64.60% for pedestrians and 80.92% for cyclists on the KITTI dataset, which is 1.33% and 2.1% higher, respectively, compared with the baseline network, enabling more accurate detection and localization of more difficult objects.


Assuntos
Veículos Autônomos , Humanos , Pedestres
9.
Biomed Eng Lett ; 14(2): 341-353, 2024 Mar.
Artigo em Inglês | MEDLINE | ID: mdl-38374903

RESUMO

Deep learning-based methods have recently shown great promise in medical image segmentation task. However, CNN-based frameworks struggle with inadequate long-range spatial dependency capture, whereas Transformers suffer from computational inefficiency and necessitate substantial volumes of labeled data for effective training. To tackle these issues, this paper introduces CI-UNet, a novel architecture that utilizes ConvNeXt as its encoder, amalgamating the computational efficiency and feature extraction capabilities. Moreover, an advanced attention mechanism is proposed to captures intricate cross-dimensional interactions and global context. Extensive experiments on two segmentation datasets, namely BCSD, and CT2USforKidneySeg, confirm the excellent performance of the proposed CI-UNet as compared to other segmentation methods.

10.
Comput Biol Med ; 177: 108592, 2024 Jul.
Artigo em Inglês | MEDLINE | ID: mdl-38781642

RESUMO

Cardiac MRI segmentation is a significant research area in medical image processing, holding immense clinical and scientific importance in assisting the diagnosis and treatment of heart diseases. Currently, existing cardiac MRI segmentation algorithms are often constrained by specific datasets and conditions, leading to a notable decrease in segmentation performance when applied to diverse datasets. These limitations affect the algorithm's overall performance and generalization capabilities. Inspired by ConvNext, we introduce a two-dimensional cardiac MRI segmentation U-shaped network called ConvNextUNet. It is the first application of a combination of ConvNext and the U-shaped architecture in the field of cardiac MRI segmentation. Firstly, we incorporate up-sampling modules into the original ConvNext architecture and combine it with the U-shaped framework to achieve accurate reconstruction. Secondly, we integrate Input Stem into ConvNext, and introduce attention mechanisms along the bridging path. By merging features extracted from both the encoder and decoder, a probability distribution is obtained through linear and nonlinear transformations, serving as attention weights, thereby enhancing the signal of the same region of interest. The resulting attention weights are applied to the decoder features, highlighting the region of interest. This allows the model to simultaneously consider local context and global details during the learning phase, fully leveraging the advantages of both global and local perception for a more comprehensive understanding of cardiac anatomical structures. Consequently, the model demonstrates a clear advantage and robust generalization capability, especially in small-region segmentation. Experimental results on the ACDC, LVQuan19, and RVSC datasets confirm that the ConvNextUNet model outperforms the current state-of-the-art models, particularly in small-region segmentation tasks. Furthermore, we conducted cross-dataset training and testing experiments, which revealed that the pre-trained model can accurately segment diverse cardiac datasets, showcasing its powerful generalization capabilities. The source code of this project is available at https://github.com/Zemin-Cai/ConvNextUNet.


Assuntos
Algoritmos , Imageamento por Ressonância Magnética , Humanos , Imageamento por Ressonância Magnética/métodos , Coração/diagnóstico por imagem , Processamento de Imagem Assistida por Computador/métodos
11.
J Food Sci ; 89(6): 3369-3383, 2024 Jun.
Artigo em Inglês | MEDLINE | ID: mdl-38720576

RESUMO

Salted egg yolks from salted duck eggs are widely utilized in the domestic and international food industry as both raw materials and ingredients. When salted egg yolks are not fully cured and matured, they exist in a fluid state, with a mixture of solid and liquid internally. Due to this composition, they are susceptible to deterioration during storage and usage, necessitating their detection and classification. In this study, a dataset specifically for salted egg yolks was established, and the ConvNeXt-T model, employed as the benchmark model, underwent two notable improvements. First, a lightweight location-aware circular convolution (ParC) was introduced, utilizing a ParC-block to replace a portion of the original ConvNeXt-T block. This enhancement aimed to overcome the limitations of convolution in extracting global feature information while integrating the global sensing capability of vision transformer and the localization capability of convolution. Additionally, the activation function was modified through substitution. These improvements resulted in the final model. Experimental results indicate that the enhanced model exhibits faster convergence on the custom salted egg yolk dataset compared to the baseline model. Furthermore, a significant reduction of model parameters by a factor of 4 led to a 2.167 percentage point improvement in the accuracy of the test set. The ParC-ConvNeXt-SMU-T model achieved an accuracy of 96.833% with 26.8 million parameters. Notably, the improved model demonstrates exceptional effectiveness in recognizing salted egg yolks. PRACTICAL APPLICATION: This study can be widely applied in the process of salted egg yolk production and quality inspection, which can improve the actual sorting efficiency of salted egg yolks and reduce the labor cost at the same time. It can also be used for nondestructive testing of salted egg yolks by governmental enterprises and other regulatory authorities.


Assuntos
Gema de Ovo , Gema de Ovo/química , Animais , Patos , Manipulação de Alimentos/métodos , Cloreto de Sódio/análise , Cloreto de Sódio/química
12.
Dermatopathology (Basel) ; 11(3): 239-252, 2024 Aug 15.
Artigo em Inglês | MEDLINE | ID: mdl-39189182

RESUMO

Skin tumors, especially melanoma, which is highly aggressive and progresses quickly to other sites, are an issue in various parts of the world. Nevertheless, the one and only way to save lives is to detect it at its initial stages. This study explores the application of advanced deep learning models for classifying benign and malignant melanoma using dermoscopic images. The aim of the study is to enhance the accuracy and efficiency of melanoma diagnosis with the ConvNeXt, Vision Transformer (ViT) Base-16, and Swin Transformer V2 Small (Swin V2 S) deep learning models. The ConvNeXt model, which integrates principles of both convolutional neural networks and transformers, demonstrated superior performance, with balanced precision and recall metrics. The dataset, sourced from Kaggle, comprises 13,900 uniformly sized images, preprocessed to standardize the inputs for the models. Experimental results revealed that ConvNeXt achieved the highest diagnostic accuracy among the tested models. Experimental results revealed that ConvNeXt achieved an accuracy of 91.5%, with balanced precision and recall rates of 90.45% and 92.8% for benign cases, and 92.61% and 90.2% for malignant cases, respectively. The F1-scores for ConvNeXt were 91.61% for benign cases and 91.39% for malignant cases. This research points out the potential of hybrid deep learning architectures in medical image analysis, particularly for early melanoma detection.

13.
Phys Med Biol ; 69(8)2024 Apr 03.
Artigo em Inglês | MEDLINE | ID: mdl-38479022

RESUMO

Objective.Multi-contrast magnetic resonance imaging (MC MRI) can obtain more comprehensive anatomical information of the same scanning object but requires a longer acquisition time than single-contrast MRI. To accelerate MC MRI speed, recent studies only collect partial k-space data of one modality (target contrast) to reconstruct the remaining non-sampled measurements using a deep learning-based model with the assistance of another fully sampled modality (reference contrast). However, MC MRI reconstruction mainly performs the image domain reconstruction with conventional CNN-based structures by full supervision. It ignores the prior information from reference contrast images in other sparse domains and requires fully sampled target contrast data. In addition, because of the limited receptive field, conventional CNN-based networks are difficult to build a high-quality non-local dependency.Approach.In the paper, we propose an Image-Wavelet domain ConvNeXt-based network (IWNeXt) for self-supervised MC MRI reconstruction. Firstly, INeXt and WNeXt based on ConvNeXt reconstruct undersampled target contrast data in the image domain and refine the initial reconstructed result in the wavelet domain respectively. To generate more tissue details in the refinement stage, reference contrast wavelet sub-bands are used as additional supplementary information for wavelet domain reconstruction. Then we design a novel attention ConvNeXt block for feature extraction, which can capture the non-local information of the MC image. Finally, the cross-domain consistency loss is designed for self-supervised learning. Especially, the frequency domain consistency loss deduces the non-sampled data, while the image and wavelet domain consistency loss retain more high-frequency information in the final reconstruction.Main results.Numerous experiments are conducted on the HCP dataset and the M4Raw dataset with different sampling trajectories. Compared with DuDoRNet, our model improves by 1.651 dB in the peak signal-to-noise ratio.Significance.IWNeXt is a potential cross-domain method that can enhance the accuracy of MC MRI reconstruction and reduce reliance on fully sampled target contrast images.


Assuntos
Processamento de Imagem Assistida por Computador , Imageamento por Ressonância Magnética , Processamento de Imagem Assistida por Computador/métodos , Tempo , Imageamento por Ressonância Magnética/métodos , Razão Sinal-Ruído
14.
Comput Biol Med ; 172: 108317, 2024 Apr.
Artigo em Inglês | MEDLINE | ID: mdl-38492455

RESUMO

Crafting effective deep learning models for medical image analysis is a complex task, particularly in cases where the medical image dataset lacks significant inter-class variation. This challenge is further aggravated when employing such datasets to generate synthetic images using generative adversarial networks (GANs), as the output of GANs heavily relies on the input data. In this research, we propose a novel filtering algorithm called Cosine Similarity-based Image Filtering (CosSIF). We leverage CosSIF to develop two distinct filtering methods: Filtering Before GAN Training (FBGT) and Filtering After GAN Training (FAGT). FBGT involves the removal of real images that exhibit similarities to images of other classes before utilizing them as the training dataset for a GAN. On the other hand, FAGT focuses on eliminating synthetic images with less discriminative features compared to real images used for training the GAN. The experimental results reveal that the utilization of either the FAGT or FBGT method reduces low inter-class variation in clinical image classification datasets and enables GANs to generate synthetic images with greater discriminative features. Moreover, modern transformer and convolutional-based models, trained with datasets that utilize these filtering methods, lead to less bias toward the majority class, more accurate predictions of samples in the minority class, and overall better generalization capabilities. Code and implementation details are available at: https://github.com/mominul-ssv/cossif.


Assuntos
Algoritmos , Extremidade Superior , Processamento de Imagem Assistida por Computador
15.
Vis Comput Ind Biomed Art ; 7(1): 2, 2024 Jan 26.
Artigo em Inglês | MEDLINE | ID: mdl-38273164

RESUMO

Accurate segmentation of breast ultrasound (BUS) images is crucial for early diagnosis and treatment of breast cancer. Further, the task of segmenting lesions in BUS images continues to pose significant challenges due to the limitations of convolutional neural networks (CNNs) in capturing long-range dependencies and obtaining global context information. Existing methods relying solely on CNNs have struggled to address these issues. Recently, ConvNeXts have emerged as a promising architecture for CNNs, while transformers have demonstrated outstanding performance in diverse computer vision tasks, including the analysis of medical images. In this paper, we propose a novel breast lesion segmentation network CS-Net that combines the strengths of ConvNeXt and Swin Transformer models to enhance the performance of the U-Net architecture. Our network operates on BUS images and adopts an end-to-end approach to perform segmentation. To address the limitations of CNNs, we design a hybrid encoder that incorporates modified ConvNeXt convolutions and Swin Transformer. Furthermore, to enhance capturing the spatial and channel attention in feature maps we incorporate the Coordinate Attention Module. Second, we design an Encoder-Decoder Features Fusion Module that facilitates the fusion of low-level features from the encoder with high-level semantic features from the decoder during the image reconstruction. Experimental results demonstrate the superiority of our network over state-of-the-art image segmentation methods for BUS lesions segmentation.

16.
Spectrochim Acta A Mol Biomol Spectrosc ; 311: 123966, 2024 Apr 15.
Artigo em Inglês | MEDLINE | ID: mdl-38335591

RESUMO

Potatoes are popular among consumers due to their high yield and delicious taste. However, due to the numerous varieties of potatoes, different varieties are suitable for different processing methods. Therefore, it is necessary to distinguish varieties after harvest to meet the needs of processing enterprises and consumers. In this study, a new visible-near-infrared spectroscopic analysis method was proposed, which can achieve detection of five potato varieties. The method measures the transmission and reflection spectra of potatoes using a spectral acquisition system, encodes one-dimensional spectra into two-dimensional images using Gramian Angular Summation Field (GASF), Gramian Angular Difference Field (GADF), Markov Transition Field (MTF) and Recurrence Plot (RP), and improves the coordinated attention mechanism module and embeds the improved module into the ConvNeXt V2 model to build the ConvNeXt V2-CAP model for potato variety classification. The results show that compared with directly using one-dimensional classification models, image encoding of spectral data for classification greatly improves the accuracy. Among them, the best accuracy of 99.54% is achieved by using GADF image encoding of transmission spectra combined with the ConvNeXt V2-CAP model for classification, which is 16.28% higher than the highest accuracy of the one-dimensional classification model. The CAP attention mechanism module improves the performance of the model, especially when the dataset is small. When the training set is reduced to 150 images, the accuracy of the model is improved by 2.33% compared to the original model. Therefore, it is feasible to classify potato varieties using visible-near infrared spectroscopy and image encoding technology.

17.
Foods ; 13(6)2024 Mar 19.
Artigo em Inglês | MEDLINE | ID: mdl-38540915

RESUMO

As a traditional delicacy in China, preserved eggs inevitably experience instances of substandard quality during the production process. Chinese preserved egg production facilities can only rely on experienced workers to select the preserved eggs. However, the manual selection of preserved eggs presents challenges such as a low efficiency, subjective judgments, high costs, and hindered industrial production processes. In response to these challenges, this study procured the transmitted imagery of preserved eggs and refined the ConvNeXt network across four pivotal dimensions: the dimensionality reduction of model feature maps, the integration of multi-scale feature fusion (MSFF), the incorporation of a global attention mechanism (GAM) module, and the amalgamation of the cross-entropy loss function with focal loss. The resultant refined model, ConvNeXt_PEgg, attained proficiency in classifying and grading preserved eggs. Notably, the improved model achieved a classification accuracy of 92.6% across the five categories of preserved eggs, with a grading accuracy of 95.9% spanning three levels. Moreover, in contrast to its predecessor, the refined model witnessed a 24.5% reduction in the parameter volume, alongside a 3.2 percentage point augmentation in the classification accuracy and a 2.8 percentage point boost in the grading accuracy. Through meticulous comparative analysis, each enhancement exhibited varying degrees of performance elevation. Evidently, the refined model outshone a plethora of classical models, underscoring its efficacy in discerning the internal quality of preserved eggs. With its potential for real-world implementation, this technology portends to heighten the economic viability of manufacturing facilities.

18.
Front Bioeng Biotechnol ; 11: 1191803, 2023.
Artigo em Inglês | MEDLINE | ID: mdl-37324431

RESUMO

Accurate segmentation of retinal layer boundaries can facilitate the detection of patients with early ophthalmic disease. Typical segmentation algorithms operate at low resolutions without fully exploiting multi-granularity visual features. Moreover, several related studies do not release their datasets that are key for the research on deep learning-based solutions. We propose a novel end-to-end retinal layer segmentation network based on ConvNeXt, which can retain more feature map details by using a new depth-efficient attention module and multi-scale structures. In addition, we provide a semantic segmentation dataset containing 206 retinal images of healthy human eyes (named NR206 dataset), which is easy to use as it does not require any additional transcoding processing. We experimentally show that our segmentation approach outperforms state-of-the-art approaches on this new dataset, achieving, on average, a Dice score of 91.3% and mIoU of 84.4%. Moreover, our approach achieves state-of-the-art performance on a glaucoma dataset and a diabetic macular edema (DME) dataset, showing that our model is also suitable for other applications. We will make our source code and the NR206 dataset publicly available at (https://github.com/Medical-Image-Analysis/Retinal-layer-segmentation).

19.
PeerJ Comput Sci ; 9: e1446, 2023.
Artigo em Inglês | MEDLINE | ID: mdl-37705628

RESUMO

Rapid developments in automatic driving technology have given rise to new experiences for passengers. Safety is a main priority in automatic driving. A strong familiarity with road-surface conditions during the day and night is essential to ensuring driving safety. Existing models used for recognizing road-surface conditions lack the required robustness and generalization abilities. Most studies only validated the performance of these models on daylight images. To address this problem, we propose a novel multi-supervised bidirectional fusion network (MBFN) model to detect weather-induced road-surface conditions on the path of automatic vehicles at both daytime and nighttime. We employed ConvNeXt to extract the basic features, which were further processed using a new bidirectional fusion module to create a fused feature. Then, the basic and fused features were concatenated to generate a refined feature with greater discriminative and generalization abilities. Finally, we designed a multi-supervised loss function to train the MBFN model based on the extracted features. Experiments were conducted using two public datasets. The results clearly demonstrated that the MBFN model could classify diverse road-surface conditions, such as dry, wet, and snowy conditions, with a satisfactory accuracy and outperform state-of-the-art baseline models. Notably, the proposed model has multiple variants that could also achieve competitive performances under different road conditions. The code for the MBFN model is shared at https://zenodo.org/badge/latestdoi/607014079.

20.
Comput Biol Med ; 166: 107541, 2023 Sep 30.
Artigo em Inglês | MEDLINE | ID: mdl-37804779

RESUMO

Colorectal cancer (CRC) holds the distinction of being the most prevalent malignant tumor affecting the digestive system. It is a formidable global health challenge, as it ranks as the fourth leading cause of cancer-related fatalities around the world. Despite considerable advancements in comprehending and addressing colorectal cancer (CRC), the likelihood of recurring tumors and metastasis remains a major cause of high morbidity and mortality rates during treatment. Currently, colonoscopy is the predominant method for CRC screening. Artificial intelligence has emerged as a promising tool in aiding the diagnosis of polyps, which have demonstrated significant potential. Unfortunately, most segmentation methods face challenges in terms of limited accuracy and generalization to different datasets, especially the slow processing and analysis speed has become a major obstacle. In this study, we propose a fast and efficient polyp segmentation framework based on the Large-Kernel Receptive Field Block (LK-RFB) and Global Parallel Partial Decoder(GPPD). Our proposed ColonNet has been extensively tested and proven effective, achieving a DICE coefficient of over 0.910 and an FPS of over 102 on the CVC-300 dataset. In comparison to the state-of-the-art (SOTA) methods, ColonNet outperforms or achieves comparable performance on five publicly available datasets, establishing a new SOTA. Compared to state-of-the-art methods, ColonNet achieves the highest FPS (over 102 FPS) while maintaining excellent segmentation results, achieving the best or comparable performance on the five public datasets. The code will be released at: https://github.com/SPECTRELWF/ColonNet.

SELEÇÃO DE REFERÊNCIAS
Detalhe da pesquisa