Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 124
Filtrar
1.
PeerJ Comput Sci ; 10: e2245, 2024.
Artigo em Inglês | MEDLINE | ID: mdl-39314694

RESUMO

Artificial intelligence (AI) and machine learning (ML) aim to mimic human intelligence and enhance decision making processes across various fields. A key performance determinant in a ML model is the ratio between the training and testing dataset. This research investigates the impact of varying train-test split ratios on machine learning model performance and generalization capabilities using the BraTS 2013 dataset. Logistic regression, random forest, k nearest neighbors, and support vector machines were trained with split ratios ranging from 60:40 to 95:05. Findings reveal significant variations in accuracies across these ratios, emphasizing the critical need to strike a balance to avoid overfitting or underfitting. The study underscores the importance of selecting an optimal train-test split ratio that considers tradeoffs such as model performance metrics, statistical measures, and resource constraints. Ultimately, these insights contribute to a deeper understanding of how ratio selection impacts the effectiveness and reliability of machine learning applications across diverse fields.

2.
J Imaging Inform Med ; 2024 Sep 16.
Artigo em Inglês | MEDLINE | ID: mdl-39284982

RESUMO

Cutaneous melanoma is a highly lethal form of cancer. Developing a medical image segmentation model capable of accurately delineating melanoma lesions with high robustness and generalization presents a formidable challenge. This study draws inspiration from cellular functional characteristics and natural selection, proposing a novel medical segmentation model named the vital characteristics cellular neural network. This model incorporates vital characteristics observed in multicellular organisms, including memory, adaptation, apoptosis, and division. Memory module enables the network to rapidly adapt to input data during the early stages of training, accelerating model convergence. Adaptation module allows neurons to select the appropriate activation function based on varying environmental conditions. Apoptosis module reduces the risk of overfitting by pruning neurons with low activation values. Division module enhances the network's learning capacity by duplicating neurons with high activation values. Experimental evaluations demonstrate the efficacy of this model in enhancing the performance of neural networks for medical image segmentation. The proposed method achieves outstanding results across numerous publicly available datasets, indicating its potential to contribute significantly to the field of medical image analysis and facilitating accurate and efficient segmentation of medical imagery. The proposed method achieves outstanding results across numerous publicly available datasets, with an F1 score of 0.901, Intersection over Union of 0.841, and Dice coefficient of 0.913, indicating its potential to contribute significantly to the field of medical image analysis and facilitating accurate and efficient segmentation of medical imagery.

3.
Med Phys ; 2024 Aug 21.
Artigo em Inglês | MEDLINE | ID: mdl-39167055

RESUMO

BACKGROUND: Adaptive radiotherapy (ART) workflows have been increasingly adopted to achieve dose escalation and tissue sparing under shifting anatomic conditions, but the necessity of recontouring and the associated time burden hinders a real-time or online ART workflow. In response to this challenge, approaches to auto-segmentation involving deformable image registration, atlas-based segmentation, and deep learning-based segmentation (DLS) have been developed. Despite the particular promise shown by DLS methods, implementing these approaches in a clinical setting remains a challenge, namely due to the difficulty of curating a data set of sufficient size and quality so as to achieve generalizability in a trained model. PURPOSE: To address this challenge, we have developed an intentional deep overfit learning (IDOL) framework tailored to the auto-segmentation task. However, certain limitations were identified, particularly the insufficiency of the personalized dataset to effectively overfit the model. In this study, we introduce a personalized hyperspace learning (PHL)-IDOL segmentation framework capable of generating datasets that induce the model to overfit specific patient characteristics for medical image segmentation. METHODS: The PHL-IDOL model is trained in two stages. In the first, a conventional, general model is trained with a diverse set of patient data (n = 100 patients) consisting of CT images and clinical contours. Following this, the general model is tuned with a data set consisting of two components: (a) selection of a subset of the patient data (m < n) using the similarity metrics (mean square error (MSE), peak signal-to-noise ratio (PSNR), structural similarity index (SSIM), and the universal quality image index (UQI) values); (b) adjust the CT and the clinical contours using a deformed vector generated from the reference patient and the selected patients using (a). After training, the general model, the continual model, the conventional IDOL model, and the proposed PHL-IDOL model were evaluated using the volumetric dice similarity coefficient (VDSC) and the Hausdorff distance 95% (HD95%) computed for 18 structures in 20 test patients. RESULTS: Implementing the PHL-IDOL framework resulted in improved segmentation performance for each patient. The Dice scores increased from 0.81 ± $ \pm $ 0.05 with the general model, 0.83 ± 0.04 $ \pm 0.04$ for the continual model, 0.83 ± 0.04 $ \pm 0.04$ for the conventional IDOL model to an average of 0.87 ± 0.03 $ \pm 0.03$ with the PHL-IDOL model. Similarly, the Hausdorff distance decreased from 3.06 ± 0.99 $ \pm 0.99$ with the general model, 2.84 ± 0.69 $ \pm 0.69$ for the continual model, 2.79 ± 0.79 $ \pm 0.79$ for the conventional IDOL model and 2.36 ± 0.52 $ \pm 0.52$ for the PHL-IDOL model. All the standard deviations were decreased by nearly half of the values comparing the general model and the PHL-IDOL model. CONCLUSION: The PHL-IDOL framework applied to the auto-segmentation task achieves improved performance compared to the general DLS approach, demonstrating the promise of leveraging patient-specific prior information in a task central to online ART workflows.

4.
Empir Softw Eng ; 29(5): 116, 2024.
Artigo em Inglês | MEDLINE | ID: mdl-39069998

RESUMO

Previous studies have shown that Automated Program Repair (apr) techniques suffer from the overfitting problem. Overfitting happens when a patch is run and the test suite does not reveal any error, but the patch actually does not fix the underlying bug or it introduces a new defect that is not covered by the test suite. Therefore, the patches generated by apr tools need to be validated by human programmers, which can be very costly, and prevents apr tool adoption in practice. Our work aims to minimize the number of plausible patches that programmers have to review, thereby reducing the time required to find a correct patch. We introduce a novel light-weight test-based patch clustering approach called xTestCluster, which clusters patches based on their dynamic behavior. xTestCluster is applied after the patch generation phase in order to analyze the generated patches from one or more repair tools and to provide more information about those patches for facilitating patch assessment. The novelty of xTestCluster lies in using information from execution of newly generated test cases to cluster patches generated by multiple APR approaches. A cluster is formed of patches that fail on the same generated test cases. The output from xTestCluster gives developers a) a way of reducing the number of patches to analyze, as they can focus on analyzing a sample of patches from each cluster, b) additional information (new test cases and their results) attached to each patch. After analyzing 902 plausible patches from 21 Java apr tools, our results show that xTestCluster is able to reduce the number of patches to review and analyze with a median of 50%. xTestCluster can save a significant amount of time for developers that have to review the multitude of patches generated by apr tools, and provides them with new test cases that expose the differences in behavior between generated patches. Moreover, xTestCluster can complement other patch assessment techniques that help detect patch misclassifications.

5.
Front Genet ; 15: 1401470, 2024.
Artigo em Inglês | MEDLINE | ID: mdl-39050246

RESUMO

As genomic selection emerges as a promising breeding method for both plants and animals, numerous methods have been introduced and applied to various real and simulated data sets. Research suggests that no single method is universally better than others; rather, performance is highly dependent on the characteristics of the data and the nature of the prediction task. This implies that each method has its strengths and weaknesses. In this study, we exploit this notion and propose a different approach. Rather than comparing multiple methods to determine the best one for a particular study, we advocate combining multiple methods to achieve better performance than each method in isolation. In pursuit of this goal, we introduce and develop a computational method of the stacked generalization within ensemble methods. In this method, the meta-model merges predictions from multiple base models to achieve improved performance. We applied this method to plant and animal data and compared its performance with currently available methods using standard performance metrics. We found that the proposed method yielded a lower or comparable mean squared error in predicting phenotypes compared to the current methods. In addition, the proposed method showed greater resistance to overfitting compared to the current methods. Further analysis included statistical hypothesis testing, which showed that the proposed method outperformed or matched the current methods. In summary, the proposed stacked generalization integrates currently available methods to achieve stable and better performance. In this context, our study provides general recommendations for effective practices in genomic selection.

6.
Neural Netw ; 177: 106389, 2024 Sep.
Artigo em Inglês | MEDLINE | ID: mdl-38788291

RESUMO

In this work we approach attractor neural networks from a machine learning perspective: we look for optimal network parameters by applying a gradient descent over a regularized loss function. Within this framework, the optimal neuron-interaction matrices turn out to be a class of matrices which correspond to Hebbian kernels revised by a reiterated unlearning protocol. Remarkably, the extent of such unlearning is proved to be related to the regularization hyperparameter of the loss function and to the training time. Thus, we can design strategies to avoid overfitting that are formulated in terms of regularization and early-stopping tuning. The generalization capabilities of these attractor networks are also investigated: analytical results are obtained for random synthetic datasets, next, the emerging picture is corroborated by numerical experiments that highlight the existence of several regimes (i.e., overfitting, failure and success) as the dataset parameters are varied.


Assuntos
Aprendizado de Máquina , Redes Neurais de Computação , Algoritmos , Neurônios/fisiologia , Humanos
7.
BMC Bioinformatics ; 25(1): 155, 2024 Apr 20.
Artigo em Inglês | MEDLINE | ID: mdl-38641616

RESUMO

BACKGROUND: Classification of binary data arises naturally in many clinical applications, such as patient risk stratification through ICD codes. One of the key practical challenges in data classification using machine learning is to avoid overfitting. Overfitting in supervised learning primarily occurs when a model learns random variations from noisy labels in training data rather than the underlying patterns. While traditional methods such as regularization and early stopping have demonstrated effectiveness in interpolation tasks, addressing overfitting in the classification of binary data, in which predictions always amount to extrapolation, demands extrapolation-enhanced strategies. One such approach is hybrid mechanistic/data-driven modeling, which integrates prior knowledge on input features into the learning process, enhancing the model's ability to extrapolate. RESULTS: We present NoiseCut, a Python package for noise-tolerant classification of binary data by employing a hybrid modeling approach that leverages solutions of defined max-cut problems. In a comparative analysis conducted on synthetically generated binary datasets, NoiseCut exhibits better overfitting prevention compared to the early stopping technique employed by different supervised machine learning algorithms. The noise tolerance of NoiseCut stems from a dropout strategy that leverages prior knowledge of input features and is further enhanced by the integration of max-cut problems into the learning process. CONCLUSIONS: NoiseCut is a Python package for the implementation of hybrid modeling for the classification of binary data. It facilitates the integration of mechanistic knowledge on the input features into learning from data in a structured manner and proves to be a valuable classification tool when the available training data is noisy and/or limited in size. This advantage is especially prominent in medical and biomedical applications where data scarcity and noise are common challenges. The codebase, illustrations, and documentation for NoiseCut are accessible for download at https://pypi.org/project/noisecut/ . The implementation detailed in this paper corresponds to the version 0.2.1 release of the software.


Assuntos
Algoritmos , Software , Humanos , Aprendizado de Máquina Supervisionado , Aprendizado de Máquina
8.
Acad Radiol ; 31(9): 3590-3596, 2024 Sep.
Artigo em Inglês | MEDLINE | ID: mdl-38614825

RESUMO

RATIONALE AND OBJECTIVES: This study demonstrates a method for quantifying the impact of overfitting on the receiving operator characteristic curve (AUC) when using standard analysis pipelines to develop imaging biomarkers. We illustrate the approach using two publicly available repositories of radiology and pathology images for breast cancer diagnosis. MATERIALS AND METHODS: For each dataset, we permuted the outcome (cancer diagnosis) values to eliminate any true association between imaging features and outcome. Seven types of classification models (logistic regression, linear discriminant analysis, Naïve Bayes, linear support vector machines, nonlinear support vector machine, random forest, and multi-layer perceptron) were fitted to each scrambled dataset and evaluated by each of four techniques (all data, hold-out, 10-fold cross-validation, and bootstrapping). After repeating this process for a total of 50 outcome permutations, we averaged the resulting AUCs. Any increase over a null AUC of 0.5 can be attributed to overfitting. RESULTS: Applying this approach and varying sample size and the number of imaging features, we found that failing to control for overfitting could result in near-perfect prediction (AUC near 1.0). Cross-validation offered greater protection against overfitting than the other evaluation techniques, and for most classification algorithms a sample size of at least 200 was required to assess as few as 10 features with less than 0.05 AUC inflation attributable to overfitting. CONCLUSION: This approach could be applied to any curated dataset to suggest the number of features and analysis approaches to limit overfitting.


Assuntos
Neoplasias da Mama , Humanos , Neoplasias da Mama/diagnóstico por imagem , Feminino , Algoritmos , Área Sob a Curva , Interpretação de Imagem Assistida por Computador/métodos
9.
Neural Netw ; 174: 106215, 2024 Jun.
Artigo em Inglês | MEDLINE | ID: mdl-38471261

RESUMO

Deep neural networks tend to suffer from the overfitting issue when the training data are not enough. In this paper, we introduce two metrics from the intra-class distribution of correct-predicted and incorrect-predicted samples to provide a new perspective on the overfitting issue. Based on it, we propose a knowledge distillation approach without pretraining a teacher model in advance named Tolerant Self-Distillation (TSD) for alleviating the overfitting issue. It introduces an online updating memory and selectively stores the class predictions of the samples from the past iterations, making it possible to distill knowledge across the iterations. Specifically, the class predictions stored in the memory bank serve as the soft labels for supervising the samples from the same class for the current iteration in a reverse way, i.e. the correct-predicted samples are supervised with the incorrect predictions while the incorrect-predicted samples are supervised with the correct predictions. Consequently, the premature convergence issue caused by the over-confident samples would be mitigated, which helps the model to converge to a better local optimum. Extensive experimental results on several image classification benchmarks, including small-scale, large-scale, and fine-grained datasets, demonstrate the superiority of the proposed TSD.


Assuntos
Benchmarking , Conhecimento , Redes Neurais de Computação
10.
J Sci Food Agric ; 104(10): 6018-6034, 2024 Aug 15.
Artigo em Inglês | MEDLINE | ID: mdl-38483173

RESUMO

BACKGROUND: The accurate recognition and early warning for plant diseases and pests are a prerequisite of intelligent prevention and control for plant diseases and pests. As a result of the phenotype similarity of the hazarded plant after plant diseases and pests occur, as well as the interference of the external environment, traditional deep learning models often face the overfitting problem in phenotype recognition of plant diseases and pests, which leads to not only the slow convergence speed of the network, but also low recognition accuracy. RESULTS: Motivated by the above problems, the present study proposes a deep learning model EResNet-support vector machine (SVM) to alleviate the overfitting for the recognition and classification of plant diseases and pests. First, the feature extraction capability of the model is improved by increasing feature extraction layers in the convolutional neural network. Second, the order-reduced modules are embedded and a sparsely activated function is introduced to reduce model complexity and alleviate overfitting. Finally, a classifier fused by SVM and fully connected layers are introduced to transforms the original non-linear classification problem into a linear classification problem in high-dimensional space to further alleviate the overfitting and improve the recognition accuracy of plant diseases and pests. The ablation experiments further demonstrate that the fused structure can effectively alleviate the overfitting and improve the recognition accuracy. The experimental recognition results for typical plant diseases and pests show that the proposed EResNet-SVM model has 99.30% test accuracy for eight conditions (seven plant diseases and one normal), which is 5.90% higher than the original ResNet18. Compared with the classic AlexNet, GoogLeNet, Xception, SqueezeNet and DenseNet201 models, the accuracy of the EResNet-SVM model has improved by 5.10%, 7%, 8.10%, 6.20% and 1.90%, respectively. The testing accuracy of the EResNet-SVM model for 6 insect pests is 100%, which is 3.90% higher than that of the original ResNet18 model. CONCLUSION: This research provides not only useful references for alleviating the overfitting problem in deep learning, but also a theoretical and technical support for the intelligent detection and control of plant diseases and pests. © 2024 Society of Chemical Industry.


Assuntos
Aprendizado Profundo , Redes Neurais de Computação , Doenças das Plantas , Máquina de Vetores de Suporte , Doenças das Plantas/parasitologia , Doenças das Plantas/prevenção & controle , Animais , Insetos , Controle de Pragas/métodos
11.
Neural Netw ; 169: 307-324, 2024 Jan.
Artigo em Inglês | MEDLINE | ID: mdl-37922714

RESUMO

Large deep learning models are impressive, but they struggle when real-time data is not available. Few-shot class-incremental learning (FSCIL) poses a significant challenge for deep neural networks to learn new tasks from just a few labeled samples without forgetting the previously learned ones. This setup can easily leads to catastrophic forgetting and overfitting problems, severely affecting model performance. Studying FSCIL helps overcome deep learning model limitations on data volume and acquisition time, while improving practicality and adaptability of machine learning models. This paper provides a comprehensive survey on FSCIL. Unlike previous surveys, we aim to synthesize few-shot learning and incremental learning, focusing on introducing FSCIL from two perspectives, while reviewing over 30 theoretical research studies and more than 20 applied research studies. From the theoretical perspective, we provide a novel categorization approach that divides the field into five subcategories, including traditional machine learning methods, meta learning-based methods, feature and feature space-based methods, replay-based methods, and dynamic network structure-based methods. We also evaluate the performance of recent theoretical research on benchmark datasets of FSCIL. From the application perspective, FSCIL has achieved impressive achievements in various fields of computer vision such as image classification, object detection, and image segmentation, as well as in natural language processing and graph. We summarize the important applications. Finally, we point out potential future research directions, including applications, problem setups, and theory development. Overall, this paper offers a comprehensive analysis of the latest advances in FSCIL from a methodological, performance, and application perspective.


Assuntos
Aprendizado de Máquina , Redes Neurais de Computação , Inquéritos e Questionários , Tempo
12.
Sensors (Basel) ; 23(23)2023 Nov 27.
Artigo em Inglês | MEDLINE | ID: mdl-38067822

RESUMO

For a fiber optic gyroscope, thermal deformation of the fiber coil can introduce additional thermal-induced phase errors, commonly referred to as thermal errors. Implementing effective thermal error compensation techniques is crucial to addressing this issue. These techniques operate based on the real-time sensing of thermal errors and subsequent correction within the output signal. Given the challenge of directly isolating thermal errors from the gyroscope's output signal, predicting thermal errors based on temperature becomes necessary. To establish a mathematical model correlating the temperature and thermal errors, this study measured synchronized data of phase errors and angular velocity for the fiber coil under various temperature conditions, aiming to model it using data-driven methods. However, due to the difficulty of conducting tests and the limited number of data samples, direct engagement in data-driven modeling poses a risk of severe overfitting. To overcome this challenge, we propose a modeling algorithm that effectively integrates theoretical models with data, referred to as the TD-model in this paper. Initially, a theoretical analysis of the phase errors caused by thermal deformation of the fiber coil is performed. Subsequently, critical parameters, such as the thermal expansion coefficient, are determined, leading to the establishment of a theoretical model. Finally, the theoretical analysis model is incorporated as a regularization term and combined with the test data to jointly participate in the regression of model coefficients. Through experimental comparative analysis, it is shown that, relative to ordinary regression models, the TD-model effectively mitigates overfitting caused by the limited number of samples, resulting in a substantial 58% improvement in predictive accuracy.

13.
Curr Stem Cell Rep ; 9(4): 67-76, 2023.
Artigo em Inglês | MEDLINE | ID: mdl-38145009

RESUMO

Purpose of Review: This article gives a broad overview of quantitative modelling approaches in biology and provides guidance on how to employ them to boost stem cell research, by helping to answer biological questions and to predict the outcome of biological processes. Recent Findings: The twenty-first century has seen a steady increase in the proportion of cell biology publications employing mathematical modelling to aid experimental research. However, quantitative modelling is often used as a rather decorative element to confirm experimental findings, an approach which often yields only marginal added value, and is in many cases scientifically questionable. Summary: Quantitative modelling can boost biological research in manifold ways, but one has to take some careful considerations before embarking on a modelling campaign, in order to maximise its added value, to avoid pitfalls that may lead to wrong results, and to be aware of its fundamental limitations, imposed by the risks of over-fitting and "universality".

14.
Metabolomics ; 20(1): 8, 2023 Dec 21.
Artigo em Inglês | MEDLINE | ID: mdl-38127222

RESUMO

INTRODUCTION: In general, two characteristics are ever present in NMR-based metabolomics studies: (1) they are assays aiming to classify the samples in different groups, and (2) the number of samples is smaller than the feature (chemical shift) number. It is also common to observe imbalanced datasets due to the sampling method and/or inclusion criteria. These situations can cause overfitting. However, appropriate feature selection and classification methods can be useful to solve this issue. OBJECTIVES: Investigate the performance of metabolomics models built from the association between feature selectors, the absence of feature selection, and classification algorithms, as well as use the best performance model as an NMR-based metabolomic method for prostate cancer diagnosis. METHODS: We evaluated the performance of NMR-based metabolomics models for prostate cancer diagnosis using seven feature selectors and five classification formalisms. We also obtained metabolomics models without feature selection. In this study, thirty-eight volunteers with a positive diagnosis of prostate cancer and twenty-three healthy volunteers were enrolled. RESULTS: Thirty-eight models obtained were evaluated using AUROC, accuracy, sensitivity, specificity, and kappa's index values. The best result was obtained when Genetic Algorithm was used with Linear Discriminant Analysis with 0.92 sensitivity, 0.83 specificity, and 0.88 accuracy. CONCLUSION: The results show that the pick of a proper feature selection method and classification model, and a resampling method can avoid overfitting in a small metabolomic dataset. Furthermore, this approach would decrease the number of biopsies and optimize patient follow-up. 1H NMR-based metabolomics promises to be a non-invasive tool in prostate cancer diagnosis.


Assuntos
Quimiometria , Neoplasias da Próstata , Masculino , Humanos , Metabolômica , Neoplasias da Próstata/diagnóstico , Imageamento por Ressonância Magnética , Algoritmos
15.
Acta Crystallogr D Struct Biol ; 79(Pt 12): 1071-1078, 2023 Dec 01.
Artigo em Inglês | MEDLINE | ID: mdl-37921807

RESUMO

Model building and refinement, and the validation of their correctness, are very effective and reliable at local resolutions better than about 2.5 Šfor both crystallography and cryo-EM. However, at local resolutions worse than 2.5 Šboth the procedures and their validation break down and do not ensure reliably correct models. This is because in the broad density at lower resolution, critical features such as protein backbone carbonyl O atoms are not just less accurate but are not seen at all, and so peptide orientations are frequently wrongly fitted by 90-180°. This puts both backbone and side chains into the wrong local energy minimum, and they are then worsened rather than improved by further refinement into a valid but incorrect rotamer or Ramachandran region. On the positive side, new tools are being developed to locate this type of pernicious error in PDB depositions, such as CaBLAM, EMRinger, Pperp diagnosis of ribose puckers, and peptide flips in PDB-REDO, while interactive modeling in Coot or ISOLDE can help to fix many of them. Another positive trend is that artificial intelligence predictions such as those made by AlphaFold2 contribute additional evidence from large multiple sequence alignments, and in high-confidence parts they provide quite good starting models for loops, termini or whole domains with otherwise ambiguous density.


Assuntos
Inteligência Artificial , Proteínas , Modelos Moleculares , Proteínas/química , Cristalografia por Raios X , Peptídeos , Microscopia Crioeletrônica/métodos , Conformação Proteica
16.
Pharmaceutics ; 15(11)2023 Nov 16.
Artigo em Inglês | MEDLINE | ID: mdl-38004608

RESUMO

Prospectively planned designs of experiments (DoEs) offer a valuable approach to preventing collinearity issues that can result in statistical confusion, leading to misinterpretation and reducing the predictability of statistical models. However, it is also possible to develop models using historical data, provided that certain guidelines are followed to enhance and ensure proper statistical modeling. This article presents a methodology for constructing a design space using process data, while avoiding the common pitfalls associated with retrospective data analysis. For this study, data from a real wet granulation process were collected to pragmatically illustrate all the concepts and methods developed in this article.

17.
Heliyon ; 9(10): e20382, 2023 Oct.
Artigo em Inglês | MEDLINE | ID: mdl-37780751

RESUMO

Currently, social networks have become the main source to acquire news about current global affairs. However, fake news appears and spreads on social media daily. This disinformation has a negative influence on several domains, such as politics, the economy, and health. In addition, it further generates detriments to societal stability. Several studies have provided effective models for detecting fake news in social networks through a variety of methods; however, there are limitations. Furthermore, since it is a critical field, the accuracy of the detection models was found to be notably insufficient. Although many review articles have addressed the repercussions of fake news, most have focused on specific and recurring aspects of fake news detection models. For example, the majority of reviews have primarily focused on dividing datasets, features, and classifiers used in this field by type. The limitations of the datasets, their features, how these features are fused, and the impact of all these factors on detection models were not investigated, especially since most detection models were based on a supervised learning approach. This review article analyzes relevant studies for the few last years and highlights the challenges faced by fake news detection models and their impact on their performance. The investigation of fake news detection studies relied on the following aspects and their impact on detection accuracy, namely datasets, overfitting/underfitting, image-based features, feature vector representation, machine learning models, and data fusion. Based on the analysis of relevant studies, the review showed that these issues significantly affect the performance and accuracy of detection models. This review aims to provide room for other researchers in the future to improve fake news detection models.

18.
Sensors (Basel) ; 23(17)2023 Aug 23.
Artigo em Inglês | MEDLINE | ID: mdl-37687802

RESUMO

Temperature sensors are widely used in industrial production and scientific research, and accurate temperature measurement is crucial for ensuring the quality and safety of production processes. To improve the accuracy and stability of temperature sensors, this paper proposed using an artificial neural network (ANN) model for calibration and explored the feasibility and effectiveness of using ANNs to calibrate temperature sensors. The experiment collected multiple sets of temperature data from standard temperature sensors in different environments and compared the calibration results of the ANN model, linear regression, and polynomial regression. The experimental results show that calibration using the ANN improved the accuracy of the temperature sensors. Compared with traditional linear regression and polynomial regression, the ANN model produced more accurate calibration. However, overfitting may occur due to a small sample size or a large amount of noise. Therefore, the key to improving calibration using the ANN model is to design reasonable training samples and adjust the model parameters. The results of this study are important for practical applications and provide reliable technical support for industrial production and scientific research.

19.
Hum Vaccin Immunother ; 19(2): 2251830, 2023 08 01.
Artigo em Inglês | MEDLINE | ID: mdl-37697867

RESUMO

Overfitting describes the phenomenon where a highly predictive model on the training data generalizes poorly to future observations. It is a common concern when applying machine learning techniques to contemporary medical applications, such as predicting vaccination response and disease status in infectious disease or cancer studies. This review examines the causes of overfitting and offers strategies to counteract it, focusing on model complexity reduction, reliable model evaluation, and harnessing data diversity. Through discussion of the underlying mathematical models and illustrative examples using both synthetic data and published real datasets, our objective is to equip analysts and bioinformaticians with the knowledge and tools necessary to detect and mitigate overfitting in their research.


Assuntos
Aprendizado de Máquina , Vacinação
20.
MethodsX ; 11: 102289, 2023 Dec.
Artigo em Inglês | MEDLINE | ID: mdl-37560402

RESUMO

Some statistical analysis techniques may require complete data matrices, but a frequent problem in the construction of databases is the incomplete collection of information for different reasons. One option to tackle the problem is to estimate and impute the missing data. This paper describes a form of imputation that mixes regression with lower rank approximations. To improve the quality of the imputations, a generalisation is proposed that replaces the singular value decomposition (SVD) of the matrix with a regularised SVD in which the regularisation parameter is estimated by cross-validation. To evaluate the performance of the proposal, ten sets of real data from multienvironment trials were used. Missing values were created in each set at four percentages of missing not at random, and three criteria were then considered to investigate the effectiveness of the proposal. The results show that the regularised method proves very competitive when compared to the original method, beating it in several of the considered scenarios. As it is a very general system, its application can be extended to all multivariate data matrices. •The imputation method is modified through the inclusion of a stable and efficient computational algorithm that replaces the classical SVD least squares criterion by a penalised criterion. This penalty produces smoothed eigenvectors and eigenvalues that avoid overfitting problems, improving the performance of the method when the penalty is necessary. The size of the penalty can be determined by minimising one of the following criteria: the prediction errors, the Procrustes similarity statistic or the critical angles between subspaces of principal components.

SELEÇÃO DE REFERÊNCIAS
DETALHE DA PESQUISA