Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Results 1 - 20 de 43
Filter
1.
Chemosphere ; 357: 141833, 2024 Jun.
Article in English | MEDLINE | ID: mdl-38579944

ABSTRACT

Experimental water research lacks clear methodology to estimate experimental error. Especially when natural waters are involved, the characterization tools bear method-specific artifacts while the varying environmental conditions prevent regular repeats. This tutorial review identifies common mistakes, and proposes a practical procedure to determine experimental errors at the example of membrane filtration. Statistical analysis is often applied to an insufficient number of repeated measurements, while not all error sources and contributions are considered. This results in an underestimation of the experimental error. Variations in relevant experimental parameters need to be investigated systematically, and the related errors are quantified as a half of the variation between the max and min values when standard deviation is not applicable. Error of calculated parameters (e.g. flux, pollutant removal and mass loss) is estimated by applying error propagation, where weighing contributions of the experimental parameters are considered. Appropriate judgment and five-time repetition of a selected experiment under identical conditions are proposed to validate the propagated experimental error. For validation, the five repeated data points should lie within the estimated error range of the error bar. The proposed error evaluation procedure is adaptable in experimental water research and intended for researchers to identify the contributing factors of an experimental error and carry out appropriate error quantification and validation. The most important aim is to raise awareness of the necessity to question error methodology and reproducibility of experimental data, to produce and publish high quality research.


Subject(s)
Filtration , Membranes, Artificial , Filtration/methods , Water Purification/methods , Water/chemistry , Reproducibility of Results , Research Design , Scientific Experimental Error/statistics & numerical data
3.
Life Sci Alliance ; 5(4)2022 04.
Article in English | MEDLINE | ID: mdl-35022248

ABSTRACT

Nucleotide sequence reagents underpin molecular techniques that have been applied across hundreds of thousands of publications. We have previously reported wrongly identified nucleotide sequence reagents in human research publications and described a semi-automated screening tool Seek & Blastn to fact-check their claimed status. We applied Seek & Blastn to screen >11,700 publications across five literature corpora, including all original publications in Gene from 2007 to 2018 and all original open-access publications in Oncology Reports from 2014 to 2018. After manually checking Seek & Blastn outputs for >3,400 human research articles, we identified 712 articles across 78 journals that described at least one wrongly identified nucleotide sequence. Verifying the claimed identities of >13,700 sequences highlighted 1,535 wrongly identified sequences, most of which were claimed targeting reagents for the analysis of 365 human protein-coding genes and 120 non-coding RNAs. The 712 problematic articles have received >17,000 citations, including citations by human clinical trials. Given our estimate that approximately one-quarter of problematic articles may misinform the future development of human therapies, urgent measures are required to address unreliable gene research articles.


Subject(s)
Base Sequence/genetics , Genetic Research , Genome, Human/genetics , Publications/statistics & numerical data , Scientific Experimental Error/statistics & numerical data , Human Genetics/standards , Humans , Proteins/genetics
5.
Am J Epidemiol ; 190(9): 1830-1840, 2021 09 01.
Article in English | MEDLINE | ID: mdl-33517416

ABSTRACT

Although variables are often measured with error, the impact of measurement error on machine-learning predictions is seldom quantified. The purpose of this study was to assess the impact of measurement error on the performance of random-forest models and variable importance. First, we assessed the impact of misclassification (i.e., measurement error of categorical variables) of predictors on random-forest model performance (e.g., accuracy, sensitivity) and variable importance (mean decrease in accuracy) using data from the National Comorbidity Survey Replication (2001-2003). Second, we created simulated data sets in which we knew the true model performance and variable importance measures and could verify that quantitative bias analysis was recovering the truth in misclassified versions of the data sets. Our findings showed that measurement error in the data used to construct random forests can distort model performance and variable importance measures and that bias analysis can recover the correct results. This study highlights the utility of applying quantitative bias analysis in machine learning to quantify the impact of measurement error on study results.


Subject(s)
Bias , Scientific Experimental Error/statistics & numerical data , Computer Simulation , Datasets as Topic , Humans , Machine Learning/statistics & numerical data , Probability , Suicide, Attempted/statistics & numerical data
6.
Cancer Res ; 81(2): 282-288, 2021 01 15.
Article in English | MEDLINE | ID: mdl-33115802

ABSTRACT

Although next-generation sequencing is widely used in cancer to profile tumors and detect variants, most somatic variant callers used in these pipelines identify variants at the lowest possible granularity, single-nucleotide variants (SNV). As a result, multiple adjacent SNVs are called individually instead of as a multi-nucleotide variants (MNV). With this approach, the amino acid change from the individual SNV within a codon could be different from the amino acid change based on the MNV that results from combining SNV, leading to incorrect conclusions about the downstream effects of the variants. Here, we analyzed 10,383 variant call files (VCF) from the Cancer Genome Atlas (TCGA) and found 12,141 incorrectly annotated MNVs. Analysis of seven commonly mutated genes from 178 studies in cBioPortal revealed that MNVs were consistently missed in 20 of these studies, whereas they were correctly annotated in 15 more recent studies. At the BRAF V600 locus, the most common example of MNV, several public datasets reported separate BRAF V600E and BRAF V600M variants instead of a single merged V600K variant. VCFs from the TCGA Mutect2 caller were used to develop a solution to merge SNV to MNV. Our custom script used the phasing information from the SNV VCF and determined whether SNVs were at the same codon and needed to be merged into MNV before variant annotation. This study shows that institutions performing NGS sequencing for cancer genomics should incorporate the step of merging MNV as a best practice in their pipelines. SIGNIFICANCE: Identification of incorrect mutation calls in TCGA, including clinically relevant BRAF V600 and KRAS G12, will influence research and potentially clinical decisions.


Subject(s)
Genome, Human , Genomics/standards , Molecular Sequence Annotation/standards , Mutation , Neoplasms/genetics , Polymorphism, Single Nucleotide , Scientific Experimental Error/statistics & numerical data , Algorithms , High-Throughput Nucleotide Sequencing/methods , Humans , Neoplasms/pathology
7.
Clin Exp Dent Res ; 6(4): 383-390, 2020 08.
Article in English | MEDLINE | ID: mdl-32233020

ABSTRACT

OBJECTIVES: The present systematic review aimed to perform an in-depth analysis of the different features of retracted publications in the dental field. MATERIAL AND METHODS: This review has been recorded in the PROSPERO database (CRD42017075634). Two independent reviewers performed an electronic search (Pubmed, Retraction Watch) for retracted articles in dental literature up to December 31, 2018. RESULTS: 180 retracted papers were identified, the first published in 2001. Retractions increased by 47% in the last four-year period (2014-2018), when compared with 2009-2013 (94 and 64 retracted publications, respectively). Author misconduct was the most common reason for retraction (65.0%), followed by honest scientific errors (12.2%) and publisher-related issues (10.6%). The majority of retracted research was conducted in Asia (55.6%), with 49 papers written in India (27.2%). 552 researchers (89%) are listed as authors in only one retracted article, while 10 researchers (1.6%) are present in five or more retracted publications. Retracted articles were cited 530 times after retraction: the great majority of these citations (89.6%) did not consider the existence of the retraction notice and treated data from retracted articles as reliable. CONCLUSIONS: Retractions in dental literature have constantly increased in recent years, with the majority of them due to misconduct and fraud. The publication of unreliable research has many negative consequences. Studies derived from such material are designed on potentially incorrect bases, waste funds and resources, and most importantly, increase risk of incorrect treatment for patients. Citation of retracted papers represents a major issue for the scientific community.


Subject(s)
Biomedical Research/standards , Dentistry/standards , Fraud/statistics & numerical data , Periodicals as Topic/statistics & numerical data , Scientific Experimental Error/statistics & numerical data , Scientific Misconduct/statistics & numerical data , Databases, Factual , Humans , Periodicals as Topic/standards , Retraction of Publication as Topic
8.
J. optom. (Internet) ; 13(1): 3-14, ene.-mar. 2020. ilus, tab
Article in English | IBECS | ID: ibc-195303

ABSTRACT

Measurement of the amplitude of accommodation is established as a procedure in a routine optometric eye examination. However, clinical methods of measurement of this basic optical function have several sources of error. They are numerous and diverse, and include depth of focus, reaction time, instrument design, specification of the measurement end-point, specification of the reference point of measurement, measurement conditions, consideration of refractive error, and psychological factors. Several of these sources of inaccuracy are composed of multiple sub-sources, and many of the sub-sources influence the common methods of measurement of amplitude of accommodation. Consideration of these sources of measurement error casts doubt on the reliability of the results of measurement, on the validity of established normative values that have been produced using these methods, and on the value of reports of the results of surgery designed to restore accommodation. Clinicians can reduce the effects of some of the sources of error by modifying techniques of measurement with existing methods, but a new method may further improve accuracy


La medición de la amplitud de acomodación se ha establecido como un procedimiento del examen optométrico ocular rutinario. Sin embargo, los métodos clínicos de medición de esta función óptica básica tienen diversas fuentes de error. Estas son numerosas y diversas, e incluyen profundidad de foco, tiempo de reacción, diseño del instrumento, especificación del punto final de la medición, especificación del punto de referencia de la medición, condiciones de la medición, consideración del error refractivo, y factores psicológicos. Algunas de estas fuentes de imprecisión se componen de múltiples sub-fuentes, muchas de las cuales influyen en los métodos comunes de medición de la amplitud de acomodación. La consideración de estas fuentes de error en la medición plantea dudas sobre la fiabilidad de los resultados de dicha medición, la validez de los valores normativos establecidos que se han producido utilizando estos métodos, y el valor de los informes sobre resultados de la cirugía diseñada para restablecer la acomodación. Los clínicos pueden reducir los efectos de algunas de las fuentes de error, modificando las técnicas de medición con ayuda de los métodos existentes, aunque el desarrollo de un nuevo método podría mejorar la precisión


Subject(s)
Humans , Accommodation, Ocular/physiology , Scientific Experimental Error/statistics & numerical data , Vision Tests/standards , Models, Statistical , Reproducibility of Results , Retinoscopy
9.
Multivariate Behav Res ; 55(6): 926-940, 2020.
Article in English | MEDLINE | ID: mdl-31795755

ABSTRACT

Researchers detecting heterogeneity of regression in a treatment outcome study including a covariate and random assignment to groups often want to investigate the simple treatment effect at the sample grand mean of the covariate and at points one standard deviation above and below that mean. The estimated variances of the simple treatment effect that have traditionally been used in such tests were derived under the assumption that the covariate values were fixed constants. We derive results appropriate for a two-group experiment that instead presume the covariate is a normally distributed random variable. A simulation study is used to confirm the validity of the analytical results and to compare error estimates and confidence intervals based on these results with those based on assuming a fixed covariate. Discrepancies between estimates for fixed and random covariates of the variability of treatment effects can be substantial. However, in situations where the extent of heterogeneity of regression is like that typically reported, presuming the covariate is random rather than fixed will generally result in only a modest increase in estimated standard errors, and in some circumstances can even result in a smaller estimated standard error. We illustrate the new methods with an empirical data set.


Subject(s)
Confidence Intervals , Scientific Experimental Error/statistics & numerical data , Statistics as Topic/methods , Algorithms , Analysis of Variance , Computer Simulation , Humans , Models, Statistical , Random Allocation , Regression Analysis , Research Design , Treatment Outcome
10.
J Optom ; 13(1): 3-14, 2020.
Article in English | MEDLINE | ID: mdl-31303551

ABSTRACT

Measurement of the amplitude of accommodation is established as a procedure in a routine optometric eye examination. However, clinical methods of measurement of this basic optical function have several sources of error. They are numerous and diverse, and include depth of focus, reaction time, instrument design, specification of the measurement end-point, specification of the reference point of measurement, measurement conditions, consideration of refractive error, and psychological factors. Several of these sources of inaccuracy are composed of multiple sub-sources, and many of the sub-sources influence the common methods of measurement of amplitude of accommodation. Consideration of these sources of measurement error casts doubt on the reliability of the results of measurement, on the validity of established normative values that have been produced using these methods, and on the value of reports of the results of surgery designed to restore accommodation. Clinicians can reduce the effects of some of the sources of error by modifying techniques of measurement with existing methods, but a new method may further improve accuracy.


Subject(s)
Accommodation, Ocular/physiology , Scientific Experimental Error/statistics & numerical data , Vision Tests/standards , Humans , Models, Statistical , Reproducibility of Results , Retinoscopy
11.
J Mass Spectrom ; 55(7): e4464, 2020 Jul.
Article in English | MEDLINE | ID: mdl-31697861

ABSTRACT

HPLC-MS/MS analysis of various human cell lines shows the presence of a major amount of bovine protein contaminants. These likely originate from fetal bovine serum (FBS), typically used in cell cultures. If evaluated against a human protein database, on average 10% of the identified human proteins will be misleading (bovine proteins, but indicated as if they were human). Bovine contaminants therefore may cause major bias in proteomic studies of cell cultures, if not considered explicitly.


Subject(s)
Cell Line/chemistry , Culture Media/chemistry , Proteins/analysis , Serum Albumin, Bovine/chemistry , Animals , Cattle , Cell Culture Techniques , Drug Contamination , HeLa Cells , Humans , Proteomics , Scientific Experimental Error/statistics & numerical data , Tandem Mass Spectrometry
12.
Medwave ; 19(7): e7687, 2019 Aug 27.
Article in Spanish, English | MEDLINE | ID: mdl-31584929

ABSTRACT

Biomedical research, particularly when it involves human beings, is always subjected to sources of error that must be recognized. Systematic error or bias is associated with problems in the methodological design or during the execu-tion phase of a research project. It affects its validity and is qualitatively ap-praised. On the other hand, random error is related to variations due to chance. It may be quantitatively expressed, but never removed. This review is the first of a methodological series on general concepts in biostatistics and clin-ical epidemiology developed by the Chair of Scientific Research Methodology at the School of Medicine, University of Valparaíso, Chile. In this article, we address the theoretical concepts of error, its evaluation, and control. Finally, we discuss some current controversies in its conceptualization that are relevant to undergraduate and graduate students of health sciences.


La investigación biomédica, particularmente la que involucra a seres humanos, está siempre sometida a fuentes de error que deben ser reconocidas. El error sistemático o sesgo, se asocia con debilidades en el diseño metodológico o de la fase de ejecución del estudio. Éste afecta su validez y se valora cualitativamente. Por su parte, el error aleatorio se relaciona con las variaciones producidas por el azar, el cual puede expresarse cuantitativamente, pero nunca eliminarse. Esta revisión es la primera entrega de una serie metodológica sobre conceptos generales en bioestadística y epidemiología clínica desarrollada por la Cátedra de Metodología de la Investigación Científica de la Universidad de Valparaíso, Chile. En este artículo se abordan los conceptos teóricos asociados al error, su evaluación y control. Finalmente, se discuten algunas controversias actuales en cuanto a su conceptualización, de relevancia para estudiantes de pre y posgrado de ciencias de la salud.


Subject(s)
Biomedical Research/statistics & numerical data , Biostatistics/methods , Epidemiology/statistics & numerical data , Bias , Humans , Research Design , Scientific Experimental Error/statistics & numerical data
13.
Stat Med ; 38(27): 5182-5196, 2019 11 30.
Article in English | MEDLINE | ID: mdl-31478240

ABSTRACT

In randomised trials, continuous endpoints are often measured with some degree of error. This study explores the impact of ignoring measurement error and proposes methods to improve statistical inference in the presence of measurement error. Three main types of measurement error in continuous endpoints are considered: classical, systematic, and differential. For each measurement error type, a corrected effect estimator is proposed. The corrected estimators and several methods for confidence interval estimation are tested in a simulation study. These methods combine information about error-prone and error-free measurements of the endpoint in individuals not included in the trial (external calibration sample). We show that, if measurement error in continuous endpoints is ignored, the treatment effect estimator is unbiased when measurement error is classical, while Type-II error is increased at a given sample size. Conversely, the estimator can be substantially biased when measurement error is systematic or differential. In those cases, bias can largely be prevented and inferences improved upon using information from an external calibration sample, of which the required sample size increases as the strength of the association between the error-prone and error-free endpoint decreases. Measurement error correction using already a small (external) calibration sample is shown to improve inferences and should be considered in trials with error-prone endpoints. Implementation of the proposed correction methods is accommodated by a new software package for R.


Subject(s)
Endpoint Determination , Randomized Controlled Trials as Topic/methods , Scientific Experimental Error , Computer Simulation , Data Interpretation, Statistical , Endpoint Determination/methods , Endpoint Determination/statistics & numerical data , Hemoglobins/analysis , Humans , Randomized Controlled Trials as Topic/standards , Sample Size , Scientific Experimental Error/statistics & numerical data
17.
Biometrics ; 75(4): 1334-1344, 2019 12.
Article in English | MEDLINE | ID: mdl-31290137

ABSTRACT

It is known that the one-sided Simes' test controls the error rate if the underlying distribution is multivariate totally positive of order 2 (MTP2), but not in general. The two-sided test also controls the error rate when the coordinate absolute value has an MTP2 distribution, which holds more generally. We prove mathematically that when the coordinate absolute value controls the error rate at level 2α, then certain kinds of truncated Simes' tests also control the one-sided error rate at level α. We also compare the closure of the truncated tests with the Holms, Hochberg, and Hommel procedures in many scenarios when the test statistics are multivariate normal.


Subject(s)
Data Interpretation, Statistical , Multivariate Analysis , Statistical Distributions , Biometry , Confidence Intervals , Humans , Scientific Experimental Error/statistics & numerical data
18.
Methods ; 169: 57-68, 2019 10 01.
Article in English | MEDLINE | ID: mdl-31302177

ABSTRACT

Tethered particle motion experiments are versatile single-molecule techniques enabling one to address in vitro the molecular properties of DNA and its interactions with various partners involved in genetic regulations. These techniques provide raw data such as the tracked particle amplitude of movement, from which relevant information about DNA conformations or states must be recovered. Solving this inverse problem appeals to specific theoretical tools that have been designed in the two last decades, together with the data pre-processing procedures that ought to be implemented to avoid biases inherent to these experimental techniques. These statistical tools and models are reviewed in this paper.


Subject(s)
DNA/chemistry , Models, Statistical , Single Molecule Imaging/methods , Markov Chains , Molecular Dynamics Simulation , Motion , Nucleic Acid Conformation , Physics , Scientific Experimental Error/statistics & numerical data
19.
PLoS One ; 14(5): e0216118, 2019.
Article in English | MEDLINE | ID: mdl-31042766

ABSTRACT

The qPCR method provides an inexpensive, rapid method for estimating relative telomere length across a set of biological samples. Like all laboratory methods, it involves some degree of measurement error. The estimation of relative telomere length is done subjecting the actual measurements made (the Cq values for telomere and a control gene) to non-linear transformations and combining them into a ratio (the TS ratio). Here, we use computer simulations, supported by mathematical analysis, to explore how errors in measurement affect qPCR estimates of relative telomere length, both in cross-sectional and longitudinal data. We show that errors introduced at the level of Cq values are magnified when the TS ratio is calculated. If the errors at the Cq level are normally distributed and independent of true telomere length, those in the TS ratio are positively skewed and proportional to true telomere length. The repeatability of the TS ratio declines sharply with increasing error in measurement of the Cq values for telomere and/or control gene. In simulated longitudinal data, measurement error alone can produce a pattern of low correlation between successive measures of relative telomere length, coupled with a strong negative dependency of the rate of change on initial relative telomere length. Our results illustrate the importance of reducing measurement error: a small increase in error in Cq values can have large consequences for the power and interpretability of qPCR estimates of relative telomere length. The findings also illustrate the importance of characterising the measurement error in each dataset-coefficients of variation are generally unhelpful, and researchers should report standard deviations of Cq values and/or repeatabilities of TS ratios-and allowing for the known effects of measurement error when interpreting patterns of TS ratio change over time.


Subject(s)
Real-Time Polymerase Chain Reaction/methods , Scientific Experimental Error/statistics & numerical data , Computer Simulation/statistics & numerical data , Cross-Sectional Studies , Humans , Telomere/genetics , Telomere Homeostasis/physiology
20.
Chemosphere ; 230: 67-75, 2019 Sep.
Article in English | MEDLINE | ID: mdl-31102873

ABSTRACT

Quinones are becoming an essential tool for refractory organics treatment, while their quantification may be not well-considered. In this paper, two kinds of potential errors in quantification were evaluated in multiple pH conditions. They were derived from the coexistence of oxidized/reduced quinone species (Type I) and pH-sensitive feature (Type II), respectively. These errors would remarkably influence the accuracy of quantification while they haven't been emphasized. Thus, to elaborate the relationship between the two types of errors and the absorbance or pH conditions, three typical quinones [Anthraquinone-1-sulfonate (α-AQS), anthraquinone-2,6-disulfonate (AQDS) and lawsone] were selected and their acid dissociation coefficients (pKa) as well as UV-Vis spectra were determined. Results revealed that, for Type I, the relative error (RE) of α-AQS concentration would exceed the limit (5%) when reduced α-AQS was below 48% of total α-AQS. Similar results were found for lawsone. However, the RE can be eliminated by the equation established in this paper. For Type II, the pH-sensitive feature was related to the pKa values of quinones. Absorbances of α-AQS and lawsone would change remarkably with pH variation. Therefore, a model for correction was established. Analog data showed high consistency with experimental data [r = 0.995 (n = 25, p < 0.01) and r = 0.997 (n = 36, p < 0.01), for lawsone and α-AQS respectively]. Especially, the determination of AQDS concentrations was noticed to be pH-independent at 437 nm under pH 4.00 to 9.18 conditions. Based on these features, a comprehensive data solution was proposed for handling these errors.


Subject(s)
Anthraquinones/analysis , Naphthoquinones/analysis , Scientific Experimental Error/statistics & numerical data , Water Purification/methods , Calibration/standards , Hydrogen-Ion Concentration , Oxidation-Reduction , Quinones/analysis , Wastewater/chemistry
SELECTION OF CITATIONS
SEARCH DETAIL