RESUMEN
Cell clustering is typically the initial step in single-cell RNA sequencing (scRNA-seq) analyses. The performance of clustering considerably impacts the validity and reproducibility of cell identification. A variety of clustering algorithms have been developed for scRNA-seq data. These algorithms generate cell label sets that assign each cell to a cluster. However, different algorithms usually yield different label sets, which can introduce variations in cell-type identification based on the generated label sets. Currently, the performance of these algorithms has not been systematically evaluated in single-cell transcriptome studies. Herein, we performed a critical assessment of seven state-of-the-art clustering algorithms including four deep learning-based clustering algorithms and commonly used methods Seurat, Cosine-based Tanimoto similarity-refined graph for community detection using Leiden's algorithm (CosTaL) and Single-cell consensus clustering (SC3). We used diverse evaluation indices based on 10 different scRNA-seq benchmarks to systematically evaluate their clustering performance. Our results show that CosTaL, Seurat, Deep Embedding for Single-cell Clustering (DESC) and SC3 consistently outperformed Single-Cell Clustering Assessment Framework and scDeepCluster based on nine effectiveness scores. Notably, CosTaL and DESC demonstrated superior performance in clustering specific cell types. The performance of the single-cell Variational Inference tools varied across different datasets, suggesting its sensitivity to certain dataset characteristics. Notably, DESC exhibited promising results for cell subtype identification and capturing cellular heterogeneity. In addition, SC3 requires more memory and exhibits slower computation speed compared to other algorithms for the same dataset. In sum, this study provides useful guidance for selecting appropriate clustering methods in scRNA-seq data analysis.
Asunto(s)
Análisis de la Célula Individual , Transcriptoma , Análisis de Secuencia de ARN/métodos , Reproducibilidad de los Resultados , Análisis de la Célula Individual/métodos , Algoritmos , Análisis por Conglomerados , Perfilación de la Expresión Génica/métodosRESUMEN
Promoters are crucial regulatory DNA regions for gene transcriptional activation. Rapid advances in next-generation sequencing technologies have accelerated the accumulation of genome sequences, providing increased training data to inform computational approaches for both prokaryotic and eukaryotic promoter prediction. However, it remains a significant challenge to accurately identify species-specific promoter sequences using computational approaches. To advance computational support for promoter prediction, in this study, we curated 58 comprehensive, up-to-date, benchmark datasets for 7 different species (i.e. Escherichia coli, Bacillus subtilis, Homo sapiens, Mus musculus, Arabidopsis thaliana, Zea mays and Drosophila melanogaster) to assist the research community to assess the relative functionality of alternative approaches and support future research on both prokaryotic and eukaryotic promoters. We revisited 106 predictors published since 2000 for promoter identification (40 for prokaryotic promoter, 61 for eukaryotic promoter, and 5 for both). We systematically evaluated their training datasets, computational methodologies, calculated features, performance and software usability. On the basis of these benchmark datasets, we benchmarked 19 predictors with functioning webservers/local tools and assessed their prediction performance. We found that deep learning and traditional machine learning-based approaches generally outperformed scoring function-based approaches. Taken together, the curated benchmark dataset repository and the benchmarking analysis in this study serve to inform the design and implementation of computational approaches for promoter prediction and facilitate more rigorous comparison of new techniques in the future.
Asunto(s)
Drosophila melanogaster , Eucariontes , Animales , Biología Computacional/métodos , Drosophila melanogaster/genética , Células Eucariotas , Ratones , Células Procariotas , Regiones Promotoras GenéticasRESUMEN
Currently, there exist no generally accepted strategies of evaluating computational models for microRNA-disease associations (MDAs). Though K-fold cross validations and case studies seem to be must-have procedures, the value of K, the evaluation metrics, and the choice of query diseases as well as the inclusion of other procedures (such as parameter sensitivity tests, ablation studies and computational cost reports) are all determined on a case-by-case basis and depending on the researchers' choices. In the current review, we include a comprehensive analysis on how 29 state-of-the-art models for predicting MDAs were evaluated. Based on the analytical results, we recommend a feasible evaluation workflow that would suit any future model to facilitate fair and systematic assessment of predictive performance.
Asunto(s)
MicroARNs , MicroARNs/genética , Biología Computacional/métodos , Algoritmos , Simulación por ComputadorRESUMEN
Optimal methods could effectively improve the accuracy of predicting and identifying candidate driver genes. Various computational methods based on mutational frequency, network and function approaches have been developed to identify mutation driver genes in cancer genomes. However, a comprehensive evaluation of the performance levels of network-, function- and frequency-based methods is lacking. In the present study, we assessed and compared eight performance criteria for eight network-based, one function-based and three frequency-based algorithms using eight benchmark datasets. Under different conditions, the performance of approaches varied in terms of network, measurement and sample size. The frequency-based driverMAPS and network-based HotNet2 methods showed the best overall performance. Network-based algorithms using protein-protein interaction networks outperformed the function- and the frequency-based approaches. Precision, F1 score and Matthews correlation coefficient were low for most approaches. Thus, most of these algorithms require stringent cutoffs to correctly distinguish driver and non-driver genes. We constructed a website named Cancer Driver Catalog (http://159.226.67.237/sun/cancer_driver/), wherein we integrated the gene scores predicted by the foregoing software programs. This resource provides valuable guidance for cancer researchers and clinical oncologists prioritizing cancer driver gene candidates by using an optimal tool.
Asunto(s)
Neoplasias , Oncogenes , Algoritmos , Biología Computacional/métodos , Redes Reguladoras de Genes , Humanos , Mutación , Neoplasias/genética , Programas InformáticosRESUMEN
A critical challenge in genetic diagnostics is the assessment of genetic variants associated with diseases, specifically variants that fall out with canonical splice sites, by altering alternative splicing. Several computational methods have been developed to prioritize variants effect on splicing; however, performance evaluation of these methods is hampered by the lack of large-scale benchmark datasets. In this study, we employed a splicing-region-specific strategy to evaluate the performance of prediction methods based on eight independent datasets. Under most conditions, we found that dbscSNV-ADA performed better in the exonic region, S-CAP performed better in the core donor and acceptor regions, S-CAP and SpliceAI performed better in the extended acceptor region and MMSplice performed better in identifying variants that caused exon skipping. However, it should be noted that the performances of prediction methods varied widely under different datasets and splicing regions, and none of these methods showed the best overall performance with all datasets. To address this, we developed a new method, machine learning-based classification of splice sites variants (MLCsplice), to predict variants effect on splicing based on individual methods. We demonstrated that MLCsplice achieved stable and superior prediction performance compared with any individual method. To facilitate the identification of the splicing effect of variants, we provided precomputed MLCsplice scores for all possible splice sites variants across human protein-coding genes (http://39.105.51.3:8090/MLCsplice/). We believe that the performance of different individual methods under eight benchmark datasets will provide tentative guidance for appropriate method selection to prioritize candidate splice-disrupting variants, thereby increasing the genetic diagnostic yield.
Asunto(s)
Empalme Alternativo , Empalme del ARN , Biología Computacional/métodos , Exones , Humanos , Aprendizaje Automático , MutaciónRESUMEN
Predicting protein properties from amino acid sequences is an important problem in biology and pharmacology. Protein-protein interactions among SARS-CoV-2 spike protein, human receptors and antibodies are key determinants of the potency of this virus and its ability to evade the human immune response. As a rapidly evolving virus, SARS-CoV-2 has already developed into many variants with considerable variation in virulence among these variants. Utilizing the proteomic data of SARS-CoV-2 to predict its viral characteristics will, therefore, greatly aid in disease control and prevention. In this paper, we review and compare recent successful prediction methods based on long short-term memory (LSTM), transformer, convolutional neural network (CNN) and a similarity-based topological regression (TR) model and offer recommendations about appropriate predictive methodology depending on the similarity between training and test datasets. We compare the effectiveness of these models in predicting the binding affinity and expression of SARS-CoV-2 spike protein sequences. We also explore how effective these predictive methods are when trained on laboratory-created data and are tasked with predicting the binding affinity of the in-the-wild SARS-CoV-2 spike protein sequences obtained from the GISAID datasets. We observe that TR is a better method when the sample size is small and test protein sequences are sufficiently similar to the training sequence. However, when the training sample size is sufficiently large and prediction requires extrapolation, LSTM embedding and CNN-based predictive model show superior performance.
Asunto(s)
COVID-19 , SARS-CoV-2 , Secuencia de Aminoácidos , COVID-19/genética , Humanos , Unión Proteica , Proteómica , SARS-CoV-2/genética , Análisis de Secuencia de Proteína , Glicoproteína de la Espiga del Coronavirus/metabolismoRESUMEN
Changes in protein sequence can have dramatic effects on how proteins fold, their stability and dynamics. Over the last 20 years, pioneering methods have been developed to try to estimate the effects of missense mutations on protein stability, leveraging growing availability of protein 3D structures. These, however, have been developed and validated using experimentally derived structures and biophysical measurements. A large proportion of protein structures remain to be experimentally elucidated and, while many studies have based their conclusions on predictions made using homology models, there has been no systematic evaluation of the reliability of these tools in the absence of experimental structural data. We have, therefore, systematically investigated the performance and robustness of ten widely used structural methods when presented with homology models built using templates at a range of sequence identity levels (from 15% to 95%) and contrasted performance with sequence-based tools, as a baseline. We found there is indeed performance deterioration on homology models built using templates with sequence identity below 40%, where sequence-based tools might become preferable. This was most marked for mutations in solvent exposed residues and stabilizing mutations. As structure prediction tools improve, the reliability of these predictors is expected to follow, however we strongly suggest that these factors should be taken into consideration when interpreting results from structure-based predictors of mutation effects on protein stability.
Asunto(s)
Biología Computacional , Proteínas , Biología Computacional/métodos , Bases de Datos de Proteínas , Mutación , Estabilidad Proteica , Proteínas/química , Proteínas/genética , Reproducibilidad de los ResultadosRESUMEN
A highly sensitive and reliable Hepatitis B virus surface antigen (HBsAg) measurement is essential to universal screening, timely diagnosis, and management of Hepatitis B virus (HBV) infection. This study aimed to evaluate the performance of MAGLUMI HBsAg chemiluminescence immunoassay (CLIA). MAGLUMI HBsAg (CLIA) was compared against ARCHITECT HBsAg. 411 HBsAg positive samples, including different stages of infection, genotypes, subtypes, mutants, and 30 seroconversion panels were tested to evaluate diagnostic sensitivity. Diagnostic specificity was evaluated by testing 205 hospitalized samples and 5101 blood donor samples. Precision, limit of blank (LoB), limit of detection (LoD), and linearity were also verified. The diagnostic sensitivity of the MAGLUMI HBsAg (CLIA) was 100% with better seroconversion sensitivity than ARCHITECT HBsAg. The MAGLUMI HBsAg (CLIA) has optimal detection efficacy for HBV subgenotypes samples. The analytical sensitivity is 0.039 IU/mL. The initial diagnostic specificity is 99.63% on blood donors and 96.59% on hospitalized samples. The verification data demonstrated high repeatability, a LoB of 0.02 IU/mL, LoD of 0.05 IU/mL and an excellent linearity of 0.050-250 IU/mL (R2 = 0.9946). The MAGLUMI HBsAg (CLIA) is proved a highly sensitive and reliable assay with optimal subgenotype detection efficacy.
Asunto(s)
Antígenos de Superficie de la Hepatitis B , Virus de la Hepatitis B , Hepatitis B , Mediciones Luminiscentes , Sensibilidad y Especificidad , Humanos , Antígenos de Superficie de la Hepatitis B/sangre , Antígenos de Superficie de la Hepatitis B/inmunología , Hepatitis B/diagnóstico , Hepatitis B/sangre , Mediciones Luminiscentes/métodos , Inmunoensayo/métodos , Inmunoensayo/normas , Virus de la Hepatitis B/inmunología , Virus de la Hepatitis B/genética , Virus de la Hepatitis B/aislamiento & purificación , Genotipo , Adulto , Femenino , Masculino , Persona de Mediana Edad , Adulto Joven , Reproducibilidad de los Resultados , Anciano , AdolescenteRESUMEN
Skin cancer can be detected through visual screening and skin analysis based on the biopsy and pathological state of the human body. The survival rate of cancer patients is low, and millions of people are diagnosed annually. By determining the different comparative analyses, the skin malignancy classification is evaluated. Using the Isomap with the vision transformer, we analyze the high-dimensional images with dimensionality reduction. Skin cancer can present with severe cases and life-threatening symptoms. Overall performance evaluation and classification tend to improve the accuracy of the high-dimensional skin lesion dataset when completed. In deep learning methodologies, the distinct phases of skin malignancy classification are determined by its accuracy, specificity, F1 recall, and sensitivity while implementing the classification methodology. A nonlinear dimensionality reduction technique called Isomap preserves the data's underlying nonlinear relationships intact. This is essential for the categorization of skin malignancies, as the features that separate malignant from benign skin lesions may not be linearly separable. Isomap decreases the data's complexity while maintaining its essential characteristics, making it simpler to analyze and explain the findings. High-dimensional datasets for skin lesions have been evaluated and classified more effectively when evaluated and classified using Isomap with the vision transformer.
Asunto(s)
Neoplasias Cutáneas , Humanos , Neoplasias Cutáneas/patología , Neoplasias Cutáneas/clasificación , Neoplasias Cutáneas/diagnóstico , Aprendizaje Profundo , Piel/patologíaRESUMEN
BACKGROUND: In clinical trials, the determination of an adequate sample size is a challenging task, mainly due to the uncertainty about the value of the effect size and nuisance parameters. One method to deal with this uncertainty is a sample size recalculation. Thereby, an interim analysis is performed based on which the sample size for the remaining trial is adapted. With few exceptions, previous literature has only examined the potential of recalculation in two-stage trials. METHODS: In our research, we address sample size recalculation in three-stage trials, i.e. trials with two pre-planned interim analyses. We show how recalculation rules from two-stage trials can be modified to be applicable to three-stage trials. We also illustrate how a performance measure, recently suggested for two-stage trial recalculation (the conditional performance score) can be applied to evaluate recalculation rules in three-stage trials, and we describe performance evaluation in those trials from the global point of view. To assess the potential of recalculation in three-stage trials, we compare, in a simulation study, two-stage group sequential designs with three-stage group sequential designs as well as multiple three-stage designs with recalculation. RESULTS: While we observe a notable favorable effect in terms of power and expected sample size by using three-stage designs compared to two-stage designs, the benefits of recalculation rules appear less clear and are dependent on the performance measures applied. CONCLUSIONS: Sample size recalculation is also applicable in three-stage designs. However, the extent to which recalculation brings benefits depends on which trial characteristics are most important to the applicants.
Asunto(s)
Ensayos Clínicos como Asunto , Proyectos de Investigación , Tamaño de la Muestra , Humanos , Ensayos Clínicos como Asunto/métodos , Ensayos Clínicos como Asunto/estadística & datos numéricos , Proyectos de Investigación/estadística & datos numéricos , Simulación por ComputadorRESUMEN
BACKGROUND: In group-sequential designs, it is typically assumed that there is no time gap between patient enrollment and outcome measurement in clinical trials. However, in practice, there is usually a lag between the two time points. This can affect the statistical analysis of the data, especially in trials with interim analyses. One approach to address delayed responses has been introduced by Hampson and Jennison (J R Stat Soc Ser B Stat Methodol 75:3-54, 2013), who proposed the use of error-spending stopping boundaries for patient enrollment, followed by critical values to reject the null hypothesis if the stopping boundaries are crossed beforehand. Regarding the choice of a trial design, it is important to consider the efficiency of trial designs, e.g. in terms of the probability of trial success (power) and required resources (sample size and time). METHODS: This article aims to shed more light on the performance comparison of group sequential clinical trial designs that account for delayed responses and designs that do not. Suitable performance measures are described and designs are evaluated using the R package rpact. By doing so, we provide insight into global performance measures, discuss the applicability of conditional performance characteristics, and finally whether performance gain justifies the use of complex trial designs that incorporate delayed responses. RESULTS: We investigated how the delayed response group sequential test (DR-GSD) design proposed by Hampson and Jennison (J R Stat Soc Ser B Stat Methodol 75:3-54, 2013) can be extended to include nonbinding lower recruitment stopping boundaries, illustrating that their original design framework can accommodate both binding and nonbinding rules when additional constraints are imposed. Our findings indicate that the performance enhancements from methods incorporating delayed responses heavily rely on the sample size at interim and the volume of data in the pipeline, with overall performance gains being limited. CONCLUSION: This research extends existing literature on group-sequential designs by offering insights into differences in performance. We conclude that, given the overall marginal differences, discussions regarding appropriate trial designs can pivot towards practical considerations of operational feasibility.
Asunto(s)
Proyectos de Investigación , Humanos , Tamaño de la Muestra , Ensayos Clínicos como Asunto/métodos , Algoritmos , Modelos Estadísticos , Interpretación Estadística de Datos , Factores de TiempoRESUMEN
Flow-electrode capacitive deionization (FCDI) is a promising technology for sustainable water treatment. However, studies on the process have thus far been limited to lab-scale conditions and select fields of application. Such limitation is induced by several shortcomings, one of which is the absence of a comprehensive process model that accurately predicts the operational performance and the energy consumption of FCDI. In this study, a simulation model is newly proposed with initial validation based on experimental data and is then utilized to elucidate the performance and the specific energy consumption (SEC) of FCDI under multiple source water conditions ranging from near-groundwater to high salinity brine. Further, simulated pilot-scale FCDI system was compared with actual brackish water reverse osmosis (BWRO) and seawater reverse osmosis (SWRO) plant data with regard to SEC to determine the feasibility of FCDI as an alternative to the conventional membrane processes. Analysis showed that FCDI is competent for operation against brackish water solutions under all possible operational conditions with respect to the BWRO. Moreover, its distinction can be extended to the SWRO for seawater conditions through optimization of its total effective membrane area via scale-up. Accordingly, future directions for the advancement of FCDI was suggested to ultimately prompt the commercialization of the FCDI process.
Asunto(s)
Cloruro de Sodio , Purificación del Agua , Filtración , Electrodos , Agua de MarRESUMEN
OBJECTIVES: To evaluate the analytical characteristics of a novel high-sensitivity cardiac troponin T (hs-cTnT) test on the automatic light-initiated chemiluminescent assay (LiCA®) system, and validated its diagnostic performance for non-ST-segment elevation myocardial infarction (NSTEMI). METHODS: Studies included an extensive analytical evaluation and established the 99th percentile upper reference limit (URL) from apparently healthy individuals, followed by a diagnostic performance validation for NSTEMI. RESULTS: Sex-specific 99th percentile URLs were 16.0â¯ng/L (1.7â¯% CV: coefficient of variation) for men (21-92 years) and 13.4â¯ng/L (2.0â¯% CV) for women (23-87 years) in serum, and 30.6â¯ng/L (0.9â¯% CV) for men (18-87 years) and 20.2â¯ng/L (1.4â¯% CV) for women (18-88 years) in heparin plasma. Detection rates in healthy individuals ranged from 98.9 to 100â¯%. An excellent agreement was identified between LiCA® and Elecsys® assays with a correlation coefficient of 0.993 and mean bias of -0.7â¯% (-1.8-0.4â¯%) across the full measuring range, while the correlation coefficient and overall bias were 0.967 and -1.1â¯% (-2.5-0.3â¯%) for the lower levels of cTnT (10-100â¯ng/L), respectively. At the specific medical decision levels (14.0 and 52.0â¯ng/L), assay difference was estimated to be <5.0â¯%. No significant difference was found between these two assays in terms of area under curve (AUC), sensitivity and specificity, negative predictive value (NPV) and positive predictive value (PPV) for the diagnosis of NSTEMI. CONCLUSIONS: LiCA® hs-cTnT is a reliable 3rd-generation (level 4) high-sensitivity assay for detecting cardiac troponin T. The assay is acceptable for practical use in the diagnosis of NSTEMI.
Asunto(s)
Infarto del Miocardio , Infarto del Miocardio sin Elevación del ST , Infarto del Miocardio con Elevación del ST , Masculino , Humanos , Femenino , Troponina T , Infarto del Miocardio/diagnóstico , Heparina , Sensibilidad y Especificidad , BiomarcadoresRESUMEN
OBJECTIVES: Clinical laboratories face limitations in implementing advanced quality control (QC) methods with existing systems. This study aimed to develop a web-based application to addresses this gap, and improve QC practices. METHODS: QC Constellation, a web application built using Python 3.11, integrates various statistical QC modules. These include Levey-Jennings charts with Westgard rules, sigma-metric calculations, exponentially weighted moving average (EWMA) and cumulative sum (CUSUM) charts, and method decision charts. Additionally, it offers a risk-based QC section and a patient-based QC module aligning with modern QC practices. The codes and the web application links for QC Constellation were shared at https://github.com/hikmetc/QC_Constellation, and http://qcconstellation.com, respectively. RESULTS: Using synthetic data, QC Constellation demonstrated effective implementation of Levey-Jennings charts with user-friendly features like checkboxes for Westgard rules and customizable moving averages graphs. Sigma-metric calculations for hypothetical performance values of serum total cholesterol were successfully performed using allowable total error and maximum allowable measurement uncertainty goals, and displayed on method decision charts. The utility of the risk-based QC module was exemplified by assessing QC plans for serum total cholesterol, showcasing the application's capability in calculating risk-based QC parameters including maximum unreliable final patient results, risk management index, and maximum run size and offering risk-based QC recommendations. Similarly, the patient-based QC and optimization modules were demonstrated using simulated sodium results. CONCLUSIONS: In conclusion, QC Constellation emerges as a pivotal tool for laboratory professionals, streamlining the management of quality control and analytical performance monitoring, while enhancing patient safety through optimized QC processes.
Asunto(s)
Laboratorios Clínicos , Control de Calidad , Programas Informáticos , Humanos , Laboratorios Clínicos/normasRESUMEN
BACKGROUND: Sepsis, an acute and potentially fatal systemic response to infection, significantly impacts global health by affecting millions annually. Prompt identification of sepsis is vital, as treatment delays lead to increased fatalities through progressive organ dysfunction. While recent studies have delved into leveraging Machine Learning (ML) for predicting sepsis, focusing on aspects such as prognosis, diagnosis, and clinical application, there remains a notable deficiency in the discourse regarding feature engineering. Specifically, the role of feature selection and extraction in enhancing model accuracy has been underexplored. OBJECTIVES: This scoping review aims to fulfill two primary objectives: To identify pivotal features for predicting sepsis across a variety of ML models, providing valuable insights for future model development, and To assess model efficacy through performance metrics including AUROC, sensitivity, and specificity. RESULTS: The analysis included 29 studies across diverse clinical settings such as Intensive Care Units (ICU), Emergency Departments, and others, encompassing 1,147,202 patients. The review highlighted the diversity in prediction strategies and timeframes. It was found that feature extraction techniques notably outperformed others in terms of sensitivity and AUROC values, thus indicating their critical role in improving sepsis prediction models. CONCLUSION: Key dynamic indicators, including vital signs and critical laboratory values, are instrumental in the early detection of sepsis. Applying feature selection methods significantly boosts model precision, with models like Random Forest and XG Boost showing promising results. Furthermore, Deep Learning models (DL) reveal unique insights, spotlighting the pivotal role of feature engineering in sepsis prediction, which could greatly benefit clinical practice.
Asunto(s)
Aprendizaje Automático , Sepsis , Humanos , Sepsis/diagnóstico , Sepsis/terapia , Aprendizaje Automático/tendencias , Aprendizaje Automático/normasRESUMEN
In this dual-center study, we assessed the BioHermes A1C EXP M13 system for point-of-care (POC) HbA1c testing against two NGSP-certified HPLC instruments, the Bio-Rad D100 and Tosoh G8. Analyzing 605 samples, we evaluated the A1C EXP's reproducibility, sensitivity, specificity and impact of anemia on HbA1c measurements. The device showed excellent reproducibility with CVs under 2.4% and high sensitivity and specificity for diabetes diagnosis-98.1% and 96.8% against D100, and 97.1% and 96.7% against G8. Passing-Bablok regression confirmed a close correlation between A1C EXP and the HPLC instruments, with equations y = 0.10625 + 0.9688x (D100) and y = 0.0000 + 0.1000x (G8), and Bland-Altman plots indicated mean relative differences of -1.4% (D100) and -0.4% (G8). However, in anemic samples, A1C EXP showed a negative bias compared to HPLC devices, suggesting that anemia may affect the accuracy of HbA1c results. The study indicates that A1C EXP is a reliable POC alternative to laboratory assays, albeit with considerations for anemic patients.
Asunto(s)
Hemoglobina Glucada , Pruebas en el Punto de Atención , Hemoglobina Glucada/análisis , Humanos , Pruebas en el Punto de Atención/normas , Reproducibilidad de los Resultados , Anemia/diagnóstico , Anemia/sangre , Cromatografía Líquida de Alta Presión , Sensibilidad y Especificidad , Diabetes Mellitus/sangre , Diabetes Mellitus/diagnóstico , Sistemas de Atención de Punto/normasRESUMEN
Numerous studies have used air quality models to estimate pollutant concentrations in the Metropolitan Area of São Paulo (MASP) by using different inputs and assumptions. Our objectives are to summarize these studies, compare their performance, configurations, and inputs, and recommend areas of further research. We examined 29 air quality modeling studies that focused on ozone (O3) and fine particulate matter (PM2.5) performed over the MASP, published from 2001 to 2023. The California Institute of Technology airshed model (CIT) was the most used offline model, while the Weather Research and Forecasting model coupled with Chemistry (WRF-Chem) was the most used online model. Because the main source of air pollution in the MASP is the vehicular fleet, it is commonly used as the only anthropogenic input emissions. Simulation periods were typically the end of winter and during spring, seasons with higher O3 and PM2.5 concentrations. Model performance for hourly ozone is good with half of the studies with Pearson correlation above 0.6 and root mean square error (RMSE) ranging from 7.7 to 27.1 ppb. Fewer studies modeled PM2.5 and their performance is not as good as ozone estimates. Lack of information on emission sources, pollutant measurements, and urban meteorology parameters is the main limitation to perform air quality modeling. Nevertheless, researchers have used measurement campaign data to update emission factors, estimate temporal emission profiles, and estimate volatile organic compounds (VOCs) and aerosol speciation. They also tested different emission spatial disaggregation approaches and transitioned to global meteorological reanalysis with a higher spatial resolution. Areas of research to explore are further evaluation of models' physics and chemical configurations, the impact of climate change on air quality, the use of satellite data, data assimilation techniques, and using model results in health impact studies. This work provides an overview of advancements in air quality modeling within the MASP and offers practical approaches for modeling air quality in other South American cities with limited data, particularly those heavily impacted by vehicle emissions.
RESUMEN
BACKGROUND: Hospitals are the biggest consumers of health system budgets and hence measuring hospital performance by quantitative or qualitative accessible and reliable indicators is crucial. This review aimed to categorize and present a set of indicators for evaluating overall hospital performance. METHODS: We conducted a literature search across three databases, i.e., PubMed, Scopus, and Web of Science, using possible keyword combinations. We included studies that explored hospital performance evaluation indicators from different dimensions. RESULTS: We included 91 English language studies published in the past 10 years. In total, 1161 indicators were extracted from the included studies. We classified the extracted indicators into 3 categories, 14 subcategories, 21 performance dimensions, and 110 main indicators. Finally, we presented a comprehensive set of indicators with regard to different performance dimensions and classified them based on what they indicate in the production process, i.e., input, process, output, outcome and impact. CONCLUSION: The findings provide a comprehensive set of indicators at different levels that can be used for hospital performance evaluation. Future studies can be conducted to validate and apply these indicators in different contexts. It seems that, depending on the specific conditions of each country, an appropriate set of indicators can be selected from this comprehensive list of indicators for use in the performance evaluation of hospitals in different settings.
Asunto(s)
Hospitales , Indicadores de Calidad de la Atención de Salud , Humanos , Hospitales/normasRESUMEN
OBJECTIVES: This study aimed to comprehensively evaluate Mexico's health system performance from 1990 to 2019 utilising the Health Access and Quality Index (HAQI) as a primary indicator. STUDY DESIGN: A retrospective ecological analysis was performed using data from the Global Burden of Diseases, Injuries and Risk Factors Study (GBD) study and the National Population Council (CONAPO). METHODS: HAQI values for 1990, 2000, 2010, 2015, and 2019 were examined for each state in Mexico and three age groups (young, working, and post-working). Additionally, the marginalisation index was employed to assess inequalities in the HAQI distribution across states. The concentration index of the HAQI for each year was estimated, and the efficiency of states in producing the HAQI was evaluated using a data envelopment approach. RESULTS: Through the analysis of national and subnational data, results indicated an overall improvement in healthcare access and quality during the study period. Although differences in the HAQI value related to state marginalisation decreased from 1990 to 2015, by 2019, the inequality had returned to a level comparable to 2000. Efficiency in producing health (HAQI values) exhibited substantial heterogeneity and fluctuations in the ranking order over time. States such as Nuevo León consistently performed well, while others, such as Guerrero, Chihuahua, Mexico City, and Puebla, consistently underperformed. CONCLUSIONS: The findings from this study emphasise the necessity for nuanced strategies to address healthcare disparities and enhance the overall system performance. The study provides valuable insights for ongoing discussions about the future of Mexico's healthcare system, aiming to inform evidence-based policy decisions and improve the nationwide delivery of healthcare services.
RESUMEN
Decision accuracy is a crucial factor in the evaluation of refereeing performance. In sports research, officials' decision-making is frequently assessed outside real games through video-based decision experiments, where they evaluate recorded game situations from a third-person perspective. This study examines whether the inclusion of the first-person perspective influences decision accuracy and certainty. Twenty-four professional officials from the first and second German basketball leagues participated in the study. The officials assessed 50 game situations from both first-person and third-person perspectives, indicating their decisions and certainty levels. The statistical analysis utilises signal detection theory to evaluate the efficacy of the first-person perspective compared to the third-person perspective in identifying rule violations and no-calls in video recordings. The findings indicate that the first-person perspective does not yield superior accuracy in identifying foul calls. However, scenes from the first-person perspective exhibit a significant 9% increase in correctly identifying no-calls. Furthermore, officials report significantly higher levels of decision certainty and comfort when using the first-person perspective. The study suggests that sports officials may benefit from incorporating additional scenes from the first-person perspective into video-based decision training. Future studies should explore whether this additional perspective improves the training effect and translates into enhanced in-game performance.