RESUMO
Spatial transcriptomics offers deep insights into cellular functional localization and communication by mapping gene expression to spatial locations. Traditional approaches that focus on selecting spatially variable genes often overlook the complexity of biological pathways and the interactions among genes. Here, we introduce a novel framework that shifts the focus towards directly identifying functional pathways associated with spatial variability by adapting the Brownian distance covariance test in an innovative manner to explore the heterogeneity of biological functions over space. Unlike most other methods, this statistical testing approach is free of gene selection and parameter selection and allows nonlinear and complex dependencies. It allows for a deeper understanding of how cells coordinate their activities across different spatial domains through biological pathways. By analyzing real human and mouse datasets, the method found significant pathways that were associated with spatial variation, as well as different pathway patterns among inner- and edge-cancer regions. This innovative framework offers a new perspective on analyzing spatial transcriptomic data, contributing to our understanding of tissue architecture and disease pathology. The implementation is publicly available at https://github.com/tianlq-prog/STpathway.
Assuntos
Perfilação da Expressão Gênica , Humanos , Camundongos , Animais , Perfilação da Expressão Gênica/métodos , Transcriptoma , Biologia Computacional/métodos , Algoritmos , Neoplasias/genética , Neoplasias/metabolismo , Neoplasias/patologia , Transdução de SinaisRESUMO
A/B testing is widely used to tune search and recommendation algorithms, to compare product variants as efficiently and effectively as possible, and even to study animal behavior. With ongoing investment, due to diminishing returns, the items produced by the new alternative B show smaller and smaller improvement in quality from the items produced by the current system A. By formalizing this observation, we develop closed-form analytical expressions for the sample efficiency of a number of widely used families of slate-based comparison tests. In empirical trials, these theoretical sample complexity results are shown to be predictive of real-world testing efficiency outcomes. These findings offer opportunities for both more cost-effective testing and a better analytical understanding of the problem.
RESUMO
The largest magnitude earthquake in a sequence is often used as a proxy for hazard estimates, as consequences are often predominately from this single event (in small seismic zones). In this article, the concept of order statistics is adapted to infer the maximum magnitude ([Formula: see text]) of an earthquake catalogue. A suite tools developed here can discern [Formula: see text] influences through hypothesis testing, quantify [Formula: see text] through maximum likelihood estimation (MLE) or select the best [Formula: see text] prediction amongst several models. The efficacy of these tools is benchmarked against synthetic and real-data tests, demonstrating their utility. Ultimately, 13 cases of induced seismicity spanning wastewater disposal, hydraulic fracturing and enhanced geothermal systems are tested for volume-based [Formula: see text]. I find that there is no evidence of volume-based processes influencing any of these cases. On the contrary, all these cases are adequately explained by an unbounded magnitude distribution. This is significant because it suggests that induced earthquake hazards should also be treated as unbounded. On the other hand, if bounded cases exist, then the tools developed here will be able to discern them, potentially changing how an operator mitigates these hazards. Overall, this suite of tools will be important for better-understanding earthquakes and managing their risks. This article is part of the theme issue 'Induced seismicity in coupled subsurface systems'.
RESUMO
TLS is nowadays often used for deformation monitoring. As it is not able to scan identical points in different time epochs, mathematical models of objects derived from point clouds have to be used. The most common geometric form to describe built objects is a plane, which can be described by four parameters. In this study, we aimed to find out how small changes in the parameters of the plane can be detected by TLS. We aimed to eliminate all possible factors that influence the scanning. Then, we shifted and tilted a finite physical representation of a plane in a controlled way. After each controlled change, the board was scanned several times and the parameters of the plane were calculated. We used two different types of scanning devices and compared their performance. The changes in the plane parameters were compared with the actual change values and statistically tested. The results show that TLS detects shifts in the millimetre range and tilts of 150â³ (for a 1 m plane). A robotic total station can achieve twice the precision of TLS despite lower density and slower performance. For deformation monitoring, we strongly recommend repeating each scan several times (i) to check for gross errors and (ii) to obtain a realistic precision estimate.
RESUMO
It is tempting to base (eco-)toxicological assay evaluation solely on statistical significance tests. The approach is stringent, objective and facilitates binary decisions. However, tests according to null hypothesis statistical testing (NHST) are thought experiments that rely heavily on assumptions. The generic and unreflected application of statistical tests has been called "mindless" by Gigerenzer. While statistical tests have an appropriate application domain, the present work investigates how unreflected testing may affect toxicological assessments. Dunnett multiple-comparison and Williams trend testing and their compatibility intervals are compared with dose-response-modelling in case studies, where data do not follow textbook behavior, nor behave as expected from a toxicological point of view. In such cases, toxicological assessments based only on p-values may be biased and biological evaluations based on plausibility may be prioritized. If confidence in a negative assay outcome cannot be established, further data may be needed for a robust toxicological assessment.
Assuntos
Interpretação Estatística de Dados , Toxicologia/estatística & dados numéricos , Relação Dose-Resposta a Droga , Modelos Biológicos , Testes de Toxicidade/estatística & dados numéricosRESUMO
Passive multistatic radars have gained a lot of interest in recent years as they offer many benefits contrary to conventional radars. Here in this research, our aim is detection of target in a passive multistatic radar system. The system contains a single transmitter and multiple spatially distributed receivers comprised of both the surveillance and reference antennas. The system consists of two main parts: 1. Local receiver, and 2. Fusion center. Each local receiver detects the signal, processes it, and passes the information to the fusion center for final detection. To take the advantage of spatial diversity, we apply major fusion techniques consisting of hard fusion and soft fusion for the case of multistatic passive radars. Hard fusion techniques are analyzed for the case of different local radar detectors. In terms of soft fusion, a blind technique called equal gain soft fusion technique with random matrix theory-based local detector is analytically and theoretically analyzed under null hypothesis along with the calculation of detection threshold. Furthermore, six novel random matrix theory-based soft fusion techniques are proposed. All the techniques are blind in nature and hence do not require any knowledge of transmitted signal or channel information. Simulation results illustrate that proposed fusion techniques increase detection performance to a reasonable extent compared to other blind fusion techniques.
RESUMO
Applying georadar (GPR) technology for detecting underground utilities is an important element of the comprehensive assessment of the location and ground infrastructure status. These works are usually connected with the conducted investment processes or serialised inventory of underground fittings. The detection of infrastructure is also crucial in implementing the BIM technology, 3D cadastre, and planned network modernization works. GPR detection accuracy depends on the type of equipment used, the selected detection method, and external factors. The multitude of techniques used for localizing underground utilities and constantly growing accuracy demands resulting from the fact that it is often necessary to detect infrastructure under challenging conditions of dense urban development leads to the need to improve the existing technologies. The factor that motivated us to start research on assessing the precision and accuracy of ground penetrating radar detection was the need to ensure the appropriate accuracy, precision, and reliability of detecting underground utilities versus different methods and analyses. The results of the multi-variant GPR were subjected to statistical testing. Various analyses were also conducted, depending on the detection method and on the current soil parameters using a unique sensor probe. When planning detection routes, we took into account regular, established grids and tracked the trajectory of movement of the equipment using GNSS receivers (internal and external ones). Moreover, a specialist probe was used to evaluate the potential influence of the changing soil conditions on the obtained detection results. Our tests were conducted in a developed area for ten months. The results confirmed a strong correlation between the obtained accuracy and the measurement method used, while the correlation with the other factors discussed here was significantly weaker.
Assuntos
Radar , Solo , Reprodutibilidade dos TestesRESUMO
Loss of cognitive ability is commonly associated with dementia, a broad category of progressive brain diseases. However, major depressive disorder may also cause temporary deterioration of one's cognition known as pseudodementia. Differentiating a true dementia and pseudodementia is still difficult even for an experienced clinician and extensive and careful examinations must be performed. Although mental disorders such as depression and dementia have been studied, there is still no solution for shorter and undemanding pseudodementia screening. This study inspects the distribution and statistical characteristics from both dementia patient and depression patient, and compared them. It is found that some acoustic features were shared in both dementia and depression, albeit their correlation was reversed. Statistical significance was also found when comparing the features. Additionally, the possibility of utilizing machine learning for automatic pseudodementia screening was explored. The machine learning part includes feature selection using LASSO algorithm and support vector machine (SVM) with linear kernel as the predictive model with age-matched symptomatic depression patient and dementia patient as the database. High accuracy, sensitivity, and specificity was obtained in both training session and testing session. The resulting model was also tested against other datasets that were not included and still performs considerably well. These results imply that dementia and depression might be both detected and differentiated based on acoustic features alone. Automated screening is also possible based on the high accuracy of machine learning results.
Assuntos
Demência/diagnóstico , Transtorno Depressivo Maior/diagnóstico , Fala , Máquina de Vetores de Suporte , Adulto , Idoso , Idoso de 80 Anos ou mais , Algoritmos , Demência/classificação , Depressão/diagnóstico , Transtorno Depressivo Maior/classificação , Feminino , Humanos , Masculino , Pessoa de Meia-IdadeRESUMO
Spatial or temporal aspects of neural organization are known to be important indices of how cognition is organized. However, measurements and estimations are often noisy and many of the algorithms used are probabilistic, which in combination have been argued to limit studies exploring the neural basis of specific aspects of cognition. Focusing on static and dynamic functional connectivity estimations, we propose to leverage this variability to improve statistical efficiency in relating these estimations to behavior. To achieve this goal, we use a procedure based on permutation testing that provides a way of combining the results from many individual tests that refer to the same hypothesis. This is needed when testing a measure whose value is obtained from a noisy process, which can be repeated multiple times, referred to as replications. Focusing on functional connectivity, this noisy process can be: (a) computational, for example, when using an approximate inference algorithm for which different runs can produce different results or (b) observational, if we have the capacity to acquire data multiple times, and the different acquired data sets can be considered noisy examples of some underlying truth. In both cases, we are not interested in the individual replications but on the unobserved process generating each replication. In this note, we show how results can be combined instead of choosing just one of the estimated models. Using both simulations and real data, we show the benefits of this approach in practice.
Assuntos
Algoritmos , Encéfalo/fisiologia , Cognição/fisiologia , Conectoma/métodos , Vias Neurais/fisiologia , Simulação por Computador , Humanos , Processamento de Imagem Assistida por Computador/métodosRESUMO
When a candidate predictive marker is available, but evidence on its predictive ability is not sufficiently reliable, all-comers trials with marker stratification are frequently conducted. We propose a framework for planning and evaluating prospective testing strategies in confirmatory, phase III marker-stratified clinical trials based on a natural assumption on heterogeneity of treatment effects across marker-defined subpopulations, where weak rather than strong control is permitted for multiple population tests. For phase III marker-stratified trials, it is expected that treatment efficacy is established in a particular patient population, possibly in a marker-defined subpopulation, and that the marker accuracy is assessed when the marker is used to restrict the indication or labelling of the treatment to a marker-based subpopulation, ie, assessment of the clinical validity of the marker. In this paper, we develop statistical testing strategies based on criteria that are explicitly designated to the marker assessment, including those examining treatment effects in marker-negative patients. As existing and developed statistical testing strategies can assert treatment efficacy for either the overall patient population or the marker-positive subpopulation, we also develop criteria for evaluating the operating characteristics of the statistical testing strategies based on the probabilities of asserting treatment efficacy across marker subpopulations. Numerical evaluations to compare the statistical testing strategies based on the developed criteria are provided.
Assuntos
Ensaios Clínicos Fase III como Assunto , Interpretação Estatística de Dados , Resultado do Tratamento , Humanos , Probabilidade , Estudos ProspectivosRESUMO
AIM: The purpose of this study was to develop and test the psychometric properties of the new Cultural and Linguistic Diversity scale, which is designed to be used with the newly validated Clinical Learning Environment, Supervision and Nurse Teacher scale for assessing international nursing students' clinical learning environments. BACKGROUND: In various developed countries, clinical placements are known to present challenges in the professional development of international nursing students. DESIGN: A cross-sectional survey. METHODS: Data were collected from eight Finnish universities of applied sciences offering nursing degree courses taught in English during 2015-2016. All the relevant students (N = 664) were invited and 50% chose to participate. Of the total data submitted by the participants, 28% were used for scale validation. The construct validity of the two scales was tested by exploratory factor analysis, while their validity with respect to convergence and discriminability was assessed using Spearman's correlation. RESULTS: Construct validation of the Clinical Learning Environment, Supervision and Nurse Teacher scale yielded an eight-factor model with 34 items, while validation of the Cultural and Linguistic Diversity scale yielded a five-factor model with 21 items. CONCLUSION: A new scale was developed to improve evidence-based mentorship of international nursing students in clinical learning environments. The instrument will be useful to educators seeking to identify factors that affect the learning of international students.
Assuntos
Educação em Enfermagem/métodos , Enfermeiros Internacionais/educação , Estudantes de Enfermagem , Adolescente , Adulto , Enfermagem Baseada em Evidências/normas , Feminino , Finlândia , Humanos , Aprendizagem , Masculino , Mentores , Pessoa de Meia-Idade , Psicometria , Inquéritos e Questionários , Adulto JovemRESUMO
Misinterpretation and abuse of statistical tests, confidence intervals, and statistical power have been decried for decades, yet remain rampant. A key problem is that there are no interpretations of these concepts that are at once simple, intuitive, correct, and foolproof. Instead, correct use and interpretation of these statistics requires an attention to detail which seems to tax the patience of working scientists. This high cognitive demand has led to an epidemic of shortcut definitions and interpretations that are simply wrong, sometimes disastrously so-and yet these misinterpretations dominate much of the scientific literature. In light of this problem, we provide definitions and a discussion of basic statistics that are more general and critical than typically found in traditional introductory expositions. Our goal is to provide a resource for instructors, researchers, and consumers of statistics whose knowledge of statistical theory and technique may be limited but who wish to avoid and spot misinterpretations. We emphasize how violation of often unstated analysis protocols (such as selecting analyses for presentation based on the P values they produce) can lead to small P values even if the declared test hypothesis is correct, and can lead to large P values even if that hypothesis is incorrect. We then provide an explanatory list of 25 misinterpretations of P values, confidence intervals, and power. We conclude with guidelines for improving statistical interpretation and reporting.
Assuntos
Intervalos de Confiança , Interpretação Estatística de Dados , Humanos , ProbabilidadeRESUMO
Current methods for thresholding functional magnetic resonance imaging (fMRI) maps are based on the well-known hypothesis-test framework, optimal for addressing novel theoretical claims. However, these methods as typically practiced have a strong bias toward protecting the null hypothesis, and thus may not provide an optimal balance between specificity and sensitivity in forming activation maps for surgical planning. Maps based on hypothesis-test thresholds are also highly sensitive to sample size and signal-to-noise ratio, whereas many clinical applications require methods that are robust to these effects. We propose a new thresholding method, optimized for surgical planning, based on normalized amplitude thresholding. We show that this method produces activation maps that are more reproducible and more predictive of postoperative cognitive outcome than maps produced with current standard thresholding methods.
Assuntos
Mapeamento Encefálico/métodos , Epilepsia/fisiopatologia , Epilepsia/cirurgia , Aumento da Imagem/métodos , Interpretação de Imagem Assistida por Computador/métodos , Imageamento por Ressonância Magnética/métodos , Cuidados Pré-Operatórios/métodos , Adulto , Algoritmos , Epilepsia/diagnóstico , Feminino , Humanos , Masculino , Seleção de Pacientes , Reconhecimento Automatizado de Padrão/métodos , Reprodutibilidade dos Testes , Sensibilidade e Especificidade , Cirurgia Assistida por Computador/métodos , Resultado do TratamentoRESUMO
Research on different machine learning (ML) has become incredibly popular during the past few decades. However, for some researchers not familiar with statistics, it might be difficult to understand how to evaluate the performance of ML models and compare them with each other. Here, we introduce the most common evaluation metrics used for the typical supervised ML tasks including binary, multi-class, and multi-label classification, regression, image segmentation, object detection, and information retrieval. We explain how to choose a suitable statistical test for comparing models, how to obtain enough values of the metric for testing, and how to perform the test and interpret its results. We also present a few practical examples about comparing convolutional neural networks used to classify X-rays with different lung infections and detect cancer tumors in positron emission tomography images.
Assuntos
Processamento de Imagem Assistida por Computador , Aprendizado de Máquina , Processamento de Imagem Assistida por Computador/métodos , Redes Neurais de Computação , Aprendizado de Máquina Supervisionado , Tomografia por Emissão de PósitronsRESUMO
Introduction: Despite the large amount of researches addressed the issue of the relationship between the intensity of preoperative symptoms of depression and/ or anxiety with their postoperative intensity and any complications after surgery, there have been almost unaddressed such subjects as how the patients perceive their own capabilities or physical attractiveness, and the emotions which are evoked by various aspects of their own bodies, including postoperative scars. These aspects play a significant role in assessing the quality of patients' life and have a significant impact on the overall assessment of the surgery as an event, in both the short- and long-term perspective. Aim: To evaluate the relationship between anxiety, pain level, self-efficacy and body esteem in the pre- and postoperative periods among patients scheduled for coronary artery bypass surgery. Material and methods: Prospective studies were carried out in a group of 50 patients scheduled for coronary artery bypass surgery, either on a planned or urgent basis. Anxiety, both as a state and as a trait, was assessed using the Polish version of the State-Trait Anxiety Inventory (STAI). The Visual Analogue Scale (VAS) was employed to evaluate pain. The Self-Efficacy Gauge measured self-efficacy, while the Body Esteem Scale assessed body esteem. Results and Conclusions: The intensity of state anxiety significantly negatively correlated with self-efficacy following CABG surgery. There was a statistically significant negative correlation between the intensity of painand self-efficacy in the postoperative period. Among female patients, the intensity of pain, both pre- and post-operatively, negatively correlated with their assessment of body esteem concerning physical condition at the respective time points. When assessing anxiety as a trait during the perioperative period, a positive correlation with pain intensification after CABG was identified.
RESUMO
BACKGROUND AND OBJECTIVE: Mediation analysis is used to gain insight into the mechanisms of exposure-outcome effects by dividing this effect into a direct and an indirect effect. One of the problems of mediation analysis is that in many situations, the standard error of the indirect effect is much lower than the standard errors of the total and direct effect. Because this problem is ignored in the epidemiological literature, the purpose of this paper was to illustrate this problem and to provide an advice regarding the statistical testing of indirect effects in mediation analysis. METHODS: To illustrate the problem of the estimation of the standard error of the indirect effect two real life datasets and several simulations are used. RESULTS: The paper shows that the problem of estimating the standard error of the indirect effect was most pronounced when the relationship between exposure and mediator and the relationship between mediator and outcome were equally strong. Furthermore, the magnitude of the estimation problem is different for different strengths of the mediation effect. CONCLUSION: The indirect effect in mediation analysis should not be tested for statistical significance but the importance of mediation should be evaluated by its clinical relevance.
Assuntos
Análise de Mediação , Humanos , Interpretação Estatística de Dados , Modelos Estatísticos , CausalidadeRESUMO
Neural systems are networks, and strategic comparisons between multiple networks are a prevalent task in many research scenarios. In this study, we construct a statistical test for the comparison of matrices representing pairwise aspects of neural networks, in particular, the correlation between spiking activity and connectivity. The "eigenangle test" quantifies the similarity of two matrices by the angles between their ranked eigenvectors. We calibrate the behavior of the test for use with correlation matrices using stochastic models of correlated spiking activity and demonstrate how it compares to classical two-sample tests, such as the Kolmogorov-Smirnov distance, in the sense that it is able to evaluate also structural aspects of pairwise measures. Furthermore, the principle of the eigenangle test can be applied to compare the similarity of adjacency matrices of certain types of networks. Thus, the approach can be used to quantitatively explore the relationship between connectivity and activity with the same metric. By applying the eigenangle test to the comparison of connectivity matrices and correlation matrices of a random balanced network model before and after a specific synaptic rewiring intervention, we gauge the influence of connectivity features on the correlated activity. Potential applications of the eigenangle test include simulation experiments, model validation, and data analysis.
Assuntos
Redes Neurais de Computação , Neurônios , Simulação por Computador , Vias Neurais , Rede NervosaRESUMO
There has been a concerted effort by the neuroimaging community to establish standards for computational methods for data analysis that promote reproducibility and portability. In particular, the Brain Imaging Data Structure (BIDS) specifies a standard for storing imaging data, and the related BIDS App methodology provides a standard for implementing containerized processing environments that include all necessary dependencies to process BIDS datasets using image processing workflows. We present the BrainSuite BIDS App, which encapsulates the core MRI processing functionality of BrainSuite within the BIDS App framework. Specifically, the BrainSuite BIDS App implements a participant-level workflow comprising three pipelines and a corresponding set of group-level analysis workflows for processing the participant-level outputs. The BrainSuite Anatomical Pipeline (BAP) extracts cortical surface models from a T1-weighted (T1w) MRI. It then performs surface-constrained volumetric registration to align the T1w MRI to a labeled anatomical atlas, which is used to delineate anatomical regions of interest in the MRI brain volume and on the cortical surface models. The BrainSuite Diffusion Pipeline (BDP) processes diffusion-weighted imaging (DWI) data, with steps that include coregistering the DWI data to the T1w scan, correcting for geometric image distortion, and fitting diffusion models to the DWI data. The BrainSuite Functional Pipeline (BFP) performs fMRI processing using a combination of FSL, AFNI, and BrainSuite tools. BFP coregisters the fMRI data to the T1w image, then transforms the data to the anatomical atlas space and to the Human Connectome Project's grayordinate space. Each of these outputs can then be processed during group-level analysis. The outputs of BAP and BDP are analyzed using the BrainSuite Statistics in R (bssr) toolbox, which provides functionality for hypothesis testing and statistical modeling. The outputs of BFP can be analyzed using atlas-based or atlas-free statistical methods during group-level processing. These analyses include the application of BrainSync, which synchronizes the time-series data temporally and enables comparison of resting-state or task-based fMRI data across scans. We also present the BrainSuite Dashboard quality control system, which provides a browser-based interface for reviewing the outputs of individual modules of the participant-level pipelines across a study in real-time as they are generated. BrainSuite Dashboard facilitates rapid review of intermediate results, enabling users to identify processing errors and make adjustments to processing parameters if necessary. The comprehensive functionality included in the BrainSuite BIDS App provides a mechanism for rapidly deploying the BrainSuite workflows into new environments to perform large-scale studies. We demonstrate the capabilities of the BrainSuite BIDS App using structural, diffusion, and functional MRI data from the Amsterdam Open MRI Collection's Population Imaging of Psychology dataset.
RESUMO
BACKGROUND: With increasing data sizes and more easily available computational methods, neurosciences rely more and more on predictive modeling with machine learning, e.g., to extract disease biomarkers. Yet, a successful prediction may capture a confounding effect correlated with the outcome instead of brain features specific to the outcome of interest. For instance, because patients tend to move more in the scanner than controls, imaging biomarkers of a disease condition may mostly reflect head motion, leading to inefficient use of resources and wrong interpretation of the biomarkers. RESULTS: Here we study how to adapt statistical methods that control for confounds to predictive modeling settings. We review how to train predictors that are not driven by such spurious effects. We also show how to measure the unbiased predictive accuracy of these biomarkers, based on a confounded dataset. For this purpose, cross-validation must be modified to account for the nuisance effect. To guide understanding and practical recommendations, we apply various strategies to assess predictive models in the presence of confounds on simulated data and population brain imaging settings. Theoretical and empirical studies show that deconfounding should not be applied to the train and test data jointly: modeling the effect of confounds, on the training data only, should instead be decoupled from removing confounds. CONCLUSIONS: Cross-validation that isolates nuisance effects gives an additional piece of information: confound-free prediction accuracy.