Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 47
Filtrar
1.
Biostatistics ; 24(2): 358-371, 2023 04 14.
Artículo en Inglés | MEDLINE | ID: mdl-34435196

RESUMEN

With mammography being the primary breast cancer screening strategy, it is essential to make full use of the mammogram imaging data to better identify women who are at higher and lower than average risk. Our primary goal in this study is to extract mammogram-based features that augment the well-established breast cancer risk factors to improve prediction accuracy. In this article, we propose a supervised functional principal component analysis (sFPCA) over triangulations method for extracting features that are ordered by the magnitude of association with the failure time outcome. The proposed method accommodates the irregular boundary issue posed by the breast area within the mammogram imaging data with flexible bivariate splines over triangulations. We also provide an eigenvalue decomposition algorithm that is computationally efficient. Compared to the conventional unsupervised FPCA method, the proposed method results in a lower Brier Score and higher area under the ROC curve (AUC) in simulation studies. We apply our method to data from the Joanne Knight Breast Health Cohort at Siteman Cancer Center. Our approach not only obtains the best prediction performance comparing to unsupervised FPCA and benchmark models but also reveals important risk patterns within the mammogram images. This demonstrates the importance of utilizing additional supervised image-based features to clarify breast cancer risk.


Asunto(s)
Neoplasias de la Mama , Mamografía , Humanos , Femenino , Neoplasias de la Mama/diagnóstico por imagen , Mamografía/métodos , Algoritmos , Análisis de Componente Principal
2.
BMC Bioinformatics ; 24(1): 271, 2023 Jun 30.
Artículo en Inglés | MEDLINE | ID: mdl-37391692

RESUMEN

BACKGROUND: Dealing with the high dimension of both neuroimaging data and genetic data is a difficult problem in the association of genetic data to neuroimaging. In this article, we tackle the latter problem with an eye toward developing solutions that are relevant for disease prediction. Supported by a vast literature on the predictive power of neural networks, our proposed solution uses neural networks to extract from neuroimaging data features that are relevant for predicting Alzheimer's Disease (AD) for subsequent relation to genetics. The neuroimaging-genetic pipeline we propose is comprised of image processing, neuroimaging feature extraction and genetic association steps. We present a neural network classifier for extracting neuroimaging features that are related with the disease. The proposed method is data-driven and requires no expert advice or a priori selection of regions of interest. We further propose a multivariate regression with priors specified in the Bayesian framework that allows for group sparsity at multiple levels including SNPs and genes. RESULTS: We find the features extracted with our proposed method are better predictors of AD than features used previously in the literature suggesting that single nucleotide polymorphisms (SNPs) related to the features extracted by our proposed method are also more relevant for AD. Our neuroimaging-genetic pipeline lead to the identification of some overlapping and more importantly some different SNPs when compared to those identified with previously used features. CONCLUSIONS: The pipeline we propose combines machine learning and statistical methods to benefit from the strong predictive performance of blackbox models to extract relevant features while preserving the interpretation provided by Bayesian models for genetic association. Finally, we argue in favour of using automatic feature extraction, such as the method we propose, in addition to ROI or voxelwise analysis to find potentially novel disease-relevant SNPs that may not be detected when using ROIs or voxels alone.


Asunto(s)
Enfermedad de Alzheimer , Neuroimagen , Humanos , Teorema de Bayes , Procesamiento de Imagen Asistido por Computador , Enfermedad de Alzheimer/diagnóstico por imagen , Enfermedad de Alzheimer/genética , Redes Neurales de la Computación
3.
Biometrics ; 79(2): 1359-1369, 2023 06.
Artículo en Inglés | MEDLINE | ID: mdl-34854477

RESUMEN

Screening mammography aims to identify breast cancer early and secondarily measures breast density to classify women at higher or lower than average risk for future breast cancer in the general population. Despite the strong association of individual mammography features to breast cancer risk, the statistical literature on mammogram imaging data is limited. While functional principal component analysis (FPCA) has been studied in the literature for extracting image-based features, it is conducted independently of the time-to-event response variable. With the consideration of building a prognostic model for precision prevention, we present a set of flexible methods, supervised FPCA (sFPCA) and functional partial least squares (FPLS), to extract image-based features associated with the failure time while accommodating the added complication from right censoring. Throughout the article, we hope to demonstrate that one method is favored over the other under different clinical setups. The proposed methods are applied to the motivating data set from the Joanne Knight Breast Health cohort at Siteman Cancer Center. Our approaches not only obtain the best prediction performance compared to the benchmark model, but also reveal different risk patterns within the mammograms.


Asunto(s)
Neoplasias de la Mama , Mamografía , Femenino , Humanos , Neoplasias de la Mama/diagnóstico , Análisis de Componente Principal , Detección Precoz del Cáncer/métodos
4.
Biometrics ; 78(2): 742-753, 2022 06.
Artículo en Inglés | MEDLINE | ID: mdl-33765325

RESUMEN

We develop a Bayesian bivariate spatial model for multivariate regression analysis applicable to studies examining the influence of genetic variation on brain structure. Our model is motivated by an imaging genetics study of the Alzheimer's Disease Neuroimaging Initiative (ADNI), where the objective is to examine the association between images of volumetric and cortical thickness values summarizing the structure of the brain as measured by magnetic resonance imaging (MRI) and a set of 486 single nucleotide polymorphism (SNPs) from 33 Alzheimer's disease (AD) candidate genes obtained from 632 subjects. A bivariate spatial process model is developed to accommodate the correlation structures typically seen in structural brain imaging data. First, we allow for spatial correlation on a graph structure in the imaging phenotypes obtained from a neighborhood matrix for measures on the same hemisphere of the brain. Second, we allow for correlation in the same measures obtained from different hemispheres (left/right) of the brain. We develop a mean-field variational Bayes algorithm and a Gibbs sampling algorithm to fit the model. We also incorporate Bayesian false discovery rate (FDR) procedures to select SNPs. We implement the methodology in a new release of the R package bgsmtr. We show that the new spatial model demonstrates superior performance over a standard model in our application. Data used in the preparation of this article were obtained from the ADNI database (https://adni.loni.usc.edu).


Asunto(s)
Enfermedad de Alzheimer , Enfermedad de Alzheimer/diagnóstico por imagen , Enfermedad de Alzheimer/genética , Enfermedad de Alzheimer/patología , Teorema de Bayes , Encéfalo/diagnóstico por imagen , Encéfalo/patología , Humanos , Imagen por Resonancia Magnética , Neuroimagen
5.
Stat Med ; 41(18): 3547-3560, 2022 08 15.
Artículo en Inglés | MEDLINE | ID: mdl-35574725

RESUMEN

Time-varying biomarkers reflect important information on disease progression over time. Dynamic prediction for event occurrence on a real-time basis, utilizing time-varying information, is crucial in making accurate clinical decisions. Functional principal component analysis (FPCA) has been widely adopted in the literature for extracting features from time-varying biomarker trajectories. However, feature extraction via FPCA is conducted independent of the time-to-event response, which may not produce optimal results when the goal lies in prediction. With this consideration, we propose a novel supervised FPCA, where the functional principal components are determined to optimize the association between the time-varying biomarker and time-to-event outcome. The proposed framework also accommodates irregularly spaced and sparse longitudinal data. Our method is empirically shown to retain better discrimination and calibration performance than the unsupervised FPCA method in simulation studies. Application of the proposed method is also illustrated in the Alzheimer's Disease Neuroimaging Initiative database.


Asunto(s)
Neuroimagen , Biomarcadores/análisis , Progresión de la Enfermedad , Humanos , Análisis de Componente Principal , Análisis de Supervivencia
6.
Stat Med ; 40(11): 2626-2649, 2021 05 20.
Artículo en Inglés | MEDLINE | ID: mdl-33650708

RESUMEN

Unlike chemotherapy, the maximum tolerated dose (MTD) of molecularly targeted agents and immunotherapy may not pose significant clinical benefit over the lower doses. By simultaneously considering both toxicity and efficacy endpoints, phase I/II trials can identify a more clinically meaningful dose for subsequent phase II trials than traditional toxicity-based phase I trials in terms of risk-benefit tradeoff. To strengthen and simplify the current practice of phase I/II trials, we propose a utility-based toxicity probability interval (uTPI) design for finding the optimal biological dose, based on a numerical utility that provides a clinically meaningful, one-dimensional summary representation of the patient's bivariate toxicity and efficacy outcome. The uTPI design does not rely on any parametric specification of the dose-response relationship, and it directly models the dose desirability through a quasi binomial likelihood. Toxicity probability intervals are used to screen out overly toxic dose levels, and then the dose escalation/de-escalation decisions are made adaptively by comparing the posterior desirability distributions of the adjacent levels of the current dose. The uTPI design is flexible in accommodating various dose desirability formulations, while only requiring minimum design parameters. It has a clear decision structure such that a dose-assignment decision table can be calculated before the trial starts and can be used throughout the trial, which simplifies the practical implementation of the design. Extensive simulation studies demonstrate that the proposed uTPI design yields desirable as well as robust performance under various scenarios.


Asunto(s)
Antineoplásicos , Teorema de Bayes , Ensayos Clínicos Fase I como Asunto , Ensayos Clínicos Fase II como Asunto , Simulación por Computador , Relación Dosis-Respuesta a Droga , Humanos , Dosis Máxima Tolerada , Modelos Estadísticos , Probabilidad , Proyectos de Investigación
7.
Stat Med ; 40(3): 712-724, 2021 02 10.
Artículo en Inglés | MEDLINE | ID: mdl-33179286

RESUMEN

In longitudinal studies, the values of biomarkers are often informatively missing due to dropout. The conventional functional principal component analysis typically disregards the missing information and simply treats the unobserved data points as missing completely at random. As a result, the estimation of the mean function and the covariance surface might be biased, resulting in a biased estimation of the functional principal components. We propose the informatively missing functional principal component analysis (imFunPCA), which is well suited for cases where the longitudinal trajectories are subject to informative missingness. Computation of the functional principal components in our approach is based on the likelihood of the data, where information of both the observed and missing data points are incorporated. We adopt a regression-based orthogonal approximation method to decompose the latent stochastic process based on a set of orthonormal empirical basis functions. Under the case of informative missingness, we show via simulation studies that the performance of our approach is superior to that of the conventional ones. We apply our method on a longitudinal dataset of kidney glomerular filtration rates for patients post renal transplantation.


Asunto(s)
Modelos Estadísticos , Humanos , Estudios Longitudinales , Análisis de Componente Principal , Probabilidad , Análisis de Regresión
8.
Stat Appl Genet Mol Biol ; 19(3)2020 08 31.
Artículo en Inglés | MEDLINE | ID: mdl-32866136

RESUMEN

We conduct an imaging genetics study to explore how effective brain connectivity in the default mode network (DMN) may be related to genetics within the context of Alzheimer's disease and mild cognitive impairment. We develop an analysis of longitudinal resting-state functional magnetic resonance imaging (rs-fMRI) and genetic data obtained from a sample of 111 subjects with a total of 319 rs-fMRI scans from the Alzheimer's Disease Neuroimaging Initiative (ADNI) database. A Dynamic Causal Model (DCM) is fit to the rs-fMRI scans to estimate effective brain connectivity within the DMN and related to a set of single nucleotide polymorphisms (SNPs) contained in an empirical disease-constrained set which is obtained out-of-sample from 663 ADNI subjects having only genome-wide data. We relate longitudinal effective brain connectivity estimated using spectral DCM to SNPs using both linear mixed effect (LME) models as well as function-on-scalar regression (FSR). In both cases we implement a parametric bootstrap for testing SNP coefficients and make comparisons with p-values obtained from asymptotic null distributions. In both networks at an initial q-value threshold of 0.1 no effects are found. We report on exploratory patterns of associations with relatively high ranks that exhibit stability to the differing assumptions made by both FSR and LME.


Asunto(s)
Enfermedad de Alzheimer/diagnóstico por imagen , Mapeo Encefálico/métodos , Encéfalo/diagnóstico por imagen , Disfunción Cognitiva/diagnóstico por imagen , Conectoma/métodos , Imagen por Resonancia Magnética/métodos , Anciano , Anciano de 80 o más Años , Enfermedad de Alzheimer/genética , Encéfalo/patología , Disfunción Cognitiva/genética , Bases de Datos Genéticas , Femenino , Humanos , Modelos Lineales , Masculino , Modelos Teóricos , Polimorfismo de Nucleótido Simple
9.
Hum Brain Mapp ; 40(5): 1507-1527, 2019 04 01.
Artículo en Inglés | MEDLINE | ID: mdl-30431208

RESUMEN

When analyzing large multicenter databases, the effects of multiple confounding covariates increase the variability in the data and may reduce the ability to detect changes due to the actual effect of interest, for example, changes due to disease. Efficient ways to evaluate the effect of covariates toward the data harmonization are therefore important. In this article, we showcase techniques to assess the "goodness of harmonization" of covariates. We analyze 7,656 MR images in the multisite, multiscanner Alzheimer's Disease Neuroimaging Initiative (ADNI) database. We present a comparison of three methods for estimating total intracranial volume to assess their robustness and correct the brain structure volumes using the residual method and the proportional (normalization by division) method. We then evaluated the distribution of brain structure volumes over the entire ADNI database before and after accounting for multiple covariates such as total intracranial volume, scanner field strength, sex, and age using two techniques: (a) Zscapes, a panoramic visualization technique to analyze the entire database and (b) empirical cumulative distributions functions. The results from this study highlight the importance of assessing the goodness of data harmonization as a necessary preprocessing step when pooling large data set with multiple covariates, prior to further statistical data analysis.


Asunto(s)
Enfermedad de Alzheimer/diagnóstico por imagen , Encéfalo/diagnóstico por imagen , Anciano , Anciano de 80 o más Años , Envejecimiento , Disfunción Cognitiva/diagnóstico por imagen , Estudios Transversales , Interpretación Estadística de Datos , Bases de Datos Factuales , Progresión de la Enfermedad , Femenino , Humanos , Procesamiento de Imagen Asistido por Computador , Estudios Longitudinales , Imagen por Resonancia Magnética , Masculino , Reproducibilidad de los Resultados , Caracteres Sexuales
10.
Eur Respir J ; 49(2)2017 02.
Artículo en Inglés | MEDLINE | ID: mdl-28179436

RESUMEN

Diffusing capacity of the lung for nitric oxide (DLNO), otherwise known as the transfer factor, was first measured in 1983. This document standardises the technique and application of single-breath DLNO This panel agrees that 1) pulmonary function systems should allow for mixing and measurement of both nitric oxide (NO) and carbon monoxide (CO) gases directly from an inspiratory reservoir just before use, with expired concentrations measured from an alveolar "collection" or continuously sampled via rapid gas analysers; 2) breath-hold time should be 10 s with chemiluminescence NO analysers, or 4-6 s to accommodate the smaller detection range of the NO electrochemical cell; 3) inspired NO and oxygen concentrations should be 40-60 ppm and close to 21%, respectively; 4) the alveolar oxygen tension (PAO2 ) should be measured by sampling the expired gas; 5) a finite specific conductance in the blood for NO (θNO) should be assumed as 4.5 mL·min-1·mmHg-1·mL-1 of blood; 6) the equation for 1/θCO should be (0.0062·PAO2 +1.16)·(ideal haemoglobin/measured haemoglobin) based on breath-holding PAO2 and adjusted to an average haemoglobin concentration (male 14.6 g·dL-1, female 13.4 g·dL-1); 7) a membrane diffusing capacity ratio (DMNO/DMCO) should be 1.97, based on tissue diffusivity.


Asunto(s)
Volumen Sanguíneo , Óxido Nítrico/sangre , Alveolos Pulmonares/irrigación sanguínea , Capacidad de Difusión Pulmonar/normas , Adolescente , Adulto , Anciano , Anciano de 80 o más Años , Permeabilidad Capilar , Monóxido de Carbono/sangre , Femenino , Hemoglobinas/análisis , Humanos , Modelos Lineales , Masculino , Persona de Mediana Edad , Oxígeno/sangre , Adulto Joven
11.
Biometrics ; 73(3): 949-959, 2017 09.
Artículo en Inglés | MEDLINE | ID: mdl-28076654

RESUMEN

Partial differential equations (PDEs) are used to model complex dynamical systems in multiple dimensions, and their parameters often have important scientific interpretations. In some applications, PDE parameters are not constant but can change depending on the values of covariates, a feature that we call varying coefficients. We propose a parameter cascading method to estimate varying coefficients in PDE models from noisy data. Our estimates of the varying coefficients are shown to be consistent and asymptotically normally distributed. The performance of our method is evaluated by a simulation study and by an empirical study estimating three varying coefficients in a PDE model arising from LIDAR data.


Asunto(s)
Modelos Teóricos , Simulación por Computador
12.
Biometrics ; 73(4): 1231-1242, 2017 12.
Artículo en Inglés | MEDLINE | ID: mdl-28369708

RESUMEN

The problem of modeling the dynamical regulation process within a gene network has been of great interest for a long time. We propose to model this dynamical system with a large number of nonlinear ordinary differential equations (ODEs), in which the regulation function is estimated directly from data without any parametric assumption. Most current research assumes the gene regulation network is static, but in reality, the connection and regulation function of the network may change with time or environment. This change is reflected in our dynamical model by allowing the regulation function varying with the gene expression and forcing this regulation function to be zero if no regulation happens. We introduce a statistical method called functional SCAD to estimate a time-varying sparse and directed gene regulation network, and simultaneously, to provide a smooth estimation of the regulation function and identify the interval in which no regulation effect exists. The finite sample performance of the proposed method is investigated in a Monte Carlo simulation study. Our method is demonstrated by estimating a time-varying directed gene regulation network of 20 genes involved in muscle development during the embryonic stage of Drosophila melanogaster.


Asunto(s)
Redes Reguladoras de Genes , Método de Montecarlo , Dinámicas no Lineales , Animales , Drosophila melanogaster/embriología , Desarrollo de Músculos , Factores de Tiempo
13.
Biometrics ; 73(3): 802-810, 2017 09.
Artículo en Inglés | MEDLINE | ID: mdl-28295173

RESUMEN

Functional principal component analysis (FPCA) is a popular approach in functional data analysis to explore major sources of variation in a sample of random curves. These major sources of variation are represented by functional principal components (FPCs). Most existing FPCA approaches use a set of flexible basis functions such as B-spline basis to represent the FPCs, and control the smoothness of the FPCs by adding roughness penalties. However, the flexible representations pose difficulties for users to understand and interpret the FPCs. In this article, we consider a variety of applications of FPCA and find that, in many situations, the shapes of top FPCs are simple enough to be approximated using simple parametric functions. We propose a parametric approach to estimate the top FPCs to enhance their interpretability for users. Our parametric approach can also circumvent the smoothing parameter selecting process in conventional nonparametric FPCA methods. In addition, our simulation study shows that the proposed parametric FPCA is more robust when outlier curves exist. The parametric FPCA method is demonstrated by analyzing several datasets from a variety of applications.


Asunto(s)
Análisis de Componente Principal
14.
Biometrics ; 72(3): 846-54, 2016 09.
Artículo en Inglés | MEDLINE | ID: mdl-26683051

RESUMEN

Functional principal component analysis (FPCA) is a popular approach to explore major sources of variation in a sample of random curves. These major sources of variation are represented by functional principal components (FPCs). The intervals where the values of FPCs are significant are interpreted as where sample curves have major variations. However, these intervals are often hard for naïve users to identify, because of the vague definition of "significant values". In this article, we develop a novel penalty-based method to derive FPCs that are only nonzero precisely in the intervals where the values of FPCs are significant, whence the derived FPCs possess better interpretability than the FPCs derived from existing methods. To compute the proposed FPCs, we devise an efficient algorithm based on projection deflation techniques. We show that the proposed interpretable FPCs are strongly consistent and asymptotically normal under mild conditions. Simulation studies confirm that with a competitive performance in explaining variations of sample curves, the proposed FPCs are more interpretable than the traditional counterparts. This advantage is demonstrated by analyzing two real datasets, namely, electroencephalography data and Canadian weather data.


Asunto(s)
Modelos Estadísticos , Análisis de Componente Principal/métodos , Algoritmos , Canadá , Simulación por Computador , Electroencefalografía/estadística & datos numéricos , Humanos , Tiempo (Meteorología)
15.
Biometrics ; 71(1): 131-138, 2015 Mar.
Artículo en Inglés | MEDLINE | ID: mdl-25287611

RESUMEN

We consider model selection and estimation in a context where there are competing ordinary differential equation (ODE) models, and all the models are special cases of a "full" model. We propose a computationally inexpensive approach that employs statistical estimation of the full model, followed by a combination of a least squares approximation (LSA) and the adaptive Lasso. We show the resulting method, here called the LSA method, to be an (asymptotically) oracle model selection method. The finite sample performance of the proposed LSA method is investigated with Monte Carlo simulations, in which we examine the percentage of selecting true ODE models, the efficiency of the parameter estimation compared to simply using the full and true models, and coverage probabilities of the estimated confidence intervals for ODE parameters, all of which have satisfactory performances. Our method is also demonstrated by selecting the best predator-prey ODE to model a lynx and hare population dynamical system among some well-known and biologically interpretable ODE models.


Asunto(s)
Biometría/métodos , Lynx/fisiología , Modelos Biológicos , Modelos Estadísticos , Conducta Predatoria/fisiología , Conejos/fisiología , Algoritmos , Animales , Simulación por Computador , Humanos
16.
Stat Med ; 32(15): 2681-94, 2013 Jul 10.
Artículo en Inglés | MEDLINE | ID: mdl-23172783

RESUMEN

Emergency department access block is an urgent problem faced by many public hospitals today. When access block occurs, patients in need of acute care cannot access inpatient wards within an optimal time frame. A widely held belief is that access block is the end product of a long causal chain, which involves poor discharge planning, insufficient bed capacity, and inadequate admission intensity to the wards. This paper studies the last link of the causal chain-the effect of admission intensity on access block, using data from a metropolitan hospital in Australia. We applied several modern statistical methods to analyze the data. First, we modeled the admission events as a nonhomogeneous Poisson process and estimated time-varying admission intensity with penalized regression splines. Next, we established a functional linear model to investigate the effect of the time-varying admission intensity on emergency department access block. Finally, we used functional principal component analysis to explore the variation in the daily time-varying admission intensities. The analyses suggest that improving admission practice during off-peak hours may have most impact on reducing the number of ED access blocks.


Asunto(s)
Bioestadística/métodos , Servicio de Urgencia en Hospital/estadística & datos numéricos , Admisión del Paciente/estadística & datos numéricos , Australia , Accesibilidad a los Servicios de Salud , Hospitales Urbanos , Humanos , Funciones de Verosimilitud , Modelos Lineales , Modelos Estadísticos , Distribución de Poisson , Análisis de Componente Principal , Factores de Tiempo
17.
Stat Comput ; 33(1): 1, 2023.
Artículo en Inglés | MEDLINE | ID: mdl-36415568

RESUMEN

The selection of smoothing parameter is central to the estimation of penalized splines. The best value of the smoothing parameter is often the one that optimizes a smoothness selection criterion, such as generalized cross-validation error (GCV) and restricted likelihood (REML). To correctly identify the global optimum rather than being trapped in an undesired local optimum, grid search is recommended for optimization. Unfortunately, the grid search method requires a pre-specified search interval that contains the unknown global optimum, yet no guideline is available for providing this interval. As a result, practitioners have to find it by trial and error. To overcome such difficulty, we develop novel algorithms to automatically find this interval. Our automatic search interval has four advantages. (i) It specifies a smoothing parameter range where the associated penalized least squares problem is numerically solvable. (ii) It is criterion-independent so that different criteria, such as GCV and REML, can be explored on the same parameter range. (iii) It is sufficiently wide to contain the global optimum of any criterion, so that for example, the global minimum of GCV and the global maximum of REML can both be identified. (iv) It is computationally cheap compared with the grid search itself, carrying no extra computational burden in practice. Our method is ready to use through our recently developed R package gps ( ≥  version 1.1). It may be embedded in more advanced statistical modeling methods that rely on penalized splines. Supplementary Information: The online version contains supplementary material available at 10.1007/s11222-022-10178-z.

18.
Stat Methods Med Res ; 32(5): 895-903, 2023 05.
Artículo en Inglés | MEDLINE | ID: mdl-36951095

RESUMEN

Screening mammography is the primary preventive strategy for early detection of breast cancer and an essential input to breast cancer risk prediction and application of prevention/risk management guidelines. Identifying regions of interest within mammogram images that are associated with 5- or 10-year breast cancer risk is therefore clinically meaningful. The problem is complicated by the irregular boundary issue posed by the semi-circular domain of the breast area within mammograms. Accommodating the irregular domain is especially crucial when identifying regions of interest, as the true signal comes only from the semi-circular domain of the breast region, and noise elsewhere. We address these challenges by introducing a proportional hazards model with imaging predictors characterized by bivariate splines over triangulation. The model sparsity is enforced with the group lasso penalty function. We apply the proposed method to the motivating Joanne Knight Breast Health Cohort to illustrate important risk patterns and show that the proposed method is able to achieve higher discriminatory performance.


Asunto(s)
Neoplasias de la Mama , Mamografía , Humanos , Femenino , Mamografía/métodos , Neoplasias de la Mama/diagnóstico , Detección Precoz del Cáncer , Riesgo , Modelos de Riesgos Proporcionales
19.
J Appl Stat ; 50(1): 43-59, 2023.
Artículo en Inglés | MEDLINE | ID: mdl-36530777

RESUMEN

In many clinical studies, longitudinal biomarkers are often used to monitor the progression of a disease. For example, in a kidney transplant study, the glomerular filtration rate (GFR) is used as a longitudinal biomarker to monitor the progression of the kidney function and the patient's state of survival that is characterized by multiple time-to-event outcomes, such as kidney transplant failure and death. It is known that the joint modelling of longitudinal and survival data leads to a more accurate and comprehensive estimation of the covariates' effect. While most joint models use the longitudinal outcome as a covariate for predicting survival, very few models consider the further decomposition of the variation within the longitudinal trajectories and its effect on survival. We develop a joint model that uses functional principal component analysis (FPCA) to extract useful features from the longitudinal trajectories and adopt the competing risk model to handle multiple time-to-event outcomes. The longitudinal trajectories and the multiple time-to-event outcomes are linked via the shared functional features. The application of our model on a real kidney transplant data set reveals the significance of these functional features, and a simulation study is carried out to validate the accurateness of the estimation method.

20.
J Agric Biol Environ Stat ; 28(1): 43-58, 2023.
Artículo en Inglés | MEDLINE | ID: mdl-36065440

RESUMEN

We address two computational issues common to open-population N-mixture models, hidden integer-valued autoregressive models, and some hidden Markov models. The first issue is computation time, which can be dramatically improved through the use of a fast Fourier transform. The second issue is tractability of the model likelihood function for large numbers of hidden states, which can be solved by improving numerical stability of calculations. As an illustrative example, we detail the application of these methods to the open-population N-mixture models. We compare computational efficiency and precision between these methods and standard methods employed by state-of-the-art ecological software. We show faster computing times (a ∼ 6 to ∼ 30 times speed improvement for population size upper bounds of 500 and 1000, respectively) over state-of-the-art ecological software for N-mixture models. We also apply our methods to compute the size of a large elk population using an N-mixture model and show that while our methods converge, previous software cannot produce estimates due to numerical issues. These solutions can be applied to many ecological models to improve precision when logs of sums exist in the likelihood function and to improve computational efficiency when convolutions are present in the likelihood function. Supplementary materials accompanying this paper appear online. Supplementary materials for this article are available at 10.1007/s13253-022-00509-y.

SELECCIÓN DE REFERENCIAS
DETALLE DE LA BÚSQUEDA