RESUMEN
Cattails (Typha latifolia L.) are naturally occurring aquatic macrophytes with significant industrial potential because of their abundance, high-quality fibers, and high fiber yields. This study is the first attempt to investigate how phenological development and plant maturity impact the quality of cattail fibers as they relate to composite applications. It was observed that fibers from all five growth stages exhibited a Weibull shape parameter greater than 1.0, with a goodness-of-fit exceeding 0.8. These calculations were performed using both the Least Square Regression (LSR) and Maximum Likelihood Estimation (MLE) methods. Among the estimators, the MLE method provided the most conservative estimation of Weibull parameters. Based on the Weibull parameters obtained with all estimators, cattail fibers from all five growth stages appear suitable for composite applications. The consistency of shape parameters across all five growth stages can be attributed to the morphological and molecular developments of cattail fiber during the vegetative period. These developments were confirmed through the presence of calcium oxalate (CaOx) plates, elemental composition, and specific infrared peaks at 2360 cm-1 contributing to the strength, cellulose peaks at 1635 cm-1, 2920 cm-1, and 3430 cm-1. In conclusion, it was found that the mechanical properties of cattail fiber remain similar when harvested multiple times in a single growing season.
RESUMEN
One essential initial step in the analysis of ancient DNA is to authenticate that the DNA sequencing reads are actually from ancient DNA. This is done by assessing if the reads exhibit typical characteristics of post-mortem damage (PMD), including cytosine deamination and nicks. We present a novel statistical method implemented in a fast multithreaded programme, ngsBriggs that enables rapid quantification of PMD by estimation of the Briggs ancient damage model parameters (Briggs parameters). Using a multinomial model with maximum likelihood fit, ngsBriggs accurately estimates the parameters of the Briggs model, quantifying the PMD signal from single and double-stranded DNA regions. We extend the original Briggs model to capture PMD signals for contemporary sequencing platforms and show that ngsBriggs accurately estimates the Briggs parameters across a variety of contamination levels. Classification of reads into ancient or modern reads, for the purpose of decontamination, is significantly more accurate using ngsBriggs than using other methods available. Furthermore, ngsBriggs is substantially faster than other state-of-the-art methods. ngsBriggs offers a practical and accurate method for researchers seeking to authenticate ancient DNA and improve the quality of their data.
RESUMEN
An alternative closed-form expression for the marginal joint probability distribution of item scores under the random effects generalized partial credit model is presented. The closed-form expression involves a cumulant generating function and is therefore subjected to convexity constraints. As a consequence, complicated moment inequalities are taken into account in maximum likelihood estimation of the parameters of the model, so that the estimation solution is always proper. Another important favorable consequence is that the likelihood function has a single local extreme point, the global maximum. Furthermore, attention is paid to expected a posteriori person parameter estimation, generalizations of the model, and testing the goodness-of-fit of the model. Procedures proposed are demonstrated in an illustrative example.
RESUMEN
A very useful modification to ranked set sampling (RSS) that allows a larger set size without significantly increasing ranking errors is the maximum ranked set sampling with unequal samples (MRSSU) approach. This article covers the parameter estimation of the inverted Kumaraswamy distribution using MRSSU and RSS designs. The maximum likelihood and Bayesian estimation techniques are considered. The regarded Bayesian estimation technique is determined in the case of non-informative and informative priors represented by Jeffreys and gamma priors, respectively. Squared error and minimum expected are the two loss functions that are employed. We presented a simulation study to evaluate the performance of the recommended estimations using root mean squared error and relative bias. The Bayes point estimates were computed using the Metropolis-Hastings algorithm. Additional conclusions have been made based on actual geological data regarding the intervals between Kiama Blowhole's 64 consecutive eruptions. Based on the same number of measured units, the results of simulation and real data analysis showed that MRSSU estimators performed much better than their RSS counterparts.
RESUMEN
Multistate transition models (MSTMs) are valuable tools depicting disease progression. However, due to the complexity of MSTMs, larger sample size and longer follow-up time in real-world data, the computation of statistical estimation and inference for MSTMs becomes challenging. A bounded Taylor series in Newton-Raphson procedure is proposed which leverages the uniformization technique to derive maximum likelihood estimates and corresponding covariance matrix. The proposed method, namely uniformization Taylor-bounded Newton-Raphson, is validated in three simulation studies, which demonstrate the accuracy in parameter estimation, the efficiency in computation time and robustness in terms of different situations. This method is also illustrated using a large electronic medical record data related to statin-induced side effects and discontinuation.
RESUMEN
One basic limitation of using the periodogram as a frequency estimator is that any of its significant peaks may result from a diffuse (or spread) frequency component rather than a pure one. Diffuse components are common in applications such as channel estimation, in which a given periodogram peak reveals the presence of a complex multipath distribution (unresolvable propagation paths or diffuse scattering, for example). We present a method to detect the presence of a diffuse component in a given peak based on analyzing the projection of the data vector onto the span of the signature's derivatives up to a given order. Fundamentally, a diffuse component is detected if the energy in the derivatives' subspace is too high at the peak's frequency, and its spread is estimated as the ratio between this last energy and the peak's energy. The method is based on exploiting the signature's Vandermonde structure through the properties of discrete Chebyshev polynomials. We also present an efficient numerical procedure for computing the data component in the derivatives' span based on barycentric interpolation. The paper contains a numerical assessment of the proposed estimator and detector.
RESUMEN
Partial accelerated life tests (PALTs) are employed when the results of accelerated life testing cannot be extended to usage circumstances. This work discusses the challenge of different estimating strategies in constant PALT with complete data. The lifetime distribution of the test item is assumed to follow the power half-logistic distribution. Several classical and Bayesian estimation techniques are presented to estimate the distribution parameters and the acceleration factor of the power half-logistic distribution. These techniques include Anderson-Darling, maximum likelihood, Cramér von-Mises, ordinary least squares, weighted least squares, maximum product of spacing and Bayesian. Additionally, the Bayesian credible intervals and approximate confidence intervals are constructed. A simulation study is provided to compare the outcomes of various estimation methods that have been provided based on mean squared error, absolute average bias, length of intervals, and coverage probabilities. This study shows that the maximum product of spacing estimation is the most effective strategy among the options in most circumstances when adopting the minimum values for MSE and average bias. In the majority of situations, Bayesian method outperforms other methods when taking into account both MSE and average bias values. When comparing approximation confidence intervals to Bayesian credible intervals, the latter have a higher coverage probability and smaller average length. Two authentic data sets are examined for illustrative purposes. Examining the two real data sets shows that the value methods are workable and applicable to certain engineering-related problems.
RESUMEN
This paper presents the exponentiated alpha-power log-logistic (EAPLL) distribution, which extends the log-logistic distribution. The EAPLL distribution emphasizes its suitability for survival data modeling by providing analytical simplicity and accommodating both monotone and non-monotone failure rates. We derive some of its mathematical properties and test eight estimation methods using an extensive simulation study. To determine the best estimation approach, we rank mean estimates, mean square errors, and average absolute biases on a partial and overall ranking. Furthermore, we use the EAPLL distribution to examine three real-life survival data sets, demonstrating its superior performance over competing log-logistic distributions. This study adds vital insights to survival analysis methodology and provides a solid framework for modeling various survival data scenarios.
RESUMEN
In recent decades, many research studies have been conducted for the development of some new modifications of different baseline distributions to cope with real-world problems. This paper proposes a novel generalization of probability distributions, called modified type- II half-logistic distribution. We have derived the new family of probability distributions using T-X family with input as a type II variant of the half-logistic distribution. For the purpose of demonstration, the Weibull distribution is considered as a sub-model. Various algebraic properties of the suggested distribution have been discussed. For efficient parameter estimation, we have used the maximum likelihood principle. Additionally, two real-world data sets from the literature have been considered to illustrate the practical usefulness and significance of the suggested model.
RESUMEN
Atmospheric phase error is the main factor affecting the accuracy of ground-based synthetic aperture radar (GB-SAR). The atmospheric phase screen (APS) may be very complicated, so the atmospheric phase correction (APC) model is very important; in particular, the parameters to be estimated in the model are the key to improving the accuracy of APC. However, the conventional APC method first performs phase unwrapping and then removes the APS based on the least-squares method (LSM), and the general phase unwrapping method is prone to introducing unwrapping error. In particular, the LSM is difficult to apply directly due to the phase wrapping of permanent scatterers (PSs). Therefore, a novel methodology for estimating parameters of the APC model based on the maximum likelihood estimation (MLE) and the Gauss-Newton algorithm is proposed in this paper, which first introduces the MLE method to provide a suitable objective function for the parameter estimation of nonlinear far-end and near-end correction models. Then, based on the Gauss-Newton algorithm, the parameters of the objective function are iteratively estimated with suitable initial values, and the Matthews and Davies algorithm is used to optimize the Gauss-Newton algorithm to improve the accuracy of parameter estimation. Finally, the parameter estimation performance is evaluated based on Monte Carlo simulation experiments. The method proposed in this paper experimentally verifies the feasibility and superiority, which avoids phase unwrapping processing unlike the conventional method.
RESUMEN
Antimicrobial heteroresistance refers to the presence of different subpopulations with heterogeneous antimicrobial responses within the same bacterial isolate, so they show reduced susceptibility compared with the main population. Though it is widely accepted that heteroresistance can play a crucial role in the outcome of antimicrobial treatments, predictive Antimicrobial Resistance (AMR) models accounting for bacterial heteroresistance are still scarce and need to be refined as the techniques to measure heteroresistance become standardised and consistent conclusions are drawn from data. In this work, we propose a multivariate Birth-Death (BD) model of bacterial heteroresistance and analyse its properties in detail. Stochasticity in the population dynamics is considered since heteroresistance is often characterised by low initial frequencies of the less susceptible subpopulations, those mediating AMR transmission and potentially leading to treatment failure. We also discuss the utility of the heteroresistance model for practical applications and calibration under realistic conditions, demonstrating that it is possible to infer the model parameters and heteroresistance distribution from time-kill data, i.e., by measuring total cell counts alone and without performing any heteroresistance test.
Asunto(s)
Farmacorresistencia Bacteriana , Modelos Biológicos , Antibacterianos/farmacología , Bacterias/efectos de los fármacos , Pruebas de Sensibilidad Microbiana/estadística & datos numéricos , Procesos Estocásticos , HumanosRESUMEN
Crashes are frequently disproportionally observed in disadvantaged areas. Despite the evident disparities in transportation safety, there has been limited exploration of quantitative approaches to incorporating equity considerations into road safety management. This study proposes a novel concept of equity-aware safety performance functions (SPFs), enabling a distinct treatment of equity-related variables such as race and income. Equity-aware SPFs introduce a fairness distance and integrate it into the log-likelihood function of the negative binomial regression as a form of partial lasso regularization. A parameter λ is used to control the importance of the regularization term. Equity-aware SPFs are developed for pedestrian-involved crashes at the census tract level in Virginia, USA, and then employed to compute the potential for safety improvement (PSI), a prevalent metric used in hotspot identification. Results show that equity-aware SPFs can diminish the effects of equity-related variables, including poverty ratio, black ratio, Asian ratio, and the ratio of households without vehicles, on the expected crash frequencies, generating higher PSIs for disadvantaged areas. Based on the results of Wilcoxon signed-rank tests, it is evident that there are significant differences in the rankings of PSIs when equity awareness is considered, especially for disadvantaged areas. This study adds to the literature a new quantitative approach to harmonize equity and effectiveness considerations, empowering more equitable decision-making in safety management, such as allocating resources for safety enhancement.
Asunto(s)
Accidentes de Tránsito , Peatones , Seguridad , Humanos , Accidentes de Tránsito/prevención & control , Accidentes de Tránsito/estadística & datos numéricos , Peatones/estadística & datos numéricos , Virginia , Funciones de Verosimilitud , Poblaciones Vulnerables , Administración de la Seguridad , RentaRESUMEN
Parameter inference and uncertainty quantification are important steps when relating mathematical models to real-world observations and when estimating uncertainty in model predictions. However, methods for doing this can be computationally expensive, particularly when the number of unknown model parameters is large. The aim of this study is to develop and test an efficient profile likelihood-based method, which takes advantage of the structure of the mathematical model being used. We do this by identifying specific parameters that affect model output in a known way, such as a linear scaling. We illustrate the method by applying it to three toy models from different areas of the life sciences: (i) a predator-prey model from ecology; (ii) a compartment-based epidemic model from health sciences; and (iii) an advection-diffusion reaction model describing the transport of dissolved solutes from environmental science. We show that the new method produces results of comparable accuracy to existing profile likelihood methods but with substantially fewer evaluations of the forward model. We conclude that our method could provide a much more efficient approach to parameter inference for models where a structured approach is feasible. Computer code to apply the new method to user-supplied models and data is provided via a publicly accessible repository.
RESUMEN
Interval-censored failure time data frequently arise in various scientific studies where each subject experiences periodical examinations for the occurrence of the failure event of interest, and the failure time is only known to lie in a specific time interval. In addition, collected data may include multiple observed variables with a certain degree of correlation, leading to severe multicollinearity issues. This work proposes a factor-augmented transformation model to analyze interval-censored failure time data while reducing model dimensionality and avoiding multicollinearity elicited by multiple correlated covariates. We provide a joint modeling framework by comprising a factor analysis model to group multiple observed variables into a few latent factors and a class of semiparametric transformation models with the augmented factors to examine their and other covariate effects on the failure event. Furthermore, we propose a nonparametric maximum likelihood estimation approach and develop a computationally stable and reliable expectation-maximization algorithm for its implementation. We establish the asymptotic properties of the proposed estimators and conduct simulation studies to assess the empirical performance of the proposed method. An application to the Alzheimer's Disease Neuroimaging Initiative (ADNI) study is provided. An R package ICTransCFA is also available for practitioners. Data used in preparation of this article were obtained from the ADNI database.
Asunto(s)
Enfermedad de Alzheimer , Simulación por Computador , Modelos Estadísticos , Humanos , Funciones de Verosimilitud , Algoritmos , Neuroimagen , Análisis Factorial , Interpretación Estadística de Datos , Factores de TiempoRESUMEN
As the population of Somaliland continues to grow rapidly, the demand for electricity is anticipated to rise exponentially over the next few decades. The provision of reliable and cost-effective electricity service is at the core of the economic and social development of Somaliland. Wind energy might offer a sustainable solution to the exceptionally high electricity prices. In this study, a techno-economic assessment of the wind energy potential in some parts of the western region of Somaliland is performed. Measured data of wind speed and wind direction for three sites around the capital city of Hargeisa are utilized to characterize the resource using Weibull distribution functions. Technical and economic performances of several commercial wind turbines are examined. Out of the three sites, Xumba Weyne stands out as the most favorable site for wind energy harnessing with average annual power and energy densities at 80 m hub height of 317 kW/m2 and 2782 kWh/m2, respectively. Wind turbines installed in Xumba Weyne yielded the lowest levelized cost of electricity (LCOE) of not more than 0.07 $/kWh, shortest payback times (i.e., less than 7.2 years) with minimum return on investment (ROI) of approximately 150%.
RESUMEN
When software systems are introduced, they are typically deployed in field environments similar to those used during development and testing. However, these systems may also be used in various other locations with different environmental conditions, making it challenging to improve software reliability. Factors such as the specific operating environment and the location of bugs in the code contribute to this difficulty. In this paper, we propose a new software reliability model that accounts for the uncertainty of operating environments. We present the explicit closed-form mean value function solution for the proposed model. The model's goodness of fit is demonstrated by comparing it to the nonhomogeneous Poisson process (NHPP) model based on Weibull model, using four sets of failure data sets from software applications. The proposed model performs well under various estimation techniques, making it a versatile tool for practitioners and researchers alike. The proposed model outperforms other existing NHPP Weibull based in terms of fitting accuracy under two different methods of estimation and provides a more detailed and precise evaluation of software reliability. Additionally, sensitivity analysis shows that the parameters of the suggested distribution significantly impact the mean value function.
RESUMEN
In competing risks data, in practice, there may be lack of information or uncertainty about the true failure type, termed as 'missing failure type', for some subjects. We consider a general pattern of missing failure type in which we observe, if not the true failure type, a set of possible failure types containing the true one. In this work, we focus on both parametric and non-parametric estimation based on current status data with two competing risks and the above-mentioned missing failure type. Here, the missing probabilities are assumed to be time-dependent, that is, dependent on both failure and monitoring time points, in addition to being dependent on the true failure type. This makes the missing mechanism non-ignorable. We carry out maximum likelihood estimation and obtain the asymptotic properties of the estimators. Simulation studies are conducted to investigate the finite sample properties of the estimators. Finally, the methods are illustrated through a data set on hearing loss.
RESUMEN
A scrutiny analysis of the COVID-19 data is required to get insights into effective strategies for pandemic control. However, there is a gap between official data and methods used to assess the effectiveness of the potential measures, which was partly addressed in an editorial-letter-type discussion on the impact of the COVID-19 passport in Lithuania. The therein-applied descriptive statistics method provides only limited evidence, while detailed analysis requires more sensitive and reliable methods. In this regard, this paper advocates a maximum likelihood compartmental modeling approach, which provides the flexibility to raise various hypotheses about infection, recovery, and mortality dynamics and to find the most likely answers given the data. Our paper is based on COVID-19 deaths, which are more reliable and essential than infection cases. It should also be noted that officially collected data are unsuitable for in-depth analyses, including compartmental modeling, as they do not capture important information. Overall, this paper does not aim to solve the underlying problems completely but rather stimulate a discussion.
RESUMEN
We introduce the bivariate unit-log-symmetric model based on the bivariate log-symmetric distribution (BLS) defined in Vila et al. [25] as a flexible family of bivariate distributions over the unit square. We then study its mathematical properties such as stochastic representations, quantiles, conditional distributions, independence of the marginal distributions and marginal moments. Maximum likelihood estimation method is discussed and examined through Monte Carlo simulation. Finally, the proposed model is used to analyze some soccer data sets.
RESUMEN
Bounded count response data arise naturally in health applications. In general, the well-known beta-binomial regression model form the basis for analyzing this data, specially when we have overdispersed data. Little attention, however, has been given to the literature on the possibility of having extreme observations and overdispersed data. We propose in this work an extension of the beta-binomial regression model, named the beta-2-binomial regression model, which provides a rather flexible approach for fitting a regression model with a wide spectrum of bounded count response data sets under the presence of overdispersion, outliers, or excess of extreme observations. This distribution possesses more skewness and kurtosis than the beta-binomial model but preserves the same mean and variance form of the beta-binomial model. Additional properties of the beta-2-binomial distribution are derived including its behavior on the limits of its parametric space. A penalized maximum likelihood approach is considered to estimate parameters of this model and a residual analysis is included to assess departures from model assumptions as well as to detect outlier observations. Simulation studies, considering the robustness to outliers, are presented confirming that the beta-2-binomial regression model is a better robust alternative, in comparison with the binomial and beta-binomial regression models. We also found that the beta-2-binomial regression model outperformed the binomial and beta-binomial regression models in our applications of predicting liver cancer development in mice and the number of inappropriate days a patient spent in a hospital.