Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 2 de 2
Filtrar
Más filtros

Bases de datos
Tipo del documento
Asunto de la revista
País de afiliación
Intervalo de año de publicación
1.
Comput Biol Med ; 37(11): 1553-64, 2007 Nov.
Artículo en Inglés | MEDLINE | ID: mdl-17434159

RESUMEN

We evaluate the effectiveness of seven Bayesian network classifiers as potential tools for the diagnosis of breast cancer using two real-world databases containing fine-needle aspiration of the breast lesion cases collected by a single observer and multiple observers, respectively. The results show a certain ingredient of subjectivity implicitly contained in these data: we get an average accuracy of 93.04% for the former and 83.31% for the latter. These findings suggest that observers see different things when looking at the samples in the microscope; a situation that significantly diminishes the performance of these classifiers in diagnosing such a disease.


Asunto(s)
Neoplasias de la Mama/diagnóstico , Diagnóstico por Computador , Algoritmos , Teorema de Bayes , Biopsia con Aguja Fina , Citodiagnóstico/estadística & datos numéricos , Bases de Datos Factuales , Diagnóstico por Computador/estadística & datos numéricos , Femenino , Humanos , Variaciones Dependientes del Observador
2.
PLoS One ; 9(3): e92866, 2014.
Artículo en Inglés | MEDLINE | ID: mdl-24671204

RESUMEN

The bias-variance dilemma is a well-known and important problem in Machine Learning. It basically relates the generalization capability (goodness of fit) of a learning method to its corresponding complexity. When we have enough data at hand, it is possible to use these data in such a way so as to minimize overfitting (the risk of selecting a complex model that generalizes poorly). Unfortunately, there are many situations where we simply do not have this required amount of data. Thus, we need to find methods capable of efficiently exploiting the available data while avoiding overfitting. Different metrics have been proposed to achieve this goal: the Minimum Description Length principle (MDL), Akaike's Information Criterion (AIC) and Bayesian Information Criterion (BIC), among others. In this paper, we focus on crude MDL and empirically evaluate its performance in selecting models with a good balance between goodness of fit and complexity: the so-called bias-variance dilemma, decomposition or tradeoff. Although the graphical interaction between these dimensions (bias and variance) is ubiquitous in the Machine Learning literature, few works present experimental evidence to recover such interaction. In our experiments, we argue that the resulting graphs allow us to gain insights that are difficult to unveil otherwise: that crude MDL naturally selects balanced models in terms of bias-variance, which not necessarily need be the gold-standard ones. We carry out these experiments using a specific model: a Bayesian network. In spite of these motivating results, we also should not overlook three other components that may significantly affect the final model selection: the search procedure, the noise rate and the sample size.


Asunto(s)
Algoritmos , Sesgo , Teorema de Bayes , Bases de Datos como Asunto , Probabilidad
SELECCIÓN DE REFERENCIAS
DETALLE DE LA BÚSQUEDA