Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 1 de 1
Filtrar
Más filtros

Banco de datos
Tipo de estudio
Tipo del documento
Intervalo de año de publicación
1.
IEEE Trans Cybern ; 52(5): 2994-3005, 2022 May.
Artículo en Inglés | MEDLINE | ID: mdl-33027021

RESUMEN

Fuzzy-rough cognitive networks (FRCNs) are recurrent neural networks (RNNs) intended for structured classification purposes in which the problem is described by an explicit set of features. The advantage of this granular neural system relies on its transparency and simplicity while being competitive to state-of-the-art classifiers. Despite their relative empirical success in terms of prediction rates, there are limited studies on FRCNs' dynamic properties and how their building blocks contribute to the algorithm's performance. In this article, we theoretically study these issues and conclude that boundary and negative neurons always converge to a unique fixed-point attractor. Moreover, we demonstrate that negative neurons have no impact on the algorithm's performance and that the ranking of positive neurons is invariant. Moved by our theoretical findings, we propose two simpler fuzzy-rough classifiers that overcome the detected issues and maintain the competitive prediction rates of this classifier. Toward the end, we present a case study concerned with image classification, in which a convolutional neural network is coupled with one of the simpler models derived from the theoretical analysis of the FRCN model. The numerical simulations suggest that once the features have been extracted, our granular neural system performs as well as other RNNs.


Asunto(s)
Lógica Difusa , Redes Neurales de la Computación , Cognición , Modelos Teóricos , Neuronas
SELECCIÓN DE REFERENCIAS
DETALLE DE LA BÚSQUEDA