Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 3 de 3
Filtrar
Mais filtros

Base de dados
Ano de publicação
Tipo de documento
Assunto da revista
País de afiliação
Intervalo de ano de publicação
1.
Artigo em Inglês | MEDLINE | ID: mdl-36409810

RESUMO

Multivariate or multidimensional visualization plays an essential role in exploratory data analysis by allowing users to derive insights and formulate hypotheses. Despite their popularity, it is usually users' responsibility to (visually) discover the data patterns, which can be cumbersome and time-consuming. Visual Analytics (VA) and machine learning techniques can be instrumental in mitigating this problem by automatically discovering and representing such patterns. One example is the integration of classification models with (visual) interpretability strategies, where models are used as surrogates for data patterns so that understanding a model enables understanding the phenomenon represented by the data. Although useful and inspiring, the few proposed solutions are based on visual representations of so-called black-box models, so the interpretation of the patterns captured by the models is not straightforward, requiring mechanisms to transform them into human-understandable pieces of information. This paper presents multiVariate dAta eXplanation (VAX), a new VA method to support identifying and visual interpreting patterns in multivariate datasets. Unlike the existing similar approaches, VAX uses the concept of Jumping Emerging Patterns, inherent interpretable logic statements representing class-variable relationships (patterns) derived from random Decision Trees. The potential of VAX is shown through use cases employing two real-world datasets covering different scenarios where intricate patterns are discovered and represented, something challenging to be done using usual exploratory approaches.

2.
Talanta ; 243: 123327, 2022 Jun 01.
Artigo em Inglês | MEDLINE | ID: mdl-35240367

RESUMO

The diagnosis of cancer and other diseases using data from non-specific sensors - such as the electronic tongues (e-tongues) - is challenging owing to the lack of selectivity, in addition to the variability of biological samples. In this study, we demonstrate that impedance data obtained with an e-tongue in saliva samples can be used to diagnose cancer in the mouth. Data taken with a single-response microfluidic e-tongue applied to the saliva of 27 individuals were treated with multidimensional projection techniques and non-supervised and supervised machine learning algorithms. The distinction between healthy individuals and patients with cancer on the floor of mouth or oral cavity could only be made with supervised learning. Accuracy above 80% was obtained for the binary classification (YES or NO for cancer) using a Support Vector Machine (SVM) with radial basis function kernel and Random Forest. In the classification considering the type of cancer, the accuracy dropped to ca. 70%. The accuracy tended to increase when clinical information such as alcohol consumption was used in conjunction with the e-tongue data. With the random forest algorithm, the rules to explain the diagnosis could be identified using the concept of Multidimensional Calibration Space. Since the training of the machine learning algorithms is believed to be more efficient when the data of a larger number of patients are employed, the approach presented here is promising for computer-assisted diagnosis.


Assuntos
Neoplasias Bucais , Saliva , Algoritmos , Nariz Eletrônico , Humanos , Aprendizado de Máquina , Neoplasias Bucais/diagnóstico , Máquina de Vetores de Suporte
3.
IEEE Trans Vis Comput Graph ; 27(2): 1427-1437, 2021 02.
Artigo em Inglês | MEDLINE | ID: mdl-33048689

RESUMO

Over the past decades, classification models have proven to be essential machine learning tools given their potential and applicability in various domains. In these years, the north of the majority of the researchers had been to improve quantitative metrics, notwithstanding the lack of information about models' decisions such metrics convey. This paradigm has recently shifted, and strategies beyond tables and numbers to assist in interpreting models' decisions are increasing in importance. Part of this trend, visualization techniques have been extensively used to support classification models' interpretability, with a significant focus on rule-based models. Despite the advances, the existing approaches present limitations in terms of visual scalability, and the visualization of large and complex models, such as the ones produced by the Random Forest (RF) technique, remains a challenge. In this paper, we propose Explainable Matrix (ExMatrix), a novel visualization method for RF interpretability that can handle models with massive quantities of rules. It employs a simple yet powerful matrix-like visual metaphor, where rows are rules, columns are features, and cells are rules predicates, enabling the analysis of entire models and auditing classification results. ExMatrix applicability is confirmed via different examples, showing how it can be used in practice to promote RF models interpretability.

SELEÇÃO DE REFERÊNCIAS
DETALHE DA PESQUISA