Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 36
Filtrar
Mais filtros

Base de dados
País/Região como assunto
Tipo de documento
Intervalo de ano de publicação
1.
Proc Natl Acad Sci U S A ; 118(49)2021 12 07.
Artigo em Inglês | MEDLINE | ID: mdl-34873056

RESUMO

Preclinical models have been the workhorse of cancer research, producing massive amounts of drug response data. Unfortunately, translating response biomarkers derived from these datasets to human tumors has proven to be particularly challenging. To address this challenge, we developed TRANSACT, a computational framework that builds a consensus space to capture biological processes common to preclinical models and human tumors and exploits this space to construct drug response predictors that robustly transfer from preclinical models to human tumors. TRANSACT performs favorably compared to four competing approaches, including two deep learning approaches, on a set of 23 drug prediction challenges on The Cancer Genome Atlas and 226 metastatic tumors from the Hartwig Medical Foundation. We demonstrate that response predictions deliver a robust performance for a number of therapies of high clinical importance: platinum-based chemotherapies, gemcitabine, and paclitaxel. In contrast to other approaches, we demonstrate the interpretability of the TRANSACT predictors by correctly identifying known biomarkers of targeted therapies, and we propose potential mechanisms that mediate the resistance to two chemotherapeutic agents.


Assuntos
Ensaios de Seleção de Medicamentos Antitumorais/métodos , Perfilação da Expressão Gênica/métodos , Animais , Antineoplásicos/uso terapêutico , Biomarcadores Farmacológicos/metabolismo , Linhagem Celular Tumoral/efeitos dos fármacos , Aprendizado Profundo , Modelos Animais de Doenças , Previsões/métodos , Xenoenxertos , Humanos , Modelos Teóricos
2.
Bioinformatics ; 35(14): i510-i519, 2019 07 15.
Artigo em Inglês | MEDLINE | ID: mdl-31510654

RESUMO

MOTIVATION: Cell lines and patient-derived xenografts (PDXs) have been used extensively to understand the molecular underpinnings of cancer. While core biological processes are typically conserved, these models also show important differences compared to human tumors, hampering the translation of findings from pre-clinical models to the human setting. In particular, employing drug response predictors generated on data derived from pre-clinical models to predict patient response remains a challenging task. As very large drug response datasets have been collected for pre-clinical models, and patient drug response data are often lacking, there is an urgent need for methods that efficiently transfer drug response predictors from pre-clinical models to the human setting. RESULTS: We show that cell lines and PDXs share common characteristics and processes with human tumors. We quantify this similarity and show that a regression model cannot simply be trained on cell lines or PDXs and then applied on tumors. We developed PRECISE, a novel methodology based on domain adaptation that captures the common information shared amongst pre-clinical models and human tumors in a consensus representation. Employing this representation, we train predictors of drug response on pre-clinical data and apply these predictors to stratify human tumors. We show that the resulting domain-invariant predictors show a small reduction in predictive performance in the pre-clinical domain but, importantly, reliably recover known associations between independent biomarkers and their companion drugs on human tumors. AVAILABILITY AND IMPLEMENTATION: PRECISE and the scripts for running our experiments are available on our GitHub page (https://github.com/NKI-CCB/PRECISE). SUPPLEMENTARY INFORMATION: Supplementary data are available at Bioinformatics online.


Assuntos
Antineoplásicos , Neoplasias , Animais , Antineoplásicos/farmacologia , Fenômenos Biológicos , Modelos Animais de Doenças , Previsões , Humanos , Neoplasias/tratamento farmacológico , Software
3.
Proc Natl Acad Sci U S A ; 117(20): 10625-10626, 2020 05 19.
Artigo em Inglês | MEDLINE | ID: mdl-32371495
4.
Hum Brain Mapp ; 35(9): 4916-31, 2014 Sep.
Artigo em Inglês | MEDLINE | ID: mdl-24700485

RESUMO

Because hypoperfusion of brain tissue precedes atrophy in dementia, the detection of dementia may be advanced by the use of perfusion information. Such information can be obtained noninvasively with arterial spin labeling (ASL), a relatively new MR technique quantifying cerebral blood flow (CBF). Using ASL and structural MRI, we evaluated diagnostic classification in 32 prospectively included presenile early stage dementia patients and 32 healthy controls. Patients were suspected of Alzheimer's disease (AD) or frontotemporal dementia. Classification was based on CBF as perfusion marker, gray matter (GM) volume as atrophy marker, and their combination. These markers were each examined using six feature extraction methods: a voxel-wise method and a region of interest (ROI)-wise approach using five ROI-sets in the GM. These ROI-sets ranged in number from 72 brain regions to a single ROI for the entire supratentorial brain. Classification was performed with a linear support vector machine classifier. For validation of the classification method on the basis of GM features, a reference dataset from the AD Neuroimaging Initiative database was used consisting of AD patients and healthy controls. In our early stage dementia population, the voxelwise feature-extraction approach achieved more accurate results (area under the curve (AUC) range = 86 - 91%) than all other approaches (AUC = 57 - 84%). Used in isolation, CBF quantified with ASL was a good diagnostic marker for dementia. However, our findings indicated only little added diagnostic value when combining ASL with the structural MRI data (AUC = 91%), which did not significantly improve over accuracy of structural MRI atrophy marker by itself.


Assuntos
Doença de Alzheimer/diagnóstico , Encéfalo/patologia , Encéfalo/fisiopatologia , Circulação Cerebrovascular , Interpretação de Imagem Assistida por Computador/métodos , Imageamento por Ressonância Magnética/métodos , Idoso , Doença de Alzheimer/patologia , Doença de Alzheimer/fisiopatologia , Área Sob a Curva , Atrofia , Diagnóstico Diferencial , Feminino , Demência Frontotemporal/diagnóstico , Demência Frontotemporal/patologia , Demência Frontotemporal/fisiopatologia , Substância Cinzenta/patologia , Humanos , Modelos Lineares , Masculino , Pessoa de Meia-Idade , Tamanho do Órgão , Estudos Prospectivos , Máquina de Vetores de Suporte
5.
IEEE Trans Pattern Anal Mach Intell ; 45(6): 7799-7819, 2023 Jun.
Artigo em Inglês | MEDLINE | ID: mdl-36350870

RESUMO

Learning curves provide insight into the dependence of a learner's generalization performance on the training set size. This important tool can be used for model selection, to predict the effect of more training data, and to reduce the computational complexity of model training and hyperparameter tuning. This review recounts the origins of the term, provides a formal definition of the learning curve, and briefly covers basics such as its estimation. Our main contribution is a comprehensive overview of the literature regarding the shape of learning curves. We discuss empirical and theoretical evidence that supports well-behaved curves that often have the shape of a power law or an exponential. We consider the learning curves of Gaussian processes, the complex shapes they can display, and the factors influencing them. We draw specific attention to examples of learning curves that are ill-behaved, showing worse learning performance with more training data. To wrap up, we point out various open problems that warrant deeper empirical and theoretical investigation. All in all, our review underscores that learning curves are surprisingly diverse and no universal model can be identified.

6.
IEEE Trans Pattern Anal Mach Intell ; 45(4): 4747-4767, 2023 Apr.
Artigo em Inglês | MEDLINE | ID: mdl-35984799

RESUMO

Semi-supervised learning is the learning setting in which we have both labeled and unlabeled data at our disposal. This survey covers theoretical results for this setting and maps out the benefits of unlabeled data in classification and regression tasks. Most methods that use unlabeled data rely on certain assumptions about the data distribution. When those assumptions are not met, including unlabeled data may actually decrease performance. For all practical purposes, it is therefore instructive to have an understanding of the underlying theory and the possible learning behavior that comes with it. This survey gathers results about the possible gains one can achieve when using semi-supervised learning as well as results about the limits of such methods. Specifically, it aims to answer the following questions: what are, in terms of improving supervised methods, the limits of semi-supervised learning? What are the assumptions of different methods? What can we achieve if the assumptions are true? As, indeed, the precise assumptions made are of the essence, this is where the survey's particular attention goes out to.

7.
IEEE Trans Pattern Anal Mach Intell ; 43(3): 766-785, 2021 03.
Artigo em Inglês | MEDLINE | ID: mdl-31603771

RESUMO

Domain adaptation has become a prominent problem setting in machine learning and related fields. This review asks the question: How can a classifier learn from a source domain and generalize to a target domain? We present a categorization of approaches, divided into, what we refer to as, sample-based, feature-based, and inference-based methods. Sample-based methods focus on weighting individual observations during training based on their importance to the target domain. Feature-based methods revolve around on mapping, projecting, and representing features such that a source classifier performs well on the target domain and inference-based methods incorporate adaptation into the parameter estimation procedure, for instance through constraints on the optimization procedure. Additionally, we review a number of conditions that allow for formulating bounds on the cross-domain generalization error. Our categorization highlights recurring ideas and raises questions important to further research.

8.
IEEE Trans Image Process ; 30: 8342-8353, 2021.
Artigo em Inglês | MEDLINE | ID: mdl-34587011

RESUMO

Resolution in deep convolutional neural networks (CNNs) is typically bounded by the receptive field size through filter sizes, and subsampling layers or strided convolutions on feature maps. The optimal resolution may vary significantly depending on the dataset. Modern CNNs hard-code their resolution hyper-parameters in the network architecture which makes tuning such hyper-parameters cumbersome. We propose to do away with hard-coded resolution hyper-parameters and aim to learn the appropriate resolution from data. We use scale-space theory to obtain a self-similar parametrization of filters and make use of the N-Jet: a truncated Taylor series to approximate a filter by a learned combination of Gaussian derivative filters. The parameter σ of the Gaussian basis controls both the amount of detail the filter encodes and the spatial extent of the filter. Since σ is a continuous parameter, we can optimize it with respect to the loss. The proposed N-Jet layer achieves comparable performance when used in state-of-the art architectures, while learning the correct resolution in each layer automatically. We evaluate our N-Jet layer on both classification and segmentation, and we show that learning σ is especially beneficial when dealing with inputs at multiple sizes.

9.
IEEE Trans Pattern Anal Mach Intell ; 30(6): 1114-5, 2008 Jun.
Artigo em Inglês | MEDLINE | ID: mdl-18421115

RESUMO

Recently, an interpretation of the whitened cosine similarity measure as a Bayes decision rule was proposed (C. Liu, "The Bayes Decision Rule Induced Similarity Measures,'' IEEE Trans. Pattern Analysis and Machine Intelligence, vol. 29, no. 6, pp. 1086-1090, June 2007. This communication makes the observation that some of the distributional assumptions made to derive this measure are very restrictive and, considered simultaneously, even inconsistent.


Assuntos
Algoritmos , Inteligência Artificial , Teorema de Bayes , Interpretação de Imagem Assistida por Computador/métodos , Armazenamento e Recuperação da Informação/métodos , Reconhecimento Automatizado de Padrão/métodos , Simulação por Computador , Aumento da Imagem/métodos , Modelos Estatísticos , Análise Numérica Assistida por Computador
10.
IEEE Trans Pattern Anal Mach Intell ; 40(2): 452-466, 2018 02.
Artigo em Inglês | MEDLINE | ID: mdl-28252390

RESUMO

We propose a template matching method for the detection of 2D image objects that are characterized by orientation patterns. Our method is based on data representations via orientation scores, which are functions on the space of positions and orientations, and which are obtained via a wavelet-type transform. This new representation allows us to detect orientation patterns in an intuitive and direct way, namely via cross-correlations. Additionally, we propose a generalized linear regression framework for the construction of suitable templates using smoothing splines. Here, it is important to recognize a curved geometry on the position-orientation domain, which we identify with the Lie group SE(2): the roto-translation group. Templates are then optimized in a B-spline basis, and smoothness is defined with respect to the curved geometry. We achieve state-of-the-art results on three different applications: detection of the optic nerve head in the retina (99.83 percent success rate on 1,737 images), of the fovea in the retina (99.32 percent success rate on 1,616 images), and of the pupil in regular camera images (95.86 percent on 1,521 images). The high performance is due to inclusion of both intensity and orientation features with effective geometric priors in the template matching. Moreover, our method is fast due to a cross-correlation based matching approach.

11.
IEEE Trans Med Imaging ; 25(5): 602-11, 2006 May.
Artigo em Inglês | MEDLINE | ID: mdl-16689264

RESUMO

The task of segmenting the posterior ribs within the lung fields of standard posteroanterior chest radiographs is considered. To this end, an iterative, pixel-based, supervised, statistical classification method is used, which is called iterated contextual pixel classification (ICPC). Starting from an initial rib segmentation obtained from pixel classification, ICPC updates it by reclassifying every pixel, based on the original features and, additionally, class label information of pixels in the neighborhood of the pixel to be reclassified. The method is evaluated on 30 radiographs taken from the JSRT (Japanese Society of Radiological Technology) database. All posterior ribs within the lung fields in these images have been traced manually by two observers. The first observer's segmentations are set as the gold standard; ICPC is trained using these segmentations. In a sixfold cross-validation experiment, ICPC achieves a classification accuracy of 0.86 +/- 0.06, as compared to 0.94 +/- 0.02 for the second human observer.


Assuntos
Neoplasias Pulmonares/diagnóstico por imagem , Reconhecimento Automatizado de Padrão/métodos , Intensificação de Imagem Radiográfica/métodos , Interpretação de Imagem Radiográfica Assistida por Computador/métodos , Radiografia Torácica/métodos , Costelas/diagnóstico por imagem , Processamento de Sinais Assistido por Computador , Algoritmos , Inteligência Artificial , Humanos , Armazenamento e Recuperação da Informação/métodos , Variações Dependentes do Observador , Reprodutibilidade dos Testes , Estudos Retrospectivos , Sensibilidade e Especificidade
12.
Med Image Anal ; 10(1): 19-40, 2006 Feb.
Artigo em Inglês | MEDLINE | ID: mdl-15919232

RESUMO

The task of segmenting the lung fields, the heart, and the clavicles in standard posterior-anterior chest radiographs is considered. Three supervised segmentation methods are compared: active shape models, active appearance models and a multi-resolution pixel classification method that employs a multi-scale filter bank of Gaussian derivatives and a k-nearest-neighbors classifier. The methods have been tested on a publicly available database of 247 chest radiographs, in which all objects have been manually segmented by two human observers. A parameter optimization for active shape models is presented, and it is shown that this optimization improves performance significantly. It is demonstrated that the standard active appearance model scheme performs poorly, but large improvements can be obtained by including areas outside the objects into the model. For lung field segmentation, all methods perform well, with pixel classification giving the best results: a paired t-test showed no significant performance difference between pixel classification and an independent human observer. For heart segmentation, all methods perform comparably, but significantly worse than a human observer. Clavicle segmentation is a hard problem for all methods; best results are obtained with active shape models, but human performance is substantially better. In addition, several hybrid systems are investigated. For heart segmentation, where the separate systems perform comparably, significantly better performance can be obtained by combining the results with majority voting. As an application, the cardio-thoracic ratio is computed automatically from the segmentation results. Bland and Altman plots indicate that all methods perform well when compared to the gold standard, with confidence intervals from pixel classification and active appearance modeling very close to those of a human observer. All results, including the manual segmentations, have been made publicly available to facilitate future comparative studies.


Assuntos
Algoritmos , Interpretação de Imagem Radiográfica Assistida por Computador , Radiografia Torácica , Bases de Dados como Assunto , Humanos , Intensificação de Imagem Radiográfica
13.
Med Image Anal ; 10(2): 247-58, 2006 Apr.
Artigo em Inglês | MEDLINE | ID: mdl-16293441

RESUMO

A computer algorithm for nodule detection in chest radiographs is presented. The algorithm consists of four main steps: (i) image preprocessing; (ii) nodule candidate detection; (iii) feature extraction; (iv) candidate classification. Two optional extensions to this scheme are tested: candidate selection and candidate segmentation. The output of step (ii) is a list of circles, which can be transformed into more detailed contours by the extra candidate segmentation step. In addition, the candidate selection step (which is a classification step using a small number of features) can be used to reduce the list of nodule candidates before step (iii). The algorithm uses multi-scale techniques in several stages of the scheme: Candidates are found by looking for local intensity maxima in Gaussian scale space; nodule boundaries are detected by tracing edge points found at large scales down to pixel scale; some of the features used for classification are taken from a multi-scale Gaussian filterbank. Experiments with this scheme (with and without the segmentation and selection steps) are carried out on a previously characterized, publicly available database, that contains a large number of very subtle nodules. For this database, counting as detections only those nodules that were indicated with a confidence level of 50% or more, radiologists previously detected 70% of the nodules. For our algorithm, it turns out that the selection step does have an added value for the system, while segmentation does not lead to a clear improvement. With the scheme with the best performance, accepting on average two false positives per image results in the identification of 51% of all nodules. For four false positives, this increases to 67%. This is close to the previously reported 70% detection rate of the radiologists.


Assuntos
Inteligência Artificial , Bases de Dados Factuais , Reconhecimento Automatizado de Padrão/métodos , Intensificação de Imagem Radiográfica/métodos , Interpretação de Imagem Radiográfica Assistida por Computador/métodos , Radiografia Torácica/métodos , Nódulo Pulmonar Solitário/diagnóstico por imagem , Algoritmos , Análise por Conglomerados , Humanos , Imageamento Tridimensional/métodos , Neoplasias Pulmonares/classificação , Neoplasias Pulmonares/diagnóstico por imagem , Sistemas Computadorizados de Registros Médicos , Reprodutibilidade dos Testes , Sensibilidade e Especificidade , Nódulo Pulmonar Solitário/classificação
14.
IEEE Trans Pattern Anal Mach Intell ; 38(3): 462-75, 2016 Mar.
Artigo em Inglês | MEDLINE | ID: mdl-27046491

RESUMO

Improvement guarantees for semi-supervised classifiers can currently only be given under restrictive conditions on the data. We propose a general way to perform semi-supervised parameter estimation for likelihood-based classifiers for which, on the full training set, the estimates are never worse than the supervised solution in terms of the log-likelihood. We argue, moreover, that we may expect these solutions to really improve upon the supervised classifier in particular cases. In a worked-out example for LDA, we take it one step further and essentially prove that its semi-supervised version is strictly better than its supervised counterpart. The two new concepts that form the core of our estimation principle are contrast and pessimism. The former refers to the fact that our objective function takes the supervised estimates into account, enabling the semi-supervised solution to explicitly control the potential improvements over this estimate. The latter refers to the fact that our estimates are conservative and therefore resilient to whatever form the true labeling of the unlabeled data takes on. Experiments demonstrate the improvements in terms of both the log-likelihood and the classification error rate on independent test sets.

15.
IEEE Trans Neural Netw Learn Syst ; 27(6): 1379-91, 2016 06.
Artigo em Inglês | MEDLINE | ID: mdl-27214351

RESUMO

In multiple instance learning, objects are sets (bags) of feature vectors (instances) rather than individual feature vectors. In this paper, we address the problem of how these bags can best be represented. Two standard approaches are to use (dis)similarities between bags and prototype bags, or between bags and prototype instances. The first approach results in a relatively low-dimensional representation, determined by the number of training bags, whereas the second approach results in a relatively high-dimensional representation, determined by the total number of instances in the training set. However, an advantage of the latter representation is that the informativeness of the prototype instances can be inferred. In this paper, a third, intermediate approach is proposed, which links the two approaches and combines their strengths. Our classifier is inspired by a random subspace ensemble, and considers subspaces of the dissimilarity space, defined by subsets of instances, as prototypes. We provide insight into the structure of some popular multiple instance problems and show state-of-the-art performances on these data sets.

16.
IEEE Trans Neural Netw Learn Syst ; 26(5): 995-1006, 2015 May.
Artigo em Inglês | MEDLINE | ID: mdl-25881369

RESUMO

We cast a semi-supervised nearest mean classifier, previously introduced by the first author, in a more principled log-likelihood formulation that is subject to constraints. This, in turn, leads us to make the important suggestion to not only investigate error rates of semi-supervised learners but also consider the risk they originally aim to optimize. We demonstrate empirically that in terms of classification error, mixed results are obtained when comparing supervised to semi-supervised nearest mean classification, while in terms of log-likelihood on the test set, the semi-supervised method consistently outperforms its supervised counterpart. Comparisons to self-learning, a standard approach in semi-supervised learning, are included to further clarify the way, in which our constrained nearest mean classifier improves over regular, supervised nearest mean classification.

17.
Neural Netw ; 17(4): 563-6, 2004 May.
Artigo em Inglês | MEDLINE | ID: mdl-15109684

RESUMO

Recently, a discrimination measure for feature extraction for two-class data, called the maximum discriminating (MDF) measure (Talukder and Casasent [Neural Networks 14 (2001) 1201-1218]), was introduced. In the present paper, it is shown that the MDF discrimination measure produces exactly the same results as the classical Fisher criterion, on the condition that the two prior probabilities are chosen to be equal. The effect of unequal priors on the efficiency of the measures is also discussed.


Assuntos
Redes Neurais de Computação , Dinâmica não Linear , Pesos e Medidas , Animais , Análise Discriminante , Humanos , Neurônios/fisiologia
18.
IEEE Trans Pattern Anal Mach Intell ; 26(6): 732-9, 2004 Jun.
Artigo em Inglês | MEDLINE | ID: mdl-18579934

RESUMO

We propose an eigenvector-based heteroscedastic linear dimension reduction (LDR) technique for multiclass data. The technique is based on a heteroscedastic two-class technique which utilizes the so-called Chernoff criterion, and successfully extends the well-known linear discriminant analysis (LDA). The latter, which is based on the Fisher criterion, is incapable of dealing with heteroscedastic data in a proper way. For the two-class case, the between-class scatter is generalized so to capture differences in (co)variances. It is shown that the classical notion of between-class scatter can be associated with Euclidean distances between class means. From this viewpoint, the between-class scatter is generalized by employing the Chernoff distance measure, leading to our proposed heteroscedastic measure. Finally, using the results from the two-class case, a multiclass extension of the Chernoff criterion is proposed. This criterion combines separation information present in the class mean as well as the class covariance matrices. Extensive experiments and a comparison with similar dimension reduction techniques are presented.


Assuntos
Algoritmos , Inteligência Artificial , Armazenamento e Recuperação da Informação/métodos , Modelos Lineares , Reconhecimento Automatizado de Padrão/métodos , Simulação por Computador , Reprodutibilidade dos Testes , Sensibilidade e Especificidade
19.
Artigo em Inglês | MEDLINE | ID: mdl-24110974

RESUMO

Using more than one classification stage and exploiting class population imbalance allows for incorporating powerful classifiers in tasks requiring large scale training data, even if these classifiers scale badly with the number of training samples. This led us to propose a two-stage classifier for segmenting tibial cartilage in knee MRI scans combining nearest neighbor classification and support vector machines (SVMs). Here we apply it to femoral cartilage segmentation. We describe the similarities and differences between segmenting these two knee cartilages. For further speeding up batch SVM training, we propose loosening the stopping condition in the quadratic program solver before considering moving on to other approximation techniques such as online SVMs. The two-stage approach reached a higher accuracy in comparison to the one-stage state-of-the-art method. It also achieved better inter-scan segmentation reproducibility when compared to a radiologist as well as the current state-of-the-art method.


Assuntos
Cartilagem/patologia , Processamento de Imagem Assistida por Computador/métodos , Imageamento por Ressonância Magnética/métodos , Reconhecimento Automatizado de Padrão/métodos , Algoritmos , Análise por Conglomerados , Fêmur/anatomia & histologia , Humanos , Articulação do Joelho , Radiologia/métodos , Reprodutibilidade dos Testes , Máquina de Vetores de Suporte
20.
Artigo em Inglês | MEDLINE | ID: mdl-23286174

RESUMO

In this study, we propose a computational diagnosis system for detecting the colorectal cancer from histopathological slices. The computational analysis was usually performed on patch level where only a small part of the slice is covered. However, slice-based classification is more realistic for histopathological diagnosis. The developed method combines both textural and structural features from patch images and proposes a two level classification scheme. In the first level, the patches in slices are classified into possible classes (adenomatous, inflamed, cancer and normal) and the distribution of the patches into these classes is considered as the information representing the slices. Then the slices are classified using a logistic linear classifier. In patch level, we obtain the correct classification accuracies of 94.36% and 96.34% for the cancer and normal classes, respectively. However, in slice level, the accuracies of the 79.17% and 92.68% are achieved for cancer and normal classes, respectively.


Assuntos
Algoritmos , Inteligência Artificial , Neoplasias Colorretais/patologia , Interpretação de Imagem Assistida por Computador/métodos , Microscopia/métodos , Reconhecimento Automatizado de Padrão/métodos , Humanos , Aumento da Imagem/métodos , Reprodutibilidade dos Testes , Sensibilidade e Especificidade
SELEÇÃO DE REFERÊNCIAS
DETALHE DA PESQUISA