Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 13 de 13
Filtrar
1.
Bioinformatics ; 40(6)2024 Jun 03.
Artigo em Inglês | MEDLINE | ID: mdl-38775680

RESUMO

MOTIVATION: The completion of the genome has paved the way for genome-wide association studies (GWAS), which explained certain proportions of heritability. GWAS are not optimally suited to detect non-linear effects in disease risk, possibly hidden in non-additive interactions (epistasis). Alternative methods for epistasis detection using, e.g. deep neural networks (DNNs) are currently under active development. However, DNNs are constrained by finite computational resources, which can be rapidly depleted due to increasing complexity with the sheer size of the genome. Besides, the curse of dimensionality complicates the task of capturing meaningful genetic patterns for DNNs; therefore necessitates dimensionality reduction. RESULTS: We propose a method to compress single nucleotide polymorphism (SNP) data, while leveraging the linkage disequilibrium (LD) structure and preserving potential epistasis. This method involves clustering correlated SNPs into haplotype blocks and training per-block autoencoders to learn a compressed representation of the block's genetic content. We provide an adjustable autoencoder design to accommodate diverse blocks and bypass extensive hyperparameter tuning. We applied this method to genotyping data from Project MinE, and achieved 99% average test reconstruction accuracy-i.e. minimal information loss-while compressing the input to nearly 10% of the original size. We demonstrate that haplotype-block based autoencoders outperform linear Principal Component Analysis (PCA) by approximately 3% chromosome-wide accuracy of reconstructed variants. To the extent of our knowledge, our approach is the first to simultaneously leverage haplotype structure and DNNs for dimensionality reduction of genetic data. AVAILABILITY AND IMPLEMENTATION: Data are available for academic use through Project MinE at https://www.projectmine.com/research/data-sharing/, contingent upon terms and requirements specified by the source studies. Code is available at https://github.com/gizem-tas/haploblock-autoencoders.


Assuntos
Estudo de Associação Genômica Ampla , Desequilíbrio de Ligação , Polimorfismo de Nucleotídeo Único , Humanos , Estudo de Associação Genômica Ampla/métodos , Epistasia Genética , Haplótipos , Redes Neurais de Computação , Algoritmos
2.
Bioinformatics ; 37(7): 905-912, 2021 05 17.
Artigo em Inglês | MEDLINE | ID: mdl-32871010

RESUMO

MOTIVATION: The microbes that live in an environment can be identified from the combined genomic material, also referred to as the metagenome. Sequencing a metagenome can result in large volumes of sequencing reads. A promising approach to reduce the size of metagenomic datasets is by clustering reads into groups based on their overlaps. Clustering reads are valuable to facilitate downstream analyses, including computationally intensive strain-aware assembly. As current read clustering approaches cannot handle the large datasets arising from high-throughput metagenome sequencing, a novel read clustering approach is needed. In this article, we propose OGRE, an Overlap Graph-based Read clustEring procedure for high-throughput sequencing data, with a focus on shotgun metagenomes. RESULTS: We show that for small datasets OGRE outperforms other read binners in terms of the number of species included in a cluster, also referred to as cluster purity, and the fraction of all reads that is placed in one of the clusters. Furthermore, OGRE is able to process metagenomic datasets that are too large for other read binners into clusters with high cluster purity. CONCLUSION: OGRE is the only method that can successfully cluster reads in species-specific clusters for large metagenomic datasets without running into computation time- or memory issues. AVAILABILITYAND IMPLEMENTATION: Code is made available on Github (https://github.com/Marleen1/OGRE). SUPPLEMENTARY INFORMATION: Supplementary data are available at Bioinformatics online.


Assuntos
Metagenoma , Software , Algoritmos , Análise por Conglomerados , Sequenciamento de Nucleotídeos em Larga Escala , Metagenômica , Análise de Sequência de DNA
3.
Bioinformatics ; 35(14): i538-i547, 2019 07 15.
Artigo em Inglês | MEDLINE | ID: mdl-31510706

RESUMO

MOTIVATION: Amyotrophic lateral sclerosis (ALS) is a neurodegenerative disease caused by aberrations in the genome. While several disease-causing variants have been identified, a major part of heritability remains unexplained. ALS is believed to have a complex genetic basis where non-additive combinations of variants constitute disease, which cannot be picked up using the linear models employed in classical genotype-phenotype association studies. Deep learning on the other hand is highly promising for identifying such complex relations. We therefore developed a deep-learning based approach for the classification of ALS patients versus healthy individuals from the Dutch cohort of the Project MinE dataset. Based on recent insight that regulatory regions harbor the majority of disease-associated variants, we employ a two-step approach: first promoter regions that are likely associated to ALS are identified, and second individuals are classified based on their genotype in the selected genomic regions. Both steps employ a deep convolutional neural network. The network architecture accounts for the structure of genome data by applying convolution only to parts of the data where this makes sense from a genomics perspective. RESULTS: Our approach identifies potentially ALS-associated promoter regions, and generally outperforms other classification methods. Test results support the hypothesis that non-additive combinations of variants contribute to ALS. Architectures and protocols developed are tailored toward processing population-scale, whole-genome data. We consider this a relevant first step toward deep learning assisted genotype-phenotype association in whole genome-sized data. AVAILABILITY AND IMPLEMENTATION: Our code will be available on Github, together with a synthetic dataset (https://github.com/byin-cwi/ALS-Deeplearning). The data used in this study is available to bona-fide researchers upon request. SUPPLEMENTARY INFORMATION: Supplementary data are available at Bioinformatics online.


Assuntos
Esclerose Lateral Amiotrófica , Genoma , Redes Neurais de Computação , Doenças Neurodegenerativas , Esclerose Lateral Amiotrófica/genética , Genótipo , Humanos
4.
Phys Med Biol ; 69(2)2024 Jan 12.
Artigo em Inglês | MEDLINE | ID: mdl-37625421

RESUMO

Objective. Increasing cancer incidence, staff shortage and high burnout rate among radiation oncologists, medical physicists and radiation technicians are putting many departments under strain. Operations research (OR) tools could optimize radiotherapy processes, however, clinical implementation of OR-tools in radiotherapy is scarce since most investigated optimization methods lack robustness against patient-to-patient variation in duration of tasks. By combining OR-tools, a method was developed that optimized deployment of radiotherapy resources by generating robust pretreatment preparation schedules that balance the expected average patient preparation time (Fmean) with the risk of working overtime (RoO). The method was evaluated for various settings of an one-stop shop (OSS) outpatient clinic for palliative radiotherapy.Approach. The OSS at our institute sees, scans and treats 3-5 patients within one day. The OSS pretreatment preparation workflow consists of a fixed sequence of tasks, which was manually optimized for radiation oncologist and CT availability. To find more optimal sequences, with shorterFmeanand lowerRoO, a genetic algorithm was developed which regards these sequences as DNA-strands. The genetic algorithm applied natural selection principles to produce new sequences. A decoder translated sequences to schedules to find the conflicting fitness parametersFmeanandRoO. For every generation, fitness of sequences was determined by the distance to the estimated Pareto front ofFmeanandRoO. Experiments were run in various OSS-settings.Main results. According to our approach, the expectedFmeanof the current clinical schedule could be reduced with 37%, without increasingRoO. Additional experiments provided insights in trade-offs betweenFmean,RoO, working shift length, number of patients treated on a single day and staff composition.Significance. Our approach demonstrated that OR-tools could optimize radiotherapy resources by robust pretreatment workflow scheduling. The results strongly support further exploration of scheduling optimization for treatment preparation also outside a one-stop shop or radiotherapy setting.


Assuntos
Neoplasias , Planejamento da Radioterapia Assistida por Computador , Humanos , Planejamento da Radioterapia Assistida por Computador/métodos , Fluxo de Trabalho , Neoplasias/radioterapia , Dosagem Radioterapêutica
5.
Front Nutr ; 10: 1239915, 2023.
Artigo em Inglês | MEDLINE | ID: mdl-37497056

RESUMO

[This corrects the article DOI: 10.3389/fnut.2022.1056205.].

6.
Front Nutr ; 9: 1056205, 2022.
Artigo em Inglês | MEDLINE | ID: mdl-36590216

RESUMO

Introduction: Linear programming (LP) is often used within diet optimization to find, from a set of available food commodities, the most affordable diet that meets the nutritional requirements of an individual or (sub)population. It is, however, not always possible to create a feasible diet, as certain nutritional requirements are difficult to meet. In that case, goal programming (GP) can be used to minimize deviations from the nutritional requirements in order to obtain a near feasible diet. With GP the cost of the diet is often overlooked or taken into account using the ε-constraint method. This method does not guarantee to find all possible trade-offs between costs and nutritional deficiency without solving many uninformative LPs. Methods: We present a method to find all trade-offs between any two linear objectives in a dietary LP context that is simple, does not solve uninformative LPs and does not need prior input from the decision maker (DM). This method is a bi-objective algorithm based on the NonInferior Set Estimation (NISE) method that finds all efficient trade-offs between two linear objectives. Results: In order to show what type of insights can be gained from this approach, two analyses are presented that investigate the relation between cost and nutritional adequacy. In the first analysis a diet with a restriction on the exact energy intake is considered where all nutrient intakes except energy are allowed to deviate from their prescription. This analysis is especially helpful in case of a restrictive budget or when a nutritionally adequate diet is either unaffordable or unattainable. The second analysis only relaxes the exact energy intake, where the other nutrients are kept within their requirements, to investigate how the energy intake affects the cost of a diet. Here, we describe in what situations the so-called more-for-less paradox takes place, which can be induced by requiring an exact energy intake. Conclusion: To the best of our knowledge, we are the first to address how to obtain all efficient trade-offs of two linear objectives in a dietary LP context and how this can be used for analyses.

7.
Med Phys ; 48(6): 3096-3108, 2021 Jun.
Artigo em Inglês | MEDLINE | ID: mdl-33721350

RESUMO

PURPOSE: Theoretical studies have shown that dose-painting-by-numbers (DPBN) could lead to large gains in tumor control probability (TCP) compared to conventional dose distributions. However, these gains may vary considerably among patients due to (a) variations in the overall radiosensitivity of the tumor, (b) variations in the 3D distribution of intra-tumor radiosensitivity within the tumor in combination with patient anatomy, (c) uncertainties of the 3D radiosensitivity maps, (d) geometrical uncertainties, and (e) temporal changes in radiosensitivity. The goal of this study was to investigate how much of the theoretical gains of DPBN remain when accounting for these factors. DPBN was compared to both a homogeneous reference dose distribution and to nonselective dose escalation (NSDE), that uses the same dose constraints as DPBN, but does not require 3D radiosensitivity maps. METHODS: A fully automated DPBN treatment planning strategy was developed and implemented in our in-house developed treatment planning system (TPS) that is robust to uncertainties in radiosensitivity and patient positioning. The method optimized the expected TCP based on 3D maps of intra-tumor radiosensitivity, while accounting for normal tissue constraints, uncertainties in radiosensitivity, and setup uncertainties. Based on FDG-PETCT scans of 12 non-small cell lung cancer (NSCLC) patients, data of 324 virtual patients were created synthetically with large variations in the aforementioned parameters. DPBN was compared to both a uniform dose distribution of 60 Gy, and NSDE. In total, 360 DPBN and 24 NSDE treatment plans were optimized. RESULTS: The average gain in TCP over all patients and radiosensitivity maps of DPBN was 0.54 ± 0.20 (range 0-0.97) compared to the 60 Gy uniform reference dose distribution, but only 0.03 ± 0.03 (range 0-0.22) compared to NSDE. The gains varied per patient depending on the radiosensitivity of the entire tumor and the 3D radiosensitivity maps. Uncertainty in radiosensitivity led to a considerable loss in TCP gain, which could be recovered almost completely by accounting for the uncertainty directly in the optimization. CONCLUSIONS: Our results suggest that the gains of DPBN can be considerable compared to a 60 Gy uniform reference dose distribution, but small compared to NSDE for most patients. Using the robust DPBN treatment planning system developed in this work, the optimal DPBN treatment plan could be derived for any patient for whom 3D intra-tumor radiosensitivity maps are known, and can be used to select patients that might benefit from DPBN. NSDE could be an effective strategy to increase TCP without requiring biological information of the tumor.


Assuntos
Carcinoma Pulmonar de Células não Pequenas , Neoplasias Pulmonares , Carcinoma Pulmonar de Células não Pequenas/diagnóstico por imagem , Carcinoma Pulmonar de Células não Pequenas/radioterapia , Humanos , Neoplasias Pulmonares/diagnóstico por imagem , Neoplasias Pulmonares/radioterapia , Tolerância a Radiação , Dosagem Radioterapêutica , Planejamento da Radioterapia Assistida por Computador
8.
PLoS One ; 16(3): e0245519, 2021.
Artigo em Inglês | MEDLINE | ID: mdl-33657128

RESUMO

Since the onset of the COVID-19 pandemic many researchers and health advisory institutions have focused on virus spread prediction through epidemiological models. Such models rely on virus- and disease characteristics of which most are uncertain or even unknown for SARS-CoV-2. This study addresses the validity of various assumptions using an epidemiological simulation model. The contributions of this work are twofold. First, we show that multiple scenarios all lead to realistic numbers of deaths and ICU admissions, two observable and verifiable metrics. Second, we test the sensitivity of estimates for the number of infected and immune individuals, and show that these vary strongly between scenarios. Note that the amount of variation measured in this study is merely a lower bound: epidemiological modeling contains uncertainty on more parameters than the four in this study, and including those as well would lead to an even larger set of possible scenarios. As the level of infection and immunity among the population are particularly important for policy makers, further research on virus and disease progression characteristics is essential. Until that time, epidemiological modeling studies cannot give conclusive results and should come with a careful analysis of several scenarios on virus- and disease characteristics.


Assuntos
COVID-19/epidemiologia , Transmissão de Doença Infecciosa/estatística & dados numéricos , Previsões/métodos , COVID-19/transmissão , Humanos , Modelos Estatísticos , Pandemias , SARS-CoV-2/patogenicidade
9.
Genome Biol ; 21(1): 31, 2020 02 07.
Artigo em Inglês | MEDLINE | ID: mdl-32033589

RESUMO

The recent boom in microfluidics and combinatorial indexing strategies, combined with low sequencing costs, has empowered single-cell sequencing technology. Thousands-or even millions-of cells analyzed in a single experiment amount to a data revolution in single-cell biology and pose unique data science problems. Here, we outline eleven challenges that will be central to bringing this emerging field of single-cell data science forward. For each challenge, we highlight motivating research questions, review prior work, and formulate open problems. This compendium is for established researchers, newcomers, and students alike, highlighting interesting and rewarding problems for the coming years.


Assuntos
Ciência de Dados/métodos , Genômica/métodos , RNA-Seq/métodos , Análise de Célula Única/métodos , Animais , Humanos
11.
Phys Med Biol ; 62(4): 1225-1247, 2017 02 21.
Artigo em Inglês | MEDLINE | ID: mdl-28052046

RESUMO

In this article we provide a method to generate the trade-off between delivery time and fluence map matching quality for dynamically delivered fluence maps. At the heart of our method lies a mathematical programming model that, for a given duration of delivery, optimizes leaf trajectories and dose rates such that the desired fluence map is reproduced as well as possible. We begin with the single fluence map case and then generalize the model and the solution technique to the delivery of sequential fluence maps. The resulting large-scale, non-convex optimization problem was solved using a heuristic approach. We test our method using a prostate case and a head and neck case, and present the resulting trade-off curves. Analysis of the leaf trajectories reveals that short time plans have larger leaf openings in general than longer delivery time plans. Our method allows one to explore the continuum of possibilities between coarse, large segment plans characteristic of direct aperture approaches and narrow field plans produced by sliding window approaches. Exposing this trade-off will allow for an informed choice between plan quality and solution time. Further research is required to speed up the optimization process to make this method clinically implementable.


Assuntos
Algoritmos , Neoplasias de Cabeça e Pescoço/radioterapia , Neoplasias da Próstata/radioterapia , Planejamento da Radioterapia Assistida por Computador/métodos , Radioterapia de Intensidade Modulada/métodos , Humanos , Masculino , Dosagem Radioterapêutica
12.
Phys Med Biol ; 60(2): 537-48, 2015 Jan 21.
Artigo em Inglês | MEDLINE | ID: mdl-25549084

RESUMO

Inverse planning algorithms for dwell time optimisation in interstitial high-dose-rate (HDR) brachytherapy may produce solutions with large dwell time variations within catheters, which may result in undesirable selective high-dose subvolumes. Extending the dwell time optimisation model with a dwell time modulation restriction (DTMR) that limits dwell time differences between neighboring dwell positions has been suggested to eliminate this problem. DTMRs may additionally reduce the sensitivity for uncertainties in dwell positions that inevitably result from catheter reconstruction errors and afterloader source positioning inaccuracies. This study quantifies the reduction of high-dose subvolumes and the robustness against these uncertainties by applying a DTMR to template-based prostate HDR brachytherapy implants. Three different DTMRs were consecutively applied to a linear dose-based penalty model (LD) and a dose-volume based model (LDV), both obtained from literature. The models were solved with DTMR levels ranging from no restriction to uniform dwell times within catheters in discrete steps. Uncertainties were simulated on clinical cases using in-house developed software, and dose-volume metrics were calculated in each simulation. For the assessment of high-dose subvolumes, the dose homogeneity index (DHI) and the contiguous dose volume histogram were analysed. Robustness was measured by the improvement of the lowest D90% of the planning target volume (PTV) observed in the simulations. For (LD), a DTMR yields an increase in DHI of approximately 30% and reduces the size of the largest high-dose volume by 2-5 cc. However, this comes at a cost of a reduction in D90% of the PTV of 10%, which often implies that it drops below the desired minimum of 100%. For (LDV), none of the DTMRs were able to improve high-dose volume measures. DTMRs were not capable of improving robustness of PTV D90% against uncertainty in dwell positions for both models.


Assuntos
Braquiterapia/métodos , Neoplasias da Próstata/radioterapia , Doses de Radiação , Planejamento da Radioterapia Assistida por Computador/métodos , Algoritmos , Humanos , Modelos Lineares , Masculino , Dosagem Radioterapêutica , Software
13.
Phys Med Biol ; 60(14): 5681-98, 2015 Jul 21.
Artigo em Inglês | MEDLINE | ID: mdl-26147184

RESUMO

Advances in precision small animal radiotherapy hardware enable the delivery of increasingly complicated dose distributions on the millimeter scale. Manual creation and evaluation of treatment plans becomes difficult or even infeasible with an increasing number of degrees of freedom for dose delivery and available image data. The goal of this work is to develop an optimisation model that determines beam-on times for a given beam configuration, and to assess the feasibility and benefits of an automated treatment planning system for small animal radiotherapy. The developed model determines a Pareto optimal solution using operator-defined weights for a multiple-objective treatment planning problem. An interactive approach allows the planner to navigate towards, and to select the Pareto optimal treatment plan that yields the most preferred trade-off of the conflicting objectives. This model was evaluated using four small animal cases based on cone-beam computed tomography images. Resulting treatment plan quality was compared to the quality of manually optimised treatment plans using dose-volume histograms and metrics. Results show that the developed framework is well capable of optimising beam-on times for 3D dose distributions and offers several advantages over manual treatment plan optimisation. For all cases but the simple flank tumour case, a similar amount of time was needed for manual and automated beam-on time optimisation. In this time frame, manual optimisation generates a single treatment plan, while the inverse planning system yields a set of Pareto optimal solutions which provides quantitative insight on the sensitivity of conflicting objectives. Treatment planning automation decreases the dependence on operator experience and allows for the use of class solutions for similar treatment scenarios. This can shorten the time required for treatment planning and therefore increase animal throughput. In addition, this can improve treatment standardisation and comparability of research data within studies and among different institutes.


Assuntos
Tomografia Computadorizada de Feixe Cônico/métodos , Imageamento Tridimensional/métodos , Neoplasias Experimentais/radioterapia , Planejamento da Radioterapia Assistida por Computador/métodos , Planejamento da Radioterapia Assistida por Computador/normas , Radioterapia de Intensidade Modulada/métodos , Animais , Camundongos , Dosagem Radioterapêutica , Distribuição Tecidual , Carga Tumoral
SELEÇÃO DE REFERÊNCIAS
DETALHE DA PESQUISA