Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 545
Filtrar
1.
BMC Bioinformatics ; 25(1): 186, 2024 May 10.
Artigo em Inglês | MEDLINE | ID: mdl-38730374

RESUMO

BACKGROUND: Commonly used next generation sequencing machines typically produce large amounts of short reads of a few hundred base-pairs in length. However, many downstream applications would generally benefit from longer reads. RESULTS: We present CAREx-an algorithm for the generation of pseudo-long reads from paired-end short-read Illumina data based on the concept of repeatedly computing multiple-sequence-alignments to extend a read until its partner is found. Our performance evaluation on both simulated data and real data shows that CAREx is able to connect significantly more read pairs (up to 99 % for simulated data) and to produce more error-free pseudo-long reads than previous approaches. When used prior to assembly it can achieve superior de novo assembly results. Furthermore, the GPU-accelerated version of CAREx exhibits the fastest execution times among all tested tools. CONCLUSION: CAREx is a new MSA-based algorithm and software for producing pseudo-long reads from paired-end short read data. It outperforms other state-of-the-art programs in terms of (i) percentage of connected read pairs, (ii) reduction of error rates of filled gaps, (iii) runtime, and (iv) downstream analysis using de novo assembly. CAREx is open-source software written in C++ (CPU version) and in CUDA/C++ (GPU version). It is licensed under GPLv3 and can be downloaded at ( https://github.com/fkallen/CAREx ).


Assuntos
Algoritmos , Sequenciamento de Nucleotídeos em Larga Escala , Software , Sequenciamento de Nucleotídeos em Larga Escala/métodos , Análise de Sequência de DNA/métodos , Humanos , Alinhamento de Sequência/métodos
2.
BMC Bioinformatics ; 25(1): 71, 2024 Feb 14.
Artigo em Inglês | MEDLINE | ID: mdl-38355413

RESUMO

BACKGROUND: Gene expression may be regulated by the DNA methylation of regulatory elements in cis, distal, and trans regions. One method to evaluate the relationship between DNA methylation and gene expression is the mapping of expression quantitative trait methylation (eQTM) loci (also called expression associated CpG loci, eCpG). However, no open-source tools are available to provide eQTM mapping. In addition, eQTM mapping can involve a large number of comparisons which may prevent the analyses due to limitations of computational resources. Here, we describe Torch-eCpG, an open-source tool to perform eQTM mapping that includes an optimized implementation that can use the graphical processing unit (GPU) to reduce runtime. RESULTS: We demonstrate the analyses using the tool are reproducible, up to 18 × faster using the GPU, and scale linearly with increasing methylation loci. CONCLUSIONS: Torch-eCpG is a fast, reliable, and scalable tool to perform eQTM mapping. Source code for Torch-eCpG is available at https://github.com/kordk/torch-ecpg .


Assuntos
Metilação de DNA , Locos de Características Quantitativas , Fenótipo , Sequências Reguladoras de Ácido Nucleico , Software
3.
J Comput Chem ; 2024 May 25.
Artigo em Inglês | MEDLINE | ID: mdl-38795375

RESUMO

The fragment molecular orbital (FMO) scheme is one of the popular fragmentation-based methods and has the potential advantage of making the circuit shallow for quantum chemical calculations on quantum computers. In this study, we used a GPU-accelerated quantum simulator (cuQuantum) to perform the electron correlation part of the FMO calculation as unitary coupled-cluster singles and doubles (UCCSD) with the variational quantum eigensolver (VQE) for hydrogen-bonded (FH) 3 $$ {}_3 $$ and (FH) 2 $$ {}_2 $$ -H 2 $$ {}_2 $$ O systems with the STO-3G basis set. VQE-UCCSD calculations were performed using both canonical and localized MO sets, and the results were examined from the point of view of size-consistency and orbital-invariance affected by the Trotter error. It was found that the use of localized MO leads to better results, especially for (FH) 2 $$ {}_2 $$ -H 2 $$ {}_2 $$ O. The GPU acceleration was substantial for the simulations with larger numbers of qubits, and was about a factor of 6.7-7.7 for 18 qubit systems.

4.
J Synchrotron Radiat ; 31(Pt 4): 851-866, 2024 Jul 01.
Artigo em Inglês | MEDLINE | ID: mdl-38771775

RESUMO

Despite the increased brilliance of the new generation synchrotron sources, there is still a challenge with high-resolution scanning of very thick and absorbing samples, such as a whole mouse brain stained with heavy elements, and, extending further, brains of primates. Samples are typically cut into smaller parts, to ensure a sufficient X-ray transmission, and scanned separately. Compared with the standard tomography setup where the sample would be cut into many pillars, the laminographic geometry operates with slab-shaped sections significantly reducing the number of sample parts to be prepared, the cutting damage and data stitching problems. In this work, a laminography pipeline for imaging large samples (>1 cm) at micrometre resolution is presented. The implementation includes a low-cost instrument setup installed at the 2-BM micro-CT beamline of the Advanced Photon Source. Additionally, sample mounting, scanning techniques, data stitching procedures, a fast reconstruction algorithm with low computational complexity, and accelerated reconstruction on multi-GPU systems for processing large-scale datasets are presented. The applicability of the whole laminography pipeline was demonstrated by imaging four sequential slabs throughout an entire mouse brain sample stained with osmium, in total generating approximately 12 TB of raw data for reconstruction.

5.
J Synchrotron Radiat ; 31(Pt 3): 517-526, 2024 May 01.
Artigo em Inglês | MEDLINE | ID: mdl-38517755

RESUMO

Physical optics simulations for beamlines and experiments allow users to test experiment feasibility and optimize beamline settings ahead of beam time in order to optimize valuable beam time at synchrotron light sources like NSLS-II. Further, such simulations also help to develop and test experimental data processing methods and software in advance. The Synchrotron Radiation Workshop (SRW) software package supports such complex simulations. We demonstrate how recent developments in SRW significantly improve the efficiency of physical optics simulations, such as end-to-end simulations of time-dependent X-ray photon correlation spectroscopy experiments with partially coherent undulator radiation (UR). The molecular dynamics simulation code LAMMPS was chosen to model the sample: a solution of silica nanoparticles in water at room temperature. Real-space distributions of nanoparticles produced by LAMMPS were imported into SRW and used to simulate scattering patterns of partially coherent hard X-ray UR from such a sample at the detector. The partially coherent UR illuminating the sample can be represented by a set of orthogonal coherent modes obtained by simulation of emission and propagation of this radiation through the coherent hard X-ray (CHX) scattering beamline followed by a coherent-mode decomposition. GPU acceleration is added for several key functions of SRW used in propagation from sample to detector, further improving the speed of the calculations. The accuracy of this simulation is benchmarked by comparison with experimental data.

6.
J Synchrotron Radiat ; 31(Pt 1): 85-94, 2024 Jan 01.
Artigo em Inglês | MEDLINE | ID: mdl-37947305

RESUMO

X-ray-based computed tomography is a well established technique for determining the three-dimensional structure of an object from its two-dimensional projections. In the past few decades, there have been significant advancements in the brightness and detector technology of tomography instruments at synchrotron sources. These advancements have led to the emergence of new observations and discoveries, with improved capabilities such as faster frame rates, larger fields of view, higher resolution and higher dimensionality. These advancements have enabled the material science community to expand the scope of tomographic measurements towards increasingly in situ and in operando measurements. In these new experiments, samples can be rapidly evolving, have complex geometries and restrictions on the field of view, limiting the number of projections that can be collected. In such cases, standard filtered back-projection often results in poor quality reconstructions. Iterative reconstruction algorithms, such as model-based iterative reconstructions (MBIR), have demonstrated considerable success in producing high-quality reconstructions under such restrictions, but typically require high-performance computing resources with hundreds of compute nodes to solve the problem in a reasonable time. Here, tomoCAM, is introduced, a new GPU-accelerated implementation of model-based iterative reconstruction that leverages non-uniform fast Fourier transforms to efficiently compute Radon and back-projection operators and asynchronous memory transfers to maximize the throughput to the GPU memory. The resulting code is significantly faster than traditional MBIR codes and delivers the reconstructive improvement offered by MBIR with affordable computing time and resources. tomoCAM has a Python front-end, allowing access from Jupyter-based frameworks, providing straightforward integration into existing workflows at synchrotron facilities.

7.
Magn Reson Med ; 91(6): 2621-2637, 2024 Jun.
Artigo em Inglês | MEDLINE | ID: mdl-38234037

RESUMO

PURPOSE: To present an open-source MR simulation framework that facilitates the incorporation of complex motion and flow for studying cardiovascular MR (CMR) acquisition and reconstruction. METHODS: CMRsim is a Python package that allows simulation of CMR images using dynamic digital phantoms with complex motion as input. Two simulation paradigms are available, namely, numerical and analytical solutions to the Bloch equations, using a common motion representation. Competitive simulation speeds are achieved using TensorFlow for GPU acceleration. To demonstrate the capability of the package, one introductory and two advanced CMR simulation experiments are presented. The latter showcase phase-contrast imaging of turbulent flow downstream of a stenotic section and cardiac diffusion tensor imaging on a contracting left ventricle. Additionally, extensive documentation and example resources are provided. RESULTS: The Bloch simulation with turbulent flow using approximately 1.5 million particles and a sequence duration of 710 ms for each of the seven different velocity encodings took a total of 29 min on a NVIDIA Titan RTX GPU. The results show characteristic phase contrast and magnitude modulation present in real data. The analytical simulation of cardiac diffusion tensor imaging with bulk-motion phase sensitivity took approximately 10 s per diffusion-weighted image, including preparation and loading steps. The results exhibit the expected alteration of diffusion metrics due to strain. CONCLUSION: CMRsim is the first simulation framework that allows one to feasibly incorporate complex motion, including turbulent flow, to systematically study advanced CMR acquisition and reconstruction approaches. The open-source package features modularity and transparency, facilitating maintainability and extensibility in support of reproducible research.


Assuntos
Imagem de Tensor de Difusão , Coração , Coração/diagnóstico por imagem , Simulação por Computador , Movimento (Física) , Imagens de Fantasmas
8.
Magn Reson Med ; 92(2): 447-458, 2024 Aug.
Artigo em Inglês | MEDLINE | ID: mdl-38469890

RESUMO

PURPOSE: To introduce a tool (TensorFit) for ultrafast and robust metabolite fitting of MRSI data based on Torch's auto-differentiation and optimization framework. METHODS: TensorFit was implemented in Python based on Torch's auto-differentiation to fit individual metabolites in MRS spectra. The underlying time domain and/or frequency domain fitting model is based on a linear combination of metabolite spectroscopic response. The computational time efficiency and accuracy of TensorFit were tested on simulated and in vivo MRS data and compared against TDFDFit and QUEST. RESULTS: TensorFit demonstrates a significant improvement in computation speed, achieving a 165-times acceleration compared with TDFDFit and 115 times against QUEST. TensorFit showed smaller percentual errors on simulated data compared with TDFDFit and QUEST. When tested on in vivo data, it performed similarly to TDFDFit with a 2% better fit in terms of mean squared error while obtaining a 169-fold speedup. CONCLUSION: TensorFit enables fast and robust metabolite fitting in large MRSI data sets compared with conventional metabolite fitting methods. This tool could boost the clinical applicability of large 3D MRSI by enabling the fitting of large MRSI data sets within computation times acceptable in a clinical environment.


Assuntos
Algoritmos , Espectroscopia de Ressonância Magnética , Humanos , Espectroscopia de Ressonância Magnética/métodos , Simulação por Computador , Software , Encéfalo/metabolismo , Encéfalo/diagnóstico por imagem , Imageamento por Ressonância Magnética/métodos , Reprodutibilidade dos Testes , Processamento de Imagem Assistida por Computador/métodos
9.
J Exp Bot ; 2024 Jul 02.
Artigo em Inglês | MEDLINE | ID: mdl-38954539

RESUMO

Linear mixed models (LMMs) are a commonly used method for genome-wide association studies (GWAS) that aim to detect associations between genetic markers and phenotypic measurements in a population of individuals while accounting for population structure and cryptic relatedness. In a standard GWAS, hundreds of thousands to millions of statistical tests are performed, requiring control for multiple hypothesis testing. Typically, static corrections that penalize the number of tests performed are used to control for the family-wise error rate, which is the probability of making at least one false positive. However, it has been shown that in practice this threshold is too conservative for normally distributed phenotypes and not stringent enough for non-normally distributed phenotypes. Therefore, permutation-based LMM approaches have recently been proposed to provide a more realistic threshold that takes phenotypic distributions into account. In this work, we will discuss the advantages of permutation-based GWAS approaches, including new simulations and results from a re-analysis of all publicly available Arabidopsis thaliana phenotypes from the AraPheno database.

10.
J Appl Clin Med Phys ; 25(1): e14208, 2024 Jan.
Artigo em Inglês | MEDLINE | ID: mdl-37987549

RESUMO

This paper presents the effort to extend a previously reported code ARCHER, a GPU-based Monte Carlo (MC) code for coupled photon and electron transport, into protons including the consideration of magnetic fields. The proton transport is modeled using a Class-II condensed-history algorithm with continuous slowing-down approximation. The model includes ionization, multiple scattering, energy straggling, elastic and inelastic nuclear interactions, as well as deflection due to the Lorentz force in magnetic fields. An additional direction change is added for protons at the end of each step in the presence of the magnetic field. Secondary charge particles, except for protons, are terminated depositing kinetic energies locally, whereas secondary neutral particles are ignored. Each proton is transported step by step until its energy drops to below 0.5 MeV or when the proton leaves the phantom. The code is implemented using the compute unified device architecture (CUDA) platform for optimized GPU thread-level parallelism and efficiency. The code is validated by comparing it against TOPAS. Comparisons of dose distributions between our code and TOPAS for several exposure scenarios, ranging from single square beams in water to patient plan with magnetic fields, show good agreement. The 3D-gamma pass rate with a 2 mm/2% criterion in the region with dose greater than 10% of the maximum dose is computed to be over 99% for all tested cases. Using a single NVIDIA TITAN V GPU card, the computational time of ARCHER is found to range from 0.82 to 4.54 seconds for 1 × 107 proton histories. Compared to a few hours running on TOPAS, this speed improvement is significant. This work presents, for the first time, the performance of a GPU-based MC code to simulate proton transportation magnetic fields, demonstrating the feasibility of accurate and efficient dose calculations in potential magnetic resonance imaging (MRI)-guided proton therapy.


Assuntos
Terapia com Prótons , Prótons , Humanos , Dosagem Radioterapêutica , Terapia com Prótons/métodos , Software , Planejamento da Radioterapia Assistida por Computador/métodos , Método de Monte Carlo , Imagens de Fantasmas , Campos Magnéticos
11.
Sensors (Basel) ; 24(5)2024 Feb 29.
Artigo em Inglês | MEDLINE | ID: mdl-38475138

RESUMO

The approach of using more than one processor to compute in order to overcome the complexity of different medical imaging methods that make up an overall job is known as GPU (graphic processing unit)-based parallel processing. It is extremely important for several medical imaging techniques such as image classification, object detection, image segmentation, registration, and content-based image retrieval, since the GPU-based parallel processing approach allows for time-efficient computation by a software, allowing multiple computations to be completed at once. On the other hand, a non-invasive imaging technology that may depict the shape of an anatomy and the biological advancements of the human body is known as magnetic resonance imaging (MRI). Implementing GPU-based parallel processing approaches in brain MRI analysis with medical imaging techniques might be helpful in achieving immediate and timely image capture. Therefore, this extended review (the extension of the IWBBIO2023 conference paper) offers a thorough overview of the literature with an emphasis on the expanding use of GPU-based parallel processing methods for the medical analysis of brain MRIs with the imaging techniques mentioned above, given the need for quicker computation to acquire early and real-time feedback in medicine. Between 2019 and 2023, we examined the articles in the literature matrix that include the tasks, techniques, MRI sequences, and processing results. As a result, the methods discussed in this review demonstrate the advancements achieved until now in minimizing computing runtime as well as the obstacles and problems still to be solved in the future.


Assuntos
Algoritmos , Gráficos por Computador , Humanos , Software , Encéfalo , Imageamento por Ressonância Magnética/métodos , Processamento de Imagem Assistida por Computador/métodos
12.
Sensors (Basel) ; 24(6)2024 Mar 18.
Artigo em Inglês | MEDLINE | ID: mdl-38544210

RESUMO

Graphics processing units (GPUs) facilitate massive parallelism and high-capacity storage, and thus are suitable for the iterative reconstruction of ultrahigh-resolution micro computed tomography (CT) scans by on-the-fly system matrix (OTFSM) calculation using ordered subsets expectation maximization (OSEM). We propose a finite state automaton (FSA) method that facilitates iterative reconstruction using a heterogeneous multi-GPU platform through parallelizing the matrix calculations derived from a ray tracing system of ordered subsets. The FSAs perform flow control for parallel threading of the heterogeneous GPUs, which minimizes the latency of launching ordered-subsets tasks, reduces the data transfer between the main system memory and local GPU memory, and solves the memory-bound of a single GPU. In the experiments, we compared the operation efficiency of OS-MLTR for three reconstruction environments. The heterogeneous multiple GPUs with job queues for high throughput calculation speed is up to five times faster than the single GPU environment, and that speed up is nine times faster than the heterogeneous multiple GPUs with the FIFO queues of the device scheduling control. Eventually, we proposed an event-triggered FSA method for iterative reconstruction using multiple heterogeneous GPUs that solves the memory-bound issue of a single GPU at ultrahigh resolutions, and the routines of the proposed method were successfully executed on each GPU simultaneously.

13.
Sensors (Basel) ; 24(5)2024 Feb 21.
Artigo em Inglês | MEDLINE | ID: mdl-38474935

RESUMO

Hyperspectral imaging (HSI) has become a very compelling technique in different scientific areas; indeed, many researchers use it in the fields of remote sensing, agriculture, forensics, and medicine. In the latter, HSI plays a crucial role as a diagnostic support and for surgery guidance. However, the computational effort in elaborating hyperspectral data is not trivial. Furthermore, the demand for detecting diseases in a short time is undeniable. In this paper, we take up this challenge by parallelizing three machine-learning methods among those that are the most intensively used: Support Vector Machine (SVM), Random Forest (RF), and eXtreme Gradient Boosting (XGB) algorithms using the Compute Unified Device Architecture (CUDA) to accelerate the classification of hyperspectral skin cancer images. They all showed a good performance in HS image classification, in particular when the size of the dataset is limited, as demonstrated in the literature. We illustrate the parallelization techniques adopted for each approach, highlighting the suitability of Graphical Processing Units (GPUs) to this aim. Experimental results show that parallel SVM and XGB algorithms significantly improve the classification times in comparison with their serial counterparts.


Assuntos
Algoritmos , Neoplasias Cutâneas , Humanos , Aprendizado de Máquina , Imageamento Hiperespectral , Aceleração , Máquina de Vetores de Suporte
14.
Sensors (Basel) ; 24(4)2024 Feb 18.
Artigo em Inglês | MEDLINE | ID: mdl-38400470

RESUMO

Cardiac CINE, a form of dynamic cardiac MRI, is indispensable in the diagnosis and treatment of heart conditions, offering detailed visualization essential for the early detection of cardiac diseases. As the demand for higher-resolution images increases, so does the volume of data requiring processing, presenting significant computational challenges that can impede the efficiency of diagnostic imaging. Our research presents an approach that takes advantage of the computational power of multiple Graphics Processing Units (GPUs) to address these challenges. GPUs are devices capable of performing large volumes of computations in a short period, and have significantly improved the cardiac MRI reconstruction process, allowing images to be produced faster. The innovation of our work resides in utilizing a multi-device system capable of processing the substantial data volumes demanded by high-resolution, five-dimensional cardiac MRI. This system surpasses the memory capacity limitations of single GPUs by partitioning large datasets into smaller, manageable segments for parallel processing, thereby preserving image integrity and accelerating reconstruction times. Utilizing OpenCL technology, our system offers adaptability and cross-platform functionality, ensuring wider applicability. The proposed multi-device approach offers an advancement in medical imaging, accelerating the reconstruction process and facilitating faster and more effective cardiac health assessment.


Assuntos
Algoritmos , Imageamento por Ressonância Magnética , Coração/diagnóstico por imagem , Aumento da Imagem/métodos , Imageamento Tridimensional/métodos
15.
J Environ Manage ; 360: 121024, 2024 Jun.
Artigo em Inglês | MEDLINE | ID: mdl-38759551

RESUMO

Urban waterlogging is a significant global issue. To achieve precisely control urban waterlogging and enhance our understanding of its causes, a novel study method was introduced. This method is based on a dynamic bidirectional coupling model that combines 1D-2D hydrodynamic and water quality simulations. The waterlogging phenomenon in densely populated metropolitan areas of Changzhi city, China, was studied. This study focused on investigating the process involved in waterlogging formation, particularly overflow at nodes induced by the design of the topological structure of the pipe network, constraints on the capacity of the underground drainage system, and the surface runoff accumulation. The complex interplay among these elements and their possible influences on waterlogging formation were clarified. The results indicated notable spatial and temporal variation in the waterlogging formation process in densely populated urban areas. Node overflow in the drainage system emerged as the key influencing factor in the waterlogging formation process, accounting for up to 71% of the total water accumulation at the peak time. The peak lag time of waterlogging during events with short return periods was primarily determined by the rainfall peak moment. In contrast, the peak time of waterlogging during events with long return periods was influenced by the rainfall peak moment, drainage capacity and topological structure of the pipe network. Notably, the access of inflow from both upstream and downstream segments of the pipe network drainage system significantly impacted the peak time of waterlogging, with upstream water potentially delaying the peak time substantially. This study not only provides new insights into urban waterlogging mechanisms but also provides practical guidance for optimizing urban drainage systems, urban planning, and disaster risk management.


Assuntos
Modelos Teóricos , China , Movimentos da Água , Chuva , Cidades , Qualidade da Água
16.
BMC Bioinformatics ; 24(1): 221, 2023 May 31.
Artigo em Inglês | MEDLINE | ID: mdl-37259021

RESUMO

BACKGROUND: As genome sequencing becomes better integrated into scientific research, government policy, and personalized medicine, the primary challenge for researchers is shifting from generating raw data to analyzing these vast datasets. Although much work has been done to reduce compute times using various configurations of traditional CPU computing infrastructures, Graphics Processing Units (GPUs) offer opportunities to accelerate genomic workflows by orders of magnitude. Here we benchmark one GPU-accelerated software suite called NVIDIA Parabricks on Amazon Web Services (AWS), Google Cloud Platform (GCP), and an NVIDIA DGX cluster. We benchmarked six variant calling pipelines, including two germline callers (HaplotypeCaller and DeepVariant) and four somatic callers (Mutect2, Muse, LoFreq, SomaticSniper). RESULTS: We achieved up to 65 × acceleration with germline variant callers, bringing HaplotypeCaller runtimes down from 36 h to 33 min on AWS, 35 min on GCP, and 24 min on the NVIDIA DGX. Somatic callers exhibited more variation between the number of GPUs and computing platforms. On cloud platforms, GPU-accelerated germline callers resulted in cost savings compared with CPU runs, whereas some somatic callers were more expensive than CPU runs because their GPU acceleration was not sufficient to overcome the increased GPU cost. CONCLUSIONS: Germline variant callers scaled well with the number of GPUs across platforms, whereas somatic variant callers exhibited more variation in the number of GPUs with the fastest runtimes, suggesting that, at least with the version of Parabricks used here, these workflows are less GPU optimized and require benchmarking on the platform of choice before being deployed at production scales. Our study demonstrates that GPUs can be used to greatly accelerate genomic workflows, thus bringing closer to grasp urgent societal advances in the areas of biosurveillance and personalized medicine.


Assuntos
Gráficos por Computador , Software , Fluxo de Trabalho , Genômica
17.
J Synchrotron Radiat ; 30(Pt 1): 179-191, 2023 Jan 01.
Artigo em Inglês | MEDLINE | ID: mdl-36601936

RESUMO

Fast 3D data analysis and steering of a tomographic experiment by changing environmental conditions or acquisition parameters require fast, close to real-time, 3D reconstruction of large data volumes. Here a performance-optimized TomocuPy package is presented as a GPU alternative to the commonly used central processing unit (CPU) based TomoPy package for tomographic reconstruction. TomocuPy utilizes modern hardware capabilities to organize a 3D asynchronous reconstruction involving parallel read/write operations with storage drives, CPU-GPU data transfers, and GPU computations. In the asynchronous reconstruction, all the operations are timely overlapped to almost fully hide all data management time. Since most cameras work with less than 16-bit digital output, the memory usage and processing speed are furthermore optimized by using 16-bit floating-point arithmetic. As a result, 3D reconstruction with TomocuPy became 20-30 times faster than its multi-threaded CPU equivalent. Full reconstruction (including read/write operations and methods initialization) of a 20483 tomographic volume takes less than 7 s on a single Nvidia Tesla A100 and PCIe 4.0 NVMe SSD, and scales almost linearly increasing the data size. To simplify operation at synchrotron beamlines, TomocuPy provides an easy-to-use command-line interface. Efficacy of the package was demonstrated during a tomographic experiment on gas-hydrate formation in porous samples, where a steering option was implemented as a lens-changing mechanism for zooming to regions of interest.


Assuntos
Gráficos por Computador , Software , Algoritmos , Processamento de Imagem Assistida por Computador/métodos
18.
J Synchrotron Radiat ; 30(Pt 1): 217-226, 2023 Jan 01.
Artigo em Inglês | MEDLINE | ID: mdl-36601940

RESUMO

FOCUS (Fast Monte CarlO approach to Coherence of Undulator Sources) is a new GPU-based simulation code to compute the transverse coherence of undulator radiation from ultra-relativistic electrons. The core structure of the code, which is written in the language C++ accelerated with CUDA, combines an analytical description of the emitted electric fields and massively parallel computations on GPUs. The combination is rigorously justified by a statistical description of synchrotron radiation based on a Fourier optics approach. FOCUS is validated by direct comparison with multi-electron Synchrotron Radiation Workshop (SRW) simulations, evidencing a reduction in computation times by up to five orders of magnitude on a consumer laptop. FOCUS is then applied to systematically study the transverse coherence in typical third- and fourth-generation facilities, highlighting peculiar features of undulator sources close to the diffraction limit. FOCUS is aimed at fast evaluation of the transverse coherence of undulator radiation as a function of the electron beam parameters, to support and help prepare more advanced and detailed numerical simulations with traditional codes like SRW.

19.
Brief Bioinform ; 22(5)2021 09 02.
Artigo em Inglês | MEDLINE | ID: mdl-33834216

RESUMO

Reverse engineering mechanistic gene regulatory network (GRN) models with a specific dynamic spatial behavior is an inverse problem without analytical solutions in general. Instead, heuristic machine learning algorithms have been proposed to infer the structure and parameters of a system of equations able to recapitulate a given gene expression pattern. However, these algorithms are computationally intensive as they need to simulate millions of candidate models, which limits their applicability and requires high computational resources. Graphics processing unit (GPU) computing is an affordable alternative for accelerating large-scale scientific computation, yet no method is currently available to exploit GPU technology for the reverse engineering of mechanistic GRNs from spatial phenotypes. Here we present an efficient methodology to parallelize evolutionary algorithms using GPU computing for the inference of mechanistic GRNs that can develop a given gene expression pattern in a multicellular tissue area or cell culture. The proposed approach is based on multi-CPU threads running the lightweight crossover, mutation and selection operators and launching GPU kernels asynchronously. Kernels can run in parallel in a single or multiple GPUs and each kernel simulates and scores the error of a model using the thread parallelism of the GPU. We tested this methodology for the inference of spatiotemporal mechanistic gene regulatory networks (GRNs)-including topology and parameters-that can develop a given 2D gene expression pattern. The results show a 700-fold speedup with respect to a single CPU implementation. This approach can streamline the extraction of knowledge from biological and medical datasets and accelerate the automatic design of GRNs for synthetic biology applications.


Assuntos
Algoritmos , Biologia Computacional/métodos , Gráficos por Computador , Perfilação da Expressão Gênica/métodos , Redes Reguladoras de Genes/genética , Modelos Genéticos , Simulação por Computador , Humanos , Reprodutibilidade dos Testes , Software , Fatores de Tempo
20.
Magn Reson Med ; 90(1): 329-342, 2023 07.
Artigo em Inglês | MEDLINE | ID: mdl-36877139

RESUMO

PURPOSE: To develop an open-source, high-performance, easy-to-use, extensible, cross-platform, and general MRI simulation framework (Koma). METHODS: Koma was developed using the Julia programming language. Like other MRI simulators, it solves the Bloch equations with CPU and GPU parallelization. The inputs are the scanner parameters, the phantom, and the pulse sequence that is Pulseq-compatible. The raw data is stored in the ISMRMRD format. For the reconstruction, MRIReco.jl is used. A graphical user interface utilizing web technologies was also designed. Two types of experiments were performed: one to compare the quality of the results and the execution speed, and the second to compare its usability. Finally, the use of Koma in quantitative imaging was demonstrated by simulating Magnetic Resonance Fingerprinting (MRF) acquisitions. RESULTS: Koma was compared to two well-known open-source MRI simulators, JEMRIS and MRiLab. Highly accurate results (with mean absolute differences below 0.1% compared to JEMRIS) and better GPU performance than MRiLab were demonstrated. In an experiment with students, Koma was proved to be easy to use, eight times faster on personal computers than JEMRIS, and 65% of test subjects recommended it. The potential for designing acquisition and reconstruction techniques was also shown through the simulation of MRF acquisitions, with conclusions that agree with the literature. CONCLUSIONS: Koma's speed and flexibility have the potential to make simulations more accessible for education and research. Koma is expected to be used for designing and testing novel pulse sequences before implementing them in the scanner with Pulseq files, and for creating synthetic data to train machine learning models.


Assuntos
Idioma , Imageamento por Ressonância Magnética , Humanos , Imageamento por Ressonância Magnética/métodos , Simulação por Computador , Imagens de Fantasmas , Aceleração
SELEÇÃO DE REFERÊNCIAS
DETALHE DA PESQUISA