Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 24
Filtrar
2.
Comput Math Methods Med ; 2020: 7231205, 2020.
Artículo en Inglés | MEDLINE | ID: mdl-32952600

RESUMEN

Although sequencing a human genome has become affordable, identifying genetic variants from whole-genome sequence data is still a hurdle for researchers without adequate computing equipment or bioinformatics support. GATK is a gold standard method for the identification of genetic variants and has been widely used in genome projects and population genetic studies for many years. This was until the Google Brain team developed a new method, DeepVariant, which utilizes deep neural networks to construct an image classification model to identify genetic variants. However, the superior accuracy of DeepVariant comes at the cost of computational intensity, largely constraining its applications. Accordingly, we present DeepVariant-on-Spark to optimize resource allocation, enable multi-GPU support, and accelerate the processing of the DeepVariant pipeline. To make DeepVariant-on-Spark more accessible to everyone, we have deployed the DeepVariant-on-Spark to the Google Cloud Platform (GCP). Users can deploy DeepVariant-on-Spark on the GCP following our instruction within 20 minutes and start to analyze at least ten whole-genome sequencing datasets using free credits provided by the GCP. DeepVaraint-on-Spark is freely available for small-scale genome analysis using a cloud-based computing framework, which is suitable for pilot testing or preliminary study, while reserving the flexibility and scalability for large-scale sequencing projects.


Asunto(s)
Nube Computacional , Aprendizaje Profundo , Variación Genética , Secuenciación Completa del Genoma/estadística & datos numéricos , Nube Computacional/economía , Biología Computacional/métodos , Análisis Costo-Beneficio , Genoma Humano , Secuenciación de Nucleótidos de Alto Rendimiento/economía , Secuenciación de Nucleótidos de Alto Rendimiento/normas , Secuenciación de Nucleótidos de Alto Rendimiento/estadística & datos numéricos , Humanos , Redes Neurales de la Computación , Programas Informáticos , Secuenciación Completa del Genoma/economía , Secuenciación Completa del Genoma/normas
3.
PLoS One ; 15(8): e0237238, 2020.
Artículo en Inglés | MEDLINE | ID: mdl-32790750

RESUMEN

Reducing energy consumption has become a critical issue in today data centers. Reducing the number of required physical and Virtual Machines results in energy-efficiency. In this paper, to avoid the disadvantages of VM migration, a static VM placement algorithm is proposed which places VMs on hosts in a Worst-Fit-Decreasing (WFD) fashion. To reduce energy consumption further, the effect of job scheduling policy on the number of VMs needed for maintaining QoS requirements is studied. Each VM is modeled by an M/M/* queue in space-shared, time-shared, and hybrid job scheduling policies, and energy consumption of real-time as well as non-real-time applications is analyzed. Numerical results show that the hybrid policy outperforms space-shared and time-shared policies, in terms of energy consumption as well as Service Level Agreement (SLA) violations. Moreover, our non-migration method outperforms three different algorithms which use VM migration, in terms of reducing both energy consumption and SLA Violations.


Asunto(s)
Nube Computacional/economía , Sistemas de Computación/economía , Algoritmos , Simulación por Computador , Programas Informáticos/economía
4.
Nat Methods ; 17(8): 793-798, 2020 08.
Artículo en Inglés | MEDLINE | ID: mdl-32719530

RESUMEN

Massively parallel single-cell and single-nucleus RNA sequencing has opened the way to systematic tissue atlases in health and disease, but as the scale of data generation is growing, so is the need for computational pipelines for scaled analysis. Here we developed Cumulus-a cloud-based framework for analyzing large-scale single-cell and single-nucleus RNA sequencing datasets. Cumulus combines the power of cloud computing with improvements in algorithm and implementation to achieve high scalability, low cost, user-friendliness and integrated support for a comprehensive set of features. We benchmark Cumulus on the Human Cell Atlas Census of Immune Cells dataset of bone marrow cells and show that it substantially improves efficiency over conventional frameworks, while maintaining or improving the quality of results, enabling large-scale studies.


Asunto(s)
Nube Computacional/economía , Biología Computacional/métodos , Secuenciación de Nucleótidos de Alto Rendimiento/métodos , Análisis de Secuencia de ARN/métodos , Análisis de la Célula Individual/métodos , Biología Computacional/economía , Secuenciación de Nucleótidos de Alto Rendimiento/economía , Análisis de Secuencia de ARN/economía
5.
J Am Med Inform Assoc ; 27(9): 1425-1430, 2020 09 01.
Artículo en Inglés | MEDLINE | ID: mdl-32719837

RESUMEN

OBJECTIVE: Advancements in human genomics have generated a surge of available data, fueling the growth and accessibility of databases for more comprehensive, in-depth genetic studies. METHODS: We provide a straightforward and innovative methodology to optimize cloud configuration in order to conduct genome-wide association studies. We utilized Spark clusters on both Google Cloud Platform and Amazon Web Services, as well as Hail (http://doi.org/10.5281/zenodo.2646680) for analysis and exploration of genomic variants dataset. RESULTS: Comparative evaluation of numerous cloud-based cluster configurations demonstrate a successful and unprecedented compromise between speed and cost for performing genome-wide association studies on 4 distinct whole-genome sequencing datasets. Results are consistent across the 2 cloud providers and could be highly useful for accelerating research in genetics. CONCLUSIONS: We present a timely piece for one of the most frequently asked questions when moving to the cloud: what is the trade-off between speed and cost?


Asunto(s)
Nube Computacional , Estudio de Asociación del Genoma Completo , Nube Computacional/economía , Redes de Comunicación de Computadores , Análisis Costo-Beneficio , Estudio de Asociación del Genoma Completo/economía , Estudio de Asociación del Genoma Completo/métodos , Genómica/métodos , Humanos
8.
J Health Popul Nutr ; 38(Suppl 1): 24, 2019 10 18.
Artículo en Inglés | MEDLINE | ID: mdl-31627751

RESUMEN

This paper examines the hosting options for electronic civil registration and vital statistics (CRVS) systems, particularly the use of data centers versus cloud-based solutions. A data center is a facility that houses computer systems and associated hardware and software components, such as network and storage systems, power supplies, environment controls, and security devices. An alternative to using a data center is cloud-based hosting, which is a virtual data center hosted by a public cloud provider. The cloud is used on a pay-as-you-go basis and does not require purchasing and maintaining of hardware for data centers. It also provides more flexibility for continuous innovation in line with evolving information and communications technology.


Asunto(s)
Nube Computacional , Estadísticas Vitales , Nube Computacional/economía , Países en Desarrollo , Humanos , Gestión de la Información/métodos , Sistema de Registros
9.
J Dent Educ ; 83(8): 895-903, 2019 Aug.
Artículo en Inglés | MEDLINE | ID: mdl-31010892

RESUMEN

Electronic health records (EHRs) are increasingly moving towards cloud-based web environments. While cloud-based EHRs claim substantial benefits at reduced cost, little cost-benefit research exists for dental schools. The aim of this study was to examine the cost-benefits of a cloud-based EHR compared to an on-premise client-server EHR in the University of Michigan School of Dentistry (U-M Dent). Data were collected in 2016 from the U-M Dent cost-benefit comparison of tangible and intangible factors associated with implementing a new EHR, using the Total Cost of Ownership (TCO) framework from EDUCAUSE. The TCO framework assessed three factors: foundational (overarching aspects: three items), qualitative (intangibles: 56 items), and quantitative (actual costs). Stakeholders performed factor grading, and relative assessment scores were derived for each item as well as the overall factor. The cloud-based EHR solution received higher foundational and qualitative factor summary scores. The overall cost of an on-premise solution over a two-year period was approximately $2,000,000 higher than a cloud-based solution. Cloud solutions did not carry any hidden costs, while such costs accounted for 8% (~$540,000) of the overall costs of the on-premise solution. Across the two-year period, both one-time and ongoing costs were higher for the on-premise solution than the cloud-based solution (by 40.5% and 20.5%, respectively). This study found that a cloud-based EHR system in the U-M Dent offered significant cost savings and unique benefits that were not available with the on-premise EHR solution. Based on cost, the U-M Dent has made a case for cloud-based EHR systems.


Asunto(s)
Nube Computacional/economía , Análisis Costo-Beneficio , Clínicas Odontológicas , Registros Electrónicos de Salud/economía , Ahorro de Costo , Análisis de Datos , Educación en Odontología , Registros Electrónicos de Salud/instrumentación , Humanos , Michigan , Innovación Organizacional/economía , Facultades de Odontología
10.
Artículo en Inglés | MEDLINE | ID: mdl-31888203

RESUMEN

Currently, the green procurement activities of private hospitals in Taiwan follow the self-built green electronic-procurement (e-procurement) system. This requires professional personnel to take the time to regularly update the green specification and software and hardware of the e-procurement system, and the information system maintenance cost is high. In the case of a green e-procurement system crash, the efficiency of green procurement activities for hospitals is affected. If the green e-procurement can be moved to a convenient and trusty cloud computing model, this will enhance the efficiency of procurement activities and reduce the information maintenance cost for private hospitals. However, implementing a cloud model is an issue of technology innovation application and the technology-organization-environment (TOE) framework has been widely applied as the theoretical framework in technology innovation application. In addition, finding the weight of factors is a multi-criteria decision-making (MCDM) issue. Therefore, the present study first collected factors influencing implementation of the cloud mode together with the TOE as the theoretical framework, by reviewing the literature. Therefore, an expert questionnaire was designed and distributed to top managers of 20 private hospitals in southern Taiwan. The fuzzy analysis hierarchical process (FAHP), which is a MCDM tool, finds the weights of the factors influencing private hospitals in southern Taiwan when they implement a cloud green e-procurement system. The research results can enable private hospitals to successfully implement a green e-procurement system through a cloud model by optimizing resource allocation according to the weight of each factor. In addition, the results of this research can help cloud service providers of green e-procurement understand users' needs and develop relevant cloud solutions and marketing strategies.


Asunto(s)
Nube Computacional/economía , Nube Computacional/estadística & datos numéricos , Administración Financiera de Hospitales/organización & administración , Administración Financiera de Hospitales/estadística & datos numéricos , Administración de Materiales de Hospital/organización & administración , Administración de Materiales de Hospital/estadística & datos numéricos , Humanos , Encuestas y Cuestionarios , Taiwán
11.
Angew Chem Int Ed Engl ; 57(46): 15128-15132, 2018 11 12.
Artículo en Inglés | MEDLINE | ID: mdl-30272384

RESUMEN

The power of the Cloud has been harnessed for pharmaceutical compound production with remote servers based in Tokyo, Japan being left to autonomously find optimal synthesis conditions for three active pharmaceutical ingredients (APIs) in laboratories in Cambridge, UK. A researcher located in Los Angeles, USA controlled the entire process via an internet connection. The constituent synthetic steps for Tramadol, Lidocaine, and Bupropion were thus optimized with minimal intervention from operators within hours, yielding conditions satisfying customizable evaluation functions for all examples.


Asunto(s)
Analgésicos Opioides/síntesis química , Anestésicos Locales/síntesis química , Antidepresivos de Segunda Generación/síntesis química , Bupropión/síntesis química , Técnicas de Química Sintética/métodos , Lidocaína/síntesis química , Tramadol/síntesis química , Técnicas de Química Sintética/economía , Técnicas de Química Sintética/instrumentación , Nube Computacional/economía , Industria Farmacéutica/economía , Industria Farmacéutica/instrumentación , Industria Farmacéutica/métodos , Diseño de Equipo , Japón , Reino Unido , Estados Unidos
13.
Radiat Oncol ; 13(1): 99, 2018 Jun 27.
Artículo en Inglés | MEDLINE | ID: mdl-29945681

RESUMEN

BACKGROUND: A new implementation has been made on CloudMC, a cloud-based platform presented in a previous work, in order to provide services for radiotherapy treatment verification by means of Monte Carlo in a fast, easy and economical way. A description of the architecture of the application and the new developments implemented is presented together with the results of the tests carried out to validate its performance. METHODS: CloudMC has been developed over Microsoft Azure cloud. It is based on a map/reduce implementation for Monte Carlo calculations distribution over a dynamic cluster of virtual machines in order to reduce calculation time. CloudMC has been updated with new methods to read and process the information related to radiotherapy treatment verification: CT image set, treatment plan, structures and dose distribution files in DICOM format. Some tests have been designed in order to determine, for the different tasks, the most suitable type of virtual machines from those available in Azure. Finally, the performance of Monte Carlo verification in CloudMC is studied through three real cases that involve different treatment techniques, linac models and Monte Carlo codes. RESULTS: Considering computational and economic factors, D1_v2 and G1 virtual machines were selected as the default type for the Worker Roles and the Reducer Role respectively. Calculation times up to 33 min and costs of 16 € were achieved for the verification cases presented when a statistical uncertainty below 2% (2σ) was required. The costs were reduced to 3-6 € when uncertainty requirements are relaxed to 4%. CONCLUSIONS: Advantages like high computational power, scalability, easy access and pay-per-usage model, make Monte Carlo cloud-based solutions, like the one presented in this work, an important step forward to solve the long-lived problem of truly introducing the Monte Carlo algorithms in the daily routine of the radiotherapy planning process.


Asunto(s)
Nube Computacional , Método de Montecarlo , Planificación de la Radioterapia Asistida por Computador/métodos , Algoritmos , Nube Computacional/economía , Humanos , Fantasmas de Imagen , Radiometría , Dosificación Radioterapéutica , Planificación de la Radioterapia Asistida por Computador/economía , Programas Informáticos
14.
BMC Genomics ; 19(Suppl 1): 959, 2018 01 19.
Artículo en Inglés | MEDLINE | ID: mdl-29363427

RESUMEN

BACKGROUND: Whole-genome sequencing (WGS) plays an increasingly important role in clinical practice and public health. Due to the big data size, WGS data analysis is usually compute-intensive and IO-intensive. Currently it usually takes 30 to 40 h to finish a 50× WGS analysis task, which is far from the ideal speed required by the industry. Furthermore, the high-end infrastructure required by WGS computing is costly in terms of time and money. In this paper, we aim to improve the time efficiency of WGS analysis and minimize the cost by elastic cloud computing. RESULTS: We developed a distributed system, GT-WGS, for large-scale WGS analyses utilizing the Amazon Web Services (AWS). Our system won the first prize on the Wind and Cloud challenge held by Genomics and Cloud Technology Alliance conference (GCTA) committee. The system makes full use of the dynamic pricing mechanism of AWS. We evaluate the performance of GT-WGS with a 55× WGS dataset (400GB fastq) provided by the GCTA 2017 competition. In the best case, it only took 18.4 min to finish the analysis and the AWS cost of the whole process is only 16.5 US dollars. The accuracy of GT-WGS is 99.9% consistent with that of the Genome Analysis Toolkit (GATK) best practice. We also evaluated the performance of GT-WGS performance on a real-world dataset provided by the XiangYa hospital, which consists of 5× whole-genome dataset with 500 samples, and on average GT-WGS managed to finish one 5× WGS analysis task in 2.4 min at a cost of $3.6. CONCLUSIONS: WGS is already playing an important role in guiding therapeutic intervention. However, its application is limited by the time cost and computing cost. GT-WGS excelled as an efficient and affordable WGS analyses tool to address this problem. The demo video and supplementary materials of GT-WGS can be accessed at https://github.com/Genetalks/wgs_analysis_demo .


Asunto(s)
Nube Computacional/economía , Genoma Humano , Genómica/métodos , Análisis de Secuencia de ADN/métodos , Programas Informáticos , Secuenciación Completa del Genoma/métodos , Análisis por Conglomerados , Humanos , Análisis de Secuencia de ADN/economía , Secuenciación Completa del Genoma/economía
15.
Comput Intell Neurosci ; 2017: 4873459, 2017.
Artículo en Inglés | MEDLINE | ID: mdl-28811819

RESUMEN

Infrastructure as a Service (IaaS) cloud provides resources as a service from a pool of compute, network, and storage resources. Cloud providers can manage their resource usage by knowing future usage demand from the current and past usage patterns of resources. Resource usage prediction is of great importance for dynamic scaling of cloud resources to achieve efficiency in terms of cost and energy consumption while keeping quality of service. The purpose of this paper is to present a real-time resource usage prediction system. The system takes real-time utilization of resources and feeds utilization values into several buffers based on the type of resources and time span size. Buffers are read by R language based statistical system. These buffers' data are checked to determine whether their data follows Gaussian distribution or not. In case of following Gaussian distribution, Autoregressive Integrated Moving Average (ARIMA) is applied; otherwise Autoregressive Neural Network (AR-NN) is applied. In ARIMA process, a model is selected based on minimum Akaike Information Criterion (AIC) values. Similarly, in AR-NN process, a network with the lowest Network Information Criterion (NIC) value is selected. We have evaluated our system with real traces of CPU utilization of an IaaS cloud of one hundred and twenty servers.


Asunto(s)
Nube Computacional/estadística & datos numéricos , Nube Computacional/economía , Sistemas de Computación/economía , Sistemas de Computación/estadística & datos numéricos , Predicción , Redes Neurales de la Computación , Distribución Normal , Programas Informáticos , Factores de Tiempo
17.
Structure ; 25(6): 951-961.e2, 2017 06 06.
Artículo en Inglés | MEDLINE | ID: mdl-28552576

RESUMEN

Cryo-electron tomography (cryo-ET) captures the 3D electron density distribution of macromolecular complexes in close to native state. With the rapid advance of cryo-ET acquisition technologies, it is possible to generate large numbers (>100,000) of subtomograms, each containing a macromolecular complex. Often, these subtomograms represent a heterogeneous sample due to variations in the structure and composition of a complex in situ form or because particles are a mixture of different complexes. In this case subtomograms must be classified. However, classification of large numbers of subtomograms is a time-intensive task and often a limiting bottleneck. This paper introduces an open source software platform, TomoMiner, for large-scale subtomogram classification, template matching, subtomogram averaging, and alignment. Its scalable and robust parallel processing allows efficient classification of tens to hundreds of thousands of subtomograms. In addition, TomoMiner provides a pre-configured TomoMinerCloud computing service permitting users without sufficient computing resources instant access to TomoMiners high-performance features.


Asunto(s)
Tomografía con Microscopio Electrónico/métodos , Programas Informáticos , Chaperonina 10/química , Chaperonina 60/química , Nube Computacional/economía , Clusterina , Procesamiento de Imagen Asistido por Computador/métodos
18.
J Am Coll Radiol ; 14(1): 130-134, 2017 Jan.
Artículo en Inglés | MEDLINE | ID: mdl-27687749

RESUMEN

As the health care environment continually changes, radiologists look to the ACR's Imaging 3.0® initiative to guide the search for value. By leveraging new technology, a cloud-based image exchange network could provide secure universal access to prior images, which were previously siloed, to facilitate accurate interpretation, improved outcomes, and reduced costs. The breast imaging department represents a viable starting point given the robust data supporting the benefit of access to prior imaging studies, existing infrastructure for image sharing, and the current workflow reliance on prior images. This concept is scalable not only to the remainder of the radiology department but also to the broader medical record.


Asunto(s)
Nube Computacional/economía , Diagnóstico por Imagen/estadística & datos numéricos , Costos de la Atención en Salud/estadística & datos numéricos , Registro Médico Coordinado , Sistemas de Información Radiológica/economía , Valores Sociales , Estados Unidos
19.
Sci Rep ; 6: 39259, 2016 12 16.
Artículo en Inglés | MEDLINE | ID: mdl-27982081

RESUMEN

Public compendia of sequencing data are now measured in petabytes. Accordingly, it is infeasible for researchers to transfer these data to local computers. Recently, the National Cancer Institute began exploring opportunities to work with molecular data in cloud-computing environments. With this approach, it becomes possible for scientists to take their tools to the data and thereby avoid large data transfers. It also becomes feasible to scale computing resources to the needs of a given analysis. We quantified transcript-expression levels for 12,307 RNA-Sequencing samples from the Cancer Cell Line Encyclopedia and The Cancer Genome Atlas. We used two cloud-based configurations and examined the performance and cost profiles of each configuration. Using preemptible virtual machines, we processed the samples for as little as $0.09 (USD) per sample. As the samples were processed, we collected performance metrics, which helped us track the duration of each processing step and quantified computational resources used at different stages of sample processing. Although the computational demands of reference alignment and expression quantification have decreased considerably, there remains a critical need for researchers to optimize preprocessing steps. We have stored the software, scripts, and processed data in a publicly accessible repository (https://osf.io/gqrz9).


Asunto(s)
Nube Computacional/economía , Neoplasias/patología , Interfaz Usuario-Computador , Línea Celular Tumoral , Biología Computacional/economía , Biología Computacional/métodos , Bases de Datos Factuales , Humanos , Internet , Neoplasias/genética , Neoplasias/metabolismo , ARN Neoplásico/química , ARN Neoplásico/metabolismo , Análisis de Secuencia de ARN
20.
PLoS One ; 11(8): e0160456, 2016.
Artículo en Inglés | MEDLINE | ID: mdl-27501046

RESUMEN

Recently, cloud computing has drawn significant attention from both industry and academia, bringing unprecedented changes to computing and information technology. The infrastructure-as-a-Service (IaaS) model offers new abilities such as the elastic provisioning and relinquishing of computing resources in response to workload fluctuations. However, because the demand for resources dynamically changes over time, the provisioning of resources in a way that a given budget is efficiently utilized while maintaining a sufficing performance remains a key challenge. This paper addresses the problem of task scheduling and resource provisioning for a set of tasks running on IaaS clouds; it presents novel provisioning and scheduling algorithms capable of executing tasks within a given budget, while minimizing the slowdown due to the budget constraint. Our simulation study demonstrates a substantial reduction up to 70% in the overall task slowdown rate by the proposed algorithms.


Asunto(s)
Nube Computacional/economía , Algoritmos , Modelos Teóricos , Carga de Trabajo
SELECCIÓN DE REFERENCIAS
DETALLE DE LA BÚSQUEDA