RESUMEN
The recent explosion of biological data and the concomitant proliferation of distributed databases make it challenging for biologists and bioinformaticians to discover the best data resources for their needs, and the most efficient way to access and use them. Despite a rapid acceleration in uptake of syntactic and semantic standards for interoperability, it is still difficult for users to find which databases support the standards and interfaces that they need. To solve these problems, several groups are developing registries of databases that capture key metadata describing the biological scope, utility, accessibility, ease-of-use and existence of web services allowing interoperability between resources. Here, we describe some of these initiatives including a novel formalism, the Database Description Framework, for describing database operations and functionality and encouraging good database practise. We expect such approaches will result in improved discovery, uptake and utilization of data resources. Database URL: http://www.casimir.org.uk/casimir_ddf.
Asunto(s)
Disciplinas de las Ciencias Biológicas/estadística & datos numéricos , Bases de Datos Factuales , Sistema de Registros , Difusión de la Información , Almacenamiento y Recuperación de la Información , InternetRESUMEN
A report by the International Society for Stem Cell Research (ISSCR)'s Task Force on Unproven Stem Cell Treatments outlines development of resources for patients, their families, and physicians seeking information on stem cell treatments.
Asunto(s)
Investigaciones con Embriones/ética , Investigaciones con Embriones/legislación & jurisprudencia , Internet , Células Madre , Guías como Asunto , Humanos , Sociedades MédicasRESUMEN
The newly available techniques for sensitive proteome analysis and the resulting amount of data require a new bioinformatics focus on automatic methods for spectrum reprocessing and peptide/protein validation. Manual validation of results in such studies is not feasible and objective enough for quality relevant interpretation. The necessity for tools enabling an automatic quality control is, therefore, important to produce reliable and comparable data in such big consortia as the Human Proteome Organization Brain Proteome Project. Standards and well-defined processing pipelines are important for these consortia. We show a way for choosing the right database model, through collecting data, processing these with a decoy database and end up with a quality controlled protein list merged from several search engines, including a known false-positive rate.