Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 18 de 18
Filtrar
2.
Stud Health Technol Inform ; 245: 1004-1008, 2017.
Artigo em Inglês | MEDLINE | ID: mdl-29295252

RESUMO

Accessing online health content of high quality and reliability presents challenges. Laypersons cannot easily differentiate trustworthy content from misinformed or manipulated content. This article describes complementary approaches for members of the general public and health professionals to find trustworthy content with as little bias as possible. These include the Khresmoi health search engine (K4E), the Health On the Net Code of Conduct (HONcode) and health trust indicator Web browser extensions.


Assuntos
Internet , Ferramenta de Busca , Informática Aplicada à Saúde dos Consumidores , Humanos , Reprodutibilidade dos Testes
3.
Artigo em Inglês | MEDLINE | ID: mdl-26384372

RESUMO

Biomedical professionals have access to a huge amount of literature, but when they use a search engine, they often have to deal with too many documents to efficiently find the appropriate information in a reasonable time. In this perspective, question-answering (QA) engines are designed to display answers, which were automatically extracted from the retrieved documents. Standard QA engines in literature process a user question, then retrieve relevant documents and finally extract some possible answers out of these documents using various named-entity recognition processes. In our study, we try to answer complex genomics questions, which can be adequately answered only using Gene Ontology (GO) concepts. Such complex answers cannot be found using state-of-the-art dictionary- and redundancy-based QA engines. We compare the effectiveness of two dictionary-based classifiers for extracting correct GO answers from a large set of 100 retrieved abstracts per question. In the same way, we also investigate the power of GOCat, a GO supervised classifier. GOCat exploits the GOA database to propose GO concepts that were annotated by curators for similar abstracts. This approach is called deep QA, as it adds an original classification step, and exploits curated biological data to infer answers, which are not explicitly mentioned in the retrieved documents. We show that for complex answers such as protein functional descriptions, the redundancy phenomenon has a limited effect. Similarly usual dictionary-based approaches are relatively ineffective. In contrast, we demonstrate how existing curated data, beyond information extraction, can be exploited by a supervised classifier, such as GOCat, to massively improve both the quantity and the quality of the answers with a +100% improvement for both recall and precision. Database URL: http://eagl.unige.ch/DeepQA4PA/.


Assuntos
Mineração de Dados/métodos , Anotação de Sequência Molecular/métodos , Análise de Sequência de Proteína/métodos , Software , Animais , Humanos
4.
Stud Health Technol Inform ; 180: 204-9, 2012.
Artigo em Inglês | MEDLINE | ID: mdl-22874181

RESUMO

Patent collections contain an important amount of medical-related knowledge, but existing tools were reported to lack of useful functionalities. We present here the development of TWINC, an advanced search engine dedicated to patent retrieval in the domain of health and life sciences. Our tool embeds two search modes: an ad hoc search to retrieve relevant patents given a short query and a related patent search to retrieve similar patents given a patent. Both search modes rely on tuning experiments performed during several patent retrieval competitions. Moreover, TWINC is enhanced with interactive modules, such as chemical query expansion, which is of prior importance to cope with various ways of naming biomedical entities. While the related patent search showed promising performances, the ad-hoc search resulted in fairly contrasted results. Nonetheless, TWINC performed well during the Chemathlon task of the PatOlympics competition and experts appreciated its usability.


Assuntos
Química Farmacêutica/métodos , Mineração de Dados/métodos , Sistemas de Gerenciamento de Base de Dados , Bases de Dados de Produtos Farmacêuticos , Internet , Patentes como Assunto , Ferramenta de Busca/métodos , Interface Usuário-Computador
5.
Stud Health Technol Inform ; 180: 210-4, 2012.
Artigo em Inglês | MEDLINE | ID: mdl-22874182

RESUMO

We present a new approach to perform biomedical documents classification and prioritization for the Comparative Toxicogenomics Database (CTD). This approach is motivated by needs such as literature curation, in particular applied to the human health environment domain. The unique integration of chemical, genes/proteins and disease data in the biomedical literature may advance the identification of exposure and disease biomarkers, mechanisms of chemical actions, and the complex aetiologies of chronic diseases. Our approach aims to assist biomedical researchers when searching for relevant articles for CTD. The task is functionally defined as a binary classification task, where selected articles must also be ranked by order of relevance. We design a SVM classifier, which combines three main feature sets: an information retrieval system (EAGLi), a biomedical named-entity recognizer (MeSH term extraction), a gene normalization (GN) service (NormaGene) and an ad-hoc keyword recognizer for diseases and chemicals. The evaluation of the gene identification module was done on BioCreativeIII test data. Disease normalization is achieved with 95% precision and 93% of recall. The evaluation of the classification was done on the corpus provided by BioCreative organizers in 2012. The approach showed promising performance on the test data.


Assuntos
Indexação e Redação de Resumos/métodos , Mineração de Dados/métodos , Bases de Dados de Compostos Químicos , Bases de Dados Genéticas , Efeitos Colaterais e Reações Adversas Relacionados a Medicamentos/classificação , Publicações Periódicas como Assunto/classificação , Toxicogenética/métodos , Sistemas de Gerenciamento de Base de Dados , Humanos , Interface Usuário-Computador
6.
Stud Health Technol Inform ; 174: 121-5, 2012.
Artigo em Inglês | MEDLINE | ID: mdl-22491124

RESUMO

Health-related information retrieval is complicated by the variety of nomenclatures available to name entities, since different communities of users will use different ways to name a same entity. We present in this report the development and evaluation of a user-friendly interactive Web application aiming at facilitating health-related patent search. Our tool, called TWINC, relies on a search engine tuned during several patent retrieval competitions, enhanced with intelligent interaction modules, such as chemical query, normalization and expansion. While the functionality of related article search showed promising performances, the ad hoc search results in fairly contrasted results. Nonetheless, TWINC performed well during the PatOlympics competition and was appreciated by intellectual property experts. This result should be balanced by the limited evaluation sample. We can also assume that it can be customized to be applied in corporate search environments to process domain and company-specific vocabularies, including non-English literature and patents reports.


Assuntos
Armazenamento e Recuperação da Informação/métodos , Internet , Patentes como Assunto , Ferramenta de Busca/métodos , Interface Usuário-Computador , Inteligência Artificial , Humanos
7.
Stud Health Technol Inform ; 169: 477-81, 2011.
Artigo em Inglês | MEDLINE | ID: mdl-21893795

RESUMO

We present exploratory investigations of multimodal mining to help designing clinical guidelines for antibiotherapy. Our approach is based on the assumption that combining various sources of data, such as the literature, a clinical datawarehouse, as well as information regarding costs will result in better recommendations. Compared to our baseline recommendation system based on a question-answering engine built on top of PubMed, an improvement of +16% is observed when clinical data (i.e. resistance profiles) are injected into the model. In complement to PubMed, an alternative search strategy is reported, which is significantly improved by the use of the combined multimodal approach. These results suggest that combining literature-based discovery with structured data mining can significantly improve effectiveness of decision-support systems for authors of clinical practice guidelines.


Assuntos
Antibacterianos/uso terapêutico , Guias de Prática Clínica como Assunto , Estatística como Assunto/métodos , Algoritmos , Antibacterianos/economia , Sistemas Computacionais , Sistemas de Apoio a Decisões Clínicas , Custos de Medicamentos , Humanos , National Institutes of Health (U.S.) , PubMed , Staphylococcus aureus/metabolismo , Staphylococcus epidermidis/metabolismo , Estados Unidos
8.
Stud Health Technol Inform ; 169: 654-8, 2011.
Artigo em Inglês | MEDLINE | ID: mdl-21893829

RESUMO

This study aims to show that health websites not asking for HONcode certification (Control sample websites A) do not respect elementary ethical standards such as the HONcode. The HONcode quality and ethical standards and the certification process have been developed by the Health on the Net Foundation to improve the transparency of the health and medical information found on the Internet. We compared the compliance with the 8 HONcode principles, and respectively the respect of principles 1 (authority), 4 (assignment), 5 (justification) and 8 (honesty in advertising and editorial policy) by certified websites (A) and by health websites which have not requested the certification (B). The assessment of the HONcode compliance was performed by HON evaluators by the same standards for all type of sites. Results shows that 0.6% of health websites not asking for HONcode certification does respect the eight HONcode ethical standards vs. 89% of certified websites. Regarding the principles 1, 4, 5 and 8, 1.2% of B respect these principles vs. 92% for A. The certification process led health websites to respect the ethical and quality standards such as the HONcode, and disclosing the production process of the health website.


Assuntos
Internet/normas , Informática Médica/métodos , Credenciamento , Ética Médica , Ética Profissional , França , Guias como Assunto , Humanos , Informática Médica/normas , Modelos Estatísticos , Editoração/normas , Controle de Qualidade , Telemedicina/normas
9.
Health Informatics J ; 17(2): 116-26, 2011 Jun.
Artigo em Inglês | MEDLINE | ID: mdl-21712355

RESUMO

We present an experimental mechanism for enriching web content with quality metadata. This mechanism is based on a simple and well-known initiative in the field of the health-related web, the HONcode. The Resource Description Framework (RDF) format and the Dublin Core Metadata Element Set were used to formalize these metadata. The model of trust proposed is based on a quality model for health-related web pages that has been tested in practice over a period of thirteen years. Our model has been explored in the context of a project to develop a research tool that automatically detects the occurrence of quality criteria in health-related web pages.


Assuntos
Bases de Dados Factuais/normas , Disseminação de Informação , Internet/normas , Ferramenta de Busca/normas , Confiança , Humanos , Armazenamento e Recuperação da Informação/normas , Ferramenta de Busca/métodos , Software
10.
Stud Health Technol Inform ; 136: 407-12, 2008.
Artigo em Inglês | MEDLINE | ID: mdl-18487765

RESUMO

Many attempts have been made in the QA domain but no system applicable to the field of health is currently available on the Internet. This paper describes a bilingual French/English question answering system adapted to the health domain and more particularly the detection of the question's model. Indeed, the Question Analyzer module for identifying the question's model has a greater effect on the rest of the QA system. Our original hypothesis for the QA is that a question can be defined by two criteria: type of answer expected and medical type. These two must appear in the step of detection of the model in order to better define the type of question and thus, the corresponding answer. For this, questions were searched on the Internet and then given to experts in order to obtain classifications according to criteria such as type of question and type of medical context as mentioned above. In addition, tests of supervised and non-supervised classification were made to determine features of questions. The result of this first step was that algorithms of classification were chosen. The results obtained showed that categorizers giving the best results were the SVM. Currently, for a set of 100 questions, 84 are well categorized in English and 68 in French according to the type of answer expected. This figures fall to less than 50% for the medical type. Evaluations have showed that the system was good to identify the type of answer expected and could be enhanced for the medical type. It leads us to use an external source of knowledge: UMLS. A future improvement will be the usage of UMLS semantic network to better categorize a query according to the medical domain.


Assuntos
Informação de Saúde ao Consumidor , Armazenamento e Recuperação da Informação , Internet , Computação em Informática Médica , Algoritmos , Inteligência Artificial , Sistemas Computacionais , Sistemas Inteligentes , Humanos , Bases de Conhecimento , Multilinguismo , Processamento de Linguagem Natural , Semântica , Vocabulário Controlado
11.
Stud Health Technol Inform ; 129(Pt 1): 147-51, 2007.
Artigo em Inglês | MEDLINE | ID: mdl-17911696

RESUMO

Legal and technologic trends are making medical records progressively more patient-accessible. In the near future, information technology may make it even easier to provide patients a chance to review their records. One may wonder, however, about the practical use of this technology by patients. Understanding his/her own health record will certainly be one of the main concerns of patients. WRAPIN has been designed to provide patients and citizens with trusted health information. It will help to determine the reliability of documents by checking the ideas contained against established benchmarks, and enable users to determine the relevance of a given document from a page of search results. First, we present what is, in our opinion, the most original and important patient-centred WRAPIN characteristics and functionalities. Then, we compare these characteristics with those of representatives of two main trends in information retrieval: systems based on the popularity of web sites, and on the clustering of web sites. This comparison shows that, even though patients are tempted to use popular search engines, these are not sufficiently specialized in the medical domain to help them understand their own HER.. Finally, we discuss the complexity of medical readings over the Internet and the efforts that are still required in this domain.


Assuntos
Armazenamento e Recuperação da Informação/métodos , Internet , Sistemas Computadorizados de Registros Médicos , Indexação e Redação de Resumos , Humanos , Medical Subject Headings , Acesso dos Pacientes aos Registros , Interface Usuário-Computador
12.
Stud Health Technol Inform ; 129(Pt 1): 705-9, 2007.
Artigo em Inglês | MEDLINE | ID: mdl-17911808

RESUMO

The number of medical websites is constantly growing [1]. Owing to the open nature of the Web, the reliability of information available on the Web is uneven. Internet users are overwhelmed by the quantity of information available on the Web. The situation is even more critical in the medical area, as the content proposed by health websites can have a direct impact on the users' well being. One way to control the reliability of health websites is to assess their quality and to make this assessment available to users. The HON Foundation has defined a set of eight ethical principles. HON's experts are working in order to manually define whether a given website complies with s the required principles. As the number of medical websites is constantly growing, manual expertise becomes insufficient and automatic systems should be used in order to help medical experts. In this paper we present the design and the evaluation of an automatic system conceived for the categorisation of medical and health documents according to he HONcode ethical principles. A first evaluation shows promising results. Currently the system shows 0.78 micro precision and 0.73 F-measure, with 0.06 errors.


Assuntos
Inteligência Artificial , Internet/normas , Informática Médica/normas , Códigos de Ética , Saúde , Humanos , Serviços de Informação/normas , Processamento de Linguagem Natural , Controle de Qualidade
13.
Stud Health Technol Inform ; 129(Pt 2): 1017-21, 2007.
Artigo em Inglês | MEDLINE | ID: mdl-17911869

RESUMO

The Internet provides a great amount of information and has become one of the communication media which is most widely used [1]. However, the problem is no longer finding information but assessing the credibility of the publishers as well as the relevance and accuracy of the documents retrieved from the web. This problem is particularly relevant in the medical area which has a direct impact on the well-being of citizens. In this paper, we assume that the quality of web pages can be controlled, even when a huge amount of documents has to be reviewed. But this must be supported by both specific automatic tools and human expertise. In this context, we present various initiatives of the Health on the Net Foundation informing the citizens about the reliability of the medical content on the web.


Assuntos
Acreditação , Saúde , Serviços de Informação/normas , Internet/normas , Garantia da Qualidade dos Cuidados de Saúde , Indexação e Redação de Resumos , Códigos de Ética , Bases de Dados Factuais , Fundações , Humanos , Serviços de Informação/ética , Armazenamento e Recuperação da Informação , Internet/ética , Medical Subject Headings , Processamento de Linguagem Natural
14.
Stud Health Technol Inform ; 129(Pt 2): 1319-23, 2007.
Artigo em Inglês | MEDLINE | ID: mdl-17911928

RESUMO

Medical institutions produce ever-increasing amount of diverse information. The digital form makes these data available for the use on more than a single patient. Images are no exception to this. However, less is known about how medical professionals search for visual medical information and how they want to use it outside of the context of a single patient. This article analyzes ten months of usage log files of the Health on the Net (HON) medical media search engine. Key words were extracted from all queries and the most frequent terms and subjects were identified. The dataset required much pre-treatment. Problems included national character sets, spelling errors and the use of terms in several languages. The results show that media search, particularly for images, was frequently used. The most common queries were for general concepts (e.g., heart, lung). To define realistic information needs for the ImageCLEFmed challenge evaluation (Cross Language Evaluation Forum medical image retrieval), we used frequent queries that were still specific enough to at least cover two of the three axes on modality, anatomic region, and pathology. Several research groups evaluated their image retrieval algorithms based on these defined topics.


Assuntos
Diagnóstico por Imagem , Armazenamento e Recuperação da Informação , Estudos de Avaliação como Assunto , Humanos , Sistemas de Informação
15.
AMIA Annu Symp Proc ; : 264-8, 2007 Oct 11.
Artigo em Inglês | MEDLINE | ID: mdl-18693839

RESUMO

The detection of ethical issues of web sites aims at selection of information helpful to the reader and is an important concern in medical informatics. Indeed, with the ever-increasing volume of online health information, coupled with its uneven reliability and quality, the public should be aware about the quality of information available online. In order to address this issue, we propose methods for the automatic detection of statements related to ethical principles such as those of the HONcode. For the detection of these statements, we combine two kinds of heterogeneous information: content-based categorizations and URL-based categorizations through application of the machine learning algorithms. Our objective is to observe the quality of categorization through URL's for web pages where categorization through content has been proven to be not precise enough. The results obtained indicate that only some of the principles were better processed.


Assuntos
Algoritmos , Inteligência Artificial , Códigos de Ética , Serviços de Informação/normas , Internet/normas , Saúde , Humanos , Processamento de Linguagem Natural , Controle de Qualidade
16.
Int J Med Inform ; 75(1): 73-85, 2006 Jan.
Artigo em Inglês | MEDLINE | ID: mdl-16377235

RESUMO

OBJECTIVE: After a review of the existing practical solution available to the citizen to retrieve eHealth document, the paper describes an original specialized search engine WRAPIN. METHOD: WRAPIN uses advanced cross lingual information retrieval technologies to check information quality by synthesizing medical concepts, conclusions and references contained in the health literature, to identify accurate, relevant sources. Thanks to MeSH terminology [1] (Medical Subject Headings from the U.S. National Library of Medicine) and advanced approaches such as conclusion extraction from structured document, reformulation of the query, WRAPIN offers to the user a privileged access to navigate through multilingual documents without language or medical prerequisites. RESULTS: The results of an evaluation conducted on the WRAPIN prototype show that results of the WRAPIN search engine are perceived as informative 65% (59% for a general-purpose search engine), reliable and trustworthy 72% (41% for the other engine) by users. But it leaves room for improvement such as the increase of database coverage, the explanation of the original functionalities and an audience adaptability. CONCLUSION: Thanks to evaluation outcomes, WRAPIN is now in exploitation on the HON web site (http://www.healthonnet.org), free of charge. Intended to the citizen it is a good alternative to general-purpose search engines when the user looks up trustworthy health and medical information or wants to check automatically a doubtful content of a Web page.


Assuntos
Armazenamento e Recuperação da Informação/métodos , Internet , Informática Médica , Europa (Continente) , Controle de Qualidade , Software
17.
Stud Health Technol Inform ; 107(Pt 1): 322-6, 2004.
Artigo em Inglês | MEDLINE | ID: mdl-15360827

RESUMO

OBJECTIVES: To cope with medical terms, which present a high variability of expression through a single natural language, in the sense that any term may be reformulated in hundred of different ways. METHODS: A typology of term variants is presented as a systematic approach in order to favour the implementation of an exhaustive solution. Then, an algorithm able to handle all variants is designed. RESULTS: Using MetaMap, single terms are analyzed with a success rate varying between 68 and 88 %; the algorithm presented in this paper improves this situation. CONCLUSIONS: This experience shows that a semantic driven method, based on a thesaurus, provides a satisfactory solution to the problem of variability of a single term. The presented typology is representative of most variants in a language.


Assuntos
Processamento de Linguagem Natural , Terminologia como Assunto , Vocabulário Controlado , Algoritmos , Armazenamento e Recuperação da Informação , Semântica
18.
Stud Health Technol Inform ; 107(Pt 1): 356-60, 2004.
Artigo em Inglês | MEDLINE | ID: mdl-15360834

RESUMO

To realize the potential of the Internet as a source of valuable healthcare information, for the general public, patients or practitioners, it is imperative to establish a validation system based on standards of quality. The WRAPIN project (World-wide online Reliable Advice to Patients and Individuals) from the European Community has this ambitious goal. WRAPIN is a federating system for medical information with an editorial policy of intelligently sharing quality and professional information. The WRAPIN project has two main axes: the efficient and intelligent search of information and the assertion of the trustworthiness of content. This article presents the scientific challenges involved in extracting the knowledge from text-based information in order to better manage the knowledge and the rest of the retrieval proc-ess. Our innovative approach is to efficiently extract MeSH terms from the analyzed documents exploiting UMLS knowledge sources. A benefit has been measured when comparing extraction results. Even if the evaluation is made with a limited corpus, this research work proposes heuristics that can be validated to the whole biomedical domain, and possibly enhanced by the adjunction of other methods.


Assuntos
Armazenamento e Recuperação da Informação/métodos , Medical Subject Headings , Processamento de Linguagem Natural , Unified Medical Language System , Indexação e Redação de Resumos , Humanos , Serviços de Informação/normas , Internet/normas
SELEÇÃO DE REFERÊNCIAS
DETALHE DA PESQUISA