Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 99
Filtrar
Más filtros

Tipo del documento
Intervalo de año de publicación
1.
Adv Health Sci Educ Theory Pract ; 29(4): 1501-1538, 2024 Sep.
Artículo en Inglés | MEDLINE | ID: mdl-38683300

RESUMEN

The learning environment (LE) includes social interactions, organizational culture, structures, and physical and virtual spaces that influence the learning experiences of students. Despite numerous studies exploring the perception of healthcare professional students (HCPS) of their LE, the validity evidence of the utilized questionnaires remains unclear. This scoping review aimed to identify questionnaires used to examine the perception of undergraduate HCPS of their LE and to assess their validity evidence. Five key concepts were used: (1) higher education; (2) questionnaire; (3) LE; (4) perception; and (5) health professions (HP). PubMed, ERIC, ProQuest, and Cochrane databases were searched for studies developing or adapting questionnaires to examine LE. This review employed the APERA standards of validity evidence and Beckman et al. (J Gen Intern Med 20:1159-1164, 2005) interpretation of these standards according to 5 categories: content, internal structure, response process, relation to other variables, and consequences. Out of 41 questionnaires included in this review, the analysis revealed a predominant emphasis on content and internal structure categories. However, less than 10% of the included questionnaires provided information in relation to other variables, consequences, and response process categories. Most of the identified questionnaires received extensive coverage in the fields of medicine and nursing, followed by dentistry. This review identified diverse questionnaires utilized for examining the perception of students of their LE across different HPs. Given the limited validity evidence for existing questionnaires, future research should prioritize the development and validation of psychometric measures. This will ultimately ensure sound and evidence-based quality improvement measures of the LE in HP education programs.


Asunto(s)
Empleos en Salud , Humanos , Encuestas y Cuestionarios , Empleos en Salud/educación , Aprendizaje , Percepción , Estudiantes del Área de la Salud/psicología , Reproducibilidad de los Resultados
2.
Eur Arch Otorhinolaryngol ; 281(4): 1905-1911, 2024 Apr.
Artículo en Inglés | MEDLINE | ID: mdl-38177897

RESUMEN

PURPOSE: This study aimed to assess the validity of simulation-based assessment of ultrasound skills for thyroid ultrasound. METHODS: The study collected validity evidence for simulation-based ultrasound assessment of thyroid ultrasound skills. Experts (n = 8) and novices (n = 21) completed a test containing two tasks and four cases on a virtual reality ultrasound simulator (U/S Mentor's Neck Ultrasound Module). Validity evidence was collected and structured according to Messick's validity framework. The assessments being evaluated included built-in simulator metrics and expert-based evaluations using the Objective Structured Assessment of Ultrasound Skills (OSAUS) scale. RESULTS: Out of 64 built-in simulator metrics, 9 (14.1%) exhibited validity evidence. The internal consistency of these metrics was strong (Cronbach's α = 0.805) with high test-retest reliability (intraclass correlation coefficient = 0.911). Novices achieved an average score of 41.9% (SD = 24.3) of the maximum, contrasting with experts at 81.9% (SD = 16.7). Time comparisons indicated minor differences between experts (median: 359 s) and novices (median: 376.5 s). All OSAUS items differed significantly between the two groups. The correlation between correctly entered clinical findings and the OSAUS scores was 0.748 (p < 0.001). The correlation between correctly entered clinical findings and the metric scores was 0.801 (p < 0.001). CONCLUSION: While simulation-based training is promising, only 14% of built-in simulator metrics could discriminate between novices and ultrasound experts. Already-established competency frameworks such as OSAUS provided strong validity evidence for the assessment of otorhinolaryngology ultrasound competence.


Asunto(s)
Competencia Clínica , Realidad Virtual , Humanos , Reproducibilidad de los Resultados , Ultrasonografía , Simulación por Computador
3.
BMC Med Educ ; 24(1): 15, 2024 Jan 03.
Artículo en Inglés | MEDLINE | ID: mdl-38172820

RESUMEN

BACKGROUND: Ultrasound is a safe and effective diagnostic tool used within several specialties. However, the quality of ultrasound scans relies on sufficiently skilled clinician operators. The aim of this study was to explore the validity of automated assessments of upper abdominal ultrasound skills using an ultrasound simulator. METHODS: Twenty five novices and five experts were recruited, all of whom completed an assessment program for the evaluation of upper abdominal ultrasound skills on a virtual reality simulator. The program included five modules that assessed different organ systems using automated simulator metrics. We used Messick's framework to explore the validity evidence of these simulator metrics to determine the contents of a final simulator test. We used the contrasting groups method to establish a pass/fail level for the final simulator test. RESULTS: Thirty seven out of 60 metrics were able to discriminate between novices and experts (p < 0.05). The median simulator score of the final simulator test including the metrics with validity evidence was 26.68% (range: 8.1-40.5%) for novices and 85.1% (range: 56.8-91.9%) for experts. The internal structure was assessed by Cronbach alpha (0.93) and intraclass correlation coefficient (0.89). The pass/fail level was determined to be 50.9%. This pass/fail criterion found no passing novices or failing experts. CONCLUSIONS: This study collected validity evidence for simulation-based assessment of upper abdominal ultrasound examinations, which is the first step toward competency-based training. Future studies may examine how competency-based training in the simulated setting translates into improvements in clinical performances.


Asunto(s)
Internado y Residencia , Realidad Virtual , Humanos , Competencia Clínica , Simulación por Computador , Ultrasonografía , Reproducibilidad de los Resultados
4.
J Surg Res ; 283: 726-732, 2023 Mar.
Artículo en Inglés | MEDLINE | ID: mdl-36463811

RESUMEN

INTRODUCTION: Despite the importance of simulation-based training for robotic surgery, there is no consensus about its training curricula. Recently, a virtual reality (VR) platform (SimNow, Intuitive, Inc) was introduced with 33 VR drills but without evidence of their validity. As part of our creating a new robotic VR curriculum, we assessed the drills' validity through content mapping and the alignment between learning goals and drill content. METHODS: Three robotically trained surgeons content-mapped all 33 drills for how well the drills incorporated 15 surgery skills and also rated the drills' difficulty, usefulness, relevance, and uniqueness. Drills were added to the new curriculum based on consensus about ratings and historic learner data. The drills were grouped according to similar skill sets and arranged in order of complexity. RESULTS: The 33 drills were judged to have 12/15 surgery skills as primary goals and 13/15 as secondary goals. Twenty of the 33 drills were selected for inclusion in the new curriculum; these had 11/15 skills as primary goals and 11/15 as secondary goals. However, skills regarding energy sources, atraumatic handling, blunt dissection, fine dissection, and running suturing were poorly represented in the drills. Three previously validated inanimate drills were added to the curriculum to address lacking skill domains. CONCLUSIONS: We identified 20 of the 33 SimNow drills as a foundation for a robotic surgery curriculum based on content-oriented evidence. We added 3 other drills to address identified gaps in drill content.


Asunto(s)
Procedimientos Quirúrgicos Robotizados , Robótica , Realidad Virtual , Procedimientos Quirúrgicos Robotizados/educación , Competencia Clínica , Robótica/educación , Curriculum , Simulación por Computador
5.
J Med Internet Res ; 24(3): e32777, 2022 03 08.
Artículo en Inglés | MEDLINE | ID: mdl-35258475

RESUMEN

BACKGROUND: Digital technologies have changed how we manage our health, and eHealth literacy is needed to engage with health technologies. Any eHealth strategy would be ineffective if users' eHealth literacy needs are not addressed. A robust measure of eHealth literacy is essential for understanding these needs. On the basis of the eHealth Literacy Framework, which identified 7 dimensions of eHealth literacy, the eHealth Literacy Questionnaire (eHLQ) was developed. The tool has demonstrated robust psychometric properties in the Danish setting, but validity testing should be an ongoing and accumulative process. OBJECTIVE: This study aims to evaluate validity evidence based on test content, response process, and internal structure of the eHLQ in the Australian community health setting. METHODS: A mixed methods approach was used with cognitive interviewing conducted to examine evidence on test content and response process, whereas a cross-sectional survey was undertaken for evidence on internal structure. Data were collected at 3 diverse community health sites in Victoria, Australia. Psychometric testing included both the classical test theory and item response theory approaches. Methods included Bayesian structural equation modeling for confirmatory factor analysis, internal consistency and test-retest for reliability, and the Bayesian multiple-indicators, multiple-causes model for testing of differential item functioning. RESULTS: Cognitive interviewing identified only 1 confusing term, which was clarified. All items were easy to read and understood as intended. A total of 525 questionnaires were included for psychometric analysis. All scales were homogenous with composite scale reliability ranging from 0.73 to 0.90. The intraclass correlation coefficient for test-retest reliability for the 7 scales ranged from 0.72 to 0.95. A 7-factor Bayesian structural equation modeling using small variance priors for cross-loadings and residual covariances was fitted to the data, and the model of interest produced a satisfactory fit (posterior productive P=.49, 95% CI for the difference between observed and replicated chi-square values -101.40 to 108.83, prior-posterior productive P=.92). All items loaded on the relevant factor, with loadings ranging from 0.36 to 0.94. No significant cross-loading was found. There was no evidence of differential item functioning for administration format, site area, and health setting. However, discriminant validity was not well established for scales 1, 3, 5, 6, and 7. Item response theory analysis found that all items provided precise information at different trait levels, except for 1 item. All items demonstrated different sensitivity to different trait levels and represented a range of difficulty levels. CONCLUSIONS: The evidence suggests that the eHLQ is a tool with robust psychometric properties and further investigation of discriminant validity is recommended. It is ready to be used to identify eHealth literacy strengths and challenges and assist the development of digital health interventions to ensure that people with limited digital access and skills are not left behind.


Asunto(s)
Alfabetización en Salud , Telemedicina , Australia , Teorema de Bayes , Estudios Transversales , Alfabetización en Salud/métodos , Humanos , Psicometría/métodos , Salud Pública , Reproducibilidad de los Resultados , Encuestas y Cuestionarios , Telemedicina/métodos
6.
BMC Med Educ ; 22(1): 177, 2022 Mar 15.
Artículo en Inglés | MEDLINE | ID: mdl-35291995

RESUMEN

BACKGROUND: Most work on the validity of clinical assessments for measuring learner performance in graduate medical education has occurred at the residency level. Minimal research exists on the validity of clinical assessments for measuring learner performance in advanced subspecialties. We sought to determine validity characteristics of cardiology fellows' assessment scores during subspecialty training, which represents the largest subspecialty of internal medicine. Validity evidence included item content, internal consistency reliability, and associations between faculty-of-fellow clinical assessments and other pertinent variables. METHODS: This was a retrospective validation study exploring the domains of content, internal structure, and relations to other variables validity evidence for scores on faculty-of-fellow clinical assessments that include the 10-item Mayo Cardiology Fellows Assessment (MCFA-10). Participants included 7 cardiology fellowship classes. The MCFA-10 item content included questions previously validated in the assessment of internal medicine residents. Internal structure evidence was assessed through Cronbach's α. The outcome for relations to other variables evidence was overall mean of faculty-of-fellow assessment score (scale 1-5). Independent variables included common measures of fellow performance. FINDINGS: Participants included 65 cardiology fellows. The overall mean ± standard deviation faculty-of-fellow assessment score was 4.07 ± 0.18. Content evidence for the MCFA-10 scores was based on published literature and core competencies. Cronbach's α was 0.98, suggesting high internal consistency reliability and offering evidence for internal structure validity. In multivariable analysis to provide relations to other variables evidence, mean assessment scores were independently associated with in-training examination scores (beta = 0.088 per 10-point increase; p = 0.05) and receiving a departmental or institutional award (beta = 0.152; p = 0.001). Assessment scores were not associated with educational conference attendance, compliance with completion of required evaluations, faculty appointment upon completion of training, or performance on the board certification exam. R2 for the multivariable model was 0.25. CONCLUSIONS: These findings provide sound validity evidence establishing item content, internal consistency reliability, and associations with other variables for faculty-of-fellow clinical assessment scores that include MCFA-10 items during cardiology fellowship. Relations to other variables evidence included associations of assessment scores with performance on the in-training examination and receipt of competitive awards. These data support the utility of the MCFA-10 as a measure of performance during cardiology training and could serve as the foundation for future research on the assessment of subspecialty learners.


Asunto(s)
Distinciones y Premios , Cardiología , Competencia Clínica , Evaluación Educacional , Humanos , Reproducibilidad de los Resultados , Estudios Retrospectivos
7.
J Child Sex Abus ; 31(5): 593-615, 2022 Jul.
Artículo en Inglés | MEDLINE | ID: mdl-35469541

RESUMEN

Child sexual abuse is a serious problem in Brazil and requires actions taken together by Justice, Security, Health and Social Welfare to effectively protect and guarantee victims' rights. Professionals working in these fields have difficulty in evaluating cases, owing to limitations of the Brazilian Welfare Network and lack of specialized training. Such difficulty may cause professionals to carry out poorly substantiated assessments and fail to properly protect victims. Instruments to measure professionals' attitudes in the assessment of situations of sexual violence are scarce. As a result, this study aimed to adapt and evaluate validity evidence of the Child Forensic Attitude Scale (CFAS) in the Brazilian context. A total of 177 professionals (86.4% females), with a mean age of 37.6 years (SD = 10.1 years) participated in the survey. The results of the confirmatory factor analysis showed that in the Brazilian context, the scale structure presents three first-order oblique factors, namely "Fear of Not Identifying Abuse" (F-Under), "Fear of Overcalling Abuse" (F-Over) and "Skepticism" (Skep). The internal consistency of the three dimensions was satisfactory (F-Under, α = 0.66, F-Over, α = 0.80, and Skep, α = 0.92). Evidence has shown that the CFAS can be used to evaluate health professionals' attitudes when assessing cases of sexual violence against children and adolescents in Brazil. This instrument can support the assessment of health professionals' attitudes, and it emphasizes the importance of qualifying Brazilian professionals in the Welfare Network services by providing training opportunities regarding work with victims of sexual abuse.


Asunto(s)
Actitud del Personal de Salud , Abuso Sexual Infantil , Encuestas y Cuestionarios , Adolescente , Adulto , Brasil , Niño , Abuso Sexual Infantil/diagnóstico , Comparación Transcultural , Análisis Factorial , Femenino , Humanos , Masculino , Persona de Mediana Edad , Reproducibilidad de los Resultados
8.
J Minim Invasive Gynecol ; 28(11): 1927-1934, 2021 11.
Artículo en Inglés | MEDLINE | ID: mdl-34010696

RESUMEN

STUDY OBJECTIVE: The "illusion of validity" is a cognitive bias in which the ability to interpret and predict surgical performance accurately is overestimated. To address this bias, we assessed participants comparing fundamentals of laparoscopic surgery (FLS) and non-FLS tasks with cadaveric vaginal cuff suturing to determine the most representative simulation task for laparoscopic vaginal cuff suturing. DESIGN: Validity (Messick framework) study comparing FLS and non-FLS tasks with cadaveric vaginal cuff suturing. SETTING: Simulation center cadaver laboratory. PARTICIPANTS: Obstetrics and gynecology residents (n = 21), minimally invasive gynecologic surgery fellows (n = 3), gynecologic surgical subspecialists (n = 4), general obstetrician/gynecologists (n = 10). INTERVENTIONS: Tasks included a simulated vaginal cuff (ipsilateral port placement), needle passage through a metal eyelet loop (contralateral and ipsilateral), and intracorporeal knot tying (contralateral and ipsilateral). Simulation task times were compared with the placement of the first cadaveric vaginal cuff suture time, as well as the in-person and blinded Global Operative Assessment of Laparoscopic Skills (GOALS) score ("relations to other variables" validity evidence). Statistical analyses included Spearman's test of correlation (continuous and ordinal variables) or Wilcoxon rank sum test (categoric variables). MEASUREMENTS AND MAIN RESULTS: There was a stronger association with cadaver cuff suturing time for simulated vaginal cuff suturing time (r = 0.73, p <.001) compared with FLS intracorporeal contralateral suturing time (r = 0.54, p <.001). Additional measures associated with cadaveric performance included subspecialty training (median: 82 vs 185 seconds, p = .002), number of total laparoscopic hysterectomies (r = -0.53, p <.001), number of laparoscopic cuff closures (r = -0.61, p <.001), number of simulated laparoscopic suturing experiences (r = -0.51, p <.001), and eyelet contralateral time (r = 0.52, p <.001). Strong agreement between the in-person and blinded GOALS (intraclass correlation coefficient = 0.80) supports response process evidence. Correlations of cadaver cuff time with in-person (Spearman's r = -0.84, p <.001) and blinded GOALS (r = -0.76, p <.001) supports relations to other variables evidence CONCLUSION: The weaker correlation between FLS suturing and cadaver cuff suturing compared with a simulated vaginal cuff model may lead to an "illusion of validity" for assessment in gynecology. Since gynecology specific validity evidence has not been well established for FLS, we recommend prioritizing the use of a simulated vaginal cuff suturing assessment in addition to FLS.


Asunto(s)
Ilusiones , Laparoscopía , Competencia Clínica , Femenino , Humanos , Técnicas de Sutura , Suturas
9.
J Med Internet Res ; 23(10): e30243, 2021 10 14.
Artículo en Inglés | MEDLINE | ID: mdl-34647897

RESUMEN

BACKGROUND: As health resources and services are increasingly delivered through digital platforms, eHealth literacy is becoming a set of essential capabilities to improve consumer health in the digital era. To understand eHealth literacy needs, a meaningful measure is required. Strong initial evidence for the reliability and construct validity of inferences drawn from the eHealth Literacy Questionnaire (eHLQ) was obtained during its development in Denmark, but validity testing for varying purposes is an ongoing and cumulative process. OBJECTIVE: This study aims to examine validity evidence based on relations to other variables-using data collected with the known-groups approach-to further explore if the eHLQ is a robust tool to understand eHealth literacy needs in different contexts. A priori hypotheses are set for the expected score differences among age, sex, education, and information and communication technology (ICT) use for each of the 7 eHealth literacy constructs represented by the 7 eHLQ scales. METHODS: A Bayesian mediated multiple indicators multiple causes model approach was used to simultaneously identify group differences and test measurement invariance through differential item functioning across the groups, with ICT use as a mediator. A sample size of 500 participants was estimated. Data were collected at 3 diverse health sites in Australia. RESULTS: Responses from 525 participants were included for analysis. Being older was significantly related to lower scores in 4 eHLQ scales, with 3. Ability to actively engage with digital services having the strongest effect (total effect -0.37; P<.001), followed by 1. Using technology to process health information (total effect -0.32; P<.001), 5. Motivated to engage with digital services (total effect -0.21; P=.01), and 7. Digital services that suit individual needs (total effect -0.21; P=.02). However, the effects were only partially mediated by ICT use. Higher education was associated with higher scores in 1. Using technology to process health information (total effect 0.22; P=.01) and 3. Ability to actively engage with digital services (total effect 0.25; P<.001), with the effects mostly mediated by ICT use. Higher ICT use was related to higher scores in all scales except 2. Understanding health concepts and language and 4. Feel safe and in control. Either no or ignorable cases of differential item functioning were found across the 4 groups. CONCLUSIONS: By using a Bayesian mediated multiple indicators multiple causes model, this study provides supportive validity evidence for the eHLQ based on relations to other variables as well as established evidence regarding internal structure related to measurement invariance across the groups for the 7 scales in the Australian community health context. This study also demonstrates that the eHLQ can be used to gain valuable insights into people's eHealth literacy needs to help optimize access and use of digital health and promote health equity.


Asunto(s)
Alfabetización en Salud , Telemedicina , Australia , Teorema de Bayes , Promoción de la Salud , Humanos , Reproducibilidad de los Resultados , Encuestas y Cuestionarios
10.
BMC Med Res Methodol ; 20(1): 130, 2020 05 26.
Artículo en Inglés | MEDLINE | ID: mdl-32456680

RESUMEN

BACKGROUND: Cross-cultural research with patient-reported outcomes measures (PROMs) assumes that the PROM in the target language will measure the same construct in the same way as the PROM in the source language. Yet translation methods are rarely used to qualitatively maximise construct equivalence or to describe the intents of each item to support common understanding within translation teams. This study aimed to systematically investigate the utility of the Translation Integrity Procedure (TIP), in particular the use of item intent descriptions, to maximise construct equivalence during the translation process, and to demonstrate how documented data from the TIP contributes evidence to a validity argument for construct equivalence between translated and source language PROMs. METHODS: Analysis of secondary data was conducted on routinely collected data in TIP Management Grids of translations (n = 9) of the Health Literacy Questionnaire (HLQ) that took place between August 2014 and August 2015: Arabic, Czech, French (Canada), French (France), Hindi, Indonesian, Slovak, Somali and Spanish (Argentina). Two researchers initially independently deductively coded the data to nine common types of translation errors. Round two of coding included an identified 10th code. Coded data were compared for discrepancies, and checked when needed with a third researcher for final code allocation. RESULTS: Across the nine translations, 259 changes were made to provisional forward translations and were coded into 10 types of errors. Most frequently coded errors were Complex word or phrase (n = 99), Semantic (n = 54) and Grammar (n = 27). Errors coded least frequently were Cultural errors (n = 7) and Printed errors (n = 5). CONCLUSIONS: To advance PROM validation practice, this study investigated a documented translation method that includes the careful specification of descriptions of item intents. Assumptions that translated PROMs have construct equivalence between linguistic contexts can be incorrect due to errors in translation. Of particular concern was the use of high level complex words by translators, which, if undetected, could cause flawed interpretation of data from people with low literacy. Item intent descriptions can support translations to maximise construct equivalence, and documented translation data can contribute evidence to justify score interpretation and use of translated PROMS in new linguistic contexts.


Asunto(s)
Alfabetización en Salud , Canadá , Francia , Humanos , Psicometría , Encuestas y Cuestionarios , Traducciones
11.
Adv Health Sci Educ Theory Pract ; 24(1): 45-63, 2019 03.
Artículo en Inglés | MEDLINE | ID: mdl-30171512

RESUMEN

Learning curves can support a competency-based approach to assessment for learning. When interpreting repeated assessment data displayed as learning curves, a key assessment question is: "How well is each learner learning?" We outline the validity argument and investigation relevant to this question, for a computer-based repeated assessment of competence in electrocardiogram (ECG) interpretation. We developed an on-line ECG learning program based on 292 anonymized ECGs collected from an electronic patient database. After diagnosing each ECG, participants received feedback including the computer interpretation, cardiologist's annotation, and correct diagnosis. In 2015, participants from a single institution, across a range of ECG skill levels, diagnosed at least 60 ECGs. We planned, collected and evaluated validity evidence under each inference of Kane's validity framework. For Scoring, three cardiologists' kappa for agreement on correct diagnosis was 0.92. There was a range of ECG difficulty across and within each diagnostic category. For Generalization, appropriate sampling was reflected in the inclusion of a typical clinical base rate of 39% normal ECGs. Applying generalizability theory presented unique challenges. Under the Extrapolation inference, group learning curves demonstrated expert-novice differences, performance increased with practice and the incremental phase of the learning curve reflected ongoing, effortful learning. A minority of learners had atypical learning curves. We did not collect Implications evidence. Our results support a preliminary validity argument for a learning curve assessment approach for repeated ECG interpretation with deliberate and mixed practice. This approach holds promise for providing educators and researchers, in collaboration with their learners, with deeper insights into how well each learner is learning.


Asunto(s)
Educación de Pregrado en Medicina/métodos , Evaluación Educacional/métodos , Electrocardiografía/métodos , Curva de Aprendizaje , Competencia Clínica , Educación Basada en Competencias , Educación a Distancia , Educación de Pregrado en Medicina/normas , Evaluación Educacional/normas , Electrocardiografía/normas , Retroalimentación Formativa , Humanos , Internet , Reproducibilidad de los Resultados
12.
Am J Drug Alcohol Abuse ; 45(4): 365-376, 2019.
Artículo en Inglés | MEDLINE | ID: mdl-30640570

RESUMEN

Background: Interpretation bias tasks such as word association tests have shown a moderate relation with substance use, but most studies have been conducted in nonclinical samples and these tasks are difficult to rate. Objectives: To provide: (1) reliability evidence of the Word Association Task for Drug Use Disorder (WAT-DUD), a novel and easy-to-rate instrument for measuring interpretation bias and (2) validity evidence based on the relationship between the WAT-DUD and variables associated with patterns of drug use and treatment outcomes. Methods: 186 patients (67 outpatients and 119 inpatients, 90% males) participated in the study. The task consisted of a simultaneous conditional discrimination where an image (either explicit or ambiguous) was the sample and two words (drug-related or not) served as comparison stimuli. The Substance Dependence Severity Scale, the Cocaine Craving Questionnaire-Now, and the Multidimensional Craving Scale were also used. Results: The ambiguous images items showed adequate reliability in terms of internal consistency (α = .80) and test-retest reliability (79.7% on average). The interpretation of images as drug-related was positively correlated with craving for cocaine (r = .20; p = .029), alcohol (r = .30; p = . 01), and alcohol withdrawal (r = .31; p = .01) along with severity of alcohol dependence (r = .23; p = .04). No relationship was found with the severity of cocaine dependence, or its symptoms of abstinence. Conclusion: WAT-DUD shows psychometric properties that support its use in research contexts, although more research is needed for its use in the clinical setting.


Asunto(s)
Psicometría/instrumentación , Trastornos Relacionados con Sustancias/psicología , Pruebas de Asociación de Palabras , Adulto , Alcoholismo/psicología , Reacción de Prevención , Sesgo , Trastornos Relacionados con Cocaína/psicología , Correlación de Datos , Ansia , Femenino , Humanos , Masculino , Persona de Mediana Edad , Datos Preliminares , Reproducibilidad de los Resultados , España , Síndrome de Abstinencia a Sustancias/psicología
13.
BMC Med Educ ; 18(1): 313, 2018 Dec 20.
Artículo en Inglés | MEDLINE | ID: mdl-30572876

RESUMEN

BACKGROUND: To prevent the problems of traditional clinical evaluation, the "Objective Structured Clinical Examination (OSCE)" was presented by Harden as a more valid and reliable assessment instrument. However, an essential condition to guarantee a high-quality and effective OSCE is the assurance of evidence to support the validity of its scores. This study examines the psychometric properties of OSCE scores, with an emphasis on consequential and internal structure validity evidence. METHODS: Fifty-three first year medical students took part in a summative OSCE at the Lebanese American University-School of Medicine. Evidence to support consequential validity was gathered by using criterion-based standard setting methods. Internal structure validity evidence was gathered by examining various psychometric measures both at the station level and across the complete OSCE. RESULTS: Compared to our actual method of computing results, the introduction of standard setting resulted in lower students' average grades and a higher cut score. Across stations, Cronbach's alpha was moderately low. CONCLUSION: Gathering consequential and internal structure validity evidence by multiple metrics provides support for or against the quality of an OSCE. It is critical that this analysis be performed routinely on local iterations of given tests, and the results used to enhance the quality of assessment.


Asunto(s)
Competencia Clínica/normas , Educación de Pregrado en Medicina/normas , Evaluación Educacional/métodos , Facultades de Medicina , Estudiantes de Medicina , Competencia Clínica/estadística & datos numéricos , Educación Basada en Competencias , Humanos , Psicometría , Reproducibilidad de los Resultados
14.
BMC Med Educ ; 18(1): 176, 2018 Aug 01.
Artículo en Inglés | MEDLINE | ID: mdl-30068394

RESUMEN

BACKGROUND: Trainees in medical subspecialties lack validated assessment scores that can be used to prepare for their licensing examination. This paper presents the development, administration, and validity evidence of a constructed-response preparatory test (CRPT) administered to meet the needs of nephrology trainees. METHODS: Learning objectives from the licensing examination were used to develop a test blueprint for the preparatory test. Messick's unified validity framework was used to gather validity evidence for content, response process, internal structure, relations to other variables, and consequences. Questionnaires were used to gather data on the trainees' perception of examination preparedness, item clarity, and curriculum adequacy. RESULTS: There were 10 trainees and 5 faculty volunteers who took the test. The majority of trainees passed the constructed-response preparatory test. However, many scored poorly on items assessing renal pathology and physiology knowledge. We gathered the following five sources of validity evidence: (1) Content: CRPT mapped to the licensing examination blueprint, with items demonstrating clarity and range of difficulty; (2) Response process: moderate rater agreement (intraclass correlation = .58); (3) Internal structure: sufficient reliability based on generalizability theory (G-coefficient = .76 and Φ-coefficient = .53); (4) Relations to other variables: CRPT scores reflected years of exposure in nephrology and clinical practice; (5) Consequences: post-assessment survey revealed that none of the test takers felt "poorly prepared" for the upcoming summative examination and that their studying would increase in duration and be adapted in terms of content focus. CONCLUSIONS: Preparatory tests using constructed response items mapped to licensure examination blueprint can be developed and used at local program settings to help prepare learners for subspecialty licensure examinations. The CRPT and questionnaire data identified shortcomings of the nephrology training program curriculum. Following the preparatory test, trainees expressed an improved sense of preparedness for their licensing examination.


Asunto(s)
Competencia Clínica , Evaluación Educacional/métodos , Concesión de Licencias , Nefrología/educación , Acreditación , Curriculum , Humanos , Proyectos Piloto , Reproducibilidad de los Resultados , Habilidades para Tomar Exámenes
15.
J Ultrasound Med ; 36(12): 2475-2483, 2017 Dec.
Artículo en Inglés | MEDLINE | ID: mdl-28646627

RESUMEN

OBJECTIVES: The aim of this study was to examine the validity of a simulator test designed to evaluate focused assessment with sonography for trauma (FAST) skills. METHODS: Participants included a group of ultrasound novices (n = 25) and ultrasound experts (n = 10). All participants had their FAST skills assessed using a virtual reality ultrasound simulator. Procedural performance on the 4 FAST windows was assessed by automated simulator metrics, which received a passing or failing score. The validity evidence for these simulator metrics was examined by a stepwise approach according to the Standards for Educational and Psychological Testing. Metrics with validity evidence were included in a simulator test, and the reliability of test scores was determined. Finally, a pass/fail level for procedural performance was established. RESULTS: Of the initial 55 metrics, 34 (61.8%) had validity evidence (P < .01). A simulator test was constructed based on the 34 metrics with established validity evidence, and test scores were calculated as percentages of the maximum score. The median simulator test scores were 14.7% (range, 0%-47.1%) and 94.1% (range, 94.1%-100%) for novices and experts, respectively (P < .001). The pass/fail level was determined to be 79.7%. CONCLUSIONS: The performance of FAST examinations can be assessed in a simulated setting using defensible performance standards, which have both good reliability and validity.


Asunto(s)
Competencia Clínica/estadística & datos numéricos , Sistemas de Atención de Punto , Ultrasonografía/normas , Adulto , Simulación por Computador , Femenino , Humanos , Masculino , Persona de Mediana Edad , Reproducibilidad de los Resultados , Ultrasonografía/métodos
16.
Surg Endosc ; 30(6): 2199-206, 2016 06.
Artículo en Inglés | MEDLINE | ID: mdl-26310528

RESUMEN

INTRODUCTION: Patients, clinicians and researchers seek an easy, reproducible and valid measure of postoperative recovery. The six-minute walk test (6MWT) is a low-cost measure of physical function, which is a relevant dimension of recovery. The aim of the present study was to contribute further evidence for the validity of the 6MWT as a measure of postoperative recovery after colorectal surgery. METHODS: This study involved a sample of 174 patients enrolled in three previous randomized controlled trials. Construct validity was assessed by testing the hypotheses that the distance walked in 6 min (6MWD) at 4 weeks after surgery is greater (1) in younger versus older patients, (2) in patients with higher preoperative physical status versus lower, (3) after laparoscopic versus open surgery, (4) in patients without postoperative complications versus with postoperative complications; and that 6MWD (5) correlates cross-sectionally with self-reported physical activity as measured with a questionnaire (CHAMPS). Statistical analysis was performed using linear regression and Spearman's correlation. The COnsensus-based Standards for the selection of health Measurement INstruments (COSMIN) checklist was used to guide the formulation of hypotheses and reporting of results. RESULTS: One hundred and fifty-one patients who completed the 6MWT at 4 weeks after surgery were included in the analysis. All hypotheses tested for construct validity were supported by the data. Older age, poorer physical status, open surgery and occurrence of postoperative complications were associated with clinically relevant reduction in 6MWD (>19 m). There was a moderate positive correlation between 6MWD and patient-reported physical activity (r = 0.46). CONCLUSIONS: This study contributes further evidence for the construct validity of the 6MWT as a measure of postoperative recovery after colorectal surgery. Results from this study support the use of the 6MWT as an outcome measure in studies evaluating interventions aimed to improve postoperative recovery.


Asunto(s)
Enfermedades del Colon/cirugía , Cirugía Colorrectal/rehabilitación , Laparoscopía/rehabilitación , Complicaciones Posoperatorias/fisiopatología , Prueba de Paso , Anciano , Enfermedades del Colon/fisiopatología , Femenino , Indicadores de Salud , Humanos , Masculino , Evaluación de Resultado en la Atención de Salud , Complicaciones Posoperatorias/rehabilitación , Periodo Posoperatorio , Reproducibilidad de los Resultados , Encuestas y Cuestionarios , Caminata
17.
Adv Health Sci Educ Theory Pract ; 20(5): 1263-89, 2015 Dec.
Artículo en Inglés | MEDLINE | ID: mdl-25808311

RESUMEN

The authors report final-year ward simulation data from the University of Dundee Medical School. Faculty who designed this assessment intend for the final score to represent an individual senior medical student's level of clinical performance. The results are included in each student's portfolio as one source of evidence of the student's capability as a practitioner, professional, and scholar. Our purpose in conducting this study was to illustrate how assessment designers who are creating assessments to evaluate clinical performance might develop propositions and then collect and examine various sources of evidence to construct and evaluate a validity argument. The data were from all 154 medical students who were in their final year of study at the University of Dundee Medical School in the 2010-2011 academic year. To the best of our knowledge, this is the first report on an analysis of senior medical students' clinical performance while they were taking responsibility for the management of a simulated ward. Using multi-facet Rasch measurement and a generalizability theory approach, we examined various sources of validity evidence that the medical school faculty have gathered for a set of six propositions needed to support their use of scores as measures of students' clinical ability. Based on our analysis of the evidence, we would conclude that, by and large, the propositions appear to be sound, and the evidence seems to support their proposed score interpretation. Given the body of evidence collected thus far, their intended interpretation seems defensible.


Asunto(s)
Competencia Clínica , Evaluación Educacional/métodos , Evaluación Educacional/normas , Simulación de Paciente , Comunicación , Femenino , Humanos , Relaciones Interprofesionales , Masculino , Seguridad del Paciente , Relaciones Médico-Paciente , Profesionalismo , Distribución Aleatoria , Reproducibilidad de los Resultados
18.
Psicol Reflex Crit ; 37(1): 18, 2024 May 07.
Artículo en Inglés | MEDLINE | ID: mdl-38710873

RESUMEN

BACKGROUND: The International Trauma Questionnaire (ITQ) is used to measure posttraumatic stress disorder (PTSD) and complex posttraumatic stress disorder (CPTSD) symptoms, and the Posttraumatic Cognitions Inventory-9 (PTCI-9) is used to measure posttraumatic cognitions. Both tools have been translated for use in Brazil. However, the psychometric properties of the Brazilian versions were not investigated, and no study has verified the invariance of these tools for many traumatic event types. OBJECTIVE: This study examined the validity, reliability, and measurement invariance of the Brazilian versions of the ITQ and the PTCI-9 for trauma type, gender, race, age group, education level, and geographical region. METHODS: A total of 2,111 people (67.74% women) participated in an online survey. The scale models were tested via confirmatory factor analyses and measurement invariance through multigroup analyses. Pearson's correlation analyses were used to examine the relationships between PTSD, CPTSD, posttraumatic cognitions, and depressive symptoms. RESULTS: Except for the affective dysregulation factor, the reliabilities of the ITQ and PTCI-9 dimensions were adequate. Models with six correlated dimensions for the ITQ and three correlated dimensions for the PTCI-9 showed adequate fit to the data. The ITQ and PTCI-9 exhibited scalar invariance for gender, race, age group, education level, and geographical region. The ITQ also demonstrated full invariance for trauma type. The factors of both instruments were related to each other and to depressive symptoms, with higher effect sizes for posttraumatic cognitions and complex posttraumatic stress disorder symptoms. CONCLUSION: We recommend using the Brazilian versions of the ITQ and PTCI-9, which are crucial tools for assessing and treating trauma-related disorders.

19.
Sci Rep ; 14(1): 20473, 2024 09 03.
Artículo en Inglés | MEDLINE | ID: mdl-39227677

RESUMEN

Breast cancer impacts the psychological well-being of women, leaving them at risk of developing depression, anxiety, and other stress-related disorders. The Depression Anxiety Stress Scales (DASS-21) is a widely used measure, although empirical evidence regarding its psychometric properties in the breast cancer population is limited. The purpose of this study was to conduct an exhaustive analysis of the psychometric properties of the DASS-21 in a sample of Spanish women diagnosed with breast cancer. Participants were 289 breast cancer patients who completed the DASS-21 and other questionnaires measuring life satisfaction, positive and negative affect, flourishing, perceived stress, and breast cancer-specific stressors. In terms of validity evidence based on the internal structure of the DASS-21, adequate fit indices were obtained for the model based on three first-order factors (depression, anxiety, stress) and one second-order factor (general psychological distress). Reliability coefficients (McDonald's omega) ranged from .84 to .95. Validity evidence based on relationships with other variables was also provided by moderate and strong correlations with well-being indicators and stress measures. The results support the use of the DASS-21 for measuring general psychological distress in the breast cancer context, where it may provide useful information for the design of psychological interventions with patients.


Asunto(s)
Ansiedad , Neoplasias de la Mama , Depresión , Psicometría , Estrés Psicológico , Humanos , Femenino , Neoplasias de la Mama/psicología , Psicometría/métodos , Persona de Mediana Edad , Depresión/diagnóstico , Depresión/psicología , Ansiedad/diagnóstico , Ansiedad/psicología , Estrés Psicológico/diagnóstico , Adulto , Anciano , Encuestas y Cuestionarios , Reproducibilidad de los Resultados , Escalas de Valoración Psiquiátrica
20.
Autism ; : 13623613241228887, 2024 Feb 17.
Artículo en Inglés | MEDLINE | ID: mdl-38366857

RESUMEN

LAY ABSTRACT: The broad autism phenotype refers to a group of behaviors related to autism spectrum disorder, but that appear to a lesser extent. Its assessment has been performed through outdated broad autism phenotype/autism spectrum disorder definitions and tests. To address this problem, this study presents the development of a new test, the Broad Autism Phenotype-International Test, a 20-item measure consisting of two dimensions, SOCIAL-BAP and RIRE-BAP, targeting the two-domain operationalization of autism spectrum disorder in Spain and the United Kingdom. Unlike the Broad Autism Phenotype Questionnaire, this test received empirical support as a quick and effective broad autism phenotype measure that can facilitate both broad autism phenotype/autism spectrum disorder research and interventions. This is the first step to studying the BAP in several Spanish and English-speaking countries.

SELECCIÓN DE REFERENCIAS
DETALLE DE LA BÚSQUEDA