Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 66
Filtrar
Mais filtros

Base de dados
País/Região como assunto
Tipo de documento
País de afiliação
Intervalo de ano de publicação
1.
Adv Health Sci Educ Theory Pract ; 25(3): 629-639, 2020 08.
Artigo em Inglês | MEDLINE | ID: mdl-31720878

RESUMO

As medical schools have changed their curricula to address foundational and clinical sciences in a more integrated fashion, teaching methods such as concept mapping have been incorporated in small group learning settings. Methods that can assess students' ability to apply such integrated knowledge are not as developed, however. The purpose of this project was to assess the validity of scores on a focused version of concept maps called mechanistic case diagrams (MCDs), which are hypothesized to enhance existing tools for assessing integrated knowledge that supports clinical reasoning. The data were from the medical school graduating class of 2018 (N = 136 students). In 2014-2015 we implemented a total of 16 case diagrams in case analysis groups within the Mechanisms of Health and Disease (MOHD) strand of the pre-clinical curriculum. These cases were based on topics being taught during the lectures and small group sessions for MOHD. We created an overall score across all 16 cases for each student. We then correlated these scores with performance in the preclinical curriculum [as assessed by overall performance in MOHD integrated foundational basic science courses and overall performance in the Clinical and Professional Skills (CAPS) courses], and standardized licensing exam scores [United States Medical Licensing Exam (USMLE)] Step 1 (following core clerkships) and Step 2 Clinical Knowledge (at the beginning of the fourth year of medical school). MCD scores correlated with students' overall basic science scores (r = .46, p = .0002) and their overall performance in Clinical and Professional Skills courses (r = .49, p < .0001). In addition, they correlated significantly with standardized exam measures, including USMLE Step 1 (r = .33, p ≤ .0001), and USMLE Step 2 CK (r = .39, p < .0001). These results provide preliminary validity evidence that MCDs may be useful in identifying students who have difficulty in integrating foundational and clinical sciences.


Assuntos
Formação de Conceito , Currículo , Internet , Ciência/educação , Integração de Sistemas , Competência Clínica , Diagnóstico Diferencial , Projetos Piloto
2.
Teach Learn Med ; 28(3): 279-85, 2016.
Artigo em Inglês | MEDLINE | ID: mdl-27092723

RESUMO

UNLABELLED: Construct/Background: Medical school grades are currently unstandardized, and their level of reliability is unknown. This means their usefulness for reporting on student achievement is also not well documented. This study investigates grade reliability within 1 medical school. APPROACH: Generalizability analyses are conducted on grades awarded. Grades from didactic and clerkship-based courses were treated as 2 levels of a fixed facet within a univariate mixed model. Grades from within the 2 levels (didactic and clerkship) were also entered in a multivariate generalizability study. RESULTS: Grades from didactic courses were shown to produce a highly reliable mean score (G = .79) when averaged over as few as 5 courses. Although the universe score correlation between didactic and clerkship courses was high (r = .80), the clerkship courses required almost twice as many grades to reach a comparable level of reliability. When grades were converted to a Pass/Fail metric, almost all information contained in the grades was lost. CONCLUSIONS: Although it has been suggested that the imprecision of medical school grades precludes their use as a reliable indicator of student achievement, these results suggest otherwise. While it is true that a Pass/Fail system of grading provides very little information about a student's level of performance, a multi-tiered grading system was shown to be a highly reliable indicator of student achievement within the medical school. Although grades awarded during the first 2 didactic years appear to be more reliable than clerkship grades, both yield useful information about student performance within the medical college.


Assuntos
Educação Médica/normas , Avaliação Educacional/normas , Logro , Humanos , Iowa , Modelos Estatísticos , Reprodutibilidade dos Testes
3.
Teach Learn Med ; 27(2): 197-200, 2015.
Artigo em Inglês | MEDLINE | ID: mdl-25893942

RESUMO

ISSUE: The research published outside of medical education journals provides an important source of validity evidence for using cognitive ability testing in medical school admissions. EVIDENCE: The cumulative body of validity research, consisting of thousands of studies and scores of meta-analyses, has conclusively demonstrated that a strong positive relationship exists between job performance and general mental ability. IMPLICATIONS: Recommendations for reducing the emphasis on or eliminating the role of general mental ability in the selection process for medical schools are not based on a consideration of the wider research evidence. Admission interventions that substantially reduce the level of academic aptitude are also likely to result in reduced professional performance.


Assuntos
Teste de Admissão Acadêmica , Valor Preditivo dos Testes , Critérios de Admissão Escolar , Faculdades de Medicina , Competência Clínica , Previsões , Humanos , Aprendizagem , Estados Unidos
4.
Anesth Analg ; 116(6): 1342-51, 2013 Jun.
Artigo em Inglês | MEDLINE | ID: mdl-23558839

RESUMO

BACKGROUND: A study by de Oliveira Filho et al. reported a validated set of 9 questions by which Brazilian anesthesia residents assessed faculty supervision in the operating room. The aim of this study was to use this question set to determine whether faculty operating room supervision scores were associated with residents' year of clinical anesthesia training and/or number of specific resident-faculty interactions. We also characterized associations between faculty operating room supervision scores and resident assessments of: (1) faculty supervision in settings other than operating rooms, (2) faculty clinical ability (family choice), and (3) faculty teaching effectiveness. Finally, we characterized the psychometric properties of the de Oliveira Filho etal. question set in an United States anesthesia residency program. METHODS: All 39 residents in the Department of Anesthesia of the University of Iowa in their first (n = 14), second (n = 13), or third (n = 12) year of clinical anesthesia training evaluated the supervision provided by all anesthesia faculty who staffed in at least 1 of 3 clinical settings (operating room [n = 49], surgical intensive care unit [n = 10], pain clinic [n = 6]). For all resident-faculty pairs, departmental billing data were used to quantitate the number of resident-faculty interactions and the interval between the last interaction and the assessment. A generalizability study was performed to determine the minimum number of resident evaluations needed for high reliability and dependability. RESULTS: There were no significant associations between faculty mean operating room supervision scores and: (1) resident-faculty patient encounters (Kendall τb = 0.01; 95% confidence interval [CI], -0.02 to +0.04; P = 0.71), (2) resident-faculty days of interaction (τb = -0.01; 95% CI, -0.05 to +0.02; P = 0.46), and (3) days since last resident-faculty interaction (τb = 0.01; 95% CI, -0.02 to 0.05; P = 0.49). Supervision scores for the operating room and surgical intensive care unit were highly correlated (τb = 0.71; 95% CI, 0.63 to 0.78; P < 0.0001). Supervision scores for the operating room also were highly correlated with family choice scores (τb = 0.77; 95% CI, 0.70 to 0.84; P < 0.0001) and teaching scores (τb = 0.87; 95% CI, 0.82 to 0.92; P < 0.0001). High reliability and dependability (both G- and ϕ-coefficients > 0.80) occurred when individual faculty anesthesiologists received assessments from 15 or more different residents. CONCLUSION: Supervision scores provided by all residents can be given equal weight when calculating an individual faculty anesthesiologist's mean supervision score. Assessments of supervision, teaching, and quality of clinical care are highly correlated. When the de Oliveira Filho et al. question set is used in a United States anesthesia residency program, supervision scores are highly reliable and dependable when at least 15 residents assess each faculty.


Assuntos
Anestesiologia/educação , Docentes de Medicina , Internato e Residência , Salas Cirúrgicas/organização & administração , Psicometria , Humanos , Organização e Administração , Reprodutibilidade dos Testes
5.
Med Educ ; 47(12): 1175-83, 2013 Dec.
Artigo em Inglês | MEDLINE | ID: mdl-24206151

RESUMO

CONTEXT: Recent reviews have claimed that the script concordance test (SCT) methodology generally produces reliable and valid assessments of clinical reasoning and that the SCT may soon be suitable for high-stakes testing. OBJECTIVES: This study is intended to describe three major threats to the validity of the SCT not yet considered in prior research and to illustrate the severity of these threats. METHODS: We conducted a review of SCT reports available through the Web of Science database. Additionally, we reanalysed scores from a previously published SCT administration to explore issues related to standard SCT scoring practice. RESULTS: Firstly, the predominant method for aggregate and partial credit scoring of SCTs introduces logical inconsistencies in the scoring key. Secondly, our literature review shows that SCT reliability studies have generally ignored inter-panel, inter-panellist and test-retest measurement error. Instead, studies have focused on observed levels of coefficient alpha, which is neither an informative index of internal structure nor a comprehensive index of reliability for SCT scores. As such, claims that SCT scores show acceptable reliability are premature. Finally, SCT criteria for item inclusion, in concert with a statistical artefact of the SCT format, cause anchors at the extremes of the scale to have less expected credit than anchors near or at the midpoint. Consequently, SCT scores are likely to reflect construct-irrelevant differences in examinees' response styles. This makes the test susceptible to bias against candidates who endorse extreme scale anchors more readily; it also makes two construct-irrelevant test taking strategies extremely effective. In our reanalysis, we found that examinees could drastically increase their scores by never endorsing extreme scale points. Furthermore, examinees who simply endorsed the scale midpoint for every item would still have outperformed most examinees who used the scale as it is intended. CONCLUSIONS: Given the severity of these threats, we conclude that aggregate scoring of SCTs cannot be recommended. Recommendations for revisions of SCT methodology are discussed.


Assuntos
Educação de Pós-Graduação em Medicina/métodos , Avaliação Educacional/métodos , Competência Clínica/normas , Tomada de Decisões , Humanos , Reprodutibilidade dos Testes
6.
Adv Health Sci Educ Theory Pract ; 18(4): 835-44, 2013 Oct.
Artigo em Inglês | MEDLINE | ID: mdl-22886140

RESUMO

Given medical education's longstanding emphasis on assessment, it seems prudent to evaluate whether our current research and development focus on testing makes sense. Since any intervention within medical education must ultimately be evaluated based upon its impact on student learning, this report seeks to provide a quantitative accounting of the learning gains attained through educational assessments. To approach this question, we estimate achieved learning within a medical school environment that optimally utilizes educational assessments. We compare this estimate to learning that might be expected in a medical school that employs no educational assessments. Effect sizes are used to estimate testing's total impact on learning by summarizing three effects; the direct effect, the indirect effect, and the selection effect. The literature is far from complete, but the available evidence strongly suggests that each of these effects is large and the net cumulative impact on learning in medical education is over two standard deviations. While additional evidence is required, the current literature shows that testing within medical education makes a strong positive contribution to learning.


Assuntos
Avaliação Educacional/métodos , Aprendizagem , Estudantes de Medicina/psicologia , Humanos
7.
Teach Learn Med ; 25(1): 103-7, 2013.
Artigo em Inglês | MEDLINE | ID: mdl-23330903

RESUMO

BACKGROUND. Admission decisions require that information about an applicant be combined using either holistic (human judges) or statistical (actuarial) methods. For optimizing a defined measureable outcome, there is a consistent body of research evidence demonstrating that statistical methods yield superior decisions compared to those generated by judges. It is possible, however, that the benefits of holistic decisions are reflected in unmeasured outcomes. If such benefits exist, they would necessarily appear as systematic variance in raters' scores that deviate from statistically-based decisions. PURPOSE. To estimate this variance, we propose a design examining the interrater reliability of difference scores (i.e., the difference between observed committee rankings and rankings based on statistical approaches). METHODS. Example calculations and G study models are presented to demonstrate how rater agreement on difference scores can be analyzed under various circumstances. High interrater reliability of difference scores would support but not prove the assertion that the holistic process adds useful information beyond that achieved by much less costly statistical approaches. Conversely, if the interrater reliability of difference scores is near zero, this would clearly demonstrate that committee judgments add random error to the decision process. RESULTS. Evidence to conduct such studies already exists within most highly selective medical schools and graduate programs and the proposed validity research could be conducted on existing data. CONCLUSIONS. Such research evidence is critical for establishing the validity of widely used holistic admission approaches.


Assuntos
Critérios de Admissão Escolar , Faculdades de Medicina , Humanos , Reprodutibilidade dos Testes , Estados Unidos
8.
Teach Learn Med ; 25 Suppl 1: S50-6, 2013.
Artigo em Inglês | MEDLINE | ID: mdl-24246107

RESUMO

Over the last 25 years a large body of research has investigated how best to select applicants to study medicine. Although these studies have inspired little actual change in admission practice, the implications of this research are substantial. Five areas of inquiry are discussed: (1) the interview and related techniques, (2) admission tests, (3) other measures of personal competencies, (4) the decision process, and (5) defining and measuring the criterion. In each of these areas we summarize consequential developments and discuss their implication for improving practice. (1) The traditional interview has been shown to lack both reliability and validity. Alternatives have been developed that display promising measurement characteristics. (2) Admission test scores have been shown to predict academic and clinical performance and are generally the most useful measures obtained about an applicant. (3) Due to the high-stakes nature of the admission decision, it is difficult to support a logical validity argument for the use of personality tests. Although standardized letters of recommendation appear to offer some promise, more research is needed. (4) The methods used to make the selection decision should be responsive to validity research on how best to utilize applicant information. (5) Few resources have been invested in obtaining valid criterion measures. Future research might profitably focus on composite score as a method for generating a measure of a physician's career success. There are a number of social and organization factors that resist evidence-based change. However, research over the last 25 years does present important findings that could be used to improve the admission process.


Assuntos
Tomada de Decisões , Critérios de Admissão Escolar , Faculdades de Medicina/organização & administração , Estudantes de Medicina/psicologia , Logro , Aptidão , Educação Pré-Médica , Avaliação Educacional , Humanos , Entrevistas como Assunto , Princípios Morais , Personalidade , Competência Profissional , Testes Psicológicos , Resiliência Psicológica , Comportamento Social
9.
Teach Learn Med ; 24(2): 101-5, 2012.
Artigo em Inglês | MEDLINE | ID: mdl-22490088

RESUMO

BACKGROUND: Although the existing psychometric literature provides guidance on the best method for acquiring a reliable clinical evaluation form (CEF)-based score, it also shows that a single CEF rating has very low reliability. PURPOSE: This study examines whether experience with rating students might act as a form of rater training and hence improve the quality of CEF ratings. METHODS: Preceptors were divided into two groups based on rater experience. The univariate and multivariate G study designs used were simple rater (r)-nested-within-person (p) [r : p and r(○) : p(•)] models, and in the univariate analysis was applied separately to CEFs completed by high and low experienced raters. RESULTS: The high experienced rater group yielded a substantially higher observed reliability in both the univariate and multivariate analyses. CONCLUSIONS: These results support the hypothesis that high experienced raters produce more reliable ratings of student performance and suggest methods for improving CEF ratings.


Assuntos
Competência Clínica/normas , Avaliação Educacional/normas , Docentes de Medicina , Preceptoria/normas , Estudantes de Medicina , Educação de Graduação em Medicina , Humanos , Iowa
10.
Teach Learn Med ; 22(3): 176-9, 2010 Jul.
Artigo em Inglês | MEDLINE | ID: mdl-20563936

RESUMO

BACKGROUND: The Accreditation Council of Graduate Medical Education now requires all pediatric residency training programs assess medical knowledge competency. PURPOSE: The goal of this project was to determine whether pediatric residency training using patient-based/experiential teaching made residents competent in the area of immunization knowledge or whether additional teaching strategies might need to be developed. METHODS: Cross-sectional and longitudinal study designs were used to determine improvement in immunization knowledge on a multiple-choice quiz over the 3 years of residency training. RESULTS: Both the cross-sectional and longitudinal data showed a statistically significant improvement in performance between residency training Years 1 and 2 but not between Years 2 and 3 on the quiz. This statistically significant relationship by year of training was seen despite the modest reliability of the short quiz and the sample size. CONCLUSIONS: This study shows that pediatric residency education using patient-based/experiential teaching is effective in teaching first year residents about immunization knowledge but is not as effective for 2nd- and 3rd-year residents. Other instructional methods such as computer-based cases could be employed during the 2nd and 3rd years.


Assuntos
Competência Clínica/normas , Conhecimentos, Atitudes e Prática em Saúde , Imunização , Internato e Residência/normas , Pediatria/educação , Análise de Variância , Competência Clínica/estatística & dados numéricos , Estudos Transversais , Currículo , Avaliação Educacional , Escolaridade , Humanos , Internato e Residência/estatística & dados numéricos , Estudos Longitudinais , Pediatria/normas , Pediatria/estatística & dados numéricos , Estatística como Assunto , Estados Unidos
11.
Teach Learn Med ; 22(1): 16-21, 2010 Jan.
Artigo em Inglês | MEDLINE | ID: mdl-20391278

RESUMO

BACKGROUND: Large-scale performance-based examinations are logistically complex and costly to run. Scores based solely on performance-based stations require extended testing time to achieve acceptable generalizability. PURPOSE: Combining scores from performance-based formats and written formats may improve test generalizability. METHODS: Data from 718 test-takers on the standardized patient-based portion of the National Board of Chiropractic Examiners Part IV examination were analyzed using a generalizability framework. RESULTS: The written and performance scores displayed a moderate universe score correlation indicating each component provided unique information about related clinical skills and the examinee's ability to manage a case. Combining the two scores provided a summary score that was more generalizable than either of the two measures individually. CONCLUSION: These results provide guidance for improving the generalizability of the standardized patient-based stations by combining the scores on the performance and written components using appropriate weighting procedures.


Assuntos
Administração de Caso/normas , Quiroprática/educação , Competência Clínica/normas , Avaliação Educacional/métodos , Análise e Desempenho de Tarefas , Humanos , Reprodutibilidade dos Testes
12.
Med Educ ; 43(4): 320-5, 2009 Apr.
Artigo em Inglês | MEDLINE | ID: mdl-19335573

RESUMO

CONTEXT: The development of a valid and reliable measure of clinical reasoning ability is a prerequisite to advancing our understanding of clinically relevant cognitive processes and to improving clinical education. A record of problem-solving performances within standardised and computerised patient simulations is often implicitly assumed to reflect clinical reasoning skills. However, the validity of this measurement method for assessing clinical reasoning is open to question. OBJECTIVES: Explicitly defining the intended clinical reasoning construct should help researchers critically evaluate current performance score interpretations. Although case-specific measurement outcomes (i.e. low correlations between cases) have led medical educators to endorse performance-based assessments of problem solving as a method of measuring clinical reasoning, the matter of low across-case generalisation is a reliability issue with validity implications and does not necessarily support a performance-based approach. Given this, it is important to critically examine whether our current performance-based testing efforts are correctly focused. To design a valid educational assessment of clinical reasoning requires a coherent argument represented as a chain of inferences supporting a clinical reasoning interpretation. DISCUSSION: Suggestions are offered for assessing how well an examinee's existing knowledge organisation accommodates the integration of new patient information, and for focusing assessments on an examinee's understanding of how new patient information changes case-related probabilities and base rates.


Assuntos
Educação Médica/métodos , Avaliação Educacional/métodos , Modelos Educacionais , Resolução de Problemas , Competência Clínica/normas , Iowa , Pensamento
14.
Med Educ ; 43(12): 1198-202, 2009 Dec.
Artigo em Inglês | MEDLINE | ID: mdl-19930511

RESUMO

CONTEXT: Some medical schools have recently replaced the medical school pre-admission interview (MSPI) with the multiple mini-interview (MMI), which utilises objective structured clinical examination (OSCE)-style measurement techniques. Their motivation for doing so stems from the superior reliabilities obtained with the OSCE-style measures. Other institutions, however, are hesitant to embrace the MMI format because of the time and costs involved in restructuring recruitment and admission procedures. OBJECTIVES: To shed light on the aetiology of the MMI's increased reliability and to explore the potential of an alternative, lower-cost interview format, this study examined the relative contributions of two facets (raters, occasions) to interview score reliability. METHODS: Institutional review board approval was obtained to conduct a study of all students who completed one or more MSPIs at a large Midwestern medical college during 2003-2007. Within this dataset, we identified 168 applicants who were interviewed twice in consecutive years and thus provided the requisite data for generalisability (G) and decision (D) studies examining these issues. RESULTS: Increasing the number of interview occasions contributed much more to score reliability than did increasing the number of raters. CONCLUSIONS: Replicating a number of interviews, each with one rater, is likely to be superior to the often recommended panel interview approach and may offer a practical, low-cost method for enhancing MSPI reliability. Whether such a method will ultimately enhance MSPI validity warrants further investigation.


Assuntos
Teste de Admissão Acadêmica , Entrevistas como Assunto , Faculdades de Medicina , Feminino , Humanos , Entrevistas como Assunto/métodos , Masculino , Reprodutibilidade dos Testes
15.
Med Educ ; 43(7): 688-94, 2009 Jul.
Artigo em Inglês | MEDLINE | ID: mdl-19573193

RESUMO

CONTEXT: Our project investigated whether trained lay observers can reliably assess the communication skills of medical students by observing their patient encounters in an out-patient clinic. METHODS: During a paediatrics clerkship, trained lay observers (standardised observers [SOs]) assessed the communication skills of Year 3 medical students while the students interviewed patients. These observers accompanied students into examination rooms in an out-patient clinic and completed a 15-item communication skills checklist during the encounter. The reliability of the communication skills scores was calculated using generalisability analysis. Students rated the experience and the validity of the assessment. The communication skills scores recorded by the SOs in the clinic were correlated with communication skills scores on a paediatrics objective structured clinical examination (OSCE). RESULTS: Standardised observers accompanied a total of 51 medical students and watched 199 of their encounters with paediatric patients. The reliability of the communication skills scores from nine observed patient encounters was calculated to be 0.80. There was substantial correlation between the communication skills scores awarded by the clinic observers and students' communication skills scores on their OSCE cases (r = 0.53, P < 0.001). Following 83.8% of the encounters, students strongly agreed that the observer had not interfered with their interaction with the patient. After 95.8% of the encounters, students agreed or strongly agreed that the observers' scoring of their communication skills was valid. CONCLUSIONS: Standardised observers can reliably assess the communication skills of medical students during clinical encounters with patients and are well accepted by students.


Assuntos
Comunicação , Educação de Graduação em Medicina/métodos , Avaliação Educacional/métodos , Pacientes Ambulatoriais/educação , Pediatria/educação , Humanos , Iowa , Satisfação do Paciente , Relações Médico-Paciente , Estatística como Assunto , Estudantes de Medicina/psicologia
16.
J Clin Anesth ; 54: 102-110, 2019 May.
Artigo em Inglês | MEDLINE | ID: mdl-30415149

RESUMO

STUDY OBJECTIVE: The first aim of this study was to test whether a 7 item evaluation scale developed by our department's certified registered nurse anesthetists (CRNAs) was psychometrically reliable. The second aim was to test whether anesthesiologists' performance changed with their years of postgraduate experience. DESIGN, SETTING, MEASUREMENTS: Sixty-two University of Iowa CRNAs evaluated 81 anesthesiologists during one weekend. Anesthesiologists' scores were adjusted for CRNA rater leniency. Anesthesiologists' scores were tested for sensitivity to CRNA-anesthesiologist case-specific variables. Scores also were tested against anesthesiologists' years of postgraduate experience. The latter association was tested for sensitivity to case-specific variables, anesthesiologists' clinical supervision scores provided by residents, and anesthesiologist clinical assignment variables. MAIN RESULTS: The 7 items demonstrated a single-factor structure, allowing calculation of mean score over the 7 items. Individual anesthesiologist scores were reliable when scores were provided by at least 10 different CRNAs. Anesthesiologists' scores (mean 3.34 [SD 0.41]) were not affected by the interval since last CRNA-anesthesiologist interaction, number of interactions, or case-specific variables. There was a negative association between leniency-adjusted anesthesiologist scores and years of anesthesiologist postgraduate practice (coefficient -0.20 per decade, t = -19.39, P < 0.0001). The association remained robust when accounting for case-specific variables, resident clinical supervision scores, and overall clinical assignment variables. CONCLUSIONS: Anesthesiologists' operating room performance can be evaluated reliably by non-physician anesthesia providers (CRNAs). The evaluation process can be done reliably and validly using an assessment scale consisting of only a few (<10) items and with evaluations by only a few individuals (≥10 CRNA raters). There is no indication evaluations provided by CRNAs were significantly influenced by the interval between interaction and evaluation, number of interactions, or other case-specific variables. From CRNAs' perspectives, on average, as anesthesiologists gain experience, anesthesiologists' behaviors in the operating room change, providing CRNAs with less direct assistance in patient care.


Assuntos
Anestesiologistas/estatística & dados numéricos , Competência Clínica/estatística & dados numéricos , Avaliação de Desempenho Profissional/estatística & dados numéricos , Enfermeiros Anestesistas/psicologia , Relações Médico-Enfermeiro , Anestesiologistas/psicologia , Avaliação de Desempenho Profissional/métodos , Humanos , Salas Cirúrgicas , Psicometria , Fatores de Tempo
17.
Teach Learn Med ; 20(3): 218-24, 2008.
Artigo em Inglês | MEDLINE | ID: mdl-18615295

RESUMO

BACKGROUND: This prospective randomized controlled study examined outpatient clinical teaching in the presence of the patient. METHODS: In 2006, patients in ambulatory internal medicine clinics at the University of Iowa were randomized to have faculty-learner presentations either in their presence or in the conference room. Staffing encounters were timed and faculty, learners and patients completed postencounter surveys. RESULTS: Participation included 254 patients and 12 faculty. Comparison of patient encounters randomized to exam room (n = 120) or conference room (n = 134) staffing demonstrated increased time spent with the patient in exam room staffing (91% vs. 54% of total staffing time; p < .0001) but no significant differences in mean total staffing time. Patients, learners, and faculty preferred exam room staffing. CONCLUSIONS: Concerns about time efficiency and patient and learner satisfaction during exam room staffing were not supported. This approach may allow attending physicians to maximize billing levels while increasing learner/patient involvement.


Assuntos
Educação Médica/métodos , Medicina Interna/educação , Avaliação de Resultados em Cuidados de Saúde , Admissão e Escalonamento de Pessoal , Exame Físico , Assistência Ambulatorial , Competência Clínica , Pesquisas sobre Atenção à Saúde , Humanos , Iowa , Satisfação do Paciente , Relações Médico-Paciente , Estudos Prospectivos
18.
Acad Med ; 93(8): 1146-1149, 2018 08.
Artigo em Inglês | MEDLINE | ID: mdl-29465452

RESUMO

PROBLEM: As medical schools move from discipline-based courses to more integrated approaches, identifying assessment tools that parallel this change is an important goal. APPROACH: The authors describe the use of test item statistics to assess the reliability and validity of web-enabled mechanistic case diagrams (MCDs) as a potential tool to assess students' ability to integrate basic science and clinical information. Students review a narrative clinical case and construct an MCD using items provided by the case author. Students identify the relationships among underlying risk factors, etiology, pathogenesis and pathophysiology, and the patients' signs and symptoms. They receive one point for each correctly identified link. OUTCOMES: In 2014-2015 and 2015-2016, case diagrams were implemented in consecutive classes of 150 medical students. The alpha reliability coefficient for the overall score, constructed using each student's mean proportion correct across all cases, was 0.82. Discrimination indices for each of the case scores with the overall score ranged from 0.23 to 0.51. In a G study using those students with complete data (n = 251) on all 16 cases, 10% of the variance was true score variance, and systematic case variance was large. Using 16 cases generated a G coefficient (relative score reliability) equal to 0.72 and a Phi equal to 0.65. NEXT STEPS: The next phase of the project will involve deploying MCDs in higher-stakes settings to determine whether similar results can be achieved. Further analyses will determine whether these assessments correlate with other measures of higher-order thinking skills.


Assuntos
Avaliação Educacional/normas , Estudantes de Medicina/psicologia , Pensamento , Competência Clínica/normas , Avaliação Educacional/métodos , Humanos , Reprodutibilidade dos Testes
19.
IISE Trans Healthc Syst Eng ; 88(2): 110-116, 2018.
Artigo em Inglês | MEDLINE | ID: mdl-29963653

RESUMO

An unbiased, repeatable process for assessing operating room performance is an important step toward quantifying the relationship between surgical training and performance. Hip fracture surgeries offer a promising first target in orthopedic trauma because they are common and they offer quantitative performance metrics that can be assessed from video recordings and intraoperative fluoroscopic images. Hip fracture repair surgeries were recorded using a head-mounted point-of-view camera. Intraoperative fluoroscopic images were also saved. The following performance metrics were analyzed: duration of wire navigation, number of fluoroscopic images collected, degree of intervention by the surgeon's supervisor, and the tip-apex distance (TAD). Two orthopedic traumatologists graded surgical performance in each video independently using an Objective Structured Assessment of Technical Skill (OSATS). Wire navigation duration correlated with weeks into residency and prior cases logged. TAD correlated with cases logged. There was no significant correlation between the OSATS total score and experience metrics. Total OSATS score correlated with duration and number of fluoroscopic images. Our results indicate that two metrics of hip fracture wire navigation performance, duration and TAD, significantly differentiate surgical experience. The methods presented have the potential to provide truly objective assessment of resident technical performance in the OR.

20.
Acad Med ; 93(8): 1212-1217, 2018 08.
Artigo em Inglês | MEDLINE | ID: mdl-29697428

RESUMO

PURPOSE: Many factors influence the reliable assessment of medical students' competencies in the clerkships. The purpose of this study was to determine how many clerkship competency assessment scores were necessary to achieve an acceptable threshold of reliability. METHOD: Clerkship student assessment data were collected during the 2015-2016 academic year as part of the medical school assessment program at the University of Michigan Medical School. Faculty and residents assigned competency assessment scores for third-year core clerkship students. Generalizability (G) and decision (D) studies were conducted using balanced, stratified, and random samples to examine the extent to which overall assessment scores could reliably differentiate between students' competency levels both within and across clerkships. RESULTS: In the across-clerkship model, the residual error accounted for the largest proportion of variance (75%), whereas the variance attributed to the student and student-clerkship effects was much smaller (7% and 10.1%, respectively). D studies indicated that generalizability estimates for eight assessors within a clerkship varied across clerkships (G coefficients range = 0.000-0.795). Within clerkships, the number of assessors needed for optimal reliability varied from 4 to 17. CONCLUSIONS: Minimal reliability was found in competency assessment scores for half of clerkships. The variability in reliability estimates across clerkships may be attributable to differences in scoring processes and assessor training. Other medical schools face similar variation in assessments of clerkship students; therefore, the authors hope this study will serve as a model for other institutions that wish to examine the reliability of their clerkship assessment scores.


Assuntos
Estágio Clínico/normas , Competência Clínica/normas , Avaliação Educacional/normas , Estágio Clínico/estatística & dados numéricos , Competência Clínica/estatística & dados numéricos , Avaliação Educacional/métodos , Avaliação Educacional/estatística & dados numéricos , Escolaridade , Humanos , Reprodutibilidade dos Testes , Estudantes de Medicina/estatística & dados numéricos
SELEÇÃO DE REFERÊNCIAS
DETALHE DA PESQUISA