The case Conference Assessment Tool (cCAT): a new workplace-based assessment ============================================================================ * Rory J O’Connor * E Diane Playford ## Abstract Rehabilitation medicine is an educational, problem-solving specialty that relies on excellent team communication, honest discussion with patients and their families, and collaborative goal setting. The case conference has been described as the technology of rehabilitation medicine because it encompasses all of these functions. Trainees should have the opportunity to develop skills in chairing case conferences through receipt of constructive feedback on their performance from their trainers. The aim of this project was to develop and evaluate the case Conference Assessment Tool (cCAT), a workplace-based assessment designed to score a trainee's performance on the key elements of chairing a case conference. Experienced rehabilitation medicine educational supervisors participated in a training workshop and then rated a series of simulated case conferences using the cCAT. Internal consistency was high (Cronbach's α **=** 0.945) and interrater reliability was acceptable (intraclass correlation coefficient range 0.673–0.777). Following feedback from the workshops, a final version of the cCAT was developed. The cCAT has now been adopted as a workplace-based assessment for specialty trainees in rehabilitation medicine by the Training Board of the Joint Royal Colleges of Physicians. Further work will explore its utility for trainees in other specialties and in communication and leadership skill training for undergraduate students. KEYWORDS:  * Workplace-based assessment * simulation * rehabilitation ## Introduction Rehabilitation medicine physicians must be confident in the medical management of people with complex disabilities, possess the knowledge and skills necessary to provide medical and functional prognoses to patients, and lead and coordinate the activities of a multidisciplinary rehabilitation team.1 A satisfactory outcome from a course of rehabilitation treatment relies on good communication among the members of the multidisciplinary team, patients and their families, and carers.2 Timely, accurate communication and involvement of all parties in collaborative goal setting enhance this outcome.3 These, and other tasks, are performed in the case conference, a joint meeting of the patient, his or her family, and the healthcare professionals involved in treating that patient. As the rehabilitation medicine physician often chairs this meeting, training in chairing case conferences is an integral part of the rehabilitation medicine postgraduate curriculum.1 Workplace-based assessments have established their value in higher medical training, but currently there is no assessment that rates a direct observation of the doctor-in-training's interaction with a patient, the patient's family and members of the multidisciplinary team. Such interactions are common in medical specialties such as rehabilitation medicine, paediatrics, palliative medicine and psychiatry, and require a particularly high level of performance from trainees in these disciplines. The direct observation of procedural skills (DOPS) is closest to the type of assessment tool that might evaluate a trainee's ability to chair a case conference.4 However, the DOPS is based on an individual trainee's interaction with a single patient – there is no record of interaction with the patient's family or the rest of the multidisciplinary team, and it is limited in its applicability to assessment of the specific skills of chairing a case conference. Measuring a trainee's performance in chairing case conferences relies on identifying the required competencies and determining these against a psychometrically robust scale.5 Standardised statistical principles for the development of workplace-based assessments should be followed when developing assessments and should include establishing appropriateness (acceptability) and estimates of error (reliability), and determination of what the scale purports to measure (validity).6 Many of the workplace-based assessments in use in postgraduate medical training do not have demonstrable psychometric properties. Where these exist, they are incomplete and from small, local samples.4,7 Feedback from assessments should improve trainees’ competence8 and compare trainees’ performances with an explicit standard.9 Unfortunately, in postgraduate medical education, feedback is infrequently provided by trainers to their trainees,10 or, if it is provided, it fails to be useful to the trainee.11 There are several initiatives to enhance trainers’ ability to provide feedback,12,13 and educational programmes for trainers have been shown to improve trainers’ ability to provide feedback.14,15 This is important because inexpert provision of negative feedback to trainees may actually worsen the trainees’ performances, possibly through decreasing motivation.16 Completing workplace-based assessments and providing feedback should improve trainees’ performances. The likely mechanism for this is through enhancing trainees’ self-regulation.17 This has been termed ‘feed-forward’ and will allow trainees to address shortcomings and incorporate suggestions for improvement into their work.18 There is some evidence to suggest that enhanced performance based on good quality feedback will lead to improved patient outcomes, which, of course, should be the aim of postgraduate medical education.19 Thus, providing feedback to trainees has a powerful educational effect that we should harness to improve patient care. The aim of this project was to develop and test a robust means of assessing a rehabilitation medicine trainee's performance in chairing a case conference and providing a structure for feedback. The project encompassed identifying the competencies in chairing a case conference and testing the psychometric properties of the case Conference Assessment Tool (cCAT). ## Methods This project comprised three phases: development of the cCAT, a pilot study and a confirmatory study. The project received institutional board approval by the University of Leeds’ Education Research Ethics Committee. ### Development of the cCAT The authors met with four groups of experienced rehabilitation professionals working with different multidisciplinary rehabilitation teams in Yorkshire and the south-east of England to develop a list of competencies that a trainee would be expected to demonstrate while chairing a case conference. During these meetings, delegates described the behaviours that they would expect a *trainee in rehabilitation* medicine to exhibit while competently chairing a case conference. Examples from their own clinical practice and their experiences of training postgraduate doctors were discussed. After each meeting, a list of competencies was drawn up and circulated to those present at the meeting for their confirmation. After four meetings no new competencies were identified. This list of competencies was matched to an assessment matrix, as is commonly used by other workplace-based assessments, such as the DOPS, to produce the pilot version of the cCAT. ### Pilot study For the pilot phase we scripted and filmed two scenarios. The scripts for these scenarios described case conferences involving patients and their families who would be typically encountered in routine clinical practice in NHS rehabilitation units in the UK and were developed together with rehabilitation clinicians and patients. An example is described in Box 1. #### Box 1. Example of a script for case conference simulation.
![][1] These two recordings were shown to two workshops in which educational supervisors, using the pilot version of the cCAT, rated the performance of each trainee. These data were collated and analysed using standardised psychometric principles for the development of workplace-based assessments.6 Items in the list of competencies that did not contribute to the assessment process or failed to meet the minimum requirements of the psychometric standards were eliminated from the final version of the cCAT (see Appendix 1). ### Confirmatory study Based on feedback from the pilot study, six further clinical scenarios were filmed with simulators playing the roles of clinicians, patients and family members. Educational supervisors representing the main training areas for rehabilitation medicine in the UK were invited to participate in workshops, at which they received training on the cCAT and the use of its scoring guide (see Appendix 2). Educational supervisors viewed the scenarios and completed a cCAT for each one. Feedback from these viewings was used to finalise the format of the cCAT and the scores were again analysed using standardised psychometrics. ### Psychometric analysis To examine the performance of the competency items in the matrix we calculated data quality, scaling assumptions and targeting.20 A principal component analysis was used to examine the scale's factors.21 Cronbach's α coefficient was used to determine internal consistency22 and interrater reliability scores (ICCs) were calculated for different raters scoring the same simulation. Convergent and discriminant construct validity were evaluated by examining the extent to which Spearman's (ρ) correlations between the rating on the cCAT and the seniority of the trainee are consistent with expectations.23 Content validity was established throughout the project from feedback by the raters. ## Results ### Pilot study Over the course of an hour, delegates at the workshop watched the two pilot recordings and provided rating scores on the performance of the trainee in the simulation; 50 initial version cCATs were completed. After initial analysis, a number of the items in the competency matrix were found not to contribute to the overall assessment of the trainee. The internal consistency of the pilot version was quite high (0.985), indicating that there may be redundancy of items. Principal component analysis identified that all of the competencies mapped onto one factor. However, a second factor was also identified linking ‘summing up’, ‘conclusion’ and ‘time keeping’, which indicates that these competencies are similar and can be combined into a single item referring to summing up agreed actions. Removal of these, and other poorly performing items, eg items with substantial floor or ceiling effects, led to the final version of the cCAT. ### Confirmatory study In total, 178 ratings of scenarios were collected at 7 workshops (Table 1). As the first workshop was a trial run of the final version of the cCAT, these ratings were excluded from the psychometric analyses. This left 171 ratings by 41 different raters for analysis. View this table: [Table 1.](http://www.rcpjournals.org/content/14/4/386/T1) Table 1. Details of meetings for the final version of the case Conference Assessment Tool (cCAT). Data quality was consistently high with less than 1% missing data across the ratings. Scaling assumptions were fulfilled (Table 2) and targeting analyses indicated that the final version of the cCAT had superior performance to the pilot version. Table 3 demonstrates that there were no significant floor or ceiling effects for the items. View this table: [Table 2.](http://www.rcpjournals.org/content/14/4/386/T2) Table 2. Statistical characteristics of the final version of the case Conference Assessment Tool (cCAT). View this table: [Table 3.](http://www.rcpjournals.org/content/14/4/386/T3) Table 3. Frequency response options for items of the final version of the case Conference Assessment Tool (cCAT). The reliability of the scale was acceptable, with an appropriate level of internal consistency and reasonable interrater reliability. As not all recordings were viewed at all sessions, the intraclass correlation coefficients for the three recordings with the greatest number of ratings are presented. These indicate that the educational supervisors watching these simulations agreed with each other about the level of competency of the trainee chairing the case conference. Validity analyses revealed that there was no correlation between the trainers’ rating of trainees’ performances and the trainees’ length of time in training (Table 4). There was also no relationship between the trainees’ rating and the trainers’ rating of the complexity of the simulated case conferences and cCAT scores. View this table: [Table 4.](http://www.rcpjournals.org/content/14/4/386/T4) Table 4. Construct validity of the final version of the case Conference Assessment Tool (cCAT). ## Discussion We have developed and tested a means of assessing and providing structured feedback to trainees in rehabilitation medicine who are learning the skills of chairing a case conference. Trainees can now use the cCAT to receive structured feedback on their performance in real clinical interactions from their trainers across the range of competencies required to chair a case conference satisfactorily. The cCAT was generated and tested using similar developmental and statistical methods to other workplace-based assessments.24 This ensures that the cCAT can be used alongside other workplace-based assessments to enable trainees to demonstrate their competence across the full range of the curriculum requirements. The validity analyses did not demonstrate a difference between trainees at distinct points in their training. This might mean that the cCAT total score has no overall meaning, with the important information in the component competency scores. For this reason, the cCAT has no rating for an overall score. The absence of a statistical relationship between length of training and cCAT score might reflect some of the trainees’ competencies already being acquired during core medical training and not requiring further development. As postgraduate medical training moves towards a competency-based model, it may become less important *when* trainees develop competencies in these areas, as long as they *do* develop the competencies, and any deficiencies in these areas are noted and remedied during training. This project has a number of limitations. The underlying premise is that educational supervisors can rate the performance of trainees in simulated case conferences and that these ratings can be used to provide statistical evidence of the reliability of the competency matrix. To try to promote verisimilitude in the recordings, patient simulators were recruited and trained, and provided with scripts developed by senior clinicians in rehabilitation medicine. Trainees rated the simulations as having a close resemblance to real case conferences, contributing to the evidence of verisimilitude. Although the individual elements of the project have been used successfully in the past, the combination of methods used in this series of studies has not previously been deployed in the development of a workplace-based assessment. It will be instructive to see how the cCAT performs in real educational situations. As this is a new assessment, and trainers have not been used to assessing trainees’ performances in case conferences in a structured way before, there may have been a learning curve for trainers using the cCAT for the first recording. We tried to reduce this effect by providing the raters at each of the viewing sessions with an introduction to the cCAT, and encouraged them to use the scoring guide when rating the simulated case conference performances. The statistical methods used in the analyses of the ratings assume that the competencies measured are stable traits of the individuals who volunteered to take part in the recordings. However, we know that these individuals, and trainees in general, may perform differently in similar situations due to the influence of many personal and professional factors operating at the time.25 This is one of the limitations of psychometric theory in general.26 Hopefully, the results from this project will help to improve methods for designing and testing workplace-based assessments and generate enthusiasm for newer statistical methods such as item response theory.27 A further limitation is that the cCAT has not demonstrated effectiveness in producing changes in trainees’ performances in chairing case conferences. This information can be collected only once the cCAT has been in use in training in real clinical situations. The work that will follow from this project will involve collecting serial data from trainees to demonstrate how the cCAT allows trainees to develop the competencies for chairing a case conference. The cCAT is potentially useful in other specialties that require a structured tool to assess trainees’ performances in chairing patient and family case conferences, such as geriatrics, paediatrics, palliative medicine and psychiatry. Further validity work will be required before the cCAT can be used by trainees in these specialties. We have also started to explore its use by medical undergraduates in interprofessional learning environments. In summary, the cCAT is an acceptable, reliable and valid workplace-based assessment that can be used to provide structured feedback to trainees in rehabilitation medicine. This project has also demonstrated that it is feasible to combine simulations, video recording and statistical testing of assessment methods to provide a novel means to develop assessments in postgraduate medical education. ## Funding The Leeds Teaching Hospitals Charitable Foundation Training and Education Special Advisory Fund and the University of Leeds Teaching Quality Enhancement Fund. ### Key points * Rehabilitation medicine demands a high degree of medical leadership * Communication among the team, patients and their families is paramount to the success of the rehabilitation programme * The case Conference Assessment Tool (cCAT) provides a structured assessment to enable trainers to help trainees develop their team leadership skills ## Acknowledgements We would like to thank Jools Symons and the Patient and Public Voice Unit at the Leeds Institute of Medical Education. We would also like to thank the trainees and educational supervisors who willingly participated in the recording and marking of the simulated case conferences. * © 2014 Royal College of Physicians ## References 1. Joint Royal Colleges of Physicians Training Board. Specialty training curriculum for rehabilitation medicine. London: Joint Royal Colleges of Physicians Training Board, 2010. 2. Barnes MP. Principles of neurological rehabilitation. J Neurol Neurosurg Psychiatry 2003;74:3–7.doi:10.1136/jnnp.74.suppl_1.i3 [CrossRef](http://www.rcpjournals.org/lookup/external-ref?access_num=10.1136/jnnp.74.suppl_1.i3&link_type=DOI) 3. Holliday RC, Ballinger C, Playford ED. Goal setting in neurological rehabilitation: patients’ perspectives. Disabil Rehabil 2007;29:389–94.doi:10.1080/09638280600841117 [CrossRef](http://www.rcpjournals.org/lookup/external-ref?access_num=10.1080/09638280600841117&link_type=DOI) [PubMed](http://www.rcpjournals.org/lookup/external-ref?access_num=17364791&link_type=MED&atom=%2Fclinmedicine%2F14%2F4%2F386.atom) [Web of Science](http://www.rcpjournals.org/lookup/external-ref?access_num=000244891900006&link_type=ISI) 4. Goff BA, Nielsen PE, Lentz GM, et al. Surgical skills assessment: a blinded examination of obstetrics and gynecology residents. Am J Obstet Gynecol 2002;186:613–17.doi:10.1067/mob.2002.122145 [CrossRef](http://www.rcpjournals.org/lookup/external-ref?access_num=10.1067/mob.2002.122145&link_type=DOI) [PubMed](http://www.rcpjournals.org/lookup/external-ref?access_num=11967481&link_type=MED&atom=%2Fclinmedicine%2F14%2F4%2F386.atom) [Web of Science](http://www.rcpjournals.org/lookup/external-ref?access_num=000175545300003&link_type=ISI) 5. Duffy FD, Gordon GH, Whelan G, et al. Assessing competence in communication and interpersonal skills: the Kalamazoo II report. Acad Med 2004;79:495–507.doi:10.1097/00001888-200406000-00002 [CrossRef](http://www.rcpjournals.org/lookup/external-ref?access_num=10.1097/00001888-200406000-00002&link_type=DOI) [PubMed](http://www.rcpjournals.org/lookup/external-ref?access_num=15165967&link_type=MED&atom=%2Fclinmedicine%2F14%2F4%2F386.atom) [Web of Science](http://www.rcpjournals.org/lookup/external-ref?access_num=000222831900002&link_type=ISI) 6. Wass V, Van der Vleuten C, Shatzer J, Jones R. Assessment of -clinical competence. Lancet 2001;357:945–9.doi:10.1016/S0140-6736(00)04221-5 [CrossRef](http://www.rcpjournals.org/lookup/external-ref?access_num=10.1016/S0140-6736(00)04221-5&link_type=DOI) [PubMed](http://www.rcpjournals.org/lookup/external-ref?access_num=11289364&link_type=MED&atom=%2Fclinmedicine%2F14%2F4%2F386.atom) [Web of Science](http://www.rcpjournals.org/lookup/external-ref?access_num=000167625300033&link_type=ISI) 7. Winckel CP, Reznick RK, Cohen R, Taylor B. Reliability and -construct validity of a structured technical skills assessment form. Am J Surg 1994;167:423–7.doi:10.1016/0002-9610(94)90128-7 [CrossRef](http://www.rcpjournals.org/lookup/external-ref?access_num=10.1016/0002-9610(94)90128-7&link_type=DOI) [PubMed](http://www.rcpjournals.org/lookup/external-ref?access_num=8179088&link_type=MED&atom=%2Fclinmedicine%2F14%2F4%2F386.atom) [Web of Science](http://www.rcpjournals.org/lookup/external-ref?access_num=A1994NJ99900010&link_type=ISI) 8. Wigton RS, Patil KD, Hoellerich VL. The effect of feedback in learning clinical diagnosis. J Med Educ 1986;61:816–22. [PubMed](http://www.rcpjournals.org/lookup/external-ref?access_num=3531519&link_type=MED&atom=%2Fclinmedicine%2F14%2F4%2F386.atom) [Web of Science](http://www.rcpjournals.org/lookup/external-ref?access_num=A1986E393500006&link_type=ISI) 9. van de Ridder JM, Stokking KM, McGaghie WC, ten Cate OT. What is feedback in clinical education? Med Educ 2008;42:189–97.doi:10.1111/j.1365-2923.2007.02973.x [CrossRef](http://www.rcpjournals.org/lookup/external-ref?access_num=10.1111/j.1365-2923.2007.02973.x&link_type=DOI) [PubMed](http://www.rcpjournals.org/lookup/external-ref?access_num=18230092&link_type=MED&atom=%2Fclinmedicine%2F14%2F4%2F386.atom) 10. Bing-You RG, Trowbridge RL. Why medical educators may be failing at feedback. JAMA 2009;302:1330–1.doi:10.1001/jama.2009.1393 [CrossRef](http://www.rcpjournals.org/lookup/external-ref?access_num=10.1001/jama.2009.1393&link_type=DOI) [PubMed](http://www.rcpjournals.org/lookup/external-ref?access_num=19773569&link_type=MED&atom=%2Fclinmedicine%2F14%2F4%2F386.atom) [Web of Science](http://www.rcpjournals.org/lookup/external-ref?access_num=000270026000016&link_type=ISI) 11. Kogan JR, Bellini LM, Shea JA. Have you had your feedback today? Acad Med 2000;75:1041.doi:10.1097/00001888-200010000-00026 [CrossRef](http://www.rcpjournals.org/lookup/external-ref?access_num=10.1097/00001888-200010000-00026&link_type=DOI) [PubMed](http://www.rcpjournals.org/lookup/external-ref?access_num=11031157&link_type=MED&atom=%2Fclinmedicine%2F14%2F4%2F386.atom) 12. Feins A, Waterman MA, Peters AS, Kim M. The teaching matrix: a tool for organizing teaching and promoting professional growth. Acad Med 1996;71:1200–3.doi:10.1097/00001888-199611000-00016 [CrossRef](http://www.rcpjournals.org/lookup/external-ref?access_num=10.1097/00001888-199611000-00016&link_type=DOI) [PubMed](http://www.rcpjournals.org/lookup/external-ref?access_num=9217508&link_type=MED&atom=%2Fclinmedicine%2F14%2F4%2F386.atom) 13. Paukert JL, Richards ML, Olney C. An encounter card system for increasing feedback to students. Am J Surg 2002;183:300–4.doi:10.1016/S0002-9610(02)00786-9 [CrossRef](http://www.rcpjournals.org/lookup/external-ref?access_num=10.1016/S0002-9610(02)00786-9&link_type=DOI) [PubMed](http://www.rcpjournals.org/lookup/external-ref?access_num=11943131&link_type=MED&atom=%2Fclinmedicine%2F14%2F4%2F386.atom) [Web of Science](http://www.rcpjournals.org/lookup/external-ref?access_num=000174915000019&link_type=ISI) 14. Menachery EP, Knight AM, Kolodner K, Wright SM. Physician characteristics associated with proficiency in feedback skills. J Gen Intern Med 2006;21:440–6.doi:10.1111/j.1525-1497.2006.00424.x [CrossRef](http://www.rcpjournals.org/lookup/external-ref?access_num=10.1111/j.1525-1497.2006.00424.x&link_type=DOI) [PubMed](http://www.rcpjournals.org/lookup/external-ref?access_num=16704385&link_type=MED&atom=%2Fclinmedicine%2F14%2F4%2F386.atom) [Web of Science](http://www.rcpjournals.org/lookup/external-ref?access_num=000237422300008&link_type=ISI) 15. Orlander JD, Bor DH, Strunin L. A structured clinical feedback exercise as a learning-to-teach practicum for medical residents. Acad Med 1994;69:18–20.doi:10.1097/00001888-199401000-00003 [CrossRef](http://www.rcpjournals.org/lookup/external-ref?access_num=10.1097/00001888-199401000-00003&link_type=DOI) [PubMed](http://www.rcpjournals.org/lookup/external-ref?access_num=8285990&link_type=MED&atom=%2Fclinmedicine%2F14%2F4%2F386.atom) 16. Litzelman DK, Stratos GA, Marriott DJ, et al. Beneficial and harmful effects of augmented feedback on physicians’ clinical-teaching performances. Acad Med 1998;73:324–32.doi:10.1097/00001888-199803000-00022 [CrossRef](http://www.rcpjournals.org/lookup/external-ref?access_num=10.1097/00001888-199803000-00022&link_type=DOI) [PubMed](http://www.rcpjournals.org/lookup/external-ref?access_num=9526460&link_type=MED&atom=%2Fclinmedicine%2F14%2F4%2F386.atom) [Web of Science](http://www.rcpjournals.org/lookup/external-ref?access_num=000072534500028&link_type=ISI) 17. Rees C, Shepherd M. The acceptability of 360-degree judgements as a method of assessing undergraduate medical students’ personal and professional behaviours. Med Educ 2005;39:49–57.doi:10.1111/j.1365-2929.2004.02032.x [CrossRef](http://www.rcpjournals.org/lookup/external-ref?access_num=10.1111/j.1365-2929.2004.02032.x&link_type=DOI) [PubMed](http://www.rcpjournals.org/lookup/external-ref?access_num=15612900&link_type=MED&atom=%2Fclinmedicine%2F14%2F4%2F386.atom) [Web of Science](http://www.rcpjournals.org/lookup/external-ref?access_num=000226255200010&link_type=ISI) 18. Sargeant J, Mann K, Sinclair D, et al. Understanding the influence of emotions and reflection upon multi-source feedback acceptance and use. Adv Health Sci Educ Theory Pract 2008;13:275–88.doi:10.1007/s10459-006-9039-x [CrossRef](http://www.rcpjournals.org/lookup/external-ref?access_num=10.1007/s10459-006-9039-x&link_type=DOI) [PubMed](http://www.rcpjournals.org/lookup/external-ref?access_num=17091339&link_type=MED&atom=%2Fclinmedicine%2F14%2F4%2F386.atom) 19. Galbraith RM, Holtman MC, Clyman SG. Use of assessment to reinforce patient safety as a habit. Qual Safety Health Care 2006;15:i30–3.doi:10.1136/qshc.2005.015941 [CrossRef](http://www.rcpjournals.org/lookup/external-ref?access_num=10.1136/qshc.2005.015941&link_type=DOI) 20. Streiner DL, Norman GR. Health measurement scales: a practical guide to their development and use. 2nd edn. Oxford: Oxford University Press, 1995. 21. Altman DG. Practical statistics for medical research. London: Chapman & Hall, 1991. 22. Cronbach LJ. Coefficient alpha and the internal structure of tests. Psychometrika 1951;16:297–334.doi:10.1007/BF02310555 [CrossRef](http://www.rcpjournals.org/lookup/external-ref?access_num=10.1007/BF02310555&link_type=DOI) [Web of Science](http://www.rcpjournals.org/lookup/external-ref?access_num=A1951YB33000004&link_type=ISI) 23. Campbell DT, Fiske DW. Convergent and discriminant validation by the multitrait-multimethod matrix. Psychol Bull 1959;56:81–105.doi:10.1037/h0046016 [CrossRef](http://www.rcpjournals.org/lookup/external-ref?access_num=10.1037/h0046016&link_type=DOI) [PubMed](http://www.rcpjournals.org/lookup/external-ref?access_num=13634291&link_type=MED&atom=%2Fclinmedicine%2F14%2F4%2F386.atom) [Web of Science](http://www.rcpjournals.org/lookup/external-ref?access_num=A1959WD16300001&link_type=ISI) 24. Van Der Vleuten CPM. The assessment of professional -competence: Developments, research and practical implications. Adv Health Sci Educ 1996;1:41–67.doi:10.1007/BF00596229 [CrossRef](http://www.rcpjournals.org/lookup/external-ref?access_num=10.1007/BF00596229&link_type=DOI) [Web of Science](http://www.rcpjournals.org/lookup/external-ref?access_num=000208245600005&link_type=ISI) 25. Schuwirth LW, van der Vleuten CP. A plea for new psychometric models in educational assessment. Med Educ 2006;40:296–300.doi:10.1111/j.1365-2929.2006.02405.x [CrossRef](http://www.rcpjournals.org/lookup/external-ref?access_num=10.1111/j.1365-2929.2006.02405.x&link_type=DOI) [PubMed](http://www.rcpjournals.org/lookup/external-ref?access_num=16573664&link_type=MED&atom=%2Fclinmedicine%2F14%2F4%2F386.atom) [Web of Science](http://www.rcpjournals.org/lookup/external-ref?access_num=000236385100005&link_type=ISI) 26. Colliver JA. Educational theory and medical education practice: a cautionary note for medical school faculty. Acad Med 2002; 77(12 Pt 1):1217–20.doi:10.1097/00001888-200212000-00012 [CrossRef](http://www.rcpjournals.org/lookup/external-ref?access_num=10.1097/00001888-200212000-00012&link_type=DOI) [PubMed](http://www.rcpjournals.org/lookup/external-ref?access_num=12480626&link_type=MED&atom=%2Fclinmedicine%2F14%2F4%2F386.atom) [Web of Science](http://www.rcpjournals.org/lookup/external-ref?access_num=000180011100009&link_type=ISI) 27. De Champlain AF. A primer on classical test theory and item response theory for assessments in medical education. Med Educ 2010;44:109–17.doi:10.1111/j.1365-2923.2009.03425.x [CrossRef](http://www.rcpjournals.org/lookup/external-ref?access_num=10.1111/j.1365-2923.2009.03425.x&link_type=DOI) [PubMed](http://www.rcpjournals.org/lookup/external-ref?access_num=20078762&link_type=MED&atom=%2Fclinmedicine%2F14%2F4%2F386.atom) [Web of Science](http://www.rcpjournals.org/lookup/external-ref?access_num=000272866400021&link_type=ISI) [1]: /embed/graphic-1.gif