Back to Journals » Advances in Medical Education and Practice » Volume 16
Evaluation of Applying Standardized Patients in Certification Assessment of Clinical Practice Ability Standardized Training for Psychiatric Resident Physicians
Authors Yao F, Chen JS, Chen LH , Ye BY , Yang CJ, Li XZ
Received 25 October 2024
Accepted for publication 16 March 2025
Published 18 March 2025 Volume 2025:16 Pages 411—418
DOI https://doi.org/10.2147/AMEP.S503006
Checked for plagiarism Yes
Review by Single anonymous peer review
Peer reviewer comments 2
Editor who approved publication: Dr Md Anwarul Azim Majumder
Fei Yao,1,2,* Jian-Song Chen,1,2,* Li-Hong Chen,1,2 Bi-Yu Ye,1,2 Chan-Juan Yang,1,2 Xuan-Zi Li1,2
1Department of Education, The Affiliated Brain Hospital, Guangzhou Medical University, Guangzhou, Guangdong, 510370, People’s Republic of China; 2School of Mental Health, Guangzhou Medical University, Guangzhou, Guangdong, 510370, People’s Republic of China
*These authors contributed equally to this work
Correspondence: Xuan-Zi Li; Chan-Juan Yang, Department of Education, The Affiliated Brain Hospital, School of Mental Health, Guangzhou Medical University, No. 36 mingxin Road, Guangzhou, Guangdong, 510370, People’s Republic of China, Tel +86 20 81268042, Email [email protected]; [email protected]
Objective: The aim of this study was to explore the application of standardized patient (SP) in the certification assessment of clinical practice ability standardized training for psychiatric resident physicians.
Methods: Candidates and examiners who participated in the graduation examination of clinical practice ability standardized training for psychiatric resident physicians in Guangdong Province from 2020 to 2023 were selected as the research participants. The assessment scores of candidates were collected. Questionnaires about satisfaction degree and SP application evaluation were designed and filled out by the examiners and candidates to evaluate their satisfaction with the clinical practice ability certification assessment.
Results: The average annual score of the practical skills certification assessment is 83.88 ± 5.78 (P< 0.001). There is no evident change in the second station (81.50 ± 6.81, P=0.0825), while the scores of the other stations fluctuate significantly (83.77 ± 6.80, 84.46 ± 6.70, and 88.31 ± 7.82; P< 0.05). A total of 185 valid questionnaires were collected, of which 97.87% of the examiners and 92.03% of the candidates are satisfied with SP, respectively, both higher than their overall satisfaction levels (97.87% vs 89.36%; 92.03% vs 85.51). Examiners are highly consistent with their evaluation of SP satisfaction (Kendall=0.829, P< 0.001), while candidates have low consistency in their evaluation of SP (Kendall=0.012, P=0.04). The differences in SP evaluation between different identities mainly include appearance, intonation, physical movements, language accuracy, and use of medical terminologies (P< 0.05).
Conclusion: The satisfaction with the application of SP in the certification assessment of clinical practice ability standardized training for psychiatric resident physicians is high. However, it has inconsistent evaluations from examiners and candidates in terms of appearance, pronunciation and intonation, physical movements, language accuracy, and use of medical terminologies. Enhancing the SP level can promote the stability of the assessment.
Keywords: standardized patient, psychiatry, standardized training for resident physicians, assessment
Since the concept of standardized patients (SP) was proposed by Professor Barrows and Abrahamson in 1964, it has been extensively applied in skills training, assessment, and development in medical education in many countries.1 SP refers to, after systematic training, the ability to accurately and realistically portray the clinical symptoms and signs of patients, serving as a triple identity of patients, evaluators, and respondents. SP can be healthy people or patients.2 Due to the particularity, operability, and safety issues of patients, the assessment of clinical practice ability in psychiatry usually utilizes the advantages of SP to replace patients. The combination of SP and objective structured clinical examination (OSCE) has been widely used in the training and assessment of resident physicians and the clinical skills examination in practicing physician assessment in the United States, Canada, and the United Kingdom. This includes psychiatric examination stations.3 SP overcomes the difficulty of finding highly characteristic cases and situations in the previous clinical practice of psychiatry while creating a safe atmosphere, which is a substantial breakthrough in psychiatry education.
Currently, China adopts a combination strategy of “regular assessment”, “phase assessment”, and “graduation assessment”. The regular assessment uses comprehensive assessment methods, including the mini-clinical evaluation exercise (Mini-CEX)4 and direct observation of procedural skills (DOPS),5 listed as the regular assessment score registration system. The phase assessment utilizes the form of the national practicing physician practical skills examination to regularly conduct skill practice operations and comprehensively evaluate students’ clinical skills and thinking abilities in psychiatry. The graduation assessment employs the OSCE mode, investigates and collects medical histories with multiple stations, and physically checks basic skills. The graduation assessment is an indispensable part of quantifying clinical competence and evaluating job competency. Moreover, the use of SP plays a crucial role in assessing evaluation consistency and stability.6 In 2020, the Health Human Resources Development Center organized experts to revise the guidance standards for the assessment of practical skills in residential training graduation and launched a new “Clinical Practice Ability Assessment Standard Plan for Standardized Training Graduation of Resident Physicians (Psychiatry)”. However, the evaluation of the application of the assessment plan is insufficient, there has been no research or evaluation on the implementation effect of this plan. This standard sets up one general examination station for cardiopulmonary resuscitation and four independent examination stations, including clinical skills, clinical thinking, clinical communication, and neurological examination. The Affiliated Brain Hospital of Guangzhou Medical University is the first specialized hospital for mental illness, the earliest clinical and teaching base for psychiatry, and one of the five national standardized training bases for psychiatric resident physicians in China. It has undertaken the graduation assessment of practical skills for psychiatric resident physicians in Guangdong Province in recent years. Since the revision of the assessment standard in 2020, it has conducted assessments combining SP with OSCE and accumulated some experience. This article mainly analyzes and summarizes the feedback on the application of SP in the certification assessment of practical skills in residential training, aiming to provide a valuable reference for peers and contribute to the continuous improvement of the quality of certification assessment of standardized training for psychiatric resident physicians.
Subject and Method
Subject
Candidates and examiners who participated in the certification assessment of clinical practice ability standardized training for psychiatric resident physicians in Guangdong Province from 2020 to 2023 were incorporated in this study as the research participants.
Method
According to the 2020 assessment standard, a multi-station assessment form was adopted, which is more relevant to psychiatry. A specific exam execution team was established to coordinate the design, preparation, and implementation of the assessment based on the characteristics and actual situation of the profession. Four examination stations were set up, including clinical skills, clinical thinking, clinical communication, and neurological examination. Among them, the first, third, and fourth stations had medical students/doctors playing the role of patients/patient families. Each station was evaluated synchronously and independently by two examiners from different professional bases who had received unified examiner training and obtained the Guangdong Province Graduation Practice Skills Examiner Certificate. The average score was taken as the assessment result. Measured by the percentage system, each project was calculated based on weights and included in the total score, with a maximum score of 100 points. The assessment content is shown in Table 1.
![]() |
Table 1 Information on Clinical Practice Skills Certification Assessment Stations |
The medical students/doctors impersonating patients/family members at each skill assessment station had received unified training. The clinical skills station included psychiatric examination, medical history collection, and necessary physical examinations. During psychiatric examination, medical students and doctors impersonated patients and family members, respectively. The clinical thinking station contained case analysis, question answering, and medical record summarizing, mainly involved in symptom and syndrome analysis, diagnostic analysis, diagnostic conclusions and criteria, risk assessment and management measures, and treatment. In the clinical communication station, examiners acted as the family members of patients and communicated with candidates based on the characteristics of relevant cases. Regarding the neurological station, candidates conducted specialized physical examinations of the nervous system on patients played by medical students.
Questionnaire Survey
After the assessment and before the announcement of the results, an online anonymous questionnaire survey was conducted on the examiners and candidates, including the rationality recognition of the exam execution, content, difficulty, and form of each station, overall satisfaction, and, most importantly, the evaluation of SP. The assessment satisfaction questionnaire set ten items for the examiners and candidates, respectively. The SP satisfaction questionnaire was composed of 20 items, of which the first ten were positive questions, followed by ten negative questions. All options were scored according to the Likert level 5, with 5 points representing extremely satisfactory/completely agree and 1 point denoting extremely unsatisfactory/completely disagree.
Statistical Processing
The data were statistically analyzed using SPSS 26.0 and GraphPad Prism 8.0. Adobe Illustrator was adopted for image processing. The scores of stations were described with mean ± standard deviation. The Kruskal–Wallis rank-sum test was employed for statistical analysis, with P <0.05 as the statistically significant difference. The questionnaire answers were described with percentages. The Kendall test7 was used to analyze the consistency of the questionnaire responses of the two groups, with P <0.05 as the statistically significant consistency. The Kendall coordination coefficient <0.2 represents poor consistency, 0.2–0.4 implies moderate consistency, 0.4–0.6 indicates moderate consistency, 0.6–0.8 represents strong consistency, and 0.8–1.0 indicates strong consistency. The Wilcoxon rank-sum test was employed to explore the differences in the questionnaires between the two groups, with P <0.05 indicating a statistically significant difference.
Results
Overall Situation of Practical Skills Certification Assessment
From 2020 to 2023, there were 523 candidates in total, including 187 males and 336 females, with an average total score of 83.88 ± 5.78. The average scores for each station are 83.77 ± 6.80, 81.50 ± 6.81, 84.46 ± 6.70, and 88.31 ± 7.82, respectively. The assessment scores for the remaining years are shown in Table 2. Except for the second station, there are significant differences in assessment scores among the other stations in different years.
![]() |
Table 2 Summary of Practical Skills Certification Assessment (Mean ± Standard Deviation, Ranging from 0 to 100) |
Evaluation of Satisfaction Degree of Examiners and Candidates with the Certification Assessment of Practical Skills
Valid questionnaires collected from examiners and candidates are 47 and 138, respectively, as shown in Table 3 and Table 4. Those who answered “Extremely Satisfactory” or “Satisfactory” are considered satisfied with the content of the items. According to the results, 89.36% of the examiners and 85.51% of the candidates have overall satisfaction with the examination execution. The satisfaction degrees of the examiners with the examination environment and equipment are lower than the overall satisfaction (87.23%, 85.11%, and 89.37%). The candidates have lower satisfaction with the examination environment, form, difficulty, and scope than the overall satisfaction (78.99%, 83.33%, 78.99%, 84.06%, and 85.51%).
![]() |
Table 3 Evaluation of the Examiners’ Satisfaction with Practical Skills Certification Assessment (%) |
![]() |
Table 4 Evaluation of the Candidates’ Satisfaction with Practical Skills Certification Assessment (%) |
Consistency of SP Satisfaction Evaluation of Examiners and Candidates
The Kendall test was adopted to analyze the consistency of SP satisfaction evaluation of examiners and candidates. As shown in Table 5, the consistency of evaluation is statistically significant, with the examiners’ consistency being higher than that of the candidates. The Kendall coordination coefficient of the examiners is 0.829, indicating high consistency, while that of the candidates is 0.012, representing low consistency.
![]() |
Table 5 Consistency Test of SP Satisfaction Evaluation |
Specific Evaluations of SP by Examiners and Candidates
The questionnaire of SP evaluation for examiners and candidates contained 20 items, of which the first were ten positive questions, followed by ten negative questions, as shown in Table 6. The intergroup comparison exhibits seven items that the satisfaction degrees of examiners and candidates are inconsistent (examiners vs candidates, the ratio of strongly agree + the ratio of agree), as shown below: tally with the appearance of patients (100% vs 92.75%), the patients proficiently respond to medical history information with appropriate pronunciation and intonation (98.08% vs 94.21%), conform to patient physical movements (98.08% vs 93.48%), patient performance is consistent and coordinated (100% vs 95.65%), the patients accurately express the language (96.16% vs 94.21%), the patients accurately express psychological status through language (98.07% vs 96.37%), and the patients use medical terminologies (17.31% vs 0%). Among the seven items, the first six are positive, while the last one is negative. Although the satisfaction ratios suggest that the examiners and candidates have consistent trends, the former is more satisfied with the appearance of patients, voice and intonation, physical movements, coherence, and language accuracy. However, in terms of whether to use medical terminologies, compared with the candidates, the examiners tend to believe that there is a use of medical terminologies in SP.
![]() |
Table 6 Evaluation of SP Satisfaction of Examiners and Candidates (%) |
Discussion
Huang et al8 disclosed that the overall prevalence of adult mental disorders in China was 17%, with approximately 16 million patients suffering from severe mental diseases, and the number of registrations had been increasing year by year. Mental health has attracted social attention. Meanwhile, the lack of professional talents has constrained the development of the mental health industry. Therefore, it is of significant importance to cultivate practical talents in mental illness and mental health that meet social needs, centering around the construction of clinical thinking and practical abilities. Clinical practice is also the continuation and perfection of theoretical knowledge.9 In 2020, the Health Human Resources Development Center of China organized experts to revise the guidance standards for the graduation assessment of practical skills in residential training and launched a “Clinical Practice Ability Assessment Standard Plan for Standardized Training Graduation of Resident Physicians (Psychiatry)”. However, there has been no application evaluation of the assessment plan yet.
As the major undertaking institution for the practical skills assessment of psychiatric resident physicians in Guangdong Province, the Affiliated Brain Hospital of Guangzhou Medical University has accumulated some experience and participated in the formulation of the assessment plan. By analyzing the results of the certification assessment in the past four years and the satisfaction degree with the assessment, from the perspective of the application of SP, it is found that (1) the scores of stations involving substitute patients and substitute family members fluctuate greatly. The standardization level of SP is essential to the stability of the certification assessment; (2) Examiners and candidates are generally satisfied with the graduation exam. The exam environment and equipment significantly impact examiners, while there is still room for improvement in the satisfaction of candidates with the exam environment, format, difficulty, and scope; (3) Examiners have consistent evaluations of SP, while the feedback from candidates is not harmonious; (4) The appearance, intonation, physical movements, coherence, and language accuracy of SP are crucial factors that influence the experience of candidates, and the use of medical terminologies by SP is a key observation point for examiners. We speculate that the above findings may be related to factors such as the candidates’ mentality before the examination (for example, eagerly wanting to obtain information supporting diagnosis from the medical history description of SP), the candidates’ shorter contact time with SP than the examiners’, and the examiners’ potential role as family members. The significance of this study is not only to understand the application of SP under the new assessment standard scheme but also to reveal the difference in experience between examiners and candidates in the current graduation assessment mode. The examination environment, hardware facilities, unifying examination difficulty, and examination scope can be rectified in the short term. However, the investment in SP rectification is enormous, making it difficult to shorten the gap between SP and actual patients. Personnel with medical backgrounds are prone to using medical terminologies, and not every doctor can be a good actor. SP training requires investing funds, energy, and time, which is costly. Unpaid SP has low persistence and is adverse to the stability and protracted nature of standardized teams.
Meanwhile, future directions should focus on the following issues: (1) Psychiatry emphasizes empathy,10 and whether the implementation of SP will weaken students’ empathy ability; (2) Since SP mostly presents classic cases or single diseases, will it affect students’ identification of comorbidities and rare diseases; (3) Whether the process of cultivating and assessing standardization restricts the divergent thinking of students; (4) This study evaluated the application situation from the perspectives of examiners and candidates and found differences between the two. Should subsequent research take the experience of SP in assessment as one of the observation subjects?
In summary, the OSCE assessment model using SP receives a high degree of satisfaction in the certification assessment of the standardized training for psychiatric resident physicians. The exploration of teaching or assessment models should adhere to the patient-centered strategy, construct rigorous clinical thinking in the cycle of discovering and solving problems, and advance independent analysis ability and clinical diagnosis and treatment level. Continuously identifying, proposing, and addressing problems in practice, supplemented by the cultivation of scientific thinking, can inject new vitality into the development of medicine. A complete and systematic assessment system is necessary and meaningful in ensuring the quality of training.
Abbreviations
SP, standardized patient; OSCE, objective structured clinical examination; Mini-CEX, mini-clinical evaluation exercise; DOPS, direct observation of procedural skills.
Data Sharing Statement
The original datasets used and analyzed during the current study are available from the corresponding authors on reasonable request.
Ethics Approval and Consent to Participate
This study was approved by the Ethics Committee of The Affiliated Brain Hospital of Guangzhou Medical University (Project No.50010724-1029). The study protocol and survey contents were conducted in accordance with the principles of the Declaration of Helsinki. Written informed consent to participate in this study was obtained from all participants.
Author Contributions
All authors made a significant contribution to the work reported, whether that is in the conception, study design, execution, acquisition of data, analysis and interpretation, or in all these areas; took part in drafting, revising or critically reviewing the article; gave final approval of the version to be published; have agreed on the journal to which the article has been submitted; and agree to be accountable for all aspects of the work.
Funding
This work was supported by the grants from Guangdong Clinical Teaching Base Teaching Reform Research Project (2021 & 2023); Guangdong Graduate Education Innovation Program (2024JGXM_149) and Guangzhou Municipal Key Discipline in Medicine (2025–2027).
Disclosure
The authors declare no competing interests.
References
1. Barrows HS, Abrahamson S. The programmed patient: a technique for appraising student performance in clinical neurology. J Med Educ. 1964;39:802–805.
2. Boulet JR, McKinley DW, Whelan GP, Hambleton RK. Quality assurance methods for performance-based assessments. Adv Health Sci Educ Theory Pract. 2003;8(1):27–47. doi:10.1023/A:1022639521218
3. Brenner AM. Uses and limitations of simulated patients in psychiatric education. Acad Psychiatry. 2009;33(2):112–119. doi:10.1176/appi.ap.33.2.112
4. Niu L, Mei Y, Xu X, et al. A novel strategy combining Mini-CEX and OSCE to assess standardized training of professional postgraduates in department of prosthodontics. BMC Med Educ. 2022;22(1):888. doi:10.1186/s12909-022-03956-w
5. Norcini J, Burch V. Workplace-based assessment as an educational tool: AMEE Guide No. 31. Med Teach. 2007;29(9):855–871. doi:10.1080/01421590701775453
6. Plakiotis C. Objective Structured Clinical Examination (OSCE) in psychiatry education: a review of its role in competency-based assessment. Adv Exp Med Biol. 2017;988:159–180.
7. Burch VC, Nash RC, Zabow T, et al. A structured assessment of newly qualified medical graduates. Med Educ. 2005;39(7):723–731. doi:10.1111/j.1365-2929.2005.02192.x
8. Huang Y, Wang Y, Wang H, et al. Prevalence of mental disorders in China: a cross-sectional epidemiological study. Lancet Psychiatry. 2019;6(3):211–224. doi:10.1016/S2215-0366(18)30511-X
9. Li W, Ng RMK, Li L. Psychiatric education in Greater China. Int Rev Psychiatry. 2020;32(2):167–171. doi:10.1080/09540261.2019.1687427
10. Esagian G, Esagian-Pouftsis S, Kaprinis SG. Empathy in psychiatry and psychotherapy. Psychiatriki. 2019;30(2):156–164. doi:10.22365/jpsych.2019.302.156
© 2025 The Author(s). This work is published and licensed by Dove Medical Press Limited. The
full terms of this license are available at https://www.dovepress.com/terms.php
and incorporate the Creative Commons Attribution
- Non Commercial (unported, 3.0) License.
By accessing the work you hereby accept the Terms. Non-commercial uses of the work are permitted
without any further permission from Dove Medical Press Limited, provided the work is properly
attributed. For permission for commercial use of this work, please see paragraphs 4.2 and 5 of our Terms.