Skip to main content

Nursing students’ clinical judgment skills in simulation and clinical placement: a comparison of student self-assessment and evaluator assessment

Abstract

Background

Clinical judgment is an important and desirable learning outcome in nursing education. Students must be able to self-assess their clinical judgment in both the simulation and clinical settings to identify knowledge gaps and further improve and develop their skills. Further investigation is needed to determine the optimal conditions for and reliability of this self-assessment.

Aims

This study aimed to compare the same group of students’ self-assessment of clinical judgment with an evaluator’s assessment in both simulation and clinical settings. The study further aimed to investigate whether the Dunning-Kruger effect is present in nursing students’ self-assessment of clinical judgment.

Methods

The study applied a quantitative comparative design. It was conducted in two learning settings: an academic simulation-based education course, and a clinical placement course in an acute care hospital. The sample consisted of 23 nursing students. The Lasater Clinical Judgment Rubric was used to collect data. The scores were compared using a t-test, intraclass correlation coefficient, Pearson’s correlation coefficient, and Bland-Altman plots. The Dunning-Kruger effect was investigated using linear regression analysis and a scatter plot.

Results

The results showed an inconsistency between student self-assessment and evaluator assessment of clinical judgment in both simulation-based education and clinical placement. Students overestimated their clinical judgment when compared to the more experienced evaluator’s assessment. Differences between students’ scores and the evaluator’s scores were larger when the evaluator’s scores were low, indicating the presence of the Dunning-Kruger effect.

Conclusion

It is vital to acknowledge that student self-assessment alone may not be a reliable predictor of a student’s clinical judgment. Students who had a lower level of clinical judgment were likely to be less aware that this was the case. For future practice and research, we recommend a combination of student self-assessment and evaluator assessment to provide a more realistic view of students’ clinical judgment skills.

Peer Review reports

Background

Clinical judgment skills are required to provide safe patient care and is therefore an important and desired learning outcome in nursing education [1,2,3,4]. The term clinical judgment skills are defined by Benner and Tanner [5] (p200) as “the ways in which nurses come to understand the problems, issues, or concerns of clients and patients, to attend to salient information, and to respond in concerned and involved ways”. Simulation-based education and clinical placement are the learning activities in nurse education most relevant to facilitating the development of students’ clinical judgment [2, 3, 6,7,8,9]. Failure for students to receive educational support and thereby achieve an adequate level of clinical judgment constitutes a major threat to patient safety, potentially leading to negative consequences for patients and society [10, 11].

Assessment of student competence is a pillar of education and is necessary to determine students’ further learning needs [12,13,14]. Hence, it is important to assess nursing students’ level of clinical judgment in the simulation setting and the clinical setting. Having the most accurate picture possible of nursing students’ clinical judgment skills can help educators identify knowledge gaps that hinder students in making sound clinical judgments [13]. By identifying students’ knowledge and skill gaps, educators can further support the development of competence to better meet professional nursing care standards for patients with multifaceted issues [6, 8, 13, 14].

Nursing students’ clinical judgment skills can be assessed by an evaluator, such as a faculty member or clinical supervisor, or by students themselves using self-assessment [8, 11, 15]. Evaluators who perform assessments must be trained in observing and mapping more objectively based on observations, as well as in the use of the instrument assessing the skills in question [16]. Andrade [17] defines self-assessment as “the act of monitoring one’s processes and products in order to make adjustments that deepen learning and enhance performance”. As used in education, self-assessment is considered to promote students’ responsibility for and self-regulation of their own learning [18]. Students’ ability to judge the quality of their own and other’s work is vital for patient safety and healthcare quality [19]. This capability can also be defined as evaluate judgment and such skill might support students’ learning after graduating [19, 20]. Self-regulatory skills such as self-assessment and evaluating judgment may support students in directing and regulating their actions towards learning outcomes and are thus necessary for the transition from novice student to lifelong learner in clinical practice [12, 18, 19, 21]. In research, self-assessment is commonly used to explore and describe students’ behaviour, skills, performance, and experiences [16]. Additionally, student self-assessment is often chosen in education and research to minimize the resources required, such as faculty and researcher staff time [16,17,18, 22].

Students’ self-assessment processes have been investigated in various ways. Consistency between different assessment methods has been found to be valuable for identifying students’ knowledge gaps and subsequently improving their nursing skills, performance, and behaviour [16]. Consistency has typically been investigated by comparing students’ self-assessments with an experienced evaluator’s assessment [21, 23]. To our knowledge, three previous studies [24,25,26] have compared nursing students’ self-assessment and evaluators’ assessment of students’ clinical judgment using the Lasater Clinical Judgment Rubric (LCJR) [6]. The use of rubrics is considered the key to reliable assessment in education and research [27]. Rubrics include assessment criteria, levels of performance and the weights of each criterion [28]. According to a recent systematic literature review, LCJR is currently the most recognized instrument for assessing nursing students’ clinical judgment [29]. Two of the three previous studies comparing students’ and evaluator assessment were conducted in the simulation setting [24, 26] while the third study [25] was conducted in the clinical setting. The overall conclusion in all these three studies is that students tend to overestimate their clinical judgment skills in both the simulation and the clinical setting when compared to an evaluator assessment. However, none of the studies investigated the same students in different settings, even though it has been argued that self-assessment should be investigated under different settings [16, 17, 21, 23]. Thus, looking at the same group of students’ self-assessment of clinical judgment skills in two settings may provide valuable knowledge.

Addressing the process of assessing clinical judgment skills, the response bias from students’ self-assessment is of interest as it may act as a barrier to reflection and learning [16]. One example of response bias is the Dunning-Kruger effect [30], which identifies that individual with low competence overestimate their competence. The Dunning-Kruger effect can be identified by a simple calculation of the difference between a student’s subjective self-assessment and a more objective assessment performed by an experienced evaluator [16, 30]. If the Dunning-Kruger effect is present among nursing students and they are unable to recognize their deficits in clinical judgment, relying heavily on student self-assessment of clinical judgment may lead to inaccurate evaluations in educational learning outcomes and research, and ultimately threaten patient safety and patient care [16, 31, 32].

Irrespective of the benefits mentioned above and the established use of self-assessment of nursing students’ clinical judgment skills in education and research, knowledge gaps concerning the assessment process still exist. As the organizational and pedagogical approaches used in the simulation setting and the clinical setting differ, it is of interest to investigate the self-assessment process in both settings. Such knowledge may enable educators to apply appropriate pedagogical approaches to further develop students’ clinical judgment. To our knowledge, there are no existing studies comparing the same group of students’ self-assessments with evaluators’ assessments in two different settings. Moreover, no previous studies have investigated whether the Dunning-Kruger effect [30] is present in nursing students’ self-assessment of clinical judgment skills.

Thus, this study aimed to compare the same students’ self-assessments of clinical judgment with evaluators’ assessments in both simulation and clinical settings. The study further aimed to investigate whether the Dunning-Kruger effect [30] is present in nursing students’ self-assessment of clinical judgment. The research questions were as follows:

  1. 1.

    Did nursing students’ self-assessment of clinical judgment in the simulation setting reflect their clinical judgment as assessed by an evaluator?

  2. 2.

    Did the same nursing students’ self-assessment of clinical judgment in the clinical placement setting reflect their clinical judgment as assessed by an evaluator?

  3. 3.

    Is the Dunning-Kruger effect present in nursing students’ self-assessment of clinical judgment in the simulation setting or the clinical setting?

Methods

Research design

This study uses a quantitative, comparative design and is reported in accordance with the STROBE guidelines (Additional file 1) and the Reporting Guidelines for Health Care Simulation Research [33]. The study is part of a larger study addressing nursing students’ professional competence and clinical judgment.

Research settings

The study took place in the second year of a three-year Bachelor of Nursing programme at a Norwegian university. This undergraduate nursing education programme entailed 180 credits in the European Credit Transfer and Accumulation System (ECTS) [34]. More specifically, the nursing students got 90 ECTS credits from theoretical courses mainly in the academic setting, minimum 75 ECTS credits from clinical placement in a variety of settings, and maximum 15 ECTS credits from simulation-based education in laboratories [34]. The study was conducted in two learning settings: a simulation centre on the university campus and an acute care hospital unit.

In the simulation setting, the students took part in a two-day simulation-based education course comprising six simulation sessions focusing on different deteriorated patient conditions and diagnoses. Nine faculty members were involved as facilitators and operators. Students were divided into groups of six to nine, alternating between the roles of nurse and observer. The simulation environment mirrored a patient room in a hospital unit and Laerdal SimMan 3G™ and ALS™ manikins were used. Each simulation session (90 min) consisted of a prebriefing (15 min), a simulated scenario (15 min), a viewing of the video recording of the simulated scenario (15 min), and a facilitator-led group debriefing (45 min). For the debriefing, the Promoting Excellence and Reflection in Simulation (PEARLS) structured and scripted debriefing [35] method was used.

After the simulation-based education course, the students attended an eight-week clinical placement course in a medical or surgical hospital unit hosting adult patients with acute, critical, and chronic conditions. Students provided nursing care under the supervision of a registered nurse working in the relevant unit. Nurse educators supervised the students in groups to promote reflection and learning and to evaluate their learning outcomes.

The learning outcomes for both courses entailed the same clinical judgment skills.

Recruitment and participants

The target group for the study was second-year nursing students. In advance of the recruitment, all students had completed theoretical courses addressing pathology and core nursing issues related to patients in need of acute care, had passed a six-week clinical placement course in a nursing home, were certified in cardiopulmonary resuscitation, and had attended compulsory classes in practical nursing skills. For recruiting, information about the study was published on the university’s digital learning platform and distributed in a pre-clinical course by the first author. Eligible participants were informed about the study aim, data collection methods, confidentiality, voluntary participation, and their right to withdraw. A sample size calculation showed that 16 student-evaluator comparisons were sufficient to identify an average 2-point difference between student and evaluator scores on the LCJR, with a standard deviation of 4 points. Due to the predetermined organisation of the simulation-based education and the clinical placement courses, it was impossible to collect data from more than one student at a time. Consequently, the study allowed for a maximum of 24 participants out of the 89 students attending the courses. Of these, the first 24 students who signed up to participate were formally invited, of which N = 23 participated. The sample consisted of 19 women (82.6%) and four men (17.4%), with ages ranging from 20 to 54 years (Mean = 28 years). None of the participants had previous experience with scenario simulation, and 78.3% had experience working in healthcare services.

Measure

The Norwegian version of the Lasater Clinical Judgment Rubric (LCJR-N) [9] was used to collect data concerning nursing students’ clinical judgment skills. The original LCJR was developed by Lasater [6] to directly observe and evaluate students’ individual performance of clinical judgment in a simulation setting. It was designed to provide a common language for learners, faculty, and preceptors to talk about learners’ thinking and to serve as a help for offering formative guidance and feedback [6, 11]. It is based on Benner’s novice to expert model [36] and Tanner’s clinical judgment model [4]. LCJR has emerged as a tool used by evaluators for observation and by students for self-assessment in both simulation and clinical settings [11, 15, 37]. The LCJR corresponded well to students’ learning outcomes in the simulation-based education course and the clinical placement course. The LCJR consists of four dimensions, called subscales in the present study, with a total of 11 items: Noticing (3 items), Interpreting (2 items), Responding (4 items), and Reflecting (2 items) [6]. The items on students’ performance were scored from 1 to 4 with higher scores indicating better clinical judgment: 1 point = beginning, 2 points = developing, 3 points = accomplished, and 4 points = exemplary [6]. The total score ranges from 11 to 44.

The LCJR has been translated into Norwegian, Swedish, German, Chinese and several other languages [9, 25, 38, 39]. In a recent review, internal consistency was supported for both evaluator and student self-assessment [15]. Regarding reliability and validity for the Norwegian version of LCJR (LCJR-N) in previous research, Cronbach’s alphas (0.74–0.91) indicated good internal consistency and face validity was verified [9]. In the current study, the Cronbach’s alphas for the LCJR-N total score ranged from 0.87 to 0.91, and from 0.69 to 0.85 for the Noticing and Responding subscales. Alpha values were not calculated for the Interpreting and Reflection subscales as these scales only had two items each.

Data collection

Data were collected from students and the evaluator in December 2019 in the simulation setting and in February 2020 in the clinical setting. Data on students’ self-assessments using the LCJR-N were collected by self-reported questionnaires in pen and paper format together with demographic information. Data from the evaluator were also collected by using LCJR-N in pen and paper format.

In the simulation setting, the evaluator completed the LCJR-N for each student while observing the student in the simulation scenario. Data for the subscale Reflecting was collected by observing the students in the debriefing. Immediately after the simulation-based education course, the LCJR-N was handed out to the students. Each student completed the LCJR-N while recalling the simulation scenario, in which they had monitored vital signs on the manikin. The questionnaires were distributed and collected by faculty members who were not otherwise engaged in the study.

In the clinical setting, the same evaluator completed the LCJR-N for each student while observing the students in a patient care situation where the student monitored a patient’s vital signs. Data for the subscale Reflecting were collected by posing three questions to each student after they left the patient’s room (“If you had to do it again, would you do anything differently?”, “What would you do then?”, and “Why would you do this differently?”). Immediately after, each student completed the LCJR-N while recalling the patient care situation. The LCJR-Ns were distributed to the students and collected by the evaluator.

The term “evaluator” in this study refers to first author, who is a registered nurse (RN) with a Master’s degree in Nursing Science (MSN), a researcher, and a faculty member. The evaluator was not involved in any of the participants’ educational activities. The evaluator has years of experience with the simulation setting and pedagogical approaches in simulation-based education, as well as with supervising and assessing students in clinical placements. Moreover, the evaluator is a clinically experienced RN having worked 15 years in acute care units entailing using clinical judgment skills when caring for deteriorating patients. The evaluator’s preparatory work for data collection included examining the concept of clinical judgment and using LCJR-N as an observation tool by testing it in a simulation scenario. During this preparation, the evaluator corresponded with the LCJR’s developer Kathie Lasater regarding the use of the LCJR in various assessments and the use of only one evaluator. Because using only one evaluator may create evaluation biases [40], this issue was carefully considered. The credibility of data collected by only one evaluator was considered acceptable and in line with reported findings in a recent review by Lee [15] demonstrating high interrater reliability metrics for the LCJR. All students had previous experience with use of the LCJR-N from having participated in an earlier research study.

Statistical analysis

Data were analysed by SPSS version 28.0. A paired-samples t-test was used to compare the students’ and the evaluator’s LCJR-N scores. Intraclass correlation coefficient (ICC) was used to investigate degrees of correlation and agreement between students’ and the evaluator’s LCJR-N scores, in line with the suggestions of Koo and Li [41]. ICC estimates and their 95% confidence intervals were based on a mean rating (k = 2), consistency, and a 2-way mixed-effects model [41]. ICC was interpreted in line with Landis and Koch [42], with values ≤ 0.20 indicating slight agreement, from 0.21 to 0.40 indicating fair agreement, 0.41 to 0.60 indicating moderate agreement, 0.61 to 0.80 indicating substantial agreement, and ≥ 0.81 indicating almost perfect agreement. Pearson’s correlation coefficient was used to investigate the relationship between students’ and the evaluator’s LCJR-N scores. Pearson’s correlation coefficients were interpreted as r = 0.10, 0.30, and 0.50 indicating a small, medium, or large correlation, respectively [43]. Bland-Altman plots were created to illustrate the average bias and to investigate whether there were systematic differences between students’ and the evaluator’s LCJR-N scores [44].

Linear regression analysis was used to investigate whether the Dunning-Kruger effect was present in nursing students’ self-assessment of clinical judgment skills in the simulation setting or the clinical placement setting. The linear regression analysis determined whether the discrepancy between student LCJR-N scores and evaluator LCJR-N scores was the same across the evaluator’s LCJR-N scores or increased with lower values on the evaluator’s LCJR-N scores. A scatter plot was created to illustrate the results of the linear regression.

The p-value for statistical significance was set at < 0.05.

Results

Comparison of student self-assessment and evaluator assessment of students’ clinical judgment in the simulation setting

In the simulation setting, students’ LCJR-N total score and subscale scores were significantly higher than the evaluator’s scores (Table 1). The Pearson’s correlation coefficients for student and evaluator assessments for both total score and subscales were quite low (-0.01 to 0.32), with none of them reaching statistical significance (Table 2). The ICC of the LCJR-N total score and the subscale Noticing ranged from − 0.01 to 0.17, indicating “slight agreement” between the students’ and the evaluator’s assessments. The ICC scores for the subscales Interpreting, Responding, and Reflecting ranged from 0.32 to 0.39, indicating “fair agreement” between the students’ and evaluator’s scores in the simulation setting (Table 2). The Bland-Altman plots showed a systematic difference and wide limits of agreement between students’ and evaluator’s LCJR-N total score and subscale scores. The Bland-Altman plots for all LCJR-N subscales and total score illustrated that students’ scores were higher than the evaluator’s score. Figure 1 shows an example of the Bland-Altman plot for the LCJR-N total score in the simulation setting.

Table 1 Comparison of students’ self-assessment and evaluator’s assessment on LCJR-N
Table 2 Agreement and correlation between students’ self-assessment and evaluator’s assessment on LCJR-N
Fig. 1
figure 1

Bland-Altman plots of students’ self-assessment and evaluator’s assessment on LCJR-N total scores in simulation-based education (SBE) and clinical placement

Comparison of student self-assessment and evaluator assessment of students’ clinical judgment in the clinical setting

In the clinical setting, students’ LCJR-N total score and subscale scores were higher than the evaluator’s scores; however, this difference was not significant (Table 1). The Pearson’s correlation coefficient (r) for student and evaluator assessments on LCJR-N total score and subscales were quite low (-0.27 to 0.19) and none of them reached statistical significance (Table 2). The ICC values of the LCJR-N total score and all subscales ranged from − 0.26 to 0.19, indicating “slight agreement” between the students’ and the evaluator’s assessments (Table 2). The Bland-Altman plots indicated a systematic difference and wide limits of agreement between students’ and the evaluator’s LCJR-N total score and all subscale scores. Each Bland-Altman plot showed that students’ scores were higher than the evaluator’s scores. Figure 1 shows an example of the Bland-Altman plot for LCJR-N total score in the clinical setting.

The Dunning-Kruger effect in students’ self-assessment of clinical judgment

In the simulation setting, the linear regression analysis of LCJR-N total score and subscales showed that the difference between the students’ scores and the evaluator’s score increased significantly as the evaluator’s score decreased (Table 3; Fig. 2). This means that the differences between student and evaluator scores were larger when the evaluator’s score was low.

Table 3 Changes in student-evaluator differences on LCJR-N by evaluators assessment on LCJR-N
Fig. 2
figure 2

Scatter plots of differences in student-evaluator LCJR-N total scores by evaluator’s LCJR-N total score in simulation-based education (SBE) and clinical placement

In the clinical setting, the linear regression analysis of LCJR-N total score and the subscales Noticing, Interpreting, and Reflecting showed that the difference between student and evaluator scores increased significantly as the evaluator’s score decreased (Table 3; Fig. 2). In absolute terms, the patterns were similar for the subscale Responding, however, this regression effect was not significant (Table 3; Fig. 2).

Discussion

Comparison of student self-assessment and evaluator assessment of students’ clinical judgment

Comparing students’ self-assessment and evaluator assessment of students’ clinical judgment skills, the overall results showed an inconsistency in both the simulation and clinical settings. Students assessed their clinical judgment as being higher than the evaluator did. When comparing the assessments using t-tests, this difference was significant in the simulation setting but not in the clinical setting. However, using supplementary statistical tests such as Pearson’s r, ICC, and scatter plots, we found the inconsistency between student and evaluator assessment to be present independently of the learning setting. These findings regarding assessment of nursing students’ clinical judgment both in the simulation and clinical settings add valuable knowledge to this research field.

Because no existing research has investigated the same group of students in two educational settings, in what follows we compare our results with findings from research conducted in one educational setting. The student-evaluator inconsistency identified in our study concurs with previous studies concluding that students tend to overestimate their clinical judgment compared to evaluators [24,25,26]. In accordance with our findings from the simulation setting, Strickland and Cheshire [26] found student self-assessment in the simulation setting to be higher than evaluator assessment, and they reported a positive, although not strong, correlation (r = 0.31) between these assessments. Likewise, Jensen [24] found that students rated themselves higher than the evaluator did in the simulation setting, although not significantly higher. Corresponding to our findings, Jensen [24] also reported weak correlations (r = -0.14–0.27) between students’ assessment and evaluators’ assessment. In accordance with our findings from the clinical placement setting, Vreugdenhil and Spek [25] found the student-evaluator difference to be systematic and significant (p = 0.020) when investigating agreement, with students tending to score themselves significantly higher (6.1%) than the evaluator did. As in our findings, Vreugdenhil and Spek [23] did not find any significant differences between students’ and evaluators’ assessments in a t-test analysis, but they did find a strong positive correlation (r = 0.78) between students’ self-assessment and evaluator assessment, which is different from our findings. Taken together, previous studies and our study show that students tend to rate their clinical judgment higher than the evaluator, regardless of being studied in the simulation or the clinical setting and regardless of being studied in one or two settings.

The inconsistency between the same students’ and same evaluator’s assessments in the simulation setting and the clinical setting in our study may have several explanations. The student-evaluator inconsistency might be due to different understandings of the concept of clinical judgment. Although students were trained in the use of the LCJR-N, the items on which corresponded to the learning outcomes in the simulation-based education course and the clinical placement course, their cognitive or linguistic representation of clinical judgment might still be limited [45]. On this issue, the use of a rubric as LCJR-N in the assessment process could make it easier for students and evaluators to recognize the expectations for clinical judgment [6]. However, assessing clinical judgment is complicated and requires metacognitive skills, the ability to think abstractly, and an in-depth understanding of nurses’ responsibilities and role in the clinical setting [29]. The students and the evaluator might have had different perceptions of clinical judgment and therefore interpreted the assessment criteria in the LCJR-N differently. The students might have focused on specific tasks more than on cognitive processes in clinical judgment. Novice students often lack capability to reflect abstractly on theoretical and practical aspects of a skill and thus tend to focus on superficial features of their performance in the self-assessment process [16, 23]. This argument aligns well with Benner’s “from novice to expert” theory [36], which identifies five levels of competence in nursing – novice, advanced beginner, competent, proficient, and expert – each of which builds upon the previous one. Benner [36] describes nursing students as being at a novice level, characterised by having little experience and understanding of the situations in which they are expected to perform, which is often the case. In the current study, students’ limited experience might have affected their understanding of the contextual meaning of clinical judgment and how to apply these skills in simulation scenarios or clinical situations, and thus also influenced their self-assessments.

The Dunning-Kruger effect in student self-assessment of clinical judgment

To investigate whether the Dunning-Kruger effect was present in students’ self-assessment of clinical judgment, the evaluator’s score for students’ clinical judgment was considered more accurate due to the evaluator’s training in the use of LCJR-N and her higher level of education, competence, and experience.

The findings indicate that the Dunning-Kruger effect was present in students’ self-assessment of clinical judgment in both the simulation setting and the clinical setting due to the larger student-evaluator differences when the evaluator’s score was low. These findings are consistent with findings reported in a recent review on the Dunning-Kruger effect in a variety of educational contexts in the social sciences [16].

An explanation for our findings might be students’ lack of metacognitive awareness of their own clinical judgment. In other words, the students who were assessed by evaluator as having a lower level of clinical judgment were unaware of having a low level, and therefore were more likely to overestimate their clinical judgment [30]. It is questionable whether novice nursing students’ halfway through a bachelor’s programme have sufficient metacognitive skills and a sufficient level of self-reflection to accurately assess their own clinical judgment. Metacognitive skills involving assessing one’s own competence develop through self-evaluation, self-reflection, and feedback from others [16, 30]. Reflection itself is vital and valued in simulation-based education and clinical placement in nursing education. Reflection in these settings comprises students’ assessment of their actions and previous clinical situations followed by the integration of new knowledge and adjustment of clinical performance [4, 46]. Such assessment might promote learning and has the potential to develop students’ evaluative judgment and further lifelong learning [19, 47]. In the simulation setting in this study, students’ self-reflection on clinical judgment was carefully promoted by using the student-centred and structured Promoting Excellence and Reflection in Simulation (PEARLS) debriefing [35]. While student-centred and structured debriefing approaches have the potential to provide students with optimal opportunities for reflection and increased activity [48, 49], a pitfall in debriefing is that facilitators do not appropriately close all relevant performance gaps [50]. Hence, the facilitator might not have been attentive to students’ performance gaps concerning clinical judgment, resulting in missed learning opportunities.

Acknowledging the potential for the Dunning-Kruger effect in students’ self-assessment of clinical judgment in nursing education offers opportunities for establishing meaningful feedback discussions while learning and improving [16]. As novice students gradually develop metacognitive skills over the course of their education, blind spots regarding their own clinical judgment might decrease [36]. Hence, students’ metacognitive skills and the potential presence of the Dunning-Kruger effect should always be considered when deciding on an assessment method for nursing student’s clinical judgment.

Limitations

The study has some limitations. The sample size and the fact that there was only one sampling site limit the generalisability of the findings. There may also be measurement errors due to the use of only one evaluator [51]. Despite the evaluator’s theoretical and practical preparation to avoid observational biases, having only one evaluator eliminated the possibility of doing interrater reliability analysis on scores between evaluators [40]. Although the evaluator was prepared for the observation and not involved with students from other learning activities, objective observation and assessment of skills such as clinical judgment is always a challenge [40, 51]. Another potential measurement error is that students’ behaviour in the data collection situations might have been atypical due to their awareness of being observed [52], also known as the Hawthorne effect [53]. Finally, there is also a risk of instrumentation bias as the LCJR-N has not been psychometrically tested for the Norwegian context.

Implications for education and future research

Although nursing students’ self-assessment is widely used and considered valuable for evaluation and learning [15, 16, 29], our findings urge caution when interpreting nursing students’ self-assessment of clinical judgment in education. Students’ ability to determine their own level of competence and identify knowledge gaps is decisive for clinical performance within the limits of their competence in a lifelong learning perspective [19, 54,55,56]. Therefore, nurse educators should facilitate students’ metacognitive skills and their evaluative judgement related to clinical judgment. Further, acknowledging the presence of the Dunning-Kruger effect among nursing students may inspire faculty to promote students’ metacognitive skills and self-reflection, thereby supporting students in their learning process [16]. Promoting nursing students’ self-reflection regarding clinical judgment by using LCJR in simulation-based education and various clinical placement settings may help students gain a deeper understanding of the concept of clinical judgment before graduating. For future educational assessment practice in simulation and clinical settings, a combination of assessment methods is recommended [13]. Student self-assessment, evaluator assessment, and feedback may offer a more realistic interpretation of students’ clinical judgment and help faculty to identify those students who require additional support during their education before graduation [24, 26, 57].

For future research, pedagogical interventions aiming to promote nursing students’ metacognitive skills in relation to clinical judgment using controlled designs should be performed. Researchers should be aware of the Dunning-Kruger effect and its potential impact on validity when having students’ self-assessments as the only data source. Moreover, the Dunning-Kruger effect among nursing students should also be investigated using larger samples and other instruments. Finally, studies using a qualitative approach to explore nursing students’ experiences from self-assessment of clinical judgment are welcomed.

Conclusion

This study contributes to the body of knowledge regarding assessment of nursing students’ clinical judgment using the LCJR-N in the field of nursing education and research. Overall, our findings indicate an inconsistency between student self-assessment and evaluator assessment in the simulation setting and in the clinical setting, with students tending to have a higher estimation of their own clinical judgment compared to an evaluator’s assessment. The findings further demonstrate that the Dunning-Kruger effect was present in our sample, as students whom the evaluator assessed as having a lower level of clinical judgment were likely to be unaware of their own low level.

For future practice and research, it is vital to acknowledge that student self-assessment alone may not be a reliable predictor of a student’s clinical judgment. Thus, we recommend a combination of student self-assessment and evaluator assessment to provide a more realistic view of students’ clinical judgment.

Data Availability

The datasets used and/or analysed during the current study are available from the corresponding author on reasonable request.

References

  1. Sterner A, Ramstrand N, Palmér L, Hagiwara MA. A study of factors that predict novice nurses’ perceived ability to provide care in acute situations. Nurs Open. 2021;8(4):1958–69.

    Article  PubMed  PubMed Central  Google Scholar 

  2. Gonzalez L, Nielsen A, Lasater K. Developing students clinical reasoning skills: a faculty guide. J Nurs Educ. 2021;60(9):485–93.

    Article  PubMed  Google Scholar 

  3. Kavanagh JM, Szweda C. A crisis in competency: the strategic and ethical imperative to assessing new graduate nurses’ clinical reasoning. Nurs Educ Perspect. 2017;38(2):57–62.

    Article  PubMed  Google Scholar 

  4. Tanner C. Thinking like a nurse: a research-based model of clinical judgment in nursing. J Nurs Educ. 2006;45(6):204–11.

    Article  PubMed  Google Scholar 

  5. Benner P, Tanner C, Chesla C. Expertise in nursing practice: caring, clinical judgment, and ethics. New York: Springer Publishing Company; 2009.

    Book  Google Scholar 

  6. Lasater K. Clinical judgment development: using simulation to create an assessment rubric. J Nurs Educ. 2007;46(11):496–503.

    Article  PubMed  Google Scholar 

  7. Theobald KA, Tutticci N, Ramsbotham J, Johnston S. Effectiveness of using simulation in the development of clinical reasoning in undergraduate nursing students: a systematic review. Nurse Educ Pract. 2021;57:103220.

    Article  PubMed  Google Scholar 

  8. Jessee MA. An update on clinical judgment in nursing and implications for education, practice, and regulation. J Nurs Regul. 2021;12(3):50–60.

    Article  Google Scholar 

  9. Høegh-Larsen AM, Ravik M, Reierson I, Husebø SIE, Gonzalez MT. PEARLS debriefing compared to standard debriefing effects on nursing students’ professional competence and clinical judgment: a quasi-experimental study. Clin Simul Nurs. 2022. https://doi.org/10.1016/j.ecns.2022.09.003.

    Article  Google Scholar 

  10. Parodis I, Andersson L, Durning SJ, Hege I, Knez J, Kononowicz AA, et al. Clinical reasoning needs to be explicitly addressed in health professions curricula: recommendations from a european consortium. Int J Environ Res Public Health. 2021;18(21):11202.

    Article  PubMed  PubMed Central  Google Scholar 

  11. Lasater K. Clinical judgment: the last frontier for evaluation. Nurse Educ Pract. 2011;11(2):86–92.

    Article  PubMed  Google Scholar 

  12. Siles-González J, Solano-Ruiz C. Self-assessment, reflection on practice and critical thinking in nursing students. Nurse Educ Today. 2016;45:132–7.

    Article  PubMed  Google Scholar 

  13. Shinnick MA, Woo M. Comparison of simulation assessments: can they identify novice/student nurses? Clin Simul Nurs. 2020;46:40–9.

    Article  Google Scholar 

  14. Immonen K, Oikarainen A, Tomietto M, Kääriäinen M, Tuomikoski A-M, Kaučič BM, et al. Assessment of nursing students’ competence in clinical practice: a systematic review of reviews. Int J Nurs Stud. 2019;100:103414.

    Article  PubMed  Google Scholar 

  15. Lee KC. The Lasater Clinical Judgment Rubric: implications for evaluating teaching effectiveness. J Nurs Educ. 2021;60(2):67–73.

    Article  CAS  PubMed  Google Scholar 

  16. Bradley CS, Dreifuerst KT, Johnson BK, Loomis A. More than a meme: the Dunning-Kruger effect as an opportunity for positive change in nursing education. Clin Simul Nurs. 2022;66:58–65.

    Article  Google Scholar 

  17. Andrade HL. A critical review of research on student self-assessment. Front Educ Res. 2019;4:87.

    Article  Google Scholar 

  18. Piper K, Morphet J, Bonnamy J. Improving student-centered feedback through self-assessment. Nurse Educ Today. 2019;83:104193.

    Article  PubMed  Google Scholar 

  19. Henderson B, Chipchase L, Aitken R, Lewis LK. Consensus marking as a grading method for the development of evaluative judgement: comparing assessor and students. Nurse Educ Pract. 2022;63:103386.

    Article  PubMed  Google Scholar 

  20. Tai J, Ajjawi R, Boud D, Dawson P, Panadero E. Developing evaluative judgement: enabling students to make decisions about the quality of work. High Educ. 2018;76(3):467–81.

    Article  Google Scholar 

  21. Brown GTL, Andrade HL, Chen F. Accuracy in student self-assessment: directions and cautions for research. Assess Educ Princ Policy Pract. 2015;22(4):444–57.

    Google Scholar 

  22. Kajander-Unkuri S, Leino-Kilpi H, Katajisto J, Meretoja R, Räisänen A, Saarikoski M, et al. Congruence between graduating nursing students’ self-assessments and mentors’ assessments of students’ nurse competence. Collegian. 2016;23(3):303–12.

    Article  Google Scholar 

  23. Ross JA. The reliability, validity, and utility of self-assessment. Pract Assess Res Eval. 2006;11:10.

    Google Scholar 

  24. Jensen R. Clinical reasoning during simulation: comparison of student and faculty ratings. Nurse Educ Pract. 2013;13(1):23.

    Article  PubMed  Google Scholar 

  25. Vreugdenhil J, Spek B. Development and validation of dutch version of Lasater Clinical Judgment Rubric in hospital practice: an instrument design study. Nurse Educ Today. 2018;62:43–51.

    Article  PubMed  Google Scholar 

  26. Strickland HP, Cheshire MH, March AL. Clinical judgment during simulation: a comparison of student and faculty scores. Nurs Educ Perspect. 2017;38(2):85–6.

    Article  PubMed  Google Scholar 

  27. Martínez-Huertas J, Jastrzebska O, Olmos R, León JA. Automated summary evaluation with inbuilt rubric method: an alternative to constructed responses and multiple-choice tests assessments. Assess Eval High Educ. 2019;44(7):1029–41.

    Article  Google Scholar 

  28. Dawson P. Assessment rubrics: towards clearer and more replicable design, research and practice. Assess Eval High Educ. 2017;42(3):347–60.

    Article  Google Scholar 

  29. Brentnall J, Thackray D, Judd B. Evaluating the clinical reasoning of student health professionals in placement and simulation settings: a systematic review. Int J Environ Res Public Health. 2022;19(2):936.

    Article  PubMed  PubMed Central  Google Scholar 

  30. Kruger J, Dunning D. Unskilled and unaware of it: how difficulties in recognizing one’s own incompetence lead to inflated self-assessments. Psychology. 2009;1:30–46.

    Google Scholar 

  31. Song Y, McCreary LL. New graduate nurses’ self-assessed competencies: an integrative review. Nurse Educ Pract. 2020;45:102801.

    Article  PubMed  Google Scholar 

  32. Wang PZT, Xie WY, Nair S, Dave S, Shatzer J, Chahine S. A comparison of guided video reflection versus self-regulated learning to teach knot tying to medical students: a pilot randomized controlled trial. J Surg Educ. 2020;77(4):805–16.

    Article  PubMed  Google Scholar 

  33. Cheng A, Kessler D, Mackinnon R, Chang TP, Nadkarni VM, Hunt EA, et al. Reporting guidelines for health care simulation research: extensions to the CONSORT and STROBE statements. Adv Simul. 2016;1(1):25.

    Article  Google Scholar 

  34. Forskrift til rammeplan. for sykepleierutdanning [Regulations for the framework plan for nursing education]. LOV-2005-04-01-15-§ 3 – 2 [cited 2022 December 13]. Available from: https://lovdata.no/dokument/SF/forskrift/2008-01-25-128

  35. Eppich W, Cheng A. Promoting Excellence and reflective learning in Simulation (PEARLS): development and rationale for a blended approach to health care simulation debriefing. Simul Healthc. 2015;10(2):106–15.

    Article  PubMed  Google Scholar 

  36. Benner P. From novice to expert: excellence and power in clinical nursing practice. Menlo Park, Calif: Addison-Wesley; 1984.

    Book  Google Scholar 

  37. Lasater K. High-fidelity simulation and the development of clinical judgment: students’ experiences. J Nurs Educ. 2007;46(6):269–76.

    Article  PubMed  Google Scholar 

  38. Kristiansen L, Häggström M, Hallin K, Andersson I, Bäckström B. Svensk översättning, kvalitativ relevansvärdering och kvantitativ reliabilitetstestning av Lasater Clinical Judgment Rubric/Swedish translation, qualitative relevance evaluation and quantitative reliability test of Lasater Clinical Judgment Rubric. Nord J Nurs Res. 2015;35(2):113–22.

    Article  Google Scholar 

  39. Yang F, Wang Y, Yang C, Zhou MH, Shu J, Fu B, et al. Improving clinical judgment by simulation: a randomized trial and validation of the Lasater Clinical Judgment Rubric in Chinese. BMC Med Educ. 2019;19(1):20.

    Article  PubMed  PubMed Central  Google Scholar 

  40. Polit DF, Beck CT. Nursing research: generating and assessing evidence for nursing practice. 11th ed. Philadelphia: Lippincott Williams & Wilkins; 2020.

    Google Scholar 

  41. Koo TK, Li MY. A guideline of selecting and reporting intraclass correlation coefficients for reliability research. J Chiropr Med. 2016;15(2):155–63.

    Article  PubMed  PubMed Central  Google Scholar 

  42. Landis JR, Koch GG. The measurement of observer agreement for categorical data. Biometrics. 1977;33(1):159–74.

    Article  CAS  PubMed  Google Scholar 

  43. Cohen J. Statistical power analysis for the behavioral sciences. Academic Press; 2013.

  44. Bland JM, Altman DG. Measuring agreement in method comparison studies. Stat Methods Med Res. 1999;8(2):135–60.

    Article  CAS  PubMed  Google Scholar 

  45. Wu K, Dunning D. Hypocognition: making sense of the landscape beyond one’s conceptual reach. Rev Genl Psychol. 2018;22(1):25–35.

    Article  Google Scholar 

  46. Husebø SE, Amp, Apos, Regan S, Nestel D. Reflective practice and its role in simulation. Clin Simul Nurs. 2015;11(8):368–75.

    Article  Google Scholar 

  47. Boud D, Soler R. Sustainable assessment revisited. Assess Eval High Educ. 2016;41(3):400–13.

    Article  Google Scholar 

  48. Decker S, Alinier G, Crawford SB, Gordon RM, Jenkins D, Wilson C. Healthcare Simulation Standards of best PracticeTM the debriefing process. Clin Simul Nurs. 2021;58:27–32.

    Article  Google Scholar 

  49. Neill MA, Wotton K. High-fidelity simulation debriefing in nursing education: a literature review. Clin Simul Nurs. 2011;7(5):e161–e8.

    Article  Google Scholar 

  50. Cheng A, Grant V, Robinson T, Catena H, Lachapelle K, Kim J, et al. The promoting Excellence and reflective learning in Simulation (PEARLS) approach to health care debriefing: a faculty development guide. Clin Simul Nurs. 2016;12(10):419–28.

    Article  Google Scholar 

  51. Mahtani K, Spencer EA, Brassey J, Heneghan C. Catalogue of bias: observer bias. BMJ Evid Based Med. 2018;23(1):23–4.

    Article  PubMed  Google Scholar 

  52. Paradis E, Sutkin G. Beyond a good story: from Hawthorne Effect to reactivity in health professions education research. Med Educ. 2017;51(1):31–9.

    Article  PubMed  Google Scholar 

  53. Waring S. Richard Gillespie. Manufacturing knowledge: a history of the Hawthorne experiments [book review]. Am Hist Rev. 1992;97(4):1301. Review of: Gillespie R. Manufacturing knowledge: a history of the Hawthorne experiments. New York: Cambridge University Press; 1991.

  54. Baxter P, Norman G. Self-assessment or self-deception? A lack of association between nursing students’ self-assessment and performance. J Adv Nurs. 2011;67(11):2406–13.

    Article  PubMed  Google Scholar 

  55. Hadid S. Factors influencing nursing student self-assessment in relation to instructor assessment. J Nurs Educ. 2017;56(2):70–6.

    Article  PubMed  Google Scholar 

  56. International Council of Nurses. The ICN code of ethics for nurses. Geneva: Internation Council of Nurses; 2012.

    Google Scholar 

  57. Forsman H, Jansson I, Leksell J, Lepp M, Sundin Andersson C, Engström M, et al. Clusters of competence: relationship between self-reported professional competence and achievement on a national examination among graduating nursing students. J Adv Nurs. 2020;76(1):199–208.

    Article  PubMed  Google Scholar 

Download references

Acknowledgements

The authors would like to acknowledge the students who participated in this study and the faculty members for their assistance in collecting data. We also like to thank Professor Leiv Sandvik at University of South-Eastern Norway for his statistical support.

Funding

This research did not receive any specific grant from funding agencies in the public, commercial, or not-for-profit sectors.

Author information

Authors and Affiliations

Authors

Contributions

AMHL: Conceptualization, Methodology, Formal analysis, Investigation, Data curation, Writing-original draft, Writing- reviewing and editing, Project administration. MTG: Conceptualization, Methodology, Writing- reviewing and editing, Supervision. IÅR: Conceptualization, Methodology, Writing- reviewing and editing, Supervision. SIEH: Conceptualization, Methodology, Writing- reviewing and editing, Supervision. DH: Conceptualization, Methodology, Formal analysis, Writing- reviewing and editing, Supervision. MR: Conceptualization, Methodology, Writing- reviewing and editing, Supervision. All authors have read and agreed to the published version of the manuscript.

Corresponding author

Correspondence to Anne Mette Høegh-Larsen.

Ethics declarations

Ethics approval and consent to participate

The Norwegian Centre for Research Data approved every aspect of this study (approval number 624052). The Faculty of Health and Social Sciences at the university, the hospital administrations, and the hospital’s data protection representative approved the study. The patients involved through the data collection during the clinical placement course in the hospital unit received written and oral information about the study from the student before the evaluator entered the room, including the level of patient involvement, voluntary nature of participation, and the right to reject or withdraw from the study. Patients consented in writing to the evaluator’s presence in the room where data were collected. All participants provided written informed consent. According to national regulations, approval from a medical ethical committee to collect this type of data was not necessary. All methods in this study were carried out in accordance with relevant guidelines and the Strengthening The Reporting of Observational Studies in Epidemiology (STROBE) guideline was used to report the study.

Consent for publication

Not applicable.

Competing interests

The authors declare that they have no competing interests.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Høegh-Larsen, A.M., Gonzalez, M.T., Reierson, I.Å. et al. Nursing students’ clinical judgment skills in simulation and clinical placement: a comparison of student self-assessment and evaluator assessment. BMC Nurs 22, 64 (2023). https://doi.org/10.1186/s12912-023-01220-0

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s12912-023-01220-0

Keywords