Background
Although the growth of point-of-care ultrasound has resulted in a proliferation of teaching programs, few data exist on the maintenance of proficiency. The aim of this study was to evaluate the retention of cardiac ultrasound skills and training in physicians up to 7 years after a formal focused curriculum in residency.
Methods
Thirty internal medicine physicians, deemed proficient at graduation and having passed a practical examination that confirms imaging skills and knowledge base when a score of ≥80% correct is attained, were retested. Twenty graduates (the NOPREP group) did not study any relevant material, and 10 graduates (the PREP group) were encouraged to study online videos. Scores were categorized by length of time (1-7 years) from graduates’ last performance of ultrasound.
Results
The mean original test score of the physicians was 90 ± 6%. With retesting NOPREP ( n = 20), seven physicians were within 1 year of their last use, and five (71%) repassed the examination, scoring 80 ± 15%. Among the remaining 13 NOPREP physicians who had >1 year of nonuse, none repassed, scoring 58 ± 17%. In retesting PREP ( n = 10), one physician was within 1 year of last use and repassed, scoring 95%. Among the remaining nine PREP physicians with >1 year since last use, three (33%) repassed ( P = .05), scoring 72 ± 21%. Diagnostic knowledge was significantly higher when good-quality images were obtained.
Conclusions
Learned skills in cardiac ultrasound diminish notably within 2 years of nonuse.
The advent of handheld ultrasound equipment has increased the application of ultrasound as a bedside, point-of-care examination technique. The subsequent rise of “nontraditional” users, pioneering trauma surgeons, and emergency and critical care medicine practitioners, has resulted in a proliferation of diverse training courses and curricula in medical school and postgraduate programs. Because the performance of ultrasound is a physical skill, the attainment and maintenance of proficiency in limited ultrasound examination requires not only a retained knowledge base but also repetition and practice. The determinants of continued competency after training could affect successful practice and hospital privileging in this burgeoning technique. Perhaps similar to the practice of cardiopulmonary resuscitation in advanced cardiac life support, a technique requiring both physical skill and a specific knowledge base, a certification and recertification methodology may be needed for point-of-care ultrasound.
Over the past decade, we have taught a cardiovascular limited ultrasound examination (CLUE) in a formal curriculum to residents in internal medicine at our institution. Unlike other centers that may enroll motivated trainees in elective ultrasound courses, our internal medicine residency mandated participation in the curriculum by all residents and included a final clinical test exercise with documentation of a score. However, many graduates who were considered proficient at graduation had been unable to continue ultrasound imaging in their subsequent practice, largely because of a lack of resources early in their careers. Because few data exist on the retention of ultrasound skills over a period of nonuse, we took this opportunity to observe the retained competency in cardiac ultrasound of previously trained internists as a function of time since their last use of ultrasound.
Methods
Study Subjects
The ultrasound training program and its integration into an accredited internal medicine residency at Scripps Mercy Hospital in San Diego, California, has been in existence for >10 years and has been described in full detail elsewhere. Since 2005, all residents who were accepted in the internal medicine program (approximately 10 residents each year) participated in the 3-year ultrasound curriculum. No resident had previous experience in ultrasound training before residency.
Ultrasound Training and Testing
The CLUE training program consisted of teaching seven core views on two-dimensional ultrasound with nine simplified, evidence-based signs of disease ( Table 1 ) through monthly lectures, Web-based learning, and clinical rotations, including two cardiology consultation and 6 intensive care unit months. Each resident typically received a minimum of 50 hours of didactic or bedside instruction and performed approximately 60 CLUE examinations.
Image acquisition and quality (score: 0-4) | Image interpretation (score: 0-2) | Knowledge of diagnostic criteria (score: 0-2) | Overall retention |
---|---|---|---|
Parasternal LV long-axis view |
|
| 90% |
Longitudinal parasagittal view of the lung apices Longitudinal coronal view of the lung bases |
|
| 69% |
Subcostal four-chamber view of the heart | Identify RV, LV, and atria | RV must be smaller than the LV; know the potential space for pericardial effusion | 56% |
Subcostal longitudinal view of the IVC | Identify IVC and neighboring descending aorta | IVC diameter as large as aorta and respiratory variation < 50% is consistent with elevated CVP Differentiate IVC from descending aorta | 80% |
Longitudinal view of the infrarenal aorta | Identify borders of infrarenal aorta | Aneurysm present when vessel has 1.5× focal dilation compared with neighboring segment | 52% |
Transverse and longitudinal view of the carotid artery | Identify carotid bulb and bifurcation and presence of plaque | Define plaque by the presence of calcific shadowing or focal wall thickening Understand implications for CHD risk | 93% |
Just before graduation from residency, all residents were required to take a skill and knowledge assessment clinical exercise on CLUE, named the CLUE-CEX. The CLUE-CEX testing procedure, modeled after the “mini-CEX” procedure endorsed by the American Board of Internal Medicine to assess resident performance, required the resident to image a patient in front of a faculty referee and to verbally answer standardized questions regarding each of the seven core views obtained on the patient. The overall CLUE-CEX score, expressed as a percentage of a perfect score, contained separate assessments of (1) image acquisition and quality, scored 0 to 4 points (0 = no image; 1 = gross, off-axis errors; 2 = suboptimal or partial image; 3 = adequate for diagnosis; 4 = optimized image) per view and accounting for 44% of total points, (2) knowledge of diagnostic criteria, scored 0 to 2 points (0 = absent, 1 = partial, 2 = complete) per view and accounting for 28% of total points, and (3) image interpretation, scored 0 to 2 points (0 = inaccurate, 1 = partial credit, 2 = accurate) per view and accounting for 28% of total points. In cases in which the patient had a difficult window, the resident’s acquisition score was normalized to that of the faculty referee’s maximal score attained on that view. Determined over the initial 3 years of the program through faculty observation of resident practice and expert opinion, an 80% correct test score was felt to represent competency in the core concepts of the CLUE curriculum and had parity to other thresholds in resident skill assessments. Therefore, the 80% threshold has been considered a passing score for the CLUE-CEX since 2008. For convenience, the residents performed their tests using either a standard echocardiograph or a portable, laptop-based or hand-carried device that was available at the time, often determined by the location of the patient they were tested upon (e.g., outpatient clinic, echocardiography laboratory, inpatient wards, intensive care unit).
Study Design
Since the inception of CLUE-CEX testing in 2008, all residents ( n = 74) have participated, with 82% successfully “passing” at graduation. Of the 61 residents who had completed the ultrasound training program and passed the CLUE-CEX, and were therefore considered proficient, 32 remained in local practice and were contacted in random order, of whom 30 agreed to take a repeat CLUE-CEX, many on the contingency that it could be performed at their offices or homes. One graduate was on maternity leave, and one could not be located. To attest to the overall competency of the 30 participating physicians, 100% of these graduates had passed upon their initial sitting for internal medicine board certification from the American Board of Internal Medicine.
The first 20 graduates were specifically told not to review any material in preparation and were designated as the NOPREP group. The subsequent 10 physicians, designated as the PREP group, were encouraged to review summary CLUE videos online, which included brief “how to” summary lessons on proper image acquisition, interpretation, and diagnostic criteria. Physicians were advised not to communicate with colleagues about the test and were reassured that their performance scores were anonymous. No incentive was given to any physician in order to motivate good performance.
Physician graduates were privately retested using the same CLUE-CEX format with either a standard echocardiographic machine (iE33; Philips Healthcare, Andover, MA) at the hospital echocardiography laboratory ( n = 20) or, if in their offices or homes ( n = 10), with a portable pocket-sized ultrasound device (Vscan Duoprobe; GE Healthcare, Wauwatosa, WI), on one of two normal volunteers with good acoustic windows. No data were available on the type of equipment the residents were originally tested on. The repeat CLUE-CEX assessments were made by the same faculty member who had performed the original CLUE-CEX, who was blinded to the graduates’ original scores. Physicians were asked to state the last time they performed cardiac ultrasound since graduation and state their current specialties accounting for the majority of their practice. The faculty referee was also allowed to write notes on the graduates’ performance for later review.
Scores from the CLUE-CEX were subcategorized as skill-based, through image acquisition quality subsection scores, and as knowledge-based, through subsection scores of diagnostic criteria, and further stratified by the six separate CLUE sections. The interpretation subsection scores were not separately analyzed, because this category was subject to the bias of imaging a normal volunteer and was somewhat dependent upon the quality of the image acquired by the physician. However, because point-of-care ultrasound requires physician interpretation while imaging, to mitigate the overlap with image acquisition skill scoring, the interpretative skill score had been assessed by having the physician identify the supposed anatomy on whatever image he or she obtained, regardless of its quality. The same original CLUE-CEX data reporting form was used for retesting and has been published.
Statistical Analysis
Overall score was calculated as percentage correct by adding image acquisition and quality, knowledge, and interpretation points for each view divided by the sum of the maximum points possible for each item, multiplied by 100%. Data are reported as mean percentage correct ± SD. After testing all the graduates, the original CLUE-CEX results for each were compared with the graduate’s current results and plotted against the time given. Retention was defined as the ratio of the recent CLUE-CEX score to the original CLUE-CEX total or subsection score and was expressed as a percentage. A 3 × 3 contingency table of image acquisition scores versus knowledge scores in the NOPREP ( n = 20) group was assessed by χ 2 analysis. The difference in the CLUE-CEX performance between the NOPREP and PREP groups was compared using Fisher’s exact test for categorical data and Student’s t test for continuous data. A P value ≤ .05 was considered to indicate statistical significance. Statistical analyses were performed using SAS version 9.2 (SAS Institute, Cary, NC) by a dedicated statistician.