Keywords
user-centered design - burnout - quality improvement - efficiency
Background and Significance
Background and Significance
For many years, electronic health record (EHR) use was infrequent in the United States.
However, the 2009 Health Information Technology for Economic and Clinical Health Act
offered major incentives for hospitals and medical offices to adopt EHRs. Adoption
was swift, and by 2019, over 90% of U.S. acute care hospitals were using certified
EHRs with most of the market share dominated by a few major companies,[1] and in the outpatient setting adoption also reached high levels, though with many
more vendors involved. The advantages of EHRs have been documented widely such as
improvements in quality of clinical notes, legibility, access to patient data, and
reduction of errors.[2]
[3]
[4]
[5] On the other hand, concerns regarding physician burnout, errors due to poor usability,
and alert fatigue persist, and the impact of these on clinicians and patient care
has been assessed in many prior studies.[6]
[7]
[8]
Usability can be formally measured. For example, the System Usability Scale (SUS)
can be used to evaluate EHRs and determine users' perspective on the EHR's ease of
use.[9] In one study with 870 physicians evaluating their EHR, the mean (± standard deviation)
SUS score was 45.9 ± 21.9 and fell under the “grade F” or “not acceptable” range for
usability.[10] In an assessment of 27 vendor-reported SUS scores, there were no statistical improvements
in EHR scores from 2014 to 2015, with SUS decreasing for 44% of vendors between the
2 years.[11]
Mass General Brigham (MGB, formerly Partners Healthcare) began its transition from
an internally developed EHR to Epic in 2015. Challenges in transitioning from one
EHR to another have been studied and recommendations to ease those transitions have
been made, but satisfaction and clinician burden remain an issue.[12]
[13] Some have hypothesized that there may be an initial decline in satisfaction during
the implementation period of EHRs but that it then increases over time, although few
empiric evaluations have been performed.[14] Several studies have assessed satisfaction months after the transition up to a couple
of years after, but few have continued long term.[15]
[16] Hanauer et al's study demonstrated the challenges of successful EHR adoption and
monitoring physician satisfaction.[14] The authors conducted a 2-year longitudinal assessment of physician perceptions
after a transition from a homegrown EHR to a vendor EHR at the University of Michigan
Health System. They hypothesized that measures of physician perception would follow
a J-curve pattern, with an initial decrease in satisfaction followed by a gradual
rise back to and potentially above baseline, which typically indicates successful
technology adoption. However, they were unable to discover a J-curve pattern for any
of the measures of physician perception as many of the measures either followed a
U-curve, L-curve, or flatlined. Another study by Krousel-Wood et al found that while
positive clinician perceptions significantly increased for items such as long-term
follow-up for patient communication and satisfaction with system reliability, items
such as overall satisfaction, clinical decision quality, productivity, and monitoring
patients significantly decreased over time (p < 0.05 for each).[12]
Monitoring clinician satisfaction and perceptions of the EHR is one way to engage
clinicians and prioritize improvements to the EHR.[17] The Rhode Island Department of Health found that physicians who agreed with the
sentiment that using an EHR added to their daily frustration also had 2.4 times the
odds of burnout in comparison to the physicians who disagreed.[18] Additionally, of the 1,792 physicians who responded to the survey, 70% of the EHR
users reported health information technology-related stress.[18] A physician's self-perceived efficacy while using an EHR was found to be the factor
most predictive of physician satisfaction and patient impact.[19] A recent cross-sectional survey conducted on the relationship between EHR use and
physician burnout revealed that 62.5% (110/176) of physicians felt that the EHR added
to their daily frustration.[20] A systematic review confirmed that a lack of available time for documentation, increased
inbox or patient call message volume, and clinicians' negative perceptions of EHRs
resulted in higher rates of clinician burnout.[21] Clinicians who have increased dissatisfaction with their EHR systems may also have
lower patient satisfaction.[22]
To address how satisfaction changes over time with a vendor EHR, we administered a
satisfaction survey at a large health care system over a 4-year period. We also identified
areas within the EHR that might be improved to increase clinician satisfaction.
Methods
We conducted a longitudinal assessment of survey data gathered between 2016 and 2019
at an academic medical center health system. The health system began its transition
to a commercial health record (EpicCare, 2010, Madison, Wisconsin, United States)
in 2015 beginning with Brigham and Women's Hospital (BWH) and rolled it out across
the entire system over a two-year period, replacing an internally developed medical
record, which was in place in both the inpatient and outpatient setting. The Institutional
Review Board reviewed the study and designated it as a quality improvement project
not requiring formal clinician consent. Clinicians were able to decline participation
in the survey or stop taking it at any time.
Sample Recruitment and Survey Distribution
Clinicians were eligible to receive a survey if they were an active user of the EHR
and a physician (MD or DO), nurse practitioner, or physician assistant in either the
inpatient or outpatient setting. All levels of physicians (residents, attendings,
interns) were invited to participate. A survey at one institution only, BWH, was administered
in 2016. The following surveys were conducted during the spring 2018 and the spring
2019 and were administered to the entire population of MGB credentialed clinicians.
Eligible clinicians across the health care system were emailed a link to the survey.
Three reminder emails were then sent to nonresponders every other week for 6 weeks.
The 2019 survey was administered using REDCap electronic data capture tool hosted
at BWH.[23]
[24] The 2016 and 2018 surveys were developed in Limesurvey.[25]
The survey administered to BWH physicians was 1.5 years postimplementation of Epic.
By the time the next survey was administered in 2018, all MGB sites had completed
their transition to Epic. Participants did not receive incentive payment for completing
the survey. Survey responses were kept confidential; no identifying information was
reported. They were not anonymous since we captured email address to track participants
to send survey reminders and link responses back to individuals for analysis.
Survey Development
The survey instrument was developed using the Primary Care Information Project survey
and Family Practice Management survey as well as the original survey developed by
the research group and subject matter experts that was used for 6 years to assess
satisfaction with the internally developed electronic record at MGB.[26]
[27] Validated usability surveys were considered during development but were not sufficient
to address the broader user experience. We worked with the EHR leadership and subject
matter experts to customize the survey to address tasks and features important to
the health system, which would allow us to identify specific areas for improvement.
The survey included 48 items split into seven matrix questions, one multiple-choice
rating question and one open-ended question ([Supplementary Appendix A], available in online version). The survey included items assessing ease/difficulty
of completing tasks on a scale (“Very easy [5],” “Easy [4],” “Neutral [3],” “Difficult
[2],” “Very difficult [1],” or “Not applicable”) in the following categories: reviewing
patient data, documentation, patient engagement, task and workflow management, and
preventive care and panel management. In addition, it included items assessing perceptions
of the EHR's value, usability, and impact on workflow and patient care. Participants
rated their agreement or disagreement with these statements by responding “Strongly
agree (5),” “Agree (4),” “Neutral (3),” “Disagree (2),” “Strongly disagree (1),” or
“Not applicable.” One item assessing overall satisfaction on a scale from “Very satisfied”
(6) to “Very dissatisfied” (1) was included in addition to one open-ended question
allowing clinicians to share any additional information about the EHR. [Table 1] identifies the task items included in the survey instrument.
Table 1
Survey items on ease or difficulty of completing tasks using epic, items on agreement
with statements assessing value, perceived usability, and impact on patient care and
workflow and item assessing overall satisfaction
Group
|
Individual items
|
Reviewing patient data
|
Reviewing a patient's clinical information
|
Reviewing any health changes since you last saw the patient
|
Documentation
|
Creating the visit note
|
Documenting allergies
|
Documenting vital signs
|
Documenting immunizations
|
Documenting family history
|
Keeping problem lists updated
|
Keeping medication lists updated
|
Documenting an ICD-10 diagnosis code for billing purposes
|
Documenting a CPT procedure code for billing purposes
|
Patient engagement
|
Providing patients with an electronic previsit form
|
Incorporating patients' requests for changes to their health record
|
Communicating with the patient
|
Task and workflow management
|
Ordering laboratory tests
|
Ordering radiology tests
|
Reviewing laboratory results
|
Reviewing radiology results
|
Identifying when a laboratory order has not been completed
|
Identifying when a radiology order has not been completed
|
Writing prescriptions
|
Renewing prescriptions
|
Monitoring medication safety at the point of prescribing (e.g., drug–allergy, drug–drug
interactions)
|
Monitoring patient medication adherence
|
Communicating referral information to specialists
|
Reviewing referral information from specialists
|
Identifying when a referral has not been completed
|
Communicating with other clinicians and office staff
|
Preventive care and panel management
|
Identifying preventive care services (e.g., cancer screening, immunizations) that
are due for the patient
|
Ordering appropriate preventive care services during the visit
|
Making a list of patients based on clinical information (e.g., problem, medication)
|
Contacting patients to remind them of care for which they are due
|
Assisting patients in self-management activities (e.g., goal setting, patient education)
|
Value, perceived usability, and impact on patient care and workflow
|
The EHR provides me with all the information I need to take care of the patient
|
I can find information I need easily with this EHR
|
The information in the EHR is presented in a useful format
|
There are too many alerts and reminders in the EHR
|
The EHR allows me to complete tasks efficiently
|
The EHR disrupts the way I normally like to do my work
|
The EHR supports team-based care workflows
|
The EHR addresses the needs of my specific specialty
|
The EHR is easy to use
|
The EHR improves the quality of patient care
|
The EHR helps to prevent medical errors
|
The EHR helps me focus on patient care rather than the computer
|
The EHR provides valuable decision support
|
The EHR helps me to provide preventive care
|
Satisfaction
|
Overall, how satisfied are you with the EHR
|
Abbreviations: CPT, Current Procedural Terminology; EHR, electronic health record;
ICD-10, International Classification of Diseases.
Analysis
We used SAS version 9.4 for the statistical analysis (Cary, North Carolina, United
States). Frequencies and means (95% confidence intervals [CIs]) were calculated for
each survey question for all years. A comparison of means was done between 2016, 2018,
and 2019 BWH responses and for all systemwide responses between 2018 and 2019. A multivariable
generalized linear model was performed to predict the outcome of overall satisfaction
on a numeric scale. We controlled for clustering by clinician since we had some clinicians
who responded to surveys in multiple years; we needed to account for this in looking
at the influence of the participant characteristics. To account for missing values,
multiple imputation was also performed. All covariate variables were entered into
the overall satisfaction model followed by a stepwise selection technique to achieve
the final overall satisfaction model.
A content analysis was conducted on the open-ended comments. Each comment was assigned
a code. Similar codes were grouped into categories by subject matter. Categories with
the most frequent feedback are reported in this study. Representative quotes were
identified for each category of feedback.
Results
The response rate systemwide was similar for 2018 and 2019 at 16 and 14%, respectively,
and 19% for BWH surveyed in 2016. Of the responders, most were primarily affiliated
with BWH and BWH–Faulkner Hospitals (2018: 39%, 2019: 33%) or Mass General Hospital
(2018: 43%, 2019: 40%). There were slightly more female than male responders in both
years, which was representative of the overall clinician population ([Table 2]).
Table 2
Demographics of respondents
|
2018 Survey (N = 1,613)
(n, %)
|
2019 Survey (N = 1,632)
(n, %)
|
Primary affiliation
|
Massachusetts General Hospital
|
688 (43)
|
646 (40)
|
Brigham and Women's and Faulkner Hospitals
|
634 (39)
|
538 (33)
|
Newton Wellesley Hospital
|
100 (6)
|
119 (7)
|
Northshore Medical Center
|
85 (5)
|
65 (4)
|
McLean Hospital
|
20 (1)
|
67 (4)
|
Other community[a]
|
86 (5)
|
197 (12)
|
Clinician type
|
Specialists
|
944 (59)
|
970 (59)
|
Primary care clinicians
|
270 (17)
|
286 (18)
|
Nurse practitioners
|
193 (12)
|
169 (10)
|
Residents
|
132 (8)
|
124 (8)
|
Physician assistants
|
71 (4)
|
79 (5)
|
Other type
|
3 (<1)
|
4 (<1)
|
Gender
|
Male
|
764 (47)
|
791 (48)
|
Female
|
849 (53)
|
841 (52)
|
How long have you been using Epic?
|
0–3 mo
|
8 (<1)
|
11 (1)
|
4–6 mo
|
56 (3)
|
30 (2)
|
7–12 mo
|
174 (11)
|
54 (3)
|
1–3 y
|
1,186 (74)
|
648 (40)
|
> 3 y
|
187 (12)
|
871 (53)
|
Missing
|
2 (<1)
|
18 (1)
|
a Includes other community hospitals affiliated with Mass General Brigham.
Overall Satisfaction
The responses from the 3 years of surveys for BWH showed a slight increase in mean
overall satisfaction over time from 2016 (2.85, 95% CI: 2.71, 2.99) to 2019 (3.21,
95% CI: 3.07, 3.34; p < 0.001). Systemwide, however, there was no significant difference (p = 0.2030) in mean response for overall satisfaction between responders of the 2018
survey (3.14, 95% CI: 3.06, 3.21) and those of the 2019 survey (3.19, 95% CI: 3.12,
3.27). Overall, 46% of responders in 2018 expressed some level of satisfaction (somewhat
satisfied, satisfied, very satisfied) and the 2019 survey showed similar levels of
satisfaction (47%). In 2018 and 2019, 19% of clinicians responded that they were very
dissatisfied with the EHR ([Table 3]). [Fig. 1] shows the mean overall satisfaction over 3 years for BWH-only clinicians and over
2 years for all systemwide clinicians.
Fig. 1 BWH versus systemwide mean overall satisfaction over time. BWH, Brigham and Women's
Hospital.
Table 3
Overall satisfaction for all respondents
|
2018 (N = 1,613)
(n, %)
|
2019 (N = 1,632)
(n, %)
|
1 = Very dissatisfied
|
311 (19)
|
308 (19)
|
2 = Dissatisfied
|
293 (18)
|
317 (19)
|
3 = Somewhat dissatisfied
|
252 (16)
|
234 (14)
|
4 = Somewhat satisfied
|
386 (24)
|
364 (22)
|
5 = Satisfied
|
296 (18)
|
337 (21)
|
6 = Very satisfied
|
52 (3)
|
69 (4)
|
Missing
|
23 (1)
|
3 (<1)
|
Satisfaction by Task
[Table 4] includes results of all survey items for clinicians systemwide in 2018 to 2019.
In both years, responders rated 15 out of 33 (45%) tasks as difficult or very difficult
(average score below 3.0).
Table 4
Average response for each task across all institutions
Group
|
Survey item
|
2018
|
2019
|
p-Value
|
Mean (95% CI)
|
Mean (95% CI)
|
Task ease/difficulty
|
Reviewing patient data
|
Reviewing a patient's clinical information
|
3.08 (3.03, 3.14)
|
3.11 (3.05, 3.16)
|
0.46
|
Reviewing any health changes since you last saw the patient
|
2.67 (2.61, 2.72)
|
2.67 (2.62, 2.72)
|
0.98
|
Documentation
|
Creating the visit note
|
3.54 (3.49, 3.60)
|
3.59 (3.53, 3.64)
|
0.19
|
Documenting allergies
|
3.55 (3.50, 3.60)
|
3.53 (3.48, 3.58)
|
0.61
|
Documenting vital signs
|
3.44 (3.38, 3.50)
|
3.40 (3.34, 3.47)
|
0.35
|
Documenting immunizations
|
2.78 (2.72, 2.88)
|
2.82 (2.75, 2.90)
|
0.61
|
Documenting family history
|
3.00 (2.93, 3.05)
|
3.04 (2.98, 3.10)
|
0.21
|
Keeping problem lists updated
|
2.78 (2.71, 2.84)
|
2.69 (2.63, 2.75)
|
0.03
|
Keeping medication lists updated
|
2.64 (2.58, 2.70)
|
2.61 (2.55, 2.67)
|
0.45
|
Documenting an ICD-10 diagnosis code for billing purposes
|
3.08 (3.02, 3.15)
|
3.10 (3.04, 3.17)
|
0.68
|
Documenting a CPT procedure code for billing purposes
|
3.00 (2.93, 3.08)
|
3.00 (2.94, 3.07)
|
0.996
|
Patient engagement
|
Providing patients with an electronic previsit form
|
2.50 (2.41, 2.60)
|
2.52 (2.44, 2.61)
|
0.70
|
Incorporating patient's requests for changes to their health record
|
2.45 (2.38, 2.52)
|
2.43 (2.37, 2.49)
|
0.70
|
Communicating with the patient
|
3.05 (2.99, 3.11)
|
3.13 (3.08, 3.19)
|
0.03
|
Task and workflow management
|
Ordering laboratory tests
|
3.33 (3.27, 3.39)
|
3.37 (3.31, 3.42)
|
0.29
|
Ordering radiology tests
|
3.08 (3.02, 3.14)
|
3.10 (3.04, 3.16)
|
0.56
|
Reviewing laboratory results
|
3.55 (3.50, 3.60)
|
3.63 (3.58, 3.68)
|
0.01
|
Reviewing radiology results
|
3.68 (3.63, 3.73)
|
3.70 (3.65, 3.75)
|
0.63
|
Identifying when a laboratory order has not been completed
|
2.59 (2.53, 2.65)
|
2.61 (2.56, 2.67)
|
0.50
|
Identifying when a radiology order has not been completed
|
2.65 (2.60, 2.71)
|
2.65 (2.59, 2.71)
|
0.92
|
Writing prescriptions
|
3.26 (3.20, 3.32)
|
3.29 (3.23, 3.34)
|
0.46
|
Renewing prescriptions
|
3.30 (3.24, 3.36)
|
3.37 (3.31, 3.42)
|
0.08
|
Monitoring medication safety at the point of prescribing (e.g., drug–allergy, drug–drug
interactions)
|
3.09 (3.04, 3.15)
|
3.16 (3.10, 3.21)
|
0.07
|
Monitoring patient medication adherence
|
1.99 (1.94, 2.05)
|
2.06 (2.01, 2.12)
|
0.05
|
Communicating referral information to specialists
|
2.87 (2.81, 2.93)
|
2.90 (2.85, 2.96)
|
0.38
|
Reviewing referral information from specialists
|
2.91 (2.84, 2.97)
|
2.92 (2.86, 2.98)
|
0.69
|
Identifying when a referral has not been completed
|
2.02 (1.97, 2.08)
|
2.30 (2.25, 2.36)
|
<0.0001
|
Communicating with other clinicians and office staff
|
3.03 (2.97, 3.09)
|
3.16 (3.11, 3.22)
|
0.0002
|
Preventive care and panel management
|
Identifying preventive care services (e.g., cancer screening, immunizations) that
are due for the patient
|
2.85 (2.78, 2.92)
|
3.02 (2.95, 3.09)
|
0.0001
|
Ordering appropriate preventive care services during the visit
|
3.01 (2.93, 3.08)
|
3.10 (3.03, 3.17)
|
0.03
|
Making a list of patients based on clinical information (e.g., problem, medication)
|
2.12 (2.05, 2.19)
|
2.31 (2.24, 2.37)
|
<0.0001
|
Contacting patients to remind them of care for which they are due
|
2.45 (2.37, 2.52)
|
2.49 (2.42, 2.56)
|
0.31
|
Assisting patients in self-management activities (e.g., goal setting, patient education)
|
2.42 (2.34, 2.49)
|
2.44 (2.37, 2.50)
|
0.63
|
Agreement/disagreement
|
Value, perceived usability, and impact on patient care and workflow
|
The EHR provides me with all the information I need to take care of the patient
|
3.04 (2.98, 3.09)
|
3.07 (3.02, 3.13)
|
0.29
|
I can find information I need easily with this EHR
|
2.51 (2.45, 2.57)
|
2.60 (2.55, 2.66)
|
0.008
|
The information in the EHR is presented in a useful format
|
2.45 (2.39, 2.51)
|
2.52 (2.46, 2.58)
|
0.05
|
There are too many alerts and reminders in the EHR
|
3.63 (3.58, 3.68)
|
3.67 (3.61, 3.72)
|
0.31
|
The EHR allows me to complete tasks efficiently
|
2.40 (2.34, 2.45)
|
2.43 (2.37, 2.49)
|
0.31
|
The EHR disrupts the way I normally like to do my work
|
3.33 (3.27, 3.40)
|
3.33 (3.26, 3.39)
|
0.84
|
The EHR supports team- based care workflows
|
2.72 (2.67, 2.78)
|
2.75 (2.69, 2.80)
|
0.52
|
The EHR addresses the needs of my specific specialty
|
2.48 (2.42, 2.54)
|
2.47 (2.41, 2.53)
|
0.84
|
The EHR is easy to use
|
2.44 (2.38, 2.50)
|
2.47 (2.41, 2.53)
|
0.36
|
The EHR improves the quality of patient care
|
2.63 (2.57, 2.69)
|
2.80 (2.74, 2.86)
|
<0.0001
|
The EHR helps to prevent medical errors
|
2.90 (2.84, 2.96)
|
3.01 (2.96, 3.07)
|
<0.001
|
The EHR helps me focus on patient care rather than the computer
|
1.66 (1.62, 1.71)
|
1.70 (1.65, 1.75)
|
0.22
|
The EHR provides valuable decision support
|
2.53 (2.47, 2.58)
|
2.60 (2.55, 2.66)
|
0.02
|
The EHR helps me to provide preventive care
|
2.69 (2.62, 2.76)
|
2.80 (2.73, 2.87)
|
0.01
|
Abbreviations: CI, confidence interval; EHR, electronic health record.
Note: Participants had the option of choosing “Very easy (5),” “Easy (4),” “Neutral
(3),” “Difficult (2),” “Very difficult (1),” or “Not applicable.” Participants rated
their agreement or disagreement with the value, perceived usability, and impact on
patient care and workflow statements by responding “Strongly agree (5),” “Agree (4),”
“Neutral (3),” “Disagree (2),” “Strongly disagree (1),” or “Not applicable.”
There was no significant difference in the mean response between 2018 and 2019 for
the two items related to Reviewing Patient Data. “Reviewing any health changes since
you last saw the patient” was rated as difficult with a mean of 2.67 in both years.
When asked to rate the ease or difficulty of tasks related to documentation from very
easy (5) to very difficult (1), clinicians reported a more difficult time keeping
problems lists updated in 2019 (2.69) versus 2018 (2.78, p = 0.03). Documentation tasks with a mean rating above a 3.5 in both years included
“creating the visit note” and “documenting allergies.”
For tasks related to patient engagement the results indicate they had a significantly
easier time communicating with patients in 2019 (3.13) than 2018 (3.05, p = 0.03). In both 2018 and 2019, “Providing patients with an electronic previsit form”
(2.50, 2.52) and “Incorporating patients' requests for changes to their health record”
were more consistently rated as difficult (2.45, 2.43).
In the area of Task and Workflow Management, there were several significant mean differences:
“reviewing laboratory results” got easier over time (2018: 3.55, 2019: 3.63, p = 0.01); “identifying when a referral has not been completed” was significantly harder
(p < 0.0001) in 2018 (2.02) versus 2019 (2.30); and “communicating with other clinicians
and office staff” got easier (2018: 3.03, 2019: 3.16, p = 0.0002). The highest rated task for ease on average in both 2018 and 2019 was “reviewing
radiology results” (3.68, 3.70). In both 2018 and 2019, the two tasks that were rated
the most difficult were “monitoring patient medication adherence” (1.99, 2.06) and
“identifying when a referral has not been completed” (2.02, 2.30).
In the area of Preventive Care and Panel Management, mean response to “identifying
preventive care services,” “ordering appropriate preventive care services during the
visit,” and “making a list of patients based on clinical information” were significant
(p < 0.05) between 2018 (2.85, 3.01, 2.12) and 2019 (3.02, 3.10, 2.31), showing a positive
increase in ease of completing the task.
Several survey items where participants rated their level of agreement (“Strongly
agree (5),” “Agree (4),” “Neutral (3),” “Disagree (2),” “Strongly disagree (1),” or
“Not applicable”) with statements related to value, perceived usability, and impact
on patient care and workflow showed significant improvement from 2018 to 2019: “I
can find information I need easily with this EHR” (2.51 vs. 2.60, p = 0.008), “the information in the EHR is presented in a useful format” (2.45 vs.
2.52, p = 0.05), “the EHR improves quality of patient care” (2.63 vs. 2.80, p < 0.0001), “the EHR helps prevent medical errors” (2.90 vs. 3.01, p < 0.001), “the EHR provides valuable decision support” (2.53 vs. 2.60, p = 0.02), and “the EHR helps provide preventive care” (2.69 vs. 2.80, p = 0.01). The strongest disagreement in both years was with the statement “the EHR
helps me focus on patient care rather than the computer” (1.66, 1.70). The strongest
agreement (above 3.5) was with the statement, “there are too many alerts and reminders
in the EHR.”
Multivariable Model
The first iteration of the multivariable model included only 49% of the sample due
to missingness across the many covariates. Consequently, we imputed the missing covariates
to better complete the dataset for the multivariable model. The results of this imputed
model adjusted for clustering by clinician ([Supplementary Appendix B], available in online version) showed that there was no difference in overall satisfaction
between 2018 and 2019 for clinicians systemwide, although other variables were significant
predictors. Some significant positive predictors of overall satisfaction included
the survey question “The EHR is easy to use” (a one-point increase in agreement that
the EHR is easy to use increased the mean overall satisfaction score by 8.12%, p < 0.0001). Other items that were positive predictors of an increase in overall satisfaction
included “The EHR improves the quality of patient care” (6.6% increase, p < 0.0001) and “The information in the EHR is presented in a useful format” (4.98%
increase, <0.0001). Variables that decreased overall satisfaction included a higher
level of agreement with “The EHR disrupts the way I normally like to do my work” (1.88%
decrease, p < 0.0001) and “There are too many alerts and reminders in the EHR” (1.44% decrease,
p < 0.001).
Clinician Comments
The most frequent comments systemwide from the open-ended question in the 2019 survey
could be grouped into the following categories: overall usability, flexibility of
use and customization, implementation-specific user experience, training and support,
in-basket, medications, and problem list/problem-based charting. [Table 5] includes representative quotes from responders in each category.
Table 5
Categories of most frequent comments from open-ended survey question with example
quote in each category
Feedback
|
Quote
|
Overall perceived usability
|
“There are multiple ways to perform nearly every task and I have to think about how
to do the simplest task each time—it takes away from my caring for the patient.”
|
Flexibility of use and customization
|
“Inpatient consult list needs to be individualized for cardiac surgery (needs more
prompts). Needs to just populate one list (not new consults and ongoing consults because
the NP is first to see patient.”
|
Implementation-specific user experience
|
“I have used and liked EPIC at both of my previous employers, but it was more customizable
to specific needs of staff and visually easier/cleared/less cluttered to read. The
EPIC system at Partners is a nightmare. It is a one size fits all, cumbersome, visually
distracting EHR.”
|
Training and support
|
“The training at the time of roll out was reasonable; however, we could all benefit
from far more 1:1 support on an ongoing basis to help improve efficiency. I also think
better collaboration with compliance department to help ensure we are not over documenting
would be helpful.”
|
In-basket
|
“I do not know when I forward vs. reply to a message and add on a new recipient if
the prior message is included since it does not appear below as it does in an email.
The various In-Basket subfolders are not named intuitively.”
|
Medications
|
“I frequently get requests to renew medications from surescripts in my inbox, but
then when I go to sign them EPIC has not identified that medication as one that's
already in the patient's list of meds and paired them. So, I get the option of an
alert telling me I'm writing a duplicate order (which I'm not!) or I get to discontinue
both orders, or I get to reply to the surescripts prescription with a reason why I'm
declining to fill it, and then renew the one in the patient's record.”
|
Problem list/problem-based charting
|
“Problem-based charting interface is the single biggest problem for me in EPIC. Adding
a problem (limited only to the choices EPIC provides, which encouraged splitting not
consolidating problems), entering the overview (additional click), then assessment
and plan (additional click) and putting the problem list in the note as a smart phrase
(additional click to refresh) and there is no easy way to order the problems or group
them together besides by priority, which requires an additional click for each individual
problem.”
|
Abbreviation: EHR, electronic health record.
Comments regarding overall usability included general statements that the EHR is cumbersome,
the user interface is not pleasing, and completing tasks takes too long and requires
too many clicks. Flexibility of use and customization included comments regarding
a lack of features available to support clinicians in their specific specialty and
challenges in easily accessing frequently used values and actions. Providers commented
on Implementation-specific experiences such as challenges accessing historical data,
printing issues, and a lack of standardization of communication and documentation
methods among providers. Comments regarding training and support included a desire
for more training, additional support, and quicker resolutions to support tickets.
Three specific areas of the EHR were commented on more than others: In-basket, medications,
and problem list and problem-based charting. Providers had challenges with organization,
redundancies, and search of in-basket. In addition, many comments centered around
medications, specifically errors and challenges organizing the medication list, issues
accessing historical prescription information, and irrelevant medication alerts. Finally,
many providers commented on the problem list and problem-based charting, citing issues
with the interface and usability of problem-based charting and challenges using the
EHR to maintain and update problem lists.
Discussion
In this multiyear survey of clinicians at a large health care system, satisfaction
with the EHR has remained low. Less than half of the clinicians in each of the 3 years
responded as very satisfied, satisfied, or somewhat satisfied with the EHR overall.
The ease/difficulty of completing most tasks did not change significantly between
2016 and 2019. Results of the multivariable analysis indicated that overall satisfaction
increased the most as agreement with “The EHR is easy to use” increased. This is the
case despite regular vendor updates.
After many years with a homegrown EHR, the transition to Epic was an emotionally charged
change for many clinicians. The previous internally developed EHR had been in place
many years and was highly functioning and customizable with significant direction
and feedback from physician users driving development and design. Many clinicians
were involved in the decision to choose a single-vendor system to replace the existing
homegrown medical record and the multiple ancillary systems, though it was a fraction
of all clinicians in the integrated delivery system. Shifting to a commercial record
was challenging for the clinicians in many ways, perhaps especially in that usability
was perceived as lower, despite upgrades postimplementation. In their review, Huang
et al identified several challenges in transitioning from one EHR to another, including
financial considerations, clinician expectations, and patient safety considerations.[13] Recognizing that change is difficult, they suggest attempting to “manage expectations
and provide additional training.” In response to dissatisfaction with the EHR, institutions
across the health system launched efforts to address EHR burden and usability. In
addition to addressing usability issues through enhancements to the vendor EHR, 1-hour
one-on-one training sessions through a third party addressed clinician challenges
with documentation, clinical review, orders, and in-basket management primarily, which
were well received but only utilized by 33% of providers at BWH. Still, some tasks
are difficult to accomplish, and designing applications with good usability is a better
strategy overall than offering more training. Other strategies that were used in the
network included implementing physician advisory committees to provide feedback on
EHR issues, and also development of scribe programs in some specialties and voice
recognition to improve clinicians' experiences with the EHR. Future work would include
assessing the impact of additional training and other programs on physician satisfaction.
The few years after the transition showed little improvement in clinicians' overall
satisfaction with the EHR. While a difficult period postimplementation is the rule,
overall satisfaction remained low, and several tasks were rated somewhat difficult.
Anecdotally, some treatments that come up relatively infrequently such as ordering
outpatient transfusions are especially hard and finding how things like this can be
done can be challenging. Other tasks such as renewing a prescription when covering
for another physician can be difficult at times. Such issues with usability of EHRs
is a source of dissatisfaction among clinicians and, in comparison to other commonly
used applications, they rate significantly lower on usability.[10]
Several efforts have been reported in the literature describing recommendations to
address some of the pain points identified in this survey. Many have sought to address
the challenges of monitoring medication interactions, medication reconciliation, and
medication ordering with EHRs.[28]
[29]
[30]
[31] Some studies have suggested using different approaches to improve usability. For
example, one study highlighted the potential of a redesign focusing on indications-based
prescribing to ease this burden, which rated high on usability.[31] Interventions designed to address usability issues related to allergy documentation
have shown potential to improve clinician satisfaction as well.[32] Clinical decision support systems attempted to address the design, usability and
alert fatigue associated with some aspects of the EHR.[30]
[32]
[33]
[34] One institution was successful in reducing alerts and associated clicks through
an initiative to reduce burnout by optimizing clinical decision support.[35] Other initiatives focused on training and evaluation have shown promise.[36] However, the changes introduced by the vendor during this interval have been more
incremental to date.
The Office of the National Coordinator has required that vendors follow a user-centered
design process to be certified. The certification requirement for usability of an
EHR vendor does not include the site-specific EHR configuration, which has created
a usability reality gap.[37] There are challenges in improving the usability of commercial systems because options
are limited for the institution by the structure of the EHR, and customizations and
implementation configurations can have an impact on the usability of the system.[38] Regardless, there still appears to be several usability concerns that need to be
addressed by the vendors in partnership with their users.[39] It is clear that we need to continue to improve EHRs so they function better for
clinicians,[40] and doing this is likely to require more than minor incremental improvements. Other
analytical approaches such as AB testing, where you test two design options to identify
which one results in better performance, should also be leveraged—this is routine
in other industries.[41]
[42]
[43]
Limitations of this study included the low response rates, though we did hear from
a substantial number of clinicians who expressed consistent concerns. The response
rate was likely low because many providers were not practicing clinically, but we
were not able to exclude them. The EHR vendor was also conducting surveys during these
years (though these were not released to frontline users), which may have led to survey
fatigue. This was the experience of one health system and one EHR implementation,
though other studies have reported similar results. Also, while many of our results
were statistically significant, they are not necessarily clinically significant. We
did not examine other factors contributing to satisfaction such as training, organizational
culture, or frequency of use and prior experience with the EHR so future research
should focus on the broader user experience and sociotechnical factors. We plan to
continue the survey effort with a focus on shorter surveys to understand how specific
changes made to the EHR impact clinician satisfaction.
Conclusion
In a series of surveys assessing clinician satisfaction with a commercial EHR, we
found a consistently moderate level of satisfaction, as well as several tasks that
clinicians rated as somewhat difficult. Improvements in vendor EHR usability and site
implementation should continue to be a focus area to reduce clinician burden and improve
satisfaction and should include more substantial changes than they have to date.
Clinical Relevance Statement
Clinical Relevance Statement
Capturing clinician experience with the EHR over time can help assess the success
or failure of changes made to address usability and clinician burden. This is important
for increasing clinician effectiveness, efficiency, and safety in practice, potentially
leading to better patient care and more satisfied clinicians.
Multiple-Choice Questions
Multiple-Choice Questions
-
What variable was the most significant positive predictor of overall satisfaction
with the EHR?
-
Agreement with the statement “The EHR disrupts the way I normally like to do my work.”
-
Agreement with the statement “The EHR is easy to use.”
-
Agreement with the statement “The EHR provides valuable decision support.”
-
Ease of “Creating the visit note.”
The correct answer is option b. Clinician agreement with the statement “The EHR is
easy to use” was the most significant positive predictor of overall satisfaction with
the EHR. A one-point increase in agreement that the EHR is easy to use increased the
mean overall satisfaction score by 8.1%.
-
How did clinician overall satisfaction change over time for responders of BWH compared
with the change for clinicians systemwide?
-
Satisfaction remained the same for BWH clinicians and clinicians systemwide.
-
Satisfaction remained the same for BWH clinicians and increased significantly for
providers systemwide.
-
Satisfaction increased slightly for BWH clinicians and did not increase significantly
for providers systemwide.
-
Satisfaction decreased slightly for BWH clinicians and increased slightly for clinicians
systemwide.
The correct answer is option c. Satisfaction from 2016 to 2019 for BWH responders
showed a slight but significant increase from 2.85 to 3.21 (p = 0.0002). For clinicians systemwide, there was no significant change in satisfaction.