Keywords
clinical decision support systems - data aggregation - neurology - organizational
efficiency - referral and consultation - inpatients
Background and Significance
Background and Significance
Since its introduction in 2009, the Health Information Technology for Economic and
Clinical Health (HITECH) Act has been associated with widespread adoption of electronic
health record (EHR) systems in hospitals across the United States.[1]
[2] Many physicians have reported generally positive experiences with EHRs,[3] but despite the documented benefits of such systems,[4] EHR systems are also associated with unintended increases in physician workload
and documentation times,[5]
[6]
[7]
[8] hospital inefficiencies,[9] and decreased time spent delivering direct patient care.[10]
[11]
[12]
[13] EHR-related increases in physicians' documentation and billing workload are key
contributors to physician dissatisfaction in the field of neurology, where physician
burnout is high relative to other specialties,[14]
[15]
[16] thereby potentially leading to compromises in patient-care quality.[17]
[18]
As a specialty, neurology entails high-EHR utilization due to several factors. First,
neurologists rely on unique, multimodal EHR data, including neuroimaging, video feeds,
and electrophysiology waveforms to support diagnostic and treatment decisions.[19] Neurology additionally entails a significant amount of consultative work,[20] which may involve review and interpretation of multiple-EHR data streams, and finally,
neurological trainees are instructed to incorporate meticulous history-taking, physical
examinations, and assessments into their patient evaluations[14] which may further drive extensive use of EHRs and their data.
Dashboards that aggregate clinical data from multiple sources and present information
within a single, centralized visualization platform may potentially address the task
of extensive EHR data review from multiple disparate sources in neurology. However,
clinical decision support requires considerable summarization of EHR data[21] from fragmented locations. By constituting forms of clinical decision support,[22]
[23] such dashboards may also extend beyond neurology in their uses which include tools
for improving quality[24]
[25] or reducing cognitive burden.[26]
[27]
Given that aggregative dashboards can potentially reduce the workload associated with
data review in neurological evaluations, we sought to investigate whether such a dashboard
might be associated with reduction in time spent performing neurological consultation.
We hypothesized that, among neurological trainees being taught to perform meticulous
evaluations, inpatient consultation turnaround time (TAT) would be shorter after dashboard
implementation than before.
Methods
Design
We conducted a retrospective analysis of consultation TAT before and after the implementation
of a web-based, neurological data dashboard at Columbia University Irving Medical
Center (CUIMC), a large academic medical center in New York that is home to a neurology
residency program. In conjunction with New York-Presbyterian (CUIMC's hospital) and
the Department of Biomedical Informatics at CUIMC has a clinical data repository that
dates back to 1988, and maintains a web-based, display-only clinical data review platform
(i-NewYork Presbyterian [iNYP];
http://inyp.nyp.org/inyp
) that is separate from the transactional EHR used by the hospital and the ancillary
systems.[28] This platform is accessible via single-factor authentication from within the hospital
intranet and through dual-factor authentication from outside the hospital intranet.
Via an interface, the iNYP Clinical Information System (iNYP) ingests data from the
main hospital EHR and admission/discharge/transfer system, as well as laboratory,
radiology, ultrasound, and neurophysiology reporting systems ([Fig. 1]). The iNYP platform also contains several aggregative “dashboard” pages that display
tailored clinical data to particular clinical user groups. Given that many inpatient
consultations at CUIMC are focused on stroke, a vascular neurology-oriented dashboard
was conceived, developed, and implemented as a clinical informatics quality improvement
project between August and December, 2017. The data review platform was chosen because
it could perform dashboarding functions that the institutional EHR could not. The
platform was also chosen because we felt it was likely to have good uptake among trainee
users due to its history as the former institutional EHR system that had been replaced
in favor of the current institutional EHR but had survived in a form that contained
only data review functionality.
Fig. 1 Simplified data flow/architecture diagram between ancillary and EHR systems at CUIMC,
contextualizing the iNYP Clinical Information System that contains the iNYP neurology
dashboard. Red arrows denote flow of result data to interface. Red arrows with two
asterisks denote the flow of medication, order, and vital sign data entered by end-users
from EHR to interface. Green arrows denote data flows back to presentation layer.
ADT, admission/discharge/transfer; CUIMC, Columbia University Irving Medical Center;
CROWN, Clinical Records Online Web Network; EEG, electroencephalogram; EHR, electronic
health record; iNYP, i-NewYork Presbyterian; PACS, picture archive communication system;
SCM, Sunrise Clinical Manager.
Between August and September 2017, one vascular neurologist completing fellowship
training in clinical informatics (BRK) gathered requirements from four attending stroke
neurologists in the department of Neurology at CUIMC to finalize the dashboard's clinical
content, which was divided into 13 rectangular page subsections, or “tiles.” The design
was based on an active iNYP dashboard for general internal medicine, and retained
five tiles from the latter, but the eight remaining tiles were tailored for stroke
consultations based on vascular neurologist domain expert input ([Table 1]; sample screenshots in [Figs. 2] and [3]). The iNYP neurology dashboard was developed between September 2017 and December
2017, and went live on December 1, 2017. In November 2017, one clinical informatics
fellow conducted two separate demonstrations of the dashboard and its functionality
to an audience of CUIMC stroke neurology faculty and all neurology residents, respectively.
One additional demonstration was repeated for the neurology residents in January 2018.
Table 1
Dashboard tile content
Tile
|
Contents[a]
|
Development process
|
Clinical alerts
|
Patient age, sex, race, smoking status; 10-year ASCVD risk calculator and statin therapy;
BMI and BP alerts[b]
|
Copied from different iNYP dashboard
SME recommendation
|
Visit history
|
12-month visit history and locations (outpatient, specialty, faculty practice, inpatient)
|
Copied from different iNYP dashboard
|
Vital signs
|
Most recent vital signs with data source (outpatient/inpatient), current admission
indicator, and no-shows
|
Copied from different iNYP dashboard
|
Neuroimaging[c]
|
PACS neuraxis reports and images (X-ray icon)
|
SME recommendation
|
Neurophysiology/Neurosonology
|
EEG, carotid and transcranial Doppler reports
|
SME recommendation
|
Cardiology
|
TTE, TEE, EKG reports
|
SME recommendation
|
General Laboratory
|
Complete blood count, serum electrolytes, renal function, coagulation profile, liver
function tests
|
Copied from different iNYP dashboard
|
Health Issues
|
Two-section table of general and neurological diagnoses, grouped by ICD10 code
|
SME recommendation
|
Stroke Laboratory
|
Hemoglobin A1c, lipid panel, TSH, fingerstick glucose
|
SME recommendation
|
Hypercoagulable/inflammatory workup
|
ESR/CRP, RF, ANA, anti-dsDNA; C3, C4; ANCA; anticardiolipin Ab; lupus anticoagulant;
anti-extractable nuclear Ab; lumbar puncture; protein C, protein S, anti-thrombin
III levels, HIV
|
SME recommendation
|
Pathology
|
Factor V Leiden, MTHFR and prothrombin gene mutations, cryoglobulin
|
SME recommendation
|
Notes[d]
|
All available neurology notes
|
SME recommendation
|
Medications
|
Inpatient, outpatient medications with dose, route, frequency
|
Copied from different iNYP dashboard
|
Abbreviations: Ab, antibody; ANA, antinuclear antibody; ANCA, antinuclear cytoplasmic
antibodies; ASCVD, atherosclerotic cardiovascular disease; BMI, body mass index; iNYP,
i-NewYork Presbyterian; BP, blood pressure; CRP, C-reactive protein; CT, computed
tomography; dsDNA, double-stranded DNA; EEG, electroencephalogram; EKG, electrocardiogram;
ESR, erythrocyte sedimentation rate; HIV, human immunodeficiency virus; MR, magnetic
resonance; MTHFR, methylene-tetrahydrofolate reductase; SME, domain expert; TEE, transesophageal
echocardiogram; TSH, thyroid stimulating hormone; TTE, transthoracic echocardiogram.
a Laboratory/test reports and notes pop-up in separate browser windows from clicking
hyperlinked report/note dates, unless otherwise specified. ASCVD 10-year risk is computed
according the American College of Cardiology Guideline on the Assessment of Cardiovascular
Risk.[42]
b Displays red triangle followed by body-mass-index and corresponding category (overweight,
obese, severely obese) if the latter is greater than 25, 30, or 40, respectively.
Configured to display red triangle if most recent systolic blood pressure or diastolic
blood pressure is greater than 140 or 90 mm Hg, respectively.
c Displays all available CT/MR spine, head/brain studies, as well as CT, MR angiograms
of head/neck, and cerebral angiography.
d Displays any note with title containing “Neuro,” “Neurology,” “Neuro ICU,” or “Epilepsy.”
Fig. 2 Screenshot of iNYP dashboard with only eight tiles shown. Statin and blood pressure
clinical decision support alerts are pictured in red boxes.
Fig. 3 Closeup of neuroimaging tile with picture archive communication system (PACS) link
icon in red box.
Population and Measurements
At CUIMC, neurology house staff physicians are responsible for receiving all inpatient
consultation requests and performing the history, physical examination, and initial
assessment, and plan for each consultation. Using the institutional EHR, we identified
all distinct, initial inpatient neurological consultations at CUIMC that were completed
between July 1, 2017 and November 30, 2017, as well as for the 5-month period, following
the dashboard go-live on December 1, 2017 ending on April 30, 2018. We also identified
all inpatient neurology consultations for the 7-month interval period between May
1, 2018 and November 30, 2018. For these time periods, we calculated consultation
TAT as the difference in time between the placement of the consultation order by the
requesting provider in the EHR, and the timestamp corresponding to consultation note
completion by the neurological consultant in the EHR. Using log data that contained
user names, titles, and departments, we counted total dashboard users, unique page
hits, and unique patient chart accesses at 5 months after the go-live and 12 months
after the go-live. Using 5-month log data only, we also determined each user's title
and affiliated department, and applied text mining to the user title field to determine
whether the user was a member of the institutional house staff (defined as resident
or fellow).
Statistical Analysis
We counted consultation volume, page hits, unique chart accesses, and user numbers;
we also determined proportions of users according to disciplines, physician type,
and department. We determined the median consult TAT and interquartile range (IQR)
for each study period and compared median consultation TAT for each of the three time
periods using the Wilcoxon's rank-sum test. All analyses were performed using R version
3.5 (R Foundation, Vienna, Austria). Statistical significance was set at a two-sided
α of 0.05.
Results
Consultation Volumes and Turnaround Times
No major unexpected organizational changes that could have significantly affected
neurological consultation volume or TAT occurred over the study period, such as modifications
to physician staffing models, note templates, order sets, EHR or data platform interfaces,
or care pathways. In the 5-month period ending on November 30, 2017, 1,434 neurology
consultations were completed which was less than the 1,672 consultations completed
over the 5-month period ending April 30, 2018, and the 2,160 consultations completed
over the 7-month period between May 1, 2018 and November 30, 2018 ([Fig. 4]). The median TAT for the 5-month period after the go-live was significantly shorter
than that of the 5-month period preceding the go-live by 0.2 hours (p = 0.001; [Fig. 5]). The median TAT for the 7-month period ending November 30, 2018 was not significantly
different from the 5-month period after the go-live (p = 0.26) but remained significantly shorter than the median TAT for the 5-month period
preceding the go-live (p = 0.02).
Fig. 4 Neurology consultation volumes at 5 months before go-live, 5 months after go-live,
and between 5 and 12 months after go-live.
Fig. 5 Median consultation turnaround time (TAT) and interquartile range for each study
period. IQR, interquartile range.
Usage Patterns
By April 30, 2018, at 5 months after the dashboard go-live, we identified 269 unique
users, 684 dashboard page hits (median hits/user = 1.0, interquartile range [IQR] = 1.0),
and 510 unique patient chart accesses. By November 30, 2018, at 12 months after the
go-live, unique user, page hits, and chart access counts had increased by approximately
1.5-fold (598, 1044, and 1336, respectively; [Fig. 6]). At 5 months after dashboard go-live, usage logs showed that 133 dashboard users
(49.4%) were physicians of whom 92 (69.2%) were house staff. Of the physician users,
35 (26.3%) were neurologists. Of the house staff users, 27 (29.3%) were neurology
house staff; however, neurology house staff comprised 77.1% of neurologist users and
73% of CUIMC's 37-resident training program during the 2017 to 2018 academic year
([Fig. 7]).
Fig. 6 Cumulative dashboard usage counts at 5 and 12 month after go-live.
Fig. 7 Proportions of dashboard users at 5 months post-go-live. Clockwise from top left:
physicians as proportion of all users; house staff as proportion of all physicians;
neurology house staff as proportion of all house staff; neurologists as proportion
of all physicians.
In aggregate, neurologists were responsible for generating 223 (35.8%) of total dashboard
page hits, and neurology house staff comprised 8 (61.5%) of the 13 users that generated
the top quintile of page hits. The Internal Medicine and Neurology departments had
the most users among the top quintile of page hits, followed by anesthesia, as well
as both pediatrics and pathology ([Fig. 8]). Of the 136 nonphysician users, the most frequently represented users in descending
order were administrative personnel (n = 49; 36.0%), nursing (n = 24; 17.6%), medical students (n = 22; 16.2%), and research staff (n = 13; 9.6%; data not shown).
Fig. 8 Top quintile by page hits per user, stratified by academic department.
Discussion
In a neurology department at a large academic medical center, we found that in comparison
to 5 months before implementation of an aggregative dashboard, inpatient neurological
consult TAT was significantly shortened by approximately 12 minutes in a 5-month period
following implementation. This time difference remained stable over the following
7 months, suggesting a sustained reduction in TAT. Further, based on log data, we
found that house staff constituted nearly two-thirds of all physician users that most
neurology users were neurology house staff, and that most neurological residents at
CUIMC were users.
Our findings suggest that our implementation was successful in targeting neurology
residents for use of the dashboard. Further, a TAT reduction of 12 minutes may be
significant, considering the daily academic obligations of residency trainees, and
the small size of the CUIMC consult service team (two to three residents) that receive
consultation requests and perform evaluations on a daily basis during daytime hours.
In the post-go-live period, the mean daily consult volume was approximately 11.1,
resulting in approximately 3.7 to 5.5 consults/day/resident depending on consult team
size. Whereas a reduction of 12 minutes per consult could translate to an approximate
aggregate time savings of 44 to 60 minutes per daytime resident shift, overnight,
and on weekends, one resident receives all inpatient consultations for the hospital,
thereby increasing the consult volume and the potential time savings.
The specifics of the CUIMC consult service are important for contextualizing our findings
and may have implications for nonacademic neurological practitioners. At CUIMC, consultation
volume, patient complexity, and the logistics of the consult service all combine to
produce a high-EHR work burden for neurology residents, who are the front-line responders
to consultation requests. An internal analysis of institutional consult volume during
the 2017 to 2018 academic year demonstrates that neurology was the most frequently
consulted inpatient service at CUIMC.[29] This volume is compounded by sparse off-hour staffing, which further reduces providers'
ability to perform timely and thorough consultations. Furthermore, many patients are
referred to CUIMC for complex conditions, or for advanced diagnostic or therapeutic
modalities, which may increase the amount of EHR data to be reviewed for a neurological
consultation. The complexity of EHR data, detailed neurological examinations, and
high consultation volume may place time pressure on consultants, but these circumstances
also create opportunities for efficiency improvements, particularly where the need
for face-to-face time with the patient and hands-on evaluations accentuate the cognitive
demands of locating and consolidating data to formulate assessments and recommendations.
While our results are immediately relevant to neurologists practicing at academic
medical centers similar to CUIMC, the significant amount of consultative work performed
by neurologists may also make our findings generalizable to neurological practitioners
as a whole.
Aggregative dashboards are but one of many tools for reducing EHR workload. Such solutions
include note templates, such as that provided by Epic SmartPhrases (Epic Systems Corporation,
Verona, Wisconsin, Unites States),[30]
[31] strategic menu design,[32]
[33] and order sets.[34] Automation-centric innovations also exist, such as automated problem list generation[35] and follow-up determination after radiographic testing,[36] as well as clinical decision support, alerts for risk stratification[37] and drug–disease interactions.[38] A study by Arndt et al of ambulatory family practitioners in the United States showed
that clinicians devote nearly 50% of their total daily time in their EHR performing
clinical documentation and chart review, with chart review comprising nearly 75% of
EHR tasks related to direct medical care.[11] Additionally, Neri et al reported that emergency physicians struggle to integrate
clinical data from disparate sources into their clinical documentation workflows which
occupy a significant portion of their work.[39] Given the significance of chart review and data integration for physicians, dashboards
are uniquely poised to facilitate these tasks by providing tailored data aggregation
and potentially reducing cognitive load. Furthermore, dashboard aggregation has also
been shown to reduce medication errors at the point of care,[40] suggesting that the benefits of such dashboards may extend beyond EHR work reductions
to include clinical benefits as well.
While we advertised the dashboard as a demonstration to the CUIMC neurology faculty
and residents, and most of the institutional neurology residents used the dashboard,
logs suggest that at least in the short term, the majority of users were not neurologists.
This finding may have resulted from the fact that demonstration of the dashboard was
too infrequent, or that dissemination made use of suboptimal venues or modalities.
The high usage of the dashboard by general internal medicine residents may be related
to the existence of iNYP dashboards with similar functionalities for internal medicine,
such as general ambulatory medicine and chronic kidney disease, which appear immediately
above the neurology dashboard link in the iNYP web page. Finally, physicians caring
for many patients admitted to internal medicine services may also be managing patients
with comorbid neurological disease, such as stroke, encephalopathy, or seizure, and
may have plausible use for the aggregative data review functions of the iNYP neurology
dashboard.
Our findings also illustrate the implementation of a process improvement solution
in a system that is external to the institutional EHR, and therefore potentially emphasize
the relative advantages of such a developmental approach. It is important to note
that our study did not specifically compare ease of change implementation in transactional
(such as CUIMC's institutional EHR) to nontransactional, derivative (such as iNYP
dashboards) systems. Nonetheless, while modifications to transactional systems can
present challenges due to direct care quality and safety implications, changes may
be relatively easier to implement in a dashboard that is used as a separate clinical
data review tool. Despite the relative ease in modifying nontransactional systems,
dashboards are also susceptible to errors other than those originating from their
source systems. Reliance on dashboards over source systems, as can occur with some
successful dashboard implementations, may encourage a lack of use and/or verification
of source system data, thereby leading to incorrect identification of dashboard bugs
or data are sues with consequent effects on medical decision making and patient care.
Limitations
This study is limited by several important factors. First, this was a retrospective
study, and we were not able to control which user groups took advantage of the dashboard
to make clinical decisions. Second, due to our reliance on user log data alone, we
were not able to determine which dashboard links, functions, and tiles were utilized
and could not measure the exact amount of time users spent performing specific EHR
tasks nor in which care setting and at which point in the workflow users accessed
and used the dashboard. We therefore could not definitively determine that the dashboard
was used to support clinical decision making, nor that use of the dashboard occurred
during inpatient neurological consultations, as intended, nor that use of the dashboard
resulted in faster data review. As such, the observed reduction in consultation TAT
may be related to factors independent from dashboard use, such as higher levels of
patient complexity in the pre-go-live period, relative inexperience of new residents
starting consultative work in the pre-go-live period, or higher volumes in the post-go-live
periods which may have caused residents to be able to spend less time per individual
consult.
A third limitation was that we calculated TAT as the difference between consult order
and consult note redaction as our measurable outcome. Due to the fact that many users
were trainees, we could not entirely exclude that users could have performed nonconsult-related
responsibilities prior to completing their consultations, such as attending academic
conferences, thereby making TAT a potentially inaccurate measure of the dashboard's
effectiveness in reducing the amount of work involved in reviewing EHR data. However,
it is typically impressed upon trainees that patient care should take precedence over
academic activities, so we believe it is unlikely that academic activities inflated
the consult TAT. Fourth, the dashboard was heavily weighted toward vascular neurology
evaluations. We did not perform a stratified analysis by patient diagnoses which could
have enabled us to determine whether the dashboard was useful for evaluation of specific
neurological diagnoses, such as stroke, seizure, or encephalopathy. Fifth, we did
not measure cognitive burden on the users, quality of care, or hospital resource utilization,
all of which are important outcomes to evaluate in an implementation such as this.
Additionally, we disseminated the dashboard only through a demonstration to users,
rather than through a pilot program with progressive extension to a larger user group,
which may have limited the amount of usage among neurology residents and potentially
limited our understanding of how to tailor the dashboard to use cases that were not
understood as part of the initial requirement gathering process. However, the majority
of the neurology house staff used the dashboard during the study period, suggesting
that the dissemination strategy we employed may have been partially successful in
reaching our audience. It should also be emphasized that the time involved in performing
a consultation is independent of the quality of medical care provided, and while our
study focused on the effect of the dashboard on consult TAT, this was not a guideline-based
outcome that might accurately reflect quality of care, such as that studied in the
PRESCRIBE cluster-randomized controlled trial.[41]
Despite these limitations, we found that a 5-month decrease in inpatient neurological
consultation TAT followed the neurological dashboard go-live at a large academic medical
center, which persisted over the following 7 months. We cannot definitively conclude
that introduction of the iNYP neurology dashboard resulted in reduced TAT, decreases
in cognitive burden, or increases in face-to-face time with patients, but we demonstrated
that significant numbers of neurological and nonneurological house staff used the
dashboard. However, a future prospective cluster-randomized trial that contrasts consultation
TAT, cognitive burden, measures of burnout, and hospital resource utilization during
conventional data review with data review using a neurological dashboard could provide
further guidance. This trial should aim to collect data on time spent and usage of
each dashboard tile, and should attempt to adjust for patient demographics, clinical
characteristics, month of evaluation, as well as resident training levels. Additionally,
further work should focus on investigating the impact of neurological dashboards on
guideline-based recommended practices, and whether aggregation of neurological data
into a single-digital access point significantly impacts cognitive overload, burnout,
and time spent with patients.
Clinical Relevance Statement
Clinical Relevance Statement
In this description of an implementation of an aggregative neurological dashboard
at a large academic center, the authors measure the neurology consultation TAT 5 months
prior to, and at both 5 months after dashboard implementation, as well as over the
7-month period following. In comparison to the 5-month period prior to implementation,
consult TAT was significantly shorter at 5 months postimplementation, and this shorter
TAT remained unchanged over the following 7 months. In addition to show high usage
among institutional neurology residents, usage logs showed that high numbers of physicians
from Internal Medicine also used the dashboard over the study period. Further work
is needed to determine whether such dashboards significantly impact cognitive load
and turnaround time in neurology.
Multiple Choice Questions
Multiple Choice Questions
-
Which answer best described this study's finding when comparing in-hospital neurology
consultation turnaround time during the 5-month period after and prior to the implementation
of a neurology dashboard?
-
Statistically significantly longer.
-
Statistically significantly shorter.
-
Longer, but not statistically significantly so.
-
Shorter, but not statistically significantly so.
Correct Answer: The correct answer is option b.
-
In this study of a neurology dashboard, approximately what percentage of the total
dashboard users were physicians?
Correct Answer: The correct answer is option b.
-
Which aspect of the study design most significantly limited the association between
dashboard use and turnaround time?
-
Medication orders were not included in the dashboard.
-
The usage logs did not capture sufficient numbers of users.
-
Dashboard usage was not tracked during and outside of consultations.
-
The dashboard was only accessible via web page.
Correct Answer: The correct answer is option c.