User login
Cost Comparison of 2 Video Laryngoscopes in a Large Academic Center
From the Department of Anesthesiology, Thomas Jefferson University and Hospitals, Sidney Kimmel Medical College, Philadelphia, PA, and Sidney Kimmel Medical College at Thomas Jefferson University, Philadelphia, PA.
Objective: Retrospective study examining hospital cost information of patients requiring endotracheal intubation with video laryngoscopy. Provide a practical cost assessment on use of the McGRATH and GlideScope video laryngoscopes (VLs).
Methods: This study examined 52 hospital locations within a single, large university hospital, with most of those locations being hospital operating rooms. A total of 34 600 endotracheal intubations performed over 24 months, of which 11 345 were video laryngoscopies. Electronic medical records containing demographic data and information related to endotracheal intubation procedures, with monthly breakdowns between GlideScope and McGRATH intubations, were reviewed. Cost information calculated for equipment, blades, batteries, repairs, and subsequent analysis performed to determine cost differences between those 2 instruments during the COVID-19 period.
Results: A total of 5501 video laryngoscopy procedures were performed using the McGRATH VL and 5305 were performed using the GlideScope VL. Costs over 24 months were $181 093 lower (55.5%) for McGRATH compared to GlideScope. The mean (SD) monthly costs for GlideScope blades were $3837 ($1050) and $3236 ($538) for years 1 and 2, respectively, vs $1652 ($663) and $2933 ($585) for McGRATH blades (P < .001). Most total cost differences were attributed to equipment and blade purchases, which were $202 595 (65.0%) higher for GlideScope. During the COVID-19 period, the use of the McGRATH increased to 61% of all video laryngoscopy cases, compared to 37% for GlideScope (P < .001). Blade cost difference for the COVID-19 period was $128 higher for the McGRATH even though 293 more intubations were performed with that device.
Conclusions: Use of the McGRATH resulted in a cost savings of 55% compared to the GlideScope, and its use was highest during the COVID-19 period, which may be explained by its more portable and practical features.
Keywords: video laryngoscope; McGRATH; GlideScope; endotracheal intubation; hospital costs; COVID-19.
Hospitals have come to rely on video laryngoscopes (VLs) for tracheal intubation as necessary tools for better visualization of airways. Modern video laryngoscopy developed in the 2000s1 as a progression from direct laryngoscopy, which began in 1852 when Horace Green used a bent tongue spatula and sunlight to examine a child.2 VLs have seen many improvements and adaptations of their own, resulting in many different styles and types circulating around hospitals. The GlideScope (Verathon Inc, Bothell, WA) and the McGRATH (Medtronic, Minneapolis, MN) are examples of such instruments, which are now widely used in the US and are the 2 VLs of choice at our institution.
A few studies have compared VLs to direct laryngoscopes. In their systematic review, Lewis et al have shown the numerous benefits of using a VL over a direct laryngoscope. Some general conclusions were that the use of video laryngoscopy reduced the number of failed intubations, decreased laryngeal trauma, and provided improved visualizations.3 Other studies have compared the different types of VLs, including the McGRATH and the GlideScope, examining factors such as intubation time and display quality of the image. Two studies found that medical students were equally successful at using both the McGRATH and the GlideScope,4,5 while another study found that care providers using the GlideScope had quicker intubation times.6 Lastly, Savoldelli et al concluded that more providers preferred the McGRATH, which provided better laryngeal views,7 while their subsequent study showed more favorable learning curves of the Airtraq compared to the McGRATH and other VLs.8
Although there have been no reported differences in safety and effectiveness of the McGRATH and GlideScope devices, cost data on the use of these 2 popular laryngoscopes are lacking. Such information is important considering the increasing costs of medical technologies and the significant financial losses experienced by health care systems due to the COVID-19 crisis. The purpose of this retrospective cohort study was to compare the cost efficiency of the McGRATH MAC and GlideScope Core VLs at a large academic center.
Methods
This retrospective study was performed under exemption from the Thomas Jefferson University Institutional Review Board. The primary data sources consisted of hospital electronic patient records (EPIC) and cost information from the device manufacturers and hospital staff. The electronic patient data were provided by the EPIC Enterprise Analytics Business Intelligence group at Thomas Jefferson University Hospital (Center City Campus, Philadelphia, PA), while device costs were obtained from Verathon, Medtronic, and departmental staff responsible for purchasing equipment. Monthly data were obtained over a 24-month period (June 2018 through May 2020) when the McGRATH VL was placed into use in the department of anesthesiology. The 2 types of VLs were made available for use in a total of 52 locations, with the majority being hospital operating rooms.
The following variables were recorded: number of endotracheal intubations performed each month with breakdown between video laryngoscopy and flexible bronchoscopy airways, frequency of use for each type of laryngoscope, blades used, and equipment costs for use of each laryngoscope. Hospital cost estimates for both the McGRATH and GlideScope laryngoscopes included batteries, handles, blades, and the devices themselves. Cost data were also collected on frequency of device failure, maintenance, and replacement of parts and lost equipment.
Analysis
De-identified electronic medical records consisted of nominal and quantitative variables, with demographic data and information related to the endotracheal intubation procedure. All data were in chronological order and sorted by date after which coding was applied, to identify device type and allocate pertinent cost information. Descriptive statistics were reported as mean (SD) and sum for costs; frequency tables were generated for intubation procedures according to device type and time periods. Data were analyzed using the χ2 test, the student t test, and the Wilcoxon Mann-Whitney U test, with a P value set at .05 for statistical significance. SPSS version 26 and GraphPad Prism version 6 were used for all statistical analyses.
Results
A total of 34 600 endotracheal intubations were performed over the 24-month study period, and 11 345 (32.8%) were video laryngoscopy procedures. Out of all video laryngoscopy procedures, 5501 (48.5%) were performed using the McGRATH VL and 5305 (46.8%) were conducted using the GlideScope VL. The difference of 539 (4.8%) cases accounts for flexible bronchoscopy procedures and endotracheal intubations using other video laryngoscopy equipment. The mean (SD) monthly number of video laryngoscopy procedures for the 24 months was 221 (54) and 229 (89) for the GlideScope and McGRATH devices, respectively. Monthly endotracheal intubation distributions over 24 months trended upward for the McGRATH VL and downward for the GlideScope, but there was no statistically significant (P = .71) difference in overall use between the 2 instruments (Figure 1).
To examine the observed usage trends between the 2 VL during the first and last 12 months, a univariate ANOVA was conducted with the 2 time periods entered as predictors in the model. Video laryngoscopy intubations were performed (P = .001) more frequently with the GlideScope during the first 12 months; however, use of the McGRATH VL increased (P < .001) during the following 12 months compared to GlideScope. The GlideScope accounted for 54% of all VL intubations during the first 12 months, with the McGRATH accounting for 58% of all video laryngoscopy procedures for months 12 to 24. Additionally, the increase in video laryngoscopy procedures with the McGRATH during the last 3 months of the study period was despite an overall reduction in surgical volume due to the COVID-19 crisis, defined for this study as March 1, 2020, to May 31, 2020 (Figure 1). There was a statistically significant (P < .001) difference in the case distribution between use of the McGRATH and GlideScope VL for that period. The anesthesia personnel’s use of the McGRATH VL increased to 61% of all video laryngoscopy cases, compared to 37% for the GlideScope (Figure 2).
The total costs calculated for equipment, blades, and repairs are presented in Table 1 and yearly total costs are shown in Figure 3. Overall costs were $181 093 lower (55.5%) for the McGRATH VL compared to the GlideScope over the 24-month period. The mean (SD) monthly costs for GlideScope VL blades were $3837 ($1050) and $3236 ($538) for years 1 and 2, respectively, vs $1652 ($663) and $2933 ($585) for the McGRATH VL blades. Most of the total cost differences were attributed to equipment and blade purchases, which were $202 595 (65.0%) higher for the GlideScope compared to the McGRATH VL. The monthly blade costs alone were higher (P < .001) for the GlideScope over the 2-year period; however, the McGRATH VL required use of disposable stylets at a cost of $10 177 for all endotracheal intubations, compared to $700 for the GlideScope device.
An analysis was performed to determine whether costs differed between those 2 instruments during the COVID-19 period. There was a statistically significant (P < .001) difference in the case distribution between use of the McGRATH and GlideScope VLs during that period. The calculated blade cost difference for the COVID period was $128 higher for the McGRATH even though 293 more intubations were performed with that device (Table 2).
Discussion
We attempted to provide useful cost estimates by presenting pricing data reflecting the approximate cost that most large institutional anesthesia practices would incur for using those 2 specific devices and related peripherals. The main findings of our analysis showed that use of the McGRATH MAC VL resulted in a 55% cost savings compared to the GlideScope, with a similar number of cases performed with each device over the 24-month study period. We believe this represents a substantial savings to the department and institution, which has prompted internal review on the use of video laryngoscopy equipment. None of the McGRATH units failed; however, the GlideScope required 3 baton replacements.
Of note, use of the McGRATH MAC increased during the COVID-19 period, which may be explained by the fact that the operators found it to be a more portable device. Several physicians in the department commented that its smaller size made the McGRATH MAC more practical during the time when a plexiglass box was being used around the patient’s head to shield the intubator from aerosolized viral particles.
Although this study demonstrated the cost-saving value of the McGRATH over the GlideScope, a suggested next step would be to examine resource utilization related to video laryngoscopy use. The more dynamic tracking of the use of these devices should facilitate the assessment of existing related resources and decision making, to optimize the benefits of this initiative. We would anticipate reduced use of anesthesia personnel, such as technicians to assist with the management of this device which could be significant. As new respiratory viruses are appearing each year, video laryngoscopy will continue to gain increasing use in operating rooms and acute care locations. The adding of protective barriers between patients and providers calls for use of the most practical and effective VL devices, to protect personnel who are at high risk of contamination from airway secretions and aerosolized particles.9,10
The COVID-19 pandemic has demonstrated the value of anesthesiology in regards to analyzing and finding solutions to effectively manage infected patients or those suspected of infection in the perioperative environment. Inexpensive products are often avoided because cheaper devices are associated with being of lower quality. However, the association with cost and quality—and the assumption that a higher price is positively correlated with higher quality—is overall inconsistent in the medical literature.11 A more effective or higher quality treatment does not necessarily cost more and may actually end up costing less,12 as was the case in this study. We have been able to directly cut departmental expenses by using a more efficient and cost-effective device for intubations, without compromising safety and efficacy. Future studies should determine whether this significant reduction in costs from video laryngoscopy intubations with the McGRATH VL will be sustained across anesthesiology departments in the Jefferson Health Enterprise Hospitals, or other health systems, as well as its impact on workflow and personnel resources.
This analysis was restricted to one of the campuses of the Jefferson Health Enterprise. However, this is the largest anesthesia practice, encompassing several locations, which should reflect the general practice patterns across other anesthesiology departments in this large institution. The costs for the devices and peripherals may vary across anesthesia practices depending on volume and contracts negotiated with the suppliers. It was not possible to estimate this variability, which could change the total costs by a few percentage points. We recognize that there may be other costs associated with securing the McGRATH VL to prevent loss from theft or misplacement, which were not included in the study. Lastly, the inability to obtain randomized samples for the 2 groups treated with each device opens up the possibility of selection bias. There were, however, multiple intubators who were free to select 1 of the devices for endotracheal intubation, which may have reduced the effect of selection bias.
Conclusion
This study demonstrated that over a 24-month period use of the McGRATH MAC VL resulted in a cost reduction of around 55% compared to using the GlideScope for endotracheal intubation procedures performed at a major academic center. Over the first 3 months of the COVID-19 crisis, which our study included, use of the McGRATH VL increased while GlideScope use decreased. This was most likely related to the portability and smaller size of the McGRATH, which better facilitated intubations of COVID-19 patients.
Acknowledgements: The authors thank Craig Smith, Senior Anesthesia Technician, for his assistance with the cost information and excellent record-keeping related to the use of video laryngoscopes.
Corresponding author: Marc C. Torjman, PhD, Professor, Department of Anesthesiology, Sidney Kimmel Medical College at Thomas Jefferson University, 111 South 11th St, Suite G-8290, Philadelphia, PA 19107; Marc.Torjman@Jefferson.edu.
Financial disclosures: Dr. Thaler has served as a consultant for Medtronic since September 2020. He has participated in 2 webinars on the routine use of video laryngoscopy.
Funding: This study was supported by the Department of Anesthesiology at Thomas Jefferson University.
1. Channa AB. Video laryngoscopes. Saudi J Anaesth. 2011;5(4):357-359.
2. Pieters BM, Eindhoven GB, Acott C, Van Zundert AAJ. Pioneers of laryngoscopy: indirect, direct and video laryngoscopy. Anaesth Intensive Care. 2015;43(suppl):4-11.
3. Lewis SR, Butler AR, Parker J, et al. Videolaryngoscopy versus direct laryngoscopy for adult patients requiring tracheal intubation. Cochrane Database Syst Rev. 2016;11(11):CD011136.
4. Kim W, Choi HJ, Lim T, Kang BS. Can the new McGrath laryngoscope rival the GlideScope Ranger portable video laryngoscope? A randomized manikin study. Am J Emerg Med. 2014;32(10):1225-1229.
5. Kim W, Choi HJ, Lim T, et al. Is McGrath MAC better than Glidescope Ranger for novice providers in the simulated difficult airway? A randomized manikin study. Resuscitation. 2014;85(suppl 1):S32.
6. Jeon WJ, Kim KH, Yeom JH, et al. A comparison of the Glidescope to the McGrath videolaryngoscope in patients. Korean J Anesthesiol. 2011;61(1):19-23.
7. Savoldelli GL, Schiffer E, Abegg C, et al. Comparison of the Glidescope, the McGrath, the Airtraq and the Macintosh laryngoscopes in simulated difficult airways. Anaesthesia. 2008;63(12):1358-1364.
8. Savoldelli GL, Schiffer E, Abegg C, et al. Learning curves of the Glidescope, the McGrath and the Airtraq laryngoscopes: a manikin study. Eur J Anaesthesiol. 2009;26(7):554-558.
9. Schumacher J, Arlidge J, Dudley D, et al. The impact of respiratory protective equipment on difficult airway management: a randomised, crossover, simulation study. Anaesthesia. 2020;75(10):1301-1306.
10. De Jong A, Pardo E, Rolle A, et al. Airway management for COVID-19: a move towards universal videolaryngoscope? Lancet Respir Med. 2020;8(6):555.
11. Hussey PS, Wertheimer S, Mehrotra A. The association between health care quality and cost: a systematic review. Ann Intern Med. 2013;158(1):27-34.
12. Mitton C, Dionne F, Peacock S, Sheps S. Quality and cost in healthcare: a relationship worth examining. Appl Health Econ Health Policy. 2006;5(4):201-208.
From the Department of Anesthesiology, Thomas Jefferson University and Hospitals, Sidney Kimmel Medical College, Philadelphia, PA, and Sidney Kimmel Medical College at Thomas Jefferson University, Philadelphia, PA.
Objective: Retrospective study examining hospital cost information of patients requiring endotracheal intubation with video laryngoscopy. Provide a practical cost assessment on use of the McGRATH and GlideScope video laryngoscopes (VLs).
Methods: This study examined 52 hospital locations within a single, large university hospital, with most of those locations being hospital operating rooms. A total of 34 600 endotracheal intubations performed over 24 months, of which 11 345 were video laryngoscopies. Electronic medical records containing demographic data and information related to endotracheal intubation procedures, with monthly breakdowns between GlideScope and McGRATH intubations, were reviewed. Cost information calculated for equipment, blades, batteries, repairs, and subsequent analysis performed to determine cost differences between those 2 instruments during the COVID-19 period.
Results: A total of 5501 video laryngoscopy procedures were performed using the McGRATH VL and 5305 were performed using the GlideScope VL. Costs over 24 months were $181 093 lower (55.5%) for McGRATH compared to GlideScope. The mean (SD) monthly costs for GlideScope blades were $3837 ($1050) and $3236 ($538) for years 1 and 2, respectively, vs $1652 ($663) and $2933 ($585) for McGRATH blades (P < .001). Most total cost differences were attributed to equipment and blade purchases, which were $202 595 (65.0%) higher for GlideScope. During the COVID-19 period, the use of the McGRATH increased to 61% of all video laryngoscopy cases, compared to 37% for GlideScope (P < .001). Blade cost difference for the COVID-19 period was $128 higher for the McGRATH even though 293 more intubations were performed with that device.
Conclusions: Use of the McGRATH resulted in a cost savings of 55% compared to the GlideScope, and its use was highest during the COVID-19 period, which may be explained by its more portable and practical features.
Keywords: video laryngoscope; McGRATH; GlideScope; endotracheal intubation; hospital costs; COVID-19.
Hospitals have come to rely on video laryngoscopes (VLs) for tracheal intubation as necessary tools for better visualization of airways. Modern video laryngoscopy developed in the 2000s1 as a progression from direct laryngoscopy, which began in 1852 when Horace Green used a bent tongue spatula and sunlight to examine a child.2 VLs have seen many improvements and adaptations of their own, resulting in many different styles and types circulating around hospitals. The GlideScope (Verathon Inc, Bothell, WA) and the McGRATH (Medtronic, Minneapolis, MN) are examples of such instruments, which are now widely used in the US and are the 2 VLs of choice at our institution.
A few studies have compared VLs to direct laryngoscopes. In their systematic review, Lewis et al have shown the numerous benefits of using a VL over a direct laryngoscope. Some general conclusions were that the use of video laryngoscopy reduced the number of failed intubations, decreased laryngeal trauma, and provided improved visualizations.3 Other studies have compared the different types of VLs, including the McGRATH and the GlideScope, examining factors such as intubation time and display quality of the image. Two studies found that medical students were equally successful at using both the McGRATH and the GlideScope,4,5 while another study found that care providers using the GlideScope had quicker intubation times.6 Lastly, Savoldelli et al concluded that more providers preferred the McGRATH, which provided better laryngeal views,7 while their subsequent study showed more favorable learning curves of the Airtraq compared to the McGRATH and other VLs.8
Although there have been no reported differences in safety and effectiveness of the McGRATH and GlideScope devices, cost data on the use of these 2 popular laryngoscopes are lacking. Such information is important considering the increasing costs of medical technologies and the significant financial losses experienced by health care systems due to the COVID-19 crisis. The purpose of this retrospective cohort study was to compare the cost efficiency of the McGRATH MAC and GlideScope Core VLs at a large academic center.
Methods
This retrospective study was performed under exemption from the Thomas Jefferson University Institutional Review Board. The primary data sources consisted of hospital electronic patient records (EPIC) and cost information from the device manufacturers and hospital staff. The electronic patient data were provided by the EPIC Enterprise Analytics Business Intelligence group at Thomas Jefferson University Hospital (Center City Campus, Philadelphia, PA), while device costs were obtained from Verathon, Medtronic, and departmental staff responsible for purchasing equipment. Monthly data were obtained over a 24-month period (June 2018 through May 2020) when the McGRATH VL was placed into use in the department of anesthesiology. The 2 types of VLs were made available for use in a total of 52 locations, with the majority being hospital operating rooms.
The following variables were recorded: number of endotracheal intubations performed each month with breakdown between video laryngoscopy and flexible bronchoscopy airways, frequency of use for each type of laryngoscope, blades used, and equipment costs for use of each laryngoscope. Hospital cost estimates for both the McGRATH and GlideScope laryngoscopes included batteries, handles, blades, and the devices themselves. Cost data were also collected on frequency of device failure, maintenance, and replacement of parts and lost equipment.
Analysis
De-identified electronic medical records consisted of nominal and quantitative variables, with demographic data and information related to the endotracheal intubation procedure. All data were in chronological order and sorted by date after which coding was applied, to identify device type and allocate pertinent cost information. Descriptive statistics were reported as mean (SD) and sum for costs; frequency tables were generated for intubation procedures according to device type and time periods. Data were analyzed using the χ2 test, the student t test, and the Wilcoxon Mann-Whitney U test, with a P value set at .05 for statistical significance. SPSS version 26 and GraphPad Prism version 6 were used for all statistical analyses.
Results
A total of 34 600 endotracheal intubations were performed over the 24-month study period, and 11 345 (32.8%) were video laryngoscopy procedures. Out of all video laryngoscopy procedures, 5501 (48.5%) were performed using the McGRATH VL and 5305 (46.8%) were conducted using the GlideScope VL. The difference of 539 (4.8%) cases accounts for flexible bronchoscopy procedures and endotracheal intubations using other video laryngoscopy equipment. The mean (SD) monthly number of video laryngoscopy procedures for the 24 months was 221 (54) and 229 (89) for the GlideScope and McGRATH devices, respectively. Monthly endotracheal intubation distributions over 24 months trended upward for the McGRATH VL and downward for the GlideScope, but there was no statistically significant (P = .71) difference in overall use between the 2 instruments (Figure 1).
To examine the observed usage trends between the 2 VL during the first and last 12 months, a univariate ANOVA was conducted with the 2 time periods entered as predictors in the model. Video laryngoscopy intubations were performed (P = .001) more frequently with the GlideScope during the first 12 months; however, use of the McGRATH VL increased (P < .001) during the following 12 months compared to GlideScope. The GlideScope accounted for 54% of all VL intubations during the first 12 months, with the McGRATH accounting for 58% of all video laryngoscopy procedures for months 12 to 24. Additionally, the increase in video laryngoscopy procedures with the McGRATH during the last 3 months of the study period was despite an overall reduction in surgical volume due to the COVID-19 crisis, defined for this study as March 1, 2020, to May 31, 2020 (Figure 1). There was a statistically significant (P < .001) difference in the case distribution between use of the McGRATH and GlideScope VL for that period. The anesthesia personnel’s use of the McGRATH VL increased to 61% of all video laryngoscopy cases, compared to 37% for the GlideScope (Figure 2).
The total costs calculated for equipment, blades, and repairs are presented in Table 1 and yearly total costs are shown in Figure 3. Overall costs were $181 093 lower (55.5%) for the McGRATH VL compared to the GlideScope over the 24-month period. The mean (SD) monthly costs for GlideScope VL blades were $3837 ($1050) and $3236 ($538) for years 1 and 2, respectively, vs $1652 ($663) and $2933 ($585) for the McGRATH VL blades. Most of the total cost differences were attributed to equipment and blade purchases, which were $202 595 (65.0%) higher for the GlideScope compared to the McGRATH VL. The monthly blade costs alone were higher (P < .001) for the GlideScope over the 2-year period; however, the McGRATH VL required use of disposable stylets at a cost of $10 177 for all endotracheal intubations, compared to $700 for the GlideScope device.
An analysis was performed to determine whether costs differed between those 2 instruments during the COVID-19 period. There was a statistically significant (P < .001) difference in the case distribution between use of the McGRATH and GlideScope VLs during that period. The calculated blade cost difference for the COVID period was $128 higher for the McGRATH even though 293 more intubations were performed with that device (Table 2).
Discussion
We attempted to provide useful cost estimates by presenting pricing data reflecting the approximate cost that most large institutional anesthesia practices would incur for using those 2 specific devices and related peripherals. The main findings of our analysis showed that use of the McGRATH MAC VL resulted in a 55% cost savings compared to the GlideScope, with a similar number of cases performed with each device over the 24-month study period. We believe this represents a substantial savings to the department and institution, which has prompted internal review on the use of video laryngoscopy equipment. None of the McGRATH units failed; however, the GlideScope required 3 baton replacements.
Of note, use of the McGRATH MAC increased during the COVID-19 period, which may be explained by the fact that the operators found it to be a more portable device. Several physicians in the department commented that its smaller size made the McGRATH MAC more practical during the time when a plexiglass box was being used around the patient’s head to shield the intubator from aerosolized viral particles.
Although this study demonstrated the cost-saving value of the McGRATH over the GlideScope, a suggested next step would be to examine resource utilization related to video laryngoscopy use. The more dynamic tracking of the use of these devices should facilitate the assessment of existing related resources and decision making, to optimize the benefits of this initiative. We would anticipate reduced use of anesthesia personnel, such as technicians to assist with the management of this device which could be significant. As new respiratory viruses are appearing each year, video laryngoscopy will continue to gain increasing use in operating rooms and acute care locations. The adding of protective barriers between patients and providers calls for use of the most practical and effective VL devices, to protect personnel who are at high risk of contamination from airway secretions and aerosolized particles.9,10
The COVID-19 pandemic has demonstrated the value of anesthesiology in regards to analyzing and finding solutions to effectively manage infected patients or those suspected of infection in the perioperative environment. Inexpensive products are often avoided because cheaper devices are associated with being of lower quality. However, the association with cost and quality—and the assumption that a higher price is positively correlated with higher quality—is overall inconsistent in the medical literature.11 A more effective or higher quality treatment does not necessarily cost more and may actually end up costing less,12 as was the case in this study. We have been able to directly cut departmental expenses by using a more efficient and cost-effective device for intubations, without compromising safety and efficacy. Future studies should determine whether this significant reduction in costs from video laryngoscopy intubations with the McGRATH VL will be sustained across anesthesiology departments in the Jefferson Health Enterprise Hospitals, or other health systems, as well as its impact on workflow and personnel resources.
This analysis was restricted to one of the campuses of the Jefferson Health Enterprise. However, this is the largest anesthesia practice, encompassing several locations, which should reflect the general practice patterns across other anesthesiology departments in this large institution. The costs for the devices and peripherals may vary across anesthesia practices depending on volume and contracts negotiated with the suppliers. It was not possible to estimate this variability, which could change the total costs by a few percentage points. We recognize that there may be other costs associated with securing the McGRATH VL to prevent loss from theft or misplacement, which were not included in the study. Lastly, the inability to obtain randomized samples for the 2 groups treated with each device opens up the possibility of selection bias. There were, however, multiple intubators who were free to select 1 of the devices for endotracheal intubation, which may have reduced the effect of selection bias.
Conclusion
This study demonstrated that over a 24-month period use of the McGRATH MAC VL resulted in a cost reduction of around 55% compared to using the GlideScope for endotracheal intubation procedures performed at a major academic center. Over the first 3 months of the COVID-19 crisis, which our study included, use of the McGRATH VL increased while GlideScope use decreased. This was most likely related to the portability and smaller size of the McGRATH, which better facilitated intubations of COVID-19 patients.
Acknowledgements: The authors thank Craig Smith, Senior Anesthesia Technician, for his assistance with the cost information and excellent record-keeping related to the use of video laryngoscopes.
Corresponding author: Marc C. Torjman, PhD, Professor, Department of Anesthesiology, Sidney Kimmel Medical College at Thomas Jefferson University, 111 South 11th St, Suite G-8290, Philadelphia, PA 19107; Marc.Torjman@Jefferson.edu.
Financial disclosures: Dr. Thaler has served as a consultant for Medtronic since September 2020. He has participated in 2 webinars on the routine use of video laryngoscopy.
Funding: This study was supported by the Department of Anesthesiology at Thomas Jefferson University.
From the Department of Anesthesiology, Thomas Jefferson University and Hospitals, Sidney Kimmel Medical College, Philadelphia, PA, and Sidney Kimmel Medical College at Thomas Jefferson University, Philadelphia, PA.
Objective: Retrospective study examining hospital cost information of patients requiring endotracheal intubation with video laryngoscopy. Provide a practical cost assessment on use of the McGRATH and GlideScope video laryngoscopes (VLs).
Methods: This study examined 52 hospital locations within a single, large university hospital, with most of those locations being hospital operating rooms. A total of 34 600 endotracheal intubations performed over 24 months, of which 11 345 were video laryngoscopies. Electronic medical records containing demographic data and information related to endotracheal intubation procedures, with monthly breakdowns between GlideScope and McGRATH intubations, were reviewed. Cost information calculated for equipment, blades, batteries, repairs, and subsequent analysis performed to determine cost differences between those 2 instruments during the COVID-19 period.
Results: A total of 5501 video laryngoscopy procedures were performed using the McGRATH VL and 5305 were performed using the GlideScope VL. Costs over 24 months were $181 093 lower (55.5%) for McGRATH compared to GlideScope. The mean (SD) monthly costs for GlideScope blades were $3837 ($1050) and $3236 ($538) for years 1 and 2, respectively, vs $1652 ($663) and $2933 ($585) for McGRATH blades (P < .001). Most total cost differences were attributed to equipment and blade purchases, which were $202 595 (65.0%) higher for GlideScope. During the COVID-19 period, the use of the McGRATH increased to 61% of all video laryngoscopy cases, compared to 37% for GlideScope (P < .001). Blade cost difference for the COVID-19 period was $128 higher for the McGRATH even though 293 more intubations were performed with that device.
Conclusions: Use of the McGRATH resulted in a cost savings of 55% compared to the GlideScope, and its use was highest during the COVID-19 period, which may be explained by its more portable and practical features.
Keywords: video laryngoscope; McGRATH; GlideScope; endotracheal intubation; hospital costs; COVID-19.
Hospitals have come to rely on video laryngoscopes (VLs) for tracheal intubation as necessary tools for better visualization of airways. Modern video laryngoscopy developed in the 2000s1 as a progression from direct laryngoscopy, which began in 1852 when Horace Green used a bent tongue spatula and sunlight to examine a child.2 VLs have seen many improvements and adaptations of their own, resulting in many different styles and types circulating around hospitals. The GlideScope (Verathon Inc, Bothell, WA) and the McGRATH (Medtronic, Minneapolis, MN) are examples of such instruments, which are now widely used in the US and are the 2 VLs of choice at our institution.
A few studies have compared VLs to direct laryngoscopes. In their systematic review, Lewis et al have shown the numerous benefits of using a VL over a direct laryngoscope. Some general conclusions were that the use of video laryngoscopy reduced the number of failed intubations, decreased laryngeal trauma, and provided improved visualizations.3 Other studies have compared the different types of VLs, including the McGRATH and the GlideScope, examining factors such as intubation time and display quality of the image. Two studies found that medical students were equally successful at using both the McGRATH and the GlideScope,4,5 while another study found that care providers using the GlideScope had quicker intubation times.6 Lastly, Savoldelli et al concluded that more providers preferred the McGRATH, which provided better laryngeal views,7 while their subsequent study showed more favorable learning curves of the Airtraq compared to the McGRATH and other VLs.8
Although there have been no reported differences in safety and effectiveness of the McGRATH and GlideScope devices, cost data on the use of these 2 popular laryngoscopes are lacking. Such information is important considering the increasing costs of medical technologies and the significant financial losses experienced by health care systems due to the COVID-19 crisis. The purpose of this retrospective cohort study was to compare the cost efficiency of the McGRATH MAC and GlideScope Core VLs at a large academic center.
Methods
This retrospective study was performed under exemption from the Thomas Jefferson University Institutional Review Board. The primary data sources consisted of hospital electronic patient records (EPIC) and cost information from the device manufacturers and hospital staff. The electronic patient data were provided by the EPIC Enterprise Analytics Business Intelligence group at Thomas Jefferson University Hospital (Center City Campus, Philadelphia, PA), while device costs were obtained from Verathon, Medtronic, and departmental staff responsible for purchasing equipment. Monthly data were obtained over a 24-month period (June 2018 through May 2020) when the McGRATH VL was placed into use in the department of anesthesiology. The 2 types of VLs were made available for use in a total of 52 locations, with the majority being hospital operating rooms.
The following variables were recorded: number of endotracheal intubations performed each month with breakdown between video laryngoscopy and flexible bronchoscopy airways, frequency of use for each type of laryngoscope, blades used, and equipment costs for use of each laryngoscope. Hospital cost estimates for both the McGRATH and GlideScope laryngoscopes included batteries, handles, blades, and the devices themselves. Cost data were also collected on frequency of device failure, maintenance, and replacement of parts and lost equipment.
Analysis
De-identified electronic medical records consisted of nominal and quantitative variables, with demographic data and information related to the endotracheal intubation procedure. All data were in chronological order and sorted by date after which coding was applied, to identify device type and allocate pertinent cost information. Descriptive statistics were reported as mean (SD) and sum for costs; frequency tables were generated for intubation procedures according to device type and time periods. Data were analyzed using the χ2 test, the student t test, and the Wilcoxon Mann-Whitney U test, with a P value set at .05 for statistical significance. SPSS version 26 and GraphPad Prism version 6 were used for all statistical analyses.
Results
A total of 34 600 endotracheal intubations were performed over the 24-month study period, and 11 345 (32.8%) were video laryngoscopy procedures. Out of all video laryngoscopy procedures, 5501 (48.5%) were performed using the McGRATH VL and 5305 (46.8%) were conducted using the GlideScope VL. The difference of 539 (4.8%) cases accounts for flexible bronchoscopy procedures and endotracheal intubations using other video laryngoscopy equipment. The mean (SD) monthly number of video laryngoscopy procedures for the 24 months was 221 (54) and 229 (89) for the GlideScope and McGRATH devices, respectively. Monthly endotracheal intubation distributions over 24 months trended upward for the McGRATH VL and downward for the GlideScope, but there was no statistically significant (P = .71) difference in overall use between the 2 instruments (Figure 1).
To examine the observed usage trends between the 2 VL during the first and last 12 months, a univariate ANOVA was conducted with the 2 time periods entered as predictors in the model. Video laryngoscopy intubations were performed (P = .001) more frequently with the GlideScope during the first 12 months; however, use of the McGRATH VL increased (P < .001) during the following 12 months compared to GlideScope. The GlideScope accounted for 54% of all VL intubations during the first 12 months, with the McGRATH accounting for 58% of all video laryngoscopy procedures for months 12 to 24. Additionally, the increase in video laryngoscopy procedures with the McGRATH during the last 3 months of the study period was despite an overall reduction in surgical volume due to the COVID-19 crisis, defined for this study as March 1, 2020, to May 31, 2020 (Figure 1). There was a statistically significant (P < .001) difference in the case distribution between use of the McGRATH and GlideScope VL for that period. The anesthesia personnel’s use of the McGRATH VL increased to 61% of all video laryngoscopy cases, compared to 37% for the GlideScope (Figure 2).
The total costs calculated for equipment, blades, and repairs are presented in Table 1 and yearly total costs are shown in Figure 3. Overall costs were $181 093 lower (55.5%) for the McGRATH VL compared to the GlideScope over the 24-month period. The mean (SD) monthly costs for GlideScope VL blades were $3837 ($1050) and $3236 ($538) for years 1 and 2, respectively, vs $1652 ($663) and $2933 ($585) for the McGRATH VL blades. Most of the total cost differences were attributed to equipment and blade purchases, which were $202 595 (65.0%) higher for the GlideScope compared to the McGRATH VL. The monthly blade costs alone were higher (P < .001) for the GlideScope over the 2-year period; however, the McGRATH VL required use of disposable stylets at a cost of $10 177 for all endotracheal intubations, compared to $700 for the GlideScope device.
An analysis was performed to determine whether costs differed between those 2 instruments during the COVID-19 period. There was a statistically significant (P < .001) difference in the case distribution between use of the McGRATH and GlideScope VLs during that period. The calculated blade cost difference for the COVID period was $128 higher for the McGRATH even though 293 more intubations were performed with that device (Table 2).
Discussion
We attempted to provide useful cost estimates by presenting pricing data reflecting the approximate cost that most large institutional anesthesia practices would incur for using those 2 specific devices and related peripherals. The main findings of our analysis showed that use of the McGRATH MAC VL resulted in a 55% cost savings compared to the GlideScope, with a similar number of cases performed with each device over the 24-month study period. We believe this represents a substantial savings to the department and institution, which has prompted internal review on the use of video laryngoscopy equipment. None of the McGRATH units failed; however, the GlideScope required 3 baton replacements.
Of note, use of the McGRATH MAC increased during the COVID-19 period, which may be explained by the fact that the operators found it to be a more portable device. Several physicians in the department commented that its smaller size made the McGRATH MAC more practical during the time when a plexiglass box was being used around the patient’s head to shield the intubator from aerosolized viral particles.
Although this study demonstrated the cost-saving value of the McGRATH over the GlideScope, a suggested next step would be to examine resource utilization related to video laryngoscopy use. The more dynamic tracking of the use of these devices should facilitate the assessment of existing related resources and decision making, to optimize the benefits of this initiative. We would anticipate reduced use of anesthesia personnel, such as technicians to assist with the management of this device which could be significant. As new respiratory viruses are appearing each year, video laryngoscopy will continue to gain increasing use in operating rooms and acute care locations. The adding of protective barriers between patients and providers calls for use of the most practical and effective VL devices, to protect personnel who are at high risk of contamination from airway secretions and aerosolized particles.9,10
The COVID-19 pandemic has demonstrated the value of anesthesiology in regards to analyzing and finding solutions to effectively manage infected patients or those suspected of infection in the perioperative environment. Inexpensive products are often avoided because cheaper devices are associated with being of lower quality. However, the association with cost and quality—and the assumption that a higher price is positively correlated with higher quality—is overall inconsistent in the medical literature.11 A more effective or higher quality treatment does not necessarily cost more and may actually end up costing less,12 as was the case in this study. We have been able to directly cut departmental expenses by using a more efficient and cost-effective device for intubations, without compromising safety and efficacy. Future studies should determine whether this significant reduction in costs from video laryngoscopy intubations with the McGRATH VL will be sustained across anesthesiology departments in the Jefferson Health Enterprise Hospitals, or other health systems, as well as its impact on workflow and personnel resources.
This analysis was restricted to one of the campuses of the Jefferson Health Enterprise. However, this is the largest anesthesia practice, encompassing several locations, which should reflect the general practice patterns across other anesthesiology departments in this large institution. The costs for the devices and peripherals may vary across anesthesia practices depending on volume and contracts negotiated with the suppliers. It was not possible to estimate this variability, which could change the total costs by a few percentage points. We recognize that there may be other costs associated with securing the McGRATH VL to prevent loss from theft or misplacement, which were not included in the study. Lastly, the inability to obtain randomized samples for the 2 groups treated with each device opens up the possibility of selection bias. There were, however, multiple intubators who were free to select 1 of the devices for endotracheal intubation, which may have reduced the effect of selection bias.
Conclusion
This study demonstrated that over a 24-month period use of the McGRATH MAC VL resulted in a cost reduction of around 55% compared to using the GlideScope for endotracheal intubation procedures performed at a major academic center. Over the first 3 months of the COVID-19 crisis, which our study included, use of the McGRATH VL increased while GlideScope use decreased. This was most likely related to the portability and smaller size of the McGRATH, which better facilitated intubations of COVID-19 patients.
Acknowledgements: The authors thank Craig Smith, Senior Anesthesia Technician, for his assistance with the cost information and excellent record-keeping related to the use of video laryngoscopes.
Corresponding author: Marc C. Torjman, PhD, Professor, Department of Anesthesiology, Sidney Kimmel Medical College at Thomas Jefferson University, 111 South 11th St, Suite G-8290, Philadelphia, PA 19107; Marc.Torjman@Jefferson.edu.
Financial disclosures: Dr. Thaler has served as a consultant for Medtronic since September 2020. He has participated in 2 webinars on the routine use of video laryngoscopy.
Funding: This study was supported by the Department of Anesthesiology at Thomas Jefferson University.
1. Channa AB. Video laryngoscopes. Saudi J Anaesth. 2011;5(4):357-359.
2. Pieters BM, Eindhoven GB, Acott C, Van Zundert AAJ. Pioneers of laryngoscopy: indirect, direct and video laryngoscopy. Anaesth Intensive Care. 2015;43(suppl):4-11.
3. Lewis SR, Butler AR, Parker J, et al. Videolaryngoscopy versus direct laryngoscopy for adult patients requiring tracheal intubation. Cochrane Database Syst Rev. 2016;11(11):CD011136.
4. Kim W, Choi HJ, Lim T, Kang BS. Can the new McGrath laryngoscope rival the GlideScope Ranger portable video laryngoscope? A randomized manikin study. Am J Emerg Med. 2014;32(10):1225-1229.
5. Kim W, Choi HJ, Lim T, et al. Is McGrath MAC better than Glidescope Ranger for novice providers in the simulated difficult airway? A randomized manikin study. Resuscitation. 2014;85(suppl 1):S32.
6. Jeon WJ, Kim KH, Yeom JH, et al. A comparison of the Glidescope to the McGrath videolaryngoscope in patients. Korean J Anesthesiol. 2011;61(1):19-23.
7. Savoldelli GL, Schiffer E, Abegg C, et al. Comparison of the Glidescope, the McGrath, the Airtraq and the Macintosh laryngoscopes in simulated difficult airways. Anaesthesia. 2008;63(12):1358-1364.
8. Savoldelli GL, Schiffer E, Abegg C, et al. Learning curves of the Glidescope, the McGrath and the Airtraq laryngoscopes: a manikin study. Eur J Anaesthesiol. 2009;26(7):554-558.
9. Schumacher J, Arlidge J, Dudley D, et al. The impact of respiratory protective equipment on difficult airway management: a randomised, crossover, simulation study. Anaesthesia. 2020;75(10):1301-1306.
10. De Jong A, Pardo E, Rolle A, et al. Airway management for COVID-19: a move towards universal videolaryngoscope? Lancet Respir Med. 2020;8(6):555.
11. Hussey PS, Wertheimer S, Mehrotra A. The association between health care quality and cost: a systematic review. Ann Intern Med. 2013;158(1):27-34.
12. Mitton C, Dionne F, Peacock S, Sheps S. Quality and cost in healthcare: a relationship worth examining. Appl Health Econ Health Policy. 2006;5(4):201-208.
1. Channa AB. Video laryngoscopes. Saudi J Anaesth. 2011;5(4):357-359.
2. Pieters BM, Eindhoven GB, Acott C, Van Zundert AAJ. Pioneers of laryngoscopy: indirect, direct and video laryngoscopy. Anaesth Intensive Care. 2015;43(suppl):4-11.
3. Lewis SR, Butler AR, Parker J, et al. Videolaryngoscopy versus direct laryngoscopy for adult patients requiring tracheal intubation. Cochrane Database Syst Rev. 2016;11(11):CD011136.
4. Kim W, Choi HJ, Lim T, Kang BS. Can the new McGrath laryngoscope rival the GlideScope Ranger portable video laryngoscope? A randomized manikin study. Am J Emerg Med. 2014;32(10):1225-1229.
5. Kim W, Choi HJ, Lim T, et al. Is McGrath MAC better than Glidescope Ranger for novice providers in the simulated difficult airway? A randomized manikin study. Resuscitation. 2014;85(suppl 1):S32.
6. Jeon WJ, Kim KH, Yeom JH, et al. A comparison of the Glidescope to the McGrath videolaryngoscope in patients. Korean J Anesthesiol. 2011;61(1):19-23.
7. Savoldelli GL, Schiffer E, Abegg C, et al. Comparison of the Glidescope, the McGrath, the Airtraq and the Macintosh laryngoscopes in simulated difficult airways. Anaesthesia. 2008;63(12):1358-1364.
8. Savoldelli GL, Schiffer E, Abegg C, et al. Learning curves of the Glidescope, the McGrath and the Airtraq laryngoscopes: a manikin study. Eur J Anaesthesiol. 2009;26(7):554-558.
9. Schumacher J, Arlidge J, Dudley D, et al. The impact of respiratory protective equipment on difficult airway management: a randomised, crossover, simulation study. Anaesthesia. 2020;75(10):1301-1306.
10. De Jong A, Pardo E, Rolle A, et al. Airway management for COVID-19: a move towards universal videolaryngoscope? Lancet Respir Med. 2020;8(6):555.
11. Hussey PS, Wertheimer S, Mehrotra A. The association between health care quality and cost: a systematic review. Ann Intern Med. 2013;158(1):27-34.
12. Mitton C, Dionne F, Peacock S, Sheps S. Quality and cost in healthcare: a relationship worth examining. Appl Health Econ Health Policy. 2006;5(4):201-208.
Phototherapy: Safe and Effective for Challenging Skin Conditions in Older Adults
Identifying safe, effective, and affordable evidence-based dermatologic treatments for older adults can be challenging because of age-related changes in the skin, comorbidities, polypharmacy, mobility issues, and cognitive changes. Phototherapy has been shown to be an effective nonpharmacologic treatment option for multiple challenging dermatologic conditions1-8; however, few studies have specifically examined its effectiveness in older adults. The challenge for older patients with psoriasis and dermatitis is that the conditions can be difficult to control and often require multiple treatment modalities.9,10 Patients with psoriasis also have a higher risk for diabetes, dyslipidemia, and cardiovascular disease compared to other older patients,11,12 which poses treatment challenges and makes nonpharmacologic treatments even more appealing.
Recent studies show that phototherapy can help decrease the use of dermatologic medications. Foerster and colleagues2 found that adults with psoriasis who were treated with phototherapy significantly decreased their use of topical steroids (24.5% fewer patients required steroid creams and 31.1% fewer patients required psoriasis-specific topicals)(P<.01) while their use of non–psoriasis-specific medications did not change. Click and colleagues13 identified a decrease in medication costs, health care utilization, and risk for immunosuppression in patients treated with phototherapy when compared to those treated with biologics and apremilast. Methotrexate is a common dermatologic medication that is highly associated with increased risks in elderly patients because of impaired immune system function and the presence of comorbidities (eg, kidney disease, obesity, diabetes, fatty liver),14 which increase in prevalence with age. Combining phototherapy with methotrexate can substantially decrease the amount of methotrexate needed to achieve disease control,15 thereby decreasing the methotrexate-associated risks. Findings from these studies suggest that a safe, effective, cost-effective, and well-tolerated nonpharmacologic alternative, such as phototherapy, is highly desirable and should be optimized. Unfortunately, most studies that report the effectiveness of phototherapy are in younger populations.
This retrospective study aimed to (1) identify the most common dermatologic conditions treated with phototherapy in older adults, (2) examine the effectiveness and safety of phototherapy in older adults
Methods
Design, Setting, Sample, and Statistical Analysis
The institutional review boards of Kaiser Permanente Washington Health Research Institute, Seattle, and the University of Washington, Seattle, approved this study. It was conducted in a large US multispecialty health care system (Group Health, Seattle, Washington [now Kaiser Permanente Washington]) serving approximately 600,000 patients, using billing records to identify all patients treated with phototherapy between January 1, 2015, and December 31, 2015, all who received narrowband UVB (NB-UVB) phototherapy. All adults 65 years and older who received phototherapy treatment during the 12-month study period were included. Patients were included regardless of comorbidities and other dermatologic treatments to maintain as much uniformity as possible between the present study and 2 prior studies examining phototherapy in older adult populations in the United Kingdom16 and Turkey.17 Demographic and clinical factors were presented using frequencies (percentages) or means and medians as appropriate. Comparisons of dermatologic conditions and clearance levels used a Fisher exact test. The number of phototherapy treatments to clearance and total number of treatments were compared between groups of patients using independent sample t tests.
Phototherapy Protocol
All patients received treatments administered by specially trained phototherapy nurses using a Daavlin UV Series (The Daavlin Company) or an Ultralite unit (Ultralite Enterprises, Inc), both with 48 lamps. All phototherapy nurses had been previously trained to provide treatments based on standardized protocols (Table 1) and to determine the patient’s level of disease clearance using a high to low clearance scale (Table 2). Daavlin’s treatment protocols were built into the software that accompanied the units and were developed based on the American Academy of Dermatology guidelines. The starting dose for an individual patient was determined based on the estimated
Results
Patients
Billing records identified 229 total patients who received phototherapy in 2015, of whom 52 (22.7%) were at least 65 years old. The median age was 70 years (range, 65–91 years). Twenty-nine (56%) were men and 35 (67%) had previously received phototherapy treatments.
Dermatologic Conditions Treated With Phototherapy
Our primary aim was to identify the most common dermatologic conditions treated with phototherapy in older adults. Psoriasis and dermatitis were the most common conditions treated in the sample (50% [26/52] and 21% [11/52], respectively), with mycosis fungoides being the third most common (10% [5/52]) and vitiligo tied with prurigo nodularis as fourth most common (6% [3/52])(Figure 1).
Effectiveness and Safety of Phototherapy
Our secondary aim was to examine the effectiveness and safety of phototherapy in older adults. Phototherapy was effective in this population, with 50 of 52 patients (96%) achieving a high or medium level of clearance. The degree of clearance for each of the dermatologic conditions is shown in Figure 2. Psoriasis and dermatitis achieved high clearance rates in 81% (21/26) and 82% (9/11) of patients, respectively. Overall, conditions did not have significant differences in clearances rates (Fisher exact test, P=.10). On average, it took patients 33 treatments to achieve medium or high rates of clearance. Psoriasis cleared more quickly, with an average of 30.4 treatments vs 36.1 treatments for other conditions, but the difference was not significant (t test, P=.26). Patients received an average of 98 total phototherapy treatments; the median number of treatments was 81 due to many being on maintenance therapy over several months. There was no relationship between a history of treatment with phototherapy and the total number of treatments needed to achieve clearance (t test, P=.40), but interestingly, those who had a history of phototherapy took approximately 5 more treatments to achieve clearance. The present study found that a slightly larger number of men were being treated for psoriasis (15 men vs 11 women), but there was no significant difference in response rate based on gender.
Side effects from phototherapy were minimal; 24 patients (46%) experienced grade 1 (mild) erythema at some point during their treatment course. Thirteen (25%) patients experienced grade 2 erythema, but this was a rare event for most patients. Only 1 (2%) patient experienced grade 3 erythema 1 time. Three patients experienced increased itching (6%). Thirteen (25%) patients had no side effects. None developed severe erythema or blisters, and none discontinued phototherapy because of side effects. Over the course of the study year, we found a high degree of acceptance of phototherapy treatments by older patients: 22 (42%) completed therapy after achieving clearance, 10 (19%) were continuing ongoing treatments (maintenance), and 15 (29%) stopped because of life circumstances (eg, other health issues, moving out of the area). Only 4 (8%) stopped because of a lack of effectiveness, and 1 (2%) patient because the treatments were burdensome.
Comparison of Outcomes
Our third aim was to compare the outcomes with similar studies in the United Kingdom16 and Turkey.17 This study confirmed that phototherapy is being used in older adults (22.7% of this study’s total patients) and is an effective treatment for older patients experiencing a range of challenging inflammatory and proliferative skin diseases similar to studies in the general population. Prior phototherapy studies in elderly patients also found psoriasis to be the most common skin condition treated, with 1 study finding that 51% (19/37) of older phototherapy patients had psoriasis,16 while another reported 58% (37/95) of older phototherapy patients had psoriasis.17 These numbers are similar to those in our study, which showed 50% (26/52) of elderly phototherapy patients had psoriasis. Psoriasis is the main indication for treatment with NB-UVB phototherapy in the general population,19 and because the risk for psoriasis increases with age,20 it is not surprising that all 3 studies found psoriasis to be the most common indication in elderly phototherapy patients. Table 3 provides further details on conditions treated in all 3 studies.
Comment
Our study found that 94% of patients with psoriasis achieved clearance with an average of 30.4 treatments, which is comparable to the reported 91% response rate with an average of 30 treatments in the United Kingdom.16 The other similar study in Turkey17 reported 73.7% of psoriasis patients achieved a 75% or more improvement from baseline with an average of 42 treatments, which may reflect underlying differences in regional skin type. Of note, the scatter chart (Figure 3) shows that several patients in the present study’s analysis are listed as not clear, but many of those patients had low treatment numbers below the mean time to clearance. Thus, the present study’s response rate may have been underestimated.
In the general population, studies show that psoriasis treated with standardized phototherapy protocols typically clears with an average of 20.6 treatments.21 The levels of clearance were similar in our study’s older population, but more treatments were required to achieve those results, with an average of 10 more treatments needed (an additional 3.3 weeks). Similar results were found in this sample for dermatitis and mycosis fungoides, indicating comparable clearance rates and levels but a need for more treatments to achieve similar results compared to the general population.
Additionally, in the current study more patients experienced grade 1 (mild) erythema (46%) and grade 2 erythema (25%) at some point in their treatment compared with the United Kingdom16 (1.89%) and Turkey17 (35%) studies, though these side effects did not impact the clearance rate. Interestingly, the current study’s scatter chart (Figure 3) illustrates that this side effect did not seem to increase with aging in this population. If anything, the erythema response was more prevalent in the median or younger patients in the sample. Erythema may have been due to the frequent use of photosensitizing medications in older adults in the United States, some of which typically get discontinued in patients 75 years and older (eg, statins). Other potential causes might include the use of phototype vs minimal erythema dose–driven protocols, the standard utilization of protocols originally designed for psoriasis vs other condition-specific protocols, missed treatments leading to increased sensitivity, or possibly shielding mishaps (eg, not wearing a prescribed face shield). Given the number of potential causes and the possibility of overlapping factors, careful analysis is important. With NB-UVB phototherapy, near-erythemogenic doses are optimal to achieve effective treatments, but this delicate balance may be more problematic for older adults. Future studies are needed to fully determine the factors at play for this population. In the interim, it is important for phototherapy-trained nurses to consider this risk carefully in the older population. They must follow the prescribed protocols that guide them to query patients about their responses to the prior treatment (eg, erythema, tenderness, itching), photosensitizing medications, missed treatments, and placement of shielding, and then adjust the treatment dosing accordingly.
Limitations
This study had several limitations. Although clinical outcomes were recorded prospectively, the analysis was retrospective, unblinded, and not placebo controlled. It was conducted in a single organization (Group Health [now Kaiser Permanente Washington]) but did analyze data from 4 medical centers in different cities with diverse demographics and a variety of nursing staff providing the treatments. Although the vitiligo treatment protocol likely slowed the response rate for those patients with vitiligo, the numbers were small (ie, only 3 of 52 patients), so the researchers chose to include them in the current study. The sample population was relatively small, but when these data are evaluated alongside the studies in the United Kingdom16 and Turkey,17 they show a consistent picture illustrating the effectiveness and safety of phototherapy in the older population. Further epidemiologic studies could be helpful to further describe the usefulness of this modality compared with other treatments for a variety of dermatoses in this age group. Supplementary analysis specifically examining the relationship between the number and type of photosensitizing medications, frequency of erythema, and time to clearance also could be useful.
Conclusion
Older adults with a variety of dermatoses respond well to phototherapy and should have the opportunity to use it, particularly considering the potential for increased complications and costs from other treatment modalities, such as commonly used immunosuppressive pharmaceuticals. However, the current study and the comparison studies indicate that it is important to carefully consider the slower clearance rates and the potential risk for increased erythema in this population and adjust patient education and treatment dosing accordingly.
Unfortunately, many dermatology centers do not offer phototherapy because of infrastructure limitations such as space and specially trained nursing staff. Increasing accessibility of phototherapy for older adults through home treatments may be an alternative, given its effectiveness in the general population.22,23 In addition, home phototherapy may be worth pursuing for the older population considering the challenges they may face with transportation to the clinic setting and their increased risk for serious illness if exposed to infections such as COVID-19. The COVID-19 pandemic has brought to light the need for reliable, safe, and effective treatments that can be utilized in the safety of patients’ homes and should therefore be considered as an option for older adults. Issues such as mobility and cognitive decline could pose some complicating factors, but with the help of a well-trained family member or caregiver, home phototherapy could be a viable option that improves accessibility for older patients. Future research opportunities include further examination of the slower but ultimately equivalent response to phototherapy in the older population, the influence of photosensitizing medications on phototherapy effects, and the impact of phototherapy on utilization of immunosuppressive pharmaceuticals in older adults.
- British Photodermatology Group. An appraisal of narrowband (TL-01) UVB phototherapy. British Photodermatology Group Workshop Report (April 1996). Br J Dermatol. 1997;137:327-330.
Foerster J, Boswell K, West J, et al. Narrowband UVB treatment is highly effective and causes a strong reduction in the use of steroid and other creams in psoriasis patients in clinical practice. PLoS ONE. 2017;12:e0181813. doi:10.1371/journal.pone.0181813 - Fernández-Guarino M, Aboin-Gonzalez S, Barchino L, et al. Treatment of moderate and severe adult chronic atopic dermatitis with narrow-band UVB and the combination of narrow-band UVB/UVA phototherapy. Dermatol Ther. 2015;29:19-23.
- Ryu HH, Choe YS, Jo S, et al. Remission period in psoriasis after multiple cycles of narrowband ultraviolet B phototherapy. J Dermatol. 2014;41:622-627.
Tintle S, Shemer A, Suárez-Fariñas M, et al. Reversal of atopic dermatitis with narrow-band UVB phototherapy and biomarkers for therapeutic response. J Allergy Clin Immunol. 2011;128:583-593. - Gambichler T, Breuckmann F, Boms S, et al. Narrowband UVB phototherapy in skin conditions beyond psoriasis. J Am Acad Dermatol. 2005;52:660-670.
- Schneider LA, Hinrichs R, Scharffetter-Kochanek K. Phototherapy and photochemotherapy. Clin Dermatol. 2008;26:464-476.
- Martin JA, Laube S, Edwards C, et al. Rate of acute adverse events for narrow-band UVB and psoralen-UVA phototherapy. Photodermatol Photoimmunol Photomed. 2007;23:68-72.
- Mokos ZB, Jovic A, Ceovic R, et al. Therapeutic challenges in the mature patient. Clin Dermatol. 2018;36:128-139.
- Di Lernia V, Goldust M. An overview of the efficacy and safety of systemic treatments for psoriasis in the elderly. Exp Opin Biol Ther. 2018;18:897-903.
- Napolitano M, Balato N, Ayala F, et al. Psoriasis in elderly and non-elderly population: clinical and molecular features. G Ital Dermatol Venereol. 2016;151:587-595.
- Grozdev IS, Van Voorhees AS, Gottlieb AB, et al. Psoriasis in the elderly: from the Medical Board of the National Psoriasis Foundation. J Am Acad Dermatol. 2011;65:537-545.
- Click J, Alabaster A, Postlethwaite D, et al. Effect of availability of at-home phototherapy on the use of systemic medications for psoriasis.
Photodermatol Photoimmunol Photomed. 2017;33:345-346. - Piaserico S, Conti A, Lo Console F, et al.
Efficacy and safety of systemic treatments for psoriasis in elderly. Acta Derm Venereol. 2014;94:293-297. - Soliman A, Nofal E, Nofal A, et al. Combination therapy of methotrexate plus NB-UVB phototherapy is more effective than methotrexate monotherapy in the treatment of chronic plaque psoriasis. J Dermatol Treat. 2015;26:528-534.
- Powell JB, Gach JE. Phototherapy in the elderly. Clin Exp Dermatol. 2015;40:605-610.
- Bulur I, Erdogan HK, Aksu AE, et al. The efficacy and safety of phototherapy in geriatric patients: a retrospective study. An Bras Dermatol. 2018;93:33-38.
- Madigan LM, Al-Jamal M, Hamzavi I. Exploring the gaps in the evidence-based application of narrowband UVB for the treatment of vitiligo. Photodermatol Photoimmunol Photomed. 2016;32:66-80.
- Ibbotson SH. A perspective on the use of NB-UVB phototherapy vs. PUVA photochemotherapy. Front Med (Lausanne). 2018;5:184.
- Bell LM, Sedlack R, Beard CM, et al. Incidence of psoriasis in Rochester, Minn, 1980-1983. Arch Dermatol. 1991;127:1184-1187.
- Totonchy MB, Chiu MW. UV-based therapy. Dermatol Clin. 2014;32:399-413.
- Cameron H, Yule S, Dawe RS, et al. Review of an established UK home phototherapy service 1998-2011: improving access to a cost-effective treatment for chronic skin disease. Public Health. 2014;128:317-324.
- Matthews SW, Simmer M, Williams L, et al. Transition of patients with psoriasis from office-based phototherapy to nurse-supported home phototherapy: a pilot study. JDNA. 2018;10:29-41.
Identifying safe, effective, and affordable evidence-based dermatologic treatments for older adults can be challenging because of age-related changes in the skin, comorbidities, polypharmacy, mobility issues, and cognitive changes. Phototherapy has been shown to be an effective nonpharmacologic treatment option for multiple challenging dermatologic conditions1-8; however, few studies have specifically examined its effectiveness in older adults. The challenge for older patients with psoriasis and dermatitis is that the conditions can be difficult to control and often require multiple treatment modalities.9,10 Patients with psoriasis also have a higher risk for diabetes, dyslipidemia, and cardiovascular disease compared to other older patients,11,12 which poses treatment challenges and makes nonpharmacologic treatments even more appealing.
Recent studies show that phototherapy can help decrease the use of dermatologic medications. Foerster and colleagues2 found that adults with psoriasis who were treated with phototherapy significantly decreased their use of topical steroids (24.5% fewer patients required steroid creams and 31.1% fewer patients required psoriasis-specific topicals)(P<.01) while their use of non–psoriasis-specific medications did not change. Click and colleagues13 identified a decrease in medication costs, health care utilization, and risk for immunosuppression in patients treated with phototherapy when compared to those treated with biologics and apremilast. Methotrexate is a common dermatologic medication that is highly associated with increased risks in elderly patients because of impaired immune system function and the presence of comorbidities (eg, kidney disease, obesity, diabetes, fatty liver),14 which increase in prevalence with age. Combining phototherapy with methotrexate can substantially decrease the amount of methotrexate needed to achieve disease control,15 thereby decreasing the methotrexate-associated risks. Findings from these studies suggest that a safe, effective, cost-effective, and well-tolerated nonpharmacologic alternative, such as phototherapy, is highly desirable and should be optimized. Unfortunately, most studies that report the effectiveness of phototherapy are in younger populations.
This retrospective study aimed to (1) identify the most common dermatologic conditions treated with phototherapy in older adults, (2) examine the effectiveness and safety of phototherapy in older adults
Methods
Design, Setting, Sample, and Statistical Analysis
The institutional review boards of Kaiser Permanente Washington Health Research Institute, Seattle, and the University of Washington, Seattle, approved this study. It was conducted in a large US multispecialty health care system (Group Health, Seattle, Washington [now Kaiser Permanente Washington]) serving approximately 600,000 patients, using billing records to identify all patients treated with phototherapy between January 1, 2015, and December 31, 2015, all who received narrowband UVB (NB-UVB) phototherapy. All adults 65 years and older who received phototherapy treatment during the 12-month study period were included. Patients were included regardless of comorbidities and other dermatologic treatments to maintain as much uniformity as possible between the present study and 2 prior studies examining phototherapy in older adult populations in the United Kingdom16 and Turkey.17 Demographic and clinical factors were presented using frequencies (percentages) or means and medians as appropriate. Comparisons of dermatologic conditions and clearance levels used a Fisher exact test. The number of phototherapy treatments to clearance and total number of treatments were compared between groups of patients using independent sample t tests.
Phototherapy Protocol
All patients received treatments administered by specially trained phototherapy nurses using a Daavlin UV Series (The Daavlin Company) or an Ultralite unit (Ultralite Enterprises, Inc), both with 48 lamps. All phototherapy nurses had been previously trained to provide treatments based on standardized protocols (Table 1) and to determine the patient’s level of disease clearance using a high to low clearance scale (Table 2). Daavlin’s treatment protocols were built into the software that accompanied the units and were developed based on the American Academy of Dermatology guidelines. The starting dose for an individual patient was determined based on the estimated
Results
Patients
Billing records identified 229 total patients who received phototherapy in 2015, of whom 52 (22.7%) were at least 65 years old. The median age was 70 years (range, 65–91 years). Twenty-nine (56%) were men and 35 (67%) had previously received phototherapy treatments.
Dermatologic Conditions Treated With Phototherapy
Our primary aim was to identify the most common dermatologic conditions treated with phototherapy in older adults. Psoriasis and dermatitis were the most common conditions treated in the sample (50% [26/52] and 21% [11/52], respectively), with mycosis fungoides being the third most common (10% [5/52]) and vitiligo tied with prurigo nodularis as fourth most common (6% [3/52])(Figure 1).
Effectiveness and Safety of Phototherapy
Our secondary aim was to examine the effectiveness and safety of phototherapy in older adults. Phototherapy was effective in this population, with 50 of 52 patients (96%) achieving a high or medium level of clearance. The degree of clearance for each of the dermatologic conditions is shown in Figure 2. Psoriasis and dermatitis achieved high clearance rates in 81% (21/26) and 82% (9/11) of patients, respectively. Overall, conditions did not have significant differences in clearances rates (Fisher exact test, P=.10). On average, it took patients 33 treatments to achieve medium or high rates of clearance. Psoriasis cleared more quickly, with an average of 30.4 treatments vs 36.1 treatments for other conditions, but the difference was not significant (t test, P=.26). Patients received an average of 98 total phototherapy treatments; the median number of treatments was 81 due to many being on maintenance therapy over several months. There was no relationship between a history of treatment with phototherapy and the total number of treatments needed to achieve clearance (t test, P=.40), but interestingly, those who had a history of phototherapy took approximately 5 more treatments to achieve clearance. The present study found that a slightly larger number of men were being treated for psoriasis (15 men vs 11 women), but there was no significant difference in response rate based on gender.
Side effects from phototherapy were minimal; 24 patients (46%) experienced grade 1 (mild) erythema at some point during their treatment course. Thirteen (25%) patients experienced grade 2 erythema, but this was a rare event for most patients. Only 1 (2%) patient experienced grade 3 erythema 1 time. Three patients experienced increased itching (6%). Thirteen (25%) patients had no side effects. None developed severe erythema or blisters, and none discontinued phototherapy because of side effects. Over the course of the study year, we found a high degree of acceptance of phototherapy treatments by older patients: 22 (42%) completed therapy after achieving clearance, 10 (19%) were continuing ongoing treatments (maintenance), and 15 (29%) stopped because of life circumstances (eg, other health issues, moving out of the area). Only 4 (8%) stopped because of a lack of effectiveness, and 1 (2%) patient because the treatments were burdensome.
Comparison of Outcomes
Our third aim was to compare the outcomes with similar studies in the United Kingdom16 and Turkey.17 This study confirmed that phototherapy is being used in older adults (22.7% of this study’s total patients) and is an effective treatment for older patients experiencing a range of challenging inflammatory and proliferative skin diseases similar to studies in the general population. Prior phototherapy studies in elderly patients also found psoriasis to be the most common skin condition treated, with 1 study finding that 51% (19/37) of older phototherapy patients had psoriasis,16 while another reported 58% (37/95) of older phototherapy patients had psoriasis.17 These numbers are similar to those in our study, which showed 50% (26/52) of elderly phototherapy patients had psoriasis. Psoriasis is the main indication for treatment with NB-UVB phototherapy in the general population,19 and because the risk for psoriasis increases with age,20 it is not surprising that all 3 studies found psoriasis to be the most common indication in elderly phototherapy patients. Table 3 provides further details on conditions treated in all 3 studies.
Comment
Our study found that 94% of patients with psoriasis achieved clearance with an average of 30.4 treatments, which is comparable to the reported 91% response rate with an average of 30 treatments in the United Kingdom.16 The other similar study in Turkey17 reported 73.7% of psoriasis patients achieved a 75% or more improvement from baseline with an average of 42 treatments, which may reflect underlying differences in regional skin type. Of note, the scatter chart (Figure 3) shows that several patients in the present study’s analysis are listed as not clear, but many of those patients had low treatment numbers below the mean time to clearance. Thus, the present study’s response rate may have been underestimated.
In the general population, studies show that psoriasis treated with standardized phototherapy protocols typically clears with an average of 20.6 treatments.21 The levels of clearance were similar in our study’s older population, but more treatments were required to achieve those results, with an average of 10 more treatments needed (an additional 3.3 weeks). Similar results were found in this sample for dermatitis and mycosis fungoides, indicating comparable clearance rates and levels but a need for more treatments to achieve similar results compared to the general population.
Additionally, in the current study more patients experienced grade 1 (mild) erythema (46%) and grade 2 erythema (25%) at some point in their treatment compared with the United Kingdom16 (1.89%) and Turkey17 (35%) studies, though these side effects did not impact the clearance rate. Interestingly, the current study’s scatter chart (Figure 3) illustrates that this side effect did not seem to increase with aging in this population. If anything, the erythema response was more prevalent in the median or younger patients in the sample. Erythema may have been due to the frequent use of photosensitizing medications in older adults in the United States, some of which typically get discontinued in patients 75 years and older (eg, statins). Other potential causes might include the use of phototype vs minimal erythema dose–driven protocols, the standard utilization of protocols originally designed for psoriasis vs other condition-specific protocols, missed treatments leading to increased sensitivity, or possibly shielding mishaps (eg, not wearing a prescribed face shield). Given the number of potential causes and the possibility of overlapping factors, careful analysis is important. With NB-UVB phototherapy, near-erythemogenic doses are optimal to achieve effective treatments, but this delicate balance may be more problematic for older adults. Future studies are needed to fully determine the factors at play for this population. In the interim, it is important for phototherapy-trained nurses to consider this risk carefully in the older population. They must follow the prescribed protocols that guide them to query patients about their responses to the prior treatment (eg, erythema, tenderness, itching), photosensitizing medications, missed treatments, and placement of shielding, and then adjust the treatment dosing accordingly.
Limitations
This study had several limitations. Although clinical outcomes were recorded prospectively, the analysis was retrospective, unblinded, and not placebo controlled. It was conducted in a single organization (Group Health [now Kaiser Permanente Washington]) but did analyze data from 4 medical centers in different cities with diverse demographics and a variety of nursing staff providing the treatments. Although the vitiligo treatment protocol likely slowed the response rate for those patients with vitiligo, the numbers were small (ie, only 3 of 52 patients), so the researchers chose to include them in the current study. The sample population was relatively small, but when these data are evaluated alongside the studies in the United Kingdom16 and Turkey,17 they show a consistent picture illustrating the effectiveness and safety of phototherapy in the older population. Further epidemiologic studies could be helpful to further describe the usefulness of this modality compared with other treatments for a variety of dermatoses in this age group. Supplementary analysis specifically examining the relationship between the number and type of photosensitizing medications, frequency of erythema, and time to clearance also could be useful.
Conclusion
Older adults with a variety of dermatoses respond well to phototherapy and should have the opportunity to use it, particularly considering the potential for increased complications and costs from other treatment modalities, such as commonly used immunosuppressive pharmaceuticals. However, the current study and the comparison studies indicate that it is important to carefully consider the slower clearance rates and the potential risk for increased erythema in this population and adjust patient education and treatment dosing accordingly.
Unfortunately, many dermatology centers do not offer phototherapy because of infrastructure limitations such as space and specially trained nursing staff. Increasing accessibility of phototherapy for older adults through home treatments may be an alternative, given its effectiveness in the general population.22,23 In addition, home phototherapy may be worth pursuing for the older population considering the challenges they may face with transportation to the clinic setting and their increased risk for serious illness if exposed to infections such as COVID-19. The COVID-19 pandemic has brought to light the need for reliable, safe, and effective treatments that can be utilized in the safety of patients’ homes and should therefore be considered as an option for older adults. Issues such as mobility and cognitive decline could pose some complicating factors, but with the help of a well-trained family member or caregiver, home phototherapy could be a viable option that improves accessibility for older patients. Future research opportunities include further examination of the slower but ultimately equivalent response to phototherapy in the older population, the influence of photosensitizing medications on phototherapy effects, and the impact of phototherapy on utilization of immunosuppressive pharmaceuticals in older adults.
Identifying safe, effective, and affordable evidence-based dermatologic treatments for older adults can be challenging because of age-related changes in the skin, comorbidities, polypharmacy, mobility issues, and cognitive changes. Phototherapy has been shown to be an effective nonpharmacologic treatment option for multiple challenging dermatologic conditions1-8; however, few studies have specifically examined its effectiveness in older adults. The challenge for older patients with psoriasis and dermatitis is that the conditions can be difficult to control and often require multiple treatment modalities.9,10 Patients with psoriasis also have a higher risk for diabetes, dyslipidemia, and cardiovascular disease compared to other older patients,11,12 which poses treatment challenges and makes nonpharmacologic treatments even more appealing.
Recent studies show that phototherapy can help decrease the use of dermatologic medications. Foerster and colleagues2 found that adults with psoriasis who were treated with phototherapy significantly decreased their use of topical steroids (24.5% fewer patients required steroid creams and 31.1% fewer patients required psoriasis-specific topicals)(P<.01) while their use of non–psoriasis-specific medications did not change. Click and colleagues13 identified a decrease in medication costs, health care utilization, and risk for immunosuppression in patients treated with phototherapy when compared to those treated with biologics and apremilast. Methotrexate is a common dermatologic medication that is highly associated with increased risks in elderly patients because of impaired immune system function and the presence of comorbidities (eg, kidney disease, obesity, diabetes, fatty liver),14 which increase in prevalence with age. Combining phototherapy with methotrexate can substantially decrease the amount of methotrexate needed to achieve disease control,15 thereby decreasing the methotrexate-associated risks. Findings from these studies suggest that a safe, effective, cost-effective, and well-tolerated nonpharmacologic alternative, such as phototherapy, is highly desirable and should be optimized. Unfortunately, most studies that report the effectiveness of phototherapy are in younger populations.
This retrospective study aimed to (1) identify the most common dermatologic conditions treated with phototherapy in older adults, (2) examine the effectiveness and safety of phototherapy in older adults
Methods
Design, Setting, Sample, and Statistical Analysis
The institutional review boards of Kaiser Permanente Washington Health Research Institute, Seattle, and the University of Washington, Seattle, approved this study. It was conducted in a large US multispecialty health care system (Group Health, Seattle, Washington [now Kaiser Permanente Washington]) serving approximately 600,000 patients, using billing records to identify all patients treated with phototherapy between January 1, 2015, and December 31, 2015, all who received narrowband UVB (NB-UVB) phototherapy. All adults 65 years and older who received phototherapy treatment during the 12-month study period were included. Patients were included regardless of comorbidities and other dermatologic treatments to maintain as much uniformity as possible between the present study and 2 prior studies examining phototherapy in older adult populations in the United Kingdom16 and Turkey.17 Demographic and clinical factors were presented using frequencies (percentages) or means and medians as appropriate. Comparisons of dermatologic conditions and clearance levels used a Fisher exact test. The number of phototherapy treatments to clearance and total number of treatments were compared between groups of patients using independent sample t tests.
Phototherapy Protocol
All patients received treatments administered by specially trained phototherapy nurses using a Daavlin UV Series (The Daavlin Company) or an Ultralite unit (Ultralite Enterprises, Inc), both with 48 lamps. All phototherapy nurses had been previously trained to provide treatments based on standardized protocols (Table 1) and to determine the patient’s level of disease clearance using a high to low clearance scale (Table 2). Daavlin’s treatment protocols were built into the software that accompanied the units and were developed based on the American Academy of Dermatology guidelines. The starting dose for an individual patient was determined based on the estimated
Results
Patients
Billing records identified 229 total patients who received phototherapy in 2015, of whom 52 (22.7%) were at least 65 years old. The median age was 70 years (range, 65–91 years). Twenty-nine (56%) were men and 35 (67%) had previously received phototherapy treatments.
Dermatologic Conditions Treated With Phototherapy
Our primary aim was to identify the most common dermatologic conditions treated with phototherapy in older adults. Psoriasis and dermatitis were the most common conditions treated in the sample (50% [26/52] and 21% [11/52], respectively), with mycosis fungoides being the third most common (10% [5/52]) and vitiligo tied with prurigo nodularis as fourth most common (6% [3/52])(Figure 1).
Effectiveness and Safety of Phototherapy
Our secondary aim was to examine the effectiveness and safety of phototherapy in older adults. Phototherapy was effective in this population, with 50 of 52 patients (96%) achieving a high or medium level of clearance. The degree of clearance for each of the dermatologic conditions is shown in Figure 2. Psoriasis and dermatitis achieved high clearance rates in 81% (21/26) and 82% (9/11) of patients, respectively. Overall, conditions did not have significant differences in clearances rates (Fisher exact test, P=.10). On average, it took patients 33 treatments to achieve medium or high rates of clearance. Psoriasis cleared more quickly, with an average of 30.4 treatments vs 36.1 treatments for other conditions, but the difference was not significant (t test, P=.26). Patients received an average of 98 total phototherapy treatments; the median number of treatments was 81 due to many being on maintenance therapy over several months. There was no relationship between a history of treatment with phototherapy and the total number of treatments needed to achieve clearance (t test, P=.40), but interestingly, those who had a history of phototherapy took approximately 5 more treatments to achieve clearance. The present study found that a slightly larger number of men were being treated for psoriasis (15 men vs 11 women), but there was no significant difference in response rate based on gender.
Side effects from phototherapy were minimal; 24 patients (46%) experienced grade 1 (mild) erythema at some point during their treatment course. Thirteen (25%) patients experienced grade 2 erythema, but this was a rare event for most patients. Only 1 (2%) patient experienced grade 3 erythema 1 time. Three patients experienced increased itching (6%). Thirteen (25%) patients had no side effects. None developed severe erythema or blisters, and none discontinued phototherapy because of side effects. Over the course of the study year, we found a high degree of acceptance of phototherapy treatments by older patients: 22 (42%) completed therapy after achieving clearance, 10 (19%) were continuing ongoing treatments (maintenance), and 15 (29%) stopped because of life circumstances (eg, other health issues, moving out of the area). Only 4 (8%) stopped because of a lack of effectiveness, and 1 (2%) patient because the treatments were burdensome.
Comparison of Outcomes
Our third aim was to compare the outcomes with similar studies in the United Kingdom16 and Turkey.17 This study confirmed that phototherapy is being used in older adults (22.7% of this study’s total patients) and is an effective treatment for older patients experiencing a range of challenging inflammatory and proliferative skin diseases similar to studies in the general population. Prior phototherapy studies in elderly patients also found psoriasis to be the most common skin condition treated, with 1 study finding that 51% (19/37) of older phototherapy patients had psoriasis,16 while another reported 58% (37/95) of older phototherapy patients had psoriasis.17 These numbers are similar to those in our study, which showed 50% (26/52) of elderly phototherapy patients had psoriasis. Psoriasis is the main indication for treatment with NB-UVB phototherapy in the general population,19 and because the risk for psoriasis increases with age,20 it is not surprising that all 3 studies found psoriasis to be the most common indication in elderly phototherapy patients. Table 3 provides further details on conditions treated in all 3 studies.
Comment
Our study found that 94% of patients with psoriasis achieved clearance with an average of 30.4 treatments, which is comparable to the reported 91% response rate with an average of 30 treatments in the United Kingdom.16 The other similar study in Turkey17 reported 73.7% of psoriasis patients achieved a 75% or more improvement from baseline with an average of 42 treatments, which may reflect underlying differences in regional skin type. Of note, the scatter chart (Figure 3) shows that several patients in the present study’s analysis are listed as not clear, but many of those patients had low treatment numbers below the mean time to clearance. Thus, the present study’s response rate may have been underestimated.
In the general population, studies show that psoriasis treated with standardized phototherapy protocols typically clears with an average of 20.6 treatments.21 The levels of clearance were similar in our study’s older population, but more treatments were required to achieve those results, with an average of 10 more treatments needed (an additional 3.3 weeks). Similar results were found in this sample for dermatitis and mycosis fungoides, indicating comparable clearance rates and levels but a need for more treatments to achieve similar results compared to the general population.
Additionally, in the current study more patients experienced grade 1 (mild) erythema (46%) and grade 2 erythema (25%) at some point in their treatment compared with the United Kingdom16 (1.89%) and Turkey17 (35%) studies, though these side effects did not impact the clearance rate. Interestingly, the current study’s scatter chart (Figure 3) illustrates that this side effect did not seem to increase with aging in this population. If anything, the erythema response was more prevalent in the median or younger patients in the sample. Erythema may have been due to the frequent use of photosensitizing medications in older adults in the United States, some of which typically get discontinued in patients 75 years and older (eg, statins). Other potential causes might include the use of phototype vs minimal erythema dose–driven protocols, the standard utilization of protocols originally designed for psoriasis vs other condition-specific protocols, missed treatments leading to increased sensitivity, or possibly shielding mishaps (eg, not wearing a prescribed face shield). Given the number of potential causes and the possibility of overlapping factors, careful analysis is important. With NB-UVB phototherapy, near-erythemogenic doses are optimal to achieve effective treatments, but this delicate balance may be more problematic for older adults. Future studies are needed to fully determine the factors at play for this population. In the interim, it is important for phototherapy-trained nurses to consider this risk carefully in the older population. They must follow the prescribed protocols that guide them to query patients about their responses to the prior treatment (eg, erythema, tenderness, itching), photosensitizing medications, missed treatments, and placement of shielding, and then adjust the treatment dosing accordingly.
Limitations
This study had several limitations. Although clinical outcomes were recorded prospectively, the analysis was retrospective, unblinded, and not placebo controlled. It was conducted in a single organization (Group Health [now Kaiser Permanente Washington]) but did analyze data from 4 medical centers in different cities with diverse demographics and a variety of nursing staff providing the treatments. Although the vitiligo treatment protocol likely slowed the response rate for those patients with vitiligo, the numbers were small (ie, only 3 of 52 patients), so the researchers chose to include them in the current study. The sample population was relatively small, but when these data are evaluated alongside the studies in the United Kingdom16 and Turkey,17 they show a consistent picture illustrating the effectiveness and safety of phototherapy in the older population. Further epidemiologic studies could be helpful to further describe the usefulness of this modality compared with other treatments for a variety of dermatoses in this age group. Supplementary analysis specifically examining the relationship between the number and type of photosensitizing medications, frequency of erythema, and time to clearance also could be useful.
Conclusion
Older adults with a variety of dermatoses respond well to phototherapy and should have the opportunity to use it, particularly considering the potential for increased complications and costs from other treatment modalities, such as commonly used immunosuppressive pharmaceuticals. However, the current study and the comparison studies indicate that it is important to carefully consider the slower clearance rates and the potential risk for increased erythema in this population and adjust patient education and treatment dosing accordingly.
Unfortunately, many dermatology centers do not offer phototherapy because of infrastructure limitations such as space and specially trained nursing staff. Increasing accessibility of phototherapy for older adults through home treatments may be an alternative, given its effectiveness in the general population.22,23 In addition, home phototherapy may be worth pursuing for the older population considering the challenges they may face with transportation to the clinic setting and their increased risk for serious illness if exposed to infections such as COVID-19. The COVID-19 pandemic has brought to light the need for reliable, safe, and effective treatments that can be utilized in the safety of patients’ homes and should therefore be considered as an option for older adults. Issues such as mobility and cognitive decline could pose some complicating factors, but with the help of a well-trained family member or caregiver, home phototherapy could be a viable option that improves accessibility for older patients. Future research opportunities include further examination of the slower but ultimately equivalent response to phototherapy in the older population, the influence of photosensitizing medications on phototherapy effects, and the impact of phototherapy on utilization of immunosuppressive pharmaceuticals in older adults.
- British Photodermatology Group. An appraisal of narrowband (TL-01) UVB phototherapy. British Photodermatology Group Workshop Report (April 1996). Br J Dermatol. 1997;137:327-330.
Foerster J, Boswell K, West J, et al. Narrowband UVB treatment is highly effective and causes a strong reduction in the use of steroid and other creams in psoriasis patients in clinical practice. PLoS ONE. 2017;12:e0181813. doi:10.1371/journal.pone.0181813 - Fernández-Guarino M, Aboin-Gonzalez S, Barchino L, et al. Treatment of moderate and severe adult chronic atopic dermatitis with narrow-band UVB and the combination of narrow-band UVB/UVA phototherapy. Dermatol Ther. 2015;29:19-23.
- Ryu HH, Choe YS, Jo S, et al. Remission period in psoriasis after multiple cycles of narrowband ultraviolet B phototherapy. J Dermatol. 2014;41:622-627.
Tintle S, Shemer A, Suárez-Fariñas M, et al. Reversal of atopic dermatitis with narrow-band UVB phototherapy and biomarkers for therapeutic response. J Allergy Clin Immunol. 2011;128:583-593. - Gambichler T, Breuckmann F, Boms S, et al. Narrowband UVB phototherapy in skin conditions beyond psoriasis. J Am Acad Dermatol. 2005;52:660-670.
- Schneider LA, Hinrichs R, Scharffetter-Kochanek K. Phototherapy and photochemotherapy. Clin Dermatol. 2008;26:464-476.
- Martin JA, Laube S, Edwards C, et al. Rate of acute adverse events for narrow-band UVB and psoralen-UVA phototherapy. Photodermatol Photoimmunol Photomed. 2007;23:68-72.
- Mokos ZB, Jovic A, Ceovic R, et al. Therapeutic challenges in the mature patient. Clin Dermatol. 2018;36:128-139.
- Di Lernia V, Goldust M. An overview of the efficacy and safety of systemic treatments for psoriasis in the elderly. Exp Opin Biol Ther. 2018;18:897-903.
- Napolitano M, Balato N, Ayala F, et al. Psoriasis in elderly and non-elderly population: clinical and molecular features. G Ital Dermatol Venereol. 2016;151:587-595.
- Grozdev IS, Van Voorhees AS, Gottlieb AB, et al. Psoriasis in the elderly: from the Medical Board of the National Psoriasis Foundation. J Am Acad Dermatol. 2011;65:537-545.
- Click J, Alabaster A, Postlethwaite D, et al. Effect of availability of at-home phototherapy on the use of systemic medications for psoriasis.
Photodermatol Photoimmunol Photomed. 2017;33:345-346. - Piaserico S, Conti A, Lo Console F, et al.
Efficacy and safety of systemic treatments for psoriasis in elderly. Acta Derm Venereol. 2014;94:293-297. - Soliman A, Nofal E, Nofal A, et al. Combination therapy of methotrexate plus NB-UVB phototherapy is more effective than methotrexate monotherapy in the treatment of chronic plaque psoriasis. J Dermatol Treat. 2015;26:528-534.
- Powell JB, Gach JE. Phototherapy in the elderly. Clin Exp Dermatol. 2015;40:605-610.
- Bulur I, Erdogan HK, Aksu AE, et al. The efficacy and safety of phototherapy in geriatric patients: a retrospective study. An Bras Dermatol. 2018;93:33-38.
- Madigan LM, Al-Jamal M, Hamzavi I. Exploring the gaps in the evidence-based application of narrowband UVB for the treatment of vitiligo. Photodermatol Photoimmunol Photomed. 2016;32:66-80.
- Ibbotson SH. A perspective on the use of NB-UVB phototherapy vs. PUVA photochemotherapy. Front Med (Lausanne). 2018;5:184.
- Bell LM, Sedlack R, Beard CM, et al. Incidence of psoriasis in Rochester, Minn, 1980-1983. Arch Dermatol. 1991;127:1184-1187.
- Totonchy MB, Chiu MW. UV-based therapy. Dermatol Clin. 2014;32:399-413.
- Cameron H, Yule S, Dawe RS, et al. Review of an established UK home phototherapy service 1998-2011: improving access to a cost-effective treatment for chronic skin disease. Public Health. 2014;128:317-324.
- Matthews SW, Simmer M, Williams L, et al. Transition of patients with psoriasis from office-based phototherapy to nurse-supported home phototherapy: a pilot study. JDNA. 2018;10:29-41.
- British Photodermatology Group. An appraisal of narrowband (TL-01) UVB phototherapy. British Photodermatology Group Workshop Report (April 1996). Br J Dermatol. 1997;137:327-330.
Foerster J, Boswell K, West J, et al. Narrowband UVB treatment is highly effective and causes a strong reduction in the use of steroid and other creams in psoriasis patients in clinical practice. PLoS ONE. 2017;12:e0181813. doi:10.1371/journal.pone.0181813 - Fernández-Guarino M, Aboin-Gonzalez S, Barchino L, et al. Treatment of moderate and severe adult chronic atopic dermatitis with narrow-band UVB and the combination of narrow-band UVB/UVA phototherapy. Dermatol Ther. 2015;29:19-23.
- Ryu HH, Choe YS, Jo S, et al. Remission period in psoriasis after multiple cycles of narrowband ultraviolet B phototherapy. J Dermatol. 2014;41:622-627.
Tintle S, Shemer A, Suárez-Fariñas M, et al. Reversal of atopic dermatitis with narrow-band UVB phototherapy and biomarkers for therapeutic response. J Allergy Clin Immunol. 2011;128:583-593. - Gambichler T, Breuckmann F, Boms S, et al. Narrowband UVB phototherapy in skin conditions beyond psoriasis. J Am Acad Dermatol. 2005;52:660-670.
- Schneider LA, Hinrichs R, Scharffetter-Kochanek K. Phototherapy and photochemotherapy. Clin Dermatol. 2008;26:464-476.
- Martin JA, Laube S, Edwards C, et al. Rate of acute adverse events for narrow-band UVB and psoralen-UVA phototherapy. Photodermatol Photoimmunol Photomed. 2007;23:68-72.
- Mokos ZB, Jovic A, Ceovic R, et al. Therapeutic challenges in the mature patient. Clin Dermatol. 2018;36:128-139.
- Di Lernia V, Goldust M. An overview of the efficacy and safety of systemic treatments for psoriasis in the elderly. Exp Opin Biol Ther. 2018;18:897-903.
- Napolitano M, Balato N, Ayala F, et al. Psoriasis in elderly and non-elderly population: clinical and molecular features. G Ital Dermatol Venereol. 2016;151:587-595.
- Grozdev IS, Van Voorhees AS, Gottlieb AB, et al. Psoriasis in the elderly: from the Medical Board of the National Psoriasis Foundation. J Am Acad Dermatol. 2011;65:537-545.
- Click J, Alabaster A, Postlethwaite D, et al. Effect of availability of at-home phototherapy on the use of systemic medications for psoriasis.
Photodermatol Photoimmunol Photomed. 2017;33:345-346. - Piaserico S, Conti A, Lo Console F, et al.
Efficacy and safety of systemic treatments for psoriasis in elderly. Acta Derm Venereol. 2014;94:293-297. - Soliman A, Nofal E, Nofal A, et al. Combination therapy of methotrexate plus NB-UVB phototherapy is more effective than methotrexate monotherapy in the treatment of chronic plaque psoriasis. J Dermatol Treat. 2015;26:528-534.
- Powell JB, Gach JE. Phototherapy in the elderly. Clin Exp Dermatol. 2015;40:605-610.
- Bulur I, Erdogan HK, Aksu AE, et al. The efficacy and safety of phototherapy in geriatric patients: a retrospective study. An Bras Dermatol. 2018;93:33-38.
- Madigan LM, Al-Jamal M, Hamzavi I. Exploring the gaps in the evidence-based application of narrowband UVB for the treatment of vitiligo. Photodermatol Photoimmunol Photomed. 2016;32:66-80.
- Ibbotson SH. A perspective on the use of NB-UVB phototherapy vs. PUVA photochemotherapy. Front Med (Lausanne). 2018;5:184.
- Bell LM, Sedlack R, Beard CM, et al. Incidence of psoriasis in Rochester, Minn, 1980-1983. Arch Dermatol. 1991;127:1184-1187.
- Totonchy MB, Chiu MW. UV-based therapy. Dermatol Clin. 2014;32:399-413.
- Cameron H, Yule S, Dawe RS, et al. Review of an established UK home phototherapy service 1998-2011: improving access to a cost-effective treatment for chronic skin disease. Public Health. 2014;128:317-324.
- Matthews SW, Simmer M, Williams L, et al. Transition of patients with psoriasis from office-based phototherapy to nurse-supported home phototherapy: a pilot study. JDNA. 2018;10:29-41.
Practice Points
- With appropriate nursing care, phototherapy can be safe and effective for a variety of conditions in elderly patients.
- Compared to younger patients, elderly patients may need more sessions to achieve comparable clearance rates.
- The increased prevalence of photosensitizing medications in the elderly population will require careful adjustments in dosing.
A Longitudinal Analysis of Functional Disability, Recovery, and Nursing Home Utilization After Hospitalization for Ambulatory Care Sensitive Conditions Among Community-Living Older Persons
Acute illnesses requiring hospitalization serve as a sentinel event, with many older adults requiring assistance with activities of daily living (ADLs) upon discharge.1-3 Older adults who are frail experience even higher rates of hospital-associated disability, and rates of recovery to baseline functional status have varied.4,5 Loss of independence in ADLs has been associated with nursing home (NH) utilization, caregiver burden, and mortality.6
To date, studies have characterized functional trajectories before and after hospitalization in older persons for broad medical conditions, noting persistence of disability and incomplete recovery to baseline functional status.7 Prior evaluations have also noted the long-term disabling impact of critical conditions such as acute myocardial infarction, stroke, and sepsis,8,9 but a knowledge gap exists regarding the subsequent functional disability, recovery, and incident NH admission among older persons who are hospitalized for ambulatory care sensitive conditions (ACSCs). Often considered potentially preventable with optimal ambulatory care,10,11 ACSCs represent acute, chronic, and vaccine-preventable conditions, including urinary tract infection, congestive heart failure, diabetes mellitus, and pneumonia. Investigating the aforementioned patient-centered measures post hospitalization could provide valuable supporting evidence for the continued recognition of ACSC-related hospitalizations in national quality payment programs set forth by the Centers for Medicare & Medicaid Services (CMS).12 Demonstrating adverse outcomes after ACSC-related hospitalizations may help support interventions that target potentially preventable ACSC-related hospitalizations, such as home-based care or telehealth, with the goal of improving functional outcomes and reducing NH admission in older persons.
To address these gaps, we evaluated ACSC-related hospitalizations among participants of the Precipitating Events Project (PEP), a 19-year longitudinal study of community-living persons who were initially nondisabled in their basic functional activities. In the 6 months following an ACSC-related hospitalization, our objectives were to describe: (1) the 6-month course of postdischarge functional disability, (2) the cumulative monthly probability of functional recovery, and (3) the cumulative monthly probability of incident NH admission.
METHODS
Study Population
Participants were drawn from the PEP study, an ongoing, prospective, longitudinal study of 754 community-dwelling persons aged 70 years or older.13 Potential participants were members of a large health plan in greater New Haven, Connecticut, and were enrolled from March 1998 through October 1999. As previously described,14 persons were oversampled if they were physically frail, as denoted by a timed score >10 seconds on the rapid gait test. Exclusion criteria included significant cognitive impairment with no available proxy, life expectancy less than 12 months, plans to leave the area, and inability to speak English. Participants were initially required to be nondisabled in four basic activities of daily living (bathing, dressing, walking across a room, and transferring from a chair). Eligibility was determined during a screening telephone interview and was confirmed during an in-home assessment. Of the eligible members, 75.2% agreed to participate in the project, and persons who declined to participate did not significantly differ in age or sex from those who were enrolled. The Yale Human Investigation Committee approved the study protocol, and all participants provided verbal informed consent.
Data Collection
From 1998 to 2017, comprehensive home-based assessments were completed by trained research nurses at baseline and at 18-month intervals over 234 months (except at 126 months), and telephone interviews were completed monthly through June 2018, to obtain information on disability over time. For participants who had significant cognitive impairment or who were unavailable, we interviewed a proxy informant using a rigorous protocol with demonstrated reliability and validity.14 All incident NH admissions, including both short- and long-term stays, were identified using the CMS Skilled Nursing Facility claims file and Long Term Care Minimum Data Set. Deaths were ascertained by review of obituaries and/or from a proxy informant, with a completion rate of 100%. A total of 688 participants (91.2%) had died after a median follow-up of 108 months, while 43 participants (5.7%) dropped out of the study after a median follow-up of 27 months. Among all participants, data were otherwise available for 99.2% of 85,531 monthly telephone interviews.
Assembly of Analytic Sample
PEP participants were considered for inclusion in the analytic sample if they had a hospitalization with an ACSC as the primary diagnosis on linked Medicare claims data. The complete list of ACSCs was defined using specifications from the Agency for Healthcare Research and Quality,15 and was assembled using the International Classification of Disease, Ninth Revision, Clinical Modification (ICD-9-CM) classification prior to October 1, 2015, and ICD Tenth Revision, Clinical Modification (ICD-10-CM) classification after October 1, 2015 (Appendix Table 1). Examples of ACSCs include congestive heart failure, dehydration, urinary tract infection, and angina without procedure. As performed previously,16,17 two ACSCs (low birthweight; asthma in younger adults 18-39 years) were not included in this analysis because they were not based on full adult populations.
ACSC-related hospitalizations were included through December 2017. Participants could contribute more than one ACSC-related hospitalization over the course of the study based on the following criteria: (1) participant did not have a prior non-ACSC-related hospitalization within an 18-month interval; (2) participant did not have a prior ACSC-related hospitalization or treat-and-release emergency department (ED) visit within an 18-month interval (to ensure independence of observations if the participant was still recovering from the prior event and because some of the characteristics within Table 1 are susceptible to change in the setting of an intervening event and, hence, would not accurately reflect the status of the participant prior to ACSC-related hospitalization); (3) participant was not admitted from a NH; (4) participant did not have an in-hospital intensive care unit (ICU) stay (because persons with critical illness are a distinct population with frequent disability and prolonged recovery, as previously described18), in-hospital death, or death before first follow-up interview (because our aim was to evaluate disability and recovery after the hospitalization7).
Assembly of the primary analytic sample is depicted in the Appendix Figure. Of the 814 patients who were identified with ACSC-related hospitalizations, 107 had a prior non-ACSC-related hospitalization and 275 had a prior ACSC-related hospitalization or a treat-and-release ED visit within an 18-month interval. Of the remaining 432 ACSC-related hospitalizations, 181 were excluded: 114 patients were admitted from a NH, 38 had an in-hospital ICU stay, 3 died in the hospital, 11 died before their first follow-up interview, and 15 had withdrawn from the study. The primary analytic sample included the remaining 251 ACSC-related hospitalizations, contributed by 196 participants. Specifically, nine participants contributed three ACSC-related hospitalizations each, 37 participants contributed two hospitalizations each, and the remaining 150 participants contributed one hospitalization each. During the 6-month follow-up period, 40 participants contributing ACSC-related hospitalizations died after a median (interquartile range [IQR]) of 4 (2-5) months, and 1 person refused continued participation.
Comprehensive Assessments
During the comprehensive in-home assessments, data were obtained on demographic characteristics. Age was measured in years at the time of the ACSC-related hospitalization. In addition, we describe factors from the comprehensive assessment immediately prior to the ACSC-related hospitalization, grouped into two additional domains related to disability19: health-related and cognitive-psychosocial. The health-related factors included nine self-reported, physician-diagnosed chronic conditions and frailty. The cognitive-psychosocial factors included social support, cognitive impairment, and depressive symptoms.
Assessment of Disability
Complete details about the assessment of disability have been previously described.13,14,19,20 Briefly, disability was assessed during the monthly telephone interviews, and included four basic activities (bathing, dressing, walking across a room, and transferring from a chair), five instrumental activities (shopping, housework, meal preparation, taking medications, and managing finances), and three mobility activities (walking a quarter mile, climbing a flight of stairs, and lifting or carrying 10 lb). Participants were asked, “At the present time, do you need help from another person to [complete the task]?” Disability was operationalized as the need for personal assistance or an inability to perform the task. Participants were also asked about a fourth mobility activity, “Have you driven a car during the past month?” Those who responded no were classified as being disabled in driving.19
The number of disabilities overall and for each functional domain (basic, instrumental, and mobility) was summed. Possible disability scores ranged from 0 to 13, with a score of 0 indicating complete independence in all of the items, and a score of 13 indicating complete dependence. Worse postdischarge disability was defined as a total disability score (0-13) at the first telephone interview after an ACSC-related hospitalization that was greater than the total disability score from the telephone interview immediately preceding hospitalization.
Outcome Measures
The primary outcome was the number of disabilities in all 13 basic, instrumental, and mobility activities in each of the 6 months following discharge from an ACSC-related hospitalization. To determine whether our findings were consistent across the three functional domains, we also evaluated the number of disabilities in the four basic, five instrumental, and four mobility activities separately. As secondary outcomes, we evaluated: (1) the cumulative probability of recovery within the 6-month follow-up time frame after an ACSC-related hospitalization, with “recovery” defined as return to the participant’s pre-ACSC-related hospitalization total disability score, and (2) the cumulative probability of incident NH admission within the 6 months after an ACSC-related hospitalization. Aligned with CMS and prior literature,21,22 we defined a short-term NH stay as ≤100 days and a long-term NH stay as >100 days.
Statistical Analysis
Pre-ACSC-related hospitalization characteristics were summarized by means (SDs) and frequencies with proportions. We determined the mean number of disabilities in each of the 6 months following hospital discharge, with the prehospitalization value included as a reference point. We also determined the mean (SD) number of disabilities for the three subscales of disability (basic activities of daily living [BADLs], instrumental activities of daily living [IADLs], and mobility activities). We calculated the cumulative probability of recovery within 6 months of hospital discharge. Finally, we determined the cumulative probability of incident NH admission during the 6 months after hospital discharge.
To test the robustness of our main results, we conducted a sensitivity analysis assessing disability scores of the 150 participants that contributed only one ACSC-related hospitalization. All analyses were performed using Stata, version 16.0, statistical software (StataCorp).
RESULTS
Table 1 shows the characteristics of the 251 ACSC-related hospitalizations immediately prior to hospitalization. Participants’ mean (SD) age was 85.1 (6.0) years, and the mean total disability score was 5.4. The majority were female, non-Hispanic White, frail, and lived alone. As shown in Appendix Table 2, the three most common reasons for ACSC-related hospitalizations were congestive heart failure (n = 69), bacterial pneumonia (n = 53), and dehydration (n = 44).
The Figure shows the disability scores during the 6-month follow-up period for total, basic, instrumental, and mobility activities, in panels A, B, C, and D, respectively. The exact values are provided in Appendix Table 3. After hospitalization, disability scores for total, basic, instrumental, and mobility activities peaked at month 1 and tended to improve modestly over the next 5 months, but remained greater, on average, than pre-hospitalization scores. Of the 40 participants who died within the 6-month follow-up period, 36 (90%) had worse disability scores in their last month of life than in the month prior to their ACSC-related hospitalization.
Table 2 shows the cumulative probability of functional recovery after ACSC-related hospitalizations. Recovery was incomplete, with only 70% (95% CI, 64%-76%) of hospitalizations achieving a return to the pre-hospitalization total disability score within 6 months of hospitalization.
Table 3 shows the cumulative probability of incident NH admission after an ACSC-related hospitalization. Of the 251 ACSC-related hospitalizations, incident NH admission was experienced by 38% (95% CI, 32%-44%) within 1 month and 50% (95% CI, 43%-56%) within 6 months of discharge. Short-term NH stays accounted for 90 (75.6%) of the 119 incident NH admissions within the 6 months after ACSC-related hospitalizations. Sensitivity analyses yielded comparable disability scores, shown in Appendix Table 4.
DISCUSSION
In this longitudinal study of community-living older persons, we evaluated functional disability, recovery, and incident NH admission within 6 months of hospitalization for an ACSC. Our study has three major findings. First, disability scores for total, basic, instrumental, and mobility activities at months 1 to 6 of follow-up were greater on average than pre-hospitalization scores. Second, functional recovery was not achieved by 3 of 10 participants after an ACSC-related hospitalization. Third, half of them experienced an incident NH admission within 6 months of discharge from an ACSC-related hospitalization, although about three-quarters of these were short-term stays. Our findings provide evidence that older persons experience clinically meaningful adverse patient-reported outcomes after ACSC-related hospitalizations.
Prior research involving ACSCs has focused largely on rates of hospitalization as a measure of access to primary care and the associated factors predictive of ACSC-related hospitalizations,23-26 and has not addressed subsequent patient-reported outcomes. The findings in this analysis highlight that older persons experience worsening disability immediately after an ACSC-related hospitalization, which persists for prolonged periods and often results in incomplete recovery. Prior research has assessed pre-hospitalization functional status through retrospective recall approaches,2 included only older adults discharged with incident disability,3 and examined functional status after all-cause medical illness hospitalizations.5 Our prospective analysis extends the literature by reliably capturing pre-hospital disability scores and uniquely assessing the cohort of older persons hospitalized with ACSCs.
Our work is relevant to the continued evaluation of ACSC-related hospitalizations in national quality measurement and payment initiatives among Medicare beneficiaries. In prior evaluations of ACSC-related quality measures, stakeholders have criticized the measures for limited validity due to a lack of evidence linking each utilization outcome to other patient-centered outcomes.10,27 Our work addresses this gap by demonstrating that ACSC-related hospitalizations are linked to persistent disability, incomplete functional recovery, and incident NH admissions. Given the large body of evidence demonstrating the priority older persons place on these patient-reported outcomes,28,29 our work should reassure policymakers seeking to transform quality measurement programs into a more patient-oriented enterprise.
Our findings have several clinical practice, research, and policy implications. First, more-effective clinical strategies to minimize the level of care required for acute exacerbations of ACSC-related illnesses may include: (1) substituting home-based care30 and telehealth interventions31 for traditional inpatient hospitalization, (2) making in-ED resources (ie, case management services, geriatric-focused advanced practice providers) more accessible for older persons with ACSC-related illnesses, thereby enhancing care transitions and follow-up to avoid potential current and subsequent hospitalizations, and (3) ensuring adequate ambulatory care access to all older persons, as prior work has shown variation in ACSC hospital admission rates dependent on population factors such as high-poverty neighborhoods,16 insurance status,16,32 and race/ethnicity.33
Clinical strategies have been narrow and not holistic for ACSCs; for example, many institutions have focused on pneumonia vaccinations to reduce hospitalizations, but our work supports the need to further evaluate the impact of preventing ACSC-related hospitalizations and their associated disabling consequences. For patients admitted to the hospital, clinical strategies, such as in-hospital or post-hospital mobility and activity programs, have been shown to be protective against hospital-associated disability.34,35 Furthermore, hospital discharge planning could include preparing older persons for anticipated functional disabilities, associated recoveries, and NH admission after ACSC-related hospitalizations. Risk factors contributing to post-hospitalization functional disability and recovery have been identified,19,20,36 but future work is needed to: (1) identify target populations (including those most likely to worsen) so that interventions can be offered earlier in the course of care to those who would benefit most, and (2) identify and learn from those who are resilient and have recovered, to better understand factors contributing to their success.
Our study has several strengths. First, the study is unique due to its longitudinal design, with monthly assessments of functional status. Since functional status was assessed prospectively before the ACSC-related hospitalization, we also have avoided any potential concern for recall bias that may be present if assessed after the hospitalization. Additionally, through the use of Medicare claims and the Minimum Data Set, the ascertainment of hospitalizations and NH admissions was likely complete for the studied population.
However, the study has limitations. First, functional measures were based on self-reports rather than objective measurements. Nevertheless, the self-report function is often used to guide coverage determinations in the Medicare program, as it has been shown to be associated with poor health outcomes.37 Second, we are unable to comment on the rate of functional decline or NH admission when an older person was not hospitalized in relation to an ACSC. Future analyses may benefit from using a control group (eg, older adults without an ACSC hospitalization or older adults with a non-ACSC hospitalization). Third, we used strict exclusion criteria to identify a population of older adults without recent hospitalizations to determine the isolated impact of ACSC hospitalization on disability, incident NH admission, and functional recovery. Considering this potential selection bias, our findings are likely conservative estimates of the patient-centered outcomes evaluated. Fourth, participants were not asked about feeding and toileting. However, the incidence of disability in these ADLs is low among nondisabled, community-living older persons, and it is highly uncommon for disability to develop in these ADLs without concurrent disability in the ADLs within this analysis.14,38
Finally, because our study participants were members of a single health plan in a small urban area and included nondisabled older persons living in the community, our findings may not be generalizable to geriatric patients in other settings. Nonetheless, the demographics of our cohort reflect those of older persons in New Haven County, Connecticut, which are similar to the demographics of the US population, with the exception of race and ethnicity. In addition, the generalizability of our results are strengthened by the study’s high participation rate and minimal attrition.
CONCLUSION
Within 6 months of ACSC-related hospitalizations, community-living older persons exhibited greater total disability scores than those immediately preceding hospitalization. In the same time frame, 3 of 10 older persons did not achieve functional recovery, and half experienced incident NH admission. These results provide evidence regarding the continued recognition of ACSC-related hospitalizations in federal quality measurement and payment programs and suggests the need for preventive and comprehensive interventions to meaningfully improve longitudinal outcomes.
Acknowledgments
We thank Denise Shepard, BSN, MBA, Andrea Benjamin, BSN, Barbara Foster, and Amy Shelton, MPH, for assistance with data collection; Geraldine Hawthorne, BS, for assistance with data entry and management; Peter Charpentier, MPH, for design and development of the study database and participant tracking system; and Joanne McGloin, MDiv, MBA, for leadership and advice as the Project Director. Each of these persons were paid employees of Yale School of Medicine during the conduct of this study.
1. Covinsky KE, Pierluissi E, Johnston CB. Hospitalization-associated disability: “She was probably able to ambulate, but I’m not sure” JAMA. 2011;306(16):1782-1793. https://doi.org/10.1001/jama.2011.1556
2. Covinsky KE, Palmer RM, Fortinsky RH, et al. Loss of independence in activities of daily living in older adults hospitalized with medical illnesses: increased vulnerability with age. J Am Geriatr Soc. 2003;51(4):451-458. https://doi.org/10.1046/j.1532-5415.2003.51152.x
3. Barnes DE, Mehta KM, Boscardin WJ, et al. Prediction of recovery, dependence or death in elders who become disabled during hospitalization. J Gen Intern Med. 2013;28(2):261-268. https://doi.org/10.1007/s11606-012-2226-y
4. Gill TM, Allore HG, Gahbauer EA, Murphy TE. Change in disability after hospitalization or restricted activity in older persons. JAMA. 2010;304(17):1919-1928. https://doi.org/10.1001/jama.2010.1568
5. Boyd CM, Landefeld CS, Counsell SR, et al. Recovery of activities of daily living in older adults after hospitalization for acute medical illness. J Am Geriatr Soc. 2008;56(12):2171-2179. https://doi.org/10.1111/j.1532-5415.2008.02023.x
6. Loyd C, Markland AD, Zhang Y, et al. Prevalence of hospital-associated disability in older adults: a meta-analysis. J Am Med Dir Assoc. 2020;21(4):455-461. https://doi.org/10.1016/j.jamda.2019.09.015
7. Dharmarajan K, Han L, Gahbauer EA, Leo-Summers LS, Gill TM. Disability and recovery after hospitalization for medical illness among community-living older persons: a prospective cohort study. J Am Geriatr Soc. 2020;68(3):486-495. https://doi.org/10.1111/jgs.16350
8. Levine DA, Davydow DS, Hough CL, Langa KM, Rogers MAM, Iwashyna TJ. Functional disability and cognitive impairment after hospitalization for myocardial infarction and stroke. Circ Cardiovasc Qual Outcomes. 2014;7(6):863-871. https://doi.org/10.1161/HCQ.0000000000000008
9. Iwashyna TJ, Ely EW, Smith DM, Langa KM. Long-term cognitive impairment and functional disability among survivors of severe sepsis. JAMA. 2010;304(16):1787-1794. https://doi.org/10.1001/jama.2010.1553
10. Hodgson K, Deeny SR, Steventon A. Ambulatory care-sensitive conditions: their potential uses and limitations. BMJ Qual Saf. 2019;28(6):429-433. https://doi.org/10.1136/bmjqs-2018-008820
11. Agency for Healthcare Research and Quality (AHRQ). Quality Indicator User Guide: Prevention Quality Indicators (PQI) Composite Measures. Version 2020. Accessed November 10, 2020. https://www.qualityindicators.ahrq.gov/modules/pqi_resources.aspx.
12. Centers for Medicare & Medicaid Services. 2016 Measure information about the hospital admissions for acute and chronic ambulatory care-sensitive condition (ACSC) composite measures, calculated for the 2018 value-based payment modified program. Accessed November 24, 2020. https://www.cms.gov/Medicare/Medicare-Fee-for-Service-Payment/PhysicianFeedbackProgram/Downloads/2016-ACSC-MIF.pdf.
13. Gill TM, Desai MM, Gahbauer EA, Holford TR, Williams CS. Restricted activity among community-living older persons: incidence, precipitants, and health care utilization. Ann Intern Med. 2001;135(5):313-321. https://doi.org/10.7326/0003-4819-135-5-200109040-00007
14. Gill TM, Hardy SE, Williams CS. Underestimation of disability in community-living older persons. J Am Geriatr Soc. 2002;50(9):1492-1497. https://doi.org/10.1046/j.1532-5415.2002.50403.x
15. Agency for Healthcare Research and Quality. Prevention Quality Indicators Technical Specifications Updates—Version v2018 and 2018.0.1 (ICD 10-CM/PCS), June 2018. Accessed February 4, 2020. https://www.qualityindicators.ahrq.gov/Modules/PQI_TechSpec_ICD10_v2018.aspx.
16. Johnson PJ, Ghildayal N, Ward AC, Westgard BC, Boland LL, Hokanson JS. Disparities in potentially avoidable emergency department (ED) care: ED visits for ambulatory care sensitive conditions. Med Care. 2012;50(12):1020-1028. https://doi.org/10.1097/MLR.0b013e318270bad4
17. Galarraga JE, Mutter R, Pines JM. Costs associated with ambulatory care sensitive conditions across hospital-based settings. Acad Emerg Med. 2015;22(2):172-181. https://doi.org/10.1111/acem.12579
18. Ferrante LE, Pisani MA, Murphy TE, Gahbauer EA, Leo-Summers LS, Gill TM. Functional trajectories among older persons before and after critical illness. JAMA Intern Med. 2015;175(4):523-529. https://doi.org/10.1001/jamainternmed.2014.7889
19. Gill TM, Gahbauer EA, Murphy TE, Han L, Allore HG. Risk factors and precipitants of long-term disability in community mobility: a cohort study of older persons. Ann Intern Med. 2012;156(2):131-140. https://doi.org/10.7326/0003-4819-156-2-201201170-00009
20. Hardy SE, Gill TM. Factors associated with recovery of independence among newly disabled older persons. Arch Intern Med. 2005;165(1):106-112. https://doi.org/10.1001/archinte.165.1.106
21. Centers for Medicare & Medicaid Services. Nursing Home Quality Initiative—Quality Measures. Accessed June 13, 2021. https://www.cms.gov/Medicare/Quality-Initiatives-Patient-Assessment-Instruments/NursingHomeQualityInits/NHQIQualityMeasures
22. Goodwin JS, Li S, Zhou J, Graham JE, Karmarkar A, Ottenbacher K. Comparison of methods to identify long term care nursing home residence with administrative data. BMC Health Serv Res. 2017;17(1):376. https://doi.org/10.1186/s12913-017-2318-9
23. Laditka, JN, Laditka SB, Probst JC. More may be better: evidence of a negative relationship between physician supply and hospitalization for ambulatory care sensitive conditions. Health Serv Res. 2005;40(4):1148-1166. https://doi.org/10.1111/j.1475-6773.2005.00403.x
24. Ansar Z, Laditka JN, Laditka SB. Access to health care and hospitalization for ambulatory care sensitive conditions. Med Care Res Rev. 2006;63(6):719-741. https://doi.org/10.1177/1077558706293637
25. Mackinko J, de Oliveira VB, Turci MA, Guanais FC, Bonolo PF, Lima-Costa MF. The influence of primary care and hospital supply on ambulatory care-sensitive hospitalizations among adults in Brazil, 1999-2007. Am J Public Health. 2011;101(10):1963-1970. https://doi.org/10.2105/AJPH.2010.198887
26. Gibson OR, Segal L, McDermott RA. A systematic review of evidence on the association between hospitalisation for chronic disease related ambulatory care sensitive conditions and primary health care resourcing. BMC Health Serv Res. 2013;13:336. https://doi.org/10.1186/1472-6963-13-336
27. Vuik SI, Fontana G, Mayer E, Darzi A. Do hospitalisations for ambulatory care sensitive conditions reflect low access to primary care? An observational cohort study of primary care usage prior to hospitalisation. BMJ Open. 2017;7(8):e015704. https://doi.org/10.1136/bmjopen-2016-015704
28. Fried TR, Tinetti M, Agostini J, Iannone L, Towle V. Health outcome prioritization to elicit preferences of older persons with multiple health conditions. Patient Educ Couns. 2011;83(2):278-282. https://doi.org/10.1016/j.pec.2010.04.032
29. Reuben DB, Tinetti ME. Goal-oriented patient care—an alternative health outcomes paradigm. N Engl J Med. 2012;366(9):777-779. https://doi.org/10.1056/NEJMp1113631
30. Federman AD, Soones T, DeCherrie LV, Leff B, Siu AL. Association of a bundled hospital-at-home and 30-day postacute transitional care program with clinical outcomes and patient experiences. JAMA Intern Med. 2018;178(8):1033-1040. https://doi.org/10.1001/jamainternmed.2018.2562
31. Shah MN, Wasserman EB, Gillespie SM, et al. High-intensity telemedicine decreases emergency department use for ambulatory care sensitive conditions by older adult senior living community residents. J Am Med Dir Assoc. 2015;16(12):1077-1081. https://doi.org/10.1016/j.jamda.2015.07.009
32. Oster A, Bindman AB. Emergency department visits for ambulatory care sensitive conditions: insights into preventable hospitalizations. Med Care. 2003;41(2):198-207. https://doi.org/10.1097/01.MLR.0000045021.70297.9F
33. O’Neil SS, Lake T, Merrill A, Wilson A, Mann DA, Bartnyska LM. Racial disparities in hospitalizations for ambulatory care-sensitive conditions. Am J Prev Med. 2010;38(4):381-388. https://doi.org/10.1016/j.amepre.2009.12.026
34. Pavon JM, Sloane RJ, Pieper RF, et al. Accelerometer-measured hospital physical activity and hospital-acquired disability in older adults. J Am Geriatr Soc. 2020;68:261-265. https://doi.org/10.1111/jgs.16231
35. Sunde S, Hesseberg K, Skelton DA, et al. Effects of a multicomponent high intensity exercise program on physical function and health-related quality of life in older adults with or at risk of mobility disability after discharge from hospital: a randomised controlled trial. BMC Geriatr. 2020;20(1):464. https://doi.org/10.1186/s12877-020-01829-9
36. Hardy SE, Gill TM. Recovery from disability among community-dwelling older persons. JAMA. 2004;291(13):1596-1602. https://doi.org/10.1001/jama.291.13.1596
37. Rotenberg J, Kinosian B, Boling P, Taler G, Independence at Home Learning Collaborative Writing Group. Home-based primary care: beyond extension of the independence at home demonstration. J Am Geriatr Soc. 2018;66(4):812-817. https://doi.org/10.1111/jgs.15314
38. Rodgers W, Miller B. A comparative analysis of ADL questions in surveys of older people. J Gerontol B Psychol Sci Soc Sci. 1997;52:21-36. https://doi.org/10.1093/geronb/52b.special_issue.21
Acute illnesses requiring hospitalization serve as a sentinel event, with many older adults requiring assistance with activities of daily living (ADLs) upon discharge.1-3 Older adults who are frail experience even higher rates of hospital-associated disability, and rates of recovery to baseline functional status have varied.4,5 Loss of independence in ADLs has been associated with nursing home (NH) utilization, caregiver burden, and mortality.6
To date, studies have characterized functional trajectories before and after hospitalization in older persons for broad medical conditions, noting persistence of disability and incomplete recovery to baseline functional status.7 Prior evaluations have also noted the long-term disabling impact of critical conditions such as acute myocardial infarction, stroke, and sepsis,8,9 but a knowledge gap exists regarding the subsequent functional disability, recovery, and incident NH admission among older persons who are hospitalized for ambulatory care sensitive conditions (ACSCs). Often considered potentially preventable with optimal ambulatory care,10,11 ACSCs represent acute, chronic, and vaccine-preventable conditions, including urinary tract infection, congestive heart failure, diabetes mellitus, and pneumonia. Investigating the aforementioned patient-centered measures post hospitalization could provide valuable supporting evidence for the continued recognition of ACSC-related hospitalizations in national quality payment programs set forth by the Centers for Medicare & Medicaid Services (CMS).12 Demonstrating adverse outcomes after ACSC-related hospitalizations may help support interventions that target potentially preventable ACSC-related hospitalizations, such as home-based care or telehealth, with the goal of improving functional outcomes and reducing NH admission in older persons.
To address these gaps, we evaluated ACSC-related hospitalizations among participants of the Precipitating Events Project (PEP), a 19-year longitudinal study of community-living persons who were initially nondisabled in their basic functional activities. In the 6 months following an ACSC-related hospitalization, our objectives were to describe: (1) the 6-month course of postdischarge functional disability, (2) the cumulative monthly probability of functional recovery, and (3) the cumulative monthly probability of incident NH admission.
METHODS
Study Population
Participants were drawn from the PEP study, an ongoing, prospective, longitudinal study of 754 community-dwelling persons aged 70 years or older.13 Potential participants were members of a large health plan in greater New Haven, Connecticut, and were enrolled from March 1998 through October 1999. As previously described,14 persons were oversampled if they were physically frail, as denoted by a timed score >10 seconds on the rapid gait test. Exclusion criteria included significant cognitive impairment with no available proxy, life expectancy less than 12 months, plans to leave the area, and inability to speak English. Participants were initially required to be nondisabled in four basic activities of daily living (bathing, dressing, walking across a room, and transferring from a chair). Eligibility was determined during a screening telephone interview and was confirmed during an in-home assessment. Of the eligible members, 75.2% agreed to participate in the project, and persons who declined to participate did not significantly differ in age or sex from those who were enrolled. The Yale Human Investigation Committee approved the study protocol, and all participants provided verbal informed consent.
Data Collection
From 1998 to 2017, comprehensive home-based assessments were completed by trained research nurses at baseline and at 18-month intervals over 234 months (except at 126 months), and telephone interviews were completed monthly through June 2018, to obtain information on disability over time. For participants who had significant cognitive impairment or who were unavailable, we interviewed a proxy informant using a rigorous protocol with demonstrated reliability and validity.14 All incident NH admissions, including both short- and long-term stays, were identified using the CMS Skilled Nursing Facility claims file and Long Term Care Minimum Data Set. Deaths were ascertained by review of obituaries and/or from a proxy informant, with a completion rate of 100%. A total of 688 participants (91.2%) had died after a median follow-up of 108 months, while 43 participants (5.7%) dropped out of the study after a median follow-up of 27 months. Among all participants, data were otherwise available for 99.2% of 85,531 monthly telephone interviews.
Assembly of Analytic Sample
PEP participants were considered for inclusion in the analytic sample if they had a hospitalization with an ACSC as the primary diagnosis on linked Medicare claims data. The complete list of ACSCs was defined using specifications from the Agency for Healthcare Research and Quality,15 and was assembled using the International Classification of Disease, Ninth Revision, Clinical Modification (ICD-9-CM) classification prior to October 1, 2015, and ICD Tenth Revision, Clinical Modification (ICD-10-CM) classification after October 1, 2015 (Appendix Table 1). Examples of ACSCs include congestive heart failure, dehydration, urinary tract infection, and angina without procedure. As performed previously,16,17 two ACSCs (low birthweight; asthma in younger adults 18-39 years) were not included in this analysis because they were not based on full adult populations.
ACSC-related hospitalizations were included through December 2017. Participants could contribute more than one ACSC-related hospitalization over the course of the study based on the following criteria: (1) participant did not have a prior non-ACSC-related hospitalization within an 18-month interval; (2) participant did not have a prior ACSC-related hospitalization or treat-and-release emergency department (ED) visit within an 18-month interval (to ensure independence of observations if the participant was still recovering from the prior event and because some of the characteristics within Table 1 are susceptible to change in the setting of an intervening event and, hence, would not accurately reflect the status of the participant prior to ACSC-related hospitalization); (3) participant was not admitted from a NH; (4) participant did not have an in-hospital intensive care unit (ICU) stay (because persons with critical illness are a distinct population with frequent disability and prolonged recovery, as previously described18), in-hospital death, or death before first follow-up interview (because our aim was to evaluate disability and recovery after the hospitalization7).
Assembly of the primary analytic sample is depicted in the Appendix Figure. Of the 814 patients who were identified with ACSC-related hospitalizations, 107 had a prior non-ACSC-related hospitalization and 275 had a prior ACSC-related hospitalization or a treat-and-release ED visit within an 18-month interval. Of the remaining 432 ACSC-related hospitalizations, 181 were excluded: 114 patients were admitted from a NH, 38 had an in-hospital ICU stay, 3 died in the hospital, 11 died before their first follow-up interview, and 15 had withdrawn from the study. The primary analytic sample included the remaining 251 ACSC-related hospitalizations, contributed by 196 participants. Specifically, nine participants contributed three ACSC-related hospitalizations each, 37 participants contributed two hospitalizations each, and the remaining 150 participants contributed one hospitalization each. During the 6-month follow-up period, 40 participants contributing ACSC-related hospitalizations died after a median (interquartile range [IQR]) of 4 (2-5) months, and 1 person refused continued participation.
Comprehensive Assessments
During the comprehensive in-home assessments, data were obtained on demographic characteristics. Age was measured in years at the time of the ACSC-related hospitalization. In addition, we describe factors from the comprehensive assessment immediately prior to the ACSC-related hospitalization, grouped into two additional domains related to disability19: health-related and cognitive-psychosocial. The health-related factors included nine self-reported, physician-diagnosed chronic conditions and frailty. The cognitive-psychosocial factors included social support, cognitive impairment, and depressive symptoms.
Assessment of Disability
Complete details about the assessment of disability have been previously described.13,14,19,20 Briefly, disability was assessed during the monthly telephone interviews, and included four basic activities (bathing, dressing, walking across a room, and transferring from a chair), five instrumental activities (shopping, housework, meal preparation, taking medications, and managing finances), and three mobility activities (walking a quarter mile, climbing a flight of stairs, and lifting or carrying 10 lb). Participants were asked, “At the present time, do you need help from another person to [complete the task]?” Disability was operationalized as the need for personal assistance or an inability to perform the task. Participants were also asked about a fourth mobility activity, “Have you driven a car during the past month?” Those who responded no were classified as being disabled in driving.19
The number of disabilities overall and for each functional domain (basic, instrumental, and mobility) was summed. Possible disability scores ranged from 0 to 13, with a score of 0 indicating complete independence in all of the items, and a score of 13 indicating complete dependence. Worse postdischarge disability was defined as a total disability score (0-13) at the first telephone interview after an ACSC-related hospitalization that was greater than the total disability score from the telephone interview immediately preceding hospitalization.
Outcome Measures
The primary outcome was the number of disabilities in all 13 basic, instrumental, and mobility activities in each of the 6 months following discharge from an ACSC-related hospitalization. To determine whether our findings were consistent across the three functional domains, we also evaluated the number of disabilities in the four basic, five instrumental, and four mobility activities separately. As secondary outcomes, we evaluated: (1) the cumulative probability of recovery within the 6-month follow-up time frame after an ACSC-related hospitalization, with “recovery” defined as return to the participant’s pre-ACSC-related hospitalization total disability score, and (2) the cumulative probability of incident NH admission within the 6 months after an ACSC-related hospitalization. Aligned with CMS and prior literature,21,22 we defined a short-term NH stay as ≤100 days and a long-term NH stay as >100 days.
Statistical Analysis
Pre-ACSC-related hospitalization characteristics were summarized by means (SDs) and frequencies with proportions. We determined the mean number of disabilities in each of the 6 months following hospital discharge, with the prehospitalization value included as a reference point. We also determined the mean (SD) number of disabilities for the three subscales of disability (basic activities of daily living [BADLs], instrumental activities of daily living [IADLs], and mobility activities). We calculated the cumulative probability of recovery within 6 months of hospital discharge. Finally, we determined the cumulative probability of incident NH admission during the 6 months after hospital discharge.
To test the robustness of our main results, we conducted a sensitivity analysis assessing disability scores of the 150 participants that contributed only one ACSC-related hospitalization. All analyses were performed using Stata, version 16.0, statistical software (StataCorp).
RESULTS
Table 1 shows the characteristics of the 251 ACSC-related hospitalizations immediately prior to hospitalization. Participants’ mean (SD) age was 85.1 (6.0) years, and the mean total disability score was 5.4. The majority were female, non-Hispanic White, frail, and lived alone. As shown in Appendix Table 2, the three most common reasons for ACSC-related hospitalizations were congestive heart failure (n = 69), bacterial pneumonia (n = 53), and dehydration (n = 44).
The Figure shows the disability scores during the 6-month follow-up period for total, basic, instrumental, and mobility activities, in panels A, B, C, and D, respectively. The exact values are provided in Appendix Table 3. After hospitalization, disability scores for total, basic, instrumental, and mobility activities peaked at month 1 and tended to improve modestly over the next 5 months, but remained greater, on average, than pre-hospitalization scores. Of the 40 participants who died within the 6-month follow-up period, 36 (90%) had worse disability scores in their last month of life than in the month prior to their ACSC-related hospitalization.
Table 2 shows the cumulative probability of functional recovery after ACSC-related hospitalizations. Recovery was incomplete, with only 70% (95% CI, 64%-76%) of hospitalizations achieving a return to the pre-hospitalization total disability score within 6 months of hospitalization.
Table 3 shows the cumulative probability of incident NH admission after an ACSC-related hospitalization. Of the 251 ACSC-related hospitalizations, incident NH admission was experienced by 38% (95% CI, 32%-44%) within 1 month and 50% (95% CI, 43%-56%) within 6 months of discharge. Short-term NH stays accounted for 90 (75.6%) of the 119 incident NH admissions within the 6 months after ACSC-related hospitalizations. Sensitivity analyses yielded comparable disability scores, shown in Appendix Table 4.
DISCUSSION
In this longitudinal study of community-living older persons, we evaluated functional disability, recovery, and incident NH admission within 6 months of hospitalization for an ACSC. Our study has three major findings. First, disability scores for total, basic, instrumental, and mobility activities at months 1 to 6 of follow-up were greater on average than pre-hospitalization scores. Second, functional recovery was not achieved by 3 of 10 participants after an ACSC-related hospitalization. Third, half of them experienced an incident NH admission within 6 months of discharge from an ACSC-related hospitalization, although about three-quarters of these were short-term stays. Our findings provide evidence that older persons experience clinically meaningful adverse patient-reported outcomes after ACSC-related hospitalizations.
Prior research involving ACSCs has focused largely on rates of hospitalization as a measure of access to primary care and the associated factors predictive of ACSC-related hospitalizations,23-26 and has not addressed subsequent patient-reported outcomes. The findings in this analysis highlight that older persons experience worsening disability immediately after an ACSC-related hospitalization, which persists for prolonged periods and often results in incomplete recovery. Prior research has assessed pre-hospitalization functional status through retrospective recall approaches,2 included only older adults discharged with incident disability,3 and examined functional status after all-cause medical illness hospitalizations.5 Our prospective analysis extends the literature by reliably capturing pre-hospital disability scores and uniquely assessing the cohort of older persons hospitalized with ACSCs.
Our work is relevant to the continued evaluation of ACSC-related hospitalizations in national quality measurement and payment initiatives among Medicare beneficiaries. In prior evaluations of ACSC-related quality measures, stakeholders have criticized the measures for limited validity due to a lack of evidence linking each utilization outcome to other patient-centered outcomes.10,27 Our work addresses this gap by demonstrating that ACSC-related hospitalizations are linked to persistent disability, incomplete functional recovery, and incident NH admissions. Given the large body of evidence demonstrating the priority older persons place on these patient-reported outcomes,28,29 our work should reassure policymakers seeking to transform quality measurement programs into a more patient-oriented enterprise.
Our findings have several clinical practice, research, and policy implications. First, more-effective clinical strategies to minimize the level of care required for acute exacerbations of ACSC-related illnesses may include: (1) substituting home-based care30 and telehealth interventions31 for traditional inpatient hospitalization, (2) making in-ED resources (ie, case management services, geriatric-focused advanced practice providers) more accessible for older persons with ACSC-related illnesses, thereby enhancing care transitions and follow-up to avoid potential current and subsequent hospitalizations, and (3) ensuring adequate ambulatory care access to all older persons, as prior work has shown variation in ACSC hospital admission rates dependent on population factors such as high-poverty neighborhoods,16 insurance status,16,32 and race/ethnicity.33
Clinical strategies have been narrow and not holistic for ACSCs; for example, many institutions have focused on pneumonia vaccinations to reduce hospitalizations, but our work supports the need to further evaluate the impact of preventing ACSC-related hospitalizations and their associated disabling consequences. For patients admitted to the hospital, clinical strategies, such as in-hospital or post-hospital mobility and activity programs, have been shown to be protective against hospital-associated disability.34,35 Furthermore, hospital discharge planning could include preparing older persons for anticipated functional disabilities, associated recoveries, and NH admission after ACSC-related hospitalizations. Risk factors contributing to post-hospitalization functional disability and recovery have been identified,19,20,36 but future work is needed to: (1) identify target populations (including those most likely to worsen) so that interventions can be offered earlier in the course of care to those who would benefit most, and (2) identify and learn from those who are resilient and have recovered, to better understand factors contributing to their success.
Our study has several strengths. First, the study is unique due to its longitudinal design, with monthly assessments of functional status. Since functional status was assessed prospectively before the ACSC-related hospitalization, we also have avoided any potential concern for recall bias that may be present if assessed after the hospitalization. Additionally, through the use of Medicare claims and the Minimum Data Set, the ascertainment of hospitalizations and NH admissions was likely complete for the studied population.
However, the study has limitations. First, functional measures were based on self-reports rather than objective measurements. Nevertheless, the self-report function is often used to guide coverage determinations in the Medicare program, as it has been shown to be associated with poor health outcomes.37 Second, we are unable to comment on the rate of functional decline or NH admission when an older person was not hospitalized in relation to an ACSC. Future analyses may benefit from using a control group (eg, older adults without an ACSC hospitalization or older adults with a non-ACSC hospitalization). Third, we used strict exclusion criteria to identify a population of older adults without recent hospitalizations to determine the isolated impact of ACSC hospitalization on disability, incident NH admission, and functional recovery. Considering this potential selection bias, our findings are likely conservative estimates of the patient-centered outcomes evaluated. Fourth, participants were not asked about feeding and toileting. However, the incidence of disability in these ADLs is low among nondisabled, community-living older persons, and it is highly uncommon for disability to develop in these ADLs without concurrent disability in the ADLs within this analysis.14,38
Finally, because our study participants were members of a single health plan in a small urban area and included nondisabled older persons living in the community, our findings may not be generalizable to geriatric patients in other settings. Nonetheless, the demographics of our cohort reflect those of older persons in New Haven County, Connecticut, which are similar to the demographics of the US population, with the exception of race and ethnicity. In addition, the generalizability of our results are strengthened by the study’s high participation rate and minimal attrition.
CONCLUSION
Within 6 months of ACSC-related hospitalizations, community-living older persons exhibited greater total disability scores than those immediately preceding hospitalization. In the same time frame, 3 of 10 older persons did not achieve functional recovery, and half experienced incident NH admission. These results provide evidence regarding the continued recognition of ACSC-related hospitalizations in federal quality measurement and payment programs and suggests the need for preventive and comprehensive interventions to meaningfully improve longitudinal outcomes.
Acknowledgments
We thank Denise Shepard, BSN, MBA, Andrea Benjamin, BSN, Barbara Foster, and Amy Shelton, MPH, for assistance with data collection; Geraldine Hawthorne, BS, for assistance with data entry and management; Peter Charpentier, MPH, for design and development of the study database and participant tracking system; and Joanne McGloin, MDiv, MBA, for leadership and advice as the Project Director. Each of these persons were paid employees of Yale School of Medicine during the conduct of this study.
Acute illnesses requiring hospitalization serve as a sentinel event, with many older adults requiring assistance with activities of daily living (ADLs) upon discharge.1-3 Older adults who are frail experience even higher rates of hospital-associated disability, and rates of recovery to baseline functional status have varied.4,5 Loss of independence in ADLs has been associated with nursing home (NH) utilization, caregiver burden, and mortality.6
To date, studies have characterized functional trajectories before and after hospitalization in older persons for broad medical conditions, noting persistence of disability and incomplete recovery to baseline functional status.7 Prior evaluations have also noted the long-term disabling impact of critical conditions such as acute myocardial infarction, stroke, and sepsis,8,9 but a knowledge gap exists regarding the subsequent functional disability, recovery, and incident NH admission among older persons who are hospitalized for ambulatory care sensitive conditions (ACSCs). Often considered potentially preventable with optimal ambulatory care,10,11 ACSCs represent acute, chronic, and vaccine-preventable conditions, including urinary tract infection, congestive heart failure, diabetes mellitus, and pneumonia. Investigating the aforementioned patient-centered measures post hospitalization could provide valuable supporting evidence for the continued recognition of ACSC-related hospitalizations in national quality payment programs set forth by the Centers for Medicare & Medicaid Services (CMS).12 Demonstrating adverse outcomes after ACSC-related hospitalizations may help support interventions that target potentially preventable ACSC-related hospitalizations, such as home-based care or telehealth, with the goal of improving functional outcomes and reducing NH admission in older persons.
To address these gaps, we evaluated ACSC-related hospitalizations among participants of the Precipitating Events Project (PEP), a 19-year longitudinal study of community-living persons who were initially nondisabled in their basic functional activities. In the 6 months following an ACSC-related hospitalization, our objectives were to describe: (1) the 6-month course of postdischarge functional disability, (2) the cumulative monthly probability of functional recovery, and (3) the cumulative monthly probability of incident NH admission.
METHODS
Study Population
Participants were drawn from the PEP study, an ongoing, prospective, longitudinal study of 754 community-dwelling persons aged 70 years or older.13 Potential participants were members of a large health plan in greater New Haven, Connecticut, and were enrolled from March 1998 through October 1999. As previously described,14 persons were oversampled if they were physically frail, as denoted by a timed score >10 seconds on the rapid gait test. Exclusion criteria included significant cognitive impairment with no available proxy, life expectancy less than 12 months, plans to leave the area, and inability to speak English. Participants were initially required to be nondisabled in four basic activities of daily living (bathing, dressing, walking across a room, and transferring from a chair). Eligibility was determined during a screening telephone interview and was confirmed during an in-home assessment. Of the eligible members, 75.2% agreed to participate in the project, and persons who declined to participate did not significantly differ in age or sex from those who were enrolled. The Yale Human Investigation Committee approved the study protocol, and all participants provided verbal informed consent.
Data Collection
From 1998 to 2017, comprehensive home-based assessments were completed by trained research nurses at baseline and at 18-month intervals over 234 months (except at 126 months), and telephone interviews were completed monthly through June 2018, to obtain information on disability over time. For participants who had significant cognitive impairment or who were unavailable, we interviewed a proxy informant using a rigorous protocol with demonstrated reliability and validity.14 All incident NH admissions, including both short- and long-term stays, were identified using the CMS Skilled Nursing Facility claims file and Long Term Care Minimum Data Set. Deaths were ascertained by review of obituaries and/or from a proxy informant, with a completion rate of 100%. A total of 688 participants (91.2%) had died after a median follow-up of 108 months, while 43 participants (5.7%) dropped out of the study after a median follow-up of 27 months. Among all participants, data were otherwise available for 99.2% of 85,531 monthly telephone interviews.
Assembly of Analytic Sample
PEP participants were considered for inclusion in the analytic sample if they had a hospitalization with an ACSC as the primary diagnosis on linked Medicare claims data. The complete list of ACSCs was defined using specifications from the Agency for Healthcare Research and Quality,15 and was assembled using the International Classification of Disease, Ninth Revision, Clinical Modification (ICD-9-CM) classification prior to October 1, 2015, and ICD Tenth Revision, Clinical Modification (ICD-10-CM) classification after October 1, 2015 (Appendix Table 1). Examples of ACSCs include congestive heart failure, dehydration, urinary tract infection, and angina without procedure. As performed previously,16,17 two ACSCs (low birthweight; asthma in younger adults 18-39 years) were not included in this analysis because they were not based on full adult populations.
ACSC-related hospitalizations were included through December 2017. Participants could contribute more than one ACSC-related hospitalization over the course of the study based on the following criteria: (1) participant did not have a prior non-ACSC-related hospitalization within an 18-month interval; (2) participant did not have a prior ACSC-related hospitalization or treat-and-release emergency department (ED) visit within an 18-month interval (to ensure independence of observations if the participant was still recovering from the prior event and because some of the characteristics within Table 1 are susceptible to change in the setting of an intervening event and, hence, would not accurately reflect the status of the participant prior to ACSC-related hospitalization); (3) participant was not admitted from a NH; (4) participant did not have an in-hospital intensive care unit (ICU) stay (because persons with critical illness are a distinct population with frequent disability and prolonged recovery, as previously described18), in-hospital death, or death before first follow-up interview (because our aim was to evaluate disability and recovery after the hospitalization7).
Assembly of the primary analytic sample is depicted in the Appendix Figure. Of the 814 patients who were identified with ACSC-related hospitalizations, 107 had a prior non-ACSC-related hospitalization and 275 had a prior ACSC-related hospitalization or a treat-and-release ED visit within an 18-month interval. Of the remaining 432 ACSC-related hospitalizations, 181 were excluded: 114 patients were admitted from a NH, 38 had an in-hospital ICU stay, 3 died in the hospital, 11 died before their first follow-up interview, and 15 had withdrawn from the study. The primary analytic sample included the remaining 251 ACSC-related hospitalizations, contributed by 196 participants. Specifically, nine participants contributed three ACSC-related hospitalizations each, 37 participants contributed two hospitalizations each, and the remaining 150 participants contributed one hospitalization each. During the 6-month follow-up period, 40 participants contributing ACSC-related hospitalizations died after a median (interquartile range [IQR]) of 4 (2-5) months, and 1 person refused continued participation.
Comprehensive Assessments
During the comprehensive in-home assessments, data were obtained on demographic characteristics. Age was measured in years at the time of the ACSC-related hospitalization. In addition, we describe factors from the comprehensive assessment immediately prior to the ACSC-related hospitalization, grouped into two additional domains related to disability19: health-related and cognitive-psychosocial. The health-related factors included nine self-reported, physician-diagnosed chronic conditions and frailty. The cognitive-psychosocial factors included social support, cognitive impairment, and depressive symptoms.
Assessment of Disability
Complete details about the assessment of disability have been previously described.13,14,19,20 Briefly, disability was assessed during the monthly telephone interviews, and included four basic activities (bathing, dressing, walking across a room, and transferring from a chair), five instrumental activities (shopping, housework, meal preparation, taking medications, and managing finances), and three mobility activities (walking a quarter mile, climbing a flight of stairs, and lifting or carrying 10 lb). Participants were asked, “At the present time, do you need help from another person to [complete the task]?” Disability was operationalized as the need for personal assistance or an inability to perform the task. Participants were also asked about a fourth mobility activity, “Have you driven a car during the past month?” Those who responded no were classified as being disabled in driving.19
The number of disabilities overall and for each functional domain (basic, instrumental, and mobility) was summed. Possible disability scores ranged from 0 to 13, with a score of 0 indicating complete independence in all of the items, and a score of 13 indicating complete dependence. Worse postdischarge disability was defined as a total disability score (0-13) at the first telephone interview after an ACSC-related hospitalization that was greater than the total disability score from the telephone interview immediately preceding hospitalization.
Outcome Measures
The primary outcome was the number of disabilities in all 13 basic, instrumental, and mobility activities in each of the 6 months following discharge from an ACSC-related hospitalization. To determine whether our findings were consistent across the three functional domains, we also evaluated the number of disabilities in the four basic, five instrumental, and four mobility activities separately. As secondary outcomes, we evaluated: (1) the cumulative probability of recovery within the 6-month follow-up time frame after an ACSC-related hospitalization, with “recovery” defined as return to the participant’s pre-ACSC-related hospitalization total disability score, and (2) the cumulative probability of incident NH admission within the 6 months after an ACSC-related hospitalization. Aligned with CMS and prior literature,21,22 we defined a short-term NH stay as ≤100 days and a long-term NH stay as >100 days.
Statistical Analysis
Pre-ACSC-related hospitalization characteristics were summarized by means (SDs) and frequencies with proportions. We determined the mean number of disabilities in each of the 6 months following hospital discharge, with the prehospitalization value included as a reference point. We also determined the mean (SD) number of disabilities for the three subscales of disability (basic activities of daily living [BADLs], instrumental activities of daily living [IADLs], and mobility activities). We calculated the cumulative probability of recovery within 6 months of hospital discharge. Finally, we determined the cumulative probability of incident NH admission during the 6 months after hospital discharge.
To test the robustness of our main results, we conducted a sensitivity analysis assessing disability scores of the 150 participants that contributed only one ACSC-related hospitalization. All analyses were performed using Stata, version 16.0, statistical software (StataCorp).
RESULTS
Table 1 shows the characteristics of the 251 ACSC-related hospitalizations immediately prior to hospitalization. Participants’ mean (SD) age was 85.1 (6.0) years, and the mean total disability score was 5.4. The majority were female, non-Hispanic White, frail, and lived alone. As shown in Appendix Table 2, the three most common reasons for ACSC-related hospitalizations were congestive heart failure (n = 69), bacterial pneumonia (n = 53), and dehydration (n = 44).
The Figure shows the disability scores during the 6-month follow-up period for total, basic, instrumental, and mobility activities, in panels A, B, C, and D, respectively. The exact values are provided in Appendix Table 3. After hospitalization, disability scores for total, basic, instrumental, and mobility activities peaked at month 1 and tended to improve modestly over the next 5 months, but remained greater, on average, than pre-hospitalization scores. Of the 40 participants who died within the 6-month follow-up period, 36 (90%) had worse disability scores in their last month of life than in the month prior to their ACSC-related hospitalization.
Table 2 shows the cumulative probability of functional recovery after ACSC-related hospitalizations. Recovery was incomplete, with only 70% (95% CI, 64%-76%) of hospitalizations achieving a return to the pre-hospitalization total disability score within 6 months of hospitalization.
Table 3 shows the cumulative probability of incident NH admission after an ACSC-related hospitalization. Of the 251 ACSC-related hospitalizations, incident NH admission was experienced by 38% (95% CI, 32%-44%) within 1 month and 50% (95% CI, 43%-56%) within 6 months of discharge. Short-term NH stays accounted for 90 (75.6%) of the 119 incident NH admissions within the 6 months after ACSC-related hospitalizations. Sensitivity analyses yielded comparable disability scores, shown in Appendix Table 4.
DISCUSSION
In this longitudinal study of community-living older persons, we evaluated functional disability, recovery, and incident NH admission within 6 months of hospitalization for an ACSC. Our study has three major findings. First, disability scores for total, basic, instrumental, and mobility activities at months 1 to 6 of follow-up were greater on average than pre-hospitalization scores. Second, functional recovery was not achieved by 3 of 10 participants after an ACSC-related hospitalization. Third, half of them experienced an incident NH admission within 6 months of discharge from an ACSC-related hospitalization, although about three-quarters of these were short-term stays. Our findings provide evidence that older persons experience clinically meaningful adverse patient-reported outcomes after ACSC-related hospitalizations.
Prior research involving ACSCs has focused largely on rates of hospitalization as a measure of access to primary care and the associated factors predictive of ACSC-related hospitalizations,23-26 and has not addressed subsequent patient-reported outcomes. The findings in this analysis highlight that older persons experience worsening disability immediately after an ACSC-related hospitalization, which persists for prolonged periods and often results in incomplete recovery. Prior research has assessed pre-hospitalization functional status through retrospective recall approaches,2 included only older adults discharged with incident disability,3 and examined functional status after all-cause medical illness hospitalizations.5 Our prospective analysis extends the literature by reliably capturing pre-hospital disability scores and uniquely assessing the cohort of older persons hospitalized with ACSCs.
Our work is relevant to the continued evaluation of ACSC-related hospitalizations in national quality measurement and payment initiatives among Medicare beneficiaries. In prior evaluations of ACSC-related quality measures, stakeholders have criticized the measures for limited validity due to a lack of evidence linking each utilization outcome to other patient-centered outcomes.10,27 Our work addresses this gap by demonstrating that ACSC-related hospitalizations are linked to persistent disability, incomplete functional recovery, and incident NH admissions. Given the large body of evidence demonstrating the priority older persons place on these patient-reported outcomes,28,29 our work should reassure policymakers seeking to transform quality measurement programs into a more patient-oriented enterprise.
Our findings have several clinical practice, research, and policy implications. First, more-effective clinical strategies to minimize the level of care required for acute exacerbations of ACSC-related illnesses may include: (1) substituting home-based care30 and telehealth interventions31 for traditional inpatient hospitalization, (2) making in-ED resources (ie, case management services, geriatric-focused advanced practice providers) more accessible for older persons with ACSC-related illnesses, thereby enhancing care transitions and follow-up to avoid potential current and subsequent hospitalizations, and (3) ensuring adequate ambulatory care access to all older persons, as prior work has shown variation in ACSC hospital admission rates dependent on population factors such as high-poverty neighborhoods,16 insurance status,16,32 and race/ethnicity.33
Clinical strategies have been narrow and not holistic for ACSCs; for example, many institutions have focused on pneumonia vaccinations to reduce hospitalizations, but our work supports the need to further evaluate the impact of preventing ACSC-related hospitalizations and their associated disabling consequences. For patients admitted to the hospital, clinical strategies, such as in-hospital or post-hospital mobility and activity programs, have been shown to be protective against hospital-associated disability.34,35 Furthermore, hospital discharge planning could include preparing older persons for anticipated functional disabilities, associated recoveries, and NH admission after ACSC-related hospitalizations. Risk factors contributing to post-hospitalization functional disability and recovery have been identified,19,20,36 but future work is needed to: (1) identify target populations (including those most likely to worsen) so that interventions can be offered earlier in the course of care to those who would benefit most, and (2) identify and learn from those who are resilient and have recovered, to better understand factors contributing to their success.
Our study has several strengths. First, the study is unique due to its longitudinal design, with monthly assessments of functional status. Since functional status was assessed prospectively before the ACSC-related hospitalization, we also have avoided any potential concern for recall bias that may be present if assessed after the hospitalization. Additionally, through the use of Medicare claims and the Minimum Data Set, the ascertainment of hospitalizations and NH admissions was likely complete for the studied population.
However, the study has limitations. First, functional measures were based on self-reports rather than objective measurements. Nevertheless, the self-report function is often used to guide coverage determinations in the Medicare program, as it has been shown to be associated with poor health outcomes.37 Second, we are unable to comment on the rate of functional decline or NH admission when an older person was not hospitalized in relation to an ACSC. Future analyses may benefit from using a control group (eg, older adults without an ACSC hospitalization or older adults with a non-ACSC hospitalization). Third, we used strict exclusion criteria to identify a population of older adults without recent hospitalizations to determine the isolated impact of ACSC hospitalization on disability, incident NH admission, and functional recovery. Considering this potential selection bias, our findings are likely conservative estimates of the patient-centered outcomes evaluated. Fourth, participants were not asked about feeding and toileting. However, the incidence of disability in these ADLs is low among nondisabled, community-living older persons, and it is highly uncommon for disability to develop in these ADLs without concurrent disability in the ADLs within this analysis.14,38
Finally, because our study participants were members of a single health plan in a small urban area and included nondisabled older persons living in the community, our findings may not be generalizable to geriatric patients in other settings. Nonetheless, the demographics of our cohort reflect those of older persons in New Haven County, Connecticut, which are similar to the demographics of the US population, with the exception of race and ethnicity. In addition, the generalizability of our results are strengthened by the study’s high participation rate and minimal attrition.
CONCLUSION
Within 6 months of ACSC-related hospitalizations, community-living older persons exhibited greater total disability scores than those immediately preceding hospitalization. In the same time frame, 3 of 10 older persons did not achieve functional recovery, and half experienced incident NH admission. These results provide evidence regarding the continued recognition of ACSC-related hospitalizations in federal quality measurement and payment programs and suggests the need for preventive and comprehensive interventions to meaningfully improve longitudinal outcomes.
Acknowledgments
We thank Denise Shepard, BSN, MBA, Andrea Benjamin, BSN, Barbara Foster, and Amy Shelton, MPH, for assistance with data collection; Geraldine Hawthorne, BS, for assistance with data entry and management; Peter Charpentier, MPH, for design and development of the study database and participant tracking system; and Joanne McGloin, MDiv, MBA, for leadership and advice as the Project Director. Each of these persons were paid employees of Yale School of Medicine during the conduct of this study.
1. Covinsky KE, Pierluissi E, Johnston CB. Hospitalization-associated disability: “She was probably able to ambulate, but I’m not sure” JAMA. 2011;306(16):1782-1793. https://doi.org/10.1001/jama.2011.1556
2. Covinsky KE, Palmer RM, Fortinsky RH, et al. Loss of independence in activities of daily living in older adults hospitalized with medical illnesses: increased vulnerability with age. J Am Geriatr Soc. 2003;51(4):451-458. https://doi.org/10.1046/j.1532-5415.2003.51152.x
3. Barnes DE, Mehta KM, Boscardin WJ, et al. Prediction of recovery, dependence or death in elders who become disabled during hospitalization. J Gen Intern Med. 2013;28(2):261-268. https://doi.org/10.1007/s11606-012-2226-y
4. Gill TM, Allore HG, Gahbauer EA, Murphy TE. Change in disability after hospitalization or restricted activity in older persons. JAMA. 2010;304(17):1919-1928. https://doi.org/10.1001/jama.2010.1568
5. Boyd CM, Landefeld CS, Counsell SR, et al. Recovery of activities of daily living in older adults after hospitalization for acute medical illness. J Am Geriatr Soc. 2008;56(12):2171-2179. https://doi.org/10.1111/j.1532-5415.2008.02023.x
6. Loyd C, Markland AD, Zhang Y, et al. Prevalence of hospital-associated disability in older adults: a meta-analysis. J Am Med Dir Assoc. 2020;21(4):455-461. https://doi.org/10.1016/j.jamda.2019.09.015
7. Dharmarajan K, Han L, Gahbauer EA, Leo-Summers LS, Gill TM. Disability and recovery after hospitalization for medical illness among community-living older persons: a prospective cohort study. J Am Geriatr Soc. 2020;68(3):486-495. https://doi.org/10.1111/jgs.16350
8. Levine DA, Davydow DS, Hough CL, Langa KM, Rogers MAM, Iwashyna TJ. Functional disability and cognitive impairment after hospitalization for myocardial infarction and stroke. Circ Cardiovasc Qual Outcomes. 2014;7(6):863-871. https://doi.org/10.1161/HCQ.0000000000000008
9. Iwashyna TJ, Ely EW, Smith DM, Langa KM. Long-term cognitive impairment and functional disability among survivors of severe sepsis. JAMA. 2010;304(16):1787-1794. https://doi.org/10.1001/jama.2010.1553
10. Hodgson K, Deeny SR, Steventon A. Ambulatory care-sensitive conditions: their potential uses and limitations. BMJ Qual Saf. 2019;28(6):429-433. https://doi.org/10.1136/bmjqs-2018-008820
11. Agency for Healthcare Research and Quality (AHRQ). Quality Indicator User Guide: Prevention Quality Indicators (PQI) Composite Measures. Version 2020. Accessed November 10, 2020. https://www.qualityindicators.ahrq.gov/modules/pqi_resources.aspx.
12. Centers for Medicare & Medicaid Services. 2016 Measure information about the hospital admissions for acute and chronic ambulatory care-sensitive condition (ACSC) composite measures, calculated for the 2018 value-based payment modified program. Accessed November 24, 2020. https://www.cms.gov/Medicare/Medicare-Fee-for-Service-Payment/PhysicianFeedbackProgram/Downloads/2016-ACSC-MIF.pdf.
13. Gill TM, Desai MM, Gahbauer EA, Holford TR, Williams CS. Restricted activity among community-living older persons: incidence, precipitants, and health care utilization. Ann Intern Med. 2001;135(5):313-321. https://doi.org/10.7326/0003-4819-135-5-200109040-00007
14. Gill TM, Hardy SE, Williams CS. Underestimation of disability in community-living older persons. J Am Geriatr Soc. 2002;50(9):1492-1497. https://doi.org/10.1046/j.1532-5415.2002.50403.x
15. Agency for Healthcare Research and Quality. Prevention Quality Indicators Technical Specifications Updates—Version v2018 and 2018.0.1 (ICD 10-CM/PCS), June 2018. Accessed February 4, 2020. https://www.qualityindicators.ahrq.gov/Modules/PQI_TechSpec_ICD10_v2018.aspx.
16. Johnson PJ, Ghildayal N, Ward AC, Westgard BC, Boland LL, Hokanson JS. Disparities in potentially avoidable emergency department (ED) care: ED visits for ambulatory care sensitive conditions. Med Care. 2012;50(12):1020-1028. https://doi.org/10.1097/MLR.0b013e318270bad4
17. Galarraga JE, Mutter R, Pines JM. Costs associated with ambulatory care sensitive conditions across hospital-based settings. Acad Emerg Med. 2015;22(2):172-181. https://doi.org/10.1111/acem.12579
18. Ferrante LE, Pisani MA, Murphy TE, Gahbauer EA, Leo-Summers LS, Gill TM. Functional trajectories among older persons before and after critical illness. JAMA Intern Med. 2015;175(4):523-529. https://doi.org/10.1001/jamainternmed.2014.7889
19. Gill TM, Gahbauer EA, Murphy TE, Han L, Allore HG. Risk factors and precipitants of long-term disability in community mobility: a cohort study of older persons. Ann Intern Med. 2012;156(2):131-140. https://doi.org/10.7326/0003-4819-156-2-201201170-00009
20. Hardy SE, Gill TM. Factors associated with recovery of independence among newly disabled older persons. Arch Intern Med. 2005;165(1):106-112. https://doi.org/10.1001/archinte.165.1.106
21. Centers for Medicare & Medicaid Services. Nursing Home Quality Initiative—Quality Measures. Accessed June 13, 2021. https://www.cms.gov/Medicare/Quality-Initiatives-Patient-Assessment-Instruments/NursingHomeQualityInits/NHQIQualityMeasures
22. Goodwin JS, Li S, Zhou J, Graham JE, Karmarkar A, Ottenbacher K. Comparison of methods to identify long term care nursing home residence with administrative data. BMC Health Serv Res. 2017;17(1):376. https://doi.org/10.1186/s12913-017-2318-9
23. Laditka, JN, Laditka SB, Probst JC. More may be better: evidence of a negative relationship between physician supply and hospitalization for ambulatory care sensitive conditions. Health Serv Res. 2005;40(4):1148-1166. https://doi.org/10.1111/j.1475-6773.2005.00403.x
24. Ansar Z, Laditka JN, Laditka SB. Access to health care and hospitalization for ambulatory care sensitive conditions. Med Care Res Rev. 2006;63(6):719-741. https://doi.org/10.1177/1077558706293637
25. Mackinko J, de Oliveira VB, Turci MA, Guanais FC, Bonolo PF, Lima-Costa MF. The influence of primary care and hospital supply on ambulatory care-sensitive hospitalizations among adults in Brazil, 1999-2007. Am J Public Health. 2011;101(10):1963-1970. https://doi.org/10.2105/AJPH.2010.198887
26. Gibson OR, Segal L, McDermott RA. A systematic review of evidence on the association between hospitalisation for chronic disease related ambulatory care sensitive conditions and primary health care resourcing. BMC Health Serv Res. 2013;13:336. https://doi.org/10.1186/1472-6963-13-336
27. Vuik SI, Fontana G, Mayer E, Darzi A. Do hospitalisations for ambulatory care sensitive conditions reflect low access to primary care? An observational cohort study of primary care usage prior to hospitalisation. BMJ Open. 2017;7(8):e015704. https://doi.org/10.1136/bmjopen-2016-015704
28. Fried TR, Tinetti M, Agostini J, Iannone L, Towle V. Health outcome prioritization to elicit preferences of older persons with multiple health conditions. Patient Educ Couns. 2011;83(2):278-282. https://doi.org/10.1016/j.pec.2010.04.032
29. Reuben DB, Tinetti ME. Goal-oriented patient care—an alternative health outcomes paradigm. N Engl J Med. 2012;366(9):777-779. https://doi.org/10.1056/NEJMp1113631
30. Federman AD, Soones T, DeCherrie LV, Leff B, Siu AL. Association of a bundled hospital-at-home and 30-day postacute transitional care program with clinical outcomes and patient experiences. JAMA Intern Med. 2018;178(8):1033-1040. https://doi.org/10.1001/jamainternmed.2018.2562
31. Shah MN, Wasserman EB, Gillespie SM, et al. High-intensity telemedicine decreases emergency department use for ambulatory care sensitive conditions by older adult senior living community residents. J Am Med Dir Assoc. 2015;16(12):1077-1081. https://doi.org/10.1016/j.jamda.2015.07.009
32. Oster A, Bindman AB. Emergency department visits for ambulatory care sensitive conditions: insights into preventable hospitalizations. Med Care. 2003;41(2):198-207. https://doi.org/10.1097/01.MLR.0000045021.70297.9F
33. O’Neil SS, Lake T, Merrill A, Wilson A, Mann DA, Bartnyska LM. Racial disparities in hospitalizations for ambulatory care-sensitive conditions. Am J Prev Med. 2010;38(4):381-388. https://doi.org/10.1016/j.amepre.2009.12.026
34. Pavon JM, Sloane RJ, Pieper RF, et al. Accelerometer-measured hospital physical activity and hospital-acquired disability in older adults. J Am Geriatr Soc. 2020;68:261-265. https://doi.org/10.1111/jgs.16231
35. Sunde S, Hesseberg K, Skelton DA, et al. Effects of a multicomponent high intensity exercise program on physical function and health-related quality of life in older adults with or at risk of mobility disability after discharge from hospital: a randomised controlled trial. BMC Geriatr. 2020;20(1):464. https://doi.org/10.1186/s12877-020-01829-9
36. Hardy SE, Gill TM. Recovery from disability among community-dwelling older persons. JAMA. 2004;291(13):1596-1602. https://doi.org/10.1001/jama.291.13.1596
37. Rotenberg J, Kinosian B, Boling P, Taler G, Independence at Home Learning Collaborative Writing Group. Home-based primary care: beyond extension of the independence at home demonstration. J Am Geriatr Soc. 2018;66(4):812-817. https://doi.org/10.1111/jgs.15314
38. Rodgers W, Miller B. A comparative analysis of ADL questions in surveys of older people. J Gerontol B Psychol Sci Soc Sci. 1997;52:21-36. https://doi.org/10.1093/geronb/52b.special_issue.21
1. Covinsky KE, Pierluissi E, Johnston CB. Hospitalization-associated disability: “She was probably able to ambulate, but I’m not sure” JAMA. 2011;306(16):1782-1793. https://doi.org/10.1001/jama.2011.1556
2. Covinsky KE, Palmer RM, Fortinsky RH, et al. Loss of independence in activities of daily living in older adults hospitalized with medical illnesses: increased vulnerability with age. J Am Geriatr Soc. 2003;51(4):451-458. https://doi.org/10.1046/j.1532-5415.2003.51152.x
3. Barnes DE, Mehta KM, Boscardin WJ, et al. Prediction of recovery, dependence or death in elders who become disabled during hospitalization. J Gen Intern Med. 2013;28(2):261-268. https://doi.org/10.1007/s11606-012-2226-y
4. Gill TM, Allore HG, Gahbauer EA, Murphy TE. Change in disability after hospitalization or restricted activity in older persons. JAMA. 2010;304(17):1919-1928. https://doi.org/10.1001/jama.2010.1568
5. Boyd CM, Landefeld CS, Counsell SR, et al. Recovery of activities of daily living in older adults after hospitalization for acute medical illness. J Am Geriatr Soc. 2008;56(12):2171-2179. https://doi.org/10.1111/j.1532-5415.2008.02023.x
6. Loyd C, Markland AD, Zhang Y, et al. Prevalence of hospital-associated disability in older adults: a meta-analysis. J Am Med Dir Assoc. 2020;21(4):455-461. https://doi.org/10.1016/j.jamda.2019.09.015
7. Dharmarajan K, Han L, Gahbauer EA, Leo-Summers LS, Gill TM. Disability and recovery after hospitalization for medical illness among community-living older persons: a prospective cohort study. J Am Geriatr Soc. 2020;68(3):486-495. https://doi.org/10.1111/jgs.16350
8. Levine DA, Davydow DS, Hough CL, Langa KM, Rogers MAM, Iwashyna TJ. Functional disability and cognitive impairment after hospitalization for myocardial infarction and stroke. Circ Cardiovasc Qual Outcomes. 2014;7(6):863-871. https://doi.org/10.1161/HCQ.0000000000000008
9. Iwashyna TJ, Ely EW, Smith DM, Langa KM. Long-term cognitive impairment and functional disability among survivors of severe sepsis. JAMA. 2010;304(16):1787-1794. https://doi.org/10.1001/jama.2010.1553
10. Hodgson K, Deeny SR, Steventon A. Ambulatory care-sensitive conditions: their potential uses and limitations. BMJ Qual Saf. 2019;28(6):429-433. https://doi.org/10.1136/bmjqs-2018-008820
11. Agency for Healthcare Research and Quality (AHRQ). Quality Indicator User Guide: Prevention Quality Indicators (PQI) Composite Measures. Version 2020. Accessed November 10, 2020. https://www.qualityindicators.ahrq.gov/modules/pqi_resources.aspx.
12. Centers for Medicare & Medicaid Services. 2016 Measure information about the hospital admissions for acute and chronic ambulatory care-sensitive condition (ACSC) composite measures, calculated for the 2018 value-based payment modified program. Accessed November 24, 2020. https://www.cms.gov/Medicare/Medicare-Fee-for-Service-Payment/PhysicianFeedbackProgram/Downloads/2016-ACSC-MIF.pdf.
13. Gill TM, Desai MM, Gahbauer EA, Holford TR, Williams CS. Restricted activity among community-living older persons: incidence, precipitants, and health care utilization. Ann Intern Med. 2001;135(5):313-321. https://doi.org/10.7326/0003-4819-135-5-200109040-00007
14. Gill TM, Hardy SE, Williams CS. Underestimation of disability in community-living older persons. J Am Geriatr Soc. 2002;50(9):1492-1497. https://doi.org/10.1046/j.1532-5415.2002.50403.x
15. Agency for Healthcare Research and Quality. Prevention Quality Indicators Technical Specifications Updates—Version v2018 and 2018.0.1 (ICD 10-CM/PCS), June 2018. Accessed February 4, 2020. https://www.qualityindicators.ahrq.gov/Modules/PQI_TechSpec_ICD10_v2018.aspx.
16. Johnson PJ, Ghildayal N, Ward AC, Westgard BC, Boland LL, Hokanson JS. Disparities in potentially avoidable emergency department (ED) care: ED visits for ambulatory care sensitive conditions. Med Care. 2012;50(12):1020-1028. https://doi.org/10.1097/MLR.0b013e318270bad4
17. Galarraga JE, Mutter R, Pines JM. Costs associated with ambulatory care sensitive conditions across hospital-based settings. Acad Emerg Med. 2015;22(2):172-181. https://doi.org/10.1111/acem.12579
18. Ferrante LE, Pisani MA, Murphy TE, Gahbauer EA, Leo-Summers LS, Gill TM. Functional trajectories among older persons before and after critical illness. JAMA Intern Med. 2015;175(4):523-529. https://doi.org/10.1001/jamainternmed.2014.7889
19. Gill TM, Gahbauer EA, Murphy TE, Han L, Allore HG. Risk factors and precipitants of long-term disability in community mobility: a cohort study of older persons. Ann Intern Med. 2012;156(2):131-140. https://doi.org/10.7326/0003-4819-156-2-201201170-00009
20. Hardy SE, Gill TM. Factors associated with recovery of independence among newly disabled older persons. Arch Intern Med. 2005;165(1):106-112. https://doi.org/10.1001/archinte.165.1.106
21. Centers for Medicare & Medicaid Services. Nursing Home Quality Initiative—Quality Measures. Accessed June 13, 2021. https://www.cms.gov/Medicare/Quality-Initiatives-Patient-Assessment-Instruments/NursingHomeQualityInits/NHQIQualityMeasures
22. Goodwin JS, Li S, Zhou J, Graham JE, Karmarkar A, Ottenbacher K. Comparison of methods to identify long term care nursing home residence with administrative data. BMC Health Serv Res. 2017;17(1):376. https://doi.org/10.1186/s12913-017-2318-9
23. Laditka, JN, Laditka SB, Probst JC. More may be better: evidence of a negative relationship between physician supply and hospitalization for ambulatory care sensitive conditions. Health Serv Res. 2005;40(4):1148-1166. https://doi.org/10.1111/j.1475-6773.2005.00403.x
24. Ansar Z, Laditka JN, Laditka SB. Access to health care and hospitalization for ambulatory care sensitive conditions. Med Care Res Rev. 2006;63(6):719-741. https://doi.org/10.1177/1077558706293637
25. Mackinko J, de Oliveira VB, Turci MA, Guanais FC, Bonolo PF, Lima-Costa MF. The influence of primary care and hospital supply on ambulatory care-sensitive hospitalizations among adults in Brazil, 1999-2007. Am J Public Health. 2011;101(10):1963-1970. https://doi.org/10.2105/AJPH.2010.198887
26. Gibson OR, Segal L, McDermott RA. A systematic review of evidence on the association between hospitalisation for chronic disease related ambulatory care sensitive conditions and primary health care resourcing. BMC Health Serv Res. 2013;13:336. https://doi.org/10.1186/1472-6963-13-336
27. Vuik SI, Fontana G, Mayer E, Darzi A. Do hospitalisations for ambulatory care sensitive conditions reflect low access to primary care? An observational cohort study of primary care usage prior to hospitalisation. BMJ Open. 2017;7(8):e015704. https://doi.org/10.1136/bmjopen-2016-015704
28. Fried TR, Tinetti M, Agostini J, Iannone L, Towle V. Health outcome prioritization to elicit preferences of older persons with multiple health conditions. Patient Educ Couns. 2011;83(2):278-282. https://doi.org/10.1016/j.pec.2010.04.032
29. Reuben DB, Tinetti ME. Goal-oriented patient care—an alternative health outcomes paradigm. N Engl J Med. 2012;366(9):777-779. https://doi.org/10.1056/NEJMp1113631
30. Federman AD, Soones T, DeCherrie LV, Leff B, Siu AL. Association of a bundled hospital-at-home and 30-day postacute transitional care program with clinical outcomes and patient experiences. JAMA Intern Med. 2018;178(8):1033-1040. https://doi.org/10.1001/jamainternmed.2018.2562
31. Shah MN, Wasserman EB, Gillespie SM, et al. High-intensity telemedicine decreases emergency department use for ambulatory care sensitive conditions by older adult senior living community residents. J Am Med Dir Assoc. 2015;16(12):1077-1081. https://doi.org/10.1016/j.jamda.2015.07.009
32. Oster A, Bindman AB. Emergency department visits for ambulatory care sensitive conditions: insights into preventable hospitalizations. Med Care. 2003;41(2):198-207. https://doi.org/10.1097/01.MLR.0000045021.70297.9F
33. O’Neil SS, Lake T, Merrill A, Wilson A, Mann DA, Bartnyska LM. Racial disparities in hospitalizations for ambulatory care-sensitive conditions. Am J Prev Med. 2010;38(4):381-388. https://doi.org/10.1016/j.amepre.2009.12.026
34. Pavon JM, Sloane RJ, Pieper RF, et al. Accelerometer-measured hospital physical activity and hospital-acquired disability in older adults. J Am Geriatr Soc. 2020;68:261-265. https://doi.org/10.1111/jgs.16231
35. Sunde S, Hesseberg K, Skelton DA, et al. Effects of a multicomponent high intensity exercise program on physical function and health-related quality of life in older adults with or at risk of mobility disability after discharge from hospital: a randomised controlled trial. BMC Geriatr. 2020;20(1):464. https://doi.org/10.1186/s12877-020-01829-9
36. Hardy SE, Gill TM. Recovery from disability among community-dwelling older persons. JAMA. 2004;291(13):1596-1602. https://doi.org/10.1001/jama.291.13.1596
37. Rotenberg J, Kinosian B, Boling P, Taler G, Independence at Home Learning Collaborative Writing Group. Home-based primary care: beyond extension of the independence at home demonstration. J Am Geriatr Soc. 2018;66(4):812-817. https://doi.org/10.1111/jgs.15314
38. Rodgers W, Miller B. A comparative analysis of ADL questions in surveys of older people. J Gerontol B Psychol Sci Soc Sci. 1997;52:21-36. https://doi.org/10.1093/geronb/52b.special_issue.21
© 2021 Society of Hospital Medicine
Excess Mortality Among Patients Hospitalized During the COVID-19 Pandemic
One of the most striking features of the early COVID-19 pandemic was the sudden and sharp reductions in emergency department (ED) visits and hospitalizations throughout the United States.1-4 Several studies have documented lower rates of hospitalization for many emergent, time-sensitive conditions, such as acute myocardial infarction, stroke, and hyperglycemic crises, starting shortly after community transmission of COVID-19 was recognized and social distancing guidelines were implemented.5-8 In most cases, hospital volumes rebounded after an initial drop, stabilizing at somewhat lower levels than those expected from historic trends.9
The observed shifts in hospital use largely have been attributed to patients’ forgoing or delaying necessary care,10 which underscores the indirect effects of the pandemic on patients without COVID-19.11 To date, the extent to which outcomes for patients without COVID-19 have been adversely affected is less well understood. Evidence suggests patients with acute and chronic illnesses have experienced increased morbidity and mortality since the onset of the pandemic. For example, in northern California, abrupt declines in ED visits for cardiac symptoms were coupled with higher rates of out-of-hospital cardiac arrest.12 Moreover, states with higher rates of COVID-19 also reported increased deaths attributed to heart disease, diabetes, and other conditions.13
To better understand these potential indirect effects, this study used data from a large, multistate health care system to examine changes in hospital volume and its relationship to in-hospital mortality for patients without COVID-19 during the first 10 months of the pandemic.
METHODS
Setting and Participants
We examined unplanned hospitalizations from January 2019 to December 2020 at 51 community hospitals across 6 states (Alaska, Washington, Montana, Oregon, California, and Texas) in the Providence St. Joseph Health system. Hospitals within the Providence system share a common standard dataset for each encounter with a centralized cloud data warehouse from which we extracted clinical and demographic data. No hospitals entered or left the system during the study period. Hospitalizations were considered unplanned if they had an “urgent” or “emergency” service type in the record; most originated in the ED. Hospitalizations for children younger than 18 years and those with evidence of COVID-19 (International Classification of Disease, Tenth Revision, Clinical Modification U07.1, a positive COVID-19 polymerase chain reaction test during the encounter, or an infection control-assigned label of COVID-19) were excluded. The Providence St. Joseph Health Institutional Review Board approved this study.
Measures
Trends in daily hospitalizations and their relationship to adjusted in-hospital mortality (percentage of patients who died during their hospital admission) were examined over time. In preliminary models using segmented regression, we identified three distinct pandemic periods with different trends in daily hospitalizations: (1) a 10-week period corresponding to the spring COVID-19 surge (March 4 to May 13, 2020; Period 1), (2) an intervening period extending over the summer and early fall (May 14 to October 19, 2020; Period 2), and (3) a second 10-week period corresponding to the fall COVID-19 surge (October 20 to December 31, 2020; Period 3). In-hospital mortality for these periods was compared with a baseline period (pre-COVID-19) from January 1, 2019 to March 3, 2020. To further assess differences in mortality by clinical condition, hospitalizations were first grouped by primary diagnosis using Clinical Classifications Software Refined (CCSR) categories from the Agency for Healthcare Research and Quality14 and ranked by the number of observed deaths and the percentage of patients who died while hospitalized in 2020. We selected common conditions that had >35 total deaths and an in-hospital mortality rate ≥1% for condition-specific analyses, of which 30 met these criteria.
Analysis
Multivariate logistic regression was used to evaluate changes in mortality for each of the pandemic periods compared with baseline for the overall cohort and selected diagnosis groups. Our main model adjusted for age, sex, race/ethnicity (White, Black, Latinx, Asian or Pacific Islander, and other), primary payor (commercial, Medicaid, Medicare, other, and self-pay), the presence or absence of 31 chronic comorbidities in the medical record, primary admitting diagnosis grouped by CCSR category (456 total diagnostic groups), and hospital fixed-effects to account for clustering. Results are expressed as the average marginal effects of each pandemic period on in-hospital mortality (eg, adjusted percentage point change in mortality over baseline). The number of excess deaths in each period was calculated by multiplying the estimated percentage point change in mortality for each period by the total number of hospitalizations. These excess deaths were subtracted from the number of observed deaths to derive the number of deaths that would be expected if pre-pandemic mortality rates persisted.
To further assess whether changes in adjusted mortality could be attributed to a smaller, sicker population of patients presenting to the hospital during the pandemic (meaning that less acutely ill patients stayed home), we conducted two sensitivity analyses. First, we tested whether substituting indicators for Medicare Severity Diagnosis Groups (MS-DRG) in lieu of CCSR categories had any impact on our results. MS-DRGs are designed to account for a patient’s illness severity and expected costs, whereas CCSR categories do not.15 MS-DRGs also better distinguish between surgical versus medical conditions. We re-ran our main model using indicators for CCSR to control for diagnostic mix, but further adjusted for severity using the DRG weight for the primary diagnosis and Modified Early Warning Score (MEWS) as continuous covariates. MEWS is a physiologic scoring system that incorporates abnormal vital signs and data related to mental status during the first 24 hours of a patient’s hospitalization into a risk-based score that has been shown to predict hospital mortality and need for intensive care.16,17 These sensitivity analyses were performed on a subset of inpatient admissions because DRG data are not available for hospitalizations billed as an observation stay, and only approximately 70% of hospitals in the sample contributed vital sign data to the Providence data warehouse. All statistical analyses were conducted with R, version 3.6.3 (R Foundation for Statistical Computing) and SAS Enterprise Guide 7.1 (SAS Institute Inc).
RESULTS
The characteristics of our sample are described in Table 1. A total of 61,300, 159,430, and 65,923 hospitalizations occurred in each of the three pandemic periods, respectively, compared with 503,190 hospitalizations in the pre-pandemic period. The mean (SD) age of patients in the study was 63.2 (19.4) years; most were women (52.4%), White (70.6%), and had Medicare as their primary payor (53.7%). Less than half (42.7%) of hospitalizations occurred in California, and just under one-quarter were observation stays (23.2%). Patient characteristics were similar in the pre-COVID-19 and COVID-19 pandemic periods.
Figure 1 shows trends in hospital volume and mortality. Overall daily hospitalizations declined abruptly from a mean of 1176 per day in the pre-pandemic period to 617 per day (47.5% relative decrease) during the first 3 weeks of Period 1. Mean daily hospitalizations began to rise over the next 2 months (Period 1), reaching steady state at <1000 hospitalizations per day (15% relative decrease from baseline) during Period 2. During Period 3, we observed a decline in mean daily hospitalizations, with a low point of 882 per day on December 31, 2020 (25% relative decrease from baseline), corresponding to the end of our study period. Although hospital volumes declined during both COVID-19 surge periods, the percentage of patients who died during their hospitalization increased. There was an initial spike in in-hospital mortality that peaked approximately 1 month into the pandemic (middle of Period 1), a return to levels at or slightly below that before the pandemic by the beginning of Period 2, and then a rise throughout the autumn COVID-19 surge in Period 3, not yet peaking by the end of the study.
Adjusted in-hospital mortality for the three COVID-19 periods compared with the pre-pandemic period is presented in Table 2. The percentage of patients who died during their hospitalization rose from 2.9% in the pre-pandemic period to 3.4% during Period 1 (absolute difference, 0.6 percentage points; 95% CI, 0.5-0.7), corresponding to a 19.3% relative increase during the spring COVID-19 surge. Among the subset of patients hospitalized with 1 of the 30 conditions selected for individual analysis, mortality increased from 5.0% to 5.9% during the same time period (absolute difference, 0.9 percentage points; 95% CI, 0.8-1.1), corresponding to an 18.9% relative increase. In Period 2, in-hospital mortality was similar to that noted pre-pandemic for the overall cohort and the 30 selected conditions. During Period 3, in-hospital mortality increased by a magnitude similar to that observed in Period 1 for all hospitalizations combined (absolute difference, 0.5 percentage points; 95% CI, 0.0-0.6; corresponding to a 16.5% relative increase) as well as the subgroup with 1 of the 30 selected conditions (0.9 percentage points; 95% CI, 0.8-1.0; corresponding to an 18% relative increase). Further adjustment for severity by swapping CCSR categories with MS-DRG indicators or inclusion of DRG weight and MEWS score as covariates in our sensitivity analyses did not change our results.
Table 3 and the Appendix Figure describe changes in volume and adjusted in-hospital mortality for the 30 conditions selected for analysis. There was a decrease in the mean daily admissions for all conditions studied. Among the 30 conditions, 26 showed increased mortality during Period 1, although the increase was only statistically significant for 16 of these conditions. Among the 10 most commonly admitted conditions (by number of daily hospital admissions during the baseline period), there was a statistically significant relative increase in mortality for patients with sepsis (20.1%), heart failure (17.6%), ischemic stroke (12.5%), device/graft/surgical complications (14.0%), cardiac dysrhythmias (14.4%), pneumonia (24.5%), respiratory failure (16.1%), and gastrointestinal hemorrhage (23.3%). In general, mortality returned to baseline or improved during Period 2. Thereafter, all 30 conditions showed increased mortality in Period 3. This increase was significant for only 16 conditions, which were not the same ones noted during Period 1. Of note, although there was higher mortality for some cardiovascular conditions (heart failure cardiac dysrhythmias), mortality for myocardial infarction remained unchanged from baseline across all 3 periods. In contrast, several solid cancer–related conditions showed progressively worsening mortality throughout the study, with 7.7% higher mortality in Period 1, 10.3% higher mortality in Period 2, and 16.5% higher mortality in Period 3, respectively, compared with baseline. Although a similar pattern was observed for acute renal failure and some neurologic conditions (traumatic brain injury, seizure, other nervous system disorders), mortality for drug poisonings and gastrointestinal bleeds improved over time.
DISCUSSION
In this study of unplanned hospitalizations from 51 community hospitals across 6 states in the US West, we found a significant increase in mortality—at a rate of approximately 5 to 6 excess deaths per 1000 hospitalizations—among patients admitted during the pandemic with a variety of non-COVID-19 illnesses and injuries. Higher in-hospital mortality was observed in the spring (March to May) and fall (October to December) of 2020 when COVID-19 case counts surged and shelter-in-place mandates were implemented. With the initial surge, higher mortality rates were largely transient, and, for most conditions evaluated, returned to baseline approximately 3 months after the pandemic onset. For the fall surge, mortality rates had not peaked by the end of the study period. Changes in mortality were closely and inversely correlated with hospital volume for non-COVID-19 illnesses during both surge periods.
Higher morbidity and mortality for patients without COVID-19 appears to be an unfortunate spillover effect that has been reported in several studies. Recent work examining national surveillance data suggest that up to one-third of excess deaths (deaths higher than those expected for season) early in the pandemic have occurred among patients without known COVID-19.13,18-20 Specifically, these studies estimate that mortality rates in the United States increased by 15% to 19% in the spring of 2020; of the identified excess deaths, only 38% to 77% could be attributed to COVID-19, with the remainder attributed to cardiovascular disease, diabetes, and Alzheimer’s disease, among others. In addition, reports from several European countries and China examining population death data have found similar trends,21-25 as well as a recent study examining excess deaths in nursing homes.26 Our results are largely consistent with these earlier studies in that we describe higher mortality in a sample of patients hospitalized with a variety of common conditions that otherwise are routinely treated in US hospitals. Reporting these indirect casualties of COVID-19 is important to fully understand the pandemic’s toll on patients and healthcare systems.
Our work builds on the current body of literature, highlighting the consistent relationship between rising COVID-19 case counts, hospital volume, and excess mortality over more than one surge period. Although several studies have looked at trends in hospital admissions or population mortality rates, few have examined the two outcomes together. The close correlation between daily hospital admissions and in-hospital mortality in this study suggests that the pandemic changed how patients use healthcare resources in ways that were important for their health and outcomes. The higher mortality rate that we and others have observed likely is related to patients’ delaying care because of fear of contracting COVID-19. In one survey, more than 4 in 10 adults in the United States reported that they avoided medical care during the early pandemic.10 Importantly, even a few days delay for many conditions, such as heart failure or sepsis, can result in precipitous declines in clinical status and outcomes.
It also is possible that we found increased rates of in-hospital mortality simply because patients with more moderate illness chose to stay home, resulting in a patient population enriched with those more likely to die. We found mixed evidence in our data that the observed increases in mortality could be attributable to a smaller, sicker population. Some characteristics that might be protective, such as a slightly younger mean age and lower mean DRG weight, were more common among those hospitalized during the pandemic. However, other characteristics, such as a slightly higher MEWS score and a greater percentage of total hospitalizations in the higher mortality subgroup, also were noted during the pandemic (Table 1). We do note, however, that the differences in these severity-related characteristics were small across the study periods. Further adjusting for these characteristics in our sensitivity analyses did not appreciably change our main findings, suggesting that the mortality increase could not be explained by changes in case-mix alone.
Other factors not dependent on patient behavior, such as barriers to accessing timely ambulatory care and impacts in the quality of care delivered, might have contributed. Shelter-in-place orders, reduced in-person access to clinicians in the ambulatory setting, slow implementation of telehealth services (with uncertainty about their equivalence to in-person exams), as well as delays in diagnostic tests and outpatient procedures could have played a role, especially during early months of the pandemic.27 Significant changes to ambulatory health care delivery might have left many patients with chronic illnesses or complex medical needs with limited care options. Importantly, these care interruptions might have had greater implications for some patients, such as those with cancer who rely on intensive, largely outpatient-based treatment.28,29 This, in part, could explain why we found persistently increased mortality among patients hospitalized with cancer after the spring surge. Later into the pandemic, however, most health systems had developed processes that allowed clinicians to resume timely care of ambulatory patients. Because of this, increases in mortality observed during the fall surge likely stem from other factors, such as patient behavior.
It is possible that care delays or changes in the quality of care delivered during the index hospitalization or pre-hospital setting might have contributed to the observed increase in mortality. This is particularly true for acute, time-sensitive conditions such as sepsis and stroke. Extra time spent donning personal protective equipment and/or new protocols instituted during the pandemic likely impacted the speed of emergency medical services transport, timeliness of ED evaluation, and delivery of definitive therapy. Although most hospitals in this study were not overwhelmed by the pandemic, the complexities associated with caring for known and suspected COVID-19 patients alongside those without the disease might have altered ideal care practices and strained healthcare teams.30 In addition, nearly all hospitalized patients during this period were deprived of in-person advocacy by family members, who were not permitted to visit.
Important limitations with this study exist. First, the data come only from hospitals in the western United States. Second, some data elements such as triage scores or vital signs were not available for the entire population, potentially limiting some risk-adjustment. Third, we were unable to determine the root cause of excess mortality based on our study design and the coded variables available. It is unknown to what extent undiagnosed COVID-19 played a role. Early in the pandemic, many community hospitals did not have access to timely COVID-19 testing, and some cases might have not been diagnosed.31 However, we do not expect this to be a significant concern in the later months of the pandemic, as testing became more widespread and hospitals implemented surveillance screening for COVID-19 for inpatients.
CONCLUSIONS
Our study indicates that the COVID-19 pandemic was associated with increased mortality among patients hospitalized for a range of clinical conditions. Although higher observed mortality rates were limited to periods of high COVID-19 activity, future studies will need to tease out the extent to which these findings relate to patient factors (ie, delayed presentation and more severe disease) or systemic factors (reduction in access or changes in quality of care). This is of key importance, and appropriate solutions will need to be developed to mitigate adverse impacts with this and future pandemics.
1. Baum A, Schwartz MD. Admissions to Veterans Affairs hospitals for emergency conditions during the COVID-19 pandemic. JAMA. 2020;324(1):96-99. https://doi.org/10.1001/jama.2020.9972
2. Hartnett KP, Kite-Powell A, DeVies J, et al; National Syndromic Surveillance Program Community of Practice. Impact of the COVID-19 pandemic on emergency department visits — United States, January 1, 2019–May 30, 2020. MMWR Morb Mortal Wkly Rep. 2020;69(23):699-704. https://doi.org/10.15585/mmwr.mm6923e1
3. Birkmeyer JD, Barnato A, Birkmeyer N, Bessler R, Skinner J. The impact of the COVID-19 pandemic on hospital admissions in the United States. Health Aff. 2020;39(11):2010-2017. https://doi.org/10.1377/hlthaff.2020.00980
4. Blecker S, Jones SA, Petrilli CM, et al. Hospitalizations for chronic disease and acute conditions in the time of COVID-19. JAMA Intern Med. 2021;181(2):269-271. https://doi.org/10.1001/jamainternmed.2020.3978
5. Bhambhvani HP, Rodrigues AJ, Yu JS, Carr JB 2nd, Hayden Gephart M. Hospital volumes of 5 medical emergencies in the COVID-19 pandemic in 2 US medical centers. JAMA Intern Med. 2021;181(2):272-274. https://doi.org/10.1001/jamainternmed.2020.3982
6. Lange SJ, Ritchey MD, Goodman AB, et al. Potential indirect effects of the COVID-19 pandemic on use of emergency departments for acute life-threatening conditions — United States, January–May 2020. MMWR Morb Mortal Wkly Rep. 2020;69(25);795-800. https://doi.org/10.15585/mmwr.mm6925e2
7. Solomon MD, McNulty EJ, Rana JS, et al. The Covid-19 pandemic and the incidence of acute myocardial infarction. N Engl J Med. 2020;383(7):691-693. https://doi.org/10.1056/NEJMc2015630
8. Kansagra AP, Goyal MS, Hamilton S, Albers GW. Collateral effect of Covid-19 on stroke evaluation in the United States. N Engl J Med. 2020;383(4):400-401. https://doi.org/10.1056/NEJMc2014816
9. Heist T, Schwartz K, Butler S. Trends in overall and non-COVID-19 hospital admissions. Kaiser Family Foundation. Accessed March 18, 2021. https://www.kff.org/health-costs/issue-brief/trends-in-overall-and-non-covid-19-hospital-admissions
10. Czeisler MÉ, Marynak K, Clarke KEN, et al. Delay or avoidance of medical care because of COVID-19–related concerns — United States, June 2020. MMWR Morb Mortal Wkly Rep. 2020;69(36);1250-1257. https://doi.org/10.15585/mmwr.mm6936a4
11. Chen J, McGeorge R. Spillover effects of the COVID-19 pandemic could drive long-term health consequences for non-COVID-19 patients. Health Affairs Blog. Accessed March 18, 2021. https://www.healthaffairs.org/do/10.1377/hblog20201020.566558/full/
12. Wong LE, Hawkins JE, Langness S, Murrell KL, Iris P, Sammann A. Where are all the patients? Addressing Covid-19 fear to encourage sick patients to seek emergency care. NEJM Catalyst. Accessed March 18, 2021. https://catalyst.nejm.org/doi/abs/10.1056/CAT.20.0193
13. Woolf SH, Chapman DA, Sabo RT, Weinberger DM, Hill L. Excess deaths from COVID-19 and other causes, March-April 2020. JAMA. 2020;324(5):510-513. https://doi.org/10.1001/jama.2020.11787
14. Clinical Classifications Software Refined (CCSR) for ICD-10-CM Diagnoses. Agency for Healthcare Research and Quality, Rockville, MD. Accessed April 22, 2021. https://www.hcup-us.ahrq.gov/toolssoftware/ccsr/dxccsr.jsp
15. MS-DRG Classifications and Software. Centers for Medicare & Medicaid Services. Accessed March 18, 2021. https://www.cms.gov/Medicare/Medicare-Fee-for-Service-Payment/AcuteInpatientPPS/MS-DRG-Classifications-and-Software
16. Jayasundera R, Neilly M, Smith TO, Myint PK. Are early warning scores useful predictors for mortality and morbidity in hospitalised acutely unwell older patients? A systematic review. J Clin Med. 2018;7(10):309. https://doi.org/10.3390/jcm7100309
17. Delgado-Hurtado JJ, Berger A, Bansal AB. Emergency department Modified Early Warning Score association with admission, admission disposition, mortality, and length of stay. J Community Hosp Intern Med Perspect. 2016;6(2):31456. https://doi.org/10.3402/jchimp.v6.31456
18. Woolf SH, Chapman DA, Sabo RT, Weinberger DM, Hill L, Taylor DDH. Excess deaths from COVID-19 and other causes, March-July 2020. JAMA. 2020;324(15):1562-1564. https://doi.org/10.1001/jama.2020.19545
19. Faust JS, Krumholz HM, Du C, et al. All-cause excess mortality and COVID-19–related mortality among US adults aged 25-44 years, March-July 2020. JAMA. 2021;325(8):785-787. https://doi.org/10.1001/jama.2020.24243
20. Weinberger DM, Chen J, Cohen T, et al. Estimation of excess deaths associated with the COVID-19 pandemic in the United States, March to May 2020. JAMA Intern Med. 2020;180(10):1336-1344. https://doi.org/10.1001/jamainternmed.2020.3391
21. Vandoros S. Excess mortality during the Covid-19 pandemic: Early evidence from England and Wales. Soc Sci Med. 2020; 258:113101. https://doi.org/10.1016/j.socscimed.2020.113101
22. Vestergaard LS, Nielsen J, Richter L, et al; ECDC Public Health Emergency Team for COVID-19. Excess all-cause mortality during the COVID-19 pandemic in Europe – preliminary pooled estimates from the EuroMOMO network, March to April 2020. Euro Surveill. 2020;25(26):2001214. https://doi.org/10.2807/1560-7917.ES.2020.25.26.2001214
23. Kontopantelis E, Mamas MA, Deanfield J, Asaria M, Doran T. Excess mortality in England and Wales during the first wave of the COVID-19 pandemic. J Epidemiol Community Health. 2021;75(3):213-223. https://doi.org/10.1136/jech-2020-214764
24. Liu J, Zhang L, Yan Y, et al. Excess mortality in Wuhan city and other parts of China during the three months of the covid-19 outbreak: findings from nationwide mortality registries. BMJ. 2021;372:n415. https://doi.org/10.1136/bmj.n415
25. Docherty KF, Butt JH, de Boer RA, et al. Excess deaths during the Covid-19 pandemic: An international comparison. Preprint. Posted online May 13, 2020. medRxiv. doi:https://doi.org/10.1101/2020.04.21.20073114
26. Barnett ML, Hu L, Martin T, Grabowski DC. Mortality, admissions, and patient census at SNFs in 3 US cities during the COVID-19 pandemic. JAMA. 2020;324(5):507-509. https://doi.org/10.1001/jama.2020.11642
27. Rosenbaum L. The untold toll — The pandemic’s effects on patients without Covid-19. N Engl J Med. 2020; 382:2368-2371 https://doi.org/10.1056/NEJMms2009984
28. Lai AG, Pasea L, Banerjee A, et al. Estimated impact of the COVID-19 pandemic on cancer services and excess 1-year mortality in people with cancer and multimorbidity: near real-time data on cancer care, cancer deaths and a population-based cohort study. BMJ Open. 2020;10(11):e043828. https://doi.org/10.1136/bmjopen-2020-043828
29. Van de Haar J, Hoes LR, Coles CE, et al. Caring for patients with cancer in the COVID-19 era. Nat Med. 2020;26(5):665-671. https://doi.org/10.1038/s41591-020-0874-8
30. Traylor AM, Tannenbaum SI, Thomas EJ, Salas E. Helping healthcare teams save lives during COVID-19: insights and countermeasures from team science. Am Psychol. 2020;76(1):1-13. https://doi.org/10.1037/amp0000750
31. Grimm CA. Hospital experiences responding to the COVID-19 pandemic: results of a National Pulse Survey March 23–27. U.S. Department of Health and Human Services Office of Inspector General; 2020. https://oig.hhs.gov/oei/reports/oei-06-20-00300.pdf
One of the most striking features of the early COVID-19 pandemic was the sudden and sharp reductions in emergency department (ED) visits and hospitalizations throughout the United States.1-4 Several studies have documented lower rates of hospitalization for many emergent, time-sensitive conditions, such as acute myocardial infarction, stroke, and hyperglycemic crises, starting shortly after community transmission of COVID-19 was recognized and social distancing guidelines were implemented.5-8 In most cases, hospital volumes rebounded after an initial drop, stabilizing at somewhat lower levels than those expected from historic trends.9
The observed shifts in hospital use largely have been attributed to patients’ forgoing or delaying necessary care,10 which underscores the indirect effects of the pandemic on patients without COVID-19.11 To date, the extent to which outcomes for patients without COVID-19 have been adversely affected is less well understood. Evidence suggests patients with acute and chronic illnesses have experienced increased morbidity and mortality since the onset of the pandemic. For example, in northern California, abrupt declines in ED visits for cardiac symptoms were coupled with higher rates of out-of-hospital cardiac arrest.12 Moreover, states with higher rates of COVID-19 also reported increased deaths attributed to heart disease, diabetes, and other conditions.13
To better understand these potential indirect effects, this study used data from a large, multistate health care system to examine changes in hospital volume and its relationship to in-hospital mortality for patients without COVID-19 during the first 10 months of the pandemic.
METHODS
Setting and Participants
We examined unplanned hospitalizations from January 2019 to December 2020 at 51 community hospitals across 6 states (Alaska, Washington, Montana, Oregon, California, and Texas) in the Providence St. Joseph Health system. Hospitals within the Providence system share a common standard dataset for each encounter with a centralized cloud data warehouse from which we extracted clinical and demographic data. No hospitals entered or left the system during the study period. Hospitalizations were considered unplanned if they had an “urgent” or “emergency” service type in the record; most originated in the ED. Hospitalizations for children younger than 18 years and those with evidence of COVID-19 (International Classification of Disease, Tenth Revision, Clinical Modification U07.1, a positive COVID-19 polymerase chain reaction test during the encounter, or an infection control-assigned label of COVID-19) were excluded. The Providence St. Joseph Health Institutional Review Board approved this study.
Measures
Trends in daily hospitalizations and their relationship to adjusted in-hospital mortality (percentage of patients who died during their hospital admission) were examined over time. In preliminary models using segmented regression, we identified three distinct pandemic periods with different trends in daily hospitalizations: (1) a 10-week period corresponding to the spring COVID-19 surge (March 4 to May 13, 2020; Period 1), (2) an intervening period extending over the summer and early fall (May 14 to October 19, 2020; Period 2), and (3) a second 10-week period corresponding to the fall COVID-19 surge (October 20 to December 31, 2020; Period 3). In-hospital mortality for these periods was compared with a baseline period (pre-COVID-19) from January 1, 2019 to March 3, 2020. To further assess differences in mortality by clinical condition, hospitalizations were first grouped by primary diagnosis using Clinical Classifications Software Refined (CCSR) categories from the Agency for Healthcare Research and Quality14 and ranked by the number of observed deaths and the percentage of patients who died while hospitalized in 2020. We selected common conditions that had >35 total deaths and an in-hospital mortality rate ≥1% for condition-specific analyses, of which 30 met these criteria.
Analysis
Multivariate logistic regression was used to evaluate changes in mortality for each of the pandemic periods compared with baseline for the overall cohort and selected diagnosis groups. Our main model adjusted for age, sex, race/ethnicity (White, Black, Latinx, Asian or Pacific Islander, and other), primary payor (commercial, Medicaid, Medicare, other, and self-pay), the presence or absence of 31 chronic comorbidities in the medical record, primary admitting diagnosis grouped by CCSR category (456 total diagnostic groups), and hospital fixed-effects to account for clustering. Results are expressed as the average marginal effects of each pandemic period on in-hospital mortality (eg, adjusted percentage point change in mortality over baseline). The number of excess deaths in each period was calculated by multiplying the estimated percentage point change in mortality for each period by the total number of hospitalizations. These excess deaths were subtracted from the number of observed deaths to derive the number of deaths that would be expected if pre-pandemic mortality rates persisted.
To further assess whether changes in adjusted mortality could be attributed to a smaller, sicker population of patients presenting to the hospital during the pandemic (meaning that less acutely ill patients stayed home), we conducted two sensitivity analyses. First, we tested whether substituting indicators for Medicare Severity Diagnosis Groups (MS-DRG) in lieu of CCSR categories had any impact on our results. MS-DRGs are designed to account for a patient’s illness severity and expected costs, whereas CCSR categories do not.15 MS-DRGs also better distinguish between surgical versus medical conditions. We re-ran our main model using indicators for CCSR to control for diagnostic mix, but further adjusted for severity using the DRG weight for the primary diagnosis and Modified Early Warning Score (MEWS) as continuous covariates. MEWS is a physiologic scoring system that incorporates abnormal vital signs and data related to mental status during the first 24 hours of a patient’s hospitalization into a risk-based score that has been shown to predict hospital mortality and need for intensive care.16,17 These sensitivity analyses were performed on a subset of inpatient admissions because DRG data are not available for hospitalizations billed as an observation stay, and only approximately 70% of hospitals in the sample contributed vital sign data to the Providence data warehouse. All statistical analyses were conducted with R, version 3.6.3 (R Foundation for Statistical Computing) and SAS Enterprise Guide 7.1 (SAS Institute Inc).
RESULTS
The characteristics of our sample are described in Table 1. A total of 61,300, 159,430, and 65,923 hospitalizations occurred in each of the three pandemic periods, respectively, compared with 503,190 hospitalizations in the pre-pandemic period. The mean (SD) age of patients in the study was 63.2 (19.4) years; most were women (52.4%), White (70.6%), and had Medicare as their primary payor (53.7%). Less than half (42.7%) of hospitalizations occurred in California, and just under one-quarter were observation stays (23.2%). Patient characteristics were similar in the pre-COVID-19 and COVID-19 pandemic periods.
Figure 1 shows trends in hospital volume and mortality. Overall daily hospitalizations declined abruptly from a mean of 1176 per day in the pre-pandemic period to 617 per day (47.5% relative decrease) during the first 3 weeks of Period 1. Mean daily hospitalizations began to rise over the next 2 months (Period 1), reaching steady state at <1000 hospitalizations per day (15% relative decrease from baseline) during Period 2. During Period 3, we observed a decline in mean daily hospitalizations, with a low point of 882 per day on December 31, 2020 (25% relative decrease from baseline), corresponding to the end of our study period. Although hospital volumes declined during both COVID-19 surge periods, the percentage of patients who died during their hospitalization increased. There was an initial spike in in-hospital mortality that peaked approximately 1 month into the pandemic (middle of Period 1), a return to levels at or slightly below that before the pandemic by the beginning of Period 2, and then a rise throughout the autumn COVID-19 surge in Period 3, not yet peaking by the end of the study.
Adjusted in-hospital mortality for the three COVID-19 periods compared with the pre-pandemic period is presented in Table 2. The percentage of patients who died during their hospitalization rose from 2.9% in the pre-pandemic period to 3.4% during Period 1 (absolute difference, 0.6 percentage points; 95% CI, 0.5-0.7), corresponding to a 19.3% relative increase during the spring COVID-19 surge. Among the subset of patients hospitalized with 1 of the 30 conditions selected for individual analysis, mortality increased from 5.0% to 5.9% during the same time period (absolute difference, 0.9 percentage points; 95% CI, 0.8-1.1), corresponding to an 18.9% relative increase. In Period 2, in-hospital mortality was similar to that noted pre-pandemic for the overall cohort and the 30 selected conditions. During Period 3, in-hospital mortality increased by a magnitude similar to that observed in Period 1 for all hospitalizations combined (absolute difference, 0.5 percentage points; 95% CI, 0.0-0.6; corresponding to a 16.5% relative increase) as well as the subgroup with 1 of the 30 selected conditions (0.9 percentage points; 95% CI, 0.8-1.0; corresponding to an 18% relative increase). Further adjustment for severity by swapping CCSR categories with MS-DRG indicators or inclusion of DRG weight and MEWS score as covariates in our sensitivity analyses did not change our results.
Table 3 and the Appendix Figure describe changes in volume and adjusted in-hospital mortality for the 30 conditions selected for analysis. There was a decrease in the mean daily admissions for all conditions studied. Among the 30 conditions, 26 showed increased mortality during Period 1, although the increase was only statistically significant for 16 of these conditions. Among the 10 most commonly admitted conditions (by number of daily hospital admissions during the baseline period), there was a statistically significant relative increase in mortality for patients with sepsis (20.1%), heart failure (17.6%), ischemic stroke (12.5%), device/graft/surgical complications (14.0%), cardiac dysrhythmias (14.4%), pneumonia (24.5%), respiratory failure (16.1%), and gastrointestinal hemorrhage (23.3%). In general, mortality returned to baseline or improved during Period 2. Thereafter, all 30 conditions showed increased mortality in Period 3. This increase was significant for only 16 conditions, which were not the same ones noted during Period 1. Of note, although there was higher mortality for some cardiovascular conditions (heart failure cardiac dysrhythmias), mortality for myocardial infarction remained unchanged from baseline across all 3 periods. In contrast, several solid cancer–related conditions showed progressively worsening mortality throughout the study, with 7.7% higher mortality in Period 1, 10.3% higher mortality in Period 2, and 16.5% higher mortality in Period 3, respectively, compared with baseline. Although a similar pattern was observed for acute renal failure and some neurologic conditions (traumatic brain injury, seizure, other nervous system disorders), mortality for drug poisonings and gastrointestinal bleeds improved over time.
DISCUSSION
In this study of unplanned hospitalizations from 51 community hospitals across 6 states in the US West, we found a significant increase in mortality—at a rate of approximately 5 to 6 excess deaths per 1000 hospitalizations—among patients admitted during the pandemic with a variety of non-COVID-19 illnesses and injuries. Higher in-hospital mortality was observed in the spring (March to May) and fall (October to December) of 2020 when COVID-19 case counts surged and shelter-in-place mandates were implemented. With the initial surge, higher mortality rates were largely transient, and, for most conditions evaluated, returned to baseline approximately 3 months after the pandemic onset. For the fall surge, mortality rates had not peaked by the end of the study period. Changes in mortality were closely and inversely correlated with hospital volume for non-COVID-19 illnesses during both surge periods.
Higher morbidity and mortality for patients without COVID-19 appears to be an unfortunate spillover effect that has been reported in several studies. Recent work examining national surveillance data suggest that up to one-third of excess deaths (deaths higher than those expected for season) early in the pandemic have occurred among patients without known COVID-19.13,18-20 Specifically, these studies estimate that mortality rates in the United States increased by 15% to 19% in the spring of 2020; of the identified excess deaths, only 38% to 77% could be attributed to COVID-19, with the remainder attributed to cardiovascular disease, diabetes, and Alzheimer’s disease, among others. In addition, reports from several European countries and China examining population death data have found similar trends,21-25 as well as a recent study examining excess deaths in nursing homes.26 Our results are largely consistent with these earlier studies in that we describe higher mortality in a sample of patients hospitalized with a variety of common conditions that otherwise are routinely treated in US hospitals. Reporting these indirect casualties of COVID-19 is important to fully understand the pandemic’s toll on patients and healthcare systems.
Our work builds on the current body of literature, highlighting the consistent relationship between rising COVID-19 case counts, hospital volume, and excess mortality over more than one surge period. Although several studies have looked at trends in hospital admissions or population mortality rates, few have examined the two outcomes together. The close correlation between daily hospital admissions and in-hospital mortality in this study suggests that the pandemic changed how patients use healthcare resources in ways that were important for their health and outcomes. The higher mortality rate that we and others have observed likely is related to patients’ delaying care because of fear of contracting COVID-19. In one survey, more than 4 in 10 adults in the United States reported that they avoided medical care during the early pandemic.10 Importantly, even a few days delay for many conditions, such as heart failure or sepsis, can result in precipitous declines in clinical status and outcomes.
It also is possible that we found increased rates of in-hospital mortality simply because patients with more moderate illness chose to stay home, resulting in a patient population enriched with those more likely to die. We found mixed evidence in our data that the observed increases in mortality could be attributable to a smaller, sicker population. Some characteristics that might be protective, such as a slightly younger mean age and lower mean DRG weight, were more common among those hospitalized during the pandemic. However, other characteristics, such as a slightly higher MEWS score and a greater percentage of total hospitalizations in the higher mortality subgroup, also were noted during the pandemic (Table 1). We do note, however, that the differences in these severity-related characteristics were small across the study periods. Further adjusting for these characteristics in our sensitivity analyses did not appreciably change our main findings, suggesting that the mortality increase could not be explained by changes in case-mix alone.
Other factors not dependent on patient behavior, such as barriers to accessing timely ambulatory care and impacts in the quality of care delivered, might have contributed. Shelter-in-place orders, reduced in-person access to clinicians in the ambulatory setting, slow implementation of telehealth services (with uncertainty about their equivalence to in-person exams), as well as delays in diagnostic tests and outpatient procedures could have played a role, especially during early months of the pandemic.27 Significant changes to ambulatory health care delivery might have left many patients with chronic illnesses or complex medical needs with limited care options. Importantly, these care interruptions might have had greater implications for some patients, such as those with cancer who rely on intensive, largely outpatient-based treatment.28,29 This, in part, could explain why we found persistently increased mortality among patients hospitalized with cancer after the spring surge. Later into the pandemic, however, most health systems had developed processes that allowed clinicians to resume timely care of ambulatory patients. Because of this, increases in mortality observed during the fall surge likely stem from other factors, such as patient behavior.
It is possible that care delays or changes in the quality of care delivered during the index hospitalization or pre-hospital setting might have contributed to the observed increase in mortality. This is particularly true for acute, time-sensitive conditions such as sepsis and stroke. Extra time spent donning personal protective equipment and/or new protocols instituted during the pandemic likely impacted the speed of emergency medical services transport, timeliness of ED evaluation, and delivery of definitive therapy. Although most hospitals in this study were not overwhelmed by the pandemic, the complexities associated with caring for known and suspected COVID-19 patients alongside those without the disease might have altered ideal care practices and strained healthcare teams.30 In addition, nearly all hospitalized patients during this period were deprived of in-person advocacy by family members, who were not permitted to visit.
Important limitations with this study exist. First, the data come only from hospitals in the western United States. Second, some data elements such as triage scores or vital signs were not available for the entire population, potentially limiting some risk-adjustment. Third, we were unable to determine the root cause of excess mortality based on our study design and the coded variables available. It is unknown to what extent undiagnosed COVID-19 played a role. Early in the pandemic, many community hospitals did not have access to timely COVID-19 testing, and some cases might have not been diagnosed.31 However, we do not expect this to be a significant concern in the later months of the pandemic, as testing became more widespread and hospitals implemented surveillance screening for COVID-19 for inpatients.
CONCLUSIONS
Our study indicates that the COVID-19 pandemic was associated with increased mortality among patients hospitalized for a range of clinical conditions. Although higher observed mortality rates were limited to periods of high COVID-19 activity, future studies will need to tease out the extent to which these findings relate to patient factors (ie, delayed presentation and more severe disease) or systemic factors (reduction in access or changes in quality of care). This is of key importance, and appropriate solutions will need to be developed to mitigate adverse impacts with this and future pandemics.
One of the most striking features of the early COVID-19 pandemic was the sudden and sharp reductions in emergency department (ED) visits and hospitalizations throughout the United States.1-4 Several studies have documented lower rates of hospitalization for many emergent, time-sensitive conditions, such as acute myocardial infarction, stroke, and hyperglycemic crises, starting shortly after community transmission of COVID-19 was recognized and social distancing guidelines were implemented.5-8 In most cases, hospital volumes rebounded after an initial drop, stabilizing at somewhat lower levels than those expected from historic trends.9
The observed shifts in hospital use largely have been attributed to patients’ forgoing or delaying necessary care,10 which underscores the indirect effects of the pandemic on patients without COVID-19.11 To date, the extent to which outcomes for patients without COVID-19 have been adversely affected is less well understood. Evidence suggests patients with acute and chronic illnesses have experienced increased morbidity and mortality since the onset of the pandemic. For example, in northern California, abrupt declines in ED visits for cardiac symptoms were coupled with higher rates of out-of-hospital cardiac arrest.12 Moreover, states with higher rates of COVID-19 also reported increased deaths attributed to heart disease, diabetes, and other conditions.13
To better understand these potential indirect effects, this study used data from a large, multistate health care system to examine changes in hospital volume and its relationship to in-hospital mortality for patients without COVID-19 during the first 10 months of the pandemic.
METHODS
Setting and Participants
We examined unplanned hospitalizations from January 2019 to December 2020 at 51 community hospitals across 6 states (Alaska, Washington, Montana, Oregon, California, and Texas) in the Providence St. Joseph Health system. Hospitals within the Providence system share a common standard dataset for each encounter with a centralized cloud data warehouse from which we extracted clinical and demographic data. No hospitals entered or left the system during the study period. Hospitalizations were considered unplanned if they had an “urgent” or “emergency” service type in the record; most originated in the ED. Hospitalizations for children younger than 18 years and those with evidence of COVID-19 (International Classification of Disease, Tenth Revision, Clinical Modification U07.1, a positive COVID-19 polymerase chain reaction test during the encounter, or an infection control-assigned label of COVID-19) were excluded. The Providence St. Joseph Health Institutional Review Board approved this study.
Measures
Trends in daily hospitalizations and their relationship to adjusted in-hospital mortality (percentage of patients who died during their hospital admission) were examined over time. In preliminary models using segmented regression, we identified three distinct pandemic periods with different trends in daily hospitalizations: (1) a 10-week period corresponding to the spring COVID-19 surge (March 4 to May 13, 2020; Period 1), (2) an intervening period extending over the summer and early fall (May 14 to October 19, 2020; Period 2), and (3) a second 10-week period corresponding to the fall COVID-19 surge (October 20 to December 31, 2020; Period 3). In-hospital mortality for these periods was compared with a baseline period (pre-COVID-19) from January 1, 2019 to March 3, 2020. To further assess differences in mortality by clinical condition, hospitalizations were first grouped by primary diagnosis using Clinical Classifications Software Refined (CCSR) categories from the Agency for Healthcare Research and Quality14 and ranked by the number of observed deaths and the percentage of patients who died while hospitalized in 2020. We selected common conditions that had >35 total deaths and an in-hospital mortality rate ≥1% for condition-specific analyses, of which 30 met these criteria.
Analysis
Multivariate logistic regression was used to evaluate changes in mortality for each of the pandemic periods compared with baseline for the overall cohort and selected diagnosis groups. Our main model adjusted for age, sex, race/ethnicity (White, Black, Latinx, Asian or Pacific Islander, and other), primary payor (commercial, Medicaid, Medicare, other, and self-pay), the presence or absence of 31 chronic comorbidities in the medical record, primary admitting diagnosis grouped by CCSR category (456 total diagnostic groups), and hospital fixed-effects to account for clustering. Results are expressed as the average marginal effects of each pandemic period on in-hospital mortality (eg, adjusted percentage point change in mortality over baseline). The number of excess deaths in each period was calculated by multiplying the estimated percentage point change in mortality for each period by the total number of hospitalizations. These excess deaths were subtracted from the number of observed deaths to derive the number of deaths that would be expected if pre-pandemic mortality rates persisted.
To further assess whether changes in adjusted mortality could be attributed to a smaller, sicker population of patients presenting to the hospital during the pandemic (meaning that less acutely ill patients stayed home), we conducted two sensitivity analyses. First, we tested whether substituting indicators for Medicare Severity Diagnosis Groups (MS-DRG) in lieu of CCSR categories had any impact on our results. MS-DRGs are designed to account for a patient’s illness severity and expected costs, whereas CCSR categories do not.15 MS-DRGs also better distinguish between surgical versus medical conditions. We re-ran our main model using indicators for CCSR to control for diagnostic mix, but further adjusted for severity using the DRG weight for the primary diagnosis and Modified Early Warning Score (MEWS) as continuous covariates. MEWS is a physiologic scoring system that incorporates abnormal vital signs and data related to mental status during the first 24 hours of a patient’s hospitalization into a risk-based score that has been shown to predict hospital mortality and need for intensive care.16,17 These sensitivity analyses were performed on a subset of inpatient admissions because DRG data are not available for hospitalizations billed as an observation stay, and only approximately 70% of hospitals in the sample contributed vital sign data to the Providence data warehouse. All statistical analyses were conducted with R, version 3.6.3 (R Foundation for Statistical Computing) and SAS Enterprise Guide 7.1 (SAS Institute Inc).
RESULTS
The characteristics of our sample are described in Table 1. A total of 61,300, 159,430, and 65,923 hospitalizations occurred in each of the three pandemic periods, respectively, compared with 503,190 hospitalizations in the pre-pandemic period. The mean (SD) age of patients in the study was 63.2 (19.4) years; most were women (52.4%), White (70.6%), and had Medicare as their primary payor (53.7%). Less than half (42.7%) of hospitalizations occurred in California, and just under one-quarter were observation stays (23.2%). Patient characteristics were similar in the pre-COVID-19 and COVID-19 pandemic periods.
Figure 1 shows trends in hospital volume and mortality. Overall daily hospitalizations declined abruptly from a mean of 1176 per day in the pre-pandemic period to 617 per day (47.5% relative decrease) during the first 3 weeks of Period 1. Mean daily hospitalizations began to rise over the next 2 months (Period 1), reaching steady state at <1000 hospitalizations per day (15% relative decrease from baseline) during Period 2. During Period 3, we observed a decline in mean daily hospitalizations, with a low point of 882 per day on December 31, 2020 (25% relative decrease from baseline), corresponding to the end of our study period. Although hospital volumes declined during both COVID-19 surge periods, the percentage of patients who died during their hospitalization increased. There was an initial spike in in-hospital mortality that peaked approximately 1 month into the pandemic (middle of Period 1), a return to levels at or slightly below that before the pandemic by the beginning of Period 2, and then a rise throughout the autumn COVID-19 surge in Period 3, not yet peaking by the end of the study.
Adjusted in-hospital mortality for the three COVID-19 periods compared with the pre-pandemic period is presented in Table 2. The percentage of patients who died during their hospitalization rose from 2.9% in the pre-pandemic period to 3.4% during Period 1 (absolute difference, 0.6 percentage points; 95% CI, 0.5-0.7), corresponding to a 19.3% relative increase during the spring COVID-19 surge. Among the subset of patients hospitalized with 1 of the 30 conditions selected for individual analysis, mortality increased from 5.0% to 5.9% during the same time period (absolute difference, 0.9 percentage points; 95% CI, 0.8-1.1), corresponding to an 18.9% relative increase. In Period 2, in-hospital mortality was similar to that noted pre-pandemic for the overall cohort and the 30 selected conditions. During Period 3, in-hospital mortality increased by a magnitude similar to that observed in Period 1 for all hospitalizations combined (absolute difference, 0.5 percentage points; 95% CI, 0.0-0.6; corresponding to a 16.5% relative increase) as well as the subgroup with 1 of the 30 selected conditions (0.9 percentage points; 95% CI, 0.8-1.0; corresponding to an 18% relative increase). Further adjustment for severity by swapping CCSR categories with MS-DRG indicators or inclusion of DRG weight and MEWS score as covariates in our sensitivity analyses did not change our results.
Table 3 and the Appendix Figure describe changes in volume and adjusted in-hospital mortality for the 30 conditions selected for analysis. There was a decrease in the mean daily admissions for all conditions studied. Among the 30 conditions, 26 showed increased mortality during Period 1, although the increase was only statistically significant for 16 of these conditions. Among the 10 most commonly admitted conditions (by number of daily hospital admissions during the baseline period), there was a statistically significant relative increase in mortality for patients with sepsis (20.1%), heart failure (17.6%), ischemic stroke (12.5%), device/graft/surgical complications (14.0%), cardiac dysrhythmias (14.4%), pneumonia (24.5%), respiratory failure (16.1%), and gastrointestinal hemorrhage (23.3%). In general, mortality returned to baseline or improved during Period 2. Thereafter, all 30 conditions showed increased mortality in Period 3. This increase was significant for only 16 conditions, which were not the same ones noted during Period 1. Of note, although there was higher mortality for some cardiovascular conditions (heart failure cardiac dysrhythmias), mortality for myocardial infarction remained unchanged from baseline across all 3 periods. In contrast, several solid cancer–related conditions showed progressively worsening mortality throughout the study, with 7.7% higher mortality in Period 1, 10.3% higher mortality in Period 2, and 16.5% higher mortality in Period 3, respectively, compared with baseline. Although a similar pattern was observed for acute renal failure and some neurologic conditions (traumatic brain injury, seizure, other nervous system disorders), mortality for drug poisonings and gastrointestinal bleeds improved over time.
DISCUSSION
In this study of unplanned hospitalizations from 51 community hospitals across 6 states in the US West, we found a significant increase in mortality—at a rate of approximately 5 to 6 excess deaths per 1000 hospitalizations—among patients admitted during the pandemic with a variety of non-COVID-19 illnesses and injuries. Higher in-hospital mortality was observed in the spring (March to May) and fall (October to December) of 2020 when COVID-19 case counts surged and shelter-in-place mandates were implemented. With the initial surge, higher mortality rates were largely transient, and, for most conditions evaluated, returned to baseline approximately 3 months after the pandemic onset. For the fall surge, mortality rates had not peaked by the end of the study period. Changes in mortality were closely and inversely correlated with hospital volume for non-COVID-19 illnesses during both surge periods.
Higher morbidity and mortality for patients without COVID-19 appears to be an unfortunate spillover effect that has been reported in several studies. Recent work examining national surveillance data suggest that up to one-third of excess deaths (deaths higher than those expected for season) early in the pandemic have occurred among patients without known COVID-19.13,18-20 Specifically, these studies estimate that mortality rates in the United States increased by 15% to 19% in the spring of 2020; of the identified excess deaths, only 38% to 77% could be attributed to COVID-19, with the remainder attributed to cardiovascular disease, diabetes, and Alzheimer’s disease, among others. In addition, reports from several European countries and China examining population death data have found similar trends,21-25 as well as a recent study examining excess deaths in nursing homes.26 Our results are largely consistent with these earlier studies in that we describe higher mortality in a sample of patients hospitalized with a variety of common conditions that otherwise are routinely treated in US hospitals. Reporting these indirect casualties of COVID-19 is important to fully understand the pandemic’s toll on patients and healthcare systems.
Our work builds on the current body of literature, highlighting the consistent relationship between rising COVID-19 case counts, hospital volume, and excess mortality over more than one surge period. Although several studies have looked at trends in hospital admissions or population mortality rates, few have examined the two outcomes together. The close correlation between daily hospital admissions and in-hospital mortality in this study suggests that the pandemic changed how patients use healthcare resources in ways that were important for their health and outcomes. The higher mortality rate that we and others have observed likely is related to patients’ delaying care because of fear of contracting COVID-19. In one survey, more than 4 in 10 adults in the United States reported that they avoided medical care during the early pandemic.10 Importantly, even a few days delay for many conditions, such as heart failure or sepsis, can result in precipitous declines in clinical status and outcomes.
It also is possible that we found increased rates of in-hospital mortality simply because patients with more moderate illness chose to stay home, resulting in a patient population enriched with those more likely to die. We found mixed evidence in our data that the observed increases in mortality could be attributable to a smaller, sicker population. Some characteristics that might be protective, such as a slightly younger mean age and lower mean DRG weight, were more common among those hospitalized during the pandemic. However, other characteristics, such as a slightly higher MEWS score and a greater percentage of total hospitalizations in the higher mortality subgroup, also were noted during the pandemic (Table 1). We do note, however, that the differences in these severity-related characteristics were small across the study periods. Further adjusting for these characteristics in our sensitivity analyses did not appreciably change our main findings, suggesting that the mortality increase could not be explained by changes in case-mix alone.
Other factors not dependent on patient behavior, such as barriers to accessing timely ambulatory care and impacts in the quality of care delivered, might have contributed. Shelter-in-place orders, reduced in-person access to clinicians in the ambulatory setting, slow implementation of telehealth services (with uncertainty about their equivalence to in-person exams), as well as delays in diagnostic tests and outpatient procedures could have played a role, especially during early months of the pandemic.27 Significant changes to ambulatory health care delivery might have left many patients with chronic illnesses or complex medical needs with limited care options. Importantly, these care interruptions might have had greater implications for some patients, such as those with cancer who rely on intensive, largely outpatient-based treatment.28,29 This, in part, could explain why we found persistently increased mortality among patients hospitalized with cancer after the spring surge. Later into the pandemic, however, most health systems had developed processes that allowed clinicians to resume timely care of ambulatory patients. Because of this, increases in mortality observed during the fall surge likely stem from other factors, such as patient behavior.
It is possible that care delays or changes in the quality of care delivered during the index hospitalization or pre-hospital setting might have contributed to the observed increase in mortality. This is particularly true for acute, time-sensitive conditions such as sepsis and stroke. Extra time spent donning personal protective equipment and/or new protocols instituted during the pandemic likely impacted the speed of emergency medical services transport, timeliness of ED evaluation, and delivery of definitive therapy. Although most hospitals in this study were not overwhelmed by the pandemic, the complexities associated with caring for known and suspected COVID-19 patients alongside those without the disease might have altered ideal care practices and strained healthcare teams.30 In addition, nearly all hospitalized patients during this period were deprived of in-person advocacy by family members, who were not permitted to visit.
Important limitations with this study exist. First, the data come only from hospitals in the western United States. Second, some data elements such as triage scores or vital signs were not available for the entire population, potentially limiting some risk-adjustment. Third, we were unable to determine the root cause of excess mortality based on our study design and the coded variables available. It is unknown to what extent undiagnosed COVID-19 played a role. Early in the pandemic, many community hospitals did not have access to timely COVID-19 testing, and some cases might have not been diagnosed.31 However, we do not expect this to be a significant concern in the later months of the pandemic, as testing became more widespread and hospitals implemented surveillance screening for COVID-19 for inpatients.
CONCLUSIONS
Our study indicates that the COVID-19 pandemic was associated with increased mortality among patients hospitalized for a range of clinical conditions. Although higher observed mortality rates were limited to periods of high COVID-19 activity, future studies will need to tease out the extent to which these findings relate to patient factors (ie, delayed presentation and more severe disease) or systemic factors (reduction in access or changes in quality of care). This is of key importance, and appropriate solutions will need to be developed to mitigate adverse impacts with this and future pandemics.
1. Baum A, Schwartz MD. Admissions to Veterans Affairs hospitals for emergency conditions during the COVID-19 pandemic. JAMA. 2020;324(1):96-99. https://doi.org/10.1001/jama.2020.9972
2. Hartnett KP, Kite-Powell A, DeVies J, et al; National Syndromic Surveillance Program Community of Practice. Impact of the COVID-19 pandemic on emergency department visits — United States, January 1, 2019–May 30, 2020. MMWR Morb Mortal Wkly Rep. 2020;69(23):699-704. https://doi.org/10.15585/mmwr.mm6923e1
3. Birkmeyer JD, Barnato A, Birkmeyer N, Bessler R, Skinner J. The impact of the COVID-19 pandemic on hospital admissions in the United States. Health Aff. 2020;39(11):2010-2017. https://doi.org/10.1377/hlthaff.2020.00980
4. Blecker S, Jones SA, Petrilli CM, et al. Hospitalizations for chronic disease and acute conditions in the time of COVID-19. JAMA Intern Med. 2021;181(2):269-271. https://doi.org/10.1001/jamainternmed.2020.3978
5. Bhambhvani HP, Rodrigues AJ, Yu JS, Carr JB 2nd, Hayden Gephart M. Hospital volumes of 5 medical emergencies in the COVID-19 pandemic in 2 US medical centers. JAMA Intern Med. 2021;181(2):272-274. https://doi.org/10.1001/jamainternmed.2020.3982
6. Lange SJ, Ritchey MD, Goodman AB, et al. Potential indirect effects of the COVID-19 pandemic on use of emergency departments for acute life-threatening conditions — United States, January–May 2020. MMWR Morb Mortal Wkly Rep. 2020;69(25);795-800. https://doi.org/10.15585/mmwr.mm6925e2
7. Solomon MD, McNulty EJ, Rana JS, et al. The Covid-19 pandemic and the incidence of acute myocardial infarction. N Engl J Med. 2020;383(7):691-693. https://doi.org/10.1056/NEJMc2015630
8. Kansagra AP, Goyal MS, Hamilton S, Albers GW. Collateral effect of Covid-19 on stroke evaluation in the United States. N Engl J Med. 2020;383(4):400-401. https://doi.org/10.1056/NEJMc2014816
9. Heist T, Schwartz K, Butler S. Trends in overall and non-COVID-19 hospital admissions. Kaiser Family Foundation. Accessed March 18, 2021. https://www.kff.org/health-costs/issue-brief/trends-in-overall-and-non-covid-19-hospital-admissions
10. Czeisler MÉ, Marynak K, Clarke KEN, et al. Delay or avoidance of medical care because of COVID-19–related concerns — United States, June 2020. MMWR Morb Mortal Wkly Rep. 2020;69(36);1250-1257. https://doi.org/10.15585/mmwr.mm6936a4
11. Chen J, McGeorge R. Spillover effects of the COVID-19 pandemic could drive long-term health consequences for non-COVID-19 patients. Health Affairs Blog. Accessed March 18, 2021. https://www.healthaffairs.org/do/10.1377/hblog20201020.566558/full/
12. Wong LE, Hawkins JE, Langness S, Murrell KL, Iris P, Sammann A. Where are all the patients? Addressing Covid-19 fear to encourage sick patients to seek emergency care. NEJM Catalyst. Accessed March 18, 2021. https://catalyst.nejm.org/doi/abs/10.1056/CAT.20.0193
13. Woolf SH, Chapman DA, Sabo RT, Weinberger DM, Hill L. Excess deaths from COVID-19 and other causes, March-April 2020. JAMA. 2020;324(5):510-513. https://doi.org/10.1001/jama.2020.11787
14. Clinical Classifications Software Refined (CCSR) for ICD-10-CM Diagnoses. Agency for Healthcare Research and Quality, Rockville, MD. Accessed April 22, 2021. https://www.hcup-us.ahrq.gov/toolssoftware/ccsr/dxccsr.jsp
15. MS-DRG Classifications and Software. Centers for Medicare & Medicaid Services. Accessed March 18, 2021. https://www.cms.gov/Medicare/Medicare-Fee-for-Service-Payment/AcuteInpatientPPS/MS-DRG-Classifications-and-Software
16. Jayasundera R, Neilly M, Smith TO, Myint PK. Are early warning scores useful predictors for mortality and morbidity in hospitalised acutely unwell older patients? A systematic review. J Clin Med. 2018;7(10):309. https://doi.org/10.3390/jcm7100309
17. Delgado-Hurtado JJ, Berger A, Bansal AB. Emergency department Modified Early Warning Score association with admission, admission disposition, mortality, and length of stay. J Community Hosp Intern Med Perspect. 2016;6(2):31456. https://doi.org/10.3402/jchimp.v6.31456
18. Woolf SH, Chapman DA, Sabo RT, Weinberger DM, Hill L, Taylor DDH. Excess deaths from COVID-19 and other causes, March-July 2020. JAMA. 2020;324(15):1562-1564. https://doi.org/10.1001/jama.2020.19545
19. Faust JS, Krumholz HM, Du C, et al. All-cause excess mortality and COVID-19–related mortality among US adults aged 25-44 years, March-July 2020. JAMA. 2021;325(8):785-787. https://doi.org/10.1001/jama.2020.24243
20. Weinberger DM, Chen J, Cohen T, et al. Estimation of excess deaths associated with the COVID-19 pandemic in the United States, March to May 2020. JAMA Intern Med. 2020;180(10):1336-1344. https://doi.org/10.1001/jamainternmed.2020.3391
21. Vandoros S. Excess mortality during the Covid-19 pandemic: Early evidence from England and Wales. Soc Sci Med. 2020; 258:113101. https://doi.org/10.1016/j.socscimed.2020.113101
22. Vestergaard LS, Nielsen J, Richter L, et al; ECDC Public Health Emergency Team for COVID-19. Excess all-cause mortality during the COVID-19 pandemic in Europe – preliminary pooled estimates from the EuroMOMO network, March to April 2020. Euro Surveill. 2020;25(26):2001214. https://doi.org/10.2807/1560-7917.ES.2020.25.26.2001214
23. Kontopantelis E, Mamas MA, Deanfield J, Asaria M, Doran T. Excess mortality in England and Wales during the first wave of the COVID-19 pandemic. J Epidemiol Community Health. 2021;75(3):213-223. https://doi.org/10.1136/jech-2020-214764
24. Liu J, Zhang L, Yan Y, et al. Excess mortality in Wuhan city and other parts of China during the three months of the covid-19 outbreak: findings from nationwide mortality registries. BMJ. 2021;372:n415. https://doi.org/10.1136/bmj.n415
25. Docherty KF, Butt JH, de Boer RA, et al. Excess deaths during the Covid-19 pandemic: An international comparison. Preprint. Posted online May 13, 2020. medRxiv. doi:https://doi.org/10.1101/2020.04.21.20073114
26. Barnett ML, Hu L, Martin T, Grabowski DC. Mortality, admissions, and patient census at SNFs in 3 US cities during the COVID-19 pandemic. JAMA. 2020;324(5):507-509. https://doi.org/10.1001/jama.2020.11642
27. Rosenbaum L. The untold toll — The pandemic’s effects on patients without Covid-19. N Engl J Med. 2020; 382:2368-2371 https://doi.org/10.1056/NEJMms2009984
28. Lai AG, Pasea L, Banerjee A, et al. Estimated impact of the COVID-19 pandemic on cancer services and excess 1-year mortality in people with cancer and multimorbidity: near real-time data on cancer care, cancer deaths and a population-based cohort study. BMJ Open. 2020;10(11):e043828. https://doi.org/10.1136/bmjopen-2020-043828
29. Van de Haar J, Hoes LR, Coles CE, et al. Caring for patients with cancer in the COVID-19 era. Nat Med. 2020;26(5):665-671. https://doi.org/10.1038/s41591-020-0874-8
30. Traylor AM, Tannenbaum SI, Thomas EJ, Salas E. Helping healthcare teams save lives during COVID-19: insights and countermeasures from team science. Am Psychol. 2020;76(1):1-13. https://doi.org/10.1037/amp0000750
31. Grimm CA. Hospital experiences responding to the COVID-19 pandemic: results of a National Pulse Survey March 23–27. U.S. Department of Health and Human Services Office of Inspector General; 2020. https://oig.hhs.gov/oei/reports/oei-06-20-00300.pdf
1. Baum A, Schwartz MD. Admissions to Veterans Affairs hospitals for emergency conditions during the COVID-19 pandemic. JAMA. 2020;324(1):96-99. https://doi.org/10.1001/jama.2020.9972
2. Hartnett KP, Kite-Powell A, DeVies J, et al; National Syndromic Surveillance Program Community of Practice. Impact of the COVID-19 pandemic on emergency department visits — United States, January 1, 2019–May 30, 2020. MMWR Morb Mortal Wkly Rep. 2020;69(23):699-704. https://doi.org/10.15585/mmwr.mm6923e1
3. Birkmeyer JD, Barnato A, Birkmeyer N, Bessler R, Skinner J. The impact of the COVID-19 pandemic on hospital admissions in the United States. Health Aff. 2020;39(11):2010-2017. https://doi.org/10.1377/hlthaff.2020.00980
4. Blecker S, Jones SA, Petrilli CM, et al. Hospitalizations for chronic disease and acute conditions in the time of COVID-19. JAMA Intern Med. 2021;181(2):269-271. https://doi.org/10.1001/jamainternmed.2020.3978
5. Bhambhvani HP, Rodrigues AJ, Yu JS, Carr JB 2nd, Hayden Gephart M. Hospital volumes of 5 medical emergencies in the COVID-19 pandemic in 2 US medical centers. JAMA Intern Med. 2021;181(2):272-274. https://doi.org/10.1001/jamainternmed.2020.3982
6. Lange SJ, Ritchey MD, Goodman AB, et al. Potential indirect effects of the COVID-19 pandemic on use of emergency departments for acute life-threatening conditions — United States, January–May 2020. MMWR Morb Mortal Wkly Rep. 2020;69(25);795-800. https://doi.org/10.15585/mmwr.mm6925e2
7. Solomon MD, McNulty EJ, Rana JS, et al. The Covid-19 pandemic and the incidence of acute myocardial infarction. N Engl J Med. 2020;383(7):691-693. https://doi.org/10.1056/NEJMc2015630
8. Kansagra AP, Goyal MS, Hamilton S, Albers GW. Collateral effect of Covid-19 on stroke evaluation in the United States. N Engl J Med. 2020;383(4):400-401. https://doi.org/10.1056/NEJMc2014816
9. Heist T, Schwartz K, Butler S. Trends in overall and non-COVID-19 hospital admissions. Kaiser Family Foundation. Accessed March 18, 2021. https://www.kff.org/health-costs/issue-brief/trends-in-overall-and-non-covid-19-hospital-admissions
10. Czeisler MÉ, Marynak K, Clarke KEN, et al. Delay or avoidance of medical care because of COVID-19–related concerns — United States, June 2020. MMWR Morb Mortal Wkly Rep. 2020;69(36);1250-1257. https://doi.org/10.15585/mmwr.mm6936a4
11. Chen J, McGeorge R. Spillover effects of the COVID-19 pandemic could drive long-term health consequences for non-COVID-19 patients. Health Affairs Blog. Accessed March 18, 2021. https://www.healthaffairs.org/do/10.1377/hblog20201020.566558/full/
12. Wong LE, Hawkins JE, Langness S, Murrell KL, Iris P, Sammann A. Where are all the patients? Addressing Covid-19 fear to encourage sick patients to seek emergency care. NEJM Catalyst. Accessed March 18, 2021. https://catalyst.nejm.org/doi/abs/10.1056/CAT.20.0193
13. Woolf SH, Chapman DA, Sabo RT, Weinberger DM, Hill L. Excess deaths from COVID-19 and other causes, March-April 2020. JAMA. 2020;324(5):510-513. https://doi.org/10.1001/jama.2020.11787
14. Clinical Classifications Software Refined (CCSR) for ICD-10-CM Diagnoses. Agency for Healthcare Research and Quality, Rockville, MD. Accessed April 22, 2021. https://www.hcup-us.ahrq.gov/toolssoftware/ccsr/dxccsr.jsp
15. MS-DRG Classifications and Software. Centers for Medicare & Medicaid Services. Accessed March 18, 2021. https://www.cms.gov/Medicare/Medicare-Fee-for-Service-Payment/AcuteInpatientPPS/MS-DRG-Classifications-and-Software
16. Jayasundera R, Neilly M, Smith TO, Myint PK. Are early warning scores useful predictors for mortality and morbidity in hospitalised acutely unwell older patients? A systematic review. J Clin Med. 2018;7(10):309. https://doi.org/10.3390/jcm7100309
17. Delgado-Hurtado JJ, Berger A, Bansal AB. Emergency department Modified Early Warning Score association with admission, admission disposition, mortality, and length of stay. J Community Hosp Intern Med Perspect. 2016;6(2):31456. https://doi.org/10.3402/jchimp.v6.31456
18. Woolf SH, Chapman DA, Sabo RT, Weinberger DM, Hill L, Taylor DDH. Excess deaths from COVID-19 and other causes, March-July 2020. JAMA. 2020;324(15):1562-1564. https://doi.org/10.1001/jama.2020.19545
19. Faust JS, Krumholz HM, Du C, et al. All-cause excess mortality and COVID-19–related mortality among US adults aged 25-44 years, March-July 2020. JAMA. 2021;325(8):785-787. https://doi.org/10.1001/jama.2020.24243
20. Weinberger DM, Chen J, Cohen T, et al. Estimation of excess deaths associated with the COVID-19 pandemic in the United States, March to May 2020. JAMA Intern Med. 2020;180(10):1336-1344. https://doi.org/10.1001/jamainternmed.2020.3391
21. Vandoros S. Excess mortality during the Covid-19 pandemic: Early evidence from England and Wales. Soc Sci Med. 2020; 258:113101. https://doi.org/10.1016/j.socscimed.2020.113101
22. Vestergaard LS, Nielsen J, Richter L, et al; ECDC Public Health Emergency Team for COVID-19. Excess all-cause mortality during the COVID-19 pandemic in Europe – preliminary pooled estimates from the EuroMOMO network, March to April 2020. Euro Surveill. 2020;25(26):2001214. https://doi.org/10.2807/1560-7917.ES.2020.25.26.2001214
23. Kontopantelis E, Mamas MA, Deanfield J, Asaria M, Doran T. Excess mortality in England and Wales during the first wave of the COVID-19 pandemic. J Epidemiol Community Health. 2021;75(3):213-223. https://doi.org/10.1136/jech-2020-214764
24. Liu J, Zhang L, Yan Y, et al. Excess mortality in Wuhan city and other parts of China during the three months of the covid-19 outbreak: findings from nationwide mortality registries. BMJ. 2021;372:n415. https://doi.org/10.1136/bmj.n415
25. Docherty KF, Butt JH, de Boer RA, et al. Excess deaths during the Covid-19 pandemic: An international comparison. Preprint. Posted online May 13, 2020. medRxiv. doi:https://doi.org/10.1101/2020.04.21.20073114
26. Barnett ML, Hu L, Martin T, Grabowski DC. Mortality, admissions, and patient census at SNFs in 3 US cities during the COVID-19 pandemic. JAMA. 2020;324(5):507-509. https://doi.org/10.1001/jama.2020.11642
27. Rosenbaum L. The untold toll — The pandemic’s effects on patients without Covid-19. N Engl J Med. 2020; 382:2368-2371 https://doi.org/10.1056/NEJMms2009984
28. Lai AG, Pasea L, Banerjee A, et al. Estimated impact of the COVID-19 pandemic on cancer services and excess 1-year mortality in people with cancer and multimorbidity: near real-time data on cancer care, cancer deaths and a population-based cohort study. BMJ Open. 2020;10(11):e043828. https://doi.org/10.1136/bmjopen-2020-043828
29. Van de Haar J, Hoes LR, Coles CE, et al. Caring for patients with cancer in the COVID-19 era. Nat Med. 2020;26(5):665-671. https://doi.org/10.1038/s41591-020-0874-8
30. Traylor AM, Tannenbaum SI, Thomas EJ, Salas E. Helping healthcare teams save lives during COVID-19: insights and countermeasures from team science. Am Psychol. 2020;76(1):1-13. https://doi.org/10.1037/amp0000750
31. Grimm CA. Hospital experiences responding to the COVID-19 pandemic: results of a National Pulse Survey March 23–27. U.S. Department of Health and Human Services Office of Inspector General; 2020. https://oig.hhs.gov/oei/reports/oei-06-20-00300.pdf
© 2021 Society of Hospital Medicine
Trends and Variation in the Use of Observation Stays at Children’s Hospitals
Payors have been refining reimbursement policies for observation and inpatient stays over the past decade, and the effects on the healthcare payment system are significant.1-4 Advocates claim that observation status could improve efficiency in the use of healthcare resources by reducing emergency department (ED) crowding and lowering costs for inpatient care.5,6 Critics consider observation status to be a cost-shifting strategy that could lead to financial burdens for patients and hospitals.7,8
Although reimbursement policies for observation stays traditionally have been set by the Centers for Medicare and Medicaid Services (CMS) in a uniform manner,4,8 state Medicaid programs and commercial health insurers have developed a variety of policies for using observation status in broader populations and hospitals.9-15 Coverage criteria and implementation timelines of these policies vary by states and commercial insurers.11-15 For example, the California Department of Health Care Services did not have a specific reimbursement rate for observation stays in 2020, while some state Medicaid programs have had reimbursement policies for observation services in place since 2010.11-15 These inconsistencies likely result in greater variation in use of observation stays across children’s hospitals than general hospitals.
Previous studies have shown rising trends in use of observation stays among adult patient populations and related implications for patients and general hospitals,16-19 but few studies have reported the trends for pediatric populations. In this study, we sought to (1) describe recent trends of observation stays for pediatric populations at children’s hospitals from 2010 through 2019 and (2) investigate features of this shifting pattern for pediatric populations and hospital-level use of observation stays.
METHODS
Study Design, Data, and Populations
We performed a retrospective analysis of the Pediatric Health Information System (PHIS), an administrative database that contains inpatient, observation, ambulatory, and ED encounter-level data from 50 not-for-profit, tertiary care children’s hospitals affiliated with the Children’s Hospital Association (CHA).20 PHIS has an indicator to classify patient types (inpatient, observation, ED visits, ambulatory surgery, clinic visit, and others). The data are de-identified at the time of submission and subjected to validity and reliability checks by CHA and Truven Health Analytics (Ann Arbor, MI) before being included in PHIS. Each encounter in PHIS has only one patient type; therefore, encounters that transition to a higher level of care are assigned to their highest level of care (eg, a patient transitions from observation to inpatient status is classified as an inpatient encounter) to avoid duplicate counting.
To ensure consistent evaluations over time, we included 29 children’s hospitals that consistently reported both inpatient and observation data to PHIS across all quarters from 2010 through 2019. We identified the 20 most common clinical conditions using the All Patients Refined Diagnosis Related Groups (APR-DRGs; 3M Corporation) based upon their total frequencies of observation and inpatient stays over the study period. Regression analyses were conducted using all encounters within the 20 most common APR-DRGs.
Because all data have been de-identified in the PHIS database, the institutional review board at Ann and Robert H. Lurie Children’s Hospital of Chicago granted this study institutional review board–exempt status.
Main Outcome and Measures
We first presented longitudinal trends of observation stays for children’s hospitals using annual percentage of observation stays defined as:
To determine whether different pediatric populations have different trends of observation stays, we measured the growth rates of observation stays for each APR-DRG. Specifically, we first calculated the percentage of observation stays by APR-DRGs and years as described, and then calculated the growth rate of observation stays for each APR-DRG:
Next, we employed prolonged length of stay (LOS) and hospitalization resource-intensity scores for kids (H-RISK) to further investigate the shifting pattern of observation stays. Because most state Medicaid and commercial policies dictate that observation stays should not last longer than 48 hours, we defined prolonged LOS as >2 days.11-15 We defined the annual percentage of observation stays with prolonged LOS for each year as:
Numerators and denominators of the three measures were obtained by pooling all children’s hospitals included in this study. H-RISK is a continuous variable developed by CHA to measure use of intensive care for children, which is comparable across various APR-DRGs.21 Changes in the empirical distribution of H-RISK from observation stays were presented over years using percentiles.
Other measures included sex, age, race, payor, and LOS. To investigate the use of observation stays among payors, we categorized payors into five groups: private, in-state Medicaid (managed care), in-state Medicaid (Children’s Health Insurance Program [CHIP]/others), other government, and all others, according to the data availability. The “private” group consisted of commercial preferred provider organizations, commercial health maintenance organizations, and commercial others. We combined both CHIP and in-state Medicaid (others), including Medicaid fee-for-service or unspecified Medicaid together as “in-state Medicaid (CHIP/others).” Detailed categorization information is summarized in Appendix Table 1. LOS was classified into four groups: 1 day (24 hours), 2 days (48 hours), 3 to 4 days, and >4 days.
Statistical Analysis
Descriptive statistics were stratified by inpatient and observation status and were summarized using frequency, percent, median, and interquartile range (IQR). Chi-square or Wilcoxon rank-sum tests were performed to examine differences between observation and inpatient status. Trends in annual percentage of observation stays and annual percentage of observation stays with prolonged LOS were estimated using first-order autoregressive models, in which year was considered a continuous variable. A nonparametric measure of rank correlation (Spearman’s rank correlation coefficient) was employed to evaluate the correlation between year and H-RISK from observation stays.
The risk-adjusted probability of being admitted as an observation stay was estimated using generalized linear mixed models by adjusting for year, age, sex, race, payor, LOS, H-RISK, and a random intercept for each hospital to control for patient clustering within a hospital (Appendix Model). Hospital-level use of observation stays was measured by risk-adjusted percent use of observation stays for each hospital using the predicted values from generalized linear mixed models. All analyses were performed using SAS software, version 9.4 (SAS Institute) and R (R Core Team, 2019), and P < .05 was considered statistically significant.
RESULTS
Increasing Trend of Observation Stays
Over the study period, there were 5,611,001 encounters, including 3,901,873 (69.5%) inpatient and 1,709,128 (30.5%) observation stays (Appendix Table 1). The number of observation stays increased from 117,246 in 2010 to 207,842 in 2019, and the number of inpatient stays slightly increased from 378,433 to 397,994 over the 10 years (Appendix Table 1). Because of different growth rates between observation and inpatient status, the annual percentage of observation stays increased from 23.7% in 2010 to 34.3% in 2019, while the annual percentage of inpatient stays decreased from 76.3% in 2010 to 65.7% in 2019 (Appendix Table 1; Figure 1, P < .001).
Different Growth Rates of Observation Stays for Various Pediatric Populations
As shown in the Table, growth rates of observation stays increased for 19 of the 20 most common APR-DRGs. The four APR-DRGs having the highest growth rates in observation stays were appendectomy, diabetes mellitus, kidney and urinary tract infections, and cellulitis and other bacterial skin infections (Appendix Figure). In particular, the annual percentage of observation stays for appendectomy increased from 19.8% in 2010 to 54.7% in 2019, with the number of observation stays growing from 2,321 to 7,876, while the number of inpatient stays decreased from 9,384 to 6,535 (Appendix Figure). The annual percentage of observation stays for diabetes mellitus increased from 8.16% in 2010 to 22.74% in 2019. Tonsil and adenoid procedures consistently held the largest numbers of observation stays across the 10 years among all the APR-DRGs, with 115,207 and 31,125 total observation and inpatient stays, respectively (Table).
Characteristics of Observation and Inpatient Stays
Patient characteristics are summarized in Appendix Table 1. There were 542,344 (32.9%) observation stays among patients with in-state Medicaid (managed care), and 241,157 (27.4%) observation stays among in-state Medicaid (CHIP/others). The percentages of observation and inpatient stays were 29.8% and 70.2% for private payor, as well as 29.6% and 70.4% for other government payor. Overall, the median (IQR) of H-RISK among observation stays was 0.79 (0.57-1.19) vs 1.23 (0.72-2.43) for inpatient stays. There were 1,410,694 (82.5%) observation stays discharged within 1 day and 243,972 (14.3%) observation stays discharged within 2 days. However, there were 47,413 (2.8%) and 7,049 (0.4%) observation stays with LOS 3 to 4 days or >4 days, respectively.
Shifting Pattern in Observation Stays
The annual percentage of observation stays with prolonged LOS (>2 days) rose from 1.1% in 2010 to 4.6% in 2019 (P < .001; Figure 2). The empirical distribution of H-RISK from observation stays by years further suggests a slightly increasing trend in intensity of care under observation stays. As shown in Appendix Table 2, although the 1st, 5th, 10th, 25th, and 99th percentiles of H-RISK were relatively stable, the 50th, 75th, 90th, and 95th percentiles of H-RISK were increasing over time. The correlation between year and intensity of care used under observation stays (H-RISK from observation stays) was found to be weak but significantly positive (Spearman correlation coefficients = 0.04; P < .001).
Interaction coefficients from our regression model demonstrate that the existing inverse association between H-RISK and odds of admission as an observation stay became less negative over the years. In 2010, the adjusted odds ratio (OR) of H-RISK was 0.57 (95% CI, 0.55-0.59). By 2017, the adjusted OR had increased to 0.65 (95% CI, 0.64-0.66). Compared with 2010, the seven adjusted ORs of H-RISK at years 2012 through 2018 were observed to be higher and statistically significant (P < .001, Appendix Table 3).
Hospitals-Level Use of Observation Stays
After adjusting for all covariates and hospital random effects, hospital-level use of observation stays increased between 2010 and 2019 for 26 out of 29 children’s hospitals. Although observation status essentially was not used at two children’s hospitals over the study period, the median hospital-level use of observation stays was 26% in 2010 (IQR, 3%-36%) and increased to 46% (IQR: 39%; 55%) in 2019. As shown in Figure 3, the number of hospitals with a low percentage of observation stays (<26%) decreased from 15 in 2010 to 4 in 2019. The number of hospitals with a high percentage of observation stays (≥51%) increased from 5 in 2010 to 10 in 2019. Nevertheless, there remained significant variation in the use of observation stays, and the hospital-level use ranged from 0% to 67% in 2019.
DISCUSSION
By 2020, observation status has become a key component of healthcare for pediatric patients, and its relevance for children’s hospitals recently has been described.22,23 However, trends in observation stays for pediatric populations are not known. This represents the first study showing temporal trends of observation stays at children’s hospitals after 2010. Our results confirm that the increase in observation stays for pediatric populations is not attributable to decreasing patient acuity at children’s hospitals. We found a weak but significantly positive correlation between year and intensity of care used under observation stays. Although this correlation might not be clinically important, it demonstrates that patient acuity in observation stays is not decreasing. Regression results suggest that observation stays now encompass patients who need relatively higher intensity of care compared with those admitted under observation status in 2010.
This study also identifies a unique pattern in the use of observation stays among pediatric populations. Earlier studies exclusively focused on observation stays that were admitted from EDs.24 Our results indicate that observation status has been used beyond a bridge from ED care to inpatient admission. In particular, observation status has expanded to include pediatric populations with more diverse clinical conditions (eg, appendicitis and diabetes mellitus), and has become a substantial component of postprocedural admissions (Appendix Figure). Looking forward, it is likely that the use of observation stays might surpass inpatient admissions for more conditions that primarily involve short-term stays.
Observation status originally was designed as a reimbursement strategy for patients who needed short stays in dedicated ED units or hospitals, but did not qualify for inpatient services.5,25 After several changes in reimbursement policies, CMS released the “two midnight rule” for Medicare beneficiaries in 2013, which replaced condition-based criteria with time-based criteria to determine an inpatient or observation stay.1 Some Medicaid programs and commercial payors have developed similar policies. Unlike the universal policy for Medicare populations, the regulations for pediatric populations vary by states and health insurers.11-15,26-28 This might partially explain the wide variation observed among children’s hospital-level use of observation stays. For example, the California Medicaid program did not have a reimbursement rate for observation services as of 2020, while the Texas Medicaid program has had a policy for observation stays since 2010.12,13 We found that two children’s hospitals in California had the lowest use of observation stays (almost zero), whereas the hospital-level use of observation stays was more than 50% for three out of four children’s hospitals in Texas. In addition to reimbursement policies, individual hospitals also might have different strategies for observation status designation. An earlier survey showed that there was lack of consistency in billing and payor-based designations of observation status at children’s hospitals.29 These findings suggest that children’s hospital-level use of observation stays likely is influenced by reimbursement policy and practical strategy for observation status determination.
Earlier studies reported that observation status could be a more efficient use of healthcare resources.5,6 However, there are still at least two concerns relevant to children’s hospitals during the last decade. The first is whether the use of observation stays can promote cost-saving or if it is just a cost-shifting strategy. An earlier study demonstrated that observation stays with prolonged LOS might increase risk of cost-sharing among adult patients.29 Our study reveals an increasing trend of observation stays with prolonged LOS for pediatric patients. Similar to adult patients, LOS exceeding 24 or 48 hours could lead to uncovered healthcare costs and financial burdens on families.30-32 Meanwhile, children’s hospitals also might take on a higher financial liability by implementing observation status. Earlier studies have indicated that resource use between observation and inpatient stays at children’s hospitals is similar, and increasing use of observation stays might lead to financial risk rather than cost effectiveness.33 Further, administrative costs of observation determination are considerably high.34 Medicaid is the major payor for pediatric patients in children’s hospitals. In this study, more than 50% of encounters were paid through Medicaid programs. It is well known that Medicaid reimbursement rates are lower than Medicare and commercial plans.35 Therefore, the cost-saving conclusion drawn from Medicare patients cannot be generalized to pediatric populations at children’s hospitals without cautious reevaluation.
A second concern with increasing use of observation stays is selection bias in public reporting and comparisons of hospital performance. Presently, four main categories of quality indicators established by the Agency for Healthcare Research and Quality rely heavily on inpatient encounters.36 In this study, we found that the range of hospital-level use of observation stays was large. In 2019, the risk-adjusted percent use of observation stays was less than 5% at three hospitals, while the percent use was greater than 60% in another three hospitals. Therefore, comparisons made without uniform accounting of observation stays might have significant implications for national rankings of children’s hospitals across the United States. These consequences have been investigated in several published studies.22,23,37-39
There are several limitations to our study. First, the study sample was limited to children’s hospitals that consistently reported inpatient and observation data over the entire study period. Eighteen hospitals (86%) excluded from this study did not consistently submit inpatient and observation data to PHIS from 2010 through 2019. The primary purpose of this study was to present temporal trends of observation stays for children’s hospitals, and it was important to build the hospital cohort based on valid and consistent data during the study period. Appendix Table 4 presents differences of hospital characteristics by included and excluded groups of hospitals. Excluded hospitals might have fewer resources (eg, fewer pediatric intensive care beds). Nonetheless, the selection of hospitals was optimized based on data availability. Second, this study was a retrospective review of an administrative database of children’s hospitals and units. The sample does not represent all children’s hospitals or pediatric patients in the United States, but there are no available data sources—that we know of—that can generate national estimates for both inpatient and observation stays. Third, we did not attempt to conclusively infer any causal effects, and several factors could explain the increasing trends, such as reimbursement policies, hospital-level implementation strategies, determination guidelines for observation status designation, as well as changes in clinical care. Further studies should investigate impact of these factors on the use of observation stays for pediatric patients and children’s hospitals.
CONCLUSION
Observation status has been increasingly used for pediatric patients with more diverse clinical conditions, and there is a rising trend of prolonged LOS among observation stays since 2010. Considerable variation exists in hospital-level use of observation stays across children’s hospitals. Observation status could be an opportunity to improve efficiency of healthcare resource use or could lead to a financial risk for patients with prolonged LOS. Future studies should explore appropriateness of observation care in clinical practice through leveraging efficient care and alleviating financial risk.
1. Centers for Medicare & Medicaid Services. Fact Sheet: Two-Midnight Rule. Accessed April 11, 2021. https://www.cms.gov/newsroom/fact-sheets/fact-sheet-two-midnight-rule-0
2. BlueCross BlueShield of Rhode Island. Payment Policy Outpaient Observation. Accessed April 11, 2021. https://www.bcbsri.com/sites/default/files/polices/Outpatient-Observation.pdf
3. Blue Cross Blue Shield of Illinois. Observation Services Tool for Applying MCG Care Guidelines Clinical Payment and Coding Policy. Accessed April 11, 2021. https://www.bcbsil.com/pdf/standards/observation_services_cpcp.pdf
4. Medicare.gov. Inpatient or outpatient hospital status affects your costs. Accessed April 11, 2021. https://www.medicare.gov/what-medicare-covers/what-part-a-covers/inpatient-or-outpatient-hospital-status
5. Ross MA, Hockenberry JM, Mutter R, Barrett M, Wheatley M, Pitts SR. Protocol-driven emergency department observation units offer savings, shorter stays, and reduced admissions. Health Aff (Millwood). 2013;32(12):2149-2156. https://doi.org/10.1377/hlthaff.2013.0662
6. Baugh CW, Venkatesh AK, Hilton JA, Samuel PA, Schuur JD, Bohan JS. Making greater use of dedicated hospital observation units for many short-stay patients could save $3.1 billion a year. Health Aff (Millwood). 2012;31(10):2314-2323. https://doi.org/10.1377/hlthaff.2011.0926
7. Sheehy AM, Graf B, Gangireddy S, et al. Hospitalized but not admitted: characteristics of patients with “observation status” at an academic medical center. JAMA Intern Med. 2013;173(21):1991-1998. https://doi.org/10.1001/jamainternmed.2013.8185
8. Baugh CW, Schuur JD. Observation care—high-value care or a cost-shifting loophole? N Engl J Med. 2013;369(4):302-305. https://doi.org/10.1056/NEJMp1304493
9. Missouri Hospital Association. A patient’s guide to observation care. Accessed April 11, 2021. https://www.mhanet.com/mhaimages/PatientsGuideToObservationCareFlyer.pdf
10. Cigna. Employee-paid hospital care coverage- summary of benefits. Accessed April 11, 2021. https://www.cigna.com/iwov-resources/national-second-sale/docs/healthy-benefits/updated-HC-benefit-summary.pdf
11. BlueCross BlueShield of Minnesota. Reimbursement policy-observation care services. Accessed April 11, 2021. https://www.bluecrossmn.com/sites/default/files/DAM/2020-07/Evaluation%20and%20Management%20004_Observation%20Care%20Services%20_09.04.17.pdf
12. California Department of Health Care Services. Public Hospital Project Frequently Asked Questions. Accessed April 11, 2021. https://www.dhcs.ca.gov/provgovpart/Documents/Public%20Hospital%20Project/PHP_Final_FAQs_January2013ADA.pdf
13. Texas Medicaid & Healthcare Partnership. Inpatient and Outpatient Hospital Servicces Handbook. Accessed May 29, 2021. https://www.tmhp.com/sites/default/files/microsites/provider-manuals/tmppm/html/TMPPM/2_Inpatient_Outpatient_Hosp_Srvs/2_Inpatient_Outpatient_Hosp_Srvs.htm
14. Alabama Medicaid. Outpatient observation. Accessed April 11, 2021. https://medicaid.alabama.gov/news_detail.aspx?ID=5121
15. NC Medicaid. Medicaid and Health Choice Clinical Coverage Policy No: 2A-1. Accessed April 11, 2021. https://files.nc.gov/ncdma/documents/files/2A-1_0.pdf
16. Feng Z, Wright B, Mor V. Sharp rise in Medicare enrollees being held in hospitals for observation raises concerns about causes and consequences. Health Aff (Millwood). 2012;31(6):1251-1259. https://doi.org/10.1377/hlthaff.2012.0129
17. Wright B, O’Shea AM, Ayyagari P, Ugwi PG, Kaboli P, Vaughan Sarrazin M. Observation rates at veterans’ hospitals more than doubled during 2005-13, similar to Medicare trends. Health Aff (Millwood). 2015;34(10):1730-1737. https://doi.org/10.1377/hlthaff.2014.1474
18. Wright B, Jung HY, Feng Z, Mor V. Hospital, patient, and local health system characteristics associated with the prevalence and duration of observation care. Health Serv Res. 2014;49(4):1088-1107. https://doi.org/10.1111/1475-6773.12166
19. Sabbatini AK, Wright B, Hall MK, Basu A. The cost of observation care for commercially insured patients visiting the emergency department. Am J Emerg Med. 2018;36(9):1591-1596. https://doi.org/10.1016/j.ajem.2018.01.040
20. Children’s Hospital Association. Pediatric health information system. Accessed April 11, 2021. https://www.childrenshospitals.org/phis
21. Richardson T, Rodean J, Harris M, Berry J, Gay JC, Hall M. Development of hospitalization resource intensity scores for kids (H-RISK) and comparison across pediatric populations. J Hosp Med. 2018;13(9):602-608. https://doi.org/10.12788/jhm.2948
22. Gay JC, Hall M, Morse R, Fieldston ES, Synhorst DC, Macy ML.Observation encounters and length of stay benchmarking in children’s hospitals. Pediatrics. 2020;146(5):e20200120. https://doi.org/10.1542/peds.2020-0120
23. Synhorst DC, Hall M, Harris M, et al. Hospital observation status and readmission rates. Pediatrics. 2020;146(5):e2020003954. https://doi.org/10.1542/peds.2020-003954
24. Macy ML, Hall M, Shah SS, et al. Pediatric observation status: are we overlooking a growing population in children’s hospitals? J Hosp Med. 2012;7(7):530-536. https://doi.org/10.1002/jhm.1923
25. Macy ML, Kim CS, Sasson C, Lozon MM, Davis MM. Pediatric observation units in the United States: a systematic review. J Hosp Med. 2010;5(3):172-182. https://doi.org/10.1002/jhm.592
26. UnitedHealthcare. Observation services policy, facility. Accessed April 11, 2021. https://www.uhcprovider.com/content/dam/provider/docs/public/policies/medicaid-comm-plan-reimbursement/UHCCP-Facility-Observation-Services-Policy-(F7106).pdf
27. Cal SB-1076§1253.7. General acute care hospitals: observation services – Health and Safety. Accessed April 11, 2021. https://leginfo.legislature.ca.gov/faces/billTextClient.xhtml?bill_id=201520160SB1076
28. Nebraska Total Care. 2021 Provider Billing Guide. Accessed April 11, 2021. https://www.nebraskatotalcare.com/content/dam/centene/Nebraska/PDFs/ProviderRelations/NTC_Nebraska_Total_Care_Provider_Billing_Guide_508.pdf
29. Macy ML, Hall M, Shah SS, et al. Differences in designations of observation care in US freestanding children’s hospitals: are they virtual or real? J Hosp Med. 2012;7(4):287-293. https://doi.org/10.1002/jhm.949
30. Hockenberry JM, Mutter R, Barrett M, Parlato J, Ross MA. Factors associated with prolonged observation services stays and the impact of long stays on patient cost. Health Serv Res. 2014;49(3):893-909. https://doi.org/10.1111/1475-6773.12143
31. Anthem BlueCross BlueShield. Ohio Provider Manual. Accessed April11, 2021. https://www11.anthem.com/provider/oh/f1/s0/t0/pw_g357368.pdf?refer=ahpprovider&state=oh
32. Humana. Provider manual for physicians, hospitals and healthcare providers. Accessed April 11, 2021. https://docushare-web.apps.cf.humana.com/Marketing/docushare-app?file=3932669
33. Fieldston ES, Shah SS, Hall M, et al. Resource utilization for observation-status stays at children’s hospitals. Pediatrics. 2013;131(6):1050-1058 https://doi.org/10.1542/peds.2012-249
34. Tejedor-Sojo J. Observation status-a name at what cost? Hosp Pediatr. 2014;4(5):321-323. https://doi.org/10.1542/hpeds.2014-0037.
35. Selden TM, Karaca Z, Keenan P, White C, Kronick R. The growing difference between public and private payment rates for inpatient hospital care. Health Aff (Millwood). 2015;34(12):2147-2150. https://doi.org/10.1377/hlthaff.2015.0706
36. Agency for Healthcare Research and Quality. AHRQ Quality Indicators. Accessed April 11, 2021. https://www.qualityindicators.ahrq.gov
37. Figueroa JF, Burke LG, Zheng J, Orav EJ, Jha AK. Trends in hospitalization vs observation stay for ambulatory care-sensitive conditions. JAMA Intern Med. 2019;179(12):1714-1716. https://doi.org/10.1001/jamainternmed.2019.3177
38. Markham JL, Hall M, Gay JC, Bettenhausen JL, Berry JG. Length of stay and cost of pediatric readmissions. Pediatrics. 2018;141(4):e20172934. https://doi.org/10.1542/peds.2017-2934.
39. Overman RA, Freburger JK, Assimon MM, Li X, Brookhart, MA. Observation stays in administrative claims databases: underestimation of hospitalized cases. Pharmacoepidemiol Drug Saf. 2014;23(9):902-910. https://doi.org/10.1002/pds.3647.
Payors have been refining reimbursement policies for observation and inpatient stays over the past decade, and the effects on the healthcare payment system are significant.1-4 Advocates claim that observation status could improve efficiency in the use of healthcare resources by reducing emergency department (ED) crowding and lowering costs for inpatient care.5,6 Critics consider observation status to be a cost-shifting strategy that could lead to financial burdens for patients and hospitals.7,8
Although reimbursement policies for observation stays traditionally have been set by the Centers for Medicare and Medicaid Services (CMS) in a uniform manner,4,8 state Medicaid programs and commercial health insurers have developed a variety of policies for using observation status in broader populations and hospitals.9-15 Coverage criteria and implementation timelines of these policies vary by states and commercial insurers.11-15 For example, the California Department of Health Care Services did not have a specific reimbursement rate for observation stays in 2020, while some state Medicaid programs have had reimbursement policies for observation services in place since 2010.11-15 These inconsistencies likely result in greater variation in use of observation stays across children’s hospitals than general hospitals.
Previous studies have shown rising trends in use of observation stays among adult patient populations and related implications for patients and general hospitals,16-19 but few studies have reported the trends for pediatric populations. In this study, we sought to (1) describe recent trends of observation stays for pediatric populations at children’s hospitals from 2010 through 2019 and (2) investigate features of this shifting pattern for pediatric populations and hospital-level use of observation stays.
METHODS
Study Design, Data, and Populations
We performed a retrospective analysis of the Pediatric Health Information System (PHIS), an administrative database that contains inpatient, observation, ambulatory, and ED encounter-level data from 50 not-for-profit, tertiary care children’s hospitals affiliated with the Children’s Hospital Association (CHA).20 PHIS has an indicator to classify patient types (inpatient, observation, ED visits, ambulatory surgery, clinic visit, and others). The data are de-identified at the time of submission and subjected to validity and reliability checks by CHA and Truven Health Analytics (Ann Arbor, MI) before being included in PHIS. Each encounter in PHIS has only one patient type; therefore, encounters that transition to a higher level of care are assigned to their highest level of care (eg, a patient transitions from observation to inpatient status is classified as an inpatient encounter) to avoid duplicate counting.
To ensure consistent evaluations over time, we included 29 children’s hospitals that consistently reported both inpatient and observation data to PHIS across all quarters from 2010 through 2019. We identified the 20 most common clinical conditions using the All Patients Refined Diagnosis Related Groups (APR-DRGs; 3M Corporation) based upon their total frequencies of observation and inpatient stays over the study period. Regression analyses were conducted using all encounters within the 20 most common APR-DRGs.
Because all data have been de-identified in the PHIS database, the institutional review board at Ann and Robert H. Lurie Children’s Hospital of Chicago granted this study institutional review board–exempt status.
Main Outcome and Measures
We first presented longitudinal trends of observation stays for children’s hospitals using annual percentage of observation stays defined as:
To determine whether different pediatric populations have different trends of observation stays, we measured the growth rates of observation stays for each APR-DRG. Specifically, we first calculated the percentage of observation stays by APR-DRGs and years as described, and then calculated the growth rate of observation stays for each APR-DRG:
Next, we employed prolonged length of stay (LOS) and hospitalization resource-intensity scores for kids (H-RISK) to further investigate the shifting pattern of observation stays. Because most state Medicaid and commercial policies dictate that observation stays should not last longer than 48 hours, we defined prolonged LOS as >2 days.11-15 We defined the annual percentage of observation stays with prolonged LOS for each year as:
Numerators and denominators of the three measures were obtained by pooling all children’s hospitals included in this study. H-RISK is a continuous variable developed by CHA to measure use of intensive care for children, which is comparable across various APR-DRGs.21 Changes in the empirical distribution of H-RISK from observation stays were presented over years using percentiles.
Other measures included sex, age, race, payor, and LOS. To investigate the use of observation stays among payors, we categorized payors into five groups: private, in-state Medicaid (managed care), in-state Medicaid (Children’s Health Insurance Program [CHIP]/others), other government, and all others, according to the data availability. The “private” group consisted of commercial preferred provider organizations, commercial health maintenance organizations, and commercial others. We combined both CHIP and in-state Medicaid (others), including Medicaid fee-for-service or unspecified Medicaid together as “in-state Medicaid (CHIP/others).” Detailed categorization information is summarized in Appendix Table 1. LOS was classified into four groups: 1 day (24 hours), 2 days (48 hours), 3 to 4 days, and >4 days.
Statistical Analysis
Descriptive statistics were stratified by inpatient and observation status and were summarized using frequency, percent, median, and interquartile range (IQR). Chi-square or Wilcoxon rank-sum tests were performed to examine differences between observation and inpatient status. Trends in annual percentage of observation stays and annual percentage of observation stays with prolonged LOS were estimated using first-order autoregressive models, in which year was considered a continuous variable. A nonparametric measure of rank correlation (Spearman’s rank correlation coefficient) was employed to evaluate the correlation between year and H-RISK from observation stays.
The risk-adjusted probability of being admitted as an observation stay was estimated using generalized linear mixed models by adjusting for year, age, sex, race, payor, LOS, H-RISK, and a random intercept for each hospital to control for patient clustering within a hospital (Appendix Model). Hospital-level use of observation stays was measured by risk-adjusted percent use of observation stays for each hospital using the predicted values from generalized linear mixed models. All analyses were performed using SAS software, version 9.4 (SAS Institute) and R (R Core Team, 2019), and P < .05 was considered statistically significant.
RESULTS
Increasing Trend of Observation Stays
Over the study period, there were 5,611,001 encounters, including 3,901,873 (69.5%) inpatient and 1,709,128 (30.5%) observation stays (Appendix Table 1). The number of observation stays increased from 117,246 in 2010 to 207,842 in 2019, and the number of inpatient stays slightly increased from 378,433 to 397,994 over the 10 years (Appendix Table 1). Because of different growth rates between observation and inpatient status, the annual percentage of observation stays increased from 23.7% in 2010 to 34.3% in 2019, while the annual percentage of inpatient stays decreased from 76.3% in 2010 to 65.7% in 2019 (Appendix Table 1; Figure 1, P < .001).
Different Growth Rates of Observation Stays for Various Pediatric Populations
As shown in the Table, growth rates of observation stays increased for 19 of the 20 most common APR-DRGs. The four APR-DRGs having the highest growth rates in observation stays were appendectomy, diabetes mellitus, kidney and urinary tract infections, and cellulitis and other bacterial skin infections (Appendix Figure). In particular, the annual percentage of observation stays for appendectomy increased from 19.8% in 2010 to 54.7% in 2019, with the number of observation stays growing from 2,321 to 7,876, while the number of inpatient stays decreased from 9,384 to 6,535 (Appendix Figure). The annual percentage of observation stays for diabetes mellitus increased from 8.16% in 2010 to 22.74% in 2019. Tonsil and adenoid procedures consistently held the largest numbers of observation stays across the 10 years among all the APR-DRGs, with 115,207 and 31,125 total observation and inpatient stays, respectively (Table).
Characteristics of Observation and Inpatient Stays
Patient characteristics are summarized in Appendix Table 1. There were 542,344 (32.9%) observation stays among patients with in-state Medicaid (managed care), and 241,157 (27.4%) observation stays among in-state Medicaid (CHIP/others). The percentages of observation and inpatient stays were 29.8% and 70.2% for private payor, as well as 29.6% and 70.4% for other government payor. Overall, the median (IQR) of H-RISK among observation stays was 0.79 (0.57-1.19) vs 1.23 (0.72-2.43) for inpatient stays. There were 1,410,694 (82.5%) observation stays discharged within 1 day and 243,972 (14.3%) observation stays discharged within 2 days. However, there were 47,413 (2.8%) and 7,049 (0.4%) observation stays with LOS 3 to 4 days or >4 days, respectively.
Shifting Pattern in Observation Stays
The annual percentage of observation stays with prolonged LOS (>2 days) rose from 1.1% in 2010 to 4.6% in 2019 (P < .001; Figure 2). The empirical distribution of H-RISK from observation stays by years further suggests a slightly increasing trend in intensity of care under observation stays. As shown in Appendix Table 2, although the 1st, 5th, 10th, 25th, and 99th percentiles of H-RISK were relatively stable, the 50th, 75th, 90th, and 95th percentiles of H-RISK were increasing over time. The correlation between year and intensity of care used under observation stays (H-RISK from observation stays) was found to be weak but significantly positive (Spearman correlation coefficients = 0.04; P < .001).
Interaction coefficients from our regression model demonstrate that the existing inverse association between H-RISK and odds of admission as an observation stay became less negative over the years. In 2010, the adjusted odds ratio (OR) of H-RISK was 0.57 (95% CI, 0.55-0.59). By 2017, the adjusted OR had increased to 0.65 (95% CI, 0.64-0.66). Compared with 2010, the seven adjusted ORs of H-RISK at years 2012 through 2018 were observed to be higher and statistically significant (P < .001, Appendix Table 3).
Hospitals-Level Use of Observation Stays
After adjusting for all covariates and hospital random effects, hospital-level use of observation stays increased between 2010 and 2019 for 26 out of 29 children’s hospitals. Although observation status essentially was not used at two children’s hospitals over the study period, the median hospital-level use of observation stays was 26% in 2010 (IQR, 3%-36%) and increased to 46% (IQR: 39%; 55%) in 2019. As shown in Figure 3, the number of hospitals with a low percentage of observation stays (<26%) decreased from 15 in 2010 to 4 in 2019. The number of hospitals with a high percentage of observation stays (≥51%) increased from 5 in 2010 to 10 in 2019. Nevertheless, there remained significant variation in the use of observation stays, and the hospital-level use ranged from 0% to 67% in 2019.
DISCUSSION
By 2020, observation status has become a key component of healthcare for pediatric patients, and its relevance for children’s hospitals recently has been described.22,23 However, trends in observation stays for pediatric populations are not known. This represents the first study showing temporal trends of observation stays at children’s hospitals after 2010. Our results confirm that the increase in observation stays for pediatric populations is not attributable to decreasing patient acuity at children’s hospitals. We found a weak but significantly positive correlation between year and intensity of care used under observation stays. Although this correlation might not be clinically important, it demonstrates that patient acuity in observation stays is not decreasing. Regression results suggest that observation stays now encompass patients who need relatively higher intensity of care compared with those admitted under observation status in 2010.
This study also identifies a unique pattern in the use of observation stays among pediatric populations. Earlier studies exclusively focused on observation stays that were admitted from EDs.24 Our results indicate that observation status has been used beyond a bridge from ED care to inpatient admission. In particular, observation status has expanded to include pediatric populations with more diverse clinical conditions (eg, appendicitis and diabetes mellitus), and has become a substantial component of postprocedural admissions (Appendix Figure). Looking forward, it is likely that the use of observation stays might surpass inpatient admissions for more conditions that primarily involve short-term stays.
Observation status originally was designed as a reimbursement strategy for patients who needed short stays in dedicated ED units or hospitals, but did not qualify for inpatient services.5,25 After several changes in reimbursement policies, CMS released the “two midnight rule” for Medicare beneficiaries in 2013, which replaced condition-based criteria with time-based criteria to determine an inpatient or observation stay.1 Some Medicaid programs and commercial payors have developed similar policies. Unlike the universal policy for Medicare populations, the regulations for pediatric populations vary by states and health insurers.11-15,26-28 This might partially explain the wide variation observed among children’s hospital-level use of observation stays. For example, the California Medicaid program did not have a reimbursement rate for observation services as of 2020, while the Texas Medicaid program has had a policy for observation stays since 2010.12,13 We found that two children’s hospitals in California had the lowest use of observation stays (almost zero), whereas the hospital-level use of observation stays was more than 50% for three out of four children’s hospitals in Texas. In addition to reimbursement policies, individual hospitals also might have different strategies for observation status designation. An earlier survey showed that there was lack of consistency in billing and payor-based designations of observation status at children’s hospitals.29 These findings suggest that children’s hospital-level use of observation stays likely is influenced by reimbursement policy and practical strategy for observation status determination.
Earlier studies reported that observation status could be a more efficient use of healthcare resources.5,6 However, there are still at least two concerns relevant to children’s hospitals during the last decade. The first is whether the use of observation stays can promote cost-saving or if it is just a cost-shifting strategy. An earlier study demonstrated that observation stays with prolonged LOS might increase risk of cost-sharing among adult patients.29 Our study reveals an increasing trend of observation stays with prolonged LOS for pediatric patients. Similar to adult patients, LOS exceeding 24 or 48 hours could lead to uncovered healthcare costs and financial burdens on families.30-32 Meanwhile, children’s hospitals also might take on a higher financial liability by implementing observation status. Earlier studies have indicated that resource use between observation and inpatient stays at children’s hospitals is similar, and increasing use of observation stays might lead to financial risk rather than cost effectiveness.33 Further, administrative costs of observation determination are considerably high.34 Medicaid is the major payor for pediatric patients in children’s hospitals. In this study, more than 50% of encounters were paid through Medicaid programs. It is well known that Medicaid reimbursement rates are lower than Medicare and commercial plans.35 Therefore, the cost-saving conclusion drawn from Medicare patients cannot be generalized to pediatric populations at children’s hospitals without cautious reevaluation.
A second concern with increasing use of observation stays is selection bias in public reporting and comparisons of hospital performance. Presently, four main categories of quality indicators established by the Agency for Healthcare Research and Quality rely heavily on inpatient encounters.36 In this study, we found that the range of hospital-level use of observation stays was large. In 2019, the risk-adjusted percent use of observation stays was less than 5% at three hospitals, while the percent use was greater than 60% in another three hospitals. Therefore, comparisons made without uniform accounting of observation stays might have significant implications for national rankings of children’s hospitals across the United States. These consequences have been investigated in several published studies.22,23,37-39
There are several limitations to our study. First, the study sample was limited to children’s hospitals that consistently reported inpatient and observation data over the entire study period. Eighteen hospitals (86%) excluded from this study did not consistently submit inpatient and observation data to PHIS from 2010 through 2019. The primary purpose of this study was to present temporal trends of observation stays for children’s hospitals, and it was important to build the hospital cohort based on valid and consistent data during the study period. Appendix Table 4 presents differences of hospital characteristics by included and excluded groups of hospitals. Excluded hospitals might have fewer resources (eg, fewer pediatric intensive care beds). Nonetheless, the selection of hospitals was optimized based on data availability. Second, this study was a retrospective review of an administrative database of children’s hospitals and units. The sample does not represent all children’s hospitals or pediatric patients in the United States, but there are no available data sources—that we know of—that can generate national estimates for both inpatient and observation stays. Third, we did not attempt to conclusively infer any causal effects, and several factors could explain the increasing trends, such as reimbursement policies, hospital-level implementation strategies, determination guidelines for observation status designation, as well as changes in clinical care. Further studies should investigate impact of these factors on the use of observation stays for pediatric patients and children’s hospitals.
CONCLUSION
Observation status has been increasingly used for pediatric patients with more diverse clinical conditions, and there is a rising trend of prolonged LOS among observation stays since 2010. Considerable variation exists in hospital-level use of observation stays across children’s hospitals. Observation status could be an opportunity to improve efficiency of healthcare resource use or could lead to a financial risk for patients with prolonged LOS. Future studies should explore appropriateness of observation care in clinical practice through leveraging efficient care and alleviating financial risk.
Payors have been refining reimbursement policies for observation and inpatient stays over the past decade, and the effects on the healthcare payment system are significant.1-4 Advocates claim that observation status could improve efficiency in the use of healthcare resources by reducing emergency department (ED) crowding and lowering costs for inpatient care.5,6 Critics consider observation status to be a cost-shifting strategy that could lead to financial burdens for patients and hospitals.7,8
Although reimbursement policies for observation stays traditionally have been set by the Centers for Medicare and Medicaid Services (CMS) in a uniform manner,4,8 state Medicaid programs and commercial health insurers have developed a variety of policies for using observation status in broader populations and hospitals.9-15 Coverage criteria and implementation timelines of these policies vary by states and commercial insurers.11-15 For example, the California Department of Health Care Services did not have a specific reimbursement rate for observation stays in 2020, while some state Medicaid programs have had reimbursement policies for observation services in place since 2010.11-15 These inconsistencies likely result in greater variation in use of observation stays across children’s hospitals than general hospitals.
Previous studies have shown rising trends in use of observation stays among adult patient populations and related implications for patients and general hospitals,16-19 but few studies have reported the trends for pediatric populations. In this study, we sought to (1) describe recent trends of observation stays for pediatric populations at children’s hospitals from 2010 through 2019 and (2) investigate features of this shifting pattern for pediatric populations and hospital-level use of observation stays.
METHODS
Study Design, Data, and Populations
We performed a retrospective analysis of the Pediatric Health Information System (PHIS), an administrative database that contains inpatient, observation, ambulatory, and ED encounter-level data from 50 not-for-profit, tertiary care children’s hospitals affiliated with the Children’s Hospital Association (CHA).20 PHIS has an indicator to classify patient types (inpatient, observation, ED visits, ambulatory surgery, clinic visit, and others). The data are de-identified at the time of submission and subjected to validity and reliability checks by CHA and Truven Health Analytics (Ann Arbor, MI) before being included in PHIS. Each encounter in PHIS has only one patient type; therefore, encounters that transition to a higher level of care are assigned to their highest level of care (eg, a patient transitions from observation to inpatient status is classified as an inpatient encounter) to avoid duplicate counting.
To ensure consistent evaluations over time, we included 29 children’s hospitals that consistently reported both inpatient and observation data to PHIS across all quarters from 2010 through 2019. We identified the 20 most common clinical conditions using the All Patients Refined Diagnosis Related Groups (APR-DRGs; 3M Corporation) based upon their total frequencies of observation and inpatient stays over the study period. Regression analyses were conducted using all encounters within the 20 most common APR-DRGs.
Because all data have been de-identified in the PHIS database, the institutional review board at Ann and Robert H. Lurie Children’s Hospital of Chicago granted this study institutional review board–exempt status.
Main Outcome and Measures
We first presented longitudinal trends of observation stays for children’s hospitals using annual percentage of observation stays defined as:
To determine whether different pediatric populations have different trends of observation stays, we measured the growth rates of observation stays for each APR-DRG. Specifically, we first calculated the percentage of observation stays by APR-DRGs and years as described, and then calculated the growth rate of observation stays for each APR-DRG:
Next, we employed prolonged length of stay (LOS) and hospitalization resource-intensity scores for kids (H-RISK) to further investigate the shifting pattern of observation stays. Because most state Medicaid and commercial policies dictate that observation stays should not last longer than 48 hours, we defined prolonged LOS as >2 days.11-15 We defined the annual percentage of observation stays with prolonged LOS for each year as:
Numerators and denominators of the three measures were obtained by pooling all children’s hospitals included in this study. H-RISK is a continuous variable developed by CHA to measure use of intensive care for children, which is comparable across various APR-DRGs.21 Changes in the empirical distribution of H-RISK from observation stays were presented over years using percentiles.
Other measures included sex, age, race, payor, and LOS. To investigate the use of observation stays among payors, we categorized payors into five groups: private, in-state Medicaid (managed care), in-state Medicaid (Children’s Health Insurance Program [CHIP]/others), other government, and all others, according to the data availability. The “private” group consisted of commercial preferred provider organizations, commercial health maintenance organizations, and commercial others. We combined both CHIP and in-state Medicaid (others), including Medicaid fee-for-service or unspecified Medicaid together as “in-state Medicaid (CHIP/others).” Detailed categorization information is summarized in Appendix Table 1. LOS was classified into four groups: 1 day (24 hours), 2 days (48 hours), 3 to 4 days, and >4 days.
Statistical Analysis
Descriptive statistics were stratified by inpatient and observation status and were summarized using frequency, percent, median, and interquartile range (IQR). Chi-square or Wilcoxon rank-sum tests were performed to examine differences between observation and inpatient status. Trends in annual percentage of observation stays and annual percentage of observation stays with prolonged LOS were estimated using first-order autoregressive models, in which year was considered a continuous variable. A nonparametric measure of rank correlation (Spearman’s rank correlation coefficient) was employed to evaluate the correlation between year and H-RISK from observation stays.
The risk-adjusted probability of being admitted as an observation stay was estimated using generalized linear mixed models by adjusting for year, age, sex, race, payor, LOS, H-RISK, and a random intercept for each hospital to control for patient clustering within a hospital (Appendix Model). Hospital-level use of observation stays was measured by risk-adjusted percent use of observation stays for each hospital using the predicted values from generalized linear mixed models. All analyses were performed using SAS software, version 9.4 (SAS Institute) and R (R Core Team, 2019), and P < .05 was considered statistically significant.
RESULTS
Increasing Trend of Observation Stays
Over the study period, there were 5,611,001 encounters, including 3,901,873 (69.5%) inpatient and 1,709,128 (30.5%) observation stays (Appendix Table 1). The number of observation stays increased from 117,246 in 2010 to 207,842 in 2019, and the number of inpatient stays slightly increased from 378,433 to 397,994 over the 10 years (Appendix Table 1). Because of different growth rates between observation and inpatient status, the annual percentage of observation stays increased from 23.7% in 2010 to 34.3% in 2019, while the annual percentage of inpatient stays decreased from 76.3% in 2010 to 65.7% in 2019 (Appendix Table 1; Figure 1, P < .001).
Different Growth Rates of Observation Stays for Various Pediatric Populations
As shown in the Table, growth rates of observation stays increased for 19 of the 20 most common APR-DRGs. The four APR-DRGs having the highest growth rates in observation stays were appendectomy, diabetes mellitus, kidney and urinary tract infections, and cellulitis and other bacterial skin infections (Appendix Figure). In particular, the annual percentage of observation stays for appendectomy increased from 19.8% in 2010 to 54.7% in 2019, with the number of observation stays growing from 2,321 to 7,876, while the number of inpatient stays decreased from 9,384 to 6,535 (Appendix Figure). The annual percentage of observation stays for diabetes mellitus increased from 8.16% in 2010 to 22.74% in 2019. Tonsil and adenoid procedures consistently held the largest numbers of observation stays across the 10 years among all the APR-DRGs, with 115,207 and 31,125 total observation and inpatient stays, respectively (Table).
Characteristics of Observation and Inpatient Stays
Patient characteristics are summarized in Appendix Table 1. There were 542,344 (32.9%) observation stays among patients with in-state Medicaid (managed care), and 241,157 (27.4%) observation stays among in-state Medicaid (CHIP/others). The percentages of observation and inpatient stays were 29.8% and 70.2% for private payor, as well as 29.6% and 70.4% for other government payor. Overall, the median (IQR) of H-RISK among observation stays was 0.79 (0.57-1.19) vs 1.23 (0.72-2.43) for inpatient stays. There were 1,410,694 (82.5%) observation stays discharged within 1 day and 243,972 (14.3%) observation stays discharged within 2 days. However, there were 47,413 (2.8%) and 7,049 (0.4%) observation stays with LOS 3 to 4 days or >4 days, respectively.
Shifting Pattern in Observation Stays
The annual percentage of observation stays with prolonged LOS (>2 days) rose from 1.1% in 2010 to 4.6% in 2019 (P < .001; Figure 2). The empirical distribution of H-RISK from observation stays by years further suggests a slightly increasing trend in intensity of care under observation stays. As shown in Appendix Table 2, although the 1st, 5th, 10th, 25th, and 99th percentiles of H-RISK were relatively stable, the 50th, 75th, 90th, and 95th percentiles of H-RISK were increasing over time. The correlation between year and intensity of care used under observation stays (H-RISK from observation stays) was found to be weak but significantly positive (Spearman correlation coefficients = 0.04; P < .001).
Interaction coefficients from our regression model demonstrate that the existing inverse association between H-RISK and odds of admission as an observation stay became less negative over the years. In 2010, the adjusted odds ratio (OR) of H-RISK was 0.57 (95% CI, 0.55-0.59). By 2017, the adjusted OR had increased to 0.65 (95% CI, 0.64-0.66). Compared with 2010, the seven adjusted ORs of H-RISK at years 2012 through 2018 were observed to be higher and statistically significant (P < .001, Appendix Table 3).
Hospitals-Level Use of Observation Stays
After adjusting for all covariates and hospital random effects, hospital-level use of observation stays increased between 2010 and 2019 for 26 out of 29 children’s hospitals. Although observation status essentially was not used at two children’s hospitals over the study period, the median hospital-level use of observation stays was 26% in 2010 (IQR, 3%-36%) and increased to 46% (IQR: 39%; 55%) in 2019. As shown in Figure 3, the number of hospitals with a low percentage of observation stays (<26%) decreased from 15 in 2010 to 4 in 2019. The number of hospitals with a high percentage of observation stays (≥51%) increased from 5 in 2010 to 10 in 2019. Nevertheless, there remained significant variation in the use of observation stays, and the hospital-level use ranged from 0% to 67% in 2019.
DISCUSSION
By 2020, observation status has become a key component of healthcare for pediatric patients, and its relevance for children’s hospitals recently has been described.22,23 However, trends in observation stays for pediatric populations are not known. This represents the first study showing temporal trends of observation stays at children’s hospitals after 2010. Our results confirm that the increase in observation stays for pediatric populations is not attributable to decreasing patient acuity at children’s hospitals. We found a weak but significantly positive correlation between year and intensity of care used under observation stays. Although this correlation might not be clinically important, it demonstrates that patient acuity in observation stays is not decreasing. Regression results suggest that observation stays now encompass patients who need relatively higher intensity of care compared with those admitted under observation status in 2010.
This study also identifies a unique pattern in the use of observation stays among pediatric populations. Earlier studies exclusively focused on observation stays that were admitted from EDs.24 Our results indicate that observation status has been used beyond a bridge from ED care to inpatient admission. In particular, observation status has expanded to include pediatric populations with more diverse clinical conditions (eg, appendicitis and diabetes mellitus), and has become a substantial component of postprocedural admissions (Appendix Figure). Looking forward, it is likely that the use of observation stays might surpass inpatient admissions for more conditions that primarily involve short-term stays.
Observation status originally was designed as a reimbursement strategy for patients who needed short stays in dedicated ED units or hospitals, but did not qualify for inpatient services.5,25 After several changes in reimbursement policies, CMS released the “two midnight rule” for Medicare beneficiaries in 2013, which replaced condition-based criteria with time-based criteria to determine an inpatient or observation stay.1 Some Medicaid programs and commercial payors have developed similar policies. Unlike the universal policy for Medicare populations, the regulations for pediatric populations vary by states and health insurers.11-15,26-28 This might partially explain the wide variation observed among children’s hospital-level use of observation stays. For example, the California Medicaid program did not have a reimbursement rate for observation services as of 2020, while the Texas Medicaid program has had a policy for observation stays since 2010.12,13 We found that two children’s hospitals in California had the lowest use of observation stays (almost zero), whereas the hospital-level use of observation stays was more than 50% for three out of four children’s hospitals in Texas. In addition to reimbursement policies, individual hospitals also might have different strategies for observation status designation. An earlier survey showed that there was lack of consistency in billing and payor-based designations of observation status at children’s hospitals.29 These findings suggest that children’s hospital-level use of observation stays likely is influenced by reimbursement policy and practical strategy for observation status determination.
Earlier studies reported that observation status could be a more efficient use of healthcare resources.5,6 However, there are still at least two concerns relevant to children’s hospitals during the last decade. The first is whether the use of observation stays can promote cost-saving or if it is just a cost-shifting strategy. An earlier study demonstrated that observation stays with prolonged LOS might increase risk of cost-sharing among adult patients.29 Our study reveals an increasing trend of observation stays with prolonged LOS for pediatric patients. Similar to adult patients, LOS exceeding 24 or 48 hours could lead to uncovered healthcare costs and financial burdens on families.30-32 Meanwhile, children’s hospitals also might take on a higher financial liability by implementing observation status. Earlier studies have indicated that resource use between observation and inpatient stays at children’s hospitals is similar, and increasing use of observation stays might lead to financial risk rather than cost effectiveness.33 Further, administrative costs of observation determination are considerably high.34 Medicaid is the major payor for pediatric patients in children’s hospitals. In this study, more than 50% of encounters were paid through Medicaid programs. It is well known that Medicaid reimbursement rates are lower than Medicare and commercial plans.35 Therefore, the cost-saving conclusion drawn from Medicare patients cannot be generalized to pediatric populations at children’s hospitals without cautious reevaluation.
A second concern with increasing use of observation stays is selection bias in public reporting and comparisons of hospital performance. Presently, four main categories of quality indicators established by the Agency for Healthcare Research and Quality rely heavily on inpatient encounters.36 In this study, we found that the range of hospital-level use of observation stays was large. In 2019, the risk-adjusted percent use of observation stays was less than 5% at three hospitals, while the percent use was greater than 60% in another three hospitals. Therefore, comparisons made without uniform accounting of observation stays might have significant implications for national rankings of children’s hospitals across the United States. These consequences have been investigated in several published studies.22,23,37-39
There are several limitations to our study. First, the study sample was limited to children’s hospitals that consistently reported inpatient and observation data over the entire study period. Eighteen hospitals (86%) excluded from this study did not consistently submit inpatient and observation data to PHIS from 2010 through 2019. The primary purpose of this study was to present temporal trends of observation stays for children’s hospitals, and it was important to build the hospital cohort based on valid and consistent data during the study period. Appendix Table 4 presents differences of hospital characteristics by included and excluded groups of hospitals. Excluded hospitals might have fewer resources (eg, fewer pediatric intensive care beds). Nonetheless, the selection of hospitals was optimized based on data availability. Second, this study was a retrospective review of an administrative database of children’s hospitals and units. The sample does not represent all children’s hospitals or pediatric patients in the United States, but there are no available data sources—that we know of—that can generate national estimates for both inpatient and observation stays. Third, we did not attempt to conclusively infer any causal effects, and several factors could explain the increasing trends, such as reimbursement policies, hospital-level implementation strategies, determination guidelines for observation status designation, as well as changes in clinical care. Further studies should investigate impact of these factors on the use of observation stays for pediatric patients and children’s hospitals.
CONCLUSION
Observation status has been increasingly used for pediatric patients with more diverse clinical conditions, and there is a rising trend of prolonged LOS among observation stays since 2010. Considerable variation exists in hospital-level use of observation stays across children’s hospitals. Observation status could be an opportunity to improve efficiency of healthcare resource use or could lead to a financial risk for patients with prolonged LOS. Future studies should explore appropriateness of observation care in clinical practice through leveraging efficient care and alleviating financial risk.
1. Centers for Medicare & Medicaid Services. Fact Sheet: Two-Midnight Rule. Accessed April 11, 2021. https://www.cms.gov/newsroom/fact-sheets/fact-sheet-two-midnight-rule-0
2. BlueCross BlueShield of Rhode Island. Payment Policy Outpaient Observation. Accessed April 11, 2021. https://www.bcbsri.com/sites/default/files/polices/Outpatient-Observation.pdf
3. Blue Cross Blue Shield of Illinois. Observation Services Tool for Applying MCG Care Guidelines Clinical Payment and Coding Policy. Accessed April 11, 2021. https://www.bcbsil.com/pdf/standards/observation_services_cpcp.pdf
4. Medicare.gov. Inpatient or outpatient hospital status affects your costs. Accessed April 11, 2021. https://www.medicare.gov/what-medicare-covers/what-part-a-covers/inpatient-or-outpatient-hospital-status
5. Ross MA, Hockenberry JM, Mutter R, Barrett M, Wheatley M, Pitts SR. Protocol-driven emergency department observation units offer savings, shorter stays, and reduced admissions. Health Aff (Millwood). 2013;32(12):2149-2156. https://doi.org/10.1377/hlthaff.2013.0662
6. Baugh CW, Venkatesh AK, Hilton JA, Samuel PA, Schuur JD, Bohan JS. Making greater use of dedicated hospital observation units for many short-stay patients could save $3.1 billion a year. Health Aff (Millwood). 2012;31(10):2314-2323. https://doi.org/10.1377/hlthaff.2011.0926
7. Sheehy AM, Graf B, Gangireddy S, et al. Hospitalized but not admitted: characteristics of patients with “observation status” at an academic medical center. JAMA Intern Med. 2013;173(21):1991-1998. https://doi.org/10.1001/jamainternmed.2013.8185
8. Baugh CW, Schuur JD. Observation care—high-value care or a cost-shifting loophole? N Engl J Med. 2013;369(4):302-305. https://doi.org/10.1056/NEJMp1304493
9. Missouri Hospital Association. A patient’s guide to observation care. Accessed April 11, 2021. https://www.mhanet.com/mhaimages/PatientsGuideToObservationCareFlyer.pdf
10. Cigna. Employee-paid hospital care coverage- summary of benefits. Accessed April 11, 2021. https://www.cigna.com/iwov-resources/national-second-sale/docs/healthy-benefits/updated-HC-benefit-summary.pdf
11. BlueCross BlueShield of Minnesota. Reimbursement policy-observation care services. Accessed April 11, 2021. https://www.bluecrossmn.com/sites/default/files/DAM/2020-07/Evaluation%20and%20Management%20004_Observation%20Care%20Services%20_09.04.17.pdf
12. California Department of Health Care Services. Public Hospital Project Frequently Asked Questions. Accessed April 11, 2021. https://www.dhcs.ca.gov/provgovpart/Documents/Public%20Hospital%20Project/PHP_Final_FAQs_January2013ADA.pdf
13. Texas Medicaid & Healthcare Partnership. Inpatient and Outpatient Hospital Servicces Handbook. Accessed May 29, 2021. https://www.tmhp.com/sites/default/files/microsites/provider-manuals/tmppm/html/TMPPM/2_Inpatient_Outpatient_Hosp_Srvs/2_Inpatient_Outpatient_Hosp_Srvs.htm
14. Alabama Medicaid. Outpatient observation. Accessed April 11, 2021. https://medicaid.alabama.gov/news_detail.aspx?ID=5121
15. NC Medicaid. Medicaid and Health Choice Clinical Coverage Policy No: 2A-1. Accessed April 11, 2021. https://files.nc.gov/ncdma/documents/files/2A-1_0.pdf
16. Feng Z, Wright B, Mor V. Sharp rise in Medicare enrollees being held in hospitals for observation raises concerns about causes and consequences. Health Aff (Millwood). 2012;31(6):1251-1259. https://doi.org/10.1377/hlthaff.2012.0129
17. Wright B, O’Shea AM, Ayyagari P, Ugwi PG, Kaboli P, Vaughan Sarrazin M. Observation rates at veterans’ hospitals more than doubled during 2005-13, similar to Medicare trends. Health Aff (Millwood). 2015;34(10):1730-1737. https://doi.org/10.1377/hlthaff.2014.1474
18. Wright B, Jung HY, Feng Z, Mor V. Hospital, patient, and local health system characteristics associated with the prevalence and duration of observation care. Health Serv Res. 2014;49(4):1088-1107. https://doi.org/10.1111/1475-6773.12166
19. Sabbatini AK, Wright B, Hall MK, Basu A. The cost of observation care for commercially insured patients visiting the emergency department. Am J Emerg Med. 2018;36(9):1591-1596. https://doi.org/10.1016/j.ajem.2018.01.040
20. Children’s Hospital Association. Pediatric health information system. Accessed April 11, 2021. https://www.childrenshospitals.org/phis
21. Richardson T, Rodean J, Harris M, Berry J, Gay JC, Hall M. Development of hospitalization resource intensity scores for kids (H-RISK) and comparison across pediatric populations. J Hosp Med. 2018;13(9):602-608. https://doi.org/10.12788/jhm.2948
22. Gay JC, Hall M, Morse R, Fieldston ES, Synhorst DC, Macy ML.Observation encounters and length of stay benchmarking in children’s hospitals. Pediatrics. 2020;146(5):e20200120. https://doi.org/10.1542/peds.2020-0120
23. Synhorst DC, Hall M, Harris M, et al. Hospital observation status and readmission rates. Pediatrics. 2020;146(5):e2020003954. https://doi.org/10.1542/peds.2020-003954
24. Macy ML, Hall M, Shah SS, et al. Pediatric observation status: are we overlooking a growing population in children’s hospitals? J Hosp Med. 2012;7(7):530-536. https://doi.org/10.1002/jhm.1923
25. Macy ML, Kim CS, Sasson C, Lozon MM, Davis MM. Pediatric observation units in the United States: a systematic review. J Hosp Med. 2010;5(3):172-182. https://doi.org/10.1002/jhm.592
26. UnitedHealthcare. Observation services policy, facility. Accessed April 11, 2021. https://www.uhcprovider.com/content/dam/provider/docs/public/policies/medicaid-comm-plan-reimbursement/UHCCP-Facility-Observation-Services-Policy-(F7106).pdf
27. Cal SB-1076§1253.7. General acute care hospitals: observation services – Health and Safety. Accessed April 11, 2021. https://leginfo.legislature.ca.gov/faces/billTextClient.xhtml?bill_id=201520160SB1076
28. Nebraska Total Care. 2021 Provider Billing Guide. Accessed April 11, 2021. https://www.nebraskatotalcare.com/content/dam/centene/Nebraska/PDFs/ProviderRelations/NTC_Nebraska_Total_Care_Provider_Billing_Guide_508.pdf
29. Macy ML, Hall M, Shah SS, et al. Differences in designations of observation care in US freestanding children’s hospitals: are they virtual or real? J Hosp Med. 2012;7(4):287-293. https://doi.org/10.1002/jhm.949
30. Hockenberry JM, Mutter R, Barrett M, Parlato J, Ross MA. Factors associated with prolonged observation services stays and the impact of long stays on patient cost. Health Serv Res. 2014;49(3):893-909. https://doi.org/10.1111/1475-6773.12143
31. Anthem BlueCross BlueShield. Ohio Provider Manual. Accessed April11, 2021. https://www11.anthem.com/provider/oh/f1/s0/t0/pw_g357368.pdf?refer=ahpprovider&state=oh
32. Humana. Provider manual for physicians, hospitals and healthcare providers. Accessed April 11, 2021. https://docushare-web.apps.cf.humana.com/Marketing/docushare-app?file=3932669
33. Fieldston ES, Shah SS, Hall M, et al. Resource utilization for observation-status stays at children’s hospitals. Pediatrics. 2013;131(6):1050-1058 https://doi.org/10.1542/peds.2012-249
34. Tejedor-Sojo J. Observation status-a name at what cost? Hosp Pediatr. 2014;4(5):321-323. https://doi.org/10.1542/hpeds.2014-0037.
35. Selden TM, Karaca Z, Keenan P, White C, Kronick R. The growing difference between public and private payment rates for inpatient hospital care. Health Aff (Millwood). 2015;34(12):2147-2150. https://doi.org/10.1377/hlthaff.2015.0706
36. Agency for Healthcare Research and Quality. AHRQ Quality Indicators. Accessed April 11, 2021. https://www.qualityindicators.ahrq.gov
37. Figueroa JF, Burke LG, Zheng J, Orav EJ, Jha AK. Trends in hospitalization vs observation stay for ambulatory care-sensitive conditions. JAMA Intern Med. 2019;179(12):1714-1716. https://doi.org/10.1001/jamainternmed.2019.3177
38. Markham JL, Hall M, Gay JC, Bettenhausen JL, Berry JG. Length of stay and cost of pediatric readmissions. Pediatrics. 2018;141(4):e20172934. https://doi.org/10.1542/peds.2017-2934.
39. Overman RA, Freburger JK, Assimon MM, Li X, Brookhart, MA. Observation stays in administrative claims databases: underestimation of hospitalized cases. Pharmacoepidemiol Drug Saf. 2014;23(9):902-910. https://doi.org/10.1002/pds.3647.
1. Centers for Medicare & Medicaid Services. Fact Sheet: Two-Midnight Rule. Accessed April 11, 2021. https://www.cms.gov/newsroom/fact-sheets/fact-sheet-two-midnight-rule-0
2. BlueCross BlueShield of Rhode Island. Payment Policy Outpaient Observation. Accessed April 11, 2021. https://www.bcbsri.com/sites/default/files/polices/Outpatient-Observation.pdf
3. Blue Cross Blue Shield of Illinois. Observation Services Tool for Applying MCG Care Guidelines Clinical Payment and Coding Policy. Accessed April 11, 2021. https://www.bcbsil.com/pdf/standards/observation_services_cpcp.pdf
4. Medicare.gov. Inpatient or outpatient hospital status affects your costs. Accessed April 11, 2021. https://www.medicare.gov/what-medicare-covers/what-part-a-covers/inpatient-or-outpatient-hospital-status
5. Ross MA, Hockenberry JM, Mutter R, Barrett M, Wheatley M, Pitts SR. Protocol-driven emergency department observation units offer savings, shorter stays, and reduced admissions. Health Aff (Millwood). 2013;32(12):2149-2156. https://doi.org/10.1377/hlthaff.2013.0662
6. Baugh CW, Venkatesh AK, Hilton JA, Samuel PA, Schuur JD, Bohan JS. Making greater use of dedicated hospital observation units for many short-stay patients could save $3.1 billion a year. Health Aff (Millwood). 2012;31(10):2314-2323. https://doi.org/10.1377/hlthaff.2011.0926
7. Sheehy AM, Graf B, Gangireddy S, et al. Hospitalized but not admitted: characteristics of patients with “observation status” at an academic medical center. JAMA Intern Med. 2013;173(21):1991-1998. https://doi.org/10.1001/jamainternmed.2013.8185
8. Baugh CW, Schuur JD. Observation care—high-value care or a cost-shifting loophole? N Engl J Med. 2013;369(4):302-305. https://doi.org/10.1056/NEJMp1304493
9. Missouri Hospital Association. A patient’s guide to observation care. Accessed April 11, 2021. https://www.mhanet.com/mhaimages/PatientsGuideToObservationCareFlyer.pdf
10. Cigna. Employee-paid hospital care coverage- summary of benefits. Accessed April 11, 2021. https://www.cigna.com/iwov-resources/national-second-sale/docs/healthy-benefits/updated-HC-benefit-summary.pdf
11. BlueCross BlueShield of Minnesota. Reimbursement policy-observation care services. Accessed April 11, 2021. https://www.bluecrossmn.com/sites/default/files/DAM/2020-07/Evaluation%20and%20Management%20004_Observation%20Care%20Services%20_09.04.17.pdf
12. California Department of Health Care Services. Public Hospital Project Frequently Asked Questions. Accessed April 11, 2021. https://www.dhcs.ca.gov/provgovpart/Documents/Public%20Hospital%20Project/PHP_Final_FAQs_January2013ADA.pdf
13. Texas Medicaid & Healthcare Partnership. Inpatient and Outpatient Hospital Servicces Handbook. Accessed May 29, 2021. https://www.tmhp.com/sites/default/files/microsites/provider-manuals/tmppm/html/TMPPM/2_Inpatient_Outpatient_Hosp_Srvs/2_Inpatient_Outpatient_Hosp_Srvs.htm
14. Alabama Medicaid. Outpatient observation. Accessed April 11, 2021. https://medicaid.alabama.gov/news_detail.aspx?ID=5121
15. NC Medicaid. Medicaid and Health Choice Clinical Coverage Policy No: 2A-1. Accessed April 11, 2021. https://files.nc.gov/ncdma/documents/files/2A-1_0.pdf
16. Feng Z, Wright B, Mor V. Sharp rise in Medicare enrollees being held in hospitals for observation raises concerns about causes and consequences. Health Aff (Millwood). 2012;31(6):1251-1259. https://doi.org/10.1377/hlthaff.2012.0129
17. Wright B, O’Shea AM, Ayyagari P, Ugwi PG, Kaboli P, Vaughan Sarrazin M. Observation rates at veterans’ hospitals more than doubled during 2005-13, similar to Medicare trends. Health Aff (Millwood). 2015;34(10):1730-1737. https://doi.org/10.1377/hlthaff.2014.1474
18. Wright B, Jung HY, Feng Z, Mor V. Hospital, patient, and local health system characteristics associated with the prevalence and duration of observation care. Health Serv Res. 2014;49(4):1088-1107. https://doi.org/10.1111/1475-6773.12166
19. Sabbatini AK, Wright B, Hall MK, Basu A. The cost of observation care for commercially insured patients visiting the emergency department. Am J Emerg Med. 2018;36(9):1591-1596. https://doi.org/10.1016/j.ajem.2018.01.040
20. Children’s Hospital Association. Pediatric health information system. Accessed April 11, 2021. https://www.childrenshospitals.org/phis
21. Richardson T, Rodean J, Harris M, Berry J, Gay JC, Hall M. Development of hospitalization resource intensity scores for kids (H-RISK) and comparison across pediatric populations. J Hosp Med. 2018;13(9):602-608. https://doi.org/10.12788/jhm.2948
22. Gay JC, Hall M, Morse R, Fieldston ES, Synhorst DC, Macy ML.Observation encounters and length of stay benchmarking in children’s hospitals. Pediatrics. 2020;146(5):e20200120. https://doi.org/10.1542/peds.2020-0120
23. Synhorst DC, Hall M, Harris M, et al. Hospital observation status and readmission rates. Pediatrics. 2020;146(5):e2020003954. https://doi.org/10.1542/peds.2020-003954
24. Macy ML, Hall M, Shah SS, et al. Pediatric observation status: are we overlooking a growing population in children’s hospitals? J Hosp Med. 2012;7(7):530-536. https://doi.org/10.1002/jhm.1923
25. Macy ML, Kim CS, Sasson C, Lozon MM, Davis MM. Pediatric observation units in the United States: a systematic review. J Hosp Med. 2010;5(3):172-182. https://doi.org/10.1002/jhm.592
26. UnitedHealthcare. Observation services policy, facility. Accessed April 11, 2021. https://www.uhcprovider.com/content/dam/provider/docs/public/policies/medicaid-comm-plan-reimbursement/UHCCP-Facility-Observation-Services-Policy-(F7106).pdf
27. Cal SB-1076§1253.7. General acute care hospitals: observation services – Health and Safety. Accessed April 11, 2021. https://leginfo.legislature.ca.gov/faces/billTextClient.xhtml?bill_id=201520160SB1076
28. Nebraska Total Care. 2021 Provider Billing Guide. Accessed April 11, 2021. https://www.nebraskatotalcare.com/content/dam/centene/Nebraska/PDFs/ProviderRelations/NTC_Nebraska_Total_Care_Provider_Billing_Guide_508.pdf
29. Macy ML, Hall M, Shah SS, et al. Differences in designations of observation care in US freestanding children’s hospitals: are they virtual or real? J Hosp Med. 2012;7(4):287-293. https://doi.org/10.1002/jhm.949
30. Hockenberry JM, Mutter R, Barrett M, Parlato J, Ross MA. Factors associated with prolonged observation services stays and the impact of long stays on patient cost. Health Serv Res. 2014;49(3):893-909. https://doi.org/10.1111/1475-6773.12143
31. Anthem BlueCross BlueShield. Ohio Provider Manual. Accessed April11, 2021. https://www11.anthem.com/provider/oh/f1/s0/t0/pw_g357368.pdf?refer=ahpprovider&state=oh
32. Humana. Provider manual for physicians, hospitals and healthcare providers. Accessed April 11, 2021. https://docushare-web.apps.cf.humana.com/Marketing/docushare-app?file=3932669
33. Fieldston ES, Shah SS, Hall M, et al. Resource utilization for observation-status stays at children’s hospitals. Pediatrics. 2013;131(6):1050-1058 https://doi.org/10.1542/peds.2012-249
34. Tejedor-Sojo J. Observation status-a name at what cost? Hosp Pediatr. 2014;4(5):321-323. https://doi.org/10.1542/hpeds.2014-0037.
35. Selden TM, Karaca Z, Keenan P, White C, Kronick R. The growing difference between public and private payment rates for inpatient hospital care. Health Aff (Millwood). 2015;34(12):2147-2150. https://doi.org/10.1377/hlthaff.2015.0706
36. Agency for Healthcare Research and Quality. AHRQ Quality Indicators. Accessed April 11, 2021. https://www.qualityindicators.ahrq.gov
37. Figueroa JF, Burke LG, Zheng J, Orav EJ, Jha AK. Trends in hospitalization vs observation stay for ambulatory care-sensitive conditions. JAMA Intern Med. 2019;179(12):1714-1716. https://doi.org/10.1001/jamainternmed.2019.3177
38. Markham JL, Hall M, Gay JC, Bettenhausen JL, Berry JG. Length of stay and cost of pediatric readmissions. Pediatrics. 2018;141(4):e20172934. https://doi.org/10.1542/peds.2017-2934.
39. Overman RA, Freburger JK, Assimon MM, Li X, Brookhart, MA. Observation stays in administrative claims databases: underestimation of hospitalized cases. Pharmacoepidemiol Drug Saf. 2014;23(9):902-910. https://doi.org/10.1002/pds.3647.
© 2021 Society of Hospital Medicine
Inpatient Glycemic Control With Sliding Scale Insulin in Noncritical Patients With Type 2 Diabetes: Who Can Slide?
Sliding scale insulin (SSI) for inpatient glycemic control was first proposed by Elliott P Joslin in 1934 when he recommended titration of insulin based on urine glucose levels.1 As bedside glucose meters became widely available, physicians transitioned to dosing SSI based on capillary blood glucose (BG) levels,2,3 and SSI became widely used for the management of inpatient hyperglycemia.1 However, during the past decade, there has been strong opposition to the use of SSI in hospitals. Many authors oppose its use, highlighting the retrospective rather than prospective nature of SSI therapy and concerns about inadequate glycemic control.4-6 In 2004, the American College of Endocrinology first released a position statement discouraging the use of SSI alone and recommended basal-bolus insulin as the preferred method of glycemic control for inpatients with type 2 diabetes (T2D).7 The American Diabetes Association (ADA) inpatient guidelines in 20058 and the Endocrine Society guidelines in 20129 also opposed SSI monotherapy and reaffirmed that a basal-bolus insulin regimen should be used for most non–critically ill patients with diabetes. Those guidelines remain in place currently.
Several randomized controlled trials (RCTs) and meta-analyses have shown that basal-bolus insulin regimens provide superior glycemic control in non–critical inpatients when compared with SSI alone.10-14 In addition, the RABBIT 2 (Randomized Study of Basal-Bolus Insulin Therapy in the Inpatient Management of Patients With Type 2 Diabetes) trial showed a significant reduction in perioperative complications10 among surgical patients when treated with basal-bolus insulin therapy. Despite these studies and strong recommendations against its use, SSI continues to be widely used in the United States. According to a 2007 survey of 44 US hospitals, 41% of noncritical patients with hyperglycemia were treated with SSI alone.15 In addition, SSI remains one of the most commonly prescribed insulin regimens in many countries around the world.16-19 The persistence of SSI use raises questions as to why clinicians continue to use a therapy that has been strongly criticized. Some authors point to convenience and fear of hypoglycemia with a basal-bolus insulin regimen.20,21 Alternatively, it is possible that SSI usage remains so pervasive because it is effective in a subset of patients. In fact, a 2018 Cochrane review concluded that existing evidence is not sufficiently robust to definitively recommend basal-bolus insulin over SSI for inpatient diabetes management of non–critically ill patients despite existing guidelines.22
Owing to the ongoing controversy and widespread use of SSI, we designed an exploratory analysis to understand the rationale for such therapy by investigating whether a certain subpopulation of hospitalized patients with T2D may achieve target glycemic control with SSI alone. We hypothesized that noncritical patients with mild hyperglycemia and admission BG <180 mg/dL would do well with SSI alone and may not require intensive treatment with basal-bolus insulin regimens. To address this question, we used electronic health records with individual-level patient data to assess inpatient glycemic control of non–critically ill patients with T2D treated with SSI alone.
METHODS
Participants
Data from 25,813 adult noncritical inpatients with T2D, with an index admission between June 1, 2010, and June 30, 2018, were obtained through the Emory Healthcare Clinical Data Warehouse infrastructure program. All patients were admitted to Emory Healthcare hospitals, including Emory University Hospital, Emory University Hospital Midtown, and Emory Saint Joseph’s Hospital, in Atlanta, Georgia. Data were extracted for each patient during the index hospitalization, including demographics, anthropometrics, and admission and inpatient laboratory values. Information was collected on daily point-of-care glucose values, hemoglobin A1c (HbA1c), hypoglycemic events, insulin doses, hospital complications, comorbidities, and hospital setting (medical vs surgical admission). International Classification of Diseases, 9th and 10th Revisions (ICD-9/10) codes were used to determine diagnosis of T2D, comorbidities, and complications.
From our initial dataset, we identified 16,366 patients who were treated with SSI during hospitalization. We excluded patients who were admitted to the intensive care unit (ICU) or placed on intravenous insulin, patients with missing admission BG values, and patients with a length of stay less than 1 day. To prevent inclusion of patients presenting in diabetic ketoacidosis or hyperosmolar hyperglycemic syndrome, we excluded patients with an admission BG >500 mg/dL. We then excluded 6,739 patients who received basal insulin within the first 2 days of hospitalization, as well as 943 patients who were treated with noninsulin (oral or injectable) antidiabetic agents. Our final dataset included 8,095 patients (Appendix Figure).
Patients in the SSI cohort included all patients who were treated with short-acting insulin only (regular insulin or rapid-acting [lispro, aspart, glulisine] insulin analogs) during the first 2 days of hospitalization. Patients who remained on only short-acting insulin during the entire hospitalization were defined as continuous SSI patients. Patients who subsequently received basal insulin after day 2 of hospitalization were defined as patients who transitioned to basal. Patients were stratified according to admission BG levels (first BG available on day of admission) and HbA1c (when available during index admission). We compared the baseline characteristics and clinical outcomes of patients who remained on SSI alone throughout the entirety of hospitalization with those of patients who required transition to basal insulin. The mean hospital BG was calculated by taking the average of all BG measurements during the hospital stay. We defined hypoglycemia as a BG <70 mg/dL and severe hypoglycemia as BG <40 mg/dL. Repeated hypoglycemia values were excluded if they occurred within a period of 2 hours.
Outcome Measures
The primary outcome was the percentage of patients with T2D achieving target glycemic control with SSI therapy, defined as mean hospital BG between 70 and 180 mg/dL without hypoglycemia <70 mg/dL during hospital stay. This threshold was determined based on 2019 ADA recommendations targeting hospital BG <180 mg/dL and avoidance of hypoglycemia.23
Statistical Analysis
Patients were stratified according to continuous SSI versus transitioned to basal treatment. Patients who remained on continuous SSI were further categorized into four categories based on admission BG: <140 mg/dL, 140 to 180 mg/dL, 180 to 250 mg/dL, and ≥250 mg/dL. Clinical characteristics were compared using Wilcoxon rank-sum tests (if continuous) and chi-square tests or Fisher exact tests (if categorical). We then compared the clinical outcomes among continuous SSI patients with different admission BG levels (<140 mg/dL, 140-180 mg/dL, 180-250 mg/dL, and ≥250 mg/dL) and with different HbA1c levels (<7%, 7%-8%, 8%-9%, ≥9%). Within each scenario, logistic regression for the outcome of poor glycemic control, defined as mean hospital BG >180 mg/dL, was performed to evaluate the HbA1c levels and admission BG levels controlling for other factors (age, gender, body mass index [BMI], race, setting [medicine versus surgery] and Charlson Comorbidity Index score). A P value < .05 was regarded as statistically significant. All analyses were performed based on available cases and conducted in SAS version 9.4 (SAS Institute Inc.).
RESULTS
Among 25,813 adult patients with T2D, 8,095 patients (31.4%) were treated with SSI alone during the first 2 days of hospitalization. Of those patients treated with SSI, 6,903 (85%) remained on continuous SSI alone during the entire hospitalization, and 1,192 (15%) were transitioned to basal insulin. The clinical characteristics of these patients on continuous SSI and those who transitioned to basal insulin are shown in Table 1. Patients who transitioned to basal insulin had significantly higher mean (SD) admission BG (191.8 [88.2] mg/dL vs 156.4 [65.4] mg/dL, P < .001) and higher mean (SD) HbA1c (8.1% [2.0%] vs 7.01% [1.5%], P < .001), compared with those who remained on continuous SSI. Patients who transitioned to basal insulin were also younger and more likely to have chronic kidney disease (CKD), but less likely to have congestive heart failure, coronary artery disease, or chronic obstructive pulmonary disease (COPD). The Charlson Comorbidity Index score was significantly higher for patients who transitioned to basal (4.4 [2.5]) than for those who remained on continuous SSI (4.1 [2.5], P < .001). There were no significant differences among sex, BMI, or glomerular filtration rate (GFR) on admission. Of those transitioned to basal insulin, 53% achieved a mean hospitalization BG <180 mg/dL, compared with 82% of those on continuous SSI. The overall rate of hypoglycemia in the continuous SSI group was 8% compared with 18% in those transitioned to basal insulin.
Of the patients who remained on continuous SSI throughout the hospitalization, 3,319 patients (48%) had admission BG <140 mg/dL, 1,671 patients (24%) had admission BG 140 to 180 mg/dL, and 1,913 patients (28%) had admission BG >180 mg/dL. Only 9% of patients who remained on continuous SSI had admission BG ≥250 mg/dL. Patients with admission BG <140 mg/dL were older, had lower BMI and HbA1c, had higher rates of COPD and CKD, and were more likely to be admitted to a surgical service compared with patients with admission BG >140 mg/dL (P < .05 for all; Table 2).
Hospital glycemic control for patients on continuous SSI according to admission BG is displayed in Table 3. Among patients who remained on continuous SSI, 96% of patients with admission BG <140 mg/dL had a mean hospital BG <180 mg/dL; of them, 86% achieved target control without hypoglycemia. Similar rates of target control were achieved in patients with admission BG 140 to 180 mg/dL (83%), in contrast to patients with admission BG ≥250 mg/dL, of whom only 18% achieved target control (P < .001). These findings parallel those seen in patients transitioned to basal insulin. Of patients in the transition group admitted with BG <140 mg/dL and <180 mg/dL, 88.5% and 84.6% had mean hospital BG <180 mg/dL, respectively, while 69.1% and 68.9% had mean BG between 70 and 180 mg/dL without hypoglycemia. The overall frequency of hypoglycemia <70 mg/dL among patients on continuous SSI was 8% and was more common in patients with admission BG <140 mg/dL (10%) compared with patients with higher admission glucose levels (BG 140-180 mg/dL [4%], 180-250 mg/dL [4%], or ≥250 mg/dL [6%], P < .001). There was no difference in rates of severe hypoglycemia <40 mg/dL among groups.
HbA1c data were available for 2,560 of the patients on continuous SSI (Table 3). Mean hospital BG increased significantly with increasing HbA1c values. Patients admitted with HbA1c <7% had lower mean (SD) hospital BG (132.2 [28.2] mg/dL) and were more likely to achieve target glucose control during hospitalization (85%) compared with those with HbA1c 7% to 8% (mean BG, 148.7 [30.8] mg/dL; 80% target control), HbA1c 8% to 9% (mean BG, 169.1 [37.9] mg/dL; 61% target control), or HbA1c ≥9% (mean BG, 194.9 [53.4] mg/dL; 38% target control) (P < .001).
In a logistic regression analysis adjusted for age, gender, BMI, race, setting (medicine vs surgery), and Charlson Comorbidity Index score, the odds of poor glycemic control increased with higher admission BG (admission BG 140-180 mg/dL: odds ratio [OR], 1.8; 95% CI, 1.5-2.2; admission BG 180-250 mg/dL: OR, 3.7; 95% CI, 3.1-4.4; admission BG ≥250 mg/dL: OR, 7.2; 95% CI, 5.8-9.0; reference admission BG <140 mg/dL; Figure). Similarly, the logistic regression analysis showed greater odds of poor in-hospital glycemic control with increasing HbA1c (OR, 6.1; 95% CI, 4.3-8.8 for HbA1c >9% compared with HbA1c <7%).
DISCUSSION
This large retrospective cohort study examined the effectiveness of SSI for glycemic control in noncritical inpatients with T2D. Our results indicate that SSI is still widely used in our hospital system, with 31.4% of our initial cohort managed with SSI alone. We found that 86% of patients with BG <140 mg/dL and 83% of patients with BG 140 to 180 mg/dL achieved glycemic control without hypoglycemia when managed with SSI alone, compared with 53% of those admitted with BG 180 to 250 mg/dL and only 18% of those with admission BG ≥250 mg/dL. This high success rate of achieving optimal BG control with SSI alone is comparable to that seen with transition to basal insulin and may explain the prevalent use of SSI for the management of patients with T2D and mild to moderate hyperglycemia.
Published clinical guideline recommendations promoting the use of basal-bolus insulin treatment algorithms are based on the results of a few RCTs that compared the efficacy of SSI vs a basal-bolus insulin regimen. These studies reported significantly lower mean daily BG concentration with basal or basal-bolus insulin therapy compared with SSI.10,11,24 However, it is interesting to note that the mean admission BG of patients treated with SSI in these RCTs ranged from 184 to 225 mg/dL. Patients in these trials were excluded if admission BG was <140 mg/dL.10,11,24 This is in contrast to our study evaluating real-world data in non–critically ill settings in which we found that 48% of patients treated with SSI had admission BG <140 mg/dL, and nearly 75% had admission BG <180 mg/dL. This suggests that by nature of study design, most RCTs excluded the population of patients who do achieve good glycemic control with SSI and may have contributed to the perception that basal insulin is preferable in all populations.
Our analysis indicates that healthcare professionals should consider admission BG when selecting the type of insulin regimen to manage patients with T2D in the hospital. Our results suggest that SSI may be appropriate for many patients with admission BG <180 mg/dL and should be avoided as monotherapy in patients with admission BG ≥180 mg/dL, as the proportion of patients achieving target control decreased with increasing admission BG. More importantly, if a patient is not controlled with SSI alone, intensification of therapy with the addition of basal insulin is indicated to achieve glycemic control. In addition, we found that the admission HbA1c is an appropriate marker to consider as well, with hospital glycemic control deteriorating with increasing HbA1c values, paralleling the admission BG. The main limitation to widespread use of HbA1c for therapeutic decision-making is access to values at time of patient admission; in our population, only 37% of patients had an HbA1c value available during the index hospitalization.
Previous publications have reported that hypoglycemia carries significant safety concerns, especially among a hospitalized population.25-27 As such, we included hypoglycemia as an important metric in our definition of target glycemic control rather than simply using mean hospital BG or number of hyperglycemic events to define treatment effectiveness. We did find a higher rate of hypoglycemia in patients with moderate admission BG treated with SSI compared with those with higher admission BG; however, few patients overall experienced clinically significant (<54 mg/dL) or severe (<40 mg/dL) hypoglycemia.
In our population, only 15% of patients started on SSI received additional basal insulin during hospitalization. This finding is similar to data reported in the Rabbit 2 trial, in which 14% of patients failed SSI alone, with a higher failure rate among those with higher BG on admission.10 Given the observational nature of this study, we cannot definitively state why certain patients in our population required additional basal insulin, but we can hypothesize that these patients admitted with BG ≥180 mg/dL had higher treatment failure rates and greater rates of hyperglycemia, therefore receiving intensified insulin therapy as clinically indicated at the discretion of the treating physician. Patients who transitioned from SSI to basal insulin had significantly higher admission BG and HbA1c compared with patients who remained on SSI alone. We noted that the rates of hypoglycemia were higher in the group that transitioned to basal (18% vs 8%) and similar to rates reported in previous RCTs.11,24
This observational study takes advantage of a large, diverse study population and a combination of medicine and surgery patients in a real-world setting. We acknowledge several limitations in our study. Our primary data were observational in nature, and as such, some baseline patient characteristics were notably different between groups, suggesting selection bias for treatment allocation to SSI. We do not know which patients were managed by primary teams compared with specialized diabetes consult services, which may also influence treatment regimens. We did not have access to information about patients’ at-home diabetes medication regimens or duration of diabetes, both of which have been shown in prior publications to affect an individual’s overall hospital glycemic control. Data on HbA1c values were available for only approximately one-third of patients. In addition, our study did not include patients without a history of diabetes who developed stress-induced hyperglycemia, a population that may benefit from conservative therapy such as SSI.28 A diagnosis of CKD was defined based on ICD 9/10 codes and not on admission estimated GFR. More specific data regarding stage of CKD or changes in renal function over the duration of hospitalization are not available, which could influence insulin prescribing practice. In addition, we defined the basal group as patients prescribed any form of basal insulin (NPH, glargine, detemir or degludec), and we do not have information on the use of prandial versus correction doses of rapid-acting insulin in the basal insulin–treated group.
CONCLUSION
In conclusion, our observational study indicates that the use of SSI results in appropriate target glycemic control for most noncritical medicine and surgery patients with admission BG <180 mg/dL. In agreement with previous RCTs, our study confirms that SSI as monotherapy is frequently inadequate in patients with significant hyperglycemia >180 mg/dL.10,11,24,29 We propose that an individualized approach to inpatient glycemic management is imperative, and cautious use of SSI may be a viable option for certain patients with mild hyperglycemia and admission BG <180 mg/dL. Further observational and randomized studies are needed to confirm the efficacy of SSI therapy in T2D patients with mild hyperglycemia. By identifying which subset of patients can be safely managed with SSI alone, we can better understand which patients will require escalation of therapy with intensive glucose management.
1. Umpierrez GE, Palacio A, Smiley D. Sliding scale insulin use: myth or insanity? Am J Med. 2007;120(7):563-567. https://doi.org/10.1016/j.amjmed.2006.05.070
2. Kitabchi AE, Ayyagari V, Guerra SM. The efficacy of low-dose versus conventional therapy of insulin for treatment of diabetic ketoacidosis. Ann Intern Med. 1976;84(6):633-638. https://doi.org/10.7326/0003-4819-84-6-633
3. Skyler JS, Skyler DL, Seigler DE, O’Sullivan MJ. Algorithms for adjustment of insulin dosage by patients who monitor blood glucose. Diabetes Care. 1981;4(2):311-318. https://doi.org/10.2337/diacare.4.2.311
4. Gearhart JG, Duncan JL 3rd, Replogle WH, Forbes RC, Walley EJ. Efficacy of sliding-scale insulin therapy: a comparison with prospective regimens. Fam Pract Res J. 1994;14(4):313-322.
5. Queale WS, Seidler AJ, Brancati FL. Glycemic control and sliding scale insulin use in medical inpatients with diabetes mellitus. Arch Intern Med. 1997;157(5):545-552.
6. Clement S, Braithwaite SS, Magee MF, et al. Management of diabetes and hyperglycemia in hospitals. Diabetes Care. 2004;27(2):553-591. https://doi.org/10.2337/diacare.27.2.553
7. Garber AJ, Moghissi ES, Bransome ED Jr, et al. American College of Endocrinology position statement on inpatient diabetes and metabolic control. Endocr Pract. 2004;10(1):78-82. https://doi.org/10.4158/EP.10.1.77
8. American Diabetes Association. Standards of medical care in diabetes. Diabetes Care. 2005;28(suppl 1):S4-S36.
9. Umpierrez GE, Hellman R, Korytkowski MT, , et al. Management of hyperglycemia in hospitalized patients in non-critical care setting: an Endocrine Society clinical practice guideline. J Clin Endocrinol Metab. 2012;97(1):16-38. https://doi.org/10.1210/jc.2011-2098
10. Umpierrez GE, Smiley D, Zisman A, et al. Randomized study of basal-bolus insulin therapy in the inpatient management of patients with type 2 diabetes. Diabetes Care. 2007;30(9):2181-2186. https://doi.org/10.2337/dc07-0295
11. Umpierrez GE, Smiley D, Jacobs S, et al. Randomized study of basal-bolus insulin therapy in the inpatient management of patients with type 2 diabetes undergoing general surgery (RABBIT 2 surgery). Diabetes Care. 2011;34(2):256-261. https://doi.org/10.2337/dc10-1407
12. Schroeder JE, Liebergall M, Raz I, Egleston R, Ben Sussan G, Peyser A. Benefits of a simple glycaemic protocol in an orthopaedic surgery ward: a randomized prospective study. Diabetes Metab Res Rev. 2012;28:71-75. https://doi.org/10.1002/dmrr.1217
13. Lee YY, Lin YM, Leu WJ, et al. Sliding-scale insulin used for blood glucose control: a meta-analysis of randomized controlled trials. Metabolism. 2015;64(9):1183-1192. https://doi.org/10.1016/j.metabol.2015.05.011
14. Christensen MB, Gotfredsen A, Nørgaard K. Efficacy of basal-bolus insulin regimens in the inpatient management of non-critically ill patients with type 2 diabetes: a systematic review and meta-analysis. Diabetes Metab Res Rev. 2017;33(5):e2885. https://doi.org/10.1002/dmrr.2885
15. Wexler DJ, Meigs JB, Cagliero E, Nathan DM, Grant RW. Prevalence of hyper- and hypoglycemia among inpatients with diabetes: a national survey of 44 U.S. hospitals. Diabetes Care. 2007;30(2):367-369. https://doi.org/10.2337/dc06-1715
16. Moreira ED Jr, Silveira PCB, Neves RCS, Souza C Jr, Nunes ZO, Almeida MdCC. Glycemic control and diabetes management in hospitalized patients in Brazil. Diabetol Metab Syndr. 2013;5(1):62. https://doi.org/10.1186/1758-5996-5-62
17. Akhtar ST, Mahmood K, Naqvi IH, Vaswani AS. Inpatient management of type 2 diabetes mellitus: does choice of insulin regimen really matter? Pakistan J Med Sci. 2014;30(4):895-898.
18. Gómez Cuervo C, Sánchez Morla A, Pérez-Jacoiste Asín MA, Bisbal Pardo O, Pérez Ordoño L, Vila Santos J. Effective adverse event reduction with bolus-basal versus sliding scale insulin therapy in patients with diabetes during conventional hospitalization: systematic review and meta-analysis. Endocrinol Nutr. 2016;63(4):145-156. https://doi.org/10.1016/j.endonu.2015.11.008
19. Bain A, Hasan SS, Babar ZUD. Interventions to improve insulin prescribing practice for people with diabetes in hospital: a systematic review. Diabet Med. 2019;36(8):948-960. https://doi.org/10.1111/dme.13982
20. Ambrus DB, O’Connor MJ. Things We Do For No Reason: sliding-scale insulin as monotherapy for glycemic control in hospitalized patients. J Hosp Med. 2019;14(2):114-116. https://doi.org/10.12788/jhm.3109
21. Nau KC, Lorenzetti RC, Cucuzzella M, Devine T, Kline J. Glycemic control in hospitalized patients not in intensive care: beyond sliding-scale insulin. Am Fam Physician. 2010;81(9):1130-1135.
22. Colunga-Lozano LE, Gonzalez Torres FJ, Delgado-Figueroa N, et al. Sliding scale insulin for non-critically ill hospitalised adults with diabetes mellitus. Cochrane Database Syst Rev. 2018;11(11):CD011296. https://doi.org/10.1002/14651858.CD011296.pub2
23. American Diabetes Association. Diabetes care in the hospital: Standards of Medical Care in Diabetes—2019. Diabetes Care. 2019;42(suppl 1):S173-S181. https://doi.org/10.2337/dc19-S015
24. Umpierrez GE, Smiley D, Hermayer K, et al. Randomized study comparing a basal-bolus with a basal plus correction management of medical and surgical patients with type 2 diabetes: basal plus trial. Diabetes Care. 2013;36(8):2169-2174. https://doi.org/10.2337/dc12-1988
25. Turchin A, Matheny ME, Shubina M, Scanlon SV, Greenwood B, Pendergrass ML. Hypoglycemia and clinical outcomes in patients with diabetes hospitalized in the general ward. Diabetes Care. 2009;32(7):1153-1157. https://doi.org/10.2337/dc08-2127
26. Garg R, Hurwitz S, Turchin A, Trivedi A. Hypoglycemia, with or without insulin therapy, is associated with increased mortality among hospitalized patients. Diabetes Care. 2013;36(5):1107-1110. https://doi.org/10.2337/dc12-1296
27. Zapatero A, Gómez-Huelgas R, González N, et al. Frequency of hypoglycemia and its impact on length of stay, mortality, and short-term readmission in patients with diabetes hospitalized in internal medicine wards. Endocr Pract. 2014;20(9):870-875. https://doi.org/10.4158/EP14006.OR
28. Umpierrez GE, Isaacs SD, Bazargan N, You X, Thaler LM, Kitabchi AE. Hyperglycemia: an independent marker of in-hospital mortality in patients with undiagnosed diabetes. J Clin Endocrinol Metab. 2002;87(3):978-982. https://doi.org/10.1210/jcem.87.3.8341
29. Dickerson LM, Ye X, Sack JL, Hueston WJ. Glycemic control in medical inpatients with type 2 diabetes mellitus receiving sliding scale insulin regimens versus routine diabetes medications: a multicenter randomized controlled trial. Ann Fam Med. 2003;1(1):29-35. https://doi.org/10.1370/afm.2
Sliding scale insulin (SSI) for inpatient glycemic control was first proposed by Elliott P Joslin in 1934 when he recommended titration of insulin based on urine glucose levels.1 As bedside glucose meters became widely available, physicians transitioned to dosing SSI based on capillary blood glucose (BG) levels,2,3 and SSI became widely used for the management of inpatient hyperglycemia.1 However, during the past decade, there has been strong opposition to the use of SSI in hospitals. Many authors oppose its use, highlighting the retrospective rather than prospective nature of SSI therapy and concerns about inadequate glycemic control.4-6 In 2004, the American College of Endocrinology first released a position statement discouraging the use of SSI alone and recommended basal-bolus insulin as the preferred method of glycemic control for inpatients with type 2 diabetes (T2D).7 The American Diabetes Association (ADA) inpatient guidelines in 20058 and the Endocrine Society guidelines in 20129 also opposed SSI monotherapy and reaffirmed that a basal-bolus insulin regimen should be used for most non–critically ill patients with diabetes. Those guidelines remain in place currently.
Several randomized controlled trials (RCTs) and meta-analyses have shown that basal-bolus insulin regimens provide superior glycemic control in non–critical inpatients when compared with SSI alone.10-14 In addition, the RABBIT 2 (Randomized Study of Basal-Bolus Insulin Therapy in the Inpatient Management of Patients With Type 2 Diabetes) trial showed a significant reduction in perioperative complications10 among surgical patients when treated with basal-bolus insulin therapy. Despite these studies and strong recommendations against its use, SSI continues to be widely used in the United States. According to a 2007 survey of 44 US hospitals, 41% of noncritical patients with hyperglycemia were treated with SSI alone.15 In addition, SSI remains one of the most commonly prescribed insulin regimens in many countries around the world.16-19 The persistence of SSI use raises questions as to why clinicians continue to use a therapy that has been strongly criticized. Some authors point to convenience and fear of hypoglycemia with a basal-bolus insulin regimen.20,21 Alternatively, it is possible that SSI usage remains so pervasive because it is effective in a subset of patients. In fact, a 2018 Cochrane review concluded that existing evidence is not sufficiently robust to definitively recommend basal-bolus insulin over SSI for inpatient diabetes management of non–critically ill patients despite existing guidelines.22
Owing to the ongoing controversy and widespread use of SSI, we designed an exploratory analysis to understand the rationale for such therapy by investigating whether a certain subpopulation of hospitalized patients with T2D may achieve target glycemic control with SSI alone. We hypothesized that noncritical patients with mild hyperglycemia and admission BG <180 mg/dL would do well with SSI alone and may not require intensive treatment with basal-bolus insulin regimens. To address this question, we used electronic health records with individual-level patient data to assess inpatient glycemic control of non–critically ill patients with T2D treated with SSI alone.
METHODS
Participants
Data from 25,813 adult noncritical inpatients with T2D, with an index admission between June 1, 2010, and June 30, 2018, were obtained through the Emory Healthcare Clinical Data Warehouse infrastructure program. All patients were admitted to Emory Healthcare hospitals, including Emory University Hospital, Emory University Hospital Midtown, and Emory Saint Joseph’s Hospital, in Atlanta, Georgia. Data were extracted for each patient during the index hospitalization, including demographics, anthropometrics, and admission and inpatient laboratory values. Information was collected on daily point-of-care glucose values, hemoglobin A1c (HbA1c), hypoglycemic events, insulin doses, hospital complications, comorbidities, and hospital setting (medical vs surgical admission). International Classification of Diseases, 9th and 10th Revisions (ICD-9/10) codes were used to determine diagnosis of T2D, comorbidities, and complications.
From our initial dataset, we identified 16,366 patients who were treated with SSI during hospitalization. We excluded patients who were admitted to the intensive care unit (ICU) or placed on intravenous insulin, patients with missing admission BG values, and patients with a length of stay less than 1 day. To prevent inclusion of patients presenting in diabetic ketoacidosis or hyperosmolar hyperglycemic syndrome, we excluded patients with an admission BG >500 mg/dL. We then excluded 6,739 patients who received basal insulin within the first 2 days of hospitalization, as well as 943 patients who were treated with noninsulin (oral or injectable) antidiabetic agents. Our final dataset included 8,095 patients (Appendix Figure).
Patients in the SSI cohort included all patients who were treated with short-acting insulin only (regular insulin or rapid-acting [lispro, aspart, glulisine] insulin analogs) during the first 2 days of hospitalization. Patients who remained on only short-acting insulin during the entire hospitalization were defined as continuous SSI patients. Patients who subsequently received basal insulin after day 2 of hospitalization were defined as patients who transitioned to basal. Patients were stratified according to admission BG levels (first BG available on day of admission) and HbA1c (when available during index admission). We compared the baseline characteristics and clinical outcomes of patients who remained on SSI alone throughout the entirety of hospitalization with those of patients who required transition to basal insulin. The mean hospital BG was calculated by taking the average of all BG measurements during the hospital stay. We defined hypoglycemia as a BG <70 mg/dL and severe hypoglycemia as BG <40 mg/dL. Repeated hypoglycemia values were excluded if they occurred within a period of 2 hours.
Outcome Measures
The primary outcome was the percentage of patients with T2D achieving target glycemic control with SSI therapy, defined as mean hospital BG between 70 and 180 mg/dL without hypoglycemia <70 mg/dL during hospital stay. This threshold was determined based on 2019 ADA recommendations targeting hospital BG <180 mg/dL and avoidance of hypoglycemia.23
Statistical Analysis
Patients were stratified according to continuous SSI versus transitioned to basal treatment. Patients who remained on continuous SSI were further categorized into four categories based on admission BG: <140 mg/dL, 140 to 180 mg/dL, 180 to 250 mg/dL, and ≥250 mg/dL. Clinical characteristics were compared using Wilcoxon rank-sum tests (if continuous) and chi-square tests or Fisher exact tests (if categorical). We then compared the clinical outcomes among continuous SSI patients with different admission BG levels (<140 mg/dL, 140-180 mg/dL, 180-250 mg/dL, and ≥250 mg/dL) and with different HbA1c levels (<7%, 7%-8%, 8%-9%, ≥9%). Within each scenario, logistic regression for the outcome of poor glycemic control, defined as mean hospital BG >180 mg/dL, was performed to evaluate the HbA1c levels and admission BG levels controlling for other factors (age, gender, body mass index [BMI], race, setting [medicine versus surgery] and Charlson Comorbidity Index score). A P value < .05 was regarded as statistically significant. All analyses were performed based on available cases and conducted in SAS version 9.4 (SAS Institute Inc.).
RESULTS
Among 25,813 adult patients with T2D, 8,095 patients (31.4%) were treated with SSI alone during the first 2 days of hospitalization. Of those patients treated with SSI, 6,903 (85%) remained on continuous SSI alone during the entire hospitalization, and 1,192 (15%) were transitioned to basal insulin. The clinical characteristics of these patients on continuous SSI and those who transitioned to basal insulin are shown in Table 1. Patients who transitioned to basal insulin had significantly higher mean (SD) admission BG (191.8 [88.2] mg/dL vs 156.4 [65.4] mg/dL, P < .001) and higher mean (SD) HbA1c (8.1% [2.0%] vs 7.01% [1.5%], P < .001), compared with those who remained on continuous SSI. Patients who transitioned to basal insulin were also younger and more likely to have chronic kidney disease (CKD), but less likely to have congestive heart failure, coronary artery disease, or chronic obstructive pulmonary disease (COPD). The Charlson Comorbidity Index score was significantly higher for patients who transitioned to basal (4.4 [2.5]) than for those who remained on continuous SSI (4.1 [2.5], P < .001). There were no significant differences among sex, BMI, or glomerular filtration rate (GFR) on admission. Of those transitioned to basal insulin, 53% achieved a mean hospitalization BG <180 mg/dL, compared with 82% of those on continuous SSI. The overall rate of hypoglycemia in the continuous SSI group was 8% compared with 18% in those transitioned to basal insulin.
Of the patients who remained on continuous SSI throughout the hospitalization, 3,319 patients (48%) had admission BG <140 mg/dL, 1,671 patients (24%) had admission BG 140 to 180 mg/dL, and 1,913 patients (28%) had admission BG >180 mg/dL. Only 9% of patients who remained on continuous SSI had admission BG ≥250 mg/dL. Patients with admission BG <140 mg/dL were older, had lower BMI and HbA1c, had higher rates of COPD and CKD, and were more likely to be admitted to a surgical service compared with patients with admission BG >140 mg/dL (P < .05 for all; Table 2).
Hospital glycemic control for patients on continuous SSI according to admission BG is displayed in Table 3. Among patients who remained on continuous SSI, 96% of patients with admission BG <140 mg/dL had a mean hospital BG <180 mg/dL; of them, 86% achieved target control without hypoglycemia. Similar rates of target control were achieved in patients with admission BG 140 to 180 mg/dL (83%), in contrast to patients with admission BG ≥250 mg/dL, of whom only 18% achieved target control (P < .001). These findings parallel those seen in patients transitioned to basal insulin. Of patients in the transition group admitted with BG <140 mg/dL and <180 mg/dL, 88.5% and 84.6% had mean hospital BG <180 mg/dL, respectively, while 69.1% and 68.9% had mean BG between 70 and 180 mg/dL without hypoglycemia. The overall frequency of hypoglycemia <70 mg/dL among patients on continuous SSI was 8% and was more common in patients with admission BG <140 mg/dL (10%) compared with patients with higher admission glucose levels (BG 140-180 mg/dL [4%], 180-250 mg/dL [4%], or ≥250 mg/dL [6%], P < .001). There was no difference in rates of severe hypoglycemia <40 mg/dL among groups.
HbA1c data were available for 2,560 of the patients on continuous SSI (Table 3). Mean hospital BG increased significantly with increasing HbA1c values. Patients admitted with HbA1c <7% had lower mean (SD) hospital BG (132.2 [28.2] mg/dL) and were more likely to achieve target glucose control during hospitalization (85%) compared with those with HbA1c 7% to 8% (mean BG, 148.7 [30.8] mg/dL; 80% target control), HbA1c 8% to 9% (mean BG, 169.1 [37.9] mg/dL; 61% target control), or HbA1c ≥9% (mean BG, 194.9 [53.4] mg/dL; 38% target control) (P < .001).
In a logistic regression analysis adjusted for age, gender, BMI, race, setting (medicine vs surgery), and Charlson Comorbidity Index score, the odds of poor glycemic control increased with higher admission BG (admission BG 140-180 mg/dL: odds ratio [OR], 1.8; 95% CI, 1.5-2.2; admission BG 180-250 mg/dL: OR, 3.7; 95% CI, 3.1-4.4; admission BG ≥250 mg/dL: OR, 7.2; 95% CI, 5.8-9.0; reference admission BG <140 mg/dL; Figure). Similarly, the logistic regression analysis showed greater odds of poor in-hospital glycemic control with increasing HbA1c (OR, 6.1; 95% CI, 4.3-8.8 for HbA1c >9% compared with HbA1c <7%).
DISCUSSION
This large retrospective cohort study examined the effectiveness of SSI for glycemic control in noncritical inpatients with T2D. Our results indicate that SSI is still widely used in our hospital system, with 31.4% of our initial cohort managed with SSI alone. We found that 86% of patients with BG <140 mg/dL and 83% of patients with BG 140 to 180 mg/dL achieved glycemic control without hypoglycemia when managed with SSI alone, compared with 53% of those admitted with BG 180 to 250 mg/dL and only 18% of those with admission BG ≥250 mg/dL. This high success rate of achieving optimal BG control with SSI alone is comparable to that seen with transition to basal insulin and may explain the prevalent use of SSI for the management of patients with T2D and mild to moderate hyperglycemia.
Published clinical guideline recommendations promoting the use of basal-bolus insulin treatment algorithms are based on the results of a few RCTs that compared the efficacy of SSI vs a basal-bolus insulin regimen. These studies reported significantly lower mean daily BG concentration with basal or basal-bolus insulin therapy compared with SSI.10,11,24 However, it is interesting to note that the mean admission BG of patients treated with SSI in these RCTs ranged from 184 to 225 mg/dL. Patients in these trials were excluded if admission BG was <140 mg/dL.10,11,24 This is in contrast to our study evaluating real-world data in non–critically ill settings in which we found that 48% of patients treated with SSI had admission BG <140 mg/dL, and nearly 75% had admission BG <180 mg/dL. This suggests that by nature of study design, most RCTs excluded the population of patients who do achieve good glycemic control with SSI and may have contributed to the perception that basal insulin is preferable in all populations.
Our analysis indicates that healthcare professionals should consider admission BG when selecting the type of insulin regimen to manage patients with T2D in the hospital. Our results suggest that SSI may be appropriate for many patients with admission BG <180 mg/dL and should be avoided as monotherapy in patients with admission BG ≥180 mg/dL, as the proportion of patients achieving target control decreased with increasing admission BG. More importantly, if a patient is not controlled with SSI alone, intensification of therapy with the addition of basal insulin is indicated to achieve glycemic control. In addition, we found that the admission HbA1c is an appropriate marker to consider as well, with hospital glycemic control deteriorating with increasing HbA1c values, paralleling the admission BG. The main limitation to widespread use of HbA1c for therapeutic decision-making is access to values at time of patient admission; in our population, only 37% of patients had an HbA1c value available during the index hospitalization.
Previous publications have reported that hypoglycemia carries significant safety concerns, especially among a hospitalized population.25-27 As such, we included hypoglycemia as an important metric in our definition of target glycemic control rather than simply using mean hospital BG or number of hyperglycemic events to define treatment effectiveness. We did find a higher rate of hypoglycemia in patients with moderate admission BG treated with SSI compared with those with higher admission BG; however, few patients overall experienced clinically significant (<54 mg/dL) or severe (<40 mg/dL) hypoglycemia.
In our population, only 15% of patients started on SSI received additional basal insulin during hospitalization. This finding is similar to data reported in the Rabbit 2 trial, in which 14% of patients failed SSI alone, with a higher failure rate among those with higher BG on admission.10 Given the observational nature of this study, we cannot definitively state why certain patients in our population required additional basal insulin, but we can hypothesize that these patients admitted with BG ≥180 mg/dL had higher treatment failure rates and greater rates of hyperglycemia, therefore receiving intensified insulin therapy as clinically indicated at the discretion of the treating physician. Patients who transitioned from SSI to basal insulin had significantly higher admission BG and HbA1c compared with patients who remained on SSI alone. We noted that the rates of hypoglycemia were higher in the group that transitioned to basal (18% vs 8%) and similar to rates reported in previous RCTs.11,24
This observational study takes advantage of a large, diverse study population and a combination of medicine and surgery patients in a real-world setting. We acknowledge several limitations in our study. Our primary data were observational in nature, and as such, some baseline patient characteristics were notably different between groups, suggesting selection bias for treatment allocation to SSI. We do not know which patients were managed by primary teams compared with specialized diabetes consult services, which may also influence treatment regimens. We did not have access to information about patients’ at-home diabetes medication regimens or duration of diabetes, both of which have been shown in prior publications to affect an individual’s overall hospital glycemic control. Data on HbA1c values were available for only approximately one-third of patients. In addition, our study did not include patients without a history of diabetes who developed stress-induced hyperglycemia, a population that may benefit from conservative therapy such as SSI.28 A diagnosis of CKD was defined based on ICD 9/10 codes and not on admission estimated GFR. More specific data regarding stage of CKD or changes in renal function over the duration of hospitalization are not available, which could influence insulin prescribing practice. In addition, we defined the basal group as patients prescribed any form of basal insulin (NPH, glargine, detemir or degludec), and we do not have information on the use of prandial versus correction doses of rapid-acting insulin in the basal insulin–treated group.
CONCLUSION
In conclusion, our observational study indicates that the use of SSI results in appropriate target glycemic control for most noncritical medicine and surgery patients with admission BG <180 mg/dL. In agreement with previous RCTs, our study confirms that SSI as monotherapy is frequently inadequate in patients with significant hyperglycemia >180 mg/dL.10,11,24,29 We propose that an individualized approach to inpatient glycemic management is imperative, and cautious use of SSI may be a viable option for certain patients with mild hyperglycemia and admission BG <180 mg/dL. Further observational and randomized studies are needed to confirm the efficacy of SSI therapy in T2D patients with mild hyperglycemia. By identifying which subset of patients can be safely managed with SSI alone, we can better understand which patients will require escalation of therapy with intensive glucose management.
Sliding scale insulin (SSI) for inpatient glycemic control was first proposed by Elliott P Joslin in 1934 when he recommended titration of insulin based on urine glucose levels.1 As bedside glucose meters became widely available, physicians transitioned to dosing SSI based on capillary blood glucose (BG) levels,2,3 and SSI became widely used for the management of inpatient hyperglycemia.1 However, during the past decade, there has been strong opposition to the use of SSI in hospitals. Many authors oppose its use, highlighting the retrospective rather than prospective nature of SSI therapy and concerns about inadequate glycemic control.4-6 In 2004, the American College of Endocrinology first released a position statement discouraging the use of SSI alone and recommended basal-bolus insulin as the preferred method of glycemic control for inpatients with type 2 diabetes (T2D).7 The American Diabetes Association (ADA) inpatient guidelines in 20058 and the Endocrine Society guidelines in 20129 also opposed SSI monotherapy and reaffirmed that a basal-bolus insulin regimen should be used for most non–critically ill patients with diabetes. Those guidelines remain in place currently.
Several randomized controlled trials (RCTs) and meta-analyses have shown that basal-bolus insulin regimens provide superior glycemic control in non–critical inpatients when compared with SSI alone.10-14 In addition, the RABBIT 2 (Randomized Study of Basal-Bolus Insulin Therapy in the Inpatient Management of Patients With Type 2 Diabetes) trial showed a significant reduction in perioperative complications10 among surgical patients when treated with basal-bolus insulin therapy. Despite these studies and strong recommendations against its use, SSI continues to be widely used in the United States. According to a 2007 survey of 44 US hospitals, 41% of noncritical patients with hyperglycemia were treated with SSI alone.15 In addition, SSI remains one of the most commonly prescribed insulin regimens in many countries around the world.16-19 The persistence of SSI use raises questions as to why clinicians continue to use a therapy that has been strongly criticized. Some authors point to convenience and fear of hypoglycemia with a basal-bolus insulin regimen.20,21 Alternatively, it is possible that SSI usage remains so pervasive because it is effective in a subset of patients. In fact, a 2018 Cochrane review concluded that existing evidence is not sufficiently robust to definitively recommend basal-bolus insulin over SSI for inpatient diabetes management of non–critically ill patients despite existing guidelines.22
Owing to the ongoing controversy and widespread use of SSI, we designed an exploratory analysis to understand the rationale for such therapy by investigating whether a certain subpopulation of hospitalized patients with T2D may achieve target glycemic control with SSI alone. We hypothesized that noncritical patients with mild hyperglycemia and admission BG <180 mg/dL would do well with SSI alone and may not require intensive treatment with basal-bolus insulin regimens. To address this question, we used electronic health records with individual-level patient data to assess inpatient glycemic control of non–critically ill patients with T2D treated with SSI alone.
METHODS
Participants
Data from 25,813 adult noncritical inpatients with T2D, with an index admission between June 1, 2010, and June 30, 2018, were obtained through the Emory Healthcare Clinical Data Warehouse infrastructure program. All patients were admitted to Emory Healthcare hospitals, including Emory University Hospital, Emory University Hospital Midtown, and Emory Saint Joseph’s Hospital, in Atlanta, Georgia. Data were extracted for each patient during the index hospitalization, including demographics, anthropometrics, and admission and inpatient laboratory values. Information was collected on daily point-of-care glucose values, hemoglobin A1c (HbA1c), hypoglycemic events, insulin doses, hospital complications, comorbidities, and hospital setting (medical vs surgical admission). International Classification of Diseases, 9th and 10th Revisions (ICD-9/10) codes were used to determine diagnosis of T2D, comorbidities, and complications.
From our initial dataset, we identified 16,366 patients who were treated with SSI during hospitalization. We excluded patients who were admitted to the intensive care unit (ICU) or placed on intravenous insulin, patients with missing admission BG values, and patients with a length of stay less than 1 day. To prevent inclusion of patients presenting in diabetic ketoacidosis or hyperosmolar hyperglycemic syndrome, we excluded patients with an admission BG >500 mg/dL. We then excluded 6,739 patients who received basal insulin within the first 2 days of hospitalization, as well as 943 patients who were treated with noninsulin (oral or injectable) antidiabetic agents. Our final dataset included 8,095 patients (Appendix Figure).
Patients in the SSI cohort included all patients who were treated with short-acting insulin only (regular insulin or rapid-acting [lispro, aspart, glulisine] insulin analogs) during the first 2 days of hospitalization. Patients who remained on only short-acting insulin during the entire hospitalization were defined as continuous SSI patients. Patients who subsequently received basal insulin after day 2 of hospitalization were defined as patients who transitioned to basal. Patients were stratified according to admission BG levels (first BG available on day of admission) and HbA1c (when available during index admission). We compared the baseline characteristics and clinical outcomes of patients who remained on SSI alone throughout the entirety of hospitalization with those of patients who required transition to basal insulin. The mean hospital BG was calculated by taking the average of all BG measurements during the hospital stay. We defined hypoglycemia as a BG <70 mg/dL and severe hypoglycemia as BG <40 mg/dL. Repeated hypoglycemia values were excluded if they occurred within a period of 2 hours.
Outcome Measures
The primary outcome was the percentage of patients with T2D achieving target glycemic control with SSI therapy, defined as mean hospital BG between 70 and 180 mg/dL without hypoglycemia <70 mg/dL during hospital stay. This threshold was determined based on 2019 ADA recommendations targeting hospital BG <180 mg/dL and avoidance of hypoglycemia.23
Statistical Analysis
Patients were stratified according to continuous SSI versus transitioned to basal treatment. Patients who remained on continuous SSI were further categorized into four categories based on admission BG: <140 mg/dL, 140 to 180 mg/dL, 180 to 250 mg/dL, and ≥250 mg/dL. Clinical characteristics were compared using Wilcoxon rank-sum tests (if continuous) and chi-square tests or Fisher exact tests (if categorical). We then compared the clinical outcomes among continuous SSI patients with different admission BG levels (<140 mg/dL, 140-180 mg/dL, 180-250 mg/dL, and ≥250 mg/dL) and with different HbA1c levels (<7%, 7%-8%, 8%-9%, ≥9%). Within each scenario, logistic regression for the outcome of poor glycemic control, defined as mean hospital BG >180 mg/dL, was performed to evaluate the HbA1c levels and admission BG levels controlling for other factors (age, gender, body mass index [BMI], race, setting [medicine versus surgery] and Charlson Comorbidity Index score). A P value < .05 was regarded as statistically significant. All analyses were performed based on available cases and conducted in SAS version 9.4 (SAS Institute Inc.).
RESULTS
Among 25,813 adult patients with T2D, 8,095 patients (31.4%) were treated with SSI alone during the first 2 days of hospitalization. Of those patients treated with SSI, 6,903 (85%) remained on continuous SSI alone during the entire hospitalization, and 1,192 (15%) were transitioned to basal insulin. The clinical characteristics of these patients on continuous SSI and those who transitioned to basal insulin are shown in Table 1. Patients who transitioned to basal insulin had significantly higher mean (SD) admission BG (191.8 [88.2] mg/dL vs 156.4 [65.4] mg/dL, P < .001) and higher mean (SD) HbA1c (8.1% [2.0%] vs 7.01% [1.5%], P < .001), compared with those who remained on continuous SSI. Patients who transitioned to basal insulin were also younger and more likely to have chronic kidney disease (CKD), but less likely to have congestive heart failure, coronary artery disease, or chronic obstructive pulmonary disease (COPD). The Charlson Comorbidity Index score was significantly higher for patients who transitioned to basal (4.4 [2.5]) than for those who remained on continuous SSI (4.1 [2.5], P < .001). There were no significant differences among sex, BMI, or glomerular filtration rate (GFR) on admission. Of those transitioned to basal insulin, 53% achieved a mean hospitalization BG <180 mg/dL, compared with 82% of those on continuous SSI. The overall rate of hypoglycemia in the continuous SSI group was 8% compared with 18% in those transitioned to basal insulin.
Of the patients who remained on continuous SSI throughout the hospitalization, 3,319 patients (48%) had admission BG <140 mg/dL, 1,671 patients (24%) had admission BG 140 to 180 mg/dL, and 1,913 patients (28%) had admission BG >180 mg/dL. Only 9% of patients who remained on continuous SSI had admission BG ≥250 mg/dL. Patients with admission BG <140 mg/dL were older, had lower BMI and HbA1c, had higher rates of COPD and CKD, and were more likely to be admitted to a surgical service compared with patients with admission BG >140 mg/dL (P < .05 for all; Table 2).
Hospital glycemic control for patients on continuous SSI according to admission BG is displayed in Table 3. Among patients who remained on continuous SSI, 96% of patients with admission BG <140 mg/dL had a mean hospital BG <180 mg/dL; of them, 86% achieved target control without hypoglycemia. Similar rates of target control were achieved in patients with admission BG 140 to 180 mg/dL (83%), in contrast to patients with admission BG ≥250 mg/dL, of whom only 18% achieved target control (P < .001). These findings parallel those seen in patients transitioned to basal insulin. Of patients in the transition group admitted with BG <140 mg/dL and <180 mg/dL, 88.5% and 84.6% had mean hospital BG <180 mg/dL, respectively, while 69.1% and 68.9% had mean BG between 70 and 180 mg/dL without hypoglycemia. The overall frequency of hypoglycemia <70 mg/dL among patients on continuous SSI was 8% and was more common in patients with admission BG <140 mg/dL (10%) compared with patients with higher admission glucose levels (BG 140-180 mg/dL [4%], 180-250 mg/dL [4%], or ≥250 mg/dL [6%], P < .001). There was no difference in rates of severe hypoglycemia <40 mg/dL among groups.
HbA1c data were available for 2,560 of the patients on continuous SSI (Table 3). Mean hospital BG increased significantly with increasing HbA1c values. Patients admitted with HbA1c <7% had lower mean (SD) hospital BG (132.2 [28.2] mg/dL) and were more likely to achieve target glucose control during hospitalization (85%) compared with those with HbA1c 7% to 8% (mean BG, 148.7 [30.8] mg/dL; 80% target control), HbA1c 8% to 9% (mean BG, 169.1 [37.9] mg/dL; 61% target control), or HbA1c ≥9% (mean BG, 194.9 [53.4] mg/dL; 38% target control) (P < .001).
In a logistic regression analysis adjusted for age, gender, BMI, race, setting (medicine vs surgery), and Charlson Comorbidity Index score, the odds of poor glycemic control increased with higher admission BG (admission BG 140-180 mg/dL: odds ratio [OR], 1.8; 95% CI, 1.5-2.2; admission BG 180-250 mg/dL: OR, 3.7; 95% CI, 3.1-4.4; admission BG ≥250 mg/dL: OR, 7.2; 95% CI, 5.8-9.0; reference admission BG <140 mg/dL; Figure). Similarly, the logistic regression analysis showed greater odds of poor in-hospital glycemic control with increasing HbA1c (OR, 6.1; 95% CI, 4.3-8.8 for HbA1c >9% compared with HbA1c <7%).
DISCUSSION
This large retrospective cohort study examined the effectiveness of SSI for glycemic control in noncritical inpatients with T2D. Our results indicate that SSI is still widely used in our hospital system, with 31.4% of our initial cohort managed with SSI alone. We found that 86% of patients with BG <140 mg/dL and 83% of patients with BG 140 to 180 mg/dL achieved glycemic control without hypoglycemia when managed with SSI alone, compared with 53% of those admitted with BG 180 to 250 mg/dL and only 18% of those with admission BG ≥250 mg/dL. This high success rate of achieving optimal BG control with SSI alone is comparable to that seen with transition to basal insulin and may explain the prevalent use of SSI for the management of patients with T2D and mild to moderate hyperglycemia.
Published clinical guideline recommendations promoting the use of basal-bolus insulin treatment algorithms are based on the results of a few RCTs that compared the efficacy of SSI vs a basal-bolus insulin regimen. These studies reported significantly lower mean daily BG concentration with basal or basal-bolus insulin therapy compared with SSI.10,11,24 However, it is interesting to note that the mean admission BG of patients treated with SSI in these RCTs ranged from 184 to 225 mg/dL. Patients in these trials were excluded if admission BG was <140 mg/dL.10,11,24 This is in contrast to our study evaluating real-world data in non–critically ill settings in which we found that 48% of patients treated with SSI had admission BG <140 mg/dL, and nearly 75% had admission BG <180 mg/dL. This suggests that by nature of study design, most RCTs excluded the population of patients who do achieve good glycemic control with SSI and may have contributed to the perception that basal insulin is preferable in all populations.
Our analysis indicates that healthcare professionals should consider admission BG when selecting the type of insulin regimen to manage patients with T2D in the hospital. Our results suggest that SSI may be appropriate for many patients with admission BG <180 mg/dL and should be avoided as monotherapy in patients with admission BG ≥180 mg/dL, as the proportion of patients achieving target control decreased with increasing admission BG. More importantly, if a patient is not controlled with SSI alone, intensification of therapy with the addition of basal insulin is indicated to achieve glycemic control. In addition, we found that the admission HbA1c is an appropriate marker to consider as well, with hospital glycemic control deteriorating with increasing HbA1c values, paralleling the admission BG. The main limitation to widespread use of HbA1c for therapeutic decision-making is access to values at time of patient admission; in our population, only 37% of patients had an HbA1c value available during the index hospitalization.
Previous publications have reported that hypoglycemia carries significant safety concerns, especially among a hospitalized population.25-27 As such, we included hypoglycemia as an important metric in our definition of target glycemic control rather than simply using mean hospital BG or number of hyperglycemic events to define treatment effectiveness. We did find a higher rate of hypoglycemia in patients with moderate admission BG treated with SSI compared with those with higher admission BG; however, few patients overall experienced clinically significant (<54 mg/dL) or severe (<40 mg/dL) hypoglycemia.
In our population, only 15% of patients started on SSI received additional basal insulin during hospitalization. This finding is similar to data reported in the Rabbit 2 trial, in which 14% of patients failed SSI alone, with a higher failure rate among those with higher BG on admission.10 Given the observational nature of this study, we cannot definitively state why certain patients in our population required additional basal insulin, but we can hypothesize that these patients admitted with BG ≥180 mg/dL had higher treatment failure rates and greater rates of hyperglycemia, therefore receiving intensified insulin therapy as clinically indicated at the discretion of the treating physician. Patients who transitioned from SSI to basal insulin had significantly higher admission BG and HbA1c compared with patients who remained on SSI alone. We noted that the rates of hypoglycemia were higher in the group that transitioned to basal (18% vs 8%) and similar to rates reported in previous RCTs.11,24
This observational study takes advantage of a large, diverse study population and a combination of medicine and surgery patients in a real-world setting. We acknowledge several limitations in our study. Our primary data were observational in nature, and as such, some baseline patient characteristics were notably different between groups, suggesting selection bias for treatment allocation to SSI. We do not know which patients were managed by primary teams compared with specialized diabetes consult services, which may also influence treatment regimens. We did not have access to information about patients’ at-home diabetes medication regimens or duration of diabetes, both of which have been shown in prior publications to affect an individual’s overall hospital glycemic control. Data on HbA1c values were available for only approximately one-third of patients. In addition, our study did not include patients without a history of diabetes who developed stress-induced hyperglycemia, a population that may benefit from conservative therapy such as SSI.28 A diagnosis of CKD was defined based on ICD 9/10 codes and not on admission estimated GFR. More specific data regarding stage of CKD or changes in renal function over the duration of hospitalization are not available, which could influence insulin prescribing practice. In addition, we defined the basal group as patients prescribed any form of basal insulin (NPH, glargine, detemir or degludec), and we do not have information on the use of prandial versus correction doses of rapid-acting insulin in the basal insulin–treated group.
CONCLUSION
In conclusion, our observational study indicates that the use of SSI results in appropriate target glycemic control for most noncritical medicine and surgery patients with admission BG <180 mg/dL. In agreement with previous RCTs, our study confirms that SSI as monotherapy is frequently inadequate in patients with significant hyperglycemia >180 mg/dL.10,11,24,29 We propose that an individualized approach to inpatient glycemic management is imperative, and cautious use of SSI may be a viable option for certain patients with mild hyperglycemia and admission BG <180 mg/dL. Further observational and randomized studies are needed to confirm the efficacy of SSI therapy in T2D patients with mild hyperglycemia. By identifying which subset of patients can be safely managed with SSI alone, we can better understand which patients will require escalation of therapy with intensive glucose management.
1. Umpierrez GE, Palacio A, Smiley D. Sliding scale insulin use: myth or insanity? Am J Med. 2007;120(7):563-567. https://doi.org/10.1016/j.amjmed.2006.05.070
2. Kitabchi AE, Ayyagari V, Guerra SM. The efficacy of low-dose versus conventional therapy of insulin for treatment of diabetic ketoacidosis. Ann Intern Med. 1976;84(6):633-638. https://doi.org/10.7326/0003-4819-84-6-633
3. Skyler JS, Skyler DL, Seigler DE, O’Sullivan MJ. Algorithms for adjustment of insulin dosage by patients who monitor blood glucose. Diabetes Care. 1981;4(2):311-318. https://doi.org/10.2337/diacare.4.2.311
4. Gearhart JG, Duncan JL 3rd, Replogle WH, Forbes RC, Walley EJ. Efficacy of sliding-scale insulin therapy: a comparison with prospective regimens. Fam Pract Res J. 1994;14(4):313-322.
5. Queale WS, Seidler AJ, Brancati FL. Glycemic control and sliding scale insulin use in medical inpatients with diabetes mellitus. Arch Intern Med. 1997;157(5):545-552.
6. Clement S, Braithwaite SS, Magee MF, et al. Management of diabetes and hyperglycemia in hospitals. Diabetes Care. 2004;27(2):553-591. https://doi.org/10.2337/diacare.27.2.553
7. Garber AJ, Moghissi ES, Bransome ED Jr, et al. American College of Endocrinology position statement on inpatient diabetes and metabolic control. Endocr Pract. 2004;10(1):78-82. https://doi.org/10.4158/EP.10.1.77
8. American Diabetes Association. Standards of medical care in diabetes. Diabetes Care. 2005;28(suppl 1):S4-S36.
9. Umpierrez GE, Hellman R, Korytkowski MT, , et al. Management of hyperglycemia in hospitalized patients in non-critical care setting: an Endocrine Society clinical practice guideline. J Clin Endocrinol Metab. 2012;97(1):16-38. https://doi.org/10.1210/jc.2011-2098
10. Umpierrez GE, Smiley D, Zisman A, et al. Randomized study of basal-bolus insulin therapy in the inpatient management of patients with type 2 diabetes. Diabetes Care. 2007;30(9):2181-2186. https://doi.org/10.2337/dc07-0295
11. Umpierrez GE, Smiley D, Jacobs S, et al. Randomized study of basal-bolus insulin therapy in the inpatient management of patients with type 2 diabetes undergoing general surgery (RABBIT 2 surgery). Diabetes Care. 2011;34(2):256-261. https://doi.org/10.2337/dc10-1407
12. Schroeder JE, Liebergall M, Raz I, Egleston R, Ben Sussan G, Peyser A. Benefits of a simple glycaemic protocol in an orthopaedic surgery ward: a randomized prospective study. Diabetes Metab Res Rev. 2012;28:71-75. https://doi.org/10.1002/dmrr.1217
13. Lee YY, Lin YM, Leu WJ, et al. Sliding-scale insulin used for blood glucose control: a meta-analysis of randomized controlled trials. Metabolism. 2015;64(9):1183-1192. https://doi.org/10.1016/j.metabol.2015.05.011
14. Christensen MB, Gotfredsen A, Nørgaard K. Efficacy of basal-bolus insulin regimens in the inpatient management of non-critically ill patients with type 2 diabetes: a systematic review and meta-analysis. Diabetes Metab Res Rev. 2017;33(5):e2885. https://doi.org/10.1002/dmrr.2885
15. Wexler DJ, Meigs JB, Cagliero E, Nathan DM, Grant RW. Prevalence of hyper- and hypoglycemia among inpatients with diabetes: a national survey of 44 U.S. hospitals. Diabetes Care. 2007;30(2):367-369. https://doi.org/10.2337/dc06-1715
16. Moreira ED Jr, Silveira PCB, Neves RCS, Souza C Jr, Nunes ZO, Almeida MdCC. Glycemic control and diabetes management in hospitalized patients in Brazil. Diabetol Metab Syndr. 2013;5(1):62. https://doi.org/10.1186/1758-5996-5-62
17. Akhtar ST, Mahmood K, Naqvi IH, Vaswani AS. Inpatient management of type 2 diabetes mellitus: does choice of insulin regimen really matter? Pakistan J Med Sci. 2014;30(4):895-898.
18. Gómez Cuervo C, Sánchez Morla A, Pérez-Jacoiste Asín MA, Bisbal Pardo O, Pérez Ordoño L, Vila Santos J. Effective adverse event reduction with bolus-basal versus sliding scale insulin therapy in patients with diabetes during conventional hospitalization: systematic review and meta-analysis. Endocrinol Nutr. 2016;63(4):145-156. https://doi.org/10.1016/j.endonu.2015.11.008
19. Bain A, Hasan SS, Babar ZUD. Interventions to improve insulin prescribing practice for people with diabetes in hospital: a systematic review. Diabet Med. 2019;36(8):948-960. https://doi.org/10.1111/dme.13982
20. Ambrus DB, O’Connor MJ. Things We Do For No Reason: sliding-scale insulin as monotherapy for glycemic control in hospitalized patients. J Hosp Med. 2019;14(2):114-116. https://doi.org/10.12788/jhm.3109
21. Nau KC, Lorenzetti RC, Cucuzzella M, Devine T, Kline J. Glycemic control in hospitalized patients not in intensive care: beyond sliding-scale insulin. Am Fam Physician. 2010;81(9):1130-1135.
22. Colunga-Lozano LE, Gonzalez Torres FJ, Delgado-Figueroa N, et al. Sliding scale insulin for non-critically ill hospitalised adults with diabetes mellitus. Cochrane Database Syst Rev. 2018;11(11):CD011296. https://doi.org/10.1002/14651858.CD011296.pub2
23. American Diabetes Association. Diabetes care in the hospital: Standards of Medical Care in Diabetes—2019. Diabetes Care. 2019;42(suppl 1):S173-S181. https://doi.org/10.2337/dc19-S015
24. Umpierrez GE, Smiley D, Hermayer K, et al. Randomized study comparing a basal-bolus with a basal plus correction management of medical and surgical patients with type 2 diabetes: basal plus trial. Diabetes Care. 2013;36(8):2169-2174. https://doi.org/10.2337/dc12-1988
25. Turchin A, Matheny ME, Shubina M, Scanlon SV, Greenwood B, Pendergrass ML. Hypoglycemia and clinical outcomes in patients with diabetes hospitalized in the general ward. Diabetes Care. 2009;32(7):1153-1157. https://doi.org/10.2337/dc08-2127
26. Garg R, Hurwitz S, Turchin A, Trivedi A. Hypoglycemia, with or without insulin therapy, is associated with increased mortality among hospitalized patients. Diabetes Care. 2013;36(5):1107-1110. https://doi.org/10.2337/dc12-1296
27. Zapatero A, Gómez-Huelgas R, González N, et al. Frequency of hypoglycemia and its impact on length of stay, mortality, and short-term readmission in patients with diabetes hospitalized in internal medicine wards. Endocr Pract. 2014;20(9):870-875. https://doi.org/10.4158/EP14006.OR
28. Umpierrez GE, Isaacs SD, Bazargan N, You X, Thaler LM, Kitabchi AE. Hyperglycemia: an independent marker of in-hospital mortality in patients with undiagnosed diabetes. J Clin Endocrinol Metab. 2002;87(3):978-982. https://doi.org/10.1210/jcem.87.3.8341
29. Dickerson LM, Ye X, Sack JL, Hueston WJ. Glycemic control in medical inpatients with type 2 diabetes mellitus receiving sliding scale insulin regimens versus routine diabetes medications: a multicenter randomized controlled trial. Ann Fam Med. 2003;1(1):29-35. https://doi.org/10.1370/afm.2
1. Umpierrez GE, Palacio A, Smiley D. Sliding scale insulin use: myth or insanity? Am J Med. 2007;120(7):563-567. https://doi.org/10.1016/j.amjmed.2006.05.070
2. Kitabchi AE, Ayyagari V, Guerra SM. The efficacy of low-dose versus conventional therapy of insulin for treatment of diabetic ketoacidosis. Ann Intern Med. 1976;84(6):633-638. https://doi.org/10.7326/0003-4819-84-6-633
3. Skyler JS, Skyler DL, Seigler DE, O’Sullivan MJ. Algorithms for adjustment of insulin dosage by patients who monitor blood glucose. Diabetes Care. 1981;4(2):311-318. https://doi.org/10.2337/diacare.4.2.311
4. Gearhart JG, Duncan JL 3rd, Replogle WH, Forbes RC, Walley EJ. Efficacy of sliding-scale insulin therapy: a comparison with prospective regimens. Fam Pract Res J. 1994;14(4):313-322.
5. Queale WS, Seidler AJ, Brancati FL. Glycemic control and sliding scale insulin use in medical inpatients with diabetes mellitus. Arch Intern Med. 1997;157(5):545-552.
6. Clement S, Braithwaite SS, Magee MF, et al. Management of diabetes and hyperglycemia in hospitals. Diabetes Care. 2004;27(2):553-591. https://doi.org/10.2337/diacare.27.2.553
7. Garber AJ, Moghissi ES, Bransome ED Jr, et al. American College of Endocrinology position statement on inpatient diabetes and metabolic control. Endocr Pract. 2004;10(1):78-82. https://doi.org/10.4158/EP.10.1.77
8. American Diabetes Association. Standards of medical care in diabetes. Diabetes Care. 2005;28(suppl 1):S4-S36.
9. Umpierrez GE, Hellman R, Korytkowski MT, , et al. Management of hyperglycemia in hospitalized patients in non-critical care setting: an Endocrine Society clinical practice guideline. J Clin Endocrinol Metab. 2012;97(1):16-38. https://doi.org/10.1210/jc.2011-2098
10. Umpierrez GE, Smiley D, Zisman A, et al. Randomized study of basal-bolus insulin therapy in the inpatient management of patients with type 2 diabetes. Diabetes Care. 2007;30(9):2181-2186. https://doi.org/10.2337/dc07-0295
11. Umpierrez GE, Smiley D, Jacobs S, et al. Randomized study of basal-bolus insulin therapy in the inpatient management of patients with type 2 diabetes undergoing general surgery (RABBIT 2 surgery). Diabetes Care. 2011;34(2):256-261. https://doi.org/10.2337/dc10-1407
12. Schroeder JE, Liebergall M, Raz I, Egleston R, Ben Sussan G, Peyser A. Benefits of a simple glycaemic protocol in an orthopaedic surgery ward: a randomized prospective study. Diabetes Metab Res Rev. 2012;28:71-75. https://doi.org/10.1002/dmrr.1217
13. Lee YY, Lin YM, Leu WJ, et al. Sliding-scale insulin used for blood glucose control: a meta-analysis of randomized controlled trials. Metabolism. 2015;64(9):1183-1192. https://doi.org/10.1016/j.metabol.2015.05.011
14. Christensen MB, Gotfredsen A, Nørgaard K. Efficacy of basal-bolus insulin regimens in the inpatient management of non-critically ill patients with type 2 diabetes: a systematic review and meta-analysis. Diabetes Metab Res Rev. 2017;33(5):e2885. https://doi.org/10.1002/dmrr.2885
15. Wexler DJ, Meigs JB, Cagliero E, Nathan DM, Grant RW. Prevalence of hyper- and hypoglycemia among inpatients with diabetes: a national survey of 44 U.S. hospitals. Diabetes Care. 2007;30(2):367-369. https://doi.org/10.2337/dc06-1715
16. Moreira ED Jr, Silveira PCB, Neves RCS, Souza C Jr, Nunes ZO, Almeida MdCC. Glycemic control and diabetes management in hospitalized patients in Brazil. Diabetol Metab Syndr. 2013;5(1):62. https://doi.org/10.1186/1758-5996-5-62
17. Akhtar ST, Mahmood K, Naqvi IH, Vaswani AS. Inpatient management of type 2 diabetes mellitus: does choice of insulin regimen really matter? Pakistan J Med Sci. 2014;30(4):895-898.
18. Gómez Cuervo C, Sánchez Morla A, Pérez-Jacoiste Asín MA, Bisbal Pardo O, Pérez Ordoño L, Vila Santos J. Effective adverse event reduction with bolus-basal versus sliding scale insulin therapy in patients with diabetes during conventional hospitalization: systematic review and meta-analysis. Endocrinol Nutr. 2016;63(4):145-156. https://doi.org/10.1016/j.endonu.2015.11.008
19. Bain A, Hasan SS, Babar ZUD. Interventions to improve insulin prescribing practice for people with diabetes in hospital: a systematic review. Diabet Med. 2019;36(8):948-960. https://doi.org/10.1111/dme.13982
20. Ambrus DB, O’Connor MJ. Things We Do For No Reason: sliding-scale insulin as monotherapy for glycemic control in hospitalized patients. J Hosp Med. 2019;14(2):114-116. https://doi.org/10.12788/jhm.3109
21. Nau KC, Lorenzetti RC, Cucuzzella M, Devine T, Kline J. Glycemic control in hospitalized patients not in intensive care: beyond sliding-scale insulin. Am Fam Physician. 2010;81(9):1130-1135.
22. Colunga-Lozano LE, Gonzalez Torres FJ, Delgado-Figueroa N, et al. Sliding scale insulin for non-critically ill hospitalised adults with diabetes mellitus. Cochrane Database Syst Rev. 2018;11(11):CD011296. https://doi.org/10.1002/14651858.CD011296.pub2
23. American Diabetes Association. Diabetes care in the hospital: Standards of Medical Care in Diabetes—2019. Diabetes Care. 2019;42(suppl 1):S173-S181. https://doi.org/10.2337/dc19-S015
24. Umpierrez GE, Smiley D, Hermayer K, et al. Randomized study comparing a basal-bolus with a basal plus correction management of medical and surgical patients with type 2 diabetes: basal plus trial. Diabetes Care. 2013;36(8):2169-2174. https://doi.org/10.2337/dc12-1988
25. Turchin A, Matheny ME, Shubina M, Scanlon SV, Greenwood B, Pendergrass ML. Hypoglycemia and clinical outcomes in patients with diabetes hospitalized in the general ward. Diabetes Care. 2009;32(7):1153-1157. https://doi.org/10.2337/dc08-2127
26. Garg R, Hurwitz S, Turchin A, Trivedi A. Hypoglycemia, with or without insulin therapy, is associated with increased mortality among hospitalized patients. Diabetes Care. 2013;36(5):1107-1110. https://doi.org/10.2337/dc12-1296
27. Zapatero A, Gómez-Huelgas R, González N, et al. Frequency of hypoglycemia and its impact on length of stay, mortality, and short-term readmission in patients with diabetes hospitalized in internal medicine wards. Endocr Pract. 2014;20(9):870-875. https://doi.org/10.4158/EP14006.OR
28. Umpierrez GE, Isaacs SD, Bazargan N, You X, Thaler LM, Kitabchi AE. Hyperglycemia: an independent marker of in-hospital mortality in patients with undiagnosed diabetes. J Clin Endocrinol Metab. 2002;87(3):978-982. https://doi.org/10.1210/jcem.87.3.8341
29. Dickerson LM, Ye X, Sack JL, Hueston WJ. Glycemic control in medical inpatients with type 2 diabetes mellitus receiving sliding scale insulin regimens versus routine diabetes medications: a multicenter randomized controlled trial. Ann Fam Med. 2003;1(1):29-35. https://doi.org/10.1370/afm.2
© 2021 Society of Hospital Medicine
Identifying the Sickest During Triage: Using Point-of-Care Severity Scores to Predict Prognosis in Emergency Department Patients With Suspected Sepsis
Sepsis is the leading cause of in-hospital mortality in the United States.1 Sepsis is present on admission in 85% of cases, and each hour delay in antibiotic treatment is associated with 4% to 7% increased odds of mortality.2,3 Prompt identification and treatment of sepsis is essential for reducing morbidity and mortality, but identifying sepsis during triage is challenging.2
Risk stratification scores that rely solely on data readily available at the bedside have been developed to quickly identify those at greatest risk of poor outcomes from sepsis in real time. The quick Sequential Organ Failure Assessment (qSOFA) score, the National Early Warning System (NEWS2), and the Shock Index are easy-to-calculate measures that use routinely collected clinical data that are not subject to laboratory delay. These scores can be incorporated into electronic health record (EHR)-based alerts and can be calculated longitudinally to track the risk of poor outcomes over time. qSOFA was developed to quantify patient risk at bedside in non-intensive care unit (ICU) settings, but there is no consensus about its ability to predict adverse outcomes such as mortality and ICU admission.4-6 The United Kingdom’s National Health Service uses NEWS2 to identify patients at risk for sepsis.7 NEWS has been shown to have similar or better sensitivity in identifying poorer outcomes in sepsis patients compared with systemic inflammatory response syndrome (SIRS) criteria and qSOFA.4,8-11 However, since the latest update of NEWS2 in 2017, there has been little study of its predictive ability. The Shock Index is a simple bedside score (heart rate divided by systolic blood pressure) that was developed to detect changes in cardiovascular performance before systemic shock onset. Although it was not developed for infection and has not been regularly applied in the sepsis literature, the Shock Index might be useful for identifying patients at increased risk of poor outcomes. Patients with higher and sustained Shock Index scores are more likely to experience morbidity, such as hyperlactatemia, vasopressor use, and organ failure, and also have an increased risk of mortality.12-14
Although the predictive abilities of these bedside risk stratification scores have been assessed individually using standard binary cut-points, the comparative performance of qSOFA, the Shock Index, and NEWS2 has not been evaluated in patients presenting to an emergency department (ED) with suspected sepsis.
METHODS
Design and Setting
We conducted a retrospective cohort study of ED patients who presented with suspected sepsis to the University of California San Francisco (UCSF) Helen Diller Medical Center at Parnassus Heights between June 1, 2012, and December 31, 2018. Our institution is a 785-bed academic teaching hospital with approximately 30,000 ED encounters per year. The study was approved with a waiver of informed consent by the UCSF Human Research Protection Program.
Participants
We use an Epic-based EHR platform (Epic 2017, Epic Systems Corporation) for clinical care, which was implemented on June 1, 2012. All data elements were obtained from Clarity, the relational database that stores Epic’s inpatient data. The study included encounters for patients age ≥18 years who had blood cultures ordered within 24 hours of ED presentation and administration of intravenous antibiotics within 24 hours. Repeat encounters were treated independently in our analysis.
Outcomes and Measures
We compared the ability of qSOFA, the Shock Index, and NEWS2 to predict in-hospital mortality and admission to the ICU from the ED (ED-to-ICU admission). We used the
We compared demographic and clinical characteristics of patients who were positive for qSOFA, the Shock Index, and NEWS2. Demographic data were extracted from the EHR and included primary language, age, sex, and insurance status. All International Classification of Diseases (ICD)-9/10 diagnosis codes were pulled from Clarity billing tables. We used the Elixhauser comorbidity groupings19 of ICD-9/10 codes present on admission to identify preexisting comorbidities and underlying organ dysfunction. To estimate burden of comorbid illnesses, we calculated the validated van Walraven comorbidity index,20 which provides an estimated risk of in-hospital death based on documented Elixhauser comorbidities. Admission level of care (acute, stepdown, or intensive care) was collected for inpatient admissions to assess initial illness severity.21 We also evaluated discharge disposition and in-hospital mortality. Index blood culture results were collected, and dates and timestamps of mechanical ventilation, fluid, vasopressor, and antibiotic administration were obtained for the duration of the encounter.
UCSF uses an automated, real-time, algorithm-based severe sepsis alert that is triggered when a patient meets ≥2 SIRS criteria and again when the patient meets severe sepsis or septic shock criteria (ie, ≥2 SIRS criteria in addition to end-organ dysfunction and/or fluid nonresponsive hypotension). This sepsis screening alert was in use for the duration of our study.22
Statistical Analysis
We performed a subgroup analysis among those who were diagnosed with sepsis, according to the 2016 Third International Consensus Definitions for Sepsis and Septic Shock (Sepsis-3) criteria.
All statistical analyses were conducted using Stata 14 (StataCorp). We summarized differences in demographic and clinical characteristics among the populations meeting each severity score but elected not to conduct hypothesis testing because patients could be positive for one or more scores. We calculated sensitivity, specificity, positive predictive value (PPV), and negative predictive value (NPV) for each score to predict in-hospital mortality and ED-to-ICU admission. To allow comparison with other studies, we also created a composite outcome of either in-hospital mortality or ED-to-ICU admission.
RESULTS
Within our sample 23,837 ED patients had blood cultures ordered within 24 hours of ED presentation and were considered to have suspected sepsis. The mean age of the cohort was 60.8 years, and 1,612 (6.8%) had positive blood cultures. A total of 12,928 patients (54.2%) were found to have sepsis. We documented 1,427 in-hospital deaths (6.0%) and 3,149 (13.2%) ED-to-ICU admissions. At ED triage 1,921 (8.1%) were qSOFA-positive, 4,273 (17.9%) were Shock Index-positive, and 11,832 (49.6%) were NEWS2-positive. At ED triage, blood pressure, heart rate, respiratory rate, and oxygen saturated were documented in >99% of patients, 93.5% had temperature documented, and 28.5% had GCS recorded. If the window of assessment was widened to 1 hour, GCS was only documented among 44.2% of those with suspected sepsis.
Demographic Characteristics and Clinical Course
qSOFA-positive patients received antibiotics more quickly than those who were Shock Index-positive or NEWS2-positive (median 1.5, 1.8, and 2.8 hours after admission, respectively). In addition, those who were qSOFA-positive were more likely to have a positive blood culture (10.9%, 9.4%, and 8.5%, respectively) and to receive an EHR-based diagnosis of sepsis (77.0%, 69.6%, and 60.9%, respectively) than those who were Shock Index- or NEWS2-positive. Those who were qSOFA-positive also were more likely to be mechanically ventilated during their hospital stay (25.4%, 19.2%, and 10.8%, respectively) and to receive vasopressors (33.5%, 22.5%, and 12.2%, respectively). In-hospital mortality also was more common among those who were qSOFA-positive at triage (23.4%, 15.3%, and 9.2%, respectively).
Because both qSOFA and NEWS2 incorporate GCS, we explored baseline characteristics of patients with GCS documented at triage (n = 6,794). These patients were older (median age 63 and 61 years, P < .0001), more likely to be male (54.9% and 53.4%, P = .0031), more likely to have renal failure (22.8% and 20.1%, P < .0001), more likely to have liver disease (14.2% and 12.8%, P = .006), had a higher van Walraven comorbidity score on presentation (median 10 and 8, P < .0001), and were more likely to go directly to the ICU from the ED (20.2% and 10.6%, P < .0001). However, among the 6,397 GCS scores documented at triage, only 1,579 (24.7%) were abnormal.
Test Characteristics of qSOFA, Shock Index, and NEWS2 for Predicting In-hospital Mortality and ED-to-ICU Admission
Among 23,837 patients with suspected sepsis, NEWS2 had the highest sensitivity for predicting in-hospital mortality (76.0%; 95% CI, 73.7%-78.2%) and ED-to-ICU admission (78.9%; 95% CI, 77.5%-80.4%) but had the lowest specificity for in-hospital mortality (52.0%; 95% CI, 51.4%-52.7%) and for ED-to-ICU admission (54.8%; 95% CI, 54.1%-55.5%) (Table 3). qSOFA had the lowest sensitivity for in-hospital mortality (31.5%; 95% CI, 29.1%-33.9%) and ED-to-ICU admission (29.3%; 95% CI, 27.7%-30.9%) but the highest specificity for in-hospital mortality (93.4%; 95% CI, 93.1%-93.8%) and ED-to-ICU admission (95.2%; 95% CI, 94.9%-95.5%). The Shock Index had a sensitivity that fell between qSOFA and NEWS2 for in-hospital mortality (45.8%; 95% CI, 43.2%-48.5%) and ED-to-ICU admission (49.2%; 95% CI, 47.5%-51.0%). The specificity of the Shock Index also was between qSOFA and NEWS2 for in-hospital mortality (83.9%; 95% CI, 83.4%-84.3%) and ED-to-ICU admission (86.8%; 95% CI, 86.4%-87.3%). All three scores exhibited relatively low PPV, ranging from 9.2% to 23.4% for in-hospital mortality and 21.0% to 48.0% for ED-to-ICU triage. Conversely, all three scores exhibited relatively high NPV, ranging from 95.5% to 97.1% for in-hospital mortality and 89.8% to 94.5% for ED-to-ICU triage.
When considering a binary cutoff, the Shock Index exhibited the highest AUROC for in-hospital mortality (0.648; 95% CI, 0.635-0.662) and had a significantly higher AUROC than qSOFA (AUROC, 0.625; 95% CI, 0.612-0.637; P = .0005), but there was no difference compared with NEWS2 (AUROC, 0.640; 95% CI, 0.628-0.652; P = .2112). NEWS2 had a significantly higher AUROC than qSOFA for predicting in-hospital mortality (P = .0227). The Shock Index also exhibited the highest AUROC for ED-to-ICU admission (0.680; 95% CI, 0.617-0.689), which was significantly higher than the AUROC for qSOFA (P < .0001) and NEWS2 (P = 0.0151). NEWS2 had a significantly higher AUROC than qSOFA for predicting ED-to-ICU admission (P < .0001). Similar findings were seen in patients found to have sepsis.
DISCUSSION
In this retrospective cohort study of 23,837 patients who presented to the ED with suspected sepsis, the standard qSOFA threshold was met least frequently, followed by the Shock Index and NEWS2. NEWS2 had the highest sensitivity but the lowest specificity for predicting in-hospital mortality and ED-to-ICU admission, making it a challenging bedside risk stratification scale for identifying patients at risk of poor clinical outcomes. When comparing predictive performance among the three scales, qSOFA had the highest specificity and the Shock Index had the highest AUROC for in-hospital mortality and ED-to-ICU admission in this cohort of patients with suspected sepsis. These trends in sensitivity, specificity, and AUROC were consistent among those who met EHR criteria for a sepsis diagnosis. In the analysis of the three scoring systems using all available cut-points, qSOFA and NEWS2 had the highest AUROCs, followed by the Shock Index.
Considering the rapid progression from organ dysfunction to death in sepsis patients, as well as the difficulty establishing a sepsis diagnosis at triage,23 providers must quickly identify patients at increased risk of poor outcomes when they present to the ED. Sepsis alerts often are built using SIRS criteria,27 including the one used for sepsis surveillance at UCSF since 2012,22 but the white blood cell count criterion is subject to a laboratory lag and could lead to a delay in identification. Implementation of a point-of-care bedside score alert that uses readily available clinical data could allow providers to identify patients at greatest risk of poor outcomes immediately at ED presentation and triage, which motivated us to explore the predictive performance of qSOFA, the Shock Index, and NEWS2.
Our study is the first to provide a head-to-head comparison of the predictive performance of qSOFA, the Shock Index, and NEWS2, three easy-to-calculate bedside risk scores that use EHR data collected among patients with suspected sepsis. The Sepsis-3 guidelines recommend qSOFA to quickly identify non-ICU patients at greatest risk of poor outcomes because the measure exhibited predictive performance similar to the more extensive SOFA score outside the ICU.16,23 Although some studies have confirmed qSOFA’s high predictive performance,28-31 our test characteristics and AUROC findings are in line with other published analyses.4,6,10,17 The UK National Health Service is using NEWS2 to screen for patients at risk of poor outcomes from sepsis. Several analyses that assessed the predictive ability of NEWS have reported estimates in line with our findings.4,10,32 The Shock Index was introduced in 1967 and provided a metric to evaluate hemodynamic stability based on heart rate and systolic blood pressure.33 The Shock Index has been studied in several contexts, including sepsis,34 and studies show that a sustained Shock Index is associated with increased odds of vasopressor administration, higher prevalence of hyperlactatemia, and increased risk of poor outcomes in the ICU.13,14
For our study, we were particularly interested in exploring how the Shock Index would compare with more frequently used severity scores such as qSOFA and NEWS2 among patients with suspected sepsis, given the simplicity of its calculation and the easy availability of required data. In our cohort of 23,837 patients, only 159 people had missing blood pressure and only 71 had omitted heart rate. In contrast, both qSOFA and NEWS2 include an assessment of level of consciousness that can be subject to variability in assessment methods and EHR documentation across institutions.11 In our cohort, GCS within 30 minutes of ED presentation was missing in 72 patients, which could have led to incomplete calculation of qSOFA and NEWS2 if a missing value was not actually within normal limits.
Several investigations relate qSOFA to NEWS but few compare qSOFA with the newer NEWS2, and even fewer evaluate the Shock Index with any of these scores.10,11,18,29,35-37 In general, studies have shown that NEWS exhibits a higher AUROC for predicting mortality, sepsis with organ dysfunction, and ICU admission, often as a composite outcome.4,11,18,37,38 A handful of studies compare the Shock Index to SIRS; however, little has been done to compare the Shock Index to qSOFA or NEWS2, scores that have been used specifically for sepsis and might be more predictive of poor outcomes than SIRS.33 In our study, the Shock Index had a higher AUROC than either qSOFA or NEWS2 for predicting in-hospital mortality and ED-to-ICU admission measured as separate outcomes and as a composite outcome using standard cut-points for these scores.
When selecting a severity score to apply in an institution, it is important to carefully evaluate the score’s test characteristics, in addition to considering the availability of reliable data. Tests with high sensitivity and NPV for the population being studied can be useful to rule out disease or risk of poor outcome, while tests with high specificity and PPV can be useful to rule in disease or risk of poor outcome.39 When considering specificity, qSOFA’s performance was superior to the Shock Index and NEWS2 in our study, but a small percentage of the population was identified using a cut-point of qSOFA ≥2. If we used qSOFA and applied this standard cut-point at our institution, we could be confident that those identified were at increased risk, but we would miss a significant number of patients who would experience a poor outcome. When considering sensitivity, performance of NEWS2 was superior to qSOFA and the Shock Index in our study, but one-half of the population was identified using a cut-point of NEWS2 ≥5. If we were to apply this standard NEWS2 cut-point at our institution, we would assume that one-half of our population was at risk, which might drive resource use towards patients who will not experience a poor outcome. Although none of the scores exhibited a robust AUROC measure, the Shock Index had the highest AUROC for in-hospital mortality and ED-to-ICU admission when using the standard binary cut-point, and its sensitivity and specificity is between that of qSOFA and NEWS2, potentially making it a score to use in settings where qSOFA and NEWS2 score components, such as altered mentation, are not reliably collected. Finally, our sensitivity analysis varying the binary cut-point of each score within our population demonstrated that the standard cut-points might not be as useful within a specific population and might need to be tailored for implementation, balancing sensitivity, specificity, PPV, and NPV to meet local priorities and ICU capacity.
Our study has limitations. It is a single-center, retrospective analysis, factors that could reduce generalizability. However, it does include a large and diverse patient population spanning several years. Missing GCS data could have affected the predictive ability of qSOFA and NEWS2 in our cohort. We could not reliably perform imputation of GCS because of the high missingness and therefore we assumed missing was normal, as was done in the Sepsis-3 derivation studies.16 Previous studies have attempted to impute GCS and have not observed improved performance of qSOFA to predict mortality.40 Because manually collected variables such as GCS are less reliably documented in the EHR, there might be limitations in their use for triage risk scores.
Although the current analysis focused on the predictive performance of qSOFA, the Shock Index, and NEWS2 at triage, performance of these scores could affect the ED team’s treatment decisions before handoff to the hospitalist team and the expected level of care the patient will receive after in-patient admission. These tests also have the advantage of being easy to calculate at the bedside over time, which could provide an objective assessment of longitudinal predicted prognosis.
CONCLUSION
Local priorities should drive selection of a screening tool, balancing sensitivity, specificity, PPV, and NPV to achieve the institution’s goals. qSOFA, Shock Index, and NEWS2 are risk stratification tools that can be easily implemented at ED triage using data available at the bedside. Although none of these scores performed strongly when comparing AUROCs, qSOFA was highly specific for identifying patients with poor outcomes, and NEWS2 was the most sensitive for ruling out those at high risk among patients with suspected sepsis. The Shock Index exhibited a sensitivity and specificity that fell between qSOFA and NEWS2 and also might be considered to identify those at increased risk, given its ease of implementation, particularly in settings where altered mentation is unreliably or inconsistently documented.
Acknowledgment
The authors thank the UCSF Division of Hospital Medicine Data Core for their assistance with data acquisition.
1. Jones SL, Ashton CM, Kiehne LB, et al. Outcomes and resource use of sepsis-associated stays by presence on admission, severity, and hospital type. Med Care. 2016;54(3):303-310. https://doi.org/10.1097/MLR.0000000000000481
2. Seymour CW, Gesten F, Prescott HC, et al. Time to treatment and mortality during mandated emergency care for sepsis. N Engl J Med. 2017;376(23):2235-2244. https://doi.org/10.1056/NEJMoa1703058
3. Kumar A, Roberts D, Wood KE, et al. Duration of hypotension before initiation of effective antimicrobial therapy is the critical determinant of survival in human septic shock. Crit Care Med. 2006;34(6):1589-1596. https://doi.org/10.1097/01.CCM.0000217961.75225.E9
4. Churpek MM, Snyder A, Sokol S, Pettit NN, Edelson DP. Investigating the impact of different suspicion of infection criteria on the accuracy of Quick Sepsis-Related Organ Failure Assessment, Systemic Inflammatory Response Syndrome, and Early Warning Scores. Crit Care Med. 2017;45(11):1805-1812. https://doi.org/10.1097/CCM.0000000000002648
5. Abdullah SMOB, Sørensen RH, Dessau RBC, Sattar SMRU, Wiese L, Nielsen FE. Prognostic accuracy of qSOFA in predicting 28-day mortality among infected patients in an emergency department: a prospective validation study. Emerg Med J. 2019;36(12):722-728. https://doi.org/10.1136/emermed-2019-208456
6. Kim KS, Suh GJ, Kim K, et al. Quick Sepsis-related Organ Failure Assessment score is not sensitive enough to predict 28-day mortality in emergency department patients with sepsis: a retrospective review. Clin Exp Emerg Med. 2019;6(1):77-83. HTTPS://DOI.ORG/ 10.15441/ceem.17.294
7. National Early Warning Score (NEWS) 2: Standardising the assessment of acute-illness severity in the NHS. Royal College of Physicians; 2017.
8. Brink A, Alsma J, Verdonschot RJCG, et al. Predicting mortality in patients with suspected sepsis at the emergency department: a retrospective cohort study comparing qSOFA, SIRS and National Early Warning Score. PLoS One. 2019;14(1):e0211133. https://doi.org/ 10.1371/journal.pone.0211133
9. Redfern OC, Smith GB, Prytherch DR, Meredith P, Inada-Kim M, Schmidt PE. A comparison of the Quick Sequential (Sepsis-Related) Organ Failure Assessment Score and the National Early Warning Score in non-ICU patients with/without infection. Crit Care Med. 2018;46(12):1923-1933. https://doi.org/10.1097/CCM.0000000000003359
10. Churpek MM, Snyder A, Han X, et al. Quick Sepsis-related Organ Failure Assessment, Systemic Inflammatory Response Syndrome, and Early Warning Scores for detecting clinical deterioration in infected patients outside the intensive care unit. Am J Respir Crit Care Med. 2017;195(7):906-911. https://doi.org/10.1164/rccm.201604-0854OC
11. Goulden R, Hoyle MC, Monis J, et al. qSOFA, SIRS and NEWS for predicting inhospital mortality and ICU admission in emergency admissions treated as sepsis. Emerg Med J. 2018;35(6):345-349. https://doi.org/10.1136/emermed-2017-207120
12. Biney I, Shepherd A, Thomas J, Mehari A. Shock Index and outcomes in patients admitted to the ICU with sepsis. Chest. 2015;148(suppl 4):337A. https://doi.org/https://doi.org/10.1378/chest.2281151
13. Wira CR, Francis MW, Bhat S, Ehrman R, Conner D, Siegel M. The shock index as a predictor of vasopressor use in emergency department patients with severe sepsis. West J Emerg Med. 2014;15(1):60-66. https://doi.org/10.5811/westjem.2013.7.18472
14. Berger T, Green J, Horeczko T, et al. Shock index and early recognition of sepsis in the emergency department: pilot study. West J Emerg Med. 2013;14(2):168-174. https://doi.org/10.5811/westjem.2012.8.11546
15. Middleton DJ, Smith TO, Bedford R, Neilly M, Myint PK. Shock Index predicts outcome in patients with suspected sepsis or community-acquired pneumonia: a systematic review. J Clin Med. 2019;8(8):1144. https://doi.org/10.3390/jcm8081144
16. Seymour CW, Liu VX, Iwashyna TJ, et al. Assessment of clinical criteria for sepsis: for the Third International Consensus Definitions for Sepsis and Septic Shock (Sepsis-3). JAMA. 2016;315(8):762-774. https://doi.org/ 10.1001/jama.2016.0288
17. Abdullah S, Sørensen RH, Dessau RBC, Sattar S, Wiese L, Nielsen FE. Prognostic accuracy of qSOFA in predicting 28-day mortality among infected patients in an emergency department: a prospective validation study. Emerg Med J. 2019;36(12):722-728. https://doi.org/10.1136/emermed-2019-208456
18. Usman OA, Usman AA, Ward MA. Comparison of SIRS, qSOFA, and NEWS for the early identification of sepsis in the Emergency Department. Am J Emerg Med. 2018;37(8):1490-1497. https://doi.org/10.1016/j.ajem.2018.10.058
19. Elixhauser A, Steiner C, Harris DR, Coffey RM. Comorbidity measures for use with administrative data. Med Care. 1998;36(1):8-27. https://doi.org/10.1097/00005650-199801000-00004
20. van Walraven C, Austin PC, Jennings A, Quan H, Forster AJ. A modification of the Elixhauser comorbidity measures into a point system for hospital death using administrative data. Med Care. 2009;47(6):626-633. https://doi.org/10.1097/MLR.0b013e31819432e5
21. Prin M, Wunsch H. The role of stepdown beds in hospital care. Am J Respir Crit Care Med. 2014;190(11):1210-1216. https://doi.org/10.1164/rccm.201406-1117PP
22. Narayanan N, Gross AK, Pintens M, Fee C, MacDougall C. Effect of an electronic medical record alert for severe sepsis among ED patients. Am J Emerg Med. 2016;34(2):185-188. https://doi.org/10.1016/j.ajem.2015.10.005
23. Singer M, Deutschman CS, Seymour CW, et al. The Third International Consensus Definitions for Sepsis and Septic Shock (Sepsis-3). JAMA. 2016;315(8):801-810. https://doi.org/10.1001/jama.2016.0287
24. Rhee C, Dantes R, Epstein L, et al. Incidence and trends of sepsis in US hospitals using clinical vs claims data, 2009-2014. JAMA. 2017;318(13):1241-1249. https://doi.org/10.1001/jama.2017.13836
25. Safari S, Baratloo A, Elfil M, Negida A. Evidence based emergency medicine; part 5 receiver operating curve and area under the curve. Emerg (Tehran). 2016;4(2):111-113.
26. DeLong ER, DeLong DM, Clarke-Pearson DL. Comparing the areas under two or more correlated receiver operating characteristic curves: a nonparametric approach. Biometrics. 1988;44(3):837-845.
27. Kangas C, Iverson L, Pierce D. Sepsis screening: combining Early Warning Scores and SIRS Criteria. Clin Nurs Res. 2021;30(1):42-49. https://doi.org/10.1177/1054773818823334.
28. Freund Y, Lemachatti N, Krastinova E, et al. Prognostic accuracy of Sepsis-3 Criteria for in-hospital mortality among patients with suspected infection presenting to the emergency department. JAMA. 2017;317(3):301-308. https://doi.org/10.1001/jama.2016.20329
29. Finkelsztein EJ, Jones DS, Ma KC, et al. Comparison of qSOFA and SIRS for predicting adverse outcomes of patients with suspicion of sepsis outside the intensive care unit. Crit Care. 2017;21(1):73. https://doi.org/10.1186/s13054-017-1658-5
30. Canet E, Taylor DM, Khor R, Krishnan V, Bellomo R. qSOFA as predictor of mortality and prolonged ICU admission in Emergency Department patients with suspected infection. J Crit Care. 2018;48:118-123. https://doi.org/10.1016/j.jcrc.2018.08.022
31. Anand V, Zhang Z, Kadri SS, Klompas M, Rhee C; CDC Prevention Epicenters Program. Epidemiology of Quick Sequential Organ Failure Assessment criteria in undifferentiated patients and association with suspected infection and sepsis. Chest. 2019;156(2):289-297. https://doi.org/10.1016/j.chest.2019.03.032
32. Hamilton F, Arnold D, Baird A, Albur M, Whiting P. Early Warning Scores do not accurately predict mortality in sepsis: A meta-analysis and systematic review of the literature. J Infect. 2018;76(3):241-248. https://doi.org/10.1016/j.jinf.2018.01.002
33. Koch E, Lovett S, Nghiem T, Riggs RA, Rech MA. Shock Index in the emergency department: utility and limitations. Open Access Emerg Med. 2019;11:179-199. https://doi.org/10.2147/OAEM.S178358
34. Yussof SJ, Zakaria MI, Mohamed FL, Bujang MA, Lakshmanan S, Asaari AH. Value of Shock Index in prognosticating the short-term outcome of death for patients presenting with severe sepsis and septic shock in the emergency department. Med J Malaysia. 2012;67(4):406-411.
35. Siddiqui S, Chua M, Kumaresh V, Choo R. A comparison of pre ICU admission SIRS, EWS and q SOFA scores for predicting mortality and length of stay in ICU. J Crit Care. 2017;41:191-193. https://doi.org/10.1016/j.jcrc.2017.05.017
36. Costa RT, Nassar AP, Caruso P. Accuracy of SOFA, qSOFA, and SIRS scores for mortality in cancer patients admitted to an intensive care unit with suspected infection. J Crit Care. 2018;45:52-57. https://doi.org/10.1016/j.jcrc.2017.12.024
37. Mellhammar L, Linder A, Tverring J, et al. NEWS2 is Superior to qSOFA in detecting sepsis with organ dysfunction in the emergency department. J Clin Med. 2019;8(8):1128. https://doi.org/10.3390/jcm8081128
38. Szakmany T, Pugh R, Kopczynska M, et al. Defining sepsis on the wards: results of a multi-centre point-prevalence study comparing two sepsis definitions. Anaesthesia. 2018;73(2):195-204. https://doi.org/10.1111/anae.14062
39. Newman TB, Kohn MA. Evidence-Based Diagnosis: An Introduction to Clinical Epidemiology. Cambridge University Press; 2009.
40. Askim Å, Moser F, Gustad LT, et al. Poor performance of quick-SOFA (qSOFA) score in predicting severe sepsis and mortality - a prospective study of patients admitted with infection to the emergency department. Scand J Trauma Resusc Emerg Med. 2017;25(1):56. https://doi.org/10.1186/s13049-017-0399-4
Sepsis is the leading cause of in-hospital mortality in the United States.1 Sepsis is present on admission in 85% of cases, and each hour delay in antibiotic treatment is associated with 4% to 7% increased odds of mortality.2,3 Prompt identification and treatment of sepsis is essential for reducing morbidity and mortality, but identifying sepsis during triage is challenging.2
Risk stratification scores that rely solely on data readily available at the bedside have been developed to quickly identify those at greatest risk of poor outcomes from sepsis in real time. The quick Sequential Organ Failure Assessment (qSOFA) score, the National Early Warning System (NEWS2), and the Shock Index are easy-to-calculate measures that use routinely collected clinical data that are not subject to laboratory delay. These scores can be incorporated into electronic health record (EHR)-based alerts and can be calculated longitudinally to track the risk of poor outcomes over time. qSOFA was developed to quantify patient risk at bedside in non-intensive care unit (ICU) settings, but there is no consensus about its ability to predict adverse outcomes such as mortality and ICU admission.4-6 The United Kingdom’s National Health Service uses NEWS2 to identify patients at risk for sepsis.7 NEWS has been shown to have similar or better sensitivity in identifying poorer outcomes in sepsis patients compared with systemic inflammatory response syndrome (SIRS) criteria and qSOFA.4,8-11 However, since the latest update of NEWS2 in 2017, there has been little study of its predictive ability. The Shock Index is a simple bedside score (heart rate divided by systolic blood pressure) that was developed to detect changes in cardiovascular performance before systemic shock onset. Although it was not developed for infection and has not been regularly applied in the sepsis literature, the Shock Index might be useful for identifying patients at increased risk of poor outcomes. Patients with higher and sustained Shock Index scores are more likely to experience morbidity, such as hyperlactatemia, vasopressor use, and organ failure, and also have an increased risk of mortality.12-14
Although the predictive abilities of these bedside risk stratification scores have been assessed individually using standard binary cut-points, the comparative performance of qSOFA, the Shock Index, and NEWS2 has not been evaluated in patients presenting to an emergency department (ED) with suspected sepsis.
METHODS
Design and Setting
We conducted a retrospective cohort study of ED patients who presented with suspected sepsis to the University of California San Francisco (UCSF) Helen Diller Medical Center at Parnassus Heights between June 1, 2012, and December 31, 2018. Our institution is a 785-bed academic teaching hospital with approximately 30,000 ED encounters per year. The study was approved with a waiver of informed consent by the UCSF Human Research Protection Program.
Participants
We use an Epic-based EHR platform (Epic 2017, Epic Systems Corporation) for clinical care, which was implemented on June 1, 2012. All data elements were obtained from Clarity, the relational database that stores Epic’s inpatient data. The study included encounters for patients age ≥18 years who had blood cultures ordered within 24 hours of ED presentation and administration of intravenous antibiotics within 24 hours. Repeat encounters were treated independently in our analysis.
Outcomes and Measures
We compared the ability of qSOFA, the Shock Index, and NEWS2 to predict in-hospital mortality and admission to the ICU from the ED (ED-to-ICU admission). We used the
We compared demographic and clinical characteristics of patients who were positive for qSOFA, the Shock Index, and NEWS2. Demographic data were extracted from the EHR and included primary language, age, sex, and insurance status. All International Classification of Diseases (ICD)-9/10 diagnosis codes were pulled from Clarity billing tables. We used the Elixhauser comorbidity groupings19 of ICD-9/10 codes present on admission to identify preexisting comorbidities and underlying organ dysfunction. To estimate burden of comorbid illnesses, we calculated the validated van Walraven comorbidity index,20 which provides an estimated risk of in-hospital death based on documented Elixhauser comorbidities. Admission level of care (acute, stepdown, or intensive care) was collected for inpatient admissions to assess initial illness severity.21 We also evaluated discharge disposition and in-hospital mortality. Index blood culture results were collected, and dates and timestamps of mechanical ventilation, fluid, vasopressor, and antibiotic administration were obtained for the duration of the encounter.
UCSF uses an automated, real-time, algorithm-based severe sepsis alert that is triggered when a patient meets ≥2 SIRS criteria and again when the patient meets severe sepsis or septic shock criteria (ie, ≥2 SIRS criteria in addition to end-organ dysfunction and/or fluid nonresponsive hypotension). This sepsis screening alert was in use for the duration of our study.22
Statistical Analysis
We performed a subgroup analysis among those who were diagnosed with sepsis, according to the 2016 Third International Consensus Definitions for Sepsis and Septic Shock (Sepsis-3) criteria.
All statistical analyses were conducted using Stata 14 (StataCorp). We summarized differences in demographic and clinical characteristics among the populations meeting each severity score but elected not to conduct hypothesis testing because patients could be positive for one or more scores. We calculated sensitivity, specificity, positive predictive value (PPV), and negative predictive value (NPV) for each score to predict in-hospital mortality and ED-to-ICU admission. To allow comparison with other studies, we also created a composite outcome of either in-hospital mortality or ED-to-ICU admission.
RESULTS
Within our sample 23,837 ED patients had blood cultures ordered within 24 hours of ED presentation and were considered to have suspected sepsis. The mean age of the cohort was 60.8 years, and 1,612 (6.8%) had positive blood cultures. A total of 12,928 patients (54.2%) were found to have sepsis. We documented 1,427 in-hospital deaths (6.0%) and 3,149 (13.2%) ED-to-ICU admissions. At ED triage 1,921 (8.1%) were qSOFA-positive, 4,273 (17.9%) were Shock Index-positive, and 11,832 (49.6%) were NEWS2-positive. At ED triage, blood pressure, heart rate, respiratory rate, and oxygen saturated were documented in >99% of patients, 93.5% had temperature documented, and 28.5% had GCS recorded. If the window of assessment was widened to 1 hour, GCS was only documented among 44.2% of those with suspected sepsis.
Demographic Characteristics and Clinical Course
qSOFA-positive patients received antibiotics more quickly than those who were Shock Index-positive or NEWS2-positive (median 1.5, 1.8, and 2.8 hours after admission, respectively). In addition, those who were qSOFA-positive were more likely to have a positive blood culture (10.9%, 9.4%, and 8.5%, respectively) and to receive an EHR-based diagnosis of sepsis (77.0%, 69.6%, and 60.9%, respectively) than those who were Shock Index- or NEWS2-positive. Those who were qSOFA-positive also were more likely to be mechanically ventilated during their hospital stay (25.4%, 19.2%, and 10.8%, respectively) and to receive vasopressors (33.5%, 22.5%, and 12.2%, respectively). In-hospital mortality also was more common among those who were qSOFA-positive at triage (23.4%, 15.3%, and 9.2%, respectively).
Because both qSOFA and NEWS2 incorporate GCS, we explored baseline characteristics of patients with GCS documented at triage (n = 6,794). These patients were older (median age 63 and 61 years, P < .0001), more likely to be male (54.9% and 53.4%, P = .0031), more likely to have renal failure (22.8% and 20.1%, P < .0001), more likely to have liver disease (14.2% and 12.8%, P = .006), had a higher van Walraven comorbidity score on presentation (median 10 and 8, P < .0001), and were more likely to go directly to the ICU from the ED (20.2% and 10.6%, P < .0001). However, among the 6,397 GCS scores documented at triage, only 1,579 (24.7%) were abnormal.
Test Characteristics of qSOFA, Shock Index, and NEWS2 for Predicting In-hospital Mortality and ED-to-ICU Admission
Among 23,837 patients with suspected sepsis, NEWS2 had the highest sensitivity for predicting in-hospital mortality (76.0%; 95% CI, 73.7%-78.2%) and ED-to-ICU admission (78.9%; 95% CI, 77.5%-80.4%) but had the lowest specificity for in-hospital mortality (52.0%; 95% CI, 51.4%-52.7%) and for ED-to-ICU admission (54.8%; 95% CI, 54.1%-55.5%) (Table 3). qSOFA had the lowest sensitivity for in-hospital mortality (31.5%; 95% CI, 29.1%-33.9%) and ED-to-ICU admission (29.3%; 95% CI, 27.7%-30.9%) but the highest specificity for in-hospital mortality (93.4%; 95% CI, 93.1%-93.8%) and ED-to-ICU admission (95.2%; 95% CI, 94.9%-95.5%). The Shock Index had a sensitivity that fell between qSOFA and NEWS2 for in-hospital mortality (45.8%; 95% CI, 43.2%-48.5%) and ED-to-ICU admission (49.2%; 95% CI, 47.5%-51.0%). The specificity of the Shock Index also was between qSOFA and NEWS2 for in-hospital mortality (83.9%; 95% CI, 83.4%-84.3%) and ED-to-ICU admission (86.8%; 95% CI, 86.4%-87.3%). All three scores exhibited relatively low PPV, ranging from 9.2% to 23.4% for in-hospital mortality and 21.0% to 48.0% for ED-to-ICU triage. Conversely, all three scores exhibited relatively high NPV, ranging from 95.5% to 97.1% for in-hospital mortality and 89.8% to 94.5% for ED-to-ICU triage.
When considering a binary cutoff, the Shock Index exhibited the highest AUROC for in-hospital mortality (0.648; 95% CI, 0.635-0.662) and had a significantly higher AUROC than qSOFA (AUROC, 0.625; 95% CI, 0.612-0.637; P = .0005), but there was no difference compared with NEWS2 (AUROC, 0.640; 95% CI, 0.628-0.652; P = .2112). NEWS2 had a significantly higher AUROC than qSOFA for predicting in-hospital mortality (P = .0227). The Shock Index also exhibited the highest AUROC for ED-to-ICU admission (0.680; 95% CI, 0.617-0.689), which was significantly higher than the AUROC for qSOFA (P < .0001) and NEWS2 (P = 0.0151). NEWS2 had a significantly higher AUROC than qSOFA for predicting ED-to-ICU admission (P < .0001). Similar findings were seen in patients found to have sepsis.
DISCUSSION
In this retrospective cohort study of 23,837 patients who presented to the ED with suspected sepsis, the standard qSOFA threshold was met least frequently, followed by the Shock Index and NEWS2. NEWS2 had the highest sensitivity but the lowest specificity for predicting in-hospital mortality and ED-to-ICU admission, making it a challenging bedside risk stratification scale for identifying patients at risk of poor clinical outcomes. When comparing predictive performance among the three scales, qSOFA had the highest specificity and the Shock Index had the highest AUROC for in-hospital mortality and ED-to-ICU admission in this cohort of patients with suspected sepsis. These trends in sensitivity, specificity, and AUROC were consistent among those who met EHR criteria for a sepsis diagnosis. In the analysis of the three scoring systems using all available cut-points, qSOFA and NEWS2 had the highest AUROCs, followed by the Shock Index.
Considering the rapid progression from organ dysfunction to death in sepsis patients, as well as the difficulty establishing a sepsis diagnosis at triage,23 providers must quickly identify patients at increased risk of poor outcomes when they present to the ED. Sepsis alerts often are built using SIRS criteria,27 including the one used for sepsis surveillance at UCSF since 2012,22 but the white blood cell count criterion is subject to a laboratory lag and could lead to a delay in identification. Implementation of a point-of-care bedside score alert that uses readily available clinical data could allow providers to identify patients at greatest risk of poor outcomes immediately at ED presentation and triage, which motivated us to explore the predictive performance of qSOFA, the Shock Index, and NEWS2.
Our study is the first to provide a head-to-head comparison of the predictive performance of qSOFA, the Shock Index, and NEWS2, three easy-to-calculate bedside risk scores that use EHR data collected among patients with suspected sepsis. The Sepsis-3 guidelines recommend qSOFA to quickly identify non-ICU patients at greatest risk of poor outcomes because the measure exhibited predictive performance similar to the more extensive SOFA score outside the ICU.16,23 Although some studies have confirmed qSOFA’s high predictive performance,28-31 our test characteristics and AUROC findings are in line with other published analyses.4,6,10,17 The UK National Health Service is using NEWS2 to screen for patients at risk of poor outcomes from sepsis. Several analyses that assessed the predictive ability of NEWS have reported estimates in line with our findings.4,10,32 The Shock Index was introduced in 1967 and provided a metric to evaluate hemodynamic stability based on heart rate and systolic blood pressure.33 The Shock Index has been studied in several contexts, including sepsis,34 and studies show that a sustained Shock Index is associated with increased odds of vasopressor administration, higher prevalence of hyperlactatemia, and increased risk of poor outcomes in the ICU.13,14
For our study, we were particularly interested in exploring how the Shock Index would compare with more frequently used severity scores such as qSOFA and NEWS2 among patients with suspected sepsis, given the simplicity of its calculation and the easy availability of required data. In our cohort of 23,837 patients, only 159 people had missing blood pressure and only 71 had omitted heart rate. In contrast, both qSOFA and NEWS2 include an assessment of level of consciousness that can be subject to variability in assessment methods and EHR documentation across institutions.11 In our cohort, GCS within 30 minutes of ED presentation was missing in 72 patients, which could have led to incomplete calculation of qSOFA and NEWS2 if a missing value was not actually within normal limits.
Several investigations relate qSOFA to NEWS but few compare qSOFA with the newer NEWS2, and even fewer evaluate the Shock Index with any of these scores.10,11,18,29,35-37 In general, studies have shown that NEWS exhibits a higher AUROC for predicting mortality, sepsis with organ dysfunction, and ICU admission, often as a composite outcome.4,11,18,37,38 A handful of studies compare the Shock Index to SIRS; however, little has been done to compare the Shock Index to qSOFA or NEWS2, scores that have been used specifically for sepsis and might be more predictive of poor outcomes than SIRS.33 In our study, the Shock Index had a higher AUROC than either qSOFA or NEWS2 for predicting in-hospital mortality and ED-to-ICU admission measured as separate outcomes and as a composite outcome using standard cut-points for these scores.
When selecting a severity score to apply in an institution, it is important to carefully evaluate the score’s test characteristics, in addition to considering the availability of reliable data. Tests with high sensitivity and NPV for the population being studied can be useful to rule out disease or risk of poor outcome, while tests with high specificity and PPV can be useful to rule in disease or risk of poor outcome.39 When considering specificity, qSOFA’s performance was superior to the Shock Index and NEWS2 in our study, but a small percentage of the population was identified using a cut-point of qSOFA ≥2. If we used qSOFA and applied this standard cut-point at our institution, we could be confident that those identified were at increased risk, but we would miss a significant number of patients who would experience a poor outcome. When considering sensitivity, performance of NEWS2 was superior to qSOFA and the Shock Index in our study, but one-half of the population was identified using a cut-point of NEWS2 ≥5. If we were to apply this standard NEWS2 cut-point at our institution, we would assume that one-half of our population was at risk, which might drive resource use towards patients who will not experience a poor outcome. Although none of the scores exhibited a robust AUROC measure, the Shock Index had the highest AUROC for in-hospital mortality and ED-to-ICU admission when using the standard binary cut-point, and its sensitivity and specificity is between that of qSOFA and NEWS2, potentially making it a score to use in settings where qSOFA and NEWS2 score components, such as altered mentation, are not reliably collected. Finally, our sensitivity analysis varying the binary cut-point of each score within our population demonstrated that the standard cut-points might not be as useful within a specific population and might need to be tailored for implementation, balancing sensitivity, specificity, PPV, and NPV to meet local priorities and ICU capacity.
Our study has limitations. It is a single-center, retrospective analysis, factors that could reduce generalizability. However, it does include a large and diverse patient population spanning several years. Missing GCS data could have affected the predictive ability of qSOFA and NEWS2 in our cohort. We could not reliably perform imputation of GCS because of the high missingness and therefore we assumed missing was normal, as was done in the Sepsis-3 derivation studies.16 Previous studies have attempted to impute GCS and have not observed improved performance of qSOFA to predict mortality.40 Because manually collected variables such as GCS are less reliably documented in the EHR, there might be limitations in their use for triage risk scores.
Although the current analysis focused on the predictive performance of qSOFA, the Shock Index, and NEWS2 at triage, performance of these scores could affect the ED team’s treatment decisions before handoff to the hospitalist team and the expected level of care the patient will receive after in-patient admission. These tests also have the advantage of being easy to calculate at the bedside over time, which could provide an objective assessment of longitudinal predicted prognosis.
CONCLUSION
Local priorities should drive selection of a screening tool, balancing sensitivity, specificity, PPV, and NPV to achieve the institution’s goals. qSOFA, Shock Index, and NEWS2 are risk stratification tools that can be easily implemented at ED triage using data available at the bedside. Although none of these scores performed strongly when comparing AUROCs, qSOFA was highly specific for identifying patients with poor outcomes, and NEWS2 was the most sensitive for ruling out those at high risk among patients with suspected sepsis. The Shock Index exhibited a sensitivity and specificity that fell between qSOFA and NEWS2 and also might be considered to identify those at increased risk, given its ease of implementation, particularly in settings where altered mentation is unreliably or inconsistently documented.
Acknowledgment
The authors thank the UCSF Division of Hospital Medicine Data Core for their assistance with data acquisition.
Sepsis is the leading cause of in-hospital mortality in the United States.1 Sepsis is present on admission in 85% of cases, and each hour delay in antibiotic treatment is associated with 4% to 7% increased odds of mortality.2,3 Prompt identification and treatment of sepsis is essential for reducing morbidity and mortality, but identifying sepsis during triage is challenging.2
Risk stratification scores that rely solely on data readily available at the bedside have been developed to quickly identify those at greatest risk of poor outcomes from sepsis in real time. The quick Sequential Organ Failure Assessment (qSOFA) score, the National Early Warning System (NEWS2), and the Shock Index are easy-to-calculate measures that use routinely collected clinical data that are not subject to laboratory delay. These scores can be incorporated into electronic health record (EHR)-based alerts and can be calculated longitudinally to track the risk of poor outcomes over time. qSOFA was developed to quantify patient risk at bedside in non-intensive care unit (ICU) settings, but there is no consensus about its ability to predict adverse outcomes such as mortality and ICU admission.4-6 The United Kingdom’s National Health Service uses NEWS2 to identify patients at risk for sepsis.7 NEWS has been shown to have similar or better sensitivity in identifying poorer outcomes in sepsis patients compared with systemic inflammatory response syndrome (SIRS) criteria and qSOFA.4,8-11 However, since the latest update of NEWS2 in 2017, there has been little study of its predictive ability. The Shock Index is a simple bedside score (heart rate divided by systolic blood pressure) that was developed to detect changes in cardiovascular performance before systemic shock onset. Although it was not developed for infection and has not been regularly applied in the sepsis literature, the Shock Index might be useful for identifying patients at increased risk of poor outcomes. Patients with higher and sustained Shock Index scores are more likely to experience morbidity, such as hyperlactatemia, vasopressor use, and organ failure, and also have an increased risk of mortality.12-14
Although the predictive abilities of these bedside risk stratification scores have been assessed individually using standard binary cut-points, the comparative performance of qSOFA, the Shock Index, and NEWS2 has not been evaluated in patients presenting to an emergency department (ED) with suspected sepsis.
METHODS
Design and Setting
We conducted a retrospective cohort study of ED patients who presented with suspected sepsis to the University of California San Francisco (UCSF) Helen Diller Medical Center at Parnassus Heights between June 1, 2012, and December 31, 2018. Our institution is a 785-bed academic teaching hospital with approximately 30,000 ED encounters per year. The study was approved with a waiver of informed consent by the UCSF Human Research Protection Program.
Participants
We use an Epic-based EHR platform (Epic 2017, Epic Systems Corporation) for clinical care, which was implemented on June 1, 2012. All data elements were obtained from Clarity, the relational database that stores Epic’s inpatient data. The study included encounters for patients age ≥18 years who had blood cultures ordered within 24 hours of ED presentation and administration of intravenous antibiotics within 24 hours. Repeat encounters were treated independently in our analysis.
Outcomes and Measures
We compared the ability of qSOFA, the Shock Index, and NEWS2 to predict in-hospital mortality and admission to the ICU from the ED (ED-to-ICU admission). We used the
We compared demographic and clinical characteristics of patients who were positive for qSOFA, the Shock Index, and NEWS2. Demographic data were extracted from the EHR and included primary language, age, sex, and insurance status. All International Classification of Diseases (ICD)-9/10 diagnosis codes were pulled from Clarity billing tables. We used the Elixhauser comorbidity groupings19 of ICD-9/10 codes present on admission to identify preexisting comorbidities and underlying organ dysfunction. To estimate burden of comorbid illnesses, we calculated the validated van Walraven comorbidity index,20 which provides an estimated risk of in-hospital death based on documented Elixhauser comorbidities. Admission level of care (acute, stepdown, or intensive care) was collected for inpatient admissions to assess initial illness severity.21 We also evaluated discharge disposition and in-hospital mortality. Index blood culture results were collected, and dates and timestamps of mechanical ventilation, fluid, vasopressor, and antibiotic administration were obtained for the duration of the encounter.
UCSF uses an automated, real-time, algorithm-based severe sepsis alert that is triggered when a patient meets ≥2 SIRS criteria and again when the patient meets severe sepsis or septic shock criteria (ie, ≥2 SIRS criteria in addition to end-organ dysfunction and/or fluid nonresponsive hypotension). This sepsis screening alert was in use for the duration of our study.22
Statistical Analysis
We performed a subgroup analysis among those who were diagnosed with sepsis, according to the 2016 Third International Consensus Definitions for Sepsis and Septic Shock (Sepsis-3) criteria.
All statistical analyses were conducted using Stata 14 (StataCorp). We summarized differences in demographic and clinical characteristics among the populations meeting each severity score but elected not to conduct hypothesis testing because patients could be positive for one or more scores. We calculated sensitivity, specificity, positive predictive value (PPV), and negative predictive value (NPV) for each score to predict in-hospital mortality and ED-to-ICU admission. To allow comparison with other studies, we also created a composite outcome of either in-hospital mortality or ED-to-ICU admission.
RESULTS
Within our sample 23,837 ED patients had blood cultures ordered within 24 hours of ED presentation and were considered to have suspected sepsis. The mean age of the cohort was 60.8 years, and 1,612 (6.8%) had positive blood cultures. A total of 12,928 patients (54.2%) were found to have sepsis. We documented 1,427 in-hospital deaths (6.0%) and 3,149 (13.2%) ED-to-ICU admissions. At ED triage 1,921 (8.1%) were qSOFA-positive, 4,273 (17.9%) were Shock Index-positive, and 11,832 (49.6%) were NEWS2-positive. At ED triage, blood pressure, heart rate, respiratory rate, and oxygen saturated were documented in >99% of patients, 93.5% had temperature documented, and 28.5% had GCS recorded. If the window of assessment was widened to 1 hour, GCS was only documented among 44.2% of those with suspected sepsis.
Demographic Characteristics and Clinical Course
qSOFA-positive patients received antibiotics more quickly than those who were Shock Index-positive or NEWS2-positive (median 1.5, 1.8, and 2.8 hours after admission, respectively). In addition, those who were qSOFA-positive were more likely to have a positive blood culture (10.9%, 9.4%, and 8.5%, respectively) and to receive an EHR-based diagnosis of sepsis (77.0%, 69.6%, and 60.9%, respectively) than those who were Shock Index- or NEWS2-positive. Those who were qSOFA-positive also were more likely to be mechanically ventilated during their hospital stay (25.4%, 19.2%, and 10.8%, respectively) and to receive vasopressors (33.5%, 22.5%, and 12.2%, respectively). In-hospital mortality also was more common among those who were qSOFA-positive at triage (23.4%, 15.3%, and 9.2%, respectively).
Because both qSOFA and NEWS2 incorporate GCS, we explored baseline characteristics of patients with GCS documented at triage (n = 6,794). These patients were older (median age 63 and 61 years, P < .0001), more likely to be male (54.9% and 53.4%, P = .0031), more likely to have renal failure (22.8% and 20.1%, P < .0001), more likely to have liver disease (14.2% and 12.8%, P = .006), had a higher van Walraven comorbidity score on presentation (median 10 and 8, P < .0001), and were more likely to go directly to the ICU from the ED (20.2% and 10.6%, P < .0001). However, among the 6,397 GCS scores documented at triage, only 1,579 (24.7%) were abnormal.
Test Characteristics of qSOFA, Shock Index, and NEWS2 for Predicting In-hospital Mortality and ED-to-ICU Admission
Among 23,837 patients with suspected sepsis, NEWS2 had the highest sensitivity for predicting in-hospital mortality (76.0%; 95% CI, 73.7%-78.2%) and ED-to-ICU admission (78.9%; 95% CI, 77.5%-80.4%) but had the lowest specificity for in-hospital mortality (52.0%; 95% CI, 51.4%-52.7%) and for ED-to-ICU admission (54.8%; 95% CI, 54.1%-55.5%) (Table 3). qSOFA had the lowest sensitivity for in-hospital mortality (31.5%; 95% CI, 29.1%-33.9%) and ED-to-ICU admission (29.3%; 95% CI, 27.7%-30.9%) but the highest specificity for in-hospital mortality (93.4%; 95% CI, 93.1%-93.8%) and ED-to-ICU admission (95.2%; 95% CI, 94.9%-95.5%). The Shock Index had a sensitivity that fell between qSOFA and NEWS2 for in-hospital mortality (45.8%; 95% CI, 43.2%-48.5%) and ED-to-ICU admission (49.2%; 95% CI, 47.5%-51.0%). The specificity of the Shock Index also was between qSOFA and NEWS2 for in-hospital mortality (83.9%; 95% CI, 83.4%-84.3%) and ED-to-ICU admission (86.8%; 95% CI, 86.4%-87.3%). All three scores exhibited relatively low PPV, ranging from 9.2% to 23.4% for in-hospital mortality and 21.0% to 48.0% for ED-to-ICU triage. Conversely, all three scores exhibited relatively high NPV, ranging from 95.5% to 97.1% for in-hospital mortality and 89.8% to 94.5% for ED-to-ICU triage.
When considering a binary cutoff, the Shock Index exhibited the highest AUROC for in-hospital mortality (0.648; 95% CI, 0.635-0.662) and had a significantly higher AUROC than qSOFA (AUROC, 0.625; 95% CI, 0.612-0.637; P = .0005), but there was no difference compared with NEWS2 (AUROC, 0.640; 95% CI, 0.628-0.652; P = .2112). NEWS2 had a significantly higher AUROC than qSOFA for predicting in-hospital mortality (P = .0227). The Shock Index also exhibited the highest AUROC for ED-to-ICU admission (0.680; 95% CI, 0.617-0.689), which was significantly higher than the AUROC for qSOFA (P < .0001) and NEWS2 (P = 0.0151). NEWS2 had a significantly higher AUROC than qSOFA for predicting ED-to-ICU admission (P < .0001). Similar findings were seen in patients found to have sepsis.
DISCUSSION
In this retrospective cohort study of 23,837 patients who presented to the ED with suspected sepsis, the standard qSOFA threshold was met least frequently, followed by the Shock Index and NEWS2. NEWS2 had the highest sensitivity but the lowest specificity for predicting in-hospital mortality and ED-to-ICU admission, making it a challenging bedside risk stratification scale for identifying patients at risk of poor clinical outcomes. When comparing predictive performance among the three scales, qSOFA had the highest specificity and the Shock Index had the highest AUROC for in-hospital mortality and ED-to-ICU admission in this cohort of patients with suspected sepsis. These trends in sensitivity, specificity, and AUROC were consistent among those who met EHR criteria for a sepsis diagnosis. In the analysis of the three scoring systems using all available cut-points, qSOFA and NEWS2 had the highest AUROCs, followed by the Shock Index.
Considering the rapid progression from organ dysfunction to death in sepsis patients, as well as the difficulty establishing a sepsis diagnosis at triage,23 providers must quickly identify patients at increased risk of poor outcomes when they present to the ED. Sepsis alerts often are built using SIRS criteria,27 including the one used for sepsis surveillance at UCSF since 2012,22 but the white blood cell count criterion is subject to a laboratory lag and could lead to a delay in identification. Implementation of a point-of-care bedside score alert that uses readily available clinical data could allow providers to identify patients at greatest risk of poor outcomes immediately at ED presentation and triage, which motivated us to explore the predictive performance of qSOFA, the Shock Index, and NEWS2.
Our study is the first to provide a head-to-head comparison of the predictive performance of qSOFA, the Shock Index, and NEWS2, three easy-to-calculate bedside risk scores that use EHR data collected among patients with suspected sepsis. The Sepsis-3 guidelines recommend qSOFA to quickly identify non-ICU patients at greatest risk of poor outcomes because the measure exhibited predictive performance similar to the more extensive SOFA score outside the ICU.16,23 Although some studies have confirmed qSOFA’s high predictive performance,28-31 our test characteristics and AUROC findings are in line with other published analyses.4,6,10,17 The UK National Health Service is using NEWS2 to screen for patients at risk of poor outcomes from sepsis. Several analyses that assessed the predictive ability of NEWS have reported estimates in line with our findings.4,10,32 The Shock Index was introduced in 1967 and provided a metric to evaluate hemodynamic stability based on heart rate and systolic blood pressure.33 The Shock Index has been studied in several contexts, including sepsis,34 and studies show that a sustained Shock Index is associated with increased odds of vasopressor administration, higher prevalence of hyperlactatemia, and increased risk of poor outcomes in the ICU.13,14
For our study, we were particularly interested in exploring how the Shock Index would compare with more frequently used severity scores such as qSOFA and NEWS2 among patients with suspected sepsis, given the simplicity of its calculation and the easy availability of required data. In our cohort of 23,837 patients, only 159 people had missing blood pressure and only 71 had omitted heart rate. In contrast, both qSOFA and NEWS2 include an assessment of level of consciousness that can be subject to variability in assessment methods and EHR documentation across institutions.11 In our cohort, GCS within 30 minutes of ED presentation was missing in 72 patients, which could have led to incomplete calculation of qSOFA and NEWS2 if a missing value was not actually within normal limits.
Several investigations relate qSOFA to NEWS but few compare qSOFA with the newer NEWS2, and even fewer evaluate the Shock Index with any of these scores.10,11,18,29,35-37 In general, studies have shown that NEWS exhibits a higher AUROC for predicting mortality, sepsis with organ dysfunction, and ICU admission, often as a composite outcome.4,11,18,37,38 A handful of studies compare the Shock Index to SIRS; however, little has been done to compare the Shock Index to qSOFA or NEWS2, scores that have been used specifically for sepsis and might be more predictive of poor outcomes than SIRS.33 In our study, the Shock Index had a higher AUROC than either qSOFA or NEWS2 for predicting in-hospital mortality and ED-to-ICU admission measured as separate outcomes and as a composite outcome using standard cut-points for these scores.
When selecting a severity score to apply in an institution, it is important to carefully evaluate the score’s test characteristics, in addition to considering the availability of reliable data. Tests with high sensitivity and NPV for the population being studied can be useful to rule out disease or risk of poor outcome, while tests with high specificity and PPV can be useful to rule in disease or risk of poor outcome.39 When considering specificity, qSOFA’s performance was superior to the Shock Index and NEWS2 in our study, but a small percentage of the population was identified using a cut-point of qSOFA ≥2. If we used qSOFA and applied this standard cut-point at our institution, we could be confident that those identified were at increased risk, but we would miss a significant number of patients who would experience a poor outcome. When considering sensitivity, performance of NEWS2 was superior to qSOFA and the Shock Index in our study, but one-half of the population was identified using a cut-point of NEWS2 ≥5. If we were to apply this standard NEWS2 cut-point at our institution, we would assume that one-half of our population was at risk, which might drive resource use towards patients who will not experience a poor outcome. Although none of the scores exhibited a robust AUROC measure, the Shock Index had the highest AUROC for in-hospital mortality and ED-to-ICU admission when using the standard binary cut-point, and its sensitivity and specificity is between that of qSOFA and NEWS2, potentially making it a score to use in settings where qSOFA and NEWS2 score components, such as altered mentation, are not reliably collected. Finally, our sensitivity analysis varying the binary cut-point of each score within our population demonstrated that the standard cut-points might not be as useful within a specific population and might need to be tailored for implementation, balancing sensitivity, specificity, PPV, and NPV to meet local priorities and ICU capacity.
Our study has limitations. It is a single-center, retrospective analysis, factors that could reduce generalizability. However, it does include a large and diverse patient population spanning several years. Missing GCS data could have affected the predictive ability of qSOFA and NEWS2 in our cohort. We could not reliably perform imputation of GCS because of the high missingness and therefore we assumed missing was normal, as was done in the Sepsis-3 derivation studies.16 Previous studies have attempted to impute GCS and have not observed improved performance of qSOFA to predict mortality.40 Because manually collected variables such as GCS are less reliably documented in the EHR, there might be limitations in their use for triage risk scores.
Although the current analysis focused on the predictive performance of qSOFA, the Shock Index, and NEWS2 at triage, performance of these scores could affect the ED team’s treatment decisions before handoff to the hospitalist team and the expected level of care the patient will receive after in-patient admission. These tests also have the advantage of being easy to calculate at the bedside over time, which could provide an objective assessment of longitudinal predicted prognosis.
CONCLUSION
Local priorities should drive selection of a screening tool, balancing sensitivity, specificity, PPV, and NPV to achieve the institution’s goals. qSOFA, Shock Index, and NEWS2 are risk stratification tools that can be easily implemented at ED triage using data available at the bedside. Although none of these scores performed strongly when comparing AUROCs, qSOFA was highly specific for identifying patients with poor outcomes, and NEWS2 was the most sensitive for ruling out those at high risk among patients with suspected sepsis. The Shock Index exhibited a sensitivity and specificity that fell between qSOFA and NEWS2 and also might be considered to identify those at increased risk, given its ease of implementation, particularly in settings where altered mentation is unreliably or inconsistently documented.
Acknowledgment
The authors thank the UCSF Division of Hospital Medicine Data Core for their assistance with data acquisition.
1. Jones SL, Ashton CM, Kiehne LB, et al. Outcomes and resource use of sepsis-associated stays by presence on admission, severity, and hospital type. Med Care. 2016;54(3):303-310. https://doi.org/10.1097/MLR.0000000000000481
2. Seymour CW, Gesten F, Prescott HC, et al. Time to treatment and mortality during mandated emergency care for sepsis. N Engl J Med. 2017;376(23):2235-2244. https://doi.org/10.1056/NEJMoa1703058
3. Kumar A, Roberts D, Wood KE, et al. Duration of hypotension before initiation of effective antimicrobial therapy is the critical determinant of survival in human septic shock. Crit Care Med. 2006;34(6):1589-1596. https://doi.org/10.1097/01.CCM.0000217961.75225.E9
4. Churpek MM, Snyder A, Sokol S, Pettit NN, Edelson DP. Investigating the impact of different suspicion of infection criteria on the accuracy of Quick Sepsis-Related Organ Failure Assessment, Systemic Inflammatory Response Syndrome, and Early Warning Scores. Crit Care Med. 2017;45(11):1805-1812. https://doi.org/10.1097/CCM.0000000000002648
5. Abdullah SMOB, Sørensen RH, Dessau RBC, Sattar SMRU, Wiese L, Nielsen FE. Prognostic accuracy of qSOFA in predicting 28-day mortality among infected patients in an emergency department: a prospective validation study. Emerg Med J. 2019;36(12):722-728. https://doi.org/10.1136/emermed-2019-208456
6. Kim KS, Suh GJ, Kim K, et al. Quick Sepsis-related Organ Failure Assessment score is not sensitive enough to predict 28-day mortality in emergency department patients with sepsis: a retrospective review. Clin Exp Emerg Med. 2019;6(1):77-83. HTTPS://DOI.ORG/ 10.15441/ceem.17.294
7. National Early Warning Score (NEWS) 2: Standardising the assessment of acute-illness severity in the NHS. Royal College of Physicians; 2017.
8. Brink A, Alsma J, Verdonschot RJCG, et al. Predicting mortality in patients with suspected sepsis at the emergency department: a retrospective cohort study comparing qSOFA, SIRS and National Early Warning Score. PLoS One. 2019;14(1):e0211133. https://doi.org/ 10.1371/journal.pone.0211133
9. Redfern OC, Smith GB, Prytherch DR, Meredith P, Inada-Kim M, Schmidt PE. A comparison of the Quick Sequential (Sepsis-Related) Organ Failure Assessment Score and the National Early Warning Score in non-ICU patients with/without infection. Crit Care Med. 2018;46(12):1923-1933. https://doi.org/10.1097/CCM.0000000000003359
10. Churpek MM, Snyder A, Han X, et al. Quick Sepsis-related Organ Failure Assessment, Systemic Inflammatory Response Syndrome, and Early Warning Scores for detecting clinical deterioration in infected patients outside the intensive care unit. Am J Respir Crit Care Med. 2017;195(7):906-911. https://doi.org/10.1164/rccm.201604-0854OC
11. Goulden R, Hoyle MC, Monis J, et al. qSOFA, SIRS and NEWS for predicting inhospital mortality and ICU admission in emergency admissions treated as sepsis. Emerg Med J. 2018;35(6):345-349. https://doi.org/10.1136/emermed-2017-207120
12. Biney I, Shepherd A, Thomas J, Mehari A. Shock Index and outcomes in patients admitted to the ICU with sepsis. Chest. 2015;148(suppl 4):337A. https://doi.org/https://doi.org/10.1378/chest.2281151
13. Wira CR, Francis MW, Bhat S, Ehrman R, Conner D, Siegel M. The shock index as a predictor of vasopressor use in emergency department patients with severe sepsis. West J Emerg Med. 2014;15(1):60-66. https://doi.org/10.5811/westjem.2013.7.18472
14. Berger T, Green J, Horeczko T, et al. Shock index and early recognition of sepsis in the emergency department: pilot study. West J Emerg Med. 2013;14(2):168-174. https://doi.org/10.5811/westjem.2012.8.11546
15. Middleton DJ, Smith TO, Bedford R, Neilly M, Myint PK. Shock Index predicts outcome in patients with suspected sepsis or community-acquired pneumonia: a systematic review. J Clin Med. 2019;8(8):1144. https://doi.org/10.3390/jcm8081144
16. Seymour CW, Liu VX, Iwashyna TJ, et al. Assessment of clinical criteria for sepsis: for the Third International Consensus Definitions for Sepsis and Septic Shock (Sepsis-3). JAMA. 2016;315(8):762-774. https://doi.org/ 10.1001/jama.2016.0288
17. Abdullah S, Sørensen RH, Dessau RBC, Sattar S, Wiese L, Nielsen FE. Prognostic accuracy of qSOFA in predicting 28-day mortality among infected patients in an emergency department: a prospective validation study. Emerg Med J. 2019;36(12):722-728. https://doi.org/10.1136/emermed-2019-208456
18. Usman OA, Usman AA, Ward MA. Comparison of SIRS, qSOFA, and NEWS for the early identification of sepsis in the Emergency Department. Am J Emerg Med. 2018;37(8):1490-1497. https://doi.org/10.1016/j.ajem.2018.10.058
19. Elixhauser A, Steiner C, Harris DR, Coffey RM. Comorbidity measures for use with administrative data. Med Care. 1998;36(1):8-27. https://doi.org/10.1097/00005650-199801000-00004
20. van Walraven C, Austin PC, Jennings A, Quan H, Forster AJ. A modification of the Elixhauser comorbidity measures into a point system for hospital death using administrative data. Med Care. 2009;47(6):626-633. https://doi.org/10.1097/MLR.0b013e31819432e5
21. Prin M, Wunsch H. The role of stepdown beds in hospital care. Am J Respir Crit Care Med. 2014;190(11):1210-1216. https://doi.org/10.1164/rccm.201406-1117PP
22. Narayanan N, Gross AK, Pintens M, Fee C, MacDougall C. Effect of an electronic medical record alert for severe sepsis among ED patients. Am J Emerg Med. 2016;34(2):185-188. https://doi.org/10.1016/j.ajem.2015.10.005
23. Singer M, Deutschman CS, Seymour CW, et al. The Third International Consensus Definitions for Sepsis and Septic Shock (Sepsis-3). JAMA. 2016;315(8):801-810. https://doi.org/10.1001/jama.2016.0287
24. Rhee C, Dantes R, Epstein L, et al. Incidence and trends of sepsis in US hospitals using clinical vs claims data, 2009-2014. JAMA. 2017;318(13):1241-1249. https://doi.org/10.1001/jama.2017.13836
25. Safari S, Baratloo A, Elfil M, Negida A. Evidence based emergency medicine; part 5 receiver operating curve and area under the curve. Emerg (Tehran). 2016;4(2):111-113.
26. DeLong ER, DeLong DM, Clarke-Pearson DL. Comparing the areas under two or more correlated receiver operating characteristic curves: a nonparametric approach. Biometrics. 1988;44(3):837-845.
27. Kangas C, Iverson L, Pierce D. Sepsis screening: combining Early Warning Scores and SIRS Criteria. Clin Nurs Res. 2021;30(1):42-49. https://doi.org/10.1177/1054773818823334.
28. Freund Y, Lemachatti N, Krastinova E, et al. Prognostic accuracy of Sepsis-3 Criteria for in-hospital mortality among patients with suspected infection presenting to the emergency department. JAMA. 2017;317(3):301-308. https://doi.org/10.1001/jama.2016.20329
29. Finkelsztein EJ, Jones DS, Ma KC, et al. Comparison of qSOFA and SIRS for predicting adverse outcomes of patients with suspicion of sepsis outside the intensive care unit. Crit Care. 2017;21(1):73. https://doi.org/10.1186/s13054-017-1658-5
30. Canet E, Taylor DM, Khor R, Krishnan V, Bellomo R. qSOFA as predictor of mortality and prolonged ICU admission in Emergency Department patients with suspected infection. J Crit Care. 2018;48:118-123. https://doi.org/10.1016/j.jcrc.2018.08.022
31. Anand V, Zhang Z, Kadri SS, Klompas M, Rhee C; CDC Prevention Epicenters Program. Epidemiology of Quick Sequential Organ Failure Assessment criteria in undifferentiated patients and association with suspected infection and sepsis. Chest. 2019;156(2):289-297. https://doi.org/10.1016/j.chest.2019.03.032
32. Hamilton F, Arnold D, Baird A, Albur M, Whiting P. Early Warning Scores do not accurately predict mortality in sepsis: A meta-analysis and systematic review of the literature. J Infect. 2018;76(3):241-248. https://doi.org/10.1016/j.jinf.2018.01.002
33. Koch E, Lovett S, Nghiem T, Riggs RA, Rech MA. Shock Index in the emergency department: utility and limitations. Open Access Emerg Med. 2019;11:179-199. https://doi.org/10.2147/OAEM.S178358
34. Yussof SJ, Zakaria MI, Mohamed FL, Bujang MA, Lakshmanan S, Asaari AH. Value of Shock Index in prognosticating the short-term outcome of death for patients presenting with severe sepsis and septic shock in the emergency department. Med J Malaysia. 2012;67(4):406-411.
35. Siddiqui S, Chua M, Kumaresh V, Choo R. A comparison of pre ICU admission SIRS, EWS and q SOFA scores for predicting mortality and length of stay in ICU. J Crit Care. 2017;41:191-193. https://doi.org/10.1016/j.jcrc.2017.05.017
36. Costa RT, Nassar AP, Caruso P. Accuracy of SOFA, qSOFA, and SIRS scores for mortality in cancer patients admitted to an intensive care unit with suspected infection. J Crit Care. 2018;45:52-57. https://doi.org/10.1016/j.jcrc.2017.12.024
37. Mellhammar L, Linder A, Tverring J, et al. NEWS2 is Superior to qSOFA in detecting sepsis with organ dysfunction in the emergency department. J Clin Med. 2019;8(8):1128. https://doi.org/10.3390/jcm8081128
38. Szakmany T, Pugh R, Kopczynska M, et al. Defining sepsis on the wards: results of a multi-centre point-prevalence study comparing two sepsis definitions. Anaesthesia. 2018;73(2):195-204. https://doi.org/10.1111/anae.14062
39. Newman TB, Kohn MA. Evidence-Based Diagnosis: An Introduction to Clinical Epidemiology. Cambridge University Press; 2009.
40. Askim Å, Moser F, Gustad LT, et al. Poor performance of quick-SOFA (qSOFA) score in predicting severe sepsis and mortality - a prospective study of patients admitted with infection to the emergency department. Scand J Trauma Resusc Emerg Med. 2017;25(1):56. https://doi.org/10.1186/s13049-017-0399-4
1. Jones SL, Ashton CM, Kiehne LB, et al. Outcomes and resource use of sepsis-associated stays by presence on admission, severity, and hospital type. Med Care. 2016;54(3):303-310. https://doi.org/10.1097/MLR.0000000000000481
2. Seymour CW, Gesten F, Prescott HC, et al. Time to treatment and mortality during mandated emergency care for sepsis. N Engl J Med. 2017;376(23):2235-2244. https://doi.org/10.1056/NEJMoa1703058
3. Kumar A, Roberts D, Wood KE, et al. Duration of hypotension before initiation of effective antimicrobial therapy is the critical determinant of survival in human septic shock. Crit Care Med. 2006;34(6):1589-1596. https://doi.org/10.1097/01.CCM.0000217961.75225.E9
4. Churpek MM, Snyder A, Sokol S, Pettit NN, Edelson DP. Investigating the impact of different suspicion of infection criteria on the accuracy of Quick Sepsis-Related Organ Failure Assessment, Systemic Inflammatory Response Syndrome, and Early Warning Scores. Crit Care Med. 2017;45(11):1805-1812. https://doi.org/10.1097/CCM.0000000000002648
5. Abdullah SMOB, Sørensen RH, Dessau RBC, Sattar SMRU, Wiese L, Nielsen FE. Prognostic accuracy of qSOFA in predicting 28-day mortality among infected patients in an emergency department: a prospective validation study. Emerg Med J. 2019;36(12):722-728. https://doi.org/10.1136/emermed-2019-208456
6. Kim KS, Suh GJ, Kim K, et al. Quick Sepsis-related Organ Failure Assessment score is not sensitive enough to predict 28-day mortality in emergency department patients with sepsis: a retrospective review. Clin Exp Emerg Med. 2019;6(1):77-83. HTTPS://DOI.ORG/ 10.15441/ceem.17.294
7. National Early Warning Score (NEWS) 2: Standardising the assessment of acute-illness severity in the NHS. Royal College of Physicians; 2017.
8. Brink A, Alsma J, Verdonschot RJCG, et al. Predicting mortality in patients with suspected sepsis at the emergency department: a retrospective cohort study comparing qSOFA, SIRS and National Early Warning Score. PLoS One. 2019;14(1):e0211133. https://doi.org/ 10.1371/journal.pone.0211133
9. Redfern OC, Smith GB, Prytherch DR, Meredith P, Inada-Kim M, Schmidt PE. A comparison of the Quick Sequential (Sepsis-Related) Organ Failure Assessment Score and the National Early Warning Score in non-ICU patients with/without infection. Crit Care Med. 2018;46(12):1923-1933. https://doi.org/10.1097/CCM.0000000000003359
10. Churpek MM, Snyder A, Han X, et al. Quick Sepsis-related Organ Failure Assessment, Systemic Inflammatory Response Syndrome, and Early Warning Scores for detecting clinical deterioration in infected patients outside the intensive care unit. Am J Respir Crit Care Med. 2017;195(7):906-911. https://doi.org/10.1164/rccm.201604-0854OC
11. Goulden R, Hoyle MC, Monis J, et al. qSOFA, SIRS and NEWS for predicting inhospital mortality and ICU admission in emergency admissions treated as sepsis. Emerg Med J. 2018;35(6):345-349. https://doi.org/10.1136/emermed-2017-207120
12. Biney I, Shepherd A, Thomas J, Mehari A. Shock Index and outcomes in patients admitted to the ICU with sepsis. Chest. 2015;148(suppl 4):337A. https://doi.org/https://doi.org/10.1378/chest.2281151
13. Wira CR, Francis MW, Bhat S, Ehrman R, Conner D, Siegel M. The shock index as a predictor of vasopressor use in emergency department patients with severe sepsis. West J Emerg Med. 2014;15(1):60-66. https://doi.org/10.5811/westjem.2013.7.18472
14. Berger T, Green J, Horeczko T, et al. Shock index and early recognition of sepsis in the emergency department: pilot study. West J Emerg Med. 2013;14(2):168-174. https://doi.org/10.5811/westjem.2012.8.11546
15. Middleton DJ, Smith TO, Bedford R, Neilly M, Myint PK. Shock Index predicts outcome in patients with suspected sepsis or community-acquired pneumonia: a systematic review. J Clin Med. 2019;8(8):1144. https://doi.org/10.3390/jcm8081144
16. Seymour CW, Liu VX, Iwashyna TJ, et al. Assessment of clinical criteria for sepsis: for the Third International Consensus Definitions for Sepsis and Septic Shock (Sepsis-3). JAMA. 2016;315(8):762-774. https://doi.org/ 10.1001/jama.2016.0288
17. Abdullah S, Sørensen RH, Dessau RBC, Sattar S, Wiese L, Nielsen FE. Prognostic accuracy of qSOFA in predicting 28-day mortality among infected patients in an emergency department: a prospective validation study. Emerg Med J. 2019;36(12):722-728. https://doi.org/10.1136/emermed-2019-208456
18. Usman OA, Usman AA, Ward MA. Comparison of SIRS, qSOFA, and NEWS for the early identification of sepsis in the Emergency Department. Am J Emerg Med. 2018;37(8):1490-1497. https://doi.org/10.1016/j.ajem.2018.10.058
19. Elixhauser A, Steiner C, Harris DR, Coffey RM. Comorbidity measures for use with administrative data. Med Care. 1998;36(1):8-27. https://doi.org/10.1097/00005650-199801000-00004
20. van Walraven C, Austin PC, Jennings A, Quan H, Forster AJ. A modification of the Elixhauser comorbidity measures into a point system for hospital death using administrative data. Med Care. 2009;47(6):626-633. https://doi.org/10.1097/MLR.0b013e31819432e5
21. Prin M, Wunsch H. The role of stepdown beds in hospital care. Am J Respir Crit Care Med. 2014;190(11):1210-1216. https://doi.org/10.1164/rccm.201406-1117PP
22. Narayanan N, Gross AK, Pintens M, Fee C, MacDougall C. Effect of an electronic medical record alert for severe sepsis among ED patients. Am J Emerg Med. 2016;34(2):185-188. https://doi.org/10.1016/j.ajem.2015.10.005
23. Singer M, Deutschman CS, Seymour CW, et al. The Third International Consensus Definitions for Sepsis and Septic Shock (Sepsis-3). JAMA. 2016;315(8):801-810. https://doi.org/10.1001/jama.2016.0287
24. Rhee C, Dantes R, Epstein L, et al. Incidence and trends of sepsis in US hospitals using clinical vs claims data, 2009-2014. JAMA. 2017;318(13):1241-1249. https://doi.org/10.1001/jama.2017.13836
25. Safari S, Baratloo A, Elfil M, Negida A. Evidence based emergency medicine; part 5 receiver operating curve and area under the curve. Emerg (Tehran). 2016;4(2):111-113.
26. DeLong ER, DeLong DM, Clarke-Pearson DL. Comparing the areas under two or more correlated receiver operating characteristic curves: a nonparametric approach. Biometrics. 1988;44(3):837-845.
27. Kangas C, Iverson L, Pierce D. Sepsis screening: combining Early Warning Scores and SIRS Criteria. Clin Nurs Res. 2021;30(1):42-49. https://doi.org/10.1177/1054773818823334.
28. Freund Y, Lemachatti N, Krastinova E, et al. Prognostic accuracy of Sepsis-3 Criteria for in-hospital mortality among patients with suspected infection presenting to the emergency department. JAMA. 2017;317(3):301-308. https://doi.org/10.1001/jama.2016.20329
29. Finkelsztein EJ, Jones DS, Ma KC, et al. Comparison of qSOFA and SIRS for predicting adverse outcomes of patients with suspicion of sepsis outside the intensive care unit. Crit Care. 2017;21(1):73. https://doi.org/10.1186/s13054-017-1658-5
30. Canet E, Taylor DM, Khor R, Krishnan V, Bellomo R. qSOFA as predictor of mortality and prolonged ICU admission in Emergency Department patients with suspected infection. J Crit Care. 2018;48:118-123. https://doi.org/10.1016/j.jcrc.2018.08.022
31. Anand V, Zhang Z, Kadri SS, Klompas M, Rhee C; CDC Prevention Epicenters Program. Epidemiology of Quick Sequential Organ Failure Assessment criteria in undifferentiated patients and association with suspected infection and sepsis. Chest. 2019;156(2):289-297. https://doi.org/10.1016/j.chest.2019.03.032
32. Hamilton F, Arnold D, Baird A, Albur M, Whiting P. Early Warning Scores do not accurately predict mortality in sepsis: A meta-analysis and systematic review of the literature. J Infect. 2018;76(3):241-248. https://doi.org/10.1016/j.jinf.2018.01.002
33. Koch E, Lovett S, Nghiem T, Riggs RA, Rech MA. Shock Index in the emergency department: utility and limitations. Open Access Emerg Med. 2019;11:179-199. https://doi.org/10.2147/OAEM.S178358
34. Yussof SJ, Zakaria MI, Mohamed FL, Bujang MA, Lakshmanan S, Asaari AH. Value of Shock Index in prognosticating the short-term outcome of death for patients presenting with severe sepsis and septic shock in the emergency department. Med J Malaysia. 2012;67(4):406-411.
35. Siddiqui S, Chua M, Kumaresh V, Choo R. A comparison of pre ICU admission SIRS, EWS and q SOFA scores for predicting mortality and length of stay in ICU. J Crit Care. 2017;41:191-193. https://doi.org/10.1016/j.jcrc.2017.05.017
36. Costa RT, Nassar AP, Caruso P. Accuracy of SOFA, qSOFA, and SIRS scores for mortality in cancer patients admitted to an intensive care unit with suspected infection. J Crit Care. 2018;45:52-57. https://doi.org/10.1016/j.jcrc.2017.12.024
37. Mellhammar L, Linder A, Tverring J, et al. NEWS2 is Superior to qSOFA in detecting sepsis with organ dysfunction in the emergency department. J Clin Med. 2019;8(8):1128. https://doi.org/10.3390/jcm8081128
38. Szakmany T, Pugh R, Kopczynska M, et al. Defining sepsis on the wards: results of a multi-centre point-prevalence study comparing two sepsis definitions. Anaesthesia. 2018;73(2):195-204. https://doi.org/10.1111/anae.14062
39. Newman TB, Kohn MA. Evidence-Based Diagnosis: An Introduction to Clinical Epidemiology. Cambridge University Press; 2009.
40. Askim Å, Moser F, Gustad LT, et al. Poor performance of quick-SOFA (qSOFA) score in predicting severe sepsis and mortality - a prospective study of patients admitted with infection to the emergency department. Scand J Trauma Resusc Emerg Med. 2017;25(1):56. https://doi.org/10.1186/s13049-017-0399-4
© 2021 Society of Hospital Medicine
Home Modifications for Rural Veterans With Disabilities
The US Department of Veterans Affairs (VA) created the Home Improvements and Structural Alterations (HISA) program to help provide necessary home modifications (HMs) to veterans with disabilities (VWDs) that will facilitate the provision of medical services at home and improve home accessibility and functional independence. The Veterans Health Administration (VHA) has more than 9 million veteran enrollees; of those, 2.7 million are classified as rural or highly rural.1 Rural veterans (RVs) possess higher rate of disability compared with that of urban veterans.2-5 RVs have unequal access to screening of ambulatory care sensitive conditions (eg, hypertension, diabetes mellitus).6 Furthermore, RVs are at risk of poor medical outcomes due to distance from health care facilities and specialist care, which can be a barrier to emergency care when issues arise. These barriers, among others, are associated with compromised health quality of life and health outcomes for RVs.3,6 The HISA program may be key to decreasing falls and other serious mishaps in the home. Therefore, understanding use of the HISA program by RVs is important. However, to date little information has been available regarding use of HISA benefits by RVs or characteristics of RVs who receive HISA benefits.
HISA Alterations Program
HISA was initially developed by VA to improve veterans’ transition from acute medical care to home.7,8 However, to obtain HISA grants currently, there is an average 3 to 6 months application process.7 Through the HISA program, VWDs can be prescribed the following HMs, including (but not limited to): flooring replacement, permanent ramps, roll-in showers, installation of central air-conditioning systems, improved lighting, kitchen/bathroom modifications, and home inspections. The HMs prescribed depend on an assessment of medical need by health care providers (HCPs).8
As time passed and the veteran population aged, the program now primarily helps ensure the ability to enter into essential areas and safety in the home.5 The amount of a HISA payment is based on whether a veteran’s health condition is related to military service as defined by the VHA service connection medical evaluation process. Barriers to obtaining a HISA HM can include difficulty in navigating the evaluation process and difficulty in finding a qualified contractor or builder to do the HM.7
This article aims to: (1) Detail the sociodemographic and clinical characteristics of rural HISA users (RHUs); (2) report on HISA usage patterns in number, types, and cost of HMs; (3) compare use amid the diverse VA medical centers (VAMCs) and related complexity levels and Veterans Integrated Service Networks (VISNs); and (4) examine the relationship between travel time/distance and HISA utilization. The long-term goal is to provide accurate information to researchers, HM administrators, health care providers and policy makers on HISA program utilization by rural VWDs, which may help improve its use and bring awareness of its users. This study was approved by the affiliate University of Florida Institutional Review Board and VA research and development committee at the North Florida/South Georgia Veterans Health System.
Methods
Data were obtained from 3 VA sources: the National Prosthetics Patient Database (NPPD), the VHA Medical Inpatient Dataset, and the VHA Outpatient Dataset.7 The NPPD is a national administrative database that contains information on prosthetic-associated products ordered by HCPs for patients, such as portable ramps, handrails, home oxygen equipment, and orthotic and prosthetic apparatus. Data obtained from the NPPD included cost of HMs, clinical characteristics, VISN, and VAMC. VA facilities are categorized into complexity levels 1a, 1b, 1c, 2, and 3. Complexity level 1a to 1c VAMCs address medical cases that entail “heightening involvedness,” meaning a larger number of patients presented with medical concerns needing medical specialists. Complexity levels 2 and 3 have fewer resources, lower patient numbers, and less medically complex patients. Finally, the VHA Medical Inpatient and Outpatient Datasets administrated by VA Informatics and Computing Infrastructure, consist of in-depth health services national data on inpatient and outpatient encounters and procedures.
The study cohort was divided into those with service-connected conditions (Class 1) or those with conditions not related to military service (Class 2). If veterans were identified in both classes, they were assigned to Class 1. The cost variable is determined by using the veterans’ classification. Class 1 veterans receive a lifetime limit of $6800, and Class 2 veterans receive a lifetime limit of $2000. A Class 2 veteran with ≥ 50% disability rating is eligible for a HISA lifetime limit of $6800. Whenever a value exceeds allowed limit of $6800 or $2000, due to data entry error or other reasons, the study team reassigned the cost value to the maximum allowed value.
Travel distance and time were derived by loading patient zip codes and HISA facility locations into the geographical information system program and using the nearest facility and find-route tools. These tools used a road network that simulates real-world driving conditions to calculate distance.
Study Variables
VWDs of any age, gender, and race/ethnicity who qualified for HISA and received HMs from fiscal year ( FY) 2015 through FY 2018 were identified (N = 30,823). Most VWDs were nonrural subjects (n = 19,970), and 43 had no Federal Information Processing System data. The final study cohort consisted of 10,810 HISA recipients. The NPPD, inpatient and outpatient data were merged by scrambled social security numbers to retrieve the following data: age, gender, race, ethnicity, marital status, Class (1 or 2), mean and total number of inpatient days, and type of HMs prescribed.
We also recorded rurality using the VA Rural-Urban Commuting Areas (RUCA) system, but we combined the rural and highly rural designation.1 Census tracts with a RUCA score of 10.0 are deemed highly rural, the remainder are considered rural except those with a RUCA score of 1.0 or 1.1. Travel time and distance from a veteran’s home to the VA facility that provided the HISA prescription were determined from zip codes. The current study focuses on VAMCs prescribing stations (affiliated sites of administrative parent medical facilities) where the HISA users obtained the HM, not the parent station (administrative parent medical facilities).
HISA Utilization
To characterize HISA utilization geographically and over time, the number of users were mapped by county. Areas where users were increasing (hot spots) or decreasing (cold spots) also were mapped. The maps were created using Environmental Systems Research Institute ArcGIS Pro 2.2.1 software. We chose to use natural breaks (Jenks) data classification method in a choropleth to symbolize the change over time map. We then used the Getis Ord GI* optimized hot spot analysis tool in the ArcGIS Pro spatial statistics tool set to generate the hot/cold spot maps. This tool identifies clusters of high values (hot spots) and low values (cold spots) creating a new output layer, RHUs by county, with a Z score, P value, and CI for each county. The Gi Bin field classifies statistically significant hot and cold spots. Counties sorted into the ± 3 category (bin) have a clustering characteristic (eg, with neighboring counties) that is statistically significant with a 99% CI; the ± 2 bin indicates a 95% CI for those county clustering sorted therein; ± 1 reflects a 90% CI; and 0 bin contains county features that have no statistical significant clustering with neighboring counties.
Data Analysis
Data were cleaned and analyzed using SAS 9.4 and R 3.5.3. Descriptive statistics are provided for sociodemographic characteristics, clinical characteristics, and class. ANOVA and t tests were used to compare continuous variables between groups, while χ2 and Fisher exact tests were used for dichotomous and categorical outcome variables. The threshold for statistical significance for these tests was set at α = .001.
Results
There were 10,810 RHUs from FY 2015 through FY 2018 and HISA utilization increased each year (Figure 1). Although some years may show usage decreases relative to previous fiscal years, the cumulative trends showed an increase relative to FY 2015 for both Classes of RVs (Figure 2). There was a 45.4% increase from FY 2015 to FY 2018 with a mean 13.6% yearly increase. Class 1 increased 21.0% and Class 2 increased 39.5% from FY 2015 to FY 2016 (Figure 3).
Most RHUs were male, White, and married. Class 1 and Class 2 RHUs differed significantly by age, race, marital status, and disability conditions: Class 1 RHUs were aged 6.6 years younger with a mean age of 69.1 years compared with 75.7 years for Class 2 users. For Class 1 RHUs, a plurality (29.4%) were aged 65 to 69 years; while a plurality (41.4%) of Class 2 users were aged ≥ 80 years. Musculoskeletal was the most common identified type of condition for all RHUs (Table 1).
To better understand HISA utilization patterns and net RHUs per county, we used a map to detail RHUs by county and change over time (Figure 4). Additionally, we compared US counties by RHUs from FY 2015 to FY 2018 and determined how clusters of high numbers of RHUs (hot spots) and low numbers of RHUs (cold spots) shifted over this period (Figure 5). While HISA utilization grew over the study period, the net count of RHUs per county varied by 9 to 20 persons/county. The population of RHUs increased over time in the Southwest, Southeast, and over much of the East/Northeast, while in the Central and Midwest regions, number of RHUs seems to decrease in population and/or use of the system. The cold spots in the Midwest and South Central US seem to increase with a significant relationship to neighboring counties having a low number of RHUs.
There were 11,166 HM prescribed to RHUs (Table 2). Bathroom HMs also were the dominant HM type for all facilities regardless of complexity levels (Table 3). The San Antonio, Texas, VAMC demonstrated the highest Class 1 vs Class 2 difference in HISA use (Class 1: 87.7% and Class 2: 12.3%). Except for the Des Moines VAMC, all other VAMCs showed HISA use > 60% by Class 1.
Cost Data
Air-conditioning installation ($5007) was the costliest HM overall (Table 4), closely followed by bathroom ($4978) and kitchen modifications ($4305). Bathroom renovations were the costliest HM type for both Class 1 and Class 2, closely followed by electrical repair and air-conditioning installation for Class 1 and driveway reconstruction and wooden ramp construction for Class 2.
The mean award received for HM was $4687 (Table 5). While the number of RHUs increased from FY 2015 to FY 2016, the average cost decreased, both overall ($280) and for Class 1 ($195) and Class 2 ($153). Except for a small decline in the number of Class 2 HISA recipients from FY 2017 to FY 2018, overall, the number of RHUs continuously grew from FY 2015 to FY 2018: 977 for the overall cohort, 678 for Class 1 and 299 for Class 2. Despite the obvious gain in the number of RHUs, the average costs did not notably change over time. VISN 21 had the highest mean cost, followed by VISNs 17, 6, 22, and 20.
Travel
Travel time and distance to the HISA prescribing facility differed significantly between Class 1 and Class 2 HISA users. RHUs had to travel about 95 minutes from their place of residence to access the HISA benefits program. There were no statistically significant differences between Class 1 and 2 users with respect to travel time and distance traveled (Table 6).
The majority of Class 1 and Class 2 veterans accessed the HISA from their nearest facility. However, nearly one-quarter of both Class 1 and 2 RHUs (24% each) did not. Among the 2598 who accessed the nonnearest facility, 97 (3.7%) accessed a facility that is ≤ 40 miles. Many (44%) users traveled 40 to 100 miles, and another 43.2% traveled 100 to 200 miles from their residence to access a HM prescription. Some 2598 users (1.1%) traveled > 500 miles to access a facility.
Discussion
Although utilization of the HISA program has steadily increased, overall participation by subpopulations such as RHUs can still be improved significantly. Veterans aged ≤ 46 years who have a disability that is common to those receiving HISA benefits have low HISA utilization. Similarly, veterans with sensory disabilities also have low use. These subpopulations are among those in great need of attention and services.
A study by Lucas and Zelaya, using the 2016 National Health Interview Survey data with an aim to measure degree of vision problems, dual sensory impairment, and hearing trouble in male veterans aged ≥ 18 years, found that veterans were more likely to report dual sensory impairment and balance difficulties when compared with nonveterans.9 The number of female veterans is growing but had very low representation in this study.10 This emerging VHA population requires information and education on their HM benefits.
Home Modifications
The most common HM prescribed for RHUs was for the bathroom. Further investigation is warranted as to why, given the diversity of HM types that the grant covers, low prescription rates exist across most of the HM types. There may be a lack of knowledge by providers and VWD as to the range of HMs that can be awarded under the grant. It is important that HCPs and veterans receive education on HISA HM options.
Semeah and colleagues pointed out the need for an assessment of the HISA HM ordering system to ensure that multiple HMs items (eg, kitchen, air conditioning, fees, driveway, and plumbing) are listed among the forced choices shown to clinicians to select from.7 Poor housing in rural America is widespread: 63% of rural dwellings need renovations and/or repairs to be accessible to individuals with disabilities, with > 6.7 million rural homes having no or faulty plumbing or kitchens; yet in this study, prescriptions for these HMs accounted for < 1%.11,12
VISN 6 had the most HISA awards with 1364, while VISN 21 had the fewest (245). Across all VISNs, Class 1 RHUs received more prescriptions than did Class 2 RHUs. Future research may seek to examine whether prescribers are fully aware of the eligibility of HM prescription to Class 2 veterans. VISN 21 ($5354); VISN 17 ($5302); and VISN 6 ($5301) had the highest mean HM expenditures. The national mean cost for HISA HMs were $4978 for bathrooms and $4305 for kitchens; for non-HISA HMs in FY 2017, the mean costs were $6362 and $12,255, respectively. A noteworthy concern is whether the maximum grant limit awards are sufficient to perform more expensive and complex HMs, such as the kitchen or major bathroom alternations.13
Facilities categorized as 1a, 1b, or 1c provided
North Florida/Sough Georgia was the highest-prescribing VAMC with 39% more HM prescriptions than the second highest prescribing facility (Durham, NC). Unfortunately, the data presented here cannot establish causality for the large variance difference between the top facilities, and the skewed distribution of total RHUs across VAMCs.
Travel-Related Variables
HISA beneficiaries face significant travel-related challenges. Just 3.6% of RHUs could access a facility within 40 miles of their home and 43.2% traveled 100 to 200 miles from their home to access a HM prescription. Further exploration is warranted to understand how travel patterns impact access to or the uptake of HISA.
RVs already have problems with accessing care because of long travel time.14,15 The choice or necessity to travel to a farther facility for HISA prescription is problematic for RVs, especially when transportation is often reported in the literature as a barrier to resources for people living in rural communities.15-17 When patients have travel barriers, they wait longer to obtain medical services and often wait for their conditions to worsen before seeking services.15,18 Once HM is completed, telerehabilitation is an effective delivery method used for delivering health care services to people in remote places.18,19 Considering that HISA use has the potential to improve quality of life, afford comfort, facilitate the accomplishment of activities of daily living for RVs, it is important that future studies examine how existing telehealth technologies can be used to improve HISA access.
Future Directions
County-level analyses is warranted in future studies exploring potential variables associated with HISA use; for example, county-level rates of primary care physicians and other HCPs. Future research should explore how long distance travel impacts the HISA application process and HM implementation. Further research also should focus on the HISA application structure and process to identify causes of delays. The HISA application process takes a mean 6 months to complete, yet the duration of hospital stays is 1 to 3 weeks, thus it is impossible to connect HISA to hospital discharge, which was the original intent of the program. Future research can examine how telehealth services can expedite HISA obtainment and coordination of the application process. Future research also may study the possible causes of the wide variations in HM prescriptions per facility. It is also important that educational programs provide information on the array of HM items that veterans can obtain.
Conclusions
In our previous study of the HISA cohort (2011-2017), we documented that an increase in utilization of the HISA program was warranted based on the low national budgetary appropriation and identification of significant low participation by vulnerable subpopulations, including veterans residing in rural areas or having returned from recent conflicts.7 The present study documents national utilization patterns, demographic profiles, and clinical characteristics of RHUs from FY 2015 through FY 2018, data that may be useful to policy makers and HISA administrators in predicting future use and users. It is important to note that the data and information presented in this article identify trends. The work in no way establishes a gold standard or any targeted goal of utilization. Future research could focus on conceptualizing or theorizing what steps are necessary to set such a gold standard of utilization rate and steps toward achievement.
Acknowledgments
This research was supported by grant 15521 from the US Department of Veterans Affairs, Office of Rural Health . Furthermore, the research was supported in part by grant K12 HD055929 from the National Institutes of Health.
1. US Department of Veterans Affairs, Veteran Health Administration, Office of Rural Health. Rural veteran health care challenges. Updated February 9, 2021. Accessed June 11, 2021. https://www.ruralhealth.va.gov/aboutus/ruralvets.asp
2. Holder, K.A. Veterans in rural America, 2011–2015. Published January 2017. Accessed June 11, 2021. https://www.census.gov/content/dam/Census/library/publications/2017/acs/acs-36.pdf
3. Pezzin LE, Bogner HR, Kurichi JE, et al. Preventable hospitalizations, barriers to care, and disability. Medicine (Baltimore). 2018;97(19):e0691. doi:10.1097/MD.0000000000010691
4. Rosenbach ML. Access and satisfaction within the disabled Medicare population. Health Care Financ Rev. 1995;17(2):147-167.
5. Semeah LM, Ganesh SP, Wang X, et al. Home modification and health services utilization in rural and urban veterans with disabilities. Housing Policy Debate. 2021. Published online: March 4, 2021. doi:10.1080/10511482.2020.1858923
6. Spoont M, Greer N, Su J, Fitzgerald P, Rutks I, and Wilt TJ. Rural vs. urban ambulatory health care: A Systematic Review. Published May 2011. Accessed June 11, 2021. https://www.hsrd.research.va.gov/publications/esp/ambulatory.pdf
7. Semeah LM, Wang X, Cowper Ripley DC, et al. Improving health through a home modification service for veterans. In: Fiedler BA, ed. Three Facets of Public Health and Paths to Improvements. Academic Press; 2020:381-416.
8. Semeah LM, Ahrentzen S, Jia H, Cowper-Ripley DC, Levy CE, Mann WC. The home improvements and structural alterations benefits program: veterans with disabilities and home accessibility. J Disability Policy Studies. 2017;28(1):43-51. doi:10.1177/1044207317696275
9. Lucas, JW, Zelaya, CE. Hearing difficulty, vision trouble, and balance problems among male veterans and nonveterans. Published June 12, 2020. Accessed June 11, 2021. https://www.cdc.gov/nchs/data/nhsr/nhsr142-508.pdf
10. US Department of Veterans Affairs, National Center for Veterans Analysis and Statistics. Women veterans report: the past, present, and future of women veterans. Published February 2017. Accessed June 11, 2021. https://www.va.gov/vetdata/docs/SpecialReports/Women_Veterans_2015_Final.pdf
11. US Department of Housing and Urban Development, Office of Policy Development and Research. Housing challenges of rural seniors. Published 2017. Accessed June 11, 2021. https://www.huduser.gov/portal/periodicals/em/summer17/highlight1.html
12. Pendall R, Goodman L, Zhu J, Gold A. The future of rural housing. Published October 2016. Accessed June 11, 202.1 https://www.urban.org/sites/default/files/publication/85101/2000972-the-future-of-rural-housing_6.pdf
13. Joint Center for Housing Studies at Harvard University. Improving America’s housing 2019. Published 2019. Accessed June 11, 2021. https://www.jchs.harvard.edu/sites/default/files/reports/files/Harvard_JCHS_Improving_Americas_Housing_2019.pdf
14. Schooley BL, Horan TA, Lee PW, West PA. Rural veteran access to healthcare services: investigating the role of information and communication technologies in overcoming spatial barriers. Perspect Health Inf Manag. 2010;7(Spring):1f. Published 2010 Apr 1.
15. Ripley DC, Kwong PL, Vogel WB, Kurichi JE, Bates BE, Davenport C. How does geographic access affect in-hospital mortality for veterans with acute ischemic stroke?. Med Care. 2015;53(6):501-509. doi:10.1097/MLR.0000000000000366
16. Cowper-Ripley DC, Reker DM, Hayes J, et al. Geographic access to VHA rehabilitation services for traumatically injured veterans. Fed Pract. 2009;26(10):28-39.
17. Smith M, Towne S, Herrera-Venson A, Cameron K, Horel S, Ory M, et al. Delivery of fall prevention interventions for at-risk older adults in rural areas: Findings from a national dissemination. International journal of environmental research and public health. 2018;15:2798. doi: 10.3390/ijerph15122798
18. Hale-Gallardo JL, Kreider CM, Jia H, et al. Telerehabilitation for Rural Veterans: A Qualitative Assessment of Barriers and Facilitators to Implementation. J Multidiscip Healthc. 2020;13:559-570. doi:10.2147/JMDH.S247267
19. Sarfo FS, Akassi J, Kyem G, et al. Long-Term Outcomes of Stroke in a Ghanaian Outpatient Clinic. J Stroke Cerebrovasc Dis. 2018;27(4):1090-1099. doi:10.1016/j.jstrokecerebrovasdis.2017.11.017
The US Department of Veterans Affairs (VA) created the Home Improvements and Structural Alterations (HISA) program to help provide necessary home modifications (HMs) to veterans with disabilities (VWDs) that will facilitate the provision of medical services at home and improve home accessibility and functional independence. The Veterans Health Administration (VHA) has more than 9 million veteran enrollees; of those, 2.7 million are classified as rural or highly rural.1 Rural veterans (RVs) possess higher rate of disability compared with that of urban veterans.2-5 RVs have unequal access to screening of ambulatory care sensitive conditions (eg, hypertension, diabetes mellitus).6 Furthermore, RVs are at risk of poor medical outcomes due to distance from health care facilities and specialist care, which can be a barrier to emergency care when issues arise. These barriers, among others, are associated with compromised health quality of life and health outcomes for RVs.3,6 The HISA program may be key to decreasing falls and other serious mishaps in the home. Therefore, understanding use of the HISA program by RVs is important. However, to date little information has been available regarding use of HISA benefits by RVs or characteristics of RVs who receive HISA benefits.
HISA Alterations Program
HISA was initially developed by VA to improve veterans’ transition from acute medical care to home.7,8 However, to obtain HISA grants currently, there is an average 3 to 6 months application process.7 Through the HISA program, VWDs can be prescribed the following HMs, including (but not limited to): flooring replacement, permanent ramps, roll-in showers, installation of central air-conditioning systems, improved lighting, kitchen/bathroom modifications, and home inspections. The HMs prescribed depend on an assessment of medical need by health care providers (HCPs).8
As time passed and the veteran population aged, the program now primarily helps ensure the ability to enter into essential areas and safety in the home.5 The amount of a HISA payment is based on whether a veteran’s health condition is related to military service as defined by the VHA service connection medical evaluation process. Barriers to obtaining a HISA HM can include difficulty in navigating the evaluation process and difficulty in finding a qualified contractor or builder to do the HM.7
This article aims to: (1) Detail the sociodemographic and clinical characteristics of rural HISA users (RHUs); (2) report on HISA usage patterns in number, types, and cost of HMs; (3) compare use amid the diverse VA medical centers (VAMCs) and related complexity levels and Veterans Integrated Service Networks (VISNs); and (4) examine the relationship between travel time/distance and HISA utilization. The long-term goal is to provide accurate information to researchers, HM administrators, health care providers and policy makers on HISA program utilization by rural VWDs, which may help improve its use and bring awareness of its users. This study was approved by the affiliate University of Florida Institutional Review Board and VA research and development committee at the North Florida/South Georgia Veterans Health System.
Methods
Data were obtained from 3 VA sources: the National Prosthetics Patient Database (NPPD), the VHA Medical Inpatient Dataset, and the VHA Outpatient Dataset.7 The NPPD is a national administrative database that contains information on prosthetic-associated products ordered by HCPs for patients, such as portable ramps, handrails, home oxygen equipment, and orthotic and prosthetic apparatus. Data obtained from the NPPD included cost of HMs, clinical characteristics, VISN, and VAMC. VA facilities are categorized into complexity levels 1a, 1b, 1c, 2, and 3. Complexity level 1a to 1c VAMCs address medical cases that entail “heightening involvedness,” meaning a larger number of patients presented with medical concerns needing medical specialists. Complexity levels 2 and 3 have fewer resources, lower patient numbers, and less medically complex patients. Finally, the VHA Medical Inpatient and Outpatient Datasets administrated by VA Informatics and Computing Infrastructure, consist of in-depth health services national data on inpatient and outpatient encounters and procedures.
The study cohort was divided into those with service-connected conditions (Class 1) or those with conditions not related to military service (Class 2). If veterans were identified in both classes, they were assigned to Class 1. The cost variable is determined by using the veterans’ classification. Class 1 veterans receive a lifetime limit of $6800, and Class 2 veterans receive a lifetime limit of $2000. A Class 2 veteran with ≥ 50% disability rating is eligible for a HISA lifetime limit of $6800. Whenever a value exceeds allowed limit of $6800 or $2000, due to data entry error or other reasons, the study team reassigned the cost value to the maximum allowed value.
Travel distance and time were derived by loading patient zip codes and HISA facility locations into the geographical information system program and using the nearest facility and find-route tools. These tools used a road network that simulates real-world driving conditions to calculate distance.
Study Variables
VWDs of any age, gender, and race/ethnicity who qualified for HISA and received HMs from fiscal year ( FY) 2015 through FY 2018 were identified (N = 30,823). Most VWDs were nonrural subjects (n = 19,970), and 43 had no Federal Information Processing System data. The final study cohort consisted of 10,810 HISA recipients. The NPPD, inpatient and outpatient data were merged by scrambled social security numbers to retrieve the following data: age, gender, race, ethnicity, marital status, Class (1 or 2), mean and total number of inpatient days, and type of HMs prescribed.
We also recorded rurality using the VA Rural-Urban Commuting Areas (RUCA) system, but we combined the rural and highly rural designation.1 Census tracts with a RUCA score of 10.0 are deemed highly rural, the remainder are considered rural except those with a RUCA score of 1.0 or 1.1. Travel time and distance from a veteran’s home to the VA facility that provided the HISA prescription were determined from zip codes. The current study focuses on VAMCs prescribing stations (affiliated sites of administrative parent medical facilities) where the HISA users obtained the HM, not the parent station (administrative parent medical facilities).
HISA Utilization
To characterize HISA utilization geographically and over time, the number of users were mapped by county. Areas where users were increasing (hot spots) or decreasing (cold spots) also were mapped. The maps were created using Environmental Systems Research Institute ArcGIS Pro 2.2.1 software. We chose to use natural breaks (Jenks) data classification method in a choropleth to symbolize the change over time map. We then used the Getis Ord GI* optimized hot spot analysis tool in the ArcGIS Pro spatial statistics tool set to generate the hot/cold spot maps. This tool identifies clusters of high values (hot spots) and low values (cold spots) creating a new output layer, RHUs by county, with a Z score, P value, and CI for each county. The Gi Bin field classifies statistically significant hot and cold spots. Counties sorted into the ± 3 category (bin) have a clustering characteristic (eg, with neighboring counties) that is statistically significant with a 99% CI; the ± 2 bin indicates a 95% CI for those county clustering sorted therein; ± 1 reflects a 90% CI; and 0 bin contains county features that have no statistical significant clustering with neighboring counties.
Data Analysis
Data were cleaned and analyzed using SAS 9.4 and R 3.5.3. Descriptive statistics are provided for sociodemographic characteristics, clinical characteristics, and class. ANOVA and t tests were used to compare continuous variables between groups, while χ2 and Fisher exact tests were used for dichotomous and categorical outcome variables. The threshold for statistical significance for these tests was set at α = .001.
Results
There were 10,810 RHUs from FY 2015 through FY 2018 and HISA utilization increased each year (Figure 1). Although some years may show usage decreases relative to previous fiscal years, the cumulative trends showed an increase relative to FY 2015 for both Classes of RVs (Figure 2). There was a 45.4% increase from FY 2015 to FY 2018 with a mean 13.6% yearly increase. Class 1 increased 21.0% and Class 2 increased 39.5% from FY 2015 to FY 2016 (Figure 3).
Most RHUs were male, White, and married. Class 1 and Class 2 RHUs differed significantly by age, race, marital status, and disability conditions: Class 1 RHUs were aged 6.6 years younger with a mean age of 69.1 years compared with 75.7 years for Class 2 users. For Class 1 RHUs, a plurality (29.4%) were aged 65 to 69 years; while a plurality (41.4%) of Class 2 users were aged ≥ 80 years. Musculoskeletal was the most common identified type of condition for all RHUs (Table 1).
To better understand HISA utilization patterns and net RHUs per county, we used a map to detail RHUs by county and change over time (Figure 4). Additionally, we compared US counties by RHUs from FY 2015 to FY 2018 and determined how clusters of high numbers of RHUs (hot spots) and low numbers of RHUs (cold spots) shifted over this period (Figure 5). While HISA utilization grew over the study period, the net count of RHUs per county varied by 9 to 20 persons/county. The population of RHUs increased over time in the Southwest, Southeast, and over much of the East/Northeast, while in the Central and Midwest regions, number of RHUs seems to decrease in population and/or use of the system. The cold spots in the Midwest and South Central US seem to increase with a significant relationship to neighboring counties having a low number of RHUs.
There were 11,166 HM prescribed to RHUs (Table 2). Bathroom HMs also were the dominant HM type for all facilities regardless of complexity levels (Table 3). The San Antonio, Texas, VAMC demonstrated the highest Class 1 vs Class 2 difference in HISA use (Class 1: 87.7% and Class 2: 12.3%). Except for the Des Moines VAMC, all other VAMCs showed HISA use > 60% by Class 1.
Cost Data
Air-conditioning installation ($5007) was the costliest HM overall (Table 4), closely followed by bathroom ($4978) and kitchen modifications ($4305). Bathroom renovations were the costliest HM type for both Class 1 and Class 2, closely followed by electrical repair and air-conditioning installation for Class 1 and driveway reconstruction and wooden ramp construction for Class 2.
The mean award received for HM was $4687 (Table 5). While the number of RHUs increased from FY 2015 to FY 2016, the average cost decreased, both overall ($280) and for Class 1 ($195) and Class 2 ($153). Except for a small decline in the number of Class 2 HISA recipients from FY 2017 to FY 2018, overall, the number of RHUs continuously grew from FY 2015 to FY 2018: 977 for the overall cohort, 678 for Class 1 and 299 for Class 2. Despite the obvious gain in the number of RHUs, the average costs did not notably change over time. VISN 21 had the highest mean cost, followed by VISNs 17, 6, 22, and 20.
Travel
Travel time and distance to the HISA prescribing facility differed significantly between Class 1 and Class 2 HISA users. RHUs had to travel about 95 minutes from their place of residence to access the HISA benefits program. There were no statistically significant differences between Class 1 and 2 users with respect to travel time and distance traveled (Table 6).
The majority of Class 1 and Class 2 veterans accessed the HISA from their nearest facility. However, nearly one-quarter of both Class 1 and 2 RHUs (24% each) did not. Among the 2598 who accessed the nonnearest facility, 97 (3.7%) accessed a facility that is ≤ 40 miles. Many (44%) users traveled 40 to 100 miles, and another 43.2% traveled 100 to 200 miles from their residence to access a HM prescription. Some 2598 users (1.1%) traveled > 500 miles to access a facility.
Discussion
Although utilization of the HISA program has steadily increased, overall participation by subpopulations such as RHUs can still be improved significantly. Veterans aged ≤ 46 years who have a disability that is common to those receiving HISA benefits have low HISA utilization. Similarly, veterans with sensory disabilities also have low use. These subpopulations are among those in great need of attention and services.
A study by Lucas and Zelaya, using the 2016 National Health Interview Survey data with an aim to measure degree of vision problems, dual sensory impairment, and hearing trouble in male veterans aged ≥ 18 years, found that veterans were more likely to report dual sensory impairment and balance difficulties when compared with nonveterans.9 The number of female veterans is growing but had very low representation in this study.10 This emerging VHA population requires information and education on their HM benefits.
Home Modifications
The most common HM prescribed for RHUs was for the bathroom. Further investigation is warranted as to why, given the diversity of HM types that the grant covers, low prescription rates exist across most of the HM types. There may be a lack of knowledge by providers and VWD as to the range of HMs that can be awarded under the grant. It is important that HCPs and veterans receive education on HISA HM options.
Semeah and colleagues pointed out the need for an assessment of the HISA HM ordering system to ensure that multiple HMs items (eg, kitchen, air conditioning, fees, driveway, and plumbing) are listed among the forced choices shown to clinicians to select from.7 Poor housing in rural America is widespread: 63% of rural dwellings need renovations and/or repairs to be accessible to individuals with disabilities, with > 6.7 million rural homes having no or faulty plumbing or kitchens; yet in this study, prescriptions for these HMs accounted for < 1%.11,12
VISN 6 had the most HISA awards with 1364, while VISN 21 had the fewest (245). Across all VISNs, Class 1 RHUs received more prescriptions than did Class 2 RHUs. Future research may seek to examine whether prescribers are fully aware of the eligibility of HM prescription to Class 2 veterans. VISN 21 ($5354); VISN 17 ($5302); and VISN 6 ($5301) had the highest mean HM expenditures. The national mean cost for HISA HMs were $4978 for bathrooms and $4305 for kitchens; for non-HISA HMs in FY 2017, the mean costs were $6362 and $12,255, respectively. A noteworthy concern is whether the maximum grant limit awards are sufficient to perform more expensive and complex HMs, such as the kitchen or major bathroom alternations.13
Facilities categorized as 1a, 1b, or 1c provided
North Florida/Sough Georgia was the highest-prescribing VAMC with 39% more HM prescriptions than the second highest prescribing facility (Durham, NC). Unfortunately, the data presented here cannot establish causality for the large variance difference between the top facilities, and the skewed distribution of total RHUs across VAMCs.
Travel-Related Variables
HISA beneficiaries face significant travel-related challenges. Just 3.6% of RHUs could access a facility within 40 miles of their home and 43.2% traveled 100 to 200 miles from their home to access a HM prescription. Further exploration is warranted to understand how travel patterns impact access to or the uptake of HISA.
RVs already have problems with accessing care because of long travel time.14,15 The choice or necessity to travel to a farther facility for HISA prescription is problematic for RVs, especially when transportation is often reported in the literature as a barrier to resources for people living in rural communities.15-17 When patients have travel barriers, they wait longer to obtain medical services and often wait for their conditions to worsen before seeking services.15,18 Once HM is completed, telerehabilitation is an effective delivery method used for delivering health care services to people in remote places.18,19 Considering that HISA use has the potential to improve quality of life, afford comfort, facilitate the accomplishment of activities of daily living for RVs, it is important that future studies examine how existing telehealth technologies can be used to improve HISA access.
Future Directions
County-level analyses is warranted in future studies exploring potential variables associated with HISA use; for example, county-level rates of primary care physicians and other HCPs. Future research should explore how long distance travel impacts the HISA application process and HM implementation. Further research also should focus on the HISA application structure and process to identify causes of delays. The HISA application process takes a mean 6 months to complete, yet the duration of hospital stays is 1 to 3 weeks, thus it is impossible to connect HISA to hospital discharge, which was the original intent of the program. Future research can examine how telehealth services can expedite HISA obtainment and coordination of the application process. Future research also may study the possible causes of the wide variations in HM prescriptions per facility. It is also important that educational programs provide information on the array of HM items that veterans can obtain.
Conclusions
In our previous study of the HISA cohort (2011-2017), we documented that an increase in utilization of the HISA program was warranted based on the low national budgetary appropriation and identification of significant low participation by vulnerable subpopulations, including veterans residing in rural areas or having returned from recent conflicts.7 The present study documents national utilization patterns, demographic profiles, and clinical characteristics of RHUs from FY 2015 through FY 2018, data that may be useful to policy makers and HISA administrators in predicting future use and users. It is important to note that the data and information presented in this article identify trends. The work in no way establishes a gold standard or any targeted goal of utilization. Future research could focus on conceptualizing or theorizing what steps are necessary to set such a gold standard of utilization rate and steps toward achievement.
Acknowledgments
This research was supported by grant 15521 from the US Department of Veterans Affairs, Office of Rural Health . Furthermore, the research was supported in part by grant K12 HD055929 from the National Institutes of Health.
The US Department of Veterans Affairs (VA) created the Home Improvements and Structural Alterations (HISA) program to help provide necessary home modifications (HMs) to veterans with disabilities (VWDs) that will facilitate the provision of medical services at home and improve home accessibility and functional independence. The Veterans Health Administration (VHA) has more than 9 million veteran enrollees; of those, 2.7 million are classified as rural or highly rural.1 Rural veterans (RVs) possess higher rate of disability compared with that of urban veterans.2-5 RVs have unequal access to screening of ambulatory care sensitive conditions (eg, hypertension, diabetes mellitus).6 Furthermore, RVs are at risk of poor medical outcomes due to distance from health care facilities and specialist care, which can be a barrier to emergency care when issues arise. These barriers, among others, are associated with compromised health quality of life and health outcomes for RVs.3,6 The HISA program may be key to decreasing falls and other serious mishaps in the home. Therefore, understanding use of the HISA program by RVs is important. However, to date little information has been available regarding use of HISA benefits by RVs or characteristics of RVs who receive HISA benefits.
HISA Alterations Program
HISA was initially developed by VA to improve veterans’ transition from acute medical care to home.7,8 However, to obtain HISA grants currently, there is an average 3 to 6 months application process.7 Through the HISA program, VWDs can be prescribed the following HMs, including (but not limited to): flooring replacement, permanent ramps, roll-in showers, installation of central air-conditioning systems, improved lighting, kitchen/bathroom modifications, and home inspections. The HMs prescribed depend on an assessment of medical need by health care providers (HCPs).8
As time passed and the veteran population aged, the program now primarily helps ensure the ability to enter into essential areas and safety in the home.5 The amount of a HISA payment is based on whether a veteran’s health condition is related to military service as defined by the VHA service connection medical evaluation process. Barriers to obtaining a HISA HM can include difficulty in navigating the evaluation process and difficulty in finding a qualified contractor or builder to do the HM.7
This article aims to: (1) Detail the sociodemographic and clinical characteristics of rural HISA users (RHUs); (2) report on HISA usage patterns in number, types, and cost of HMs; (3) compare use amid the diverse VA medical centers (VAMCs) and related complexity levels and Veterans Integrated Service Networks (VISNs); and (4) examine the relationship between travel time/distance and HISA utilization. The long-term goal is to provide accurate information to researchers, HM administrators, health care providers and policy makers on HISA program utilization by rural VWDs, which may help improve its use and bring awareness of its users. This study was approved by the affiliate University of Florida Institutional Review Board and VA research and development committee at the North Florida/South Georgia Veterans Health System.
Methods
Data were obtained from 3 VA sources: the National Prosthetics Patient Database (NPPD), the VHA Medical Inpatient Dataset, and the VHA Outpatient Dataset.7 The NPPD is a national administrative database that contains information on prosthetic-associated products ordered by HCPs for patients, such as portable ramps, handrails, home oxygen equipment, and orthotic and prosthetic apparatus. Data obtained from the NPPD included cost of HMs, clinical characteristics, VISN, and VAMC. VA facilities are categorized into complexity levels 1a, 1b, 1c, 2, and 3. Complexity level 1a to 1c VAMCs address medical cases that entail “heightening involvedness,” meaning a larger number of patients presented with medical concerns needing medical specialists. Complexity levels 2 and 3 have fewer resources, lower patient numbers, and less medically complex patients. Finally, the VHA Medical Inpatient and Outpatient Datasets administrated by VA Informatics and Computing Infrastructure, consist of in-depth health services national data on inpatient and outpatient encounters and procedures.
The study cohort was divided into those with service-connected conditions (Class 1) or those with conditions not related to military service (Class 2). If veterans were identified in both classes, they were assigned to Class 1. The cost variable is determined by using the veterans’ classification. Class 1 veterans receive a lifetime limit of $6800, and Class 2 veterans receive a lifetime limit of $2000. A Class 2 veteran with ≥ 50% disability rating is eligible for a HISA lifetime limit of $6800. Whenever a value exceeds allowed limit of $6800 or $2000, due to data entry error or other reasons, the study team reassigned the cost value to the maximum allowed value.
Travel distance and time were derived by loading patient zip codes and HISA facility locations into the geographical information system program and using the nearest facility and find-route tools. These tools used a road network that simulates real-world driving conditions to calculate distance.
Study Variables
VWDs of any age, gender, and race/ethnicity who qualified for HISA and received HMs from fiscal year ( FY) 2015 through FY 2018 were identified (N = 30,823). Most VWDs were nonrural subjects (n = 19,970), and 43 had no Federal Information Processing System data. The final study cohort consisted of 10,810 HISA recipients. The NPPD, inpatient and outpatient data were merged by scrambled social security numbers to retrieve the following data: age, gender, race, ethnicity, marital status, Class (1 or 2), mean and total number of inpatient days, and type of HMs prescribed.
We also recorded rurality using the VA Rural-Urban Commuting Areas (RUCA) system, but we combined the rural and highly rural designation.1 Census tracts with a RUCA score of 10.0 are deemed highly rural, the remainder are considered rural except those with a RUCA score of 1.0 or 1.1. Travel time and distance from a veteran’s home to the VA facility that provided the HISA prescription were determined from zip codes. The current study focuses on VAMCs prescribing stations (affiliated sites of administrative parent medical facilities) where the HISA users obtained the HM, not the parent station (administrative parent medical facilities).
HISA Utilization
To characterize HISA utilization geographically and over time, the number of users were mapped by county. Areas where users were increasing (hot spots) or decreasing (cold spots) also were mapped. The maps were created using Environmental Systems Research Institute ArcGIS Pro 2.2.1 software. We chose to use natural breaks (Jenks) data classification method in a choropleth to symbolize the change over time map. We then used the Getis Ord GI* optimized hot spot analysis tool in the ArcGIS Pro spatial statistics tool set to generate the hot/cold spot maps. This tool identifies clusters of high values (hot spots) and low values (cold spots) creating a new output layer, RHUs by county, with a Z score, P value, and CI for each county. The Gi Bin field classifies statistically significant hot and cold spots. Counties sorted into the ± 3 category (bin) have a clustering characteristic (eg, with neighboring counties) that is statistically significant with a 99% CI; the ± 2 bin indicates a 95% CI for those county clustering sorted therein; ± 1 reflects a 90% CI; and 0 bin contains county features that have no statistical significant clustering with neighboring counties.
Data Analysis
Data were cleaned and analyzed using SAS 9.4 and R 3.5.3. Descriptive statistics are provided for sociodemographic characteristics, clinical characteristics, and class. ANOVA and t tests were used to compare continuous variables between groups, while χ2 and Fisher exact tests were used for dichotomous and categorical outcome variables. The threshold for statistical significance for these tests was set at α = .001.
Results
There were 10,810 RHUs from FY 2015 through FY 2018 and HISA utilization increased each year (Figure 1). Although some years may show usage decreases relative to previous fiscal years, the cumulative trends showed an increase relative to FY 2015 for both Classes of RVs (Figure 2). There was a 45.4% increase from FY 2015 to FY 2018 with a mean 13.6% yearly increase. Class 1 increased 21.0% and Class 2 increased 39.5% from FY 2015 to FY 2016 (Figure 3).
Most RHUs were male, White, and married. Class 1 and Class 2 RHUs differed significantly by age, race, marital status, and disability conditions: Class 1 RHUs were aged 6.6 years younger with a mean age of 69.1 years compared with 75.7 years for Class 2 users. For Class 1 RHUs, a plurality (29.4%) were aged 65 to 69 years; while a plurality (41.4%) of Class 2 users were aged ≥ 80 years. Musculoskeletal was the most common identified type of condition for all RHUs (Table 1).
To better understand HISA utilization patterns and net RHUs per county, we used a map to detail RHUs by county and change over time (Figure 4). Additionally, we compared US counties by RHUs from FY 2015 to FY 2018 and determined how clusters of high numbers of RHUs (hot spots) and low numbers of RHUs (cold spots) shifted over this period (Figure 5). While HISA utilization grew over the study period, the net count of RHUs per county varied by 9 to 20 persons/county. The population of RHUs increased over time in the Southwest, Southeast, and over much of the East/Northeast, while in the Central and Midwest regions, number of RHUs seems to decrease in population and/or use of the system. The cold spots in the Midwest and South Central US seem to increase with a significant relationship to neighboring counties having a low number of RHUs.
There were 11,166 HM prescribed to RHUs (Table 2). Bathroom HMs also were the dominant HM type for all facilities regardless of complexity levels (Table 3). The San Antonio, Texas, VAMC demonstrated the highest Class 1 vs Class 2 difference in HISA use (Class 1: 87.7% and Class 2: 12.3%). Except for the Des Moines VAMC, all other VAMCs showed HISA use > 60% by Class 1.
Cost Data
Air-conditioning installation ($5007) was the costliest HM overall (Table 4), closely followed by bathroom ($4978) and kitchen modifications ($4305). Bathroom renovations were the costliest HM type for both Class 1 and Class 2, closely followed by electrical repair and air-conditioning installation for Class 1 and driveway reconstruction and wooden ramp construction for Class 2.
The mean award received for HM was $4687 (Table 5). While the number of RHUs increased from FY 2015 to FY 2016, the average cost decreased, both overall ($280) and for Class 1 ($195) and Class 2 ($153). Except for a small decline in the number of Class 2 HISA recipients from FY 2017 to FY 2018, overall, the number of RHUs continuously grew from FY 2015 to FY 2018: 977 for the overall cohort, 678 for Class 1 and 299 for Class 2. Despite the obvious gain in the number of RHUs, the average costs did not notably change over time. VISN 21 had the highest mean cost, followed by VISNs 17, 6, 22, and 20.
Travel
Travel time and distance to the HISA prescribing facility differed significantly between Class 1 and Class 2 HISA users. RHUs had to travel about 95 minutes from their place of residence to access the HISA benefits program. There were no statistically significant differences between Class 1 and 2 users with respect to travel time and distance traveled (Table 6).
The majority of Class 1 and Class 2 veterans accessed the HISA from their nearest facility. However, nearly one-quarter of both Class 1 and 2 RHUs (24% each) did not. Among the 2598 who accessed the nonnearest facility, 97 (3.7%) accessed a facility that is ≤ 40 miles. Many (44%) users traveled 40 to 100 miles, and another 43.2% traveled 100 to 200 miles from their residence to access a HM prescription. Some 2598 users (1.1%) traveled > 500 miles to access a facility.
Discussion
Although utilization of the HISA program has steadily increased, overall participation by subpopulations such as RHUs can still be improved significantly. Veterans aged ≤ 46 years who have a disability that is common to those receiving HISA benefits have low HISA utilization. Similarly, veterans with sensory disabilities also have low use. These subpopulations are among those in great need of attention and services.
A study by Lucas and Zelaya, using the 2016 National Health Interview Survey data with an aim to measure degree of vision problems, dual sensory impairment, and hearing trouble in male veterans aged ≥ 18 years, found that veterans were more likely to report dual sensory impairment and balance difficulties when compared with nonveterans.9 The number of female veterans is growing but had very low representation in this study.10 This emerging VHA population requires information and education on their HM benefits.
Home Modifications
The most common HM prescribed for RHUs was for the bathroom. Further investigation is warranted as to why, given the diversity of HM types that the grant covers, low prescription rates exist across most of the HM types. There may be a lack of knowledge by providers and VWD as to the range of HMs that can be awarded under the grant. It is important that HCPs and veterans receive education on HISA HM options.
Semeah and colleagues pointed out the need for an assessment of the HISA HM ordering system to ensure that multiple HMs items (eg, kitchen, air conditioning, fees, driveway, and plumbing) are listed among the forced choices shown to clinicians to select from.7 Poor housing in rural America is widespread: 63% of rural dwellings need renovations and/or repairs to be accessible to individuals with disabilities, with > 6.7 million rural homes having no or faulty plumbing or kitchens; yet in this study, prescriptions for these HMs accounted for < 1%.11,12
VISN 6 had the most HISA awards with 1364, while VISN 21 had the fewest (245). Across all VISNs, Class 1 RHUs received more prescriptions than did Class 2 RHUs. Future research may seek to examine whether prescribers are fully aware of the eligibility of HM prescription to Class 2 veterans. VISN 21 ($5354); VISN 17 ($5302); and VISN 6 ($5301) had the highest mean HM expenditures. The national mean cost for HISA HMs were $4978 for bathrooms and $4305 for kitchens; for non-HISA HMs in FY 2017, the mean costs were $6362 and $12,255, respectively. A noteworthy concern is whether the maximum grant limit awards are sufficient to perform more expensive and complex HMs, such as the kitchen or major bathroom alternations.13
Facilities categorized as 1a, 1b, or 1c provided
North Florida/Sough Georgia was the highest-prescribing VAMC with 39% more HM prescriptions than the second highest prescribing facility (Durham, NC). Unfortunately, the data presented here cannot establish causality for the large variance difference between the top facilities, and the skewed distribution of total RHUs across VAMCs.
Travel-Related Variables
HISA beneficiaries face significant travel-related challenges. Just 3.6% of RHUs could access a facility within 40 miles of their home and 43.2% traveled 100 to 200 miles from their home to access a HM prescription. Further exploration is warranted to understand how travel patterns impact access to or the uptake of HISA.
RVs already have problems with accessing care because of long travel time.14,15 The choice or necessity to travel to a farther facility for HISA prescription is problematic for RVs, especially when transportation is often reported in the literature as a barrier to resources for people living in rural communities.15-17 When patients have travel barriers, they wait longer to obtain medical services and often wait for their conditions to worsen before seeking services.15,18 Once HM is completed, telerehabilitation is an effective delivery method used for delivering health care services to people in remote places.18,19 Considering that HISA use has the potential to improve quality of life, afford comfort, facilitate the accomplishment of activities of daily living for RVs, it is important that future studies examine how existing telehealth technologies can be used to improve HISA access.
Future Directions
County-level analyses is warranted in future studies exploring potential variables associated with HISA use; for example, county-level rates of primary care physicians and other HCPs. Future research should explore how long distance travel impacts the HISA application process and HM implementation. Further research also should focus on the HISA application structure and process to identify causes of delays. The HISA application process takes a mean 6 months to complete, yet the duration of hospital stays is 1 to 3 weeks, thus it is impossible to connect HISA to hospital discharge, which was the original intent of the program. Future research can examine how telehealth services can expedite HISA obtainment and coordination of the application process. Future research also may study the possible causes of the wide variations in HM prescriptions per facility. It is also important that educational programs provide information on the array of HM items that veterans can obtain.
Conclusions
In our previous study of the HISA cohort (2011-2017), we documented that an increase in utilization of the HISA program was warranted based on the low national budgetary appropriation and identification of significant low participation by vulnerable subpopulations, including veterans residing in rural areas or having returned from recent conflicts.7 The present study documents national utilization patterns, demographic profiles, and clinical characteristics of RHUs from FY 2015 through FY 2018, data that may be useful to policy makers and HISA administrators in predicting future use and users. It is important to note that the data and information presented in this article identify trends. The work in no way establishes a gold standard or any targeted goal of utilization. Future research could focus on conceptualizing or theorizing what steps are necessary to set such a gold standard of utilization rate and steps toward achievement.
Acknowledgments
This research was supported by grant 15521 from the US Department of Veterans Affairs, Office of Rural Health . Furthermore, the research was supported in part by grant K12 HD055929 from the National Institutes of Health.
1. US Department of Veterans Affairs, Veteran Health Administration, Office of Rural Health. Rural veteran health care challenges. Updated February 9, 2021. Accessed June 11, 2021. https://www.ruralhealth.va.gov/aboutus/ruralvets.asp
2. Holder, K.A. Veterans in rural America, 2011–2015. Published January 2017. Accessed June 11, 2021. https://www.census.gov/content/dam/Census/library/publications/2017/acs/acs-36.pdf
3. Pezzin LE, Bogner HR, Kurichi JE, et al. Preventable hospitalizations, barriers to care, and disability. Medicine (Baltimore). 2018;97(19):e0691. doi:10.1097/MD.0000000000010691
4. Rosenbach ML. Access and satisfaction within the disabled Medicare population. Health Care Financ Rev. 1995;17(2):147-167.
5. Semeah LM, Ganesh SP, Wang X, et al. Home modification and health services utilization in rural and urban veterans with disabilities. Housing Policy Debate. 2021. Published online: March 4, 2021. doi:10.1080/10511482.2020.1858923
6. Spoont M, Greer N, Su J, Fitzgerald P, Rutks I, and Wilt TJ. Rural vs. urban ambulatory health care: A Systematic Review. Published May 2011. Accessed June 11, 2021. https://www.hsrd.research.va.gov/publications/esp/ambulatory.pdf
7. Semeah LM, Wang X, Cowper Ripley DC, et al. Improving health through a home modification service for veterans. In: Fiedler BA, ed. Three Facets of Public Health and Paths to Improvements. Academic Press; 2020:381-416.
8. Semeah LM, Ahrentzen S, Jia H, Cowper-Ripley DC, Levy CE, Mann WC. The home improvements and structural alterations benefits program: veterans with disabilities and home accessibility. J Disability Policy Studies. 2017;28(1):43-51. doi:10.1177/1044207317696275
9. Lucas, JW, Zelaya, CE. Hearing difficulty, vision trouble, and balance problems among male veterans and nonveterans. Published June 12, 2020. Accessed June 11, 2021. https://www.cdc.gov/nchs/data/nhsr/nhsr142-508.pdf
10. US Department of Veterans Affairs, National Center for Veterans Analysis and Statistics. Women veterans report: the past, present, and future of women veterans. Published February 2017. Accessed June 11, 2021. https://www.va.gov/vetdata/docs/SpecialReports/Women_Veterans_2015_Final.pdf
11. US Department of Housing and Urban Development, Office of Policy Development and Research. Housing challenges of rural seniors. Published 2017. Accessed June 11, 2021. https://www.huduser.gov/portal/periodicals/em/summer17/highlight1.html
12. Pendall R, Goodman L, Zhu J, Gold A. The future of rural housing. Published October 2016. Accessed June 11, 202.1 https://www.urban.org/sites/default/files/publication/85101/2000972-the-future-of-rural-housing_6.pdf
13. Joint Center for Housing Studies at Harvard University. Improving America’s housing 2019. Published 2019. Accessed June 11, 2021. https://www.jchs.harvard.edu/sites/default/files/reports/files/Harvard_JCHS_Improving_Americas_Housing_2019.pdf
14. Schooley BL, Horan TA, Lee PW, West PA. Rural veteran access to healthcare services: investigating the role of information and communication technologies in overcoming spatial barriers. Perspect Health Inf Manag. 2010;7(Spring):1f. Published 2010 Apr 1.
15. Ripley DC, Kwong PL, Vogel WB, Kurichi JE, Bates BE, Davenport C. How does geographic access affect in-hospital mortality for veterans with acute ischemic stroke?. Med Care. 2015;53(6):501-509. doi:10.1097/MLR.0000000000000366
16. Cowper-Ripley DC, Reker DM, Hayes J, et al. Geographic access to VHA rehabilitation services for traumatically injured veterans. Fed Pract. 2009;26(10):28-39.
17. Smith M, Towne S, Herrera-Venson A, Cameron K, Horel S, Ory M, et al. Delivery of fall prevention interventions for at-risk older adults in rural areas: Findings from a national dissemination. International journal of environmental research and public health. 2018;15:2798. doi: 10.3390/ijerph15122798
18. Hale-Gallardo JL, Kreider CM, Jia H, et al. Telerehabilitation for Rural Veterans: A Qualitative Assessment of Barriers and Facilitators to Implementation. J Multidiscip Healthc. 2020;13:559-570. doi:10.2147/JMDH.S247267
19. Sarfo FS, Akassi J, Kyem G, et al. Long-Term Outcomes of Stroke in a Ghanaian Outpatient Clinic. J Stroke Cerebrovasc Dis. 2018;27(4):1090-1099. doi:10.1016/j.jstrokecerebrovasdis.2017.11.017
1. US Department of Veterans Affairs, Veteran Health Administration, Office of Rural Health. Rural veteran health care challenges. Updated February 9, 2021. Accessed June 11, 2021. https://www.ruralhealth.va.gov/aboutus/ruralvets.asp
2. Holder, K.A. Veterans in rural America, 2011–2015. Published January 2017. Accessed June 11, 2021. https://www.census.gov/content/dam/Census/library/publications/2017/acs/acs-36.pdf
3. Pezzin LE, Bogner HR, Kurichi JE, et al. Preventable hospitalizations, barriers to care, and disability. Medicine (Baltimore). 2018;97(19):e0691. doi:10.1097/MD.0000000000010691
4. Rosenbach ML. Access and satisfaction within the disabled Medicare population. Health Care Financ Rev. 1995;17(2):147-167.
5. Semeah LM, Ganesh SP, Wang X, et al. Home modification and health services utilization in rural and urban veterans with disabilities. Housing Policy Debate. 2021. Published online: March 4, 2021. doi:10.1080/10511482.2020.1858923
6. Spoont M, Greer N, Su J, Fitzgerald P, Rutks I, and Wilt TJ. Rural vs. urban ambulatory health care: A Systematic Review. Published May 2011. Accessed June 11, 2021. https://www.hsrd.research.va.gov/publications/esp/ambulatory.pdf
7. Semeah LM, Wang X, Cowper Ripley DC, et al. Improving health through a home modification service for veterans. In: Fiedler BA, ed. Three Facets of Public Health and Paths to Improvements. Academic Press; 2020:381-416.
8. Semeah LM, Ahrentzen S, Jia H, Cowper-Ripley DC, Levy CE, Mann WC. The home improvements and structural alterations benefits program: veterans with disabilities and home accessibility. J Disability Policy Studies. 2017;28(1):43-51. doi:10.1177/1044207317696275
9. Lucas, JW, Zelaya, CE. Hearing difficulty, vision trouble, and balance problems among male veterans and nonveterans. Published June 12, 2020. Accessed June 11, 2021. https://www.cdc.gov/nchs/data/nhsr/nhsr142-508.pdf
10. US Department of Veterans Affairs, National Center for Veterans Analysis and Statistics. Women veterans report: the past, present, and future of women veterans. Published February 2017. Accessed June 11, 2021. https://www.va.gov/vetdata/docs/SpecialReports/Women_Veterans_2015_Final.pdf
11. US Department of Housing and Urban Development, Office of Policy Development and Research. Housing challenges of rural seniors. Published 2017. Accessed June 11, 2021. https://www.huduser.gov/portal/periodicals/em/summer17/highlight1.html
12. Pendall R, Goodman L, Zhu J, Gold A. The future of rural housing. Published October 2016. Accessed June 11, 202.1 https://www.urban.org/sites/default/files/publication/85101/2000972-the-future-of-rural-housing_6.pdf
13. Joint Center for Housing Studies at Harvard University. Improving America’s housing 2019. Published 2019. Accessed June 11, 2021. https://www.jchs.harvard.edu/sites/default/files/reports/files/Harvard_JCHS_Improving_Americas_Housing_2019.pdf
14. Schooley BL, Horan TA, Lee PW, West PA. Rural veteran access to healthcare services: investigating the role of information and communication technologies in overcoming spatial barriers. Perspect Health Inf Manag. 2010;7(Spring):1f. Published 2010 Apr 1.
15. Ripley DC, Kwong PL, Vogel WB, Kurichi JE, Bates BE, Davenport C. How does geographic access affect in-hospital mortality for veterans with acute ischemic stroke?. Med Care. 2015;53(6):501-509. doi:10.1097/MLR.0000000000000366
16. Cowper-Ripley DC, Reker DM, Hayes J, et al. Geographic access to VHA rehabilitation services for traumatically injured veterans. Fed Pract. 2009;26(10):28-39.
17. Smith M, Towne S, Herrera-Venson A, Cameron K, Horel S, Ory M, et al. Delivery of fall prevention interventions for at-risk older adults in rural areas: Findings from a national dissemination. International journal of environmental research and public health. 2018;15:2798. doi: 10.3390/ijerph15122798
18. Hale-Gallardo JL, Kreider CM, Jia H, et al. Telerehabilitation for Rural Veterans: A Qualitative Assessment of Barriers and Facilitators to Implementation. J Multidiscip Healthc. 2020;13:559-570. doi:10.2147/JMDH.S247267
19. Sarfo FS, Akassi J, Kyem G, et al. Long-Term Outcomes of Stroke in a Ghanaian Outpatient Clinic. J Stroke Cerebrovasc Dis. 2018;27(4):1090-1099. doi:10.1016/j.jstrokecerebrovasdis.2017.11.017
Preoperative Care Assessment of Need Scores Are Associated With Postoperative Mortality and Length of Stay in Veterans Undergoing Knee Replacement
Risk calculators can be of great value in guiding clinical decision making, patient-centered precision medicine, and resource allocation.1 Several perioperative risk prediction models have emerged in recent decades that estimate specific hazards (eg, cardiovascular complications after noncardiac surgery) with varying accuracy and utility. In the perioperative sphere, the time windows are often limited to an index hospitalization or 30 days following surgery or discharge.2-9 Although longer periods are of interest to patients, families, and health systems, few widely used or validated models are designed to look beyond this very narrow window.10,11 In addition, perioperative risk prediction models do not routinely incorporate parameters of a wide variety of health or demographic domains, such as patterns of health care, health care utilization, or medication use.
In 2013, in response to the need for near real-time information to guide delivery of enhanced care management services, the Veterans Health Administration (VHA) Office of Informatics and Analytics developed automated risk prediction models that used detailed electronic health record (EHR) data. These models were used to report Care Assessment Need (CAN) scores each week for all VHA enrollees and include data from a wide array of health domains. These CAN scores predict the risk for hospitalization, death, or either event within 90 days and 1 year.12,13 Each score is reported as both a predicted probability (0-1) and as a percentile in relation to all other VHA enrollees (a value between 1 and 99).13 The data used to calculate CAN scores are listed in Table 1.12
Surgical procedures or admissions would not be differentiated from nonsurgical admissions or other procedural clinic visits, and as such, it is not possible to isolate the effect of undergoing a surgical procedure from another health-related event on the CAN score. At the same time though, a short-term increase in system utilization caused by an elective surgical procedure such as a total knee replacement (TKR) would presumably be reflected in a change in CAN score, but this has not been studied.
Since their introduction, CAN scores have been routinely accessed by primary care teams and used to facilitate care coordination for thousands of VHA patients. However, these CAN scores are currently not available to VHA surgeons, anesthesiologists, or other perioperative clinicians. In this study, we examine the distributions of preoperative CAN scores and explore the relationships of preoperative CAN 1-year mortality scores with 1-year survival following discharge and length of stay (LOS) during index hospitalization in a cohort of US veterans who underwent TKR, the most common elective operation performed within the VHA system.
Methods
Following approval of the Durham Veterans Affairs Medical Center Institutional Review Board, all necessary data were extracted from the VHA Corporate Data Warehouse (CDW) repository.14 Informed consent was waived due to the minimal risk nature of the study.
We used Current Procedural Terminology codes (27438, 27446, 27447, 27486, 27487, 27488) and International Classification of Diseases, 9th edition clinical modification procedure codes (81.54, 81.55, 81.59, 00.80-00.84) to identify all veterans who had undergone primary or revision TKR between July 2014 and December 2015 in VHA Veterans Integrated Service Network 1 (Maine, Vermont, New Hampshire, Massachusetts, Connecticut, Rhode Island, New York, Pennsylvania, West Virginia, Virginia, North Carolina). Because we focused on outcomes following hospital discharge, patients who died before discharge were excluded from the analysis. Preoperative CAN 1-year mortality score was chosen as the measure under the assumption that long-term survival may be the most meaningful of the 4 possible CAN score measures.
Our primary objective was to determine distribution of preoperative CAN scores in the study population. Our secondary was to study relationships among the preoperative CAN 1-year mortality scores and 1-year mortality and hospital LOS.
Study Variables
For each patient, we extracted the date of index surgery. The primary exposure or independent variable was the CAN score in the week prior to this date. Because prior study has shown that CAN scores trajectories do not significantly change over time, the date-stamped CAN scores in the week before surgery represent what would have been available to clinicians in a preoperative setting.15 Since CAN scores are refreshed and overwritten every week, we extracted archived scores from the CDW.
For the 1-year survival outcome, the primary dependent variable, we queried the vital status files in the CDW for the date of death if applicable. We confirmed survival beyond 1 year by examining vital signs in the CDW for a minimum of 2 independent encounters beyond 1 year after the date of discharge. To compute the index LOS, the secondary outcome, we computed the difference between the date of admission and date of hospital discharge.
Statistical Methods
The parameters and performance of the multivariable logistic regression models developed to compute the various CAN mortality and hospitalization risk scores have been previously described.12 Briefly, Wang and colleagues created parsimonious regression models using backward selection. Model discrimination was evaluated using C (concordance)-statistic. Model calibration was assessed by comparing predicted vs observed event rates by risk deciles and performing Cox proportional hazards regression.
We plotted histograms to display preoperative CAN scores as a simple measure of distribution (Figure 1). We also examined the cumulative proportion of patients at each preoperative CAN 1-year mortality score.
Using a conventional t test, we compared means of preoperative CAN 1-year mortality scores in patients who survived vs those who died within 1 year. We also constructed a plot of the proportion of patients who had died within 1 year vs preoperative CAN 1-year mortality scores. Kaplan-Meier curves were then constructed examining 1-year survival by CAN 1-year mortality score by terciles.
Finally, we examined the relationship between preoperative CAN 1-year mortality scores and index LOS in 2 ways: We plotted LOS across CAN scores, and we constructed a
Results
We identified 8206 patients who had undergone a TKR over the 18-month study period. The overall mean (SD) for age was 65 (8.41) years; 93% were male, and 78% were White veterans. Patient demographics are well described in a previous publication.16,17
In terms of model parameters for the CAN score models, C-statistics for the 90-day outcome models were as follows: 0.833 for the model predicting hospitalization (95% CI, 0.832-0.834); 0.865 for the model predicting death (95% CI, 0.863-0.876); and 0.811 for the model predicting either event (95% CI, 0.810-0.812). C-statistics for the 1-year outcome models were 0.809 for the model predicting hospitalization (95% CI, 0.808-0.810); 0.851 for the model predicting death (95% CI, 0.849-0.852); and 0.787 for the model predicting either event (95% CI, 0.786-0.787). Models were well calibrated with α = 0 and β = 1, demonstrating strong agreement between observed and predicted event rates.
The distribution of preoperative CAN 1-year mortality scores was close to normal (median, 50; interquartile range, 40; mean [SD] 48 [25.6]) (eTable). The original CAN score models were developed having an equal number of patients in each strata and as such, are normally distributed.12 Our cohort was similar in pattern of distribution. Distributions of the remaining preoperative CAN scores (90-day mortality, 1-year hospitalization, 90-day hospitalization) are shown in Figures 2, 3, and 4. Not surprisingly, histograms for both 90-day and 1-year hospitalization were skewed toward higher scores, indicating that these patients were expected to be hospitalized in the near future.
Overall, 1.4% (110/8096) of patients died within 1 year of surgery. Comparing 1-year mortality CAN scores in survivors vs nonsurvivors, we found statistically significant differences in means (47 vs 66 respectively, P < .001) and medians (45 vs 75 respectively, P < .001) (Table 2). In the plot examining the relationship between preoperative 1-year mortality CAN scores and 1-year mortality, the percentage who died within 1 year increased initially for patients with CAN scores > 60 and again exponentially for patients with CAN scores > 80. Examining Kaplan-Meier curves, we found that survivors and nonsurvivors separated early after surgery, and the differences between the top tercile and the middle/lower terciles were statistically significant (P < .001). Mortality rates were about 0.5% in the lower and middle terciles but about 2% in the upper tercile (Figure 5).
In the plot examining the relationship between CAN scores and index LOS, the LOS rose significantly beyond a CAN score of 60 and dramatically beyond a CAN score of 80 (Figure 6). LOESS curves also showed 2 inflection points suggesting an incremental and sequential rise in the LOS with increasing CAN scores (Figure 7). Mean (SD) LOS in days for the lowest to highest terciles was 2.6 (1.7), 2.8 (2.1), and 3.6 (2.2), respectively.
Discussion
CAN scores are automatically generated each week by EHR-based multivariable risk models. These scores have excellent predictive accuracy for 90-day and 1-year mortality and hospitalization and are routinely used by VHA primary care teams to assist with clinical operations.13 We studied the distribution of CAN 1-year mortality scores in a preoperative context and examined relationships of the preoperative CAN 1-year mortality scores with postoperative mortality and LOS in 8206 veterans who underwent TKR.
There are several noteworthy findings. First, the overall 1-year mortality rate observed following TKR (1.4%) was similar to other published reports.18,19 Not surprisingly, preoperative CAN 1-year mortality scores were significantly higher in veterans who died compared with those of survivors. The majority of patients who died had a preoperative CAN 1-year mortality score > 75 while most who survived had a preoperative CAN 1-year mortality score < 45 (P < .001). Interestingly, the same scores showed a nonlinear correlation with LOS. Index LOS was about 4 days in patients in the highest tercile of CAN scores vs 2.5 days in the lowest tercile, but the initial increase in LOS was detected at a CAN score of about 55 to 60.
In addition, mortality rate varied widely in different segments of the population when grouped according to preoperative CAN scores. One-year mortality rates in the highest tercile reached 2%, about 4-fold higher than that of lower terciles (0.5%). Examination of the Kaplan-Meier curves showed that this difference in mortality between the highest tercile and the lower 2 groups appears soon after discharge and continues to increase over time, suggesting that the factors contributing to the increased mortality are present at the time of discharge and persist beyond the postoperative period. In summary, although CAN scores were not designed for use in the perioperative context, we found that preoperative CAN 1-year mortality scores are broadly predictive of mortality, but especially for increases in LOS following elective TKA, both increases in hospital LOS following elective TKA and mortality over the year after TKA.
Our findings raise several important questions. The decision to undergo elective surgery is complex. Arguably, individuals who undergo elective knee replacement should be healthy enough to undergo, recover, and reap the benefits from a procedure that does not extend life. The distribution of preoperative CAN 1-year mortality scores for our study population was similar to that of the general VHA enrollee population with similar measured mortality rates (≤ 0.5% vs ≥ 1.7% in the low and high terciles, respectively).1 Further study comparing outcomes in matched cohorts who did and did not undergo joint replacement would be of interest. In lieu of this, though, the association of high but not extreme CAN scores with increased hospital LOS may potentially be used to guide allocation of resources to this group, obviating the increased cost and risk to which this group is exposed. And the additional insight afforded by CAN scores may enhance shared decision-making models by identifying patients at the very highest risk (eg, 1-year mortality CAN score ≥ 90), patients who conceivably might not survive long enough to recover from and enjoy their reconstructed knee, who might in the long run be harmed by undergoing the procedure.
Many total joint arthroplasties are performed in older patients, a population in which frailty is increasingly recognized as a significant risk factor for poor outcomes.20,21 CAN scores reliably identify high-risk patients and have been shown to correlate with frailty in this group.22 Multiple authors have reported improved outcomes with cost reductions after implementation of programs targeting modifiable risk factors in high-risk surgical candidates.23-25 A preoperative assessment that includes the CAN score may be valuable in identifying patients who would benefit most from prehabilitation programs or other interventions designed to blunt the impact of frailty. It is true that many elements used to calculate the CAN score would not be considered modifiable, especially in the short term. However, specific contributors to frailty, such as nutritional status and polypharmacy might be potential candidates. As with all multivariable risk prediction models, there are multiple paths to a high CAN score, and further research to identify clinically relevant subgroups may help inform efforts to improve perioperative care within this population.
Hospital LOS is of intense interest for many reasons, not least its utility as a surrogate for cost and increased risk for immediate perioperative adverse events, such as multidrug-resistant hospital acquired infections, need for postacute facility-based rehabilitation, and deconditioning that increase risks of falls and fractures in the older population.26-29 In addition, its importance is magnified due to the COVID-19 pandemic context in which restarting elective surgery programs has changed traditional criteria by which patients are scheduled for surgery.
We have shown that elevated CAN scores are able to identify patients at risk for extended hospital stays and, as such, may be useful additional data in allocating scarce operating room time and other resources for optimal patient and health care provider safety.30,31 Individual surgeons and hospital systems would, of course, decide which patients should be triaged to go first, based on local priorities; however, choosing lower risk patients with minimal risk of morbidity and mortality while pursuing prehabilitation for higher risk patients is a reasonable approach.
Limitations
Our study has several limitations. Only a single surgical procedure was included, albeit the most common one performed in the VHA. In addition, no information was available concerning the precise clinical course for these patients, such as the duration of surgery, anesthetic technique, and management of acute, perioperative course. Although the assumption was made that patients received standard care in a manner such that these factors would not significantly affect either their mortality or their LOS out of proportion to their preoperative clinical status, confounding cannot be excluded. Therefore, further study is necessary to determine whether CAN scores can accurately predict mortality and/or LOS for patients undergoing other procedures. Further, a clinical trial is required to assess whether systematic provision of the CAN score at the point of surgery would impact care and, more important, impact outcomes. In addition, multivariable analyses were not performed, including and excluding various components of the CAN score models. Currently, CAN scores could be made available to the surgical/anesthesia communities at minimal or no cost and are updated automatically. Model calibration and discrimination in this particular setting were not validated.
Because our interest is in leveraging an existing resource to a current clinical and operational problem rather than in creating or validating a new tool, we chose to test the simple bivariate relationship between preoperative CAN scores and outcomes. We chose the preoperative 1-year mortality CAN score from among the 4 options under the assumption that long-term survival is the most meaningful of the 4 candidate outcomes. Finally, while the CAN scores are currently only calculated and generated for patients cared for within the VHA, few data elements are unavailable to civilian health systems. The most problematic would be documentation of actual prescription filling, but this is a topic of increasing interest to the medical and academic communities and access to such information we hope will improve.32-34
Conclusions
Although designed for use by VHA primary care teams, CAN scores also may have value for perioperative clinicians, predicting mortality and prolonged hospital LOS in those with elevated 1-year mortality scores. Advantages of CAN scores relative to other perioperative risk calculators lies in their ability to predict long-term rather than 30-day survival and that they are automatically generated on a near-real-time basis for all patients who receive care in VHA ambulatory clinics. Further study is needed to determine practical utility in shared decision making, preoperative evaluation and optimization, and perioperative resource allocation.
Acknowledgments
This work was supported by the US Department of Veterans Affairs (VA) National Center for Patient Safety, Field Office 10A4E, through the Patient Safety Center of Inquiry at the Durham VA Medical Center in North Carolina. The study also received support from the Center of Innovation to Accelerate Discovery and Practice Transformation (CIN 13-410) at the Durham VA Health Care System.
1. McNair AGK, MacKichan F, Donovan JL, et al. What surgeons tell patients and what patients want to know before major cancer surgery: a qualitative study. BMC Cancer. 2016;16:258. doi:10.1186/s12885-016-2292-3
2. Grover FL, Hammermeister KE, Burchfiel C. Initial report of the Veterans Administration Preoperative Risk Assessment Study for Cardiac Surgery. Ann Thorac Surg. 1990;50(1):12-26; discussion 27-18. doi:10.1016/0003-4975(90)90073-f
3. Khuri SF, Daley J, Henderson W, et al. The National Veterans Administration Surgical Risk Study: risk adjustment for the comparative assessment of the quality of surgical care. J Am Coll Surg. 1995;180(5):519-531.
4. Glance LG, Lustik SJ, Hannan EL, et al. The Surgical Mortality Probability Model: derivation and validation of a simple simple risk prediction rule for noncardiac surgery. Ann Surg. 2012;255(4):696-702. doi:10.1097/SLA.0b013e31824b45af
5. Keller DS, Kroll D, Papaconstantinou HT, Ellis CN. Development and validation of a methodology to reduce mortality using the veterans affairs surgical quality improvement program risk calculator. J Am Coll Surg. 2017;224(4):602-607. doi:10.1016/j.jamcollsurg.2016.12.033
6. Bilimoria KY, Liu Y, Paruch JL, et al. Development and evaluation of the universal ACS NSQIP surgical risk calculator: a decision aid and informed consent tool for patients and surgeons. J Am Coll Surg. 2013;217(5):833-842.e831-833. doi:10.1016/j.jamcollsurg.2013.07.385
7. Ford MK, Beattie WS, Wijeysundera DN. Systematic review: prediction of perioperative cardiac complications and mortality by the revised cardiac risk index. Ann Intern Med. 2010;152(1):26-35. doi:10.7326/0003-4819-152-1-201001050-00007
8. Gupta PK, Gupta H, Sundaram A, et al. Development and validation of a risk calculator for prediction of cardiac risk after surgery. Circulation. 2011;124(4):381-387. doi:10.1161/CIRCULATIONAHA.110.015701
9. Lee TH, Marcantonio ER, Mangione CM, et al. Derivation and prospective validation of a simple index for prediction of cardiac risk of major noncardiac surgery. Circulation. 1999;100(10):1043-1049. doi:10.1161/01.cir.100.10.1043
10. Smith T, Li X, Nylander W, Gunnar W. Thirty-day postoperative mortality risk estimates and 1-year survival in Veterans Health Administration surgery patients. JAMA Surg. 2016;151(5):417-422. doi:10.1001/jamasurg.2015.4882
11. Damhuis RA, Wijnhoven BP, Plaisier PW, Kirkels WJ, Kranse R, van Lanschot JJ. Comparison of 30-day, 90- day and in-hospital postoperative mortality for eight different cancer types. Br J Surg. 2012;99(8):1149-1154. doi:10.1002/bjs.8813
12. Wang L, Porter B, Maynard C, et al. Predicting risk of hospitalization or death among patients receiving primary care in the Veterans Health Administration. Med Care. 2013;51(4):368-373. doi:10.1016/j.amjcard.2012.06.038
13. Fihn SD, Francis J, Clancy C, et al. Insights from advanced analytics at the Veterans Health Administration. Health Aff (Millwood). 2014;33(7):1203-1211. doi:10.1377/hlthaff.2014.0054
14. Noël PH, Copeland LA, Perrin RA, et al. VHA Corporate Data Warehouse height and weight data: opportunities and challenges for health services research. J Rehabil Res Dev. 2010;47(8):739-750. doi:10.1682/jrrd.2009.08.0110
15. Wong ES, Yoon J, Piegari RI, Rosland AM, Fihn SD, Chang ET. Identifying latent subgroups of high-risk patients using risk score trajectories. J Gen Intern Med. 2018;33(12):2120-2126. doi:10.1007/s11606-018-4653-x
16. Chen Q, Hsia HL, Overman R, et al. Impact of an opioid safety initiative on patients undergoing total knee arthroplasty: a time series analysis. Anesthesiology. 2019;131(2):369-380. doi:10.1097/ALN.0000000000002771
17. Hsia HL, Takemoto S, van de Ven T, et al. Acute pain is associated with chronic opioid use after total knee arthroplasty. Reg Anesth Pain Med. 2018;43(7):705-711. doi:10.1097/AAP.0000000000000831
18. Inacio MCS, Dillon MT, Miric A, Navarro RA, Paxton EW. Mortality after total knee and total hip arthroplasty in a large integrated health care system. Perm J. 2017;21:16-171. doi:10.7812/TPP/16-171
19. Lee QJ, Mak WP, Wong YC. Mortality following primary total knee replacement in public hospitals in Hong Kong. Hong Kong Med J. 2016;22(3):237-241. doi:10.12809/hkmj154712
20. Lin HS, Watts JN, Peel NM, Hubbard RE. Frailty and post-operative outcomes in older surgical patients: a systematic review. BMC Geriatr. 2016;16(1):157. doi:10.1186/s12877-016-0329-8
21. Shinall MC Jr, Arya S, Youk A, et al. Association of preoperative patient frailty and operative stress with postoperative mortality. JAMA Surg. 2019;155(1):e194620. doi:10.1001/jamasurg.2019.4620
22. Ruiz JG, Priyadarshni S, Rahaman Z, et al. Validation of an automatically generated screening score for frailty: the care assessment need (CAN) score. BMC Geriatr. 2018;18(1):106. doi:10.1186/s12877-018-0802-7
23. Bernstein DN, Liu TC, Winegar AL, et al. Evaluation of a preoperative optimization protocol for primary hip and knee arthroplasty patients. J Arthroplasty. 2018;33(12):3642- 3648. doi:10.1016/j.arth.2018.08.018
24. Sodhi N, Anis HK, Coste M, et al. A nationwide analysis of preoperative planning on operative times and postoperative complications in total knee arthroplasty. J Knee Surg. 2019;32(11):1040-1045. doi:10.1055/s-0039-1677790
25. Krause A, Sayeed Z, El-Othmani M, Pallekonda V, Mihalko W, Saleh KJ. Outpatient total knee arthroplasty: are we there yet? (part 1). Orthop Clin North Am. 2018;49(1):1-6. doi:10.1016/j.ocl.2017.08.002
26. Barrasa-Villar JI, Aibar-Remón C, Prieto-Andrés P, Mareca- Doñate R, Moliner-Lahoz J. Impact on morbidity, mortality, and length of stay of hospital-acquired infections by resistant microorganisms. Clin Infect Dis. 2017;65(4):644-652. doi:10.1093/cid/cix411
27. Nikkel LE, Kates SL, Schreck M, Maceroli M, Mahmood B, Elfar JC. Length of hospital stay after hip fracture and risk of early mortality after discharge in New York state: retrospective cohort study. BMJ. 2015;351:h6246. doi:10.1136/bmj.h6246
28. Marfil-Garza BA, Belaunzarán-Zamudio PF, Gulias-Herrero A, et al. Risk factors associated with prolonged hospital length-of-stay: 18-year retrospective study of hospitalizations in a tertiary healthcare center in Mexico. PLoS One. 2018;13(11):e0207203. doi:10.1371/journal.pone.0207203
29. Hirsch CH, Sommers L, Olsen A, Mullen L, Winograd CH. The natural history of functional morbidity in hospitalized older patients. J Am Geriatr Soc. 1990;38(12):1296-1303. doi:10.1111/j.1532-5415.1990.tb03451.x
30. Iyengar KP, Jain VK, Vaish A, Vaishya R, Maini L, Lal H. Post COVID-19: planning strategies to resume orthopaedic surgery -challenges and considerations. J Clin Orthop Trauma. 2020;11(suppl 3):S291-S295. doi:10.1016/j.jcot.2020.04.028
31. O’Connor CM, Anoushiravani AA, DiCaprio MR, Healy WL, Iorio R. Economic recovery after the COVID-19 pandemic: resuming elective orthopedic surgery and total joint arthroplasty. J Arthroplasty. 2020;35(suppl 7):S32-S36. doi:10.1016/j.arth.2020.04.038.
32. Mauseth SA, Skurtveit S, Skovlund E, Langhammer A, Spigset O. Medication use and association with urinary incontinence in women: data from the Norwegian Prescription Database and the HUNT study. Neurourol Urodyn. 2018;37(4):1448-1457. doi:10.1002/nau.23473
33. Sultan RS, Correll CU, Schoenbaum M, King M, Walkup JT, Olfson M. National patterns of commonly prescribed psychotropic medications to young people. J Child Adolesc Psychopharmacol. 2018;28(3):158-165. doi:10.1089/cap.2017.0077
34. McCoy RG, Dykhoff HJ, Sangaralingham L, et al. Adoption of new glucose-lowering medications in the U.S.-the case of SGLT2 inhibitors: nationwide cohort study. Diabetes Technol Ther. 2019;21(12):702-712. doi:10.1089/dia.2019.0213
Risk calculators can be of great value in guiding clinical decision making, patient-centered precision medicine, and resource allocation.1 Several perioperative risk prediction models have emerged in recent decades that estimate specific hazards (eg, cardiovascular complications after noncardiac surgery) with varying accuracy and utility. In the perioperative sphere, the time windows are often limited to an index hospitalization or 30 days following surgery or discharge.2-9 Although longer periods are of interest to patients, families, and health systems, few widely used or validated models are designed to look beyond this very narrow window.10,11 In addition, perioperative risk prediction models do not routinely incorporate parameters of a wide variety of health or demographic domains, such as patterns of health care, health care utilization, or medication use.
In 2013, in response to the need for near real-time information to guide delivery of enhanced care management services, the Veterans Health Administration (VHA) Office of Informatics and Analytics developed automated risk prediction models that used detailed electronic health record (EHR) data. These models were used to report Care Assessment Need (CAN) scores each week for all VHA enrollees and include data from a wide array of health domains. These CAN scores predict the risk for hospitalization, death, or either event within 90 days and 1 year.12,13 Each score is reported as both a predicted probability (0-1) and as a percentile in relation to all other VHA enrollees (a value between 1 and 99).13 The data used to calculate CAN scores are listed in Table 1.12
Surgical procedures or admissions would not be differentiated from nonsurgical admissions or other procedural clinic visits, and as such, it is not possible to isolate the effect of undergoing a surgical procedure from another health-related event on the CAN score. At the same time though, a short-term increase in system utilization caused by an elective surgical procedure such as a total knee replacement (TKR) would presumably be reflected in a change in CAN score, but this has not been studied.
Since their introduction, CAN scores have been routinely accessed by primary care teams and used to facilitate care coordination for thousands of VHA patients. However, these CAN scores are currently not available to VHA surgeons, anesthesiologists, or other perioperative clinicians. In this study, we examine the distributions of preoperative CAN scores and explore the relationships of preoperative CAN 1-year mortality scores with 1-year survival following discharge and length of stay (LOS) during index hospitalization in a cohort of US veterans who underwent TKR, the most common elective operation performed within the VHA system.
Methods
Following approval of the Durham Veterans Affairs Medical Center Institutional Review Board, all necessary data were extracted from the VHA Corporate Data Warehouse (CDW) repository.14 Informed consent was waived due to the minimal risk nature of the study.
We used Current Procedural Terminology codes (27438, 27446, 27447, 27486, 27487, 27488) and International Classification of Diseases, 9th edition clinical modification procedure codes (81.54, 81.55, 81.59, 00.80-00.84) to identify all veterans who had undergone primary or revision TKR between July 2014 and December 2015 in VHA Veterans Integrated Service Network 1 (Maine, Vermont, New Hampshire, Massachusetts, Connecticut, Rhode Island, New York, Pennsylvania, West Virginia, Virginia, North Carolina). Because we focused on outcomes following hospital discharge, patients who died before discharge were excluded from the analysis. Preoperative CAN 1-year mortality score was chosen as the measure under the assumption that long-term survival may be the most meaningful of the 4 possible CAN score measures.
Our primary objective was to determine distribution of preoperative CAN scores in the study population. Our secondary was to study relationships among the preoperative CAN 1-year mortality scores and 1-year mortality and hospital LOS.
Study Variables
For each patient, we extracted the date of index surgery. The primary exposure or independent variable was the CAN score in the week prior to this date. Because prior study has shown that CAN scores trajectories do not significantly change over time, the date-stamped CAN scores in the week before surgery represent what would have been available to clinicians in a preoperative setting.15 Since CAN scores are refreshed and overwritten every week, we extracted archived scores from the CDW.
For the 1-year survival outcome, the primary dependent variable, we queried the vital status files in the CDW for the date of death if applicable. We confirmed survival beyond 1 year by examining vital signs in the CDW for a minimum of 2 independent encounters beyond 1 year after the date of discharge. To compute the index LOS, the secondary outcome, we computed the difference between the date of admission and date of hospital discharge.
Statistical Methods
The parameters and performance of the multivariable logistic regression models developed to compute the various CAN mortality and hospitalization risk scores have been previously described.12 Briefly, Wang and colleagues created parsimonious regression models using backward selection. Model discrimination was evaluated using C (concordance)-statistic. Model calibration was assessed by comparing predicted vs observed event rates by risk deciles and performing Cox proportional hazards regression.
We plotted histograms to display preoperative CAN scores as a simple measure of distribution (Figure 1). We also examined the cumulative proportion of patients at each preoperative CAN 1-year mortality score.
Using a conventional t test, we compared means of preoperative CAN 1-year mortality scores in patients who survived vs those who died within 1 year. We also constructed a plot of the proportion of patients who had died within 1 year vs preoperative CAN 1-year mortality scores. Kaplan-Meier curves were then constructed examining 1-year survival by CAN 1-year mortality score by terciles.
Finally, we examined the relationship between preoperative CAN 1-year mortality scores and index LOS in 2 ways: We plotted LOS across CAN scores, and we constructed a
Results
We identified 8206 patients who had undergone a TKR over the 18-month study period. The overall mean (SD) for age was 65 (8.41) years; 93% were male, and 78% were White veterans. Patient demographics are well described in a previous publication.16,17
In terms of model parameters for the CAN score models, C-statistics for the 90-day outcome models were as follows: 0.833 for the model predicting hospitalization (95% CI, 0.832-0.834); 0.865 for the model predicting death (95% CI, 0.863-0.876); and 0.811 for the model predicting either event (95% CI, 0.810-0.812). C-statistics for the 1-year outcome models were 0.809 for the model predicting hospitalization (95% CI, 0.808-0.810); 0.851 for the model predicting death (95% CI, 0.849-0.852); and 0.787 for the model predicting either event (95% CI, 0.786-0.787). Models were well calibrated with α = 0 and β = 1, demonstrating strong agreement between observed and predicted event rates.
The distribution of preoperative CAN 1-year mortality scores was close to normal (median, 50; interquartile range, 40; mean [SD] 48 [25.6]) (eTable). The original CAN score models were developed having an equal number of patients in each strata and as such, are normally distributed.12 Our cohort was similar in pattern of distribution. Distributions of the remaining preoperative CAN scores (90-day mortality, 1-year hospitalization, 90-day hospitalization) are shown in Figures 2, 3, and 4. Not surprisingly, histograms for both 90-day and 1-year hospitalization were skewed toward higher scores, indicating that these patients were expected to be hospitalized in the near future.
Overall, 1.4% (110/8096) of patients died within 1 year of surgery. Comparing 1-year mortality CAN scores in survivors vs nonsurvivors, we found statistically significant differences in means (47 vs 66 respectively, P < .001) and medians (45 vs 75 respectively, P < .001) (Table 2). In the plot examining the relationship between preoperative 1-year mortality CAN scores and 1-year mortality, the percentage who died within 1 year increased initially for patients with CAN scores > 60 and again exponentially for patients with CAN scores > 80. Examining Kaplan-Meier curves, we found that survivors and nonsurvivors separated early after surgery, and the differences between the top tercile and the middle/lower terciles were statistically significant (P < .001). Mortality rates were about 0.5% in the lower and middle terciles but about 2% in the upper tercile (Figure 5).
In the plot examining the relationship between CAN scores and index LOS, the LOS rose significantly beyond a CAN score of 60 and dramatically beyond a CAN score of 80 (Figure 6). LOESS curves also showed 2 inflection points suggesting an incremental and sequential rise in the LOS with increasing CAN scores (Figure 7). Mean (SD) LOS in days for the lowest to highest terciles was 2.6 (1.7), 2.8 (2.1), and 3.6 (2.2), respectively.
Discussion
CAN scores are automatically generated each week by EHR-based multivariable risk models. These scores have excellent predictive accuracy for 90-day and 1-year mortality and hospitalization and are routinely used by VHA primary care teams to assist with clinical operations.13 We studied the distribution of CAN 1-year mortality scores in a preoperative context and examined relationships of the preoperative CAN 1-year mortality scores with postoperative mortality and LOS in 8206 veterans who underwent TKR.
There are several noteworthy findings. First, the overall 1-year mortality rate observed following TKR (1.4%) was similar to other published reports.18,19 Not surprisingly, preoperative CAN 1-year mortality scores were significantly higher in veterans who died compared with those of survivors. The majority of patients who died had a preoperative CAN 1-year mortality score > 75 while most who survived had a preoperative CAN 1-year mortality score < 45 (P < .001). Interestingly, the same scores showed a nonlinear correlation with LOS. Index LOS was about 4 days in patients in the highest tercile of CAN scores vs 2.5 days in the lowest tercile, but the initial increase in LOS was detected at a CAN score of about 55 to 60.
In addition, mortality rate varied widely in different segments of the population when grouped according to preoperative CAN scores. One-year mortality rates in the highest tercile reached 2%, about 4-fold higher than that of lower terciles (0.5%). Examination of the Kaplan-Meier curves showed that this difference in mortality between the highest tercile and the lower 2 groups appears soon after discharge and continues to increase over time, suggesting that the factors contributing to the increased mortality are present at the time of discharge and persist beyond the postoperative period. In summary, although CAN scores were not designed for use in the perioperative context, we found that preoperative CAN 1-year mortality scores are broadly predictive of mortality, but especially for increases in LOS following elective TKA, both increases in hospital LOS following elective TKA and mortality over the year after TKA.
Our findings raise several important questions. The decision to undergo elective surgery is complex. Arguably, individuals who undergo elective knee replacement should be healthy enough to undergo, recover, and reap the benefits from a procedure that does not extend life. The distribution of preoperative CAN 1-year mortality scores for our study population was similar to that of the general VHA enrollee population with similar measured mortality rates (≤ 0.5% vs ≥ 1.7% in the low and high terciles, respectively).1 Further study comparing outcomes in matched cohorts who did and did not undergo joint replacement would be of interest. In lieu of this, though, the association of high but not extreme CAN scores with increased hospital LOS may potentially be used to guide allocation of resources to this group, obviating the increased cost and risk to which this group is exposed. And the additional insight afforded by CAN scores may enhance shared decision-making models by identifying patients at the very highest risk (eg, 1-year mortality CAN score ≥ 90), patients who conceivably might not survive long enough to recover from and enjoy their reconstructed knee, who might in the long run be harmed by undergoing the procedure.
Many total joint arthroplasties are performed in older patients, a population in which frailty is increasingly recognized as a significant risk factor for poor outcomes.20,21 CAN scores reliably identify high-risk patients and have been shown to correlate with frailty in this group.22 Multiple authors have reported improved outcomes with cost reductions after implementation of programs targeting modifiable risk factors in high-risk surgical candidates.23-25 A preoperative assessment that includes the CAN score may be valuable in identifying patients who would benefit most from prehabilitation programs or other interventions designed to blunt the impact of frailty. It is true that many elements used to calculate the CAN score would not be considered modifiable, especially in the short term. However, specific contributors to frailty, such as nutritional status and polypharmacy might be potential candidates. As with all multivariable risk prediction models, there are multiple paths to a high CAN score, and further research to identify clinically relevant subgroups may help inform efforts to improve perioperative care within this population.
Hospital LOS is of intense interest for many reasons, not least its utility as a surrogate for cost and increased risk for immediate perioperative adverse events, such as multidrug-resistant hospital acquired infections, need for postacute facility-based rehabilitation, and deconditioning that increase risks of falls and fractures in the older population.26-29 In addition, its importance is magnified due to the COVID-19 pandemic context in which restarting elective surgery programs has changed traditional criteria by which patients are scheduled for surgery.
We have shown that elevated CAN scores are able to identify patients at risk for extended hospital stays and, as such, may be useful additional data in allocating scarce operating room time and other resources for optimal patient and health care provider safety.30,31 Individual surgeons and hospital systems would, of course, decide which patients should be triaged to go first, based on local priorities; however, choosing lower risk patients with minimal risk of morbidity and mortality while pursuing prehabilitation for higher risk patients is a reasonable approach.
Limitations
Our study has several limitations. Only a single surgical procedure was included, albeit the most common one performed in the VHA. In addition, no information was available concerning the precise clinical course for these patients, such as the duration of surgery, anesthetic technique, and management of acute, perioperative course. Although the assumption was made that patients received standard care in a manner such that these factors would not significantly affect either their mortality or their LOS out of proportion to their preoperative clinical status, confounding cannot be excluded. Therefore, further study is necessary to determine whether CAN scores can accurately predict mortality and/or LOS for patients undergoing other procedures. Further, a clinical trial is required to assess whether systematic provision of the CAN score at the point of surgery would impact care and, more important, impact outcomes. In addition, multivariable analyses were not performed, including and excluding various components of the CAN score models. Currently, CAN scores could be made available to the surgical/anesthesia communities at minimal or no cost and are updated automatically. Model calibration and discrimination in this particular setting were not validated.
Because our interest is in leveraging an existing resource to a current clinical and operational problem rather than in creating or validating a new tool, we chose to test the simple bivariate relationship between preoperative CAN scores and outcomes. We chose the preoperative 1-year mortality CAN score from among the 4 options under the assumption that long-term survival is the most meaningful of the 4 candidate outcomes. Finally, while the CAN scores are currently only calculated and generated for patients cared for within the VHA, few data elements are unavailable to civilian health systems. The most problematic would be documentation of actual prescription filling, but this is a topic of increasing interest to the medical and academic communities and access to such information we hope will improve.32-34
Conclusions
Although designed for use by VHA primary care teams, CAN scores also may have value for perioperative clinicians, predicting mortality and prolonged hospital LOS in those with elevated 1-year mortality scores. Advantages of CAN scores relative to other perioperative risk calculators lies in their ability to predict long-term rather than 30-day survival and that they are automatically generated on a near-real-time basis for all patients who receive care in VHA ambulatory clinics. Further study is needed to determine practical utility in shared decision making, preoperative evaluation and optimization, and perioperative resource allocation.
Acknowledgments
This work was supported by the US Department of Veterans Affairs (VA) National Center for Patient Safety, Field Office 10A4E, through the Patient Safety Center of Inquiry at the Durham VA Medical Center in North Carolina. The study also received support from the Center of Innovation to Accelerate Discovery and Practice Transformation (CIN 13-410) at the Durham VA Health Care System.
Risk calculators can be of great value in guiding clinical decision making, patient-centered precision medicine, and resource allocation.1 Several perioperative risk prediction models have emerged in recent decades that estimate specific hazards (eg, cardiovascular complications after noncardiac surgery) with varying accuracy and utility. In the perioperative sphere, the time windows are often limited to an index hospitalization or 30 days following surgery or discharge.2-9 Although longer periods are of interest to patients, families, and health systems, few widely used or validated models are designed to look beyond this very narrow window.10,11 In addition, perioperative risk prediction models do not routinely incorporate parameters of a wide variety of health or demographic domains, such as patterns of health care, health care utilization, or medication use.
In 2013, in response to the need for near real-time information to guide delivery of enhanced care management services, the Veterans Health Administration (VHA) Office of Informatics and Analytics developed automated risk prediction models that used detailed electronic health record (EHR) data. These models were used to report Care Assessment Need (CAN) scores each week for all VHA enrollees and include data from a wide array of health domains. These CAN scores predict the risk for hospitalization, death, or either event within 90 days and 1 year.12,13 Each score is reported as both a predicted probability (0-1) and as a percentile in relation to all other VHA enrollees (a value between 1 and 99).13 The data used to calculate CAN scores are listed in Table 1.12
Surgical procedures or admissions would not be differentiated from nonsurgical admissions or other procedural clinic visits, and as such, it is not possible to isolate the effect of undergoing a surgical procedure from another health-related event on the CAN score. At the same time though, a short-term increase in system utilization caused by an elective surgical procedure such as a total knee replacement (TKR) would presumably be reflected in a change in CAN score, but this has not been studied.
Since their introduction, CAN scores have been routinely accessed by primary care teams and used to facilitate care coordination for thousands of VHA patients. However, these CAN scores are currently not available to VHA surgeons, anesthesiologists, or other perioperative clinicians. In this study, we examine the distributions of preoperative CAN scores and explore the relationships of preoperative CAN 1-year mortality scores with 1-year survival following discharge and length of stay (LOS) during index hospitalization in a cohort of US veterans who underwent TKR, the most common elective operation performed within the VHA system.
Methods
Following approval of the Durham Veterans Affairs Medical Center Institutional Review Board, all necessary data were extracted from the VHA Corporate Data Warehouse (CDW) repository.14 Informed consent was waived due to the minimal risk nature of the study.
We used Current Procedural Terminology codes (27438, 27446, 27447, 27486, 27487, 27488) and International Classification of Diseases, 9th edition clinical modification procedure codes (81.54, 81.55, 81.59, 00.80-00.84) to identify all veterans who had undergone primary or revision TKR between July 2014 and December 2015 in VHA Veterans Integrated Service Network 1 (Maine, Vermont, New Hampshire, Massachusetts, Connecticut, Rhode Island, New York, Pennsylvania, West Virginia, Virginia, North Carolina). Because we focused on outcomes following hospital discharge, patients who died before discharge were excluded from the analysis. Preoperative CAN 1-year mortality score was chosen as the measure under the assumption that long-term survival may be the most meaningful of the 4 possible CAN score measures.
Our primary objective was to determine distribution of preoperative CAN scores in the study population. Our secondary was to study relationships among the preoperative CAN 1-year mortality scores and 1-year mortality and hospital LOS.
Study Variables
For each patient, we extracted the date of index surgery. The primary exposure or independent variable was the CAN score in the week prior to this date. Because prior study has shown that CAN scores trajectories do not significantly change over time, the date-stamped CAN scores in the week before surgery represent what would have been available to clinicians in a preoperative setting.15 Since CAN scores are refreshed and overwritten every week, we extracted archived scores from the CDW.
For the 1-year survival outcome, the primary dependent variable, we queried the vital status files in the CDW for the date of death if applicable. We confirmed survival beyond 1 year by examining vital signs in the CDW for a minimum of 2 independent encounters beyond 1 year after the date of discharge. To compute the index LOS, the secondary outcome, we computed the difference between the date of admission and date of hospital discharge.
Statistical Methods
The parameters and performance of the multivariable logistic regression models developed to compute the various CAN mortality and hospitalization risk scores have been previously described.12 Briefly, Wang and colleagues created parsimonious regression models using backward selection. Model discrimination was evaluated using C (concordance)-statistic. Model calibration was assessed by comparing predicted vs observed event rates by risk deciles and performing Cox proportional hazards regression.
We plotted histograms to display preoperative CAN scores as a simple measure of distribution (Figure 1). We also examined the cumulative proportion of patients at each preoperative CAN 1-year mortality score.
Using a conventional t test, we compared means of preoperative CAN 1-year mortality scores in patients who survived vs those who died within 1 year. We also constructed a plot of the proportion of patients who had died within 1 year vs preoperative CAN 1-year mortality scores. Kaplan-Meier curves were then constructed examining 1-year survival by CAN 1-year mortality score by terciles.
Finally, we examined the relationship between preoperative CAN 1-year mortality scores and index LOS in 2 ways: We plotted LOS across CAN scores, and we constructed a
Results
We identified 8206 patients who had undergone a TKR over the 18-month study period. The overall mean (SD) for age was 65 (8.41) years; 93% were male, and 78% were White veterans. Patient demographics are well described in a previous publication.16,17
In terms of model parameters for the CAN score models, C-statistics for the 90-day outcome models were as follows: 0.833 for the model predicting hospitalization (95% CI, 0.832-0.834); 0.865 for the model predicting death (95% CI, 0.863-0.876); and 0.811 for the model predicting either event (95% CI, 0.810-0.812). C-statistics for the 1-year outcome models were 0.809 for the model predicting hospitalization (95% CI, 0.808-0.810); 0.851 for the model predicting death (95% CI, 0.849-0.852); and 0.787 for the model predicting either event (95% CI, 0.786-0.787). Models were well calibrated with α = 0 and β = 1, demonstrating strong agreement between observed and predicted event rates.
The distribution of preoperative CAN 1-year mortality scores was close to normal (median, 50; interquartile range, 40; mean [SD] 48 [25.6]) (eTable). The original CAN score models were developed having an equal number of patients in each strata and as such, are normally distributed.12 Our cohort was similar in pattern of distribution. Distributions of the remaining preoperative CAN scores (90-day mortality, 1-year hospitalization, 90-day hospitalization) are shown in Figures 2, 3, and 4. Not surprisingly, histograms for both 90-day and 1-year hospitalization were skewed toward higher scores, indicating that these patients were expected to be hospitalized in the near future.
Overall, 1.4% (110/8096) of patients died within 1 year of surgery. Comparing 1-year mortality CAN scores in survivors vs nonsurvivors, we found statistically significant differences in means (47 vs 66 respectively, P < .001) and medians (45 vs 75 respectively, P < .001) (Table 2). In the plot examining the relationship between preoperative 1-year mortality CAN scores and 1-year mortality, the percentage who died within 1 year increased initially for patients with CAN scores > 60 and again exponentially for patients with CAN scores > 80. Examining Kaplan-Meier curves, we found that survivors and nonsurvivors separated early after surgery, and the differences between the top tercile and the middle/lower terciles were statistically significant (P < .001). Mortality rates were about 0.5% in the lower and middle terciles but about 2% in the upper tercile (Figure 5).
In the plot examining the relationship between CAN scores and index LOS, the LOS rose significantly beyond a CAN score of 60 and dramatically beyond a CAN score of 80 (Figure 6). LOESS curves also showed 2 inflection points suggesting an incremental and sequential rise in the LOS with increasing CAN scores (Figure 7). Mean (SD) LOS in days for the lowest to highest terciles was 2.6 (1.7), 2.8 (2.1), and 3.6 (2.2), respectively.
Discussion
CAN scores are automatically generated each week by EHR-based multivariable risk models. These scores have excellent predictive accuracy for 90-day and 1-year mortality and hospitalization and are routinely used by VHA primary care teams to assist with clinical operations.13 We studied the distribution of CAN 1-year mortality scores in a preoperative context and examined relationships of the preoperative CAN 1-year mortality scores with postoperative mortality and LOS in 8206 veterans who underwent TKR.
There are several noteworthy findings. First, the overall 1-year mortality rate observed following TKR (1.4%) was similar to other published reports.18,19 Not surprisingly, preoperative CAN 1-year mortality scores were significantly higher in veterans who died compared with those of survivors. The majority of patients who died had a preoperative CAN 1-year mortality score > 75 while most who survived had a preoperative CAN 1-year mortality score < 45 (P < .001). Interestingly, the same scores showed a nonlinear correlation with LOS. Index LOS was about 4 days in patients in the highest tercile of CAN scores vs 2.5 days in the lowest tercile, but the initial increase in LOS was detected at a CAN score of about 55 to 60.
In addition, mortality rate varied widely in different segments of the population when grouped according to preoperative CAN scores. One-year mortality rates in the highest tercile reached 2%, about 4-fold higher than that of lower terciles (0.5%). Examination of the Kaplan-Meier curves showed that this difference in mortality between the highest tercile and the lower 2 groups appears soon after discharge and continues to increase over time, suggesting that the factors contributing to the increased mortality are present at the time of discharge and persist beyond the postoperative period. In summary, although CAN scores were not designed for use in the perioperative context, we found that preoperative CAN 1-year mortality scores are broadly predictive of mortality, but especially for increases in LOS following elective TKA, both increases in hospital LOS following elective TKA and mortality over the year after TKA.
Our findings raise several important questions. The decision to undergo elective surgery is complex. Arguably, individuals who undergo elective knee replacement should be healthy enough to undergo, recover, and reap the benefits from a procedure that does not extend life. The distribution of preoperative CAN 1-year mortality scores for our study population was similar to that of the general VHA enrollee population with similar measured mortality rates (≤ 0.5% vs ≥ 1.7% in the low and high terciles, respectively).1 Further study comparing outcomes in matched cohorts who did and did not undergo joint replacement would be of interest. In lieu of this, though, the association of high but not extreme CAN scores with increased hospital LOS may potentially be used to guide allocation of resources to this group, obviating the increased cost and risk to which this group is exposed. And the additional insight afforded by CAN scores may enhance shared decision-making models by identifying patients at the very highest risk (eg, 1-year mortality CAN score ≥ 90), patients who conceivably might not survive long enough to recover from and enjoy their reconstructed knee, who might in the long run be harmed by undergoing the procedure.
Many total joint arthroplasties are performed in older patients, a population in which frailty is increasingly recognized as a significant risk factor for poor outcomes.20,21 CAN scores reliably identify high-risk patients and have been shown to correlate with frailty in this group.22 Multiple authors have reported improved outcomes with cost reductions after implementation of programs targeting modifiable risk factors in high-risk surgical candidates.23-25 A preoperative assessment that includes the CAN score may be valuable in identifying patients who would benefit most from prehabilitation programs or other interventions designed to blunt the impact of frailty. It is true that many elements used to calculate the CAN score would not be considered modifiable, especially in the short term. However, specific contributors to frailty, such as nutritional status and polypharmacy might be potential candidates. As with all multivariable risk prediction models, there are multiple paths to a high CAN score, and further research to identify clinically relevant subgroups may help inform efforts to improve perioperative care within this population.
Hospital LOS is of intense interest for many reasons, not least its utility as a surrogate for cost and increased risk for immediate perioperative adverse events, such as multidrug-resistant hospital acquired infections, need for postacute facility-based rehabilitation, and deconditioning that increase risks of falls and fractures in the older population.26-29 In addition, its importance is magnified due to the COVID-19 pandemic context in which restarting elective surgery programs has changed traditional criteria by which patients are scheduled for surgery.
We have shown that elevated CAN scores are able to identify patients at risk for extended hospital stays and, as such, may be useful additional data in allocating scarce operating room time and other resources for optimal patient and health care provider safety.30,31 Individual surgeons and hospital systems would, of course, decide which patients should be triaged to go first, based on local priorities; however, choosing lower risk patients with minimal risk of morbidity and mortality while pursuing prehabilitation for higher risk patients is a reasonable approach.
Limitations
Our study has several limitations. Only a single surgical procedure was included, albeit the most common one performed in the VHA. In addition, no information was available concerning the precise clinical course for these patients, such as the duration of surgery, anesthetic technique, and management of acute, perioperative course. Although the assumption was made that patients received standard care in a manner such that these factors would not significantly affect either their mortality or their LOS out of proportion to their preoperative clinical status, confounding cannot be excluded. Therefore, further study is necessary to determine whether CAN scores can accurately predict mortality and/or LOS for patients undergoing other procedures. Further, a clinical trial is required to assess whether systematic provision of the CAN score at the point of surgery would impact care and, more important, impact outcomes. In addition, multivariable analyses were not performed, including and excluding various components of the CAN score models. Currently, CAN scores could be made available to the surgical/anesthesia communities at minimal or no cost and are updated automatically. Model calibration and discrimination in this particular setting were not validated.
Because our interest is in leveraging an existing resource to a current clinical and operational problem rather than in creating or validating a new tool, we chose to test the simple bivariate relationship between preoperative CAN scores and outcomes. We chose the preoperative 1-year mortality CAN score from among the 4 options under the assumption that long-term survival is the most meaningful of the 4 candidate outcomes. Finally, while the CAN scores are currently only calculated and generated for patients cared for within the VHA, few data elements are unavailable to civilian health systems. The most problematic would be documentation of actual prescription filling, but this is a topic of increasing interest to the medical and academic communities and access to such information we hope will improve.32-34
Conclusions
Although designed for use by VHA primary care teams, CAN scores also may have value for perioperative clinicians, predicting mortality and prolonged hospital LOS in those with elevated 1-year mortality scores. Advantages of CAN scores relative to other perioperative risk calculators lies in their ability to predict long-term rather than 30-day survival and that they are automatically generated on a near-real-time basis for all patients who receive care in VHA ambulatory clinics. Further study is needed to determine practical utility in shared decision making, preoperative evaluation and optimization, and perioperative resource allocation.
Acknowledgments
This work was supported by the US Department of Veterans Affairs (VA) National Center for Patient Safety, Field Office 10A4E, through the Patient Safety Center of Inquiry at the Durham VA Medical Center in North Carolina. The study also received support from the Center of Innovation to Accelerate Discovery and Practice Transformation (CIN 13-410) at the Durham VA Health Care System.
1. McNair AGK, MacKichan F, Donovan JL, et al. What surgeons tell patients and what patients want to know before major cancer surgery: a qualitative study. BMC Cancer. 2016;16:258. doi:10.1186/s12885-016-2292-3
2. Grover FL, Hammermeister KE, Burchfiel C. Initial report of the Veterans Administration Preoperative Risk Assessment Study for Cardiac Surgery. Ann Thorac Surg. 1990;50(1):12-26; discussion 27-18. doi:10.1016/0003-4975(90)90073-f
3. Khuri SF, Daley J, Henderson W, et al. The National Veterans Administration Surgical Risk Study: risk adjustment for the comparative assessment of the quality of surgical care. J Am Coll Surg. 1995;180(5):519-531.
4. Glance LG, Lustik SJ, Hannan EL, et al. The Surgical Mortality Probability Model: derivation and validation of a simple simple risk prediction rule for noncardiac surgery. Ann Surg. 2012;255(4):696-702. doi:10.1097/SLA.0b013e31824b45af
5. Keller DS, Kroll D, Papaconstantinou HT, Ellis CN. Development and validation of a methodology to reduce mortality using the veterans affairs surgical quality improvement program risk calculator. J Am Coll Surg. 2017;224(4):602-607. doi:10.1016/j.jamcollsurg.2016.12.033
6. Bilimoria KY, Liu Y, Paruch JL, et al. Development and evaluation of the universal ACS NSQIP surgical risk calculator: a decision aid and informed consent tool for patients and surgeons. J Am Coll Surg. 2013;217(5):833-842.e831-833. doi:10.1016/j.jamcollsurg.2013.07.385
7. Ford MK, Beattie WS, Wijeysundera DN. Systematic review: prediction of perioperative cardiac complications and mortality by the revised cardiac risk index. Ann Intern Med. 2010;152(1):26-35. doi:10.7326/0003-4819-152-1-201001050-00007
8. Gupta PK, Gupta H, Sundaram A, et al. Development and validation of a risk calculator for prediction of cardiac risk after surgery. Circulation. 2011;124(4):381-387. doi:10.1161/CIRCULATIONAHA.110.015701
9. Lee TH, Marcantonio ER, Mangione CM, et al. Derivation and prospective validation of a simple index for prediction of cardiac risk of major noncardiac surgery. Circulation. 1999;100(10):1043-1049. doi:10.1161/01.cir.100.10.1043
10. Smith T, Li X, Nylander W, Gunnar W. Thirty-day postoperative mortality risk estimates and 1-year survival in Veterans Health Administration surgery patients. JAMA Surg. 2016;151(5):417-422. doi:10.1001/jamasurg.2015.4882
11. Damhuis RA, Wijnhoven BP, Plaisier PW, Kirkels WJ, Kranse R, van Lanschot JJ. Comparison of 30-day, 90- day and in-hospital postoperative mortality for eight different cancer types. Br J Surg. 2012;99(8):1149-1154. doi:10.1002/bjs.8813
12. Wang L, Porter B, Maynard C, et al. Predicting risk of hospitalization or death among patients receiving primary care in the Veterans Health Administration. Med Care. 2013;51(4):368-373. doi:10.1016/j.amjcard.2012.06.038
13. Fihn SD, Francis J, Clancy C, et al. Insights from advanced analytics at the Veterans Health Administration. Health Aff (Millwood). 2014;33(7):1203-1211. doi:10.1377/hlthaff.2014.0054
14. Noël PH, Copeland LA, Perrin RA, et al. VHA Corporate Data Warehouse height and weight data: opportunities and challenges for health services research. J Rehabil Res Dev. 2010;47(8):739-750. doi:10.1682/jrrd.2009.08.0110
15. Wong ES, Yoon J, Piegari RI, Rosland AM, Fihn SD, Chang ET. Identifying latent subgroups of high-risk patients using risk score trajectories. J Gen Intern Med. 2018;33(12):2120-2126. doi:10.1007/s11606-018-4653-x
16. Chen Q, Hsia HL, Overman R, et al. Impact of an opioid safety initiative on patients undergoing total knee arthroplasty: a time series analysis. Anesthesiology. 2019;131(2):369-380. doi:10.1097/ALN.0000000000002771
17. Hsia HL, Takemoto S, van de Ven T, et al. Acute pain is associated with chronic opioid use after total knee arthroplasty. Reg Anesth Pain Med. 2018;43(7):705-711. doi:10.1097/AAP.0000000000000831
18. Inacio MCS, Dillon MT, Miric A, Navarro RA, Paxton EW. Mortality after total knee and total hip arthroplasty in a large integrated health care system. Perm J. 2017;21:16-171. doi:10.7812/TPP/16-171
19. Lee QJ, Mak WP, Wong YC. Mortality following primary total knee replacement in public hospitals in Hong Kong. Hong Kong Med J. 2016;22(3):237-241. doi:10.12809/hkmj154712
20. Lin HS, Watts JN, Peel NM, Hubbard RE. Frailty and post-operative outcomes in older surgical patients: a systematic review. BMC Geriatr. 2016;16(1):157. doi:10.1186/s12877-016-0329-8
21. Shinall MC Jr, Arya S, Youk A, et al. Association of preoperative patient frailty and operative stress with postoperative mortality. JAMA Surg. 2019;155(1):e194620. doi:10.1001/jamasurg.2019.4620
22. Ruiz JG, Priyadarshni S, Rahaman Z, et al. Validation of an automatically generated screening score for frailty: the care assessment need (CAN) score. BMC Geriatr. 2018;18(1):106. doi:10.1186/s12877-018-0802-7
23. Bernstein DN, Liu TC, Winegar AL, et al. Evaluation of a preoperative optimization protocol for primary hip and knee arthroplasty patients. J Arthroplasty. 2018;33(12):3642- 3648. doi:10.1016/j.arth.2018.08.018
24. Sodhi N, Anis HK, Coste M, et al. A nationwide analysis of preoperative planning on operative times and postoperative complications in total knee arthroplasty. J Knee Surg. 2019;32(11):1040-1045. doi:10.1055/s-0039-1677790
25. Krause A, Sayeed Z, El-Othmani M, Pallekonda V, Mihalko W, Saleh KJ. Outpatient total knee arthroplasty: are we there yet? (part 1). Orthop Clin North Am. 2018;49(1):1-6. doi:10.1016/j.ocl.2017.08.002
26. Barrasa-Villar JI, Aibar-Remón C, Prieto-Andrés P, Mareca- Doñate R, Moliner-Lahoz J. Impact on morbidity, mortality, and length of stay of hospital-acquired infections by resistant microorganisms. Clin Infect Dis. 2017;65(4):644-652. doi:10.1093/cid/cix411
27. Nikkel LE, Kates SL, Schreck M, Maceroli M, Mahmood B, Elfar JC. Length of hospital stay after hip fracture and risk of early mortality after discharge in New York state: retrospective cohort study. BMJ. 2015;351:h6246. doi:10.1136/bmj.h6246
28. Marfil-Garza BA, Belaunzarán-Zamudio PF, Gulias-Herrero A, et al. Risk factors associated with prolonged hospital length-of-stay: 18-year retrospective study of hospitalizations in a tertiary healthcare center in Mexico. PLoS One. 2018;13(11):e0207203. doi:10.1371/journal.pone.0207203
29. Hirsch CH, Sommers L, Olsen A, Mullen L, Winograd CH. The natural history of functional morbidity in hospitalized older patients. J Am Geriatr Soc. 1990;38(12):1296-1303. doi:10.1111/j.1532-5415.1990.tb03451.x
30. Iyengar KP, Jain VK, Vaish A, Vaishya R, Maini L, Lal H. Post COVID-19: planning strategies to resume orthopaedic surgery -challenges and considerations. J Clin Orthop Trauma. 2020;11(suppl 3):S291-S295. doi:10.1016/j.jcot.2020.04.028
31. O’Connor CM, Anoushiravani AA, DiCaprio MR, Healy WL, Iorio R. Economic recovery after the COVID-19 pandemic: resuming elective orthopedic surgery and total joint arthroplasty. J Arthroplasty. 2020;35(suppl 7):S32-S36. doi:10.1016/j.arth.2020.04.038.
32. Mauseth SA, Skurtveit S, Skovlund E, Langhammer A, Spigset O. Medication use and association with urinary incontinence in women: data from the Norwegian Prescription Database and the HUNT study. Neurourol Urodyn. 2018;37(4):1448-1457. doi:10.1002/nau.23473
33. Sultan RS, Correll CU, Schoenbaum M, King M, Walkup JT, Olfson M. National patterns of commonly prescribed psychotropic medications to young people. J Child Adolesc Psychopharmacol. 2018;28(3):158-165. doi:10.1089/cap.2017.0077
34. McCoy RG, Dykhoff HJ, Sangaralingham L, et al. Adoption of new glucose-lowering medications in the U.S.-the case of SGLT2 inhibitors: nationwide cohort study. Diabetes Technol Ther. 2019;21(12):702-712. doi:10.1089/dia.2019.0213
1. McNair AGK, MacKichan F, Donovan JL, et al. What surgeons tell patients and what patients want to know before major cancer surgery: a qualitative study. BMC Cancer. 2016;16:258. doi:10.1186/s12885-016-2292-3
2. Grover FL, Hammermeister KE, Burchfiel C. Initial report of the Veterans Administration Preoperative Risk Assessment Study for Cardiac Surgery. Ann Thorac Surg. 1990;50(1):12-26; discussion 27-18. doi:10.1016/0003-4975(90)90073-f
3. Khuri SF, Daley J, Henderson W, et al. The National Veterans Administration Surgical Risk Study: risk adjustment for the comparative assessment of the quality of surgical care. J Am Coll Surg. 1995;180(5):519-531.
4. Glance LG, Lustik SJ, Hannan EL, et al. The Surgical Mortality Probability Model: derivation and validation of a simple simple risk prediction rule for noncardiac surgery. Ann Surg. 2012;255(4):696-702. doi:10.1097/SLA.0b013e31824b45af
5. Keller DS, Kroll D, Papaconstantinou HT, Ellis CN. Development and validation of a methodology to reduce mortality using the veterans affairs surgical quality improvement program risk calculator. J Am Coll Surg. 2017;224(4):602-607. doi:10.1016/j.jamcollsurg.2016.12.033
6. Bilimoria KY, Liu Y, Paruch JL, et al. Development and evaluation of the universal ACS NSQIP surgical risk calculator: a decision aid and informed consent tool for patients and surgeons. J Am Coll Surg. 2013;217(5):833-842.e831-833. doi:10.1016/j.jamcollsurg.2013.07.385
7. Ford MK, Beattie WS, Wijeysundera DN. Systematic review: prediction of perioperative cardiac complications and mortality by the revised cardiac risk index. Ann Intern Med. 2010;152(1):26-35. doi:10.7326/0003-4819-152-1-201001050-00007
8. Gupta PK, Gupta H, Sundaram A, et al. Development and validation of a risk calculator for prediction of cardiac risk after surgery. Circulation. 2011;124(4):381-387. doi:10.1161/CIRCULATIONAHA.110.015701
9. Lee TH, Marcantonio ER, Mangione CM, et al. Derivation and prospective validation of a simple index for prediction of cardiac risk of major noncardiac surgery. Circulation. 1999;100(10):1043-1049. doi:10.1161/01.cir.100.10.1043
10. Smith T, Li X, Nylander W, Gunnar W. Thirty-day postoperative mortality risk estimates and 1-year survival in Veterans Health Administration surgery patients. JAMA Surg. 2016;151(5):417-422. doi:10.1001/jamasurg.2015.4882
11. Damhuis RA, Wijnhoven BP, Plaisier PW, Kirkels WJ, Kranse R, van Lanschot JJ. Comparison of 30-day, 90- day and in-hospital postoperative mortality for eight different cancer types. Br J Surg. 2012;99(8):1149-1154. doi:10.1002/bjs.8813
12. Wang L, Porter B, Maynard C, et al. Predicting risk of hospitalization or death among patients receiving primary care in the Veterans Health Administration. Med Care. 2013;51(4):368-373. doi:10.1016/j.amjcard.2012.06.038
13. Fihn SD, Francis J, Clancy C, et al. Insights from advanced analytics at the Veterans Health Administration. Health Aff (Millwood). 2014;33(7):1203-1211. doi:10.1377/hlthaff.2014.0054
14. Noël PH, Copeland LA, Perrin RA, et al. VHA Corporate Data Warehouse height and weight data: opportunities and challenges for health services research. J Rehabil Res Dev. 2010;47(8):739-750. doi:10.1682/jrrd.2009.08.0110
15. Wong ES, Yoon J, Piegari RI, Rosland AM, Fihn SD, Chang ET. Identifying latent subgroups of high-risk patients using risk score trajectories. J Gen Intern Med. 2018;33(12):2120-2126. doi:10.1007/s11606-018-4653-x
16. Chen Q, Hsia HL, Overman R, et al. Impact of an opioid safety initiative on patients undergoing total knee arthroplasty: a time series analysis. Anesthesiology. 2019;131(2):369-380. doi:10.1097/ALN.0000000000002771
17. Hsia HL, Takemoto S, van de Ven T, et al. Acute pain is associated with chronic opioid use after total knee arthroplasty. Reg Anesth Pain Med. 2018;43(7):705-711. doi:10.1097/AAP.0000000000000831
18. Inacio MCS, Dillon MT, Miric A, Navarro RA, Paxton EW. Mortality after total knee and total hip arthroplasty in a large integrated health care system. Perm J. 2017;21:16-171. doi:10.7812/TPP/16-171
19. Lee QJ, Mak WP, Wong YC. Mortality following primary total knee replacement in public hospitals in Hong Kong. Hong Kong Med J. 2016;22(3):237-241. doi:10.12809/hkmj154712
20. Lin HS, Watts JN, Peel NM, Hubbard RE. Frailty and post-operative outcomes in older surgical patients: a systematic review. BMC Geriatr. 2016;16(1):157. doi:10.1186/s12877-016-0329-8
21. Shinall MC Jr, Arya S, Youk A, et al. Association of preoperative patient frailty and operative stress with postoperative mortality. JAMA Surg. 2019;155(1):e194620. doi:10.1001/jamasurg.2019.4620
22. Ruiz JG, Priyadarshni S, Rahaman Z, et al. Validation of an automatically generated screening score for frailty: the care assessment need (CAN) score. BMC Geriatr. 2018;18(1):106. doi:10.1186/s12877-018-0802-7
23. Bernstein DN, Liu TC, Winegar AL, et al. Evaluation of a preoperative optimization protocol for primary hip and knee arthroplasty patients. J Arthroplasty. 2018;33(12):3642- 3648. doi:10.1016/j.arth.2018.08.018
24. Sodhi N, Anis HK, Coste M, et al. A nationwide analysis of preoperative planning on operative times and postoperative complications in total knee arthroplasty. J Knee Surg. 2019;32(11):1040-1045. doi:10.1055/s-0039-1677790
25. Krause A, Sayeed Z, El-Othmani M, Pallekonda V, Mihalko W, Saleh KJ. Outpatient total knee arthroplasty: are we there yet? (part 1). Orthop Clin North Am. 2018;49(1):1-6. doi:10.1016/j.ocl.2017.08.002
26. Barrasa-Villar JI, Aibar-Remón C, Prieto-Andrés P, Mareca- Doñate R, Moliner-Lahoz J. Impact on morbidity, mortality, and length of stay of hospital-acquired infections by resistant microorganisms. Clin Infect Dis. 2017;65(4):644-652. doi:10.1093/cid/cix411
27. Nikkel LE, Kates SL, Schreck M, Maceroli M, Mahmood B, Elfar JC. Length of hospital stay after hip fracture and risk of early mortality after discharge in New York state: retrospective cohort study. BMJ. 2015;351:h6246. doi:10.1136/bmj.h6246
28. Marfil-Garza BA, Belaunzarán-Zamudio PF, Gulias-Herrero A, et al. Risk factors associated with prolonged hospital length-of-stay: 18-year retrospective study of hospitalizations in a tertiary healthcare center in Mexico. PLoS One. 2018;13(11):e0207203. doi:10.1371/journal.pone.0207203
29. Hirsch CH, Sommers L, Olsen A, Mullen L, Winograd CH. The natural history of functional morbidity in hospitalized older patients. J Am Geriatr Soc. 1990;38(12):1296-1303. doi:10.1111/j.1532-5415.1990.tb03451.x
30. Iyengar KP, Jain VK, Vaish A, Vaishya R, Maini L, Lal H. Post COVID-19: planning strategies to resume orthopaedic surgery -challenges and considerations. J Clin Orthop Trauma. 2020;11(suppl 3):S291-S295. doi:10.1016/j.jcot.2020.04.028
31. O’Connor CM, Anoushiravani AA, DiCaprio MR, Healy WL, Iorio R. Economic recovery after the COVID-19 pandemic: resuming elective orthopedic surgery and total joint arthroplasty. J Arthroplasty. 2020;35(suppl 7):S32-S36. doi:10.1016/j.arth.2020.04.038.
32. Mauseth SA, Skurtveit S, Skovlund E, Langhammer A, Spigset O. Medication use and association with urinary incontinence in women: data from the Norwegian Prescription Database and the HUNT study. Neurourol Urodyn. 2018;37(4):1448-1457. doi:10.1002/nau.23473
33. Sultan RS, Correll CU, Schoenbaum M, King M, Walkup JT, Olfson M. National patterns of commonly prescribed psychotropic medications to young people. J Child Adolesc Psychopharmacol. 2018;28(3):158-165. doi:10.1089/cap.2017.0077
34. McCoy RG, Dykhoff HJ, Sangaralingham L, et al. Adoption of new glucose-lowering medications in the U.S.-the case of SGLT2 inhibitors: nationwide cohort study. Diabetes Technol Ther. 2019;21(12):702-712. doi:10.1089/dia.2019.0213
Twenty Years of Breast Reduction Surgery at a Veterans Affairs Medical Center
Women make up an estimated 10% of the veteran population.1 The US Department of Veterans Affairs (VA) projected that there would be an increase of 18,000 female veterans per year for 10 years based on 2015 data. The number of women veterans enrolled in the VA health care increased from 397,024 to 729,989 (83.9%) between 2005 and 2015.2 This rise in the number of enrolled women veterans also increased the demand for female-specific health care services, such as breast reduction surgery, a reconstructive procedure provided at the Malcom Randall VA Medical Center (MRVAMC) federal teaching hospital in Gainesville, Florida.
Patients who experience symptomatic macromastia will report a history of neck and shoulder pain, shoulder grooving from bra straps, inframammary intertrigo, difficulty finding clothes that fit, and discomfort participating in sports. For the treatment of symptomatic macromastia, patients report a high satisfaction rate after breast reduction surgery.3-5 Unfortunately, the complications from the surgery can significantly disrupt a woman’s life due to previously unplanned hospital admissions, clinic appointments, wound care, time off work, and poor aesthetic outcome. Faculty awareness of a large number of complications for patients after breast reduction surgery prompted the MRVAMC Plastic Surgery Service to establish a stricter surgical screening protocol using body mass index (BMI) values and negative nicotine status to help patients be healthier and reduce the potential risk before offering surgery. A medical literature search did not find an existing study on veteran-specific breast reduction surgery.
Methods
The University of Florida and North Florida/South Georgia Veterans Health System Institutional Review Board approved a retrospective chart review of all breast reduction surgeries performed at MRVAMC over a 20-year period (July 1, 2000-June 30, 2020). Electronic health records were queried for all primary bilateral breast reduction surgeries performed for symptomatic macromastia using Current Procedural Terminology code 19318. Potentially modifiable or predictable risk factors for wound complications were recorded: nicotine status, BMI, diabetes mellitus (DM) status, skin incision pattern, and pedicle location. Skin incision patterns were either vertical (periareolar plus a vertical scar from the areola to the inframammary fold) or traditional Wise pattern (also known as anchor pattern: periareolar scar, vertical scar to inframammary fold, plus a horizontal scar along the inframammary fold) as seen in Figures 1 and 2. The pedicle is the source of blood supply to the nipple, which was documented as either from the inferior aspect or the superior or superior/medial aspect.
For this study, the blood supply from the superior and superior/medial was logged in the same category. Records were reviewed 3 months after surgery for documentation of local wound complications, such as hematoma, infection, wound breakdown, skin necrosis, and nipple necrosis. Major complications were defined as requiring an unplanned hospital admission or urgent return to the operating room. A χ2 test using a P value of < .05 was used to determine statistical significance between the incidence of wound complications and the individually identifiable variables.
Results
One hundred fifteen bilateral breast reduction surgeries were performed at MRVAMC over a 20-year period. Patient median age was 43 years. Median combined specimen weight was 1272 g. Forty-eight (41.7%) wound complications were documented, including 8 (7%) major complications. Most complications were identified in the first 7 years of the study before the new protocol and consult template became active. The new template resulted in the local complication rate dropping from 62% (July 2000-June 2007) to 26% (July 2007-June 2020). BMI > 32 (P = .03) and active nicotine use (P = .004) were found to be statistically significant independent risk factors for wound complications. Median BMI for all patients was 30. DM status (P = .22), skin incision pattern (P = .25), and pedicle location (P = .13) were not found to be predictors of wound complications (Table). There was no significant change in the incidence of major complications before and after the new protocols were enforced.
Discussion
Breast reduction surgery is an elective reconstructive option to treat symptomatic macromastia. There are several accepted ways to do the reduction surgical procedure where the blood supply (pedicle) to the nipple can vary and the visible scars can be in a horizontal, vertical, or Wise pattern. Technique is usually based on surgeon training, comfort, and preference. There are several known complications specific to this operation that include asymmetry, changes in nipple sensation, unattractive scars, diminished ability to breastfeed, and wound complications.5-7 Wound complications include seroma, hematoma, dehiscence, infection, wound breakdown, skin necrosis, and nipple necrosis.
This study focused on wound complications with the objective of identifying and modifying risk factors. Two known risk factors documented in the literature, nicotine use and obesity, already had been addressed by our service, and results were known anecdotally but had not been previously verified. This study also looked at other potential risk factors, including the pedicle location, skin incision, and DM status.
Residents or fellows participated in all the surgeries. An outcome analysis from The American College of Surgeons National Surgical Quality Improvement Program database from 2005 to 2011 found that resident participation was associated with morbidity, including wound complications.8 This study was performed at a federal hospital with a complexity level 1a rating, which is designated based on the highest level of patient volume, risk, teaching, research, intensive care unit beds, and specialty services.9 The hospital is closely affiliated with a level 1 trauma center and teaching hospital; therefore, resident and fellow participation is not a modifiable risk factor.
This study did not find an increased risk of wound complications in patients with DM, which has been found to be an independent risk factor in a prior study.10 DM status was indicated in only 3 histories, and they all had perioperative hemoglobin A1c levels < 8%. There is documentation of patients receiving perioperative antibiotics in 99 out of 116 of the surgical records; however, we did not include this in the analysis because the operative reports from the first year of the study were incomplete.
Smoking is a known risk factor for local wound complications in breast reduction surgery.10-15 The VA has a smoking cessation program through its mental health service that provides counseling and medication treatment options, including nicotine replacement, bupropion, and varenicline. We require patients to be at least 4 weeks nicotine free before surgery, which has been previously recommended in the literature.16
Existing studies that compare the traditional Wise pattern/inferior pedicle with vertical pattern/superior medial pedicle did not find an increased risk of wound complications.17-19 Our study separated the different incisions from the pedicle because the surgical technique among the different surgeons in the study varied, where sometimes the traditional Wise pattern was combined with the less traditional superior-medial pedicle. We did not find a statistical difference when comparing the incisions and pedicle location, which suggests that the incision type and source of blood supply to the nipple are not the determining factors for wound complications in the early postoperative period.
Obesity is a known risk factor for local wound complications.12,13,15,20-22 Studies have shown that patients who are obese benefit from breast reduction surgery; authors have argued against restricting surgery to these higher risk patients.4,23-25 Patients usually report decades of macromastia symptoms at consultation; so, we believe delaying the surgical procedure to get patients to a safer risk profile is in their best interest. We chose a cutoff BMI of 32 as a realistic value rather than 30, which is considered the definition of obesity. Patients at MRVAMC have access to MOVE!, a weight loss management program through primary care. We believe in being reasonable; so if a patient makes a significant improvement in her health but falls short of the required cutoff, we will still consider offering the surgical procedure after a full explanation of the surgical risks.
Wound complications, especially those that require admission or frequent appointments can seriously disrupt a patient’s life, creating unnecessary hardships and expense in time lost from work, travel, and child care. MRVAMC has a catchment area the size of North Carolina; so many of our patients travel hours for their appointments. The added scars and deformity from wound dehiscence and debridement can lead to asymmetry, widened scars, and future revision operations. Multiple clinic appointments for wound care not only impact that individual patient, but also has the effect of limiting access for all patients in a health care environment with high patient volume and limited providers, operating room time, and clinic appointments. As a result, minimizing predictable wound complications benefits the entire system.
Limitations and Strengths
This retrospective review comprised multiple different surgeons, including faculty and trainees, who were involved in the consultation, surgery, and postoperative care of the patients over a 20-year period; therefore, consistency in documentation is lacking. In addition, we were limited to only the information available on the charts. For example, wound size and laterality were not consistently documented. The MRVAMC complication rate was consistent with the current literature (range, 14-52%).12,18,20,24
The major strength of the study is that the veterans tend to stay within the VA, which makes complications easier to identify and follow. Patients who do not present initially to their surgeon due to travel limitations will typically contact their primary care provider or present to their local VA urgent care or emergency department provider, who will route the patient back to the surgical specialty service through the electronic health record.
Conclusions
Breast reduction surgery has a high wound complication rate, which can be predicted and improved on so that patients can receive their indicated surgical procedure with minimal inconvenience and downtime. This review confirms that preoperative weight loss and nicotine cessation were the appropriate focus of the MRVAMC plastic surgery service’s efforts to achieve a safer surgical experience. We will continue to enforce our protocol and encourage patients who are interested in breast reduction surgery and fall outside the requirements to work with their primary care provider on smoking cessation and weight loss through better nutrition and physical activity.
Acknowledgment
This manuscript is the result of work supported with resources and the use of facilities at the North Florida/South Georgia Veterans Health System in Gainesville, Florida.
1. US Department of Veterans Affairs. Statistics at a glance. Published February 2020. Accessed June 18, 2021. https://www.va.gov/vetdata/docs/Quickfacts/Homepage_slideshow_4_6_20.PDF
2. US Department of Veterans Affairs, National Center for Veterans Analysis and Statistics. Women veterans report: the past, present, and future of women veterans. Published February 2017. Accessed June 18, 2020. https://www.va.gov/vetdata/docs/specialreports/women_veterans_2015_final.pdf
3. Crittenden TA, Watson DI, Ratcliffe J, Griffin PA, Dean NR. Outcomes of breast reduction surgery using the breast-q: a prospective study and comparison with normative data. Plast Reconstr Surg. 2019;144(5):1034-1044. doi:10.1097/PRS.0000000000006114
4. Thoma A, Sprague S, Veltri K, Duku E, Furlong W. A prospective study of patients undergoing breast reduction surgery: health-related quality of life and clinical outcomes. Plast Reconstr Surg. 2007;120(1):13-26. doi:10.1097/01.prs.0000263370.94191.90
5. Nuzzi LC, Firriolo JM, Pike CM, DiVasta AD, Labow BI. Complications and quality of life following reduction mammaplasty in adolescents and young women.Plast Reconstr Surg. 2019;144(3):572-581. doi:10.1097/PRS.0000000000005907
6. Hall-Findlay EJ, Shestak KC. Breast reduction. Plast Reconstr Surg. 2015;136(4):531e-544e. doi:10.1097/PRS.0000000000001622
7. Kraut RY, Brown E, Korownyk C, et al. The impact of breast reduction surgery on breastfeeding: systematic review of observational studies. PLoS One. 2017;12(10):e0186591. doi:10.1371/journal.pone.0186591
8. Fischer JP, Wes AM, Kovach SJ. The impact of surgical resident participation in breast reduction surgery--outcome analysis from the 2005-2011 ACS-NSQIP datasets. J Plast Surg Hand Surg. 2014;48(5):315-321. doi:10.3109/2000656X.2014.882345
9. Site Facility Name and Complexity Summary of VHA Facility. Accessed June 18, 2021. https://www.vendorportal.ecms.va.gov/FBODocumentServer/DocumentServer.aspx?DocumentId=2793591&FileName=VA118-16-R-1059-A00002002.docx
10. Lewin R, Göransson M, Elander A, Thorarinsson A, Lundberg J, Lidén M. Risk factors for complications after breast reduction surgery. J Plast Surg Hand Surg. 2014;48(1):10-14. doi:10.3109/2000656X.2013.791625
11. Cunningham BL, Gear AJ, Kerrigan CL, Collins ED. Analysis of breast reduction complications derived from the BRAVO study. Plast Reconstr Surg. 2005;115(6):1597-1604. doi:10.1097/01.prs.0000160695.33457.db
12. Karamanos E, Wei B, Siddiqui A, Rubinfeld I. Tobacco use and body mass index as predictors of outcomes in patients undergoing breast reduction mammoplasty. Ann Plast Surg. 2015;75(4):383-387. doi:10.1097/SAP.0000000000000192
13. Manahan MA, Buretta KJ, Chang D, Mithani SK, Mallalieu J, Shermak MA. An outcomes analysis of 2142 breast reduction procedures. Ann Plast Surg. 2015;74(3):289-292. doi:10.1097/SAP.0b013e31829d2261
14. Hillam JS, Borsting EA, Chim JH, Thaller SR. Smoking as a risk factor for breast reduction: an analysis of 13,503 cases. J Plast Reconstr Aesthet Surg. 2017;70(6):734-740. doi:10.1016/j.bjps.2016.12.012
15. Zhang MX, Chen CY, Fang QQ, et al. Risk factors for complications after reduction mammoplasty: a meta-analysis. PLoS One. 2016;11(12):e0167746. doi:10.1371/journal.pone.0167746
16. Sørensen LT. Wound healing and infection in surgery: the pathophysiological impact of smoking, smoking cessation, and nicotine replacement therapy: a systematic review. Ann Surg. 2012;255(6):1069-1079.doi:10.1097/SLA.0b013e31824f632d
17. Antony AK, Yegiyants SS, Danielson KK, et al. A matched cohort study of superomedial pedicle vertical scar breast reduction (100 breasts) and traditional inferior pedicle Wise-pattern reduction (100 breasts): an outcomes study over 3 years. Plast Reconstr Surg. 2013;132(5):1068-1076. doi:10.1097/PRS.0b013e3182a48b2d
18. Hunter-Smith DJ, Smoll NR, Marne B, Maung H, Findlay MW. Comparing breast-reduction techniques: time-to-event analysis and recommendations. Aesthetic Plast Surg. 2012;36(3):600-606. doi:10.1007/s00266-011-9860-3
19. Ogunleye AA, Leroux O, Morrison N, Preminger AB. Complications after reduction mammaplasty: a comparison of wise pattern/inferior pedicle and vertical scar/superomedial pedicle. Ann Plast Surg. 2017;79(1):13-16. doi:10.1097/SAP.0000000000001059
20. Bauermeister AJ, Gill K, Zuriarrain A, Earle SA, Newman MI. Reduction mammaplasty with superomedial pedicle technique: a literature review and retrospective analysis of 938 consecutive breast reductions. J Plast Reconstr Aesthet Surg. 2019;72(3):410-418. doi:10.1016/j.bjps.2018.12.004
21. Nelson JA, Fischer JP, Chung CU, et al. Obesity and early complications following reduction mammaplasty: an analysis of 4545 patients from the 2005-2011 NSQIP datasets. J Plast Surg Hand Surg. 2014;48(5):334-339. doi:10.3109/2000656X.2014.886582
22. Kreithen J, Caffee H, Rosenberg J, et al. A comparison of the LeJour and Wise pattern methods of breast reduction. Ann Plast Surg. 2005;54(3):236-241. doi:10.3109/2000656X.2014.886582
23. Güemes A, Pérez E, Sousa R, et al. Quality of life and alleviation of symptoms after breast reduction for macromastia in obese patients: is surgery worth it? Aesthetic Plast Surg. 2016;40(1):62-70. doi:10.1007/s00266-015-0601-x
24. Setälä L, Papp A, Joukainen S, et al. Obesity and complications in breast reduction surgery: are restrictions justified? J Plast Reconstr Aesthet Surg. 2009;62(2):195-199. doi:10.1016/j.bjps.2007.10.043
25. Shah R, Al-Ajam Y, Stott D, Kang N. Obesity in mammaplasty: a study of complications following breast reduction. J Plast Reconstr Aesthet Surg. 2011;64(4):508-514. doi:10.1016/j.bjps.2007.10.043
Women make up an estimated 10% of the veteran population.1 The US Department of Veterans Affairs (VA) projected that there would be an increase of 18,000 female veterans per year for 10 years based on 2015 data. The number of women veterans enrolled in the VA health care increased from 397,024 to 729,989 (83.9%) between 2005 and 2015.2 This rise in the number of enrolled women veterans also increased the demand for female-specific health care services, such as breast reduction surgery, a reconstructive procedure provided at the Malcom Randall VA Medical Center (MRVAMC) federal teaching hospital in Gainesville, Florida.
Patients who experience symptomatic macromastia will report a history of neck and shoulder pain, shoulder grooving from bra straps, inframammary intertrigo, difficulty finding clothes that fit, and discomfort participating in sports. For the treatment of symptomatic macromastia, patients report a high satisfaction rate after breast reduction surgery.3-5 Unfortunately, the complications from the surgery can significantly disrupt a woman’s life due to previously unplanned hospital admissions, clinic appointments, wound care, time off work, and poor aesthetic outcome. Faculty awareness of a large number of complications for patients after breast reduction surgery prompted the MRVAMC Plastic Surgery Service to establish a stricter surgical screening protocol using body mass index (BMI) values and negative nicotine status to help patients be healthier and reduce the potential risk before offering surgery. A medical literature search did not find an existing study on veteran-specific breast reduction surgery.
Methods
The University of Florida and North Florida/South Georgia Veterans Health System Institutional Review Board approved a retrospective chart review of all breast reduction surgeries performed at MRVAMC over a 20-year period (July 1, 2000-June 30, 2020). Electronic health records were queried for all primary bilateral breast reduction surgeries performed for symptomatic macromastia using Current Procedural Terminology code 19318. Potentially modifiable or predictable risk factors for wound complications were recorded: nicotine status, BMI, diabetes mellitus (DM) status, skin incision pattern, and pedicle location. Skin incision patterns were either vertical (periareolar plus a vertical scar from the areola to the inframammary fold) or traditional Wise pattern (also known as anchor pattern: periareolar scar, vertical scar to inframammary fold, plus a horizontal scar along the inframammary fold) as seen in Figures 1 and 2. The pedicle is the source of blood supply to the nipple, which was documented as either from the inferior aspect or the superior or superior/medial aspect.
For this study, the blood supply from the superior and superior/medial was logged in the same category. Records were reviewed 3 months after surgery for documentation of local wound complications, such as hematoma, infection, wound breakdown, skin necrosis, and nipple necrosis. Major complications were defined as requiring an unplanned hospital admission or urgent return to the operating room. A χ2 test using a P value of < .05 was used to determine statistical significance between the incidence of wound complications and the individually identifiable variables.
Results
One hundred fifteen bilateral breast reduction surgeries were performed at MRVAMC over a 20-year period. Patient median age was 43 years. Median combined specimen weight was 1272 g. Forty-eight (41.7%) wound complications were documented, including 8 (7%) major complications. Most complications were identified in the first 7 years of the study before the new protocol and consult template became active. The new template resulted in the local complication rate dropping from 62% (July 2000-June 2007) to 26% (July 2007-June 2020). BMI > 32 (P = .03) and active nicotine use (P = .004) were found to be statistically significant independent risk factors for wound complications. Median BMI for all patients was 30. DM status (P = .22), skin incision pattern (P = .25), and pedicle location (P = .13) were not found to be predictors of wound complications (Table). There was no significant change in the incidence of major complications before and after the new protocols were enforced.
Discussion
Breast reduction surgery is an elective reconstructive option to treat symptomatic macromastia. There are several accepted ways to do the reduction surgical procedure where the blood supply (pedicle) to the nipple can vary and the visible scars can be in a horizontal, vertical, or Wise pattern. Technique is usually based on surgeon training, comfort, and preference. There are several known complications specific to this operation that include asymmetry, changes in nipple sensation, unattractive scars, diminished ability to breastfeed, and wound complications.5-7 Wound complications include seroma, hematoma, dehiscence, infection, wound breakdown, skin necrosis, and nipple necrosis.
This study focused on wound complications with the objective of identifying and modifying risk factors. Two known risk factors documented in the literature, nicotine use and obesity, already had been addressed by our service, and results were known anecdotally but had not been previously verified. This study also looked at other potential risk factors, including the pedicle location, skin incision, and DM status.
Residents or fellows participated in all the surgeries. An outcome analysis from The American College of Surgeons National Surgical Quality Improvement Program database from 2005 to 2011 found that resident participation was associated with morbidity, including wound complications.8 This study was performed at a federal hospital with a complexity level 1a rating, which is designated based on the highest level of patient volume, risk, teaching, research, intensive care unit beds, and specialty services.9 The hospital is closely affiliated with a level 1 trauma center and teaching hospital; therefore, resident and fellow participation is not a modifiable risk factor.
This study did not find an increased risk of wound complications in patients with DM, which has been found to be an independent risk factor in a prior study.10 DM status was indicated in only 3 histories, and they all had perioperative hemoglobin A1c levels < 8%. There is documentation of patients receiving perioperative antibiotics in 99 out of 116 of the surgical records; however, we did not include this in the analysis because the operative reports from the first year of the study were incomplete.
Smoking is a known risk factor for local wound complications in breast reduction surgery.10-15 The VA has a smoking cessation program through its mental health service that provides counseling and medication treatment options, including nicotine replacement, bupropion, and varenicline. We require patients to be at least 4 weeks nicotine free before surgery, which has been previously recommended in the literature.16
Existing studies that compare the traditional Wise pattern/inferior pedicle with vertical pattern/superior medial pedicle did not find an increased risk of wound complications.17-19 Our study separated the different incisions from the pedicle because the surgical technique among the different surgeons in the study varied, where sometimes the traditional Wise pattern was combined with the less traditional superior-medial pedicle. We did not find a statistical difference when comparing the incisions and pedicle location, which suggests that the incision type and source of blood supply to the nipple are not the determining factors for wound complications in the early postoperative period.
Obesity is a known risk factor for local wound complications.12,13,15,20-22 Studies have shown that patients who are obese benefit from breast reduction surgery; authors have argued against restricting surgery to these higher risk patients.4,23-25 Patients usually report decades of macromastia symptoms at consultation; so, we believe delaying the surgical procedure to get patients to a safer risk profile is in their best interest. We chose a cutoff BMI of 32 as a realistic value rather than 30, which is considered the definition of obesity. Patients at MRVAMC have access to MOVE!, a weight loss management program through primary care. We believe in being reasonable; so if a patient makes a significant improvement in her health but falls short of the required cutoff, we will still consider offering the surgical procedure after a full explanation of the surgical risks.
Wound complications, especially those that require admission or frequent appointments can seriously disrupt a patient’s life, creating unnecessary hardships and expense in time lost from work, travel, and child care. MRVAMC has a catchment area the size of North Carolina; so many of our patients travel hours for their appointments. The added scars and deformity from wound dehiscence and debridement can lead to asymmetry, widened scars, and future revision operations. Multiple clinic appointments for wound care not only impact that individual patient, but also has the effect of limiting access for all patients in a health care environment with high patient volume and limited providers, operating room time, and clinic appointments. As a result, minimizing predictable wound complications benefits the entire system.
Limitations and Strengths
This retrospective review comprised multiple different surgeons, including faculty and trainees, who were involved in the consultation, surgery, and postoperative care of the patients over a 20-year period; therefore, consistency in documentation is lacking. In addition, we were limited to only the information available on the charts. For example, wound size and laterality were not consistently documented. The MRVAMC complication rate was consistent with the current literature (range, 14-52%).12,18,20,24
The major strength of the study is that the veterans tend to stay within the VA, which makes complications easier to identify and follow. Patients who do not present initially to their surgeon due to travel limitations will typically contact their primary care provider or present to their local VA urgent care or emergency department provider, who will route the patient back to the surgical specialty service through the electronic health record.
Conclusions
Breast reduction surgery has a high wound complication rate, which can be predicted and improved on so that patients can receive their indicated surgical procedure with minimal inconvenience and downtime. This review confirms that preoperative weight loss and nicotine cessation were the appropriate focus of the MRVAMC plastic surgery service’s efforts to achieve a safer surgical experience. We will continue to enforce our protocol and encourage patients who are interested in breast reduction surgery and fall outside the requirements to work with their primary care provider on smoking cessation and weight loss through better nutrition and physical activity.
Acknowledgment
This manuscript is the result of work supported with resources and the use of facilities at the North Florida/South Georgia Veterans Health System in Gainesville, Florida.
Women make up an estimated 10% of the veteran population.1 The US Department of Veterans Affairs (VA) projected that there would be an increase of 18,000 female veterans per year for 10 years based on 2015 data. The number of women veterans enrolled in the VA health care increased from 397,024 to 729,989 (83.9%) between 2005 and 2015.2 This rise in the number of enrolled women veterans also increased the demand for female-specific health care services, such as breast reduction surgery, a reconstructive procedure provided at the Malcom Randall VA Medical Center (MRVAMC) federal teaching hospital in Gainesville, Florida.
Patients who experience symptomatic macromastia will report a history of neck and shoulder pain, shoulder grooving from bra straps, inframammary intertrigo, difficulty finding clothes that fit, and discomfort participating in sports. For the treatment of symptomatic macromastia, patients report a high satisfaction rate after breast reduction surgery.3-5 Unfortunately, the complications from the surgery can significantly disrupt a woman’s life due to previously unplanned hospital admissions, clinic appointments, wound care, time off work, and poor aesthetic outcome. Faculty awareness of a large number of complications for patients after breast reduction surgery prompted the MRVAMC Plastic Surgery Service to establish a stricter surgical screening protocol using body mass index (BMI) values and negative nicotine status to help patients be healthier and reduce the potential risk before offering surgery. A medical literature search did not find an existing study on veteran-specific breast reduction surgery.
Methods
The University of Florida and North Florida/South Georgia Veterans Health System Institutional Review Board approved a retrospective chart review of all breast reduction surgeries performed at MRVAMC over a 20-year period (July 1, 2000-June 30, 2020). Electronic health records were queried for all primary bilateral breast reduction surgeries performed for symptomatic macromastia using Current Procedural Terminology code 19318. Potentially modifiable or predictable risk factors for wound complications were recorded: nicotine status, BMI, diabetes mellitus (DM) status, skin incision pattern, and pedicle location. Skin incision patterns were either vertical (periareolar plus a vertical scar from the areola to the inframammary fold) or traditional Wise pattern (also known as anchor pattern: periareolar scar, vertical scar to inframammary fold, plus a horizontal scar along the inframammary fold) as seen in Figures 1 and 2. The pedicle is the source of blood supply to the nipple, which was documented as either from the inferior aspect or the superior or superior/medial aspect.
For this study, the blood supply from the superior and superior/medial was logged in the same category. Records were reviewed 3 months after surgery for documentation of local wound complications, such as hematoma, infection, wound breakdown, skin necrosis, and nipple necrosis. Major complications were defined as requiring an unplanned hospital admission or urgent return to the operating room. A χ2 test using a P value of < .05 was used to determine statistical significance between the incidence of wound complications and the individually identifiable variables.
Results
One hundred fifteen bilateral breast reduction surgeries were performed at MRVAMC over a 20-year period. Patient median age was 43 years. Median combined specimen weight was 1272 g. Forty-eight (41.7%) wound complications were documented, including 8 (7%) major complications. Most complications were identified in the first 7 years of the study before the new protocol and consult template became active. The new template resulted in the local complication rate dropping from 62% (July 2000-June 2007) to 26% (July 2007-June 2020). BMI > 32 (P = .03) and active nicotine use (P = .004) were found to be statistically significant independent risk factors for wound complications. Median BMI for all patients was 30. DM status (P = .22), skin incision pattern (P = .25), and pedicle location (P = .13) were not found to be predictors of wound complications (Table). There was no significant change in the incidence of major complications before and after the new protocols were enforced.
Discussion
Breast reduction surgery is an elective reconstructive option to treat symptomatic macromastia. There are several accepted ways to do the reduction surgical procedure where the blood supply (pedicle) to the nipple can vary and the visible scars can be in a horizontal, vertical, or Wise pattern. Technique is usually based on surgeon training, comfort, and preference. There are several known complications specific to this operation that include asymmetry, changes in nipple sensation, unattractive scars, diminished ability to breastfeed, and wound complications.5-7 Wound complications include seroma, hematoma, dehiscence, infection, wound breakdown, skin necrosis, and nipple necrosis.
This study focused on wound complications with the objective of identifying and modifying risk factors. Two known risk factors documented in the literature, nicotine use and obesity, already had been addressed by our service, and results were known anecdotally but had not been previously verified. This study also looked at other potential risk factors, including the pedicle location, skin incision, and DM status.
Residents or fellows participated in all the surgeries. An outcome analysis from The American College of Surgeons National Surgical Quality Improvement Program database from 2005 to 2011 found that resident participation was associated with morbidity, including wound complications.8 This study was performed at a federal hospital with a complexity level 1a rating, which is designated based on the highest level of patient volume, risk, teaching, research, intensive care unit beds, and specialty services.9 The hospital is closely affiliated with a level 1 trauma center and teaching hospital; therefore, resident and fellow participation is not a modifiable risk factor.
This study did not find an increased risk of wound complications in patients with DM, which has been found to be an independent risk factor in a prior study.10 DM status was indicated in only 3 histories, and they all had perioperative hemoglobin A1c levels < 8%. There is documentation of patients receiving perioperative antibiotics in 99 out of 116 of the surgical records; however, we did not include this in the analysis because the operative reports from the first year of the study were incomplete.
Smoking is a known risk factor for local wound complications in breast reduction surgery.10-15 The VA has a smoking cessation program through its mental health service that provides counseling and medication treatment options, including nicotine replacement, bupropion, and varenicline. We require patients to be at least 4 weeks nicotine free before surgery, which has been previously recommended in the literature.16
Existing studies that compare the traditional Wise pattern/inferior pedicle with vertical pattern/superior medial pedicle did not find an increased risk of wound complications.17-19 Our study separated the different incisions from the pedicle because the surgical technique among the different surgeons in the study varied, where sometimes the traditional Wise pattern was combined with the less traditional superior-medial pedicle. We did not find a statistical difference when comparing the incisions and pedicle location, which suggests that the incision type and source of blood supply to the nipple are not the determining factors for wound complications in the early postoperative period.
Obesity is a known risk factor for local wound complications.12,13,15,20-22 Studies have shown that patients who are obese benefit from breast reduction surgery; authors have argued against restricting surgery to these higher risk patients.4,23-25 Patients usually report decades of macromastia symptoms at consultation; so, we believe delaying the surgical procedure to get patients to a safer risk profile is in their best interest. We chose a cutoff BMI of 32 as a realistic value rather than 30, which is considered the definition of obesity. Patients at MRVAMC have access to MOVE!, a weight loss management program through primary care. We believe in being reasonable; so if a patient makes a significant improvement in her health but falls short of the required cutoff, we will still consider offering the surgical procedure after a full explanation of the surgical risks.
Wound complications, especially those that require admission or frequent appointments can seriously disrupt a patient’s life, creating unnecessary hardships and expense in time lost from work, travel, and child care. MRVAMC has a catchment area the size of North Carolina; so many of our patients travel hours for their appointments. The added scars and deformity from wound dehiscence and debridement can lead to asymmetry, widened scars, and future revision operations. Multiple clinic appointments for wound care not only impact that individual patient, but also has the effect of limiting access for all patients in a health care environment with high patient volume and limited providers, operating room time, and clinic appointments. As a result, minimizing predictable wound complications benefits the entire system.
Limitations and Strengths
This retrospective review comprised multiple different surgeons, including faculty and trainees, who were involved in the consultation, surgery, and postoperative care of the patients over a 20-year period; therefore, consistency in documentation is lacking. In addition, we were limited to only the information available on the charts. For example, wound size and laterality were not consistently documented. The MRVAMC complication rate was consistent with the current literature (range, 14-52%).12,18,20,24
The major strength of the study is that the veterans tend to stay within the VA, which makes complications easier to identify and follow. Patients who do not present initially to their surgeon due to travel limitations will typically contact their primary care provider or present to their local VA urgent care or emergency department provider, who will route the patient back to the surgical specialty service through the electronic health record.
Conclusions
Breast reduction surgery has a high wound complication rate, which can be predicted and improved on so that patients can receive their indicated surgical procedure with minimal inconvenience and downtime. This review confirms that preoperative weight loss and nicotine cessation were the appropriate focus of the MRVAMC plastic surgery service’s efforts to achieve a safer surgical experience. We will continue to enforce our protocol and encourage patients who are interested in breast reduction surgery and fall outside the requirements to work with their primary care provider on smoking cessation and weight loss through better nutrition and physical activity.
Acknowledgment
This manuscript is the result of work supported with resources and the use of facilities at the North Florida/South Georgia Veterans Health System in Gainesville, Florida.
1. US Department of Veterans Affairs. Statistics at a glance. Published February 2020. Accessed June 18, 2021. https://www.va.gov/vetdata/docs/Quickfacts/Homepage_slideshow_4_6_20.PDF
2. US Department of Veterans Affairs, National Center for Veterans Analysis and Statistics. Women veterans report: the past, present, and future of women veterans. Published February 2017. Accessed June 18, 2020. https://www.va.gov/vetdata/docs/specialreports/women_veterans_2015_final.pdf
3. Crittenden TA, Watson DI, Ratcliffe J, Griffin PA, Dean NR. Outcomes of breast reduction surgery using the breast-q: a prospective study and comparison with normative data. Plast Reconstr Surg. 2019;144(5):1034-1044. doi:10.1097/PRS.0000000000006114
4. Thoma A, Sprague S, Veltri K, Duku E, Furlong W. A prospective study of patients undergoing breast reduction surgery: health-related quality of life and clinical outcomes. Plast Reconstr Surg. 2007;120(1):13-26. doi:10.1097/01.prs.0000263370.94191.90
5. Nuzzi LC, Firriolo JM, Pike CM, DiVasta AD, Labow BI. Complications and quality of life following reduction mammaplasty in adolescents and young women.Plast Reconstr Surg. 2019;144(3):572-581. doi:10.1097/PRS.0000000000005907
6. Hall-Findlay EJ, Shestak KC. Breast reduction. Plast Reconstr Surg. 2015;136(4):531e-544e. doi:10.1097/PRS.0000000000001622
7. Kraut RY, Brown E, Korownyk C, et al. The impact of breast reduction surgery on breastfeeding: systematic review of observational studies. PLoS One. 2017;12(10):e0186591. doi:10.1371/journal.pone.0186591
8. Fischer JP, Wes AM, Kovach SJ. The impact of surgical resident participation in breast reduction surgery--outcome analysis from the 2005-2011 ACS-NSQIP datasets. J Plast Surg Hand Surg. 2014;48(5):315-321. doi:10.3109/2000656X.2014.882345
9. Site Facility Name and Complexity Summary of VHA Facility. Accessed June 18, 2021. https://www.vendorportal.ecms.va.gov/FBODocumentServer/DocumentServer.aspx?DocumentId=2793591&FileName=VA118-16-R-1059-A00002002.docx
10. Lewin R, Göransson M, Elander A, Thorarinsson A, Lundberg J, Lidén M. Risk factors for complications after breast reduction surgery. J Plast Surg Hand Surg. 2014;48(1):10-14. doi:10.3109/2000656X.2013.791625
11. Cunningham BL, Gear AJ, Kerrigan CL, Collins ED. Analysis of breast reduction complications derived from the BRAVO study. Plast Reconstr Surg. 2005;115(6):1597-1604. doi:10.1097/01.prs.0000160695.33457.db
12. Karamanos E, Wei B, Siddiqui A, Rubinfeld I. Tobacco use and body mass index as predictors of outcomes in patients undergoing breast reduction mammoplasty. Ann Plast Surg. 2015;75(4):383-387. doi:10.1097/SAP.0000000000000192
13. Manahan MA, Buretta KJ, Chang D, Mithani SK, Mallalieu J, Shermak MA. An outcomes analysis of 2142 breast reduction procedures. Ann Plast Surg. 2015;74(3):289-292. doi:10.1097/SAP.0b013e31829d2261
14. Hillam JS, Borsting EA, Chim JH, Thaller SR. Smoking as a risk factor for breast reduction: an analysis of 13,503 cases. J Plast Reconstr Aesthet Surg. 2017;70(6):734-740. doi:10.1016/j.bjps.2016.12.012
15. Zhang MX, Chen CY, Fang QQ, et al. Risk factors for complications after reduction mammoplasty: a meta-analysis. PLoS One. 2016;11(12):e0167746. doi:10.1371/journal.pone.0167746
16. Sørensen LT. Wound healing and infection in surgery: the pathophysiological impact of smoking, smoking cessation, and nicotine replacement therapy: a systematic review. Ann Surg. 2012;255(6):1069-1079.doi:10.1097/SLA.0b013e31824f632d
17. Antony AK, Yegiyants SS, Danielson KK, et al. A matched cohort study of superomedial pedicle vertical scar breast reduction (100 breasts) and traditional inferior pedicle Wise-pattern reduction (100 breasts): an outcomes study over 3 years. Plast Reconstr Surg. 2013;132(5):1068-1076. doi:10.1097/PRS.0b013e3182a48b2d
18. Hunter-Smith DJ, Smoll NR, Marne B, Maung H, Findlay MW. Comparing breast-reduction techniques: time-to-event analysis and recommendations. Aesthetic Plast Surg. 2012;36(3):600-606. doi:10.1007/s00266-011-9860-3
19. Ogunleye AA, Leroux O, Morrison N, Preminger AB. Complications after reduction mammaplasty: a comparison of wise pattern/inferior pedicle and vertical scar/superomedial pedicle. Ann Plast Surg. 2017;79(1):13-16. doi:10.1097/SAP.0000000000001059
20. Bauermeister AJ, Gill K, Zuriarrain A, Earle SA, Newman MI. Reduction mammaplasty with superomedial pedicle technique: a literature review and retrospective analysis of 938 consecutive breast reductions. J Plast Reconstr Aesthet Surg. 2019;72(3):410-418. doi:10.1016/j.bjps.2018.12.004
21. Nelson JA, Fischer JP, Chung CU, et al. Obesity and early complications following reduction mammaplasty: an analysis of 4545 patients from the 2005-2011 NSQIP datasets. J Plast Surg Hand Surg. 2014;48(5):334-339. doi:10.3109/2000656X.2014.886582
22. Kreithen J, Caffee H, Rosenberg J, et al. A comparison of the LeJour and Wise pattern methods of breast reduction. Ann Plast Surg. 2005;54(3):236-241. doi:10.3109/2000656X.2014.886582
23. Güemes A, Pérez E, Sousa R, et al. Quality of life and alleviation of symptoms after breast reduction for macromastia in obese patients: is surgery worth it? Aesthetic Plast Surg. 2016;40(1):62-70. doi:10.1007/s00266-015-0601-x
24. Setälä L, Papp A, Joukainen S, et al. Obesity and complications in breast reduction surgery: are restrictions justified? J Plast Reconstr Aesthet Surg. 2009;62(2):195-199. doi:10.1016/j.bjps.2007.10.043
25. Shah R, Al-Ajam Y, Stott D, Kang N. Obesity in mammaplasty: a study of complications following breast reduction. J Plast Reconstr Aesthet Surg. 2011;64(4):508-514. doi:10.1016/j.bjps.2007.10.043
1. US Department of Veterans Affairs. Statistics at a glance. Published February 2020. Accessed June 18, 2021. https://www.va.gov/vetdata/docs/Quickfacts/Homepage_slideshow_4_6_20.PDF
2. US Department of Veterans Affairs, National Center for Veterans Analysis and Statistics. Women veterans report: the past, present, and future of women veterans. Published February 2017. Accessed June 18, 2020. https://www.va.gov/vetdata/docs/specialreports/women_veterans_2015_final.pdf
3. Crittenden TA, Watson DI, Ratcliffe J, Griffin PA, Dean NR. Outcomes of breast reduction surgery using the breast-q: a prospective study and comparison with normative data. Plast Reconstr Surg. 2019;144(5):1034-1044. doi:10.1097/PRS.0000000000006114
4. Thoma A, Sprague S, Veltri K, Duku E, Furlong W. A prospective study of patients undergoing breast reduction surgery: health-related quality of life and clinical outcomes. Plast Reconstr Surg. 2007;120(1):13-26. doi:10.1097/01.prs.0000263370.94191.90
5. Nuzzi LC, Firriolo JM, Pike CM, DiVasta AD, Labow BI. Complications and quality of life following reduction mammaplasty in adolescents and young women.Plast Reconstr Surg. 2019;144(3):572-581. doi:10.1097/PRS.0000000000005907
6. Hall-Findlay EJ, Shestak KC. Breast reduction. Plast Reconstr Surg. 2015;136(4):531e-544e. doi:10.1097/PRS.0000000000001622
7. Kraut RY, Brown E, Korownyk C, et al. The impact of breast reduction surgery on breastfeeding: systematic review of observational studies. PLoS One. 2017;12(10):e0186591. doi:10.1371/journal.pone.0186591
8. Fischer JP, Wes AM, Kovach SJ. The impact of surgical resident participation in breast reduction surgery--outcome analysis from the 2005-2011 ACS-NSQIP datasets. J Plast Surg Hand Surg. 2014;48(5):315-321. doi:10.3109/2000656X.2014.882345
9. Site Facility Name and Complexity Summary of VHA Facility. Accessed June 18, 2021. https://www.vendorportal.ecms.va.gov/FBODocumentServer/DocumentServer.aspx?DocumentId=2793591&FileName=VA118-16-R-1059-A00002002.docx
10. Lewin R, Göransson M, Elander A, Thorarinsson A, Lundberg J, Lidén M. Risk factors for complications after breast reduction surgery. J Plast Surg Hand Surg. 2014;48(1):10-14. doi:10.3109/2000656X.2013.791625
11. Cunningham BL, Gear AJ, Kerrigan CL, Collins ED. Analysis of breast reduction complications derived from the BRAVO study. Plast Reconstr Surg. 2005;115(6):1597-1604. doi:10.1097/01.prs.0000160695.33457.db
12. Karamanos E, Wei B, Siddiqui A, Rubinfeld I. Tobacco use and body mass index as predictors of outcomes in patients undergoing breast reduction mammoplasty. Ann Plast Surg. 2015;75(4):383-387. doi:10.1097/SAP.0000000000000192
13. Manahan MA, Buretta KJ, Chang D, Mithani SK, Mallalieu J, Shermak MA. An outcomes analysis of 2142 breast reduction procedures. Ann Plast Surg. 2015;74(3):289-292. doi:10.1097/SAP.0b013e31829d2261
14. Hillam JS, Borsting EA, Chim JH, Thaller SR. Smoking as a risk factor for breast reduction: an analysis of 13,503 cases. J Plast Reconstr Aesthet Surg. 2017;70(6):734-740. doi:10.1016/j.bjps.2016.12.012
15. Zhang MX, Chen CY, Fang QQ, et al. Risk factors for complications after reduction mammoplasty: a meta-analysis. PLoS One. 2016;11(12):e0167746. doi:10.1371/journal.pone.0167746
16. Sørensen LT. Wound healing and infection in surgery: the pathophysiological impact of smoking, smoking cessation, and nicotine replacement therapy: a systematic review. Ann Surg. 2012;255(6):1069-1079.doi:10.1097/SLA.0b013e31824f632d
17. Antony AK, Yegiyants SS, Danielson KK, et al. A matched cohort study of superomedial pedicle vertical scar breast reduction (100 breasts) and traditional inferior pedicle Wise-pattern reduction (100 breasts): an outcomes study over 3 years. Plast Reconstr Surg. 2013;132(5):1068-1076. doi:10.1097/PRS.0b013e3182a48b2d
18. Hunter-Smith DJ, Smoll NR, Marne B, Maung H, Findlay MW. Comparing breast-reduction techniques: time-to-event analysis and recommendations. Aesthetic Plast Surg. 2012;36(3):600-606. doi:10.1007/s00266-011-9860-3
19. Ogunleye AA, Leroux O, Morrison N, Preminger AB. Complications after reduction mammaplasty: a comparison of wise pattern/inferior pedicle and vertical scar/superomedial pedicle. Ann Plast Surg. 2017;79(1):13-16. doi:10.1097/SAP.0000000000001059
20. Bauermeister AJ, Gill K, Zuriarrain A, Earle SA, Newman MI. Reduction mammaplasty with superomedial pedicle technique: a literature review and retrospective analysis of 938 consecutive breast reductions. J Plast Reconstr Aesthet Surg. 2019;72(3):410-418. doi:10.1016/j.bjps.2018.12.004
21. Nelson JA, Fischer JP, Chung CU, et al. Obesity and early complications following reduction mammaplasty: an analysis of 4545 patients from the 2005-2011 NSQIP datasets. J Plast Surg Hand Surg. 2014;48(5):334-339. doi:10.3109/2000656X.2014.886582
22. Kreithen J, Caffee H, Rosenberg J, et al. A comparison of the LeJour and Wise pattern methods of breast reduction. Ann Plast Surg. 2005;54(3):236-241. doi:10.3109/2000656X.2014.886582
23. Güemes A, Pérez E, Sousa R, et al. Quality of life and alleviation of symptoms after breast reduction for macromastia in obese patients: is surgery worth it? Aesthetic Plast Surg. 2016;40(1):62-70. doi:10.1007/s00266-015-0601-x
24. Setälä L, Papp A, Joukainen S, et al. Obesity and complications in breast reduction surgery: are restrictions justified? J Plast Reconstr Aesthet Surg. 2009;62(2):195-199. doi:10.1016/j.bjps.2007.10.043
25. Shah R, Al-Ajam Y, Stott D, Kang N. Obesity in mammaplasty: a study of complications following breast reduction. J Plast Reconstr Aesthet Surg. 2011;64(4):508-514. doi:10.1016/j.bjps.2007.10.043