Skip to main content

Database derived from an electronic medical record-based surveillance network of US emergency department patients with acute respiratory illness

Abstract

Background

For surveillance of episodic illness, the emergency department (ED) represents one of the largest interfaces for generalizable data about segments of the US public experiencing a need for unscheduled care. This protocol manuscript describes the development and operation of a national network linking symptom, clinical, laboratory and disposition data that provides a public database dedicated to the surveillance of acute respiratory infections (ARIs) in EDs.

Methods

The Respiratory Virus Laboratory Emergency Department Network Surveillance (RESP-LENS) network includes 26 academic investigators, from 24 sites, with 91 hospitals, and the Centers for Disease Control and Prevention (CDC) to survey viral infections. All data originate from electronic medical records (EMRs) accessed by structured query language (SQL) coding. Each Tuesday, data are imported into the standard data form for ARI visits that occurred the prior week (termed the index file); outcomes at 30 days and ED volume are also recorded. Up to 325 data fields can be populated for each case. Data are transferred from sites into an encrypted Google Cloud Platform, then programmatically checked for compliance, parsed, and aggregated into a central database housed on a second cloud platform prior to transfer to CDC.

Results

As of August, 2023, the network has reported data on over 870,000 ARI cases selected from approximately 5.2 million ED encounters. Post-contracting challenges to network execution have included local shifts in testing policies and platforms, delays in ICD-10 coding to detect ARI cases, and site-level personnel turnover. The network is addressing these challenges and is poised to begin streaming weekly data for dissemination.

Conclusions

The RESP-LENS network provides a weekly updated database that is a public health resource to survey the epidemiology, viral causes, and outcomes of ED patients with acute respiratory infections.

Peer Review reports

Background

The emergency department (ED) serves as a valuable site for surveillance because of its large patient volume and diverse representation of illness. At least 120 million Americans visit an ED each year, and about 5–10% of these patients present with an acute respiratory illness (ARI) [1]. This report describes the methodology for the creation and operation of a national network of EDs providing linked symptom, clinical, laboratory and disposition data weekly for patients presenting to the ED with ARI, aggregated in a single database. This work began as a retrospective registry known as RECOVER (REgistry of suspected COVID-19 in EmeRgency care) in March 2020 with the explicit purpose of tracking clinical features and outcomes of patients tested for SARS-CoV-2 [2]. That registry led to the development of the present network and database, funded by contract 75D30121C11813, and named the Respiratory Virus Laboratory Emergency Department Network Surveillance (RESP-LENS). A primary and unique approach of this surveillance network–that distinguishes it from existing networks–is the use of weekly extractions from electronic medical records to provide nearly real-time insight into the dynamics of viral transmission and severity of illness in patients seeking unscheduled care. Unlike other networks, RESP-LENS will link patient level data from the ED visit, such as vital signs, with results of laboratory testing, and then with 30 day outcome. These data are especially relevant to tracking the severity of illness associated with outbreaks of new SARS-CoV-2 variants [3, 4]. and comparing the relative impact of SARS-CoV-2, influenza and RSV circulation.

Construction and content

Overview

The database input is entirely dependent upon the RESP-LENS network, which represents a consortium of 24 emergency medicine-based clinician investigators from 21 US states. These investigators lead 24 sites (network hubs) that report data from 109 hospitals (network nodes), including 5 children’s hospitals. The sites and hospitals represent all 10 Department of Health and Human Services regions (Fig. 1 and Supplemental Table 1). All sites were recruited from the larger RECOVER network [2]. In terms of representativeness, the network includes hospitals that serve inner city urban populations (e.g., Detroit MI, San Francisco CA and New York City), many metropolitan and suburban regions, rural areas (e.g. Morganton, WV and Iowa City, IA) as well as cities at higher altitude (Colorado Springs CO and Salt Lake City, UT). Five hospitals are children’s hospitals. The network lacks any representation from the island regions such as Hawaii.

Fig. 1
figure 1

Locations of the 24 sites participating in RECOVER-CDC. Many sites have more than one hospital

For RESP-LENS, the definition of an ARI comes from one or more of the 130 ICD-10 diagnosis codes associated with the ED visit (Supplemental Table 2). These ICD-10 codes were derived from a consensus effort by investigators and CDC personnel and remain comprehensive of symptoms of illness caused by SARS-CoV-2 variants and other respiratory viruses [5]. All ED visits are identified electronically, using either Epic (Verona, WI) or Cerner (Kansas City, MO) electronic medical record systems. Because many patients have more than one visit within a week, hereafter we use the nomenclature “case” to define any given patient visit that satisfied the ICD definition of an ARI.

All case data are acquired from the electronic health record and auto-populate a database via structured query language (SQL), using either the Clarity or Electronic Data Warehouse (HealtheEDWSM, Kansas City, MO) data analytical systems [6]. Fig. 2 shows the overview of the data transmission workflow which begins at each site and is then transmitted to a centralized Google Cloud (Mountain View, CA) storage owned by the coordinating site, analyzed, parsed, and aggregated in the Study Maker (Newton, MA) cloud platform, and then ultimately transferred to CDC. This process runs weekly: sites transmit data on Mondays or Tuesdays to Google Cloud and the coordinating site transmits to CDC on Wednesdays.

Fig. 2
figure 2

Architecture of the RECOVER-CDC surveillance network

Network management

Wayne State University in Detroit, MI functions as the overall network coordinating center, which encompasses the roles of data coordinating center, site recruitment, contract management, creation of the REDCap data forms and their content, site payment, and repository of network documents including the manual of operations, data management plan, and study protocol. The network communicates via listserv email, quarterly all-site meetings via teleconference, and using Slack Teams™ (San Francisco, CA). The kick-off meeting occurred in September of 2021 via teleconference. All-site meetings always included representatives from CDC. After the program launched, sites proceeded to share information, including SQL code, toward creating the methods required for weekly data abstraction from the EMRs of the 24 sites. Each site requires, at minimum, a principal investigator (MD or PhD), and an informatics/coding professional.

Roles and responsibilities

The original protocol that described the role of Wayne State as the network and data coordinator was written by the Contact PI at Wayne State and submitted in response to a publicly posted funding opportunity announcement from CDC. With input from CDC, the Contact PI oversaw the recruitment of sites to represent all 10 DHHS regions, their subcontracting, and the content of the REDCap case report form. The Contact PI sets the agenda and leads the quarterly all-site meetings, answers questions about field definitions, and handles unexpected problems. The Contact PI wrote a data use agreement that was used by most sites and attends weekly organizational meetings with CDC. This meeting is open to all network participants. The Wayne State network data manager was responsible for setting up the secure Google Cloud platform and encrypted, password-protected folders for each site and maintaining a living field definition document that explains the standardization of codes to signify missing data (999 and -999 for string or numeric data respectively), unexpected characters (e.g., > or <), and out of range data (e.g., age > 120 years). The Wayne State project manager oversees communications, including scheduling, audiovisual recordings of meetings, maintenance of contact information and delegation of authority logs for all sites, storing documents and other files, and maintaining a manual of operations.

At each site, the site PI initiates and takes responsibility for executing the terms of the subcontract, data use agreement, and other regulatory aspects (i.e., IRB), attending all site meetings and addressing problems such as updating the SQL code in response to local changes in viral platform testing/ordering, and interruptions in data transfer because of local personnel changes. Additionally, Wayne State has subcontracted with StudyMaker to provide technical support and software systems for data ingestion, verification, export, and transmission, as well as to maintain a contemporaneous aggregated dataset mirrored back to Wayne State and delivered via exports to CDC.

Site analysts are responsible for writing and updating the SQL code and setting up the data transfer process to the Wayne Google Cloud using an API. The site analyst assists with communications with the Wayne State project manager and ensures timely data transfer on a weekly basis.

Representatives of CDC have been involved in all stages from the outset of creating the data collection forms, site selection, all-site meetings, and weekly feedback to sites. CDC is responsible for general oversight and execution of the contract supporting this work and managing the project to ensure data needed to meet CDC objectives are obtained weekly. To accomplish this, CDC staff were involved in developing all aspects of the data collection protocol and forms, provided input on site selection as needed, analyzed weekly data and provided feedback to the coordinating site regarding data anomalies and questions to ensure accurate data interpretation.

Overall governance is run by an executive committee consisting of the contact PI, four members of CDC, and the two consultants (a StudyMaker representative and the outside epidemiologist). Data sharing is overseen by the data use committee, comprising all 24 site PIs.

Regulatory issues and registrations

The protocol has been reviewed by all site institutional review boards and deemed non-human subjects research, exempted from review or approved with expedited review. Except for patient ages and dates of service, the data transmitted for RESP-LENS contains no other protected health information. The data management plan has been registered using the DMP tool and is publicly available (https://dmptool.org/plans/70148). The University of Colorado was instrumental in this project's development of the first SQL script for use with Epic; unfortunately, the State of Colorado had a constitutional statement that barred the University of Colorado from remaining in the network because of risks associated with Federal Acquisition Regulation (FAR) clauses in the Prime agreement.

Case data acquisition

Figure 3 shows a schematic overview of the network and flow of data. REDCap was utilized to develop a shareable mechanism for data standardization and validation (in the form of CSV project data dictionaries or XML project files). It also provided a facile approach for distributing project updates as revisions were made to the data collection process.

Fig. 3
figure 3

Schematic of the flow and processing of data

The REDCAP database was constructed in three data collection forms: (1) the 173 field “ndex visit” form that captures initial ED visit data (clinical, laboratory, disposition and demographic) from ARI cases, (2) the “30-day” form that includes 136 fields for results of new or repeated laboratory and diagnostic testing, as well as, admission and survival outcome for ARI cases; and (3) the third file is the “ED volume” form that includes overall ED visit data, including age and admission status, for both ARI and non-ARI cases. Data are captured at the encounter level rather than patient level with each qualifying visit logged consecutively, without regard to prior entry while still maintaining patient-level identifiers. Sites transmit encrypted or hashed identifiers (e.g., Financial record number[FIN] and the Medical record number [MRN]) rather than the original identifiers to protect confidentiality while aligning index and 30-day data and facilitating the tracking of individual subjects over time. The index and 30-day files include 325 fields shown in Supplemental File 3. For sex, race and ethnicity fields, all sites used the designations provided by the patient. The ED volume file shows the age and admission status of each ED patient seen on an individual row.

Each site’s local copy of the network-wide REDCap projects provided them with the means to perform automated data validation on their EMR query results. All reports generated by site EMR queries (possibly subject to post-processing) can be tested for compliance via a manual or API directed import into the site’s local REDCap projects. If query files load successfully into the site’s unaltered copy of the network-wide REDCap projects, they will load successfully into the corresponding projects housed in Wayne State’s REDCap environment.

Case data submission

On a weekly basis, site analysts upload the “index visit,” “30-day,” and “ED volume” files from their local REDCAP database to a Google Cloud Storage (GCS) environment maintained by Wayne State. Each site has one or more designated individuals who have been granted access to a secure, specific bucket in the GCS environment. Each site’s bucket contains three primary folders for each file type reported from sites – “index visit,” “30-day”, and “ED volume.” The three subfolders each contain respective subfolders for storing ingestion log information and archiving site files that have been successfully loaded or rejected according to whether they meet the data specification.

Data processing

In the initial phase of the study, REDCap was utilized as the study data repository and transport mechanism until the volume of data generated exceeded the REDCAP server's operational capacity. At the onset of the study, all files uploaded from sites to the GCS buckets were processed using a series of Python utilities. The first was a validation stage in which an attempt was made to load incoming site files to a test project via an “Import Records” REDCap API method. The result of this attempted load was logged and recorded in the GCS bucket subfolder for each respective site. Files loaded successfully were transferred into a queue for the second processing stage, namely the transfer to CDC. In cases when files failed to load, the log was reviewed to determine the cause, or causes, of failure, and these details were communicated back to sites. Load failure could be caused in all file types (Index, 30-Day, or ED Volume) by data related errors, incorrect field names, additional fields not listed in the network-wide project specification, data type inconsistencies (i.e. transmission of text data for numeric fields), and data coding inconsistencies (i.e. using improper numeric coding for categorical responses). Thirty-Day Follow-up files could also fail validation due to the presence of unmatched records. By definition, 30-day follow-up data must provide information for an existing index record. If a record in a 30-day follow-up file does not have a matching index record in the cumulative database, it either represents a new, previously unidentified case, or it contains a malformed record ID. These newly discovered index cases often occur because of delay in charting by emergency physicians, especially on visits made near the end of the week, followed by delays from coding and billing services. By protocol, the index files are “trued up” on at least a quarterly basis to include these visits that were newly discovered at 30-day follow-up.

As more sites executed their contracts and completed their SQL coding, the weekly volume of data began to escalate. The volume of data eventually began to affect the performance of the required API methods for import and export. This resulted in frequent 503 HTTP errors where the resource demand generated by the API exceeded the REDCap server’s operational capacity. Even after adjustments to the REDCap server memory allocation and timeout settings, resource related errors occurred too frequently for REDCap to remain a tenable solution for data storage and transfer.

To overcome bottlenecks and extend data management capabilities, the central functions of REDCap at Wayne State (data validation, data aggregation, and data export generation) were replaced with a dedicated data parser and connected data handling and data management tools in the StudyMaker cloud platform. Accordingly, Parquet was deployed to parse data and prepare data exports in convenient formats to improve data transmission to the CDC central data lake (software and systems maintained by StudyMaker, Newton, MA). This new architecture retains full compatibility with REDCap, so that no changes were required by the Sites, and so that Sites can continue to use REDCap locally for data validation and assessment.

Assessing data fidelity

For this report, fidelity refers to the consistency and accuracy of data upload from each site. To monitor the successful transmission and accuracy of data submitted by sites, files uploaded from sites are automatically detected, tested for compliance and presence of mandatory data. In the event of a missing or late file, the system automatically generates an email to the site PI and lead coder. We produced a dashboard-based monitoring system that demonstrates all inbound data from networks, including the time of upload and the percentage of fields reported as missing or empty (that latter is also known as a null value). Fields could be missing or empty for least three reasons: (1) Truly missing: data that should have been present in the EMR but were absent either from clerical omission or the data were unknown (an example of the latter, could be for an unconscious patient with no prior medical record) (2) Empty or null: failure to find the data because of SQL coding error, or (3) Pseudomissing: if the patient only had an index visit, and no other contact within 30 days, most of the questions at 30 days had no data to fill.These dashboards are provided in a secure study portal for participants to (1) track their own data quality performance, (2) track trends and key endpoints from live data submitted to the network, and (3) explore the full dataset for research purposes through a graphical interface. Figure 4 shows an export of files loaded from week 48 of 2022 in the “Weekly counter” dashboard.

Fig. 4
figure 4

Representation of the weekly counter verification and validation dashboard, with visualization of outliers

Production

All files that successfully completed the validation phase were placed in the queue for transfer to the production environment. The production phase again utilized the REDCap “Import Records” API method to import validated records into the production environment. The production environment served as the cumulative data repository, and it was also the source from which CDC pulled the weekly records. CDC’s weekly data export was performed using a Python script (coded at WSU) to implement an “Export Records” REDCap API method. The CDC representative was granted a key to access the WSU REDCap production project via the API. The export script was written to dynamically filter the exported record set to contain records that were new, or updated, since the last export. The export script also dynamically modified a SAS → template program, stored at CDC, to convert the raw REDCap export into a SAS formatted dataset.

Field coding

It is important to distinguish RESP-LENS from a retrospective human-based extraction of data. Instead, this is an entirely computerized method designed to reduce human error from routine data extraction and entry. Each index and 30-day file transferred contains fields that exist in two categories based upon logic. The first category includes fields that should always have data present (such as age, gender, and race) while the second category includes fields that may or may not have data present (such as the results of viral testing). However, for the first category, the reality of large datasets is that the fields that should always be present are not always present; as may be the case for noncommunicative patients with no personal identification who satisfied an ARI definition but cannot state their age, gender or race. In both cases, the EMR leaves the field as empty, (also referred to as a “null” value). Thus, all data fields have three states: coded and known, coded and uncertain (in the case of a John Doe, where it is a known unknown), and missing (or null, which raises the possibility of other entry error). Additionally, fields may contain either categorical (test done or not done, result positive or negative) or numerical values (e.g., vital signs). To address this configuration, as a convention, we adopted a convention of coding “999” as an indication of “not done” for categorical fields and “-999” to indicate not done for numeric fields. The negative sign was added to allow analysts to easily exclude all missing results as none of the measurements recorded in the study would legitimately have values less than zero. For missing dates “1/1/1900” is input. Additional codes were added to address indeterminate results of tests. A deeper ambiguity occurs in the rare case where a test is ordered, and no result can be found. The leadership deemed that this situation was equivalent to no test done, and as such, these results fields are coded as “999” or “-999” as appropriate. These conventions all represent tradeoffs to capture as much robust data as possible while limiting the complexity and total number of data fields collected.

Site-specific variations

Although all sites employ viral assays that use polymerase chain reaction detection, each site had numerous pathways within the EMR to order viral testing, and these order mechanisms could change with time, depending on supply chain issues. For example, one of the 24 sites has 18 unique orders to test for SARS-CoV-2, while another site only has four orders. Sites vary in their testing protocols including the use of multiplex viral testing and testing requirements for admitted patients. In some cases, the testing protocols change from weekdays to weekends. As part of the weekly organizational calls, representatives from CDC and the network management review relative changes in viral test ordering, and if a site has an outlier of more than a 50% change from prior weeks, an inquiry is made to the site. This has led to discovery of unrecognized changes in the electronic location of test ordering in 13 sites. As new diagnostic pathways and orders are developed at the site level, it will remain the sites' responsibility to identify these to avoid missing cases and diagnostic data. Additionally, many sites have had challenges with workforce turnover, particularly on the informatics side, which has led to occasional interruptions in data transmission. These examples reinforce the need for centralized, weekly human oversight to ensure complete and valid data.

Unexpected lessons learned

One issue that delayed initial contracting was that certain sites, particularly those that represented large networks of hospitals, required their own data use agreements, which added up to six months to the onboarding process. Some patients have multiple identifiers in the EHR due to registration errors or a lack of identifiers during an emergency presentation (e.g., "Jane or John Doe”). Accordingly, we used the FIN which is unique to each individual and visit as the internal tracking mechanism for the index visit. This was used to then update the hash code at the 30-day follow-up as the unique patient identifier. This 30-day follow-up leads to one of the most frequent challenges to the detection of ARI cases, namely the delay to ICD coding associated with the action of a patient being discharged or admitted to the hospital. This results in an increase in ARI cases at the time of 30-day follow-up—meaning that four weeks later, there are approximately 4% more cases discovered than had an ARI-defining ED diagnosis. Other anomalies include out-of-range variables that persistently remain coded in the EMR, probably representing human error. One example is age > 120 years, observed in 0.001% of cases. Coding errors required ongoing editing such as the mistaken use of “999” instead of “-999” and “1/1/1990″ to document missing numerals and dates, respectively.

Preliminary results

The publicly available interactive dashboard for RESP-LENS is available at https://www.cdc.gov/surveillance/resp-lens/dashboard.html. Although data collection began in late 2021, this dashboard includes data starting in October, 2022. As of August 20, 2023, the network has tracked 5,192,826 ED patients, of whom 873,905 (16.8%) had an ARI-defining ICD-10 code. Of these 873,905 patients, 63% had a diagnostic test for SARS-CoV-2 and 53.5% had a test for influenza; positivity for SARS-CoV-2 was 13.9% and for influenza A was 3.3%, respectively. The average age of patients was 41 years (SD 16), and 52% were reported in the EMR as female gender. The racial distribution of patients was diverse: 52% White, 27% Black, 2.3% Asian, 0.9% native Hawaiian or other Pacific origin, and 0.5% American Indian or Alaskan origin; 13.5% reported other or unknown race and 2.0% reported more than one raceMissing or empty data has occurred in < 1% of demographic and viral testing data. Fields that are most commonly missing from the EMR (i.e., no data to find) include body mass index (not documented in 49%) and vaccination status (not documented in 59%).

Utility and discussion

RESP-LENS has two broad aims: 1) provide timely, systematic national surveillance for ARI including both virologic and clinical data for ED visits on a weekly basis, and 2) create the infrastructure and methodology to report the epidemiology of viral infections, and their subtypes where possible, and associate them with specific diagnoses and outcomes. The network provides data to detect new outbreaks, understand the magnitude and distribution of respiratory viral infections in patients seeking unscheduled care, follow the natural history and sequela of viral infections, and evaluate control strategies [7]. Among the first outcomes investigated will be the strength of the association between viral infections and the new diagnosis of venous thromboembolism. The data will allow stratification of illness severity, based upon admission location, need for intubation (from CPT coding), and mortality within 30 days. Additionally, with relatively minor changes in SQL coding at each site, this network can serve as a platform for surveillance of other syndromes.

The COVID-19 era highlighted current deficiencies in US surveillance networks. The 2022 Senate Homeland Security and Governmental Affairs Committee report on COVID-19 concluded that the U.S. failed to sufficiently invest in public health preparedness across multiple administrations, and that U.S. public health surveillance systems for monitoring and detecting emerging infectious diseases are inadequate, antiquated and fragmented [9]. To our knowledge, RESP-LENS is the first ED-based, national surveillance network to link virologic, clinical and disposition data for multiple respiratory viruses in a single platform that provides data on a weekly basis. The advantage of this method includes the consistency and transparency offered by the coding strategy and the ability to link clinical data elements at the visit and patient level and provide the information weekly to inform situational awareness and guide public health action. Weekly data collection from hospital ED EMRs likely captures a different patient sample (in demographic and infection severity) than may be captured by health departments. The methodology used also allows the comparison of the outcomes of cases with ARI-defining ED diagnoses who are tested for viruses with those who are not tested. While new standards for data interchange from EMRs (notably the FHIR standard) provide promise to simplify future surveillance and data aggregation, the methodologies used in RESP-LENS extend the existing clinical research infrastructure to capture quality data at scale with continuous oversight and input of clinician-scientists [10].

Planned work and future directions

We will be able to integrate county-level CDC SVI and other health equity relevant exposure data, but not at the census tract level. This effort will leverage methodology used in Wayne State’s PHOENIX Virtual Data Warehouse, which has already assembled troves of social determinants of health data for ZIP codes, and has established standard operating procedures for integrating the information with investigator-generated person-level data [8]. We hope that publication of this protocol (and subsequent original research manuscripts) will enhance the visibility of RESP-LENS and encourage further investments to pursue deeper data, including viral genomic sequencing. Depending upon funding, we hope to link records to state immunization databases and compare our trends with other CDC databases.

Limitations

In addition to the unexpected lessons described above, other challenges include site-to-site variations and constant changes in test ordering, and local guidance on which patients should receive testing. The current methodology also remains vulnerable to late or missing uploads because it requires a human to remember to upload the files. Accordingly, a next step will be the creation of methodology to automate the upload process at each site. RESP-LENS is constrained by what is hard-coded in Cerner and Epic; their “other” or “more than one” race categories provide only those designations, and do not provide further racial or ethnic background. We plan to link records to valid state immunization databases, but until then, vaccination status relies upon documentation in the electronic record, which is either missing, incomplete, or lacking detail such as date of administration and vaccine manufacturer. The 30-day outcomes are restricted to the results in the medical record, and local data health exchanges. Clearly, we could miss deaths or severe outcomes that occurred in facilities outside the local systems. Other limitations include lack of viral genotyping, and outcomes only at 30 days, and lack of control over the clinical decision to perform viral testing [11].

Conclusion

The RESP-LENS national network provides weekly data describing the frequency and outcomes of patients with ARI-defining ED diagnoses in all 10 DHHS regions in the US. This network addresses the critical gap in accurate, timely pan-respiratory surveillance linking virologic and clinical data.

Availability of data and materials

Consistent with the purpose of surveillance, key situational awareness variables including weekly frequency data for viral testing and results by age group and HHS region are made publicly available from the CDC on several venues. The datasets are available at https://recover-cdc.com/login, and a login and password will be provided by the corresponding author on reasonable request. Researchers interested in accessing the data for publication will be asked to submit a project proposal which the data use committee will review. Requests are handled using a structured form that reports the title of the work, the investigators, the purpose and the analytical plan. Authorship order is decided before granting access to data, with the order determined by prior and anticipated contributions by the proposed authors. The data use committee will oversee authorship.

References

  1. National Hospital Ambulatory Medical Care Survey Emergency Department Data. https://www.cdc.gov/nchs/dhcs/ed-visits/index.htm: Centers for Disease Control.

  2. Kline JA, Pettit KL, Kabrhel C, Courtney DM, Nordenholz KE, Camargo CA. Multicenter registry of United States emergency department patients tested for SARS-CoV-2. J Am Coll Emerg Physicians Open. 2020;1(6):1341–8.

    Article  PubMed  PubMed Central  Google Scholar 

  3. Lewnard JA, Hong VX, Patel MM, Kahn R, Lipsitch M, Tartof SY. Clinical outcomes associated with SARS-CoV-2 Omicron (B.1.1.529) variant and BA.1/BA.1.1 or BA.2 subvariant infection in Southern California. Nature Medicine. 2022;28(9):1933–43.

  4. Nyberg T, Ferguson NM, Nash SG, Webster HH, Flaxman S, Andrews N, et al. Comparative analysis of the risks of hospitalisation and death associated with SARS-CoV-2 omicron (B.1.1.529) and delta (B.1.617.2) variants in England: a cohort study. Lancet. 2022;399(10332):1303–12.

  5. Geismar C, Nguyen V, Fragaszy E, Shrotri M, Navaratnam AMD, Beale S, et al. Symptom profiles of community cases infected by influenza, RSV, rhinovirus, seasonal coronavirus, and SARS-CoV-2 variants of concern. Sci Rep. 2023;13(1):12511.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  6. Harris PA, Taylor R, Thielke R, Payne J, Gonzalez N, Conde JG. Research electronic data capture (REDCap)–a metadata-driven methodology and workflow process for providing translational research informatics support. J Biomed Inform. 2009;42(2):377–81.

    Article  PubMed  Google Scholar 

  7. Teutsch S, Churchill R. Principles and Practice of Public Health Surveillance. 2nd ed. New York: Oxford University Press; 2000.

    Book  Google Scholar 

  8. Korzeniewski SJ, Bezold C, Carbone JT, Danagoulian S, Foster B, Misra D, et al. The Population Health OutcomEs aNd Information EXchange (PHOENIX) Program - A Transformative Approach to Reduce the Burden of Chronic Disease. Online J Public Health Inform. 2020;12(1): e3.

    Article  PubMed  PubMed Central  Google Scholar 

  9. Senate U. Historically unprepared: Examination of the Federal Government's Pandemic Prepardness and Initial COVID-19 response. file:///C:/Users/gr5041/Downloads/873516%20(3).pdf: HSGAC Majority Staff Report; 2022.

  10. Duda SN, Kennedy N, Conway D, Cheng AC, Nguyen V, Zayas-Cabán T, et al. HL7 FHIR-based tools and initiatives to support clinical research: a scoping review. J Am Med Inform Assoc. 2022;29(9):1642–53.

    Article  PubMed  PubMed Central  Google Scholar 

  11. Klann JG, Strasser ZH, Hutch MR, Kennedy CJ, Marwaha JS, Morris M, et al. Distinguishing Admissions Specifically for COVID-19 From Incidental SARS-CoV-2 Admissions: National Retrospective Electronic Health Record Study. J Med Internet Res. 2022;24(5): e37931.

    Article  PubMed  PubMed Central  Google Scholar 

Download references

Acknowledgements

Site investigators involved in RESP-LENS effort (to be accessible by PubMed):

PI

Institution (alphabetical)

Michael J. Roshon

Centura Health

Danielle E. Turner-Lawrence

Corewell Health William Beaumont University Hospital

Michael A. Puskarich

Hennepin Healthcare

Ka Ming Gordon Ngai

Icah School of Medicine at Mount Sinai

Benton R. Hunter

Indiana University School of Medicine

Joseph Bledsoe

Intermountain Healthcare

James K. d'Ettiene

JPS Health Network

Stephen S. Lim

LCMC Health

Christopher W. Kabrhel

Massachusetts General Hospital

Esther J. Choo

Oregon Health & Science University

Steven MH Moore

Penn State Health

Katherine R. Buck

The Ohio State University Wexner Medical Center

David M. Beiser

The University of Chicago Department of Medicine

James G. Galbraith

The University of Mississippi Medical Center

Joby E. Thoppil

The University of Texas, Southwestern

Troy Madsen

The University of Utah

Edward W. Castillo

University of California, San Diego

Israel E. Green-Hopkins

University of California, San Francisco

Kristen S. Nordenholz

University of Colorado Anschutz Medical Campus

Faheem C. Guirgis

University of Florida Health

Bryan Wilson

University of Iowa

Michael L. Pulia

University of Wisconsin-Madison

Stacey House

Washington University School of Medicine in St. Louis

Justine M. Pagenhardt

West Virginia University

Disclaimer

The findings and conclusions in this report are those of the authors and do not necessarily represent the official position of the Centers for Disease Control and Prevention, U.S. Department of Health and Human Services.

Funding

Funded by contract 75D30121C11813, and named the Respiratory Virus Laboratory Emergency Department Network Surveillance (RESP-LENS).

Author information

Authors and Affiliations

Authors

Contributions

JAK conceived the work, obtained funding, organized the network and wrote the manuscript BR co-created and oversees the data cleaning, transmission and storage system storage for Wayne State University and has edited this manuscript. AF co-created and oversees the systems and methods for data transmission and storage system storage and created the system for data checking and query prior to transfer to CDC and has edited this manuscript. NA assisted in site-specific setup for data collection, weekly data transmission, regular participation in network meetings, and edited this manuscript. NF assisted in site-specific setup for data collection, weekly data transmission, regular participation in network meetings, and edited this manuscript. MB, assisted in site-specific setup for data collection, weekly data transmission, regular participation in network meetings, and edited this manuscript. AM, assisted in site-specific setup for data collection, weekly data transmission, regular participation in network meetings, and edited this manuscript. JWG assisted in site-specific setup for data collection, weekly data transmission, regular participation in network meetings, and edited this manuscript. AB, assisted in CDC-specific setup for data collection, weekly data transmission, regular participation in network meetings, and edited this manuscript. AW, assisted in CDC-specific setup for data collection, weekly data transmission, regular participation in network meetings, and edited this manuscript. EP assisted in CDC-specific setup for data collection, weekly data transmission, regular participation in network meetings, and edited this manuscript. CAC co-conceived the work, assisted with funding, and provides technical advice for the conduct of the network and database content and edited the manuscript. All authors reviewed the manuscript.

Corresponding author

Correspondence to Jeffrey A. Kline.

Ethics declarations

Ethics approval and consent to participate

The protocol has been reviewed by all site institutional review boards and deemed non-human subjects research, exempted from review or approved with expedited review.

Institution

Named Institutional and/or Licensing Committee (From subcontract)

IRB review or approval

Corewell Health William Beaumont University Hospital

William Beaumont Hospital Research Institute

Beaumont Research Institute Institutional Review Board

Hennepin Healthcare

Hennepin Healthcare Research Institute

Hennepin Healthcare Human Research Protection Office

Intermountain Healthcare

IHC Health Services, Inc

Intermountain Healthcare Research Institional Review Board

Massachusetts General Hospital

Massachusetts General Hospital

Mass General Brigham IRB

Icahn School of Medicine at Mount Sinai

Icahn School of Medicine at Mt. Sinai

Program for the Protetion of Human Subject IRBs

JPS Health Network

Integrative Emergency Services

North Texas Regional IRB

Oregon Health & Science University

Oregon Health & Science University

OHSU Research Integrity Office

Penn State Health

The Pennsylvania State University

Office for Research Protection The Pennslyvania State Univ

University of Florida Health

University of Florida Board of Trustees

University of Florida Institutional Review Board

Centura Health

Centura Health Corporation

CommonSpirit Health Research Institute Institutional Review Board (CSHRI IRB)

The Ohio State University Wexner Medical Center

The Ohio State University

OSU IRB

Indiana University School of Medicine

Indiana University School of Medicine

Eskenazi Health Research Department

University of California, San Diego

The Regents of the University of California; University of California San Diego

UCSD HRPPP

The University of Chicago Department of Medicine

University of Chicago

UofC Biological Sciences Division UofC Medical Center

University of Iowa

University of Iowa

UIHC Data Governance Committee

University of Mississippi Medical Center

University of Mississippi Medical Center

U of Miss Medical Center Institutioinal Review Board

The University of Utah

University of Utah

The U of Utah Institutional Review Board

University of Wisconsin-Madison

Board of Regents of University of Wisconsin System

UofW Madison Minial Rish Research IRB

LCMC Health

University Medical Center Management Corporation d/b/a University Medical Center New Orleans

LSUHSC-NO IRB

The University of Texas, Southwestern

University of Texas Southwestern Medical Center

UT Southwestern Medical Center

Washinton University School of Medicine in St. Louis

The Washington University

Washington University in St. Louis Human Research Protection Office

West Virginia University

West Virginia University Research Corporation

WVU Office of Human Research Protections

University of California, San Francisco

Regents of the University of California, San Francisco

UCSF Human Research Protections Program

George Washington University

George Washington University

The George Washington University Office of Human Research

Consent for publication

Not applicable

Competing interests

Employees of the funding agency (Centers for Disease Control) contributed to the project design and writing of the manuscript. Alex Frost is employed by StudyMaker. All other authors declare that they have no competing interests.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Supplementary Information

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Kline, J.A., Reed, B., Frost, A. et al. Database derived from an electronic medical record-based surveillance network of US emergency department patients with acute respiratory illness. BMC Med Inform Decis Mak 23, 224 (2023). https://0-doi-org.brum.beds.ac.uk/10.1186/s12911-023-02310-4

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://0-doi-org.brum.beds.ac.uk/10.1186/s12911-023-02310-4

Keywords