Skip to main content

Eaton Julie

Description

Distribute is a national emergency department syndromic surveillance project developed by the International Society for Disease Surveillance for influenza-like-illness (ILI) that integrates data from existing state and local public health department surveillance systems. The Distribute project provides graphic comparisons of both ILI-related clinical visits across jurisdictions and a national picture of ILI. Unlike other surveillance systems, Distribute is designed to work solely with summarized (aggregated) data which cannot be traced back to the un-aggregated 'raw' data. This and the distributed, voluntary nature of the project creates some unique data quality issues, with considerable site to site variability. Together with the ISDS, the University of Washington has developed processes and tools to address these challenges, mirroring work done by others in the Distribute community.

Objective

To present exploratory tools and methods developed as part of the data quality monitoring of Distribute data, and discuss these tools and their applications with other participants.

Submitted by elamb on
Description

Distribute is a national emergency department syndromic surveillance project developed by the International Society for Disease Surveillance (ISDS) for influenza-like-illness (ILI) that integrates data from existing state and local public health department surveillance systems. The Distribute is a national emergency department syndromic surveillance project developed by the International Society for Disease Surveillance (ISDS) for influenza-like-illness (ILI) that integrates data from existing state and local public health department surveillance systems. The Distribute project provides graphic comparisons of both ILI-related clinical visits across jurisdictions and a national picture of ILI. Unlike other surveillance systems, Distribute is designed to work solely with summarized (aggregated) data which cannot be traced back to the un-aggregated 'raw' data. This and the distributed, voluntary nature of the project create some unique data quality issues, with considerable site to site variability. Together with the ISDS, the University of Washington has developed processes and tools to address these challenges, mirroring work done by others in the Distribute community.

Objective

The goal of this session will be to briefly present two methods for comparing aggregate data quality and invite continued discussion on data quality from other surveillance practitioners, and to present the range of data quality results across participating Distribute sites.

Referenced File
Submitted by elamb on
Description

The utility of specific sources of data for surveillance, and the quality of those data, are an ingoing issue in public health(1). Syndromic surveillance is typically conducted as a secondary use of data collected as part of routine clinical practice, and as such the data can be of high quality for the clinical use but of lower quality for the purpose of surveillance. A major data quality issue with surveillance data is that of timeliness. Data used in surveillance typically arrive as a periodic process, inherently creating a delay in the availability of the data for surveillance purposes. Surveillance data are often collected from multiple sources, each with their own processes and delays, creating a situation where the data available for surveillance are accrued piecemeal.

Objective

This abstract discusses the quality issues identified in using Distribute. From 2006 to 2012, the ISDS ran Distribute (2), a surveillance system for monitoring influenza like illness (ILI) and gastroenteritis (GI) ED visits on a nationwide basis. This system collected counts for ILI, GI and total ED visits, aggregated to the level of jurisdiction. The primary data quality issue faced with the Distribute system was that of timeliness due to accrual lag; variable delays in the receipt of surveillance data from sources by jurisdictions together with variable delays in the reporting of aggregate data from jurisdictions to Distribute resulted in data which accrued over time(3).

Submitted by knowledge_repo… on
Description

Data consisting of counts or indicators aggregated from multiple sources pose particular problems for data quality monitoring when the users of the aggregate data are blind to the individual sources. This arises when agencies wish to share data but for privacy or contractual reasons are only able to share data at an aggregate level. If the aggregators of the data are unable to guarantee the quality of either the sources of the data or the aggregation process then the quality of the aggregate data may be compromised. This situation arose in the Distribute surveillance system (1). Distribute was a national emergency department syndromic surveillance project developed by the International Society for Disease Surveillance for influenza-like-illness (ILI) that integrated data from existing state and local public health department surveillance systems, and operated from 2006 until mid 2012. Distribute was designed to work solely with aggregated data, with sites providing data aggregated from sources within their jurisdiction, and for which detailed information on the un-aggregated ‘raw’ data was unavailable. Previous work (2) on Distribute data quality identified several issues caused in part by the nature of the system: transient problems due to inconsistent uploads, problems associated with transient or long-term changes in the source make up of the reporting sites and lack of data timeliness due to individual site data accruing over time rather than in batch. Data timeliness was addressed using prediction intervals to assess the reliability of the partially accrued data (3). The types of data quality issues present in the Distribute data are likely to appear to some extent in any aggregate data surveillance system where direct control over the quality of the source data is not possible.

Objective

In this work we present methods for detecting both transient and long-term changes in the source data makeup.

 

Submitted by uysz on

This webinar will present a set of tools developed for visualizing data quality problems in aggregate surveillance data, in particular for data which accrues over a period of time. This work is based on a data quality analysis of aggregate data used for ILI surveillance within the Distribute system formerly operated by the ISDS. We will present a method developed as a result of this analysis to ‘nowcast’ complete data from incomplete, partially accruing data, as an example of how forecasting methods can be used to mitigate data quality problems.

Presenters