It is often of interest to estimate disease incidence. For this purpose, we typically rely on repeated measurements of participants’ disease status over time in a longitudinal framework, where disease status may be determined through lab tests. During the COVID-19 pandemic, accurate estimates of the incidence of infection with SARS-CoV-2 have been critical for public health policy decision-making.1–3 There are multiple challenges to achieving this goal that include a sub-population that is asymptomatic and a dynamic landscape of disease variants and diagnostic tests. Specifically, several SARS-CoV-2 test kits (rt-nucleic acid amplification, antigen, and serological tests) were approved under Emergency Use Authorization for clinical use without rigorous validation.4,5 The SARS-CoV-2 literature demonstrated how critical it was to adjust for the varying performance of these laboratory tests (sensitivity and specificity) to draw an accurate inference on prevalence and incidence.6,7 Indeed, for cross-sectional studies designed to estimate prevalence, the “standard correction” method based on Bayes’ rules6,8 was often applied to account for the laboratory test performance.9–13 Both frequentist8,9,14 and Bayesian approaches11–13,15 were proposed for this purpose. For example, Bajema et al.9 and Havers et al.10 proposed a two-stage nonparametric bootstrapping approach that resampled the false positive and negative cases to account for the uncertainty of the sensitivity and specificity estimates. Meireles et al.11 used a Bayesian approach with uniform prior distributions of sensitivity and specificity, whereas Sahlu and Whittaker12 considered an informative beta prior and Meyer et al.13 proposed more specific beta prior distributions that incorporated information by geographical regions. Under scenarios where knowledge of test performance was limited, Burstyn et al.16 proposed a Bayesian approach to inform the sensitivity and specificity in the population using publicly available time-series data.
Because incidence measures the instantaneous probability of being infected at a given time point, it is considered a more relevant measure for informing timely decisions on public health and healthcare resource allocation to arrest SARS-CoV-2 transmission, as it provides additional insight into the current “momentum” of the epidemic.17–20 In contrast to prevalence, incidence is often estimated with a longitudinal study design, which presents challenges in accounting for test performance.17 Given the longitudinal nature of the study required to estimate incidence, participants need to be repeatedly tested over time, yielding a high volume of tests performed. This means that even a small percentage of laboratory testing errors can result in a high absolute number of false positive or negative test results, eventually leading to a biased estimate of incidence, while the direction and magnitude of the bias would depend on the underlying true incidence and prevalence of the disease. For the purposes of our study, we defined incidence as the initial infection for a participant, where participants were censored at a positive result with no further follow-up tests scheduled. The implications of censoring as a function of the test status are that the test performance of one assay will affect whether the patient will be censored at subsequent visits, affecting the participant’s length of time considered at risk.
In the current literature, there are no guidelines or a “gold standard” approach to adjust for laboratory test errors when estimating incidence within a repeated measures longitudinal framework, which is a fairly common goal. For example, work by Becker and Britton (1999)17 and by Gan and Bain (1998)21 examined the estimation of the incidence risk under a similar longitudinal design. A key difference with the current article is our consideration of the testing error in identifying the true cases, quantified by imperfect sensitivity and specificity. Previous studies proposed methods to incorporate sensitivity and specificity by adjusting the number of positive or negative events.22,23 However, in those studies, the incidence was modeled as a single probability with only one follow-up measurement. Further, the methods proposed assumed consistent follow-up times and nondifferential time at risk and loss to follow-up across participants. Novel methods that challenged or relaxed these assumptions were proposed in the field of HIV. For example, McDougal et al.24 derived a correction factor for incidence adjustment—the probability of being infected divided by the probability of being in the window (P(T0)/P(W))—that was subsequently criticized because the correction factor took on the value of 1 under commonly occurring circumstances that did not make empirical sense.24,25 Hargrove et al.,26 therefore, proposed a modified version of a correction factor (ε)—the probability of being in the window period if infected at least twice the duration of the time interval (2μ) earlier.26 Similar to McDougal’s method, however, this approach could be misleading due to its underlying mathematical inconsistency that nonzero ε values could lead to anomalous results.25 The methods mentioned above all failed to adequately leverage the repeated nature of testing by assuming a uniform distribution of testing across all time points. Thus, there is a critical gap in the current literature for addressing unbiased estimation of incidence in a typical longitudinal framework. Flexible methods are needed that adjust for laboratory test performance in the common scenario where incidence is estimated from repeated measurements.
Parent Study: The TrackCOVID StudyOur work is motivated by a public health surveillance initiative to estimate the incidence and prevalence of SARS-CoV-2 infection and the associated risk factors in the San Francisco Bay Area (TrackCOVID study). The TrackCOVID study was conducted from July 2020 to March 2021, during the outbreak of the COVID-19 pandemic. The longitudinal study relied on a sampling framework to randomly select residents from six Bay Area counties using census tract data.27 We requested participants in the study to come into the clinic for a baseline visit and, subsequently, for monthly visits for up to 6 months. Evidence of infection was defined as a positive test using two types of assays measured repeatedly: (1) an rt-PCR test of a nasopharyngeal swab and (2) serologic testing of blood. In this way, we can capture infections that may have been missed by one of the assays, and particularly those infections we may miss in between monthly visits. Thus, our definition of infection relied on the performances of multiple assays. Importantly, to account for (1) the sampling framework, (2) the probability of selection from the household, and (3) nonresponse bias, we estimated and incorporated a combined weight variable for each participant when calculating quantities of interest (such as incidence).27
The TrackCOVID study was designated as a public health surveillance study and not human subjects research under 45 CFR 46.102(l) by the Stanford University School of Medicine Administrative Panel on Human Subjects in Medical Research and the University of California, San Francisco Institutional Review Board.
In this article, we proposed a new statistical framework to address the problem of estimating incidence while adjusting for laboratory test performance in a longitudinal repeated measures framework using a maximum likelihood estimation (MLE)-based approach. Our method can be extended to a variety of study design scenarios under relaxed assumptions. We presented the likelihood function using the motivating example, assessed the properties of the method with a simulation study, and applied our approach to the real-world data collected in our parent study.
METHODSTo better understand the underlying principles of our method, we illustrated ideas through the TrackCOVID study.
Possible Trajectories of Observed Longitudinal Test ResultIn the TrackCOVID study, each participant could have at most seven visits (including the baseline visit). Considering loss to follow-up and censoring after a given visit, we may observe a total of 14 possible test result trajectories as illustrated in Table 1. For example, scenario 4 reflects the situation where an individual has three negative tests followed by a positive test at visit 4, and scenario 12 represents a trajectory where there are four consecutive negative tests immediately following enrollment with loss to follow-up afterward. In practice, observed data could be represented by either an unweighted or weighted number of participants, whose test results fall into each of these possible trajectories or scenarios, denoted as nk,k=1,⋯,14. If there is any intermittent missing value, we consider the observed data to be those preceding this value.
TABLE 1. - All Possible Trajectories of Observed Longitudinal Test Result at the Participant Levela Baseline Follow-upa+ indicates a positive test result; − indicates a negative test result; and blank indicates loss to follow-up or censoring.
There are eight true infection status trajectories for the participants enrolled in our parent study (Table 2). Here, we assume that once a participant’s underlying true disease status is positive, their status remains positive until the last follow-up visit. Participant’s test status at each visit is a binary variable (positive or negative) that can be determined by either a single test result or multiple test results combined under prespecified rules. First, we define the following parameters:
TABLE 2. - All Possible Trajectories of True Longitudinal Infection Status at the Participant Levela Baseline Follow-upq1=π
True status 2 − + + + + + +q2=(1−π)p
True status 3 − − + + + + +q3=(1−π)(1−p)p
True status 4 − − − + + + +q4=(1−π)(1−p)2p
True status 5 − − − − + + +q5=(1−π)(1−p)3p
True status 6 − − − − − + +q6=(1−π)(1−p)4p
True status 7 − − − − − − +q7=(1−π)(1−p)5p
True status 8 − − − − − − −q8=(1−π)(1−p)6
a+ indicates a positive infection; − indicates no infection.
We can express the probabilities of the eight true infection status trajectories—denoted as qi(i=1, 2, 3,…, 8)—as a function of π and p (Table 2). Note that the incidence is assumed to be constant for simplicity. Later, we extend the method to allow nonconstant incidence risk.
Likelihood Function at the Participant LevelThe values of sensitivity (sen) and specificity (spe) of the laboratory test are considered to be given, and the observed data are . Given this, our interest lies in estimating (π,p). We account for test performance via the maximum likelihood estimation, which requires expressing the likelihood function. According to Tables 1 and 2, we expect to have a total of 14 × 8 = 112 combinations of the test results and the true infection status trajectories for the study cohort. We express the likelihood function for each combination as a function of π, p, and r (Tables 3 and 4).
TABLE 3. - Probability of Combination of Each Test Result and True Infection Status (True Status 1–4) True Status 1 True Status 2 True Status 3 True Status 4 Probability Test result scenario 1sen
(1−spe)
(1−spe)
(1−spe)
qj
Test result scenario 2(1−sen)sen
spesen
spe(1−spe)
spe(1−spe)
(1−r^)qj
Test result scenario 3(1−sen)2sen
spe(1−sen)sen
spe2sen
spe2(1−spe)
(1−r^)2qj
Test result scenario 4(1−sen)3sen
spe(1−sen)2sen
spe2(1−sen)sen
spe3sen
(1−r^)3qj
. Test result scenario 5(1−sen)4sen
spe(1−sen)3sen
spe2(1−sen)2sen
spe3(1−sen)sen
(1−r^)4qj
Test result scenario 6(1−sen)5sen
spe(1−sen)4sen
spe2(1−sen)3sen
spe3(1−sen)2sen
(1−r^)5qj
Test result scenario 7(1−sen)6sen
spe(1−sen)5sen
spe2(1−sen)4sen
spe3(1−sen)3sen
(1−r^)6qj
Test result scenario 8(1−sen)7
spe(1−sen)6
spe2(1−sen)5
spe3(1−sen)4
(1−r^)6qj
Test result scenario 9(1−sen)
spe
spe
spe
r^qj
Test result enario 10(1−sen)2
留言 (0)