Hostname: page-component-586b7cd67f-vdxz6 Total loading time: 0 Render date: 2024-11-27T22:02:23.191Z Has data issue: false hasContentIssue false

Bias correction methods for test-negative designs in the presence of misclassification

Published online by Cambridge University Press:  08 September 2020

A. Endo*
Affiliation:
Department of Infectious Disease Epidemiology, London School of Hygiene & Tropical Medicine, Keppel St., LondonWC1E 7HT, UK Centre for the Mathematical Modelling of Infectious Diseases, London School of Hygiene & Tropical Medicine, Keppel St., LondonWC1E 7HT, UK The Alan Turing Institue, Euston Rd., LondonNW1 2DB, UK
S. Funk
Affiliation:
Department of Infectious Disease Epidemiology, London School of Hygiene & Tropical Medicine, Keppel St., LondonWC1E 7HT, UK Centre for the Mathematical Modelling of Infectious Diseases, London School of Hygiene & Tropical Medicine, Keppel St., LondonWC1E 7HT, UK
A. J. Kucharski
Affiliation:
Department of Infectious Disease Epidemiology, London School of Hygiene & Tropical Medicine, Keppel St., LondonWC1E 7HT, UK Centre for the Mathematical Modelling of Infectious Diseases, London School of Hygiene & Tropical Medicine, Keppel St., LondonWC1E 7HT, UK
*
Author for correspondence: A. Endo, E-mail: [email protected]
Rights & Permissions [Opens in a new window]

Abstract

The test-negative design (TND) has become a standard approach for vaccine effectiveness (VE) studies. However, previous studies suggested that it may be more vulnerable than other designs to misclassification of disease outcome caused by imperfect diagnostic tests. This could be a particular limitation in VE studies where simple tests (e.g. rapid influenza diagnostic tests) are used for logistical convenience. To address this issue, we derived a mathematical representation of the TND with imperfect tests, then developed a bias correction framework for possible misclassification. TND studies usually include multiple covariates other than vaccine history to adjust for potential confounders; our methods can also address multivariate analyses and be easily coupled with existing estimation tools. We validated the performance of these methods using simulations of common scenarios for vaccine efficacy and were able to obtain unbiased estimates in a variety of parameter settings.

Type
Original Paper
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted re-use, distribution, and reproduction in any medium, provided the original work is properly cited.
Copyright
Copyright © The Author(s), 2020. Published by Cambridge University Press

Introduction

Vaccine effectiveness (VE) is typically estimated as the vaccine-induced risk reduction of the target disease (TD) and has been traditionally studied using cohort or case–control designs. However, the test-negative design (TND) is becoming a popular alternative design for VE studies [Reference De Serres1, Reference Fukushima and Hirota2]. This is a modified version of the case–control study with an alternative definition of the control group; traditional case–control studies usually define controls as non-disease individuals in the study population, while TND studies use individuals with similar symptoms to the TD but presenting negative test results (i.e. patients of non-target diseases; ND). The TND can therefore minimise ascertainment bias by including only medically-attended patients in both case and control groups. Many TND studies have focused on influenza vaccination, but recent studies have also considered other diseases including pneumococcal disease [Reference Suzuki3, Reference Prato4] and rotavirus disease [Reference Araki5Reference Muhsen7].

Despite its increasing popularity, a TND can be more vulnerable than other study designs to misclassification of disease outcome. Multiple studies have shown that VE is underestimated when the diagnostic tests used in the study are imperfect (i.e. have a sensitivity and/or a specificity less than 100%) [Reference Orenstein8Reference De Smedt10]. This can be a particular issue when simple tests (e.g. rapid diagnostic tests) are used for logistical convenience, as simple tests tend to have lower diagnostic performance than more advanced tests (e.g. polymerase chain reaction; PCR). Previous studies evaluated the expected degree of bias and concluded that specificity had a more important effect on bias than sensitivity [Reference Orenstein8Reference Haber11]. These findings appear to support the use of rapid tests, despite limited sensitivity, because the specificity of these tests is typically high [Reference Fukushima and Hirota2]. However, theoretical studies to date have been based on a limited range of assumptions about efficacy and pathogen epidemiology; it is therefore unclear whether such conclusions hold for all plausible combinations of scenarios.

If a study is expected to generate a non-negligible bias in estimation, such bias needs to be assessed and – if possible – corrected before the estimate is reported. Greenland [Reference Greenland12] proposed a bias correction method for cohort studies where the sensitivity and specificity of the test are known (or at least assumed). However, it has been pointed out that bias correction in case–control studies is in general difficult because of differential recruitment, whereby the probability of recruiting (test-positive) cases and (negative) controls may be different [Reference Greenland12, Reference Kleinbaum, Kupper and Morgenstern13]. Although TND studies are often considered to be special cases of case–control studies, they are free from the issue of differential recruitment because the recruitment and classification are mutually-independent [Reference Jackson and Nelson14]. This means that, while Greenland's method does not apply to TND as-is, another type of bias correction may still be possible. For example, De Smedt et al. have characterised the misclassification bias in VE in the TND in a simulation study [Reference De Smedt10]. One limitation of their formulation was it relies on the unobserved âœtrueâ disease risk being known, where in reality this is not usually measurable in field studies. As a result, bias correction methods for TND studies that are directly applicable to field data have not yet been proposed. Moreover, previous analysis of misclassification bias has not considered the impact of multivariate analysis, where potential confounders (e.g. age and sex) are also included in the model used to estimate VE.

To address these issues, we develop a bias correction method for the test-negative VE studies that uses only data commonly available in field studies. We also apply these methods to multivariate analyses. As our approach uses the so-called multiple overimputation (MO) framework (generalisation of multiple imputation) [Reference Blackwell, Honaker and King15], it can easily be coupled with a wide range of estimation tools without modifying their inside algorithms. Finally, we evaluate the performance of our methods by simulations of plausible epidemiological scenarios.

Methods and results

Characterising bias in TND studies

First, we consider the case where only vaccination history is included as a risk factor of acquiring the TD (i.e. the univariate setting). Following the approach of Haber et al. [Reference Haber16], we consider four steps in the case reporting process: vaccination, onset of symptoms, seeking of medical care and diagnosis. For simplicity, let us assume that occurrence of TD and ND are mutually independent, where their prevalences in the unvaccinated population are represented as r 1 and r 0, respectivelyFootnote 1. Let n V and n U be the vaccinated and unvaccinated population size. The target variable in VE studies is γ, the relative risk of TD in the vaccinated population relative to the unvaccinated (i.e. VE = 1 − γ). Vaccinated and unvaccinated population can have different likelihoods of seeking medical treatment given disease. We denote by m V and m U the probability of medical attendance given ND in vaccinated and unvaccinated population, respectively.

As our focus in the present study is the bias in VE estimation caused by imperfect tests, we made two key assumptions following Haber et al. [Reference Haber16]. One assumption is that vaccines have no effect on the risk of ND. This enables ND patients to be eligible for a control group and is a key assumption in TND studies. The other assumption is that the probability of medical attendance in vaccinated and unvaccinated population given infection is constant regardless of the disease (TD or ND). The probability of medical attendance given TD may be different from ND (m V and m U), potentially due to difference in severity; we assume that these probabilities are obtained by multiplying a constant factor μ (i.e. μm V and μm U)Footnote 2. These assumptions may not always hold and TND can be biased in such cases. However, we assume that they do in the following analysis to keep our focus on misclassification bias; namely, the study was assumed to be able to provide an unbiased VE estimate if tests are perfect.

Following the above notations, we can classify the vaccinated and unvaccinated population into multiple categories shown in Table 1. We can characterise different VE study designs (cohort, case–control and test-negative) by the categories in Table 1 from which each design tries to sample: the cohort design samples from populations n V and n U and follows them up to see what proportions fall into x V and x V; the case–control design samples from medically-attended cases (x V + x U) and non-diseased controls (n V − x V + n U − x U) and calculate the odds ratio to approximate the relative risk (however, the actual studies can mismeasure these variables when misclassificaiton is present).

Table 1. Population classified into different categories of interest in VE studies

In TND studies, medically-attended patients (x V + x U + y V + y U) are sampled and classified into four categories based on the test result and vaccine history. Let q be the proportion sampled relative to the population. Denoting the observed case counts with misclassification by X and Y, the process of data collection in TND can be represented by the following matrix expression:

(1)$$\left[{\matrix{ {X_V} \hfill & {X_U} \hfill \cr {Y_V} \hfill & {Y_U} \hfill \cr } } \right] = q\left[{\matrix{ \alpha \hfill & {1-\beta } \hfill \cr {1-\alpha } \hfill & \beta \hfill \cr } } \right]\left[{\matrix{ {x_V} \hfill & {x_U} \hfill \cr {y_V} \hfill & {y_U} \hfill \cr } } \right]\comma \;$$

where α and β are the sensitivity and specificity of the test, respectively. Matrix

$$C = \left[{\matrix{ \alpha \hfill & {1-\beta } \hfill \cr {1-\alpha } \hfill & \beta \hfill \cr } } \right]$$

describes the conversion from the true disease state to the observed result. We hereafter refer to C as the classification matrix. The determinant c = |C| = α + β − 1 is the Youden index of the test and satisfies 0 < c ≤ 1 (if c < 0, the test is not predictive and the definitions of positive/negative should be swapped). Youden index indicates the level of information retained in the potentially misclassified test results. Youden index of 0 indicates that the information is completely lost and the test is no better than random guesses.

We define bias in the VE estimate to be the absolute difference between the (raw) estimate, derived from the misclassified observation, and the true value. Let δ = ((r 1μ)/r 0) be the odds of the (medically-attended) TD in the unvaccinated population. Then the expected bias B is given as a function of four independent parameters, α, β, γ and δ:

(2)$$\eqalign{B\lpar \alpha \comma \;\beta \comma \;\gamma \comma \;\delta \rpar =\, &{\rm V}{\rm E}_{{\rm raw}}-{\rm V}{\rm E}_{{\rm true}} \cr =\, &\lpar 1-\gamma _{{\rm raw}}\rpar -\lpar 1-\gamma _{{\rm true}}\rpar \cr =\, &\gamma -\displaystyle{{[ {\alpha \gamma \delta + \lpar 1-\beta \rpar } ] [ {\lpar 1-\alpha \rpar \delta + \beta } ] } \over {[ {\lpar 1-\alpha \rpar \gamma \delta + \beta } ] [ {\alpha \delta + \lpar 1-\beta \rpar } ] }}.} $$

This suggests that the influence of sensitivity/specificity on the degree of bias varies depending on the case ratio δ/(1 + δ), i.e. the ratio between the incidence of medical attendance for TD and ND in the unvaccinated study populationFootnote 3 (Fig. 1). The degree of bias also depends on γ but is independent of m V and m U. The degree of bias is largely determined by the test specificity when the case ratio is small, but the influence of sensitivity and specificity is almost equivalent to a case ratio of 0.6. It is notable that high specificity does not always assure that the bias is negligible. This may be true if specificity is strictly 100% and the case ratio is low to moderate, but a slight decline to 97% can cause a bias up to 10–15 percentage points. The effect of sensitivity is also non-negligible when the case ratio is high.

Fig. 1. Bias in VE estimates caused by misclassification for different combinations of parameter values. (a)–(c) Estimated VE plotted against sensitivity. (a) True case ratio (the true ratio between TD and ND cases included in the study) = 0.2 (b) 0.4 (c) 0.6. Each two sets of lines respectively correspond to different true VEs (80% and 40%, denoted by the dotted lines). (d)–(f) Estimated VE plotted against specificity. (d) True case ratio = 0.2 (e) 0.4 (f) 0.6.

When the expected bias is plotted against the case ratio with various combinations of test performance, we find that VE estimates can be substantially biased for certain case ratios (especially when the ratio is far from 1:1), even with reasonably high sensitivity and specificity (Fig. 2b). In TND studies, researchers have no control over the case ratio because the study design requires that all tested individuals be included in the study. We found that the proportion of TD-positive patients in previous TND studies (retrieved from three systematic reviews [Reference Leung21Reference Darvishian23]) varied considerably, ranging from 10% to 70% (Fig. 2a)Footnote 4. Because of this large variation in the case ratio, it would be difficult to predict the degree of bias before data collection. Post-hoc assessment and correction therefore need to be considered. See the Supplementary Document for further analysis of the degree of bias.

Fig. 2. Biased VE estimates with varying case ratio and the observed proportion of positive patients. (a) The proportion of test-positive patients in TND studies from systematic reviews. The proportions were retrieved from three systematic reviews [Reference Leung21Reference Darvishian23]. (b) Estimated VE plotted against case ratio. Two sets of lines respectively correspond to different true VEs (80% and 40%, denoted by the dotted lines). The histogram in Panel (a) is overlaid on the x-axis.

Bias correction in univariate analysis

Corrected odds ratio

Although TND has sometimes been referred to as a special case of the case–control design, there is a distinct feature in the sampling procedure of TND. It has been pointed out that the adjustment methods for misclassification bias developed for cohort studies do not apply to case–control studies because the sampling ratio in case–control studies (differential recruitment) varies between case and control groups [Reference Greenland12, Reference Kleinbaum, Kupper and Morgenstern13]. However, as we have shown in Section ‘Characterising bias in TND studies’, both cases (TD patients) and controls (ND patients) are sampled at the same ratio (q) in TND studies. This suggests that the existing bias correction formulas developed for a hypothetical setting [Reference Kleinbaum, Kupper and Morgenstern13] where the whole population is evenly sampled (which is unrealistic in traditional studies) may be applicable to TND studies.

By left-multiplying Equation (1) with the inverted matrix C −1, we can obtain the corrected odds ratio γ* as

(3)$$\gamma ^\ast\! = \,\displaystyle{{X_V-\lpar \lpar 1-\beta \rpar /\beta \rpar Y_V} \over {Y_V-\lpar \lpar 1-\alpha \rpar /\alpha \rpar X_V}}\cdot \displaystyle{{Y_U-\lpar \lpar 1-\alpha \rpar /\alpha \rpar X_U} \over {X_U-\lpar \lpar 1-\beta \rpar /\beta \rpar Y_U}}\comma \;$$

which adjusts for misclassification to give an asymptotically-unbiased estimate of γ. This result can also be derived by maximising the likelihood accounting for misclassification in the observed TND data (see the Supplementary Document).

All four components of (3) (two numerators and two denominators) are usually expected to be non-negative with moderate VE (less than 100%) because these components are considered to be proportional to reconstructed true case counts. However, in some (relatively rare) cases, one or more components may become negative due to random fluctuations in observation. Theoretically, negative values are not permitted as true case counts, and thus such negative quantities would need to be truncated to 0. As a result, the corrected odds ratio can be either 0 or infinity. It is unrealistic in clinical settings that vaccines have absolute 100% or −100% effectiveness. Uncertainty around such MLEs should be carefully considered; increasing sample size or redesigning the study might be recommended where possible. Alternatively, the Bayesian framework may be used to yield an interval estimate with the likelihood shown in section ‘Direct likelihood method for the logistic regression model’ adapted for a univariate model.

The confidence interval for VE can be obtained by assuming log-normality of the odds ratio γ, i.e.

$$\gamma = \gamma ^\ast \exp \lpar { \pm 1.96\sigma^\ast } \rpar \comma \;$$

where σ is the shape parameter of the log-normal distribution and is empirically given as

(4)$$\eqalign{\sigma ^\ast{} = &\,SD\lpar {\log \lpar {\gamma^\ast } \rpar } \rpar \cr = &\,c\sqrt {\matrix{{\displaystyle{{X_VY_V\lpar {X_V + Y_V} \rpar } \over {{[ {\alpha Y_V-\lpar {1-\alpha } \rpar X_V} ] }^2{[ {\beta X_V-\lpar {1-\beta } \rpar Y_V} ] }^2}}+}\cr{ \displaystyle{{X_UY_U\lpar {X_U + Y_U} \rpar } \over {{[ {\alpha Y_U-\lpar {1-\alpha } \rpar X_U} ] }^2{[ {\beta X_U-\lpar {1-\beta } \rpar Y_U} ] }^2}}}}} .} $$

See the Supplementary Document for details of the MLE and confidence intervals.

Simulation

To assess the performance of the corrected odds ratio given in Equation (3) and uncertainty around it, we used simulation studies. TND study datasets were drawn from Poisson distributions (see the Supplementary Document for model settings and the likelihood function) as it is a reasonable assumption when medically-attended cases are recruited over the study period. We parameterised the mean incidence in the dataset by the ‘baseline medical attendance’ λ V = qm V(r 1μ + r 0)n V and λ U = qm U(r 1μ + r 0)n U, so that λ V and λ U correspond to the mean number of vaccinated/unvaccinated patients when vaccine has no effect (i.e. γ = 1 and VE = 0). The mean total sample size (given as ((1 + γδ)/(1 + δ))λ V + λ U) was set to be 3000. Parameter values were chosen according to a range of scenarios shown in Table 2, and the true VE=1−γ was compared with the estimates obtained from the simulated data. For each scenario, simulation was repeated 500 times to yield the distribution of estimates. Reproducible codes (including those for simulations in later sections) are reposited on GitHub (https://github.com/akira-endo/TND-biascorrection/).

Table 2. Simulation settings

We found that the uncorrected estimates, directly obtained from the raw case counts that were potentially misclassified, exhibited substantial underestimation of VE for most parameter values (Fig. 3). On the other hand, our bias correction method was able to yield unbiased estimates in every setting, whose median almost correspond to the true VE. Although the corrected and uncorrected distributions were similar (with a difference in median ~5%) when VE is relatively low (40%) and the test has sufficiently high sensitivity and specificity (95% and 97%, respectively), they became distinguishable with a higher VE (80%). With lower test performances, the bias in the VE estimates can be up to 10–20%, which may be beyond the level of acceptance in VE studies.

Fig. 3. Bias correction for simulated data in the univariate setting. The distributions of bias-corrected VE estimates (boxplots in blue) are compared with those of raw VE estimates without correction (red). Five hundred independent datasets were randomly generated for each set of parameter values, and the corrected and uncorrected VE estimates are compared with the true value (black solid line). See Table 2 for parameter settings in each scenario.

Bias correction of VEs reported in previous studies

We have seen that the degree of bias for uncorrected VE estimates depends on parameter values. To explore the possible degree of bias in existing VE studies, we extracted the reported crude VEs (i.e. VEs unadjusted for potential confounders) from two systematic reviews [Reference Leung21, Reference Darvishian23]Footnote 5 and applied our bias correction method assuming different levels of test sensitivity and specificity. The case counts for each study summarised in the reviews were considered eligible for the analysis if the total sample size exceeded 200. Varying the assumed sensitivity and specificity, we investigated the possible discrepancy between the reported VE (or crude VE derived from the case counts if unreported in the reviews) and bias-corrected VE. We did not consider correcting adjusted VEs because it requires access to the original datasets.

Figure 4 displays the discrepancy between the reported VE and bias-corrected VE corresponding to a range of assumptions on the test performance. Many of the extracted studies employed PCR for the diagnostic test, which is expected to have a high performance. However, the true performance of PCR cannot be definitively measured as there is currently no other gold-standard test available. Figure 4 suggests that even a slight decline in the test performance can introduce a non-negligible bias in some parameter settings. Our bias correction methods may therefore also be useful in TND studies using PCR, which would enable a sensitivity analysis accounting for potential misdiagnosis by PCR tests. In this light, it is useful that the corrected odds ratio

$$\gamma ^\ast{ = } \displaystyle{{X_V-\lpar \lpar 1-\beta \rpar /\beta \rpar Y_V} \over {Y_V-\lpar \lpar 1-\alpha \rpar /\alpha \rpar X_V}}\cdot \displaystyle{{Y_U-\lpar \lpar 1-\alpha \rpar /\alpha \rpar X_U} \over {X_U-\lpar \lpar 1-\beta \rpar /\beta \rpar Y_U}}$$

is a monotonic function of both α and β (given that all the four components are positive). The possible range of VE in a sensitivity analysis is obtained by supplying γ* with the assumed upper/lower limits of sensitivity and specificity.

Fig. 4. Bias correction method applied to published VE estimates assuming various test sensitivity and specificity. Case count data were extracted from two systematic reviews [Reference Leung21, Reference Darvishian23]. Each connected set of dots show how (crude) VE estimates reported in the review varies when imperfect sensitivity and specificity are assumed. Black dots on the grey diagonal line denote the original VEs reported in the reviews. This should correspond to the true value if sensitivity = specificity = 1. Coloured dots show the bias-corrected VE considering potential misclassification.

Bias correction in multivariate analysis

Theoretical framework

TND studies often employ a multivariate regression framework to address potential confounding variables such as age. The most widespread approach is to use generalised linear models (e.g. logistic regression) and include vaccination history as well as other confounding variables as covariates. The estimated linear coefficient for vaccination history can then be converted VE (in the logistic regression model, the linear coefficient for vaccination history corresponds to log (1 − VE)). In this situation, the likelihood function now reflects a regression model and thus the bias-corrected estimate in the univariate analysis (Equation (3)) is no longer applicable. We therefore need to develop a separate multivariate TND study framework to correct for bias in multivariate analysis.

Suppose that covariates ξ = (ξ 1, ξ 2, …, ξ n) are included in the model, and that ξ 1 corresponds to vaccination history (1: vaccinated, 0: unvaccinated). These covariates ξ i, as well as outcome variable Z i (i.e. test results) are available for each individual i included in the study. In TND studies, it is often convenient to model the binomial probability for the true outcome p 1(ξ i), i.e. the conditional probability that the true outcome is TD as opposed to ND given an individual has covariates ξ i. Let us use parameter set θ to model the binomial probabilities p 1 (and p 0 = 1 − p 1). Using the binomial probability $\pi _{Z_i}$ for observed (potentially misclassified) outcome Z i, we can obtain the MLE for θ by maximising

(5)$$\eqalign{& {\cal L}\left( {\theta ; D} \right) = \mathop \prod \limits_S^{i = 1} \pi _{Z_i}\left( {\xi _i;\theta } \right) \cr & \,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\, = \mathop \prod \limits_{i\in \left\{ + \right\}} \left[ {\alpha p_1\left( {\xi _i;\theta } \right) + \left( {1-\beta } \right)p_0\left( {\xi _i;\theta } \right)} \right] \cr & \,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\, \times \mathop \prod \limits_{i\in \left\{ - \right\}} \left[ {\left( {1-\alpha } \right)p_1\left( {\xi _i;\theta } \right) + \beta p_0\left( {\xi _i;\theta } \right)} \right].} $$

With the estimate θ*, the VE estimate for an individual with covariates ξ 2:n = (ξ 2, ξ 3, …, ξ n) is given as (1−odds ratio):

(6)$${\rm VE}\lpar {\xi^{2\colon n}} \rpar = 1-\displaystyle{{\,p_1\lpar {\xi^1 = 1\comma \;\xi^{2\colon n}\semicolon \;\theta^\ast } \rpar } \over {\,p_0\lpar {\xi^1 = 1\comma \;\xi^{2\colon n}\semicolon \;\theta^\ast } \rpar }}{\rm /}\displaystyle{{\,p_1\lpar {\xi^1 = 0\comma \;\xi^{2\colon n}\semicolon \;\theta^\ast } \rpar } \over {\,p_0\lpar {\xi^1 = 0\comma \;\xi^{2\colon n}\semicolon \;\theta^\ast } \rpar }}.$$

See the Supplementary Document for further details.

Direct likelihood method for the logistic regression model

The logistic regression model is well-suited for modelling binomial probabilities p 1 and p 0. The log-odds ($\log \left({{{p_1} \over {p_0}}} \right)$) is characterised by a linear predictor as:

(7)$$\log \left({\displaystyle{{\,p_1\lpar {\xi \semicolon \;\theta } \rpar } \over {\,p_0\lpar {\xi \semicolon \;\theta } \rpar }}} \right) = \theta _0 + \theta _1\xi ^1 + \cdots + \theta _n\xi ^n.$$

In the logistic regression model where covariate ξ 1 indicates vaccination history, the corresponding coefficient θ 1 gives the VE estimate: VE = 1 − exp (θ 1). Due to the assumed linearity, the estimated VE value is common across individuals regardless of covariates ξ 2:n.

We can employ the direct likelihood method by combining Equations (5) and (7). The usual logistic regression optimises θ by assuming that the test results follow Bernoulli distributions Z i ~ Bernoulli(p 1(ξ i;θ)) (Z i = 1 for positive test results and 0 for negative). To correct the misclassification bias, we instead need to use the modified probabilities to construct the likelihood accounting for diagnostic error, i.e.

(8)$$Z_i\sim {\rm Bernoulli}\lpar {{\pi }_ + \lpar {\xi_i\semicolon \;\theta } \rpar } \rpar = {\rm Bernoulli}\lpar {\alpha p_1\lpar {\xi_i\semicolon \;\theta } \rpar + \lpar {1-\beta } \rpar p_0\lpar {\xi_i\semicolon \;\theta } \rpar } \rpar.$$

Parameter θ is estimated by directly maximising the probability of observing {Z i} based on Equation (8)

Note that as long as the binomial probability is the modelling target, other type of models (e.g. machine learning classifiers) could also be employed under a similar framework.

Multiple overimputation combined with existing tools

The direct likelihood method presented in the previous section is the most rigorous MLE approach and would therefore be preferable whenever possible. However, it is often technically-demanding to implement such approaches as it involves re-defining the likelihood; if we wanted to use existing tools for logistic regression (or other models), for example, we would need to modify the internal algorithm of such tools. This is in particular complicated in tools for generalised linear models including logistic regression, whose standard algorithm is the iteratively reweighted least squares method [Reference Sidney Burrus24], which does not involve the explicit likelihood. To ensure that our correction methods can be employed without losing access to substantial existing software resources, we also propose another method, which employs a MO framework [Reference Blackwell, Honaker and King15] to account for misclassification. Whereas multiple imputation only considers missing values, MO is proposed as a more general concept which includes overwriting mismeasured values in the dataset by imputation. In our multivariate bias correction method, test results in the dataset (which are potentially misclassified) are randomly overimputed.

Let M be an existing estimation software tool whose likelihood specification cannot be reprogrammed. Given data d = {z i, ξ i}i=1,2,…S, where z i denotes the true disease state (z = 1 for TD and z = 0 for ND), M would be expected to return at least the following two elements: the point estimate of VE (ɛ d) and the predicted binomial probability $\hat{p}_1\lpar {\xi_i} \rpar $ for each individual i. From the original observed dataset D, J copies of imputed datasets $\lcub {{\tilde{D}}^j} \rcub = \lcub {{\tilde{D}}^1\comma \;{\tilde{D}}^2\comma \;\ldots \comma \;{\tilde{D}}^J} \rcub $ are generated by the following procedure.

  1. (1) For i = 1, 2, …, S, impute disease state $\tilde{z}_i^j $ based on the test result Z i. Each $\tilde{z}_i^j $ is sampled from a Bernoulli distribution conditional to Z i:

    (9)$$\tilde{z}_i^j \sim \left\{{\matrix{ {{\rm Bernoulli}\lpar {1-{\tilde{\varphi }}_{i + }} \rpar } \hfill & {\lpar {Z_i = 1} \rpar } \hfill \cr {{\rm Bernoulli}\lpar {{\tilde{\varphi }}_{i-}} \rpar } \hfill & {\lpar {Z_i = 0} \rpar } \hfill \cr } } \right..$$
  2. (2) $\tilde{\varphi }_{i + }$ and $\tilde{\varphi }_{i-}$ are estimated probabilities that the test result for individual i is incorrect (i.e. z i ≠ Z i) given Z i. The sampling procedure (9) is therefore interpreted as the test result Z i being ‘flipped’ at a probability $\tilde{\varphi }_{i + }$ or $\tilde{\varphi }_{i-}$. Later we will discuss possible procedures to obtain these probabilities.

  3. (3) Apply M to $\tilde{D}^j = \lcub {\tilde{z}_i^j \comma \;\xi_i} \rcub $ to yield a point estimate of VE (ɛ j).

  4. (4) Repeat (1) and (2) for j = 1, 2, …, J to yield MO estimates {ɛ j}j=1,…J.

Once MO estimates {ɛ j} are obtained, the pooled estimate and confidence intervals of VE are obtained by appropriate summary statistics, e.g. Rubin's rules [Reference Rubin25]. As long as the estimated ‘flipping’ probabilities $\tilde{\varphi }_{i \pm } = \lpar {{\tilde{\varphi }}_{i + }\comma \;{\tilde{\varphi }}_{i-}} \rpar $ are well chosen, this MO procedure should provide an unbiased estimate of VE with a sufficiently large number of iterations J.

As a method to estimate the flipping probability $\tilde{\varphi }_{i \pm }$, here we propose the parametric bootstrapping described as follows. Given the observed test result Z i, $\tilde{\varphi }_{i \pm }$ is given as the Bayesian probability

(10)$$\eqalign{&{\cal P}\lpar z_i = Z_i\vert Z_i\rpar \cr &\!\!= \!\left\{\!\!\!\!{\matrix{ {\displaystyle{{\alpha {\cal P}\lpar z_i = 1\rpar } \over {\alpha {\cal P}\lpar z_i = 1\rpar + \lpar 1-\beta \rpar {\cal P}\lpar z_i = 0\rpar }} = \displaystyle{{\alpha p_1\lpar \xi_i\rpar } \over {\alpha p_1\lpar \xi_i\rpar + \lpar 1-\beta \rpar p_0\lpar \xi_i\rpar }}} & \!\!\!\!{\lpar Z_i = 1\rpar } \cr {\displaystyle{{\beta {\cal P}\lpar z_i = 0\rpar } \over {\lpar 1-a\rpar {\cal P}\lpar z_i = 1\rpar + \beta {\cal P}\lpar z_i = 0\rpar }} = \displaystyle{{\beta p_0\lpar \xi_i\rpar } \over {\lpar 1-\alpha \rpar p_1\lpar \xi_i\rpar + \beta p_0\lpar \xi_i\rpar }}} & \!\!\!\!\!{\lpar Z_i = 0\rpar } \cr } } \right.$$

Although the true binomial probabilities p 0(ξ i), p 1(ξ i) are unknown, their estimators are derived with the inverted classification matrix in the same manner as Equation (3). By substituting

$$\left[{\matrix{ {\,p_1\lpar {\xi_i} \rpar } \cr {\,p_0\lpar {\xi_i} \rpar } \cr } } \right]$$

with

$$C^{{-}1}\left[{\matrix{ {\pi_ + \lpar {\xi_i} \rpar } \cr {\pi_-\lpar {\xi_i} \rpar } \cr } } \right]\comma \;$$

we get

(11)$$\matrix{ {{\tilde{\varphi }}_{i + } = 1-{\cal P}\lpar {z_i = 1\vert Z_i = 1} \rpar = \displaystyle{{1-\beta } \over {\alpha + \beta -1}}} & {\left[{\alpha \cdot \displaystyle{{\pi_-\lpar {\xi_i} \rpar } \over {\pi_ + \lpar {\xi_i} \rpar }}-\lpar {1-\alpha } \rpar } \right]} \cr {{\tilde{\varphi }}_{i-} = 1-{\cal P}\lpar {z_i = 0\vert Z_i = 0} \rpar = \displaystyle{{1-\alpha } \over {\alpha + \beta -1}}} & {\left[{\beta \cdot \displaystyle{{\pi_ + \lpar {\xi_i} \rpar } \over {\pi_-\lpar {\xi_i} \rpar }}-\lpar {1-\beta } \rpar } \right]} \cr } $$

These probabilities can be computed provided the odds of the test results π + (ξ i)/π (ξ i). We approximate this odds by applying estimation tool M to the original data D; i.e. the predicted binomial probability $\hat{p}_1\lpar {\xi_i} \rpar $ obtained from D is used as a proxy of π + (ξ i). Generally it is not assured that true and observed probabilities p 1(ξ i) and π + (ξ i) have the same mechanistic structure captured by M; however, when our concern is limited to the use of model-predicted probabilities to smooth the data D, we may expect for M to provide a sufficiently good approximation. The above framework can be regarded as a variant of parametric bootstrapping methods as MO datasets are generated from data D assuming a parametric model M. The whole bias correction procedure is presented in pseudocode (Fig. 5); sample R code is also available on GitHub (https://github.com/akira-endo/TND-biascorrection/).

Fig. 5. Multiple imputation with parametric bootstrapping.

EM algorithm

Another possible approach to addressing misclassification is the use of the EM algorithm, which has been proposed for case–control studies in a previous study (where differential recruitment was not considered) [Reference Magder and Hughes26]. Because of its methodological similarity, the algorithm can also be applied to the TND. The original EM algorithm presented in [Reference Magder and Hughes26] would produce, if properly implemented, the result equivalent to the direct likelihood approach. However, the original EM algorithm requires that the model can handle non-integer sample weights (which may not always be assured). Moreover, computing confidence intervals in EM algorithm can be complicated. We therefore recommend parametric bootstrapping as the first choice of bias correction method when the direct likelihood approach is inconvenient.

Simulation of bias correction with parametric bootstrapping

To assess the performance of this method, we used the same simulation framework as in the univariate analysis (Table 2). In addition to vaccination history (denoted by ξ 1), we consider one categorical and one continuous covariate. Let us assume that ξ 2 represents the age group (categorical; 1: child, 0: adult) and ξ 3 the pre-infection antibody titre against TD (continuous). Suppose that the population ratio between children and adults is 1:2, and that ξ 3 is scaled so that it is standard normally distributed in the population. For simplicity, we assumed that all the covariates are mutually independent with regard to the distribution and effects (i.e. no association between covariates and no interaction effects). The relative risk of children was set to be 2 and 1.5 for TD and ND, respectively, and a unit increase in the antibody titre was assumed to halve the risk of TD (and not to affect the risk of ND). The mean total sample size λ was set to be 3000, and 500 sets of simulation data were generated for each scenario. VE estimates were corrected by the parametric bootstrapping approach (the number of iterations J = 100) and were compared with the raw (uncorrected) VE estimates.

Figure 6 shows the distributions of estimates with and without bias correction in the multivariate setting. Our bias correction (parametric bootstrapping) provided unbiased estimates for all the scenarios considered. Overall, biases in the uncorrected estimates were larger than those in the univariate setting. In some scenarios, the standard error of the bias-corrected estimates was extremely wide. This was primarily because of the uncertainty already introduced before misclassification rather than the failure of bias correction (as can be seen in the Supplementary Fig. S3). Larger sample size is required to yield accurate estimates in those settings, as the information loss due to misclassification will be added on top of the inherent uncertainty in the true data.

Fig. 6. Bias correction for simulated data in the multivariate setting. The distributions of bias-corrected (blue) and uncorrected (red) VE estimates from 500 simulations are compared. Dotted lines denote median and black solid lines denote the true VE. The parametric bootstrapping bias correction method was used for bias correction.

The number of confounding variables

We investigated how the bias in uncorrected VE estimates can be affected by the number of confounding variables. In addition to the vaccine history ξ 1, we added a set of categorical/continuous confounding variables to the model and assessed the degree of bias caused by misclassification. The characteristics of the variables were inherited from those in section ‘Simulation of bias correction with parametric bootstrapping’: categorical variable ‘age’ and continuous variable ‘pre-infection antibody titre’. That is, individuals were assigned multiple covariates (e.g. ‘categorical variable A’, ‘categorical variable B’, …, ‘continuous variable A’, ‘continuous variable B’, …) whose distribution and effect were identical to ‘age’ (for categorical variables) and ‘antibody titre’ (for continuous variables) in section ‘Simulation of bias correction with parametric bootstrapping’. No interaction between covariates was assumed. The covariate set in section ‘Simulation of bias correction with parametric bootstrapping’ being baseline (the number of covariates: (vaccine, categorical, continuous) = (1, 1, 1)), we employed two more scenarios with a larger number of covariates: (1, 3, 3) and (1, 5, 5).

The simulation results are presented in Figure 7. Overall, additional confounding variables led to more severe bias in the uncorrected VE estimates towards underestimation. These results further highlight the importance of bias correction when heterogeneous disease risks are expected; VE estimates adjusted for many confounding variables can exhibit substantial misclassification bias.

Fig. 7. Bias in raw VE estimates from simulated data in the presence of different numbers of confounding variables. The distributions in red, purple and blue correspond to uncorrected VE estimates in the presence of 2, 6 and 10 confounding variables in addition to the vaccination history.

Discussion

Misclassification caused by imperfect diagnostic tests can potentially lead to substantial biases in TND studies. By considering the processes involved in VE estimation, we have characterised the degree of bias potentially caused by misclassification in different parameter settings, finding that VE can be noticeably underestimated, particularly when the ratio between TD and ND cases in the study data is unbalanced. To address this potential bias, we developed multiple bias correction methods that provide unbiased VE estimates in both univariate and multivariate settings. When the test sensitivity and specificity are known or assumed, those values can be used to restore the true VE estimate by a relatively simple statistical procedure. Using simulations, we showed that our methods could successfully eliminate the bias in VE estimates obtained from misclassified data, although some uncertainty was introduced as a result of the information loss.

We believe that our methods could therefore enable researchers to report unbiased VE estimates even when imperfect tests had to be used. Such methods could also help in the scaling up of TND studies, as tests with limited performance are usually inexpensive and logistically convenient.

Although TND is a relatively new study design, first appearing in a publication in 2005 [Reference Skowronski27], it has gained broad popularity and is becoming a standard approach in VE studies. One of the largest factors that have contributed to its widespread use is the fact that data collection can be completed within clinical setups [Reference De Serres1]. Whereas cohort or case–control studies usually require additional efforts including follow-up or recruitment of non-patients, TND studies only involve patients visiting healthcare facilities with suspects of certain diseases and thus routinely collected clinical data can be easily adapted for analysis. VE studies of influenza, for which TND is most frequently used, often use PCR as a diagnostic tool for better data quality [Reference Darvishian23]. However, such studies usually involve intensive effort and cost, and thus may only be feasible by large-scale research bodies. Our bias correction methods may open a possibility of wider use of clinical data especially in settings where rapid tests are routinely used for diagnosis. For example, rapid influenza diagnostic tests are routinely used for outpatient clinics and hospitals in Japan, and such clinical data have facilitated a number of TND studies [Reference Seki28Reference Sugaya33]. Such studies based on rapid tests could benefit from our methods, as it would provide strong support for the validity of their estimates. Our methods may also be useful in resource-limited settings or for diseases without high-performance diagnostic tools.

Even in resourceful settings where high-performance tests are available, the slight possibility of misclassification might not always be neglected. Although PCR tests are currently used as a gold-standard for influenza diagnosis, their sensitivity and specificity may not be exact 100%; especially, the sensitivity of the test depends not only on microbiological technique but also on the quality of swab samples. In addition, it is suggested that the sensitivity of PCR tests may change during the time course of infection [Reference Sullivan, Feng and Cowling34] and be sufficiently high only during a limited time window. Our simulation study also indicated that a high heterogeneity in individual characteristics in the study population might increase the bias. Our methods could enable researchers to implement sensitivity analysis by assuming the possible test sensitivity and specificity in such cases.

Our bias correction methods are also intended to be reasonably straightforward for researchers to introduce. Existing estimation tools including software libraries and packages are often used in epidemiological analyses. Incorporating the MO approach, our parametric bootstrapping bias correction method only involves data manipulation and does not require modification of the estimation algorithm. Once multiple sets of data are randomly generated, any type of analysis can be performed as long as the results can be summarised over the MO datasets. Of particular note is that our methods for multivariate analysis allow stratification of sensitivity and specificity among individuals. Therefore, the users can employ more complex misclassification mechanisms including time-varying test performance or test performance affected by individual characteristics. Datasets with a mixture of different diagnostic tools [Reference Suzuki3, Reference Franke35] can also be handled by applying different values for each test.

There are some limitations to our study. We only focused on misclassification of diagnosis (i.e. misclassified outcomes) and did not consider misclassification of covariates (e.g. vaccine history and other confounding variables), which is another important type of misclassification in TND studies [Reference De Smedt10]. Further, it is generally not easy to plausibly estimate the sensitivity and specificity for measurement of covariates (e.g. recall bias), which must be known or assumed to implement bias correction. However, if reliable estimates are available, an extension of our approach may yield bias-corrected VE estimates in the presence of covariate misclassification. Moreover, to keep our focus only on diagnostic misclassification, our methods rested on the assumption that other sources of bias in TND studies are non-existent or properly addressed. Potential sources of bias in TND studies have been discussed elsewhere [Reference Haber16, Reference Sullivan, Tchetgen and Cowling36, Reference Lewnard37], and the researchers conducting TND studies need to carefully consider the possibility of such biases in addition to the diagnostic misclassification. Lastly, it must be noted that our methods depend on the assumed test sensitivity and specificity, and that misspecifying those values can result in an improper correction. The sensitivity and specificity of tests are usually reported by manufacturers in a comparison of the test results with gold-standard tests; however, when such gold-standard tests themselves are not fully reliable or when no available test has satisfactory performance to be regarded as gold-standard, specifying sensitivity and specificity of a test is in principle impossible. Further, test performances reported by manufacturers might lack sufficient sample size or might not be identical to those in the actual study settings. Use of composite reference standards [Reference Limmathurotsakul38, Reference Naaktgeboren39] or external/internal validation approaches [Reference Kahn40] may help overcome these problems.

Although the presence of imperfect diagnosis limits the quality of clinical data, such data can still hold useful information, and this information can be transformed into useful insights by appropriate statistical processing. Our bias correction methods were developed primarily for TND studies, but a similar approach could be applied to broader classes of estimation problems with misclassification. Potential areas for future analysis include extension to test data involving continuous quantitative measurements, and coupling with dynamic transmission models. The value of routinely collected data in healthcare settings has become widely recognised with the advancement of data infrastructure, and we believe our methods could help support the effective use of such data.

Conclusion

Bias correction methods for the TND studies were developed to address potential misclassification bias due to imperfect tests.

Supplementary material

The supplementary material for this article can be found at https://doi.org/10.1017/S0950268820002058.

Acknowledgement

We thank Sheena G. Sullivan for helpful feedback on our early work.

Financial support

This work was supported by the Wellcome Trust (S.F., grant number 210758), (A.J.K., grant number 206250); The Nakajima Foundation (A.E.); the Alan Turing Institue (A.E.); and Lnest Grant Taisho Pharmaceutical Award (A.E.).

Conflict of interest

AE received funding from Taisho Pharmaceutical Co., Ltd.

Data availability statements

This study did not involve original data collection and replication codes are available on GitHub under the MIT License (https://github.com/akira-endo/TND-biascorrection/).

Appendix

Mathematical notations in section ‘Methods and results’

Characterising bias in TND studies

r 1, r 0:

prevalence of TD/ND in the unvaccinated population

n V, n U:

vaccinated and unvaccinated population size

γ:

the relative risk of TD in the vaccinated population relative to the unvaccinated (VE = 1 − γ)

m V, m U:

the probability of medical attendance given ND in vaccinated/unvaccinated population

μ:

factor for the probability of medical attendance given TD in relative to that given ND

x V, x U, x +V, x +U, x V, x U, y V, y U, y +V, y +U, y V, y U:

the expected number of cases in the population; see Table 1 for definitions

X V, X U, Y V, Y U:

the observed case counts subject to misclassification

q:

the proportion of study samples relative to the total population

α, β:

sensitivity and specificity of the test

C:

the classification matrix

c:

the Youden index of the test; the determinant of C δ: odds of medically-attended TD in the unvaccinated population

Bias correction in univariate analysis

γ*:

corrected odds ratio

σ*:

shape parameter of the log-normal distribution that gives the confidence interval of γ* (see Equation (4))

λ V, λ U:

‘baseline medical attendance’, the mean number of vaccinated/unvaccinated patients when vaccine has no effect (VE = 0)

Bias correction in multivariate analysis

ξ = (ξ 1, ξ 2, …, ξ n):

covariates included in the model, where xi 1 denotes vaccination history

Z i, z i:

observed test result and true disease state for individual i

p 1(ξ), p 0(ξ):

probability that the true test result is positive/negative for an individual with covariates ξ

θ, θ*:

model parameter for p 1(ξ) and its estimate

$\pi _{Z_i}\lpar {\xi_i} \rpar $:

probability for the observed test result Z i

$\tilde{z}_i^j $:

imputed disease state of individual i in the j-th imputed dataset $\tilde{D}^j$

$\tilde{\phi }_{i + }\comma \;\tilde{\phi }_{i-}$:

estimated probability that the test result for individual i is incorrect

ɛ j:

point estimate for VE from the j-th imputed dataset $\tilde{D}^j$

Footnotes

1 It has been suggested that a possible violation of this assumption occur as a result of virus interference [Reference Cowling and Nishiura17], but conclusive evidence for this is currently lacking [Reference Cowling18, Reference Sundaram19] and the effect on VE estimates may be limited in any case [Reference Suzuki, Camacho and Ariyoshi20].

2 This may not be true, for example, if vaccination reduces the severity of TD and hence reduces the likelihood of medical attendance.

3 For example, a case ratio of 0.5 indicates TD:ND = 1:1 in the unvaccinated. The value is smaller than 0.5 when TD < ND and greater than 0.5 when TD > ND.

4 Strictly speaking, proportion positive is a different quantity from case ratio, but it should serve as a reasonable proxy of the case ratio in most settings.

5 Young et al. [Reference Young22] was not included because they did not report case counts.

References

De Serres, G et al. (2013) The test-negative design: validity, accuracy and precision of vaccine efficacy estimates compared to the gold standard of randomised placebo-controlled clinical trials. Eurosurveillance.CrossRefGoogle ScholarPubMed
Fukushima, W and Hirota, Y (2017) Basic principles of test-negative design in evaluating influenza vaccine effectiveness. Vaccine.CrossRefGoogle ScholarPubMed
Suzuki, M et al. (2017) Serotype-specific effectiveness of 23-valent pneumococcal polysaccharide vaccine against pneumococcal pneumonia in adults aged 65 years or older: a multicentre, prospective, test-negative design study. The Lancet Infectious Diseases 17, 313321.CrossRefGoogle ScholarPubMed
Prato, R et al. (2018) Effectiveness of the 13-valent pneumococcal conjugate vaccine against adult pneumonia in Italy: a case–control study in a 2-year prospective cohort. BMJ Open 8.CrossRefGoogle Scholar
Araki, K et al. (2018) Effectiveness of monovalent and pentavalent rotavirus vaccines in Japanese children. Vaccine 36, 51875193.CrossRefGoogle ScholarPubMed
Lopez, AL et al. (2018) Effectiveness of monovalent rotavirus vaccine in the Philippines. Scientific Reports 8, 14291.CrossRefGoogle ScholarPubMed
Muhsen, K et al. (2018) Effectiveness of rotavirus pentavalent vaccine under a universal immunization programme in Israel, 2011–2015: a case control study. Clinical Microbiology and Infection 24, 5359.CrossRefGoogle Scholar
Orenstein, EW et al. (2007) Methodologic issues regarding the use of three observational study designs to assess influenza vaccine effectiveness. International Journal of Epidemiology.CrossRefGoogle ScholarPubMed
Jackson, ML and Rothman, KJ (2015) Effects of imperfect test sensitivity and specificity on observational studies of influenza vaccine effectiveness. Vaccine.CrossRefGoogle ScholarPubMed
De Smedt, T et al. (2018) Bias due to differential and non-differential disease – and exposure misclassification in studies of vaccine effectiveness. PLoS ONE.CrossRefGoogle ScholarPubMed
Haber, M et al. (2018) A comparison of the test-negative and traditional case-control study designs with respect to the bias of estimates of rotavirus vaccine effectiveness. Vaccine 36, 50715076.CrossRefGoogle ScholarPubMed
Greenland, S (1996) Basic methods for sensitivity analysis of biases.CrossRefGoogle Scholar
Kleinbaum, DG, Kupper, LL and Morgenstern, H (1982) Epidemiologic Research: Principles and Quantitative Methods. John Wiley & Sons.Google Scholar
Jackson, ML and Nelson, JC (2013) The test-negative design for estimating influenza vaccine effectiveness. Vaccine 31, 21652168.CrossRefGoogle ScholarPubMed
Blackwell, M, Honaker, J and King, G (2017) A unified approach to measurement error and missing data: overview and applications. Sociological Methods and Research.Google Scholar
Haber, M et al. (2015) A probability model for evaluating the bias and precision of influenza vaccine effectiveness estimates from case-control studies. Epidemiology and Infection 143, 14171426.CrossRefGoogle ScholarPubMed
Cowling, BJ and Nishiura, H (2012) Virus interference and estimates of influenza vaccine effectiveness from test-negative studies. Epidemiology (Cambridge, Mass.).CrossRefGoogle ScholarPubMed
Cowling, BJ et al. (2012) Increased risk of noninfluenza respiratory virus infections associated with receipt of inactivated influenza vaccine. Clinical Infectious Diseases.CrossRefGoogle ScholarPubMed
Sundaram, ME et al. (2013) Influenza vaccination is not associated with detection of noninfluenza respiratory viruses in seasonal studies of influenza vaccine effectiveness. Clinical Infectious Diseases.CrossRefGoogle ScholarPubMed
Suzuki, M, Camacho, A and Ariyoshi, K (2014) Potential effect of virus interference on influenza vaccine effectiveness estimates in test-negative designs. Epidemiology and Infection.CrossRefGoogle ScholarPubMed
Leung, VK et al. (2016) Concordance of interim and final estimates of influenza vaccine effectiveness: a systematic review. Eurosurveillance.CrossRefGoogle ScholarPubMed
Young, B et al. (2018) Duration of influenza vaccine effectiveness: a systematic review, meta-analysis, and meta-regression of test-negative design case-control studies. Journal of Infectious Diseases.CrossRefGoogle ScholarPubMed
Darvishian, M et al. (2014) Effectiveness of seasonal influenza vaccine in community-dwelling elderly people: a meta-analysis of test-negative design case-control studies. The Lancet Infectious Diseases.CrossRefGoogle ScholarPubMed
Sidney Burrus, C (2012) Iterative reweighted least squares.Google Scholar
Rubin, DB (2004) Multiple Imputation for Nonresponse in Surveys. Wiley Classics Library. Wiley.Google Scholar
Magder, LS and Hughes, JP (1997) Logistic regression when the outcome is measured with uncertainty. American Journal of Epidemiology 146, 195203.CrossRefGoogle ScholarPubMed
Skowronski, DM et al. (2005) Effectiveness of vaccine against medical consultation due to laboratory-confirmed influenza: results from a sentinel physician pilot project in British Columbia, 2004–2005. Canada Communicable Disease Report.Google Scholar
Seki, Y et al. (2016) [Effectiveness of influenza vaccine in adults using A test-negative, case-control design −2013/2014 and 2014/2015 seasons-] (Japanese). Kansenshogaku zasshi. The Journal of the Japanese Association for Infectious Diseases.Google Scholar
Seki, Y, Onose, A and Sugaya, N (2017) Influenza vaccine effectiveness in adults based on the rapid influenza diagnostic test results, during the 2015/16 season. Journal of Infection and Chemotherapy.CrossRefGoogle ScholarPubMed
Saito, N et al. (2017) Negative impact of prior influenza vaccination on current influenza vaccination among people infected and not infected in prior season: a test-negative case-control study in Japan. Vaccine.CrossRefGoogle Scholar
Shinjoh, M et al. (2018) Inactivated influenza vaccine effectiveness and an analysis of repeated vaccination for children during the 2016/17 season. Vaccine.CrossRefGoogle Scholar
Ando, S (2018) Effectiveness of quadrivalent influenza vaccine based on the test-negative control study in children during the 2016–2017 season. Journal of Infection and Chemotherapy.CrossRefGoogle ScholarPubMed
Sugaya, N et al. (2018) Three-season effectiveness of inactivated influenza vaccine in preventing influenza illness and hospitalization in children in Japan, 2013–2016. Vaccine.CrossRefGoogle Scholar
Sullivan, SG, Feng, S and Cowling, BJ (2014) Potential of the test-negative design for measuring influenza vaccine effectiveness: a systematic review. Expert Review of Vaccines.CrossRefGoogle ScholarPubMed
Franke, MF et al. (2017) Comparison of two control groups for estimation of oral cholera vaccine effectiveness using a case-control study design. Vaccine.CrossRefGoogle ScholarPubMed
Sullivan, SG, Tchetgen, EJ and Cowling, BJ (2016) Theoretical basis of the test-negative study design for assessment of influenza vaccine effectiveness. American Journal of Epidemiology 184, 345353.CrossRefGoogle ScholarPubMed
Lewnard, JA et al. (2018) Measurement of vaccine direct effects under the test-negative design. American Journal of Epidemiology.CrossRefGoogle ScholarPubMed
Limmathurotsakul, D et al. (2012) Fool's gold: why imperfect reference tests are undermining the evaluation of novel diagnostics: a reevaluation of 5 diagnostic tests for leptospirosis. Clinical Infectious Diseases.CrossRefGoogle ScholarPubMed
Naaktgeboren, CA et al. (2013) Value of composite reference standards in diagnostic research. BMJ (Clinical research ed.).Google ScholarPubMed
Kahn, SR et al. (2018) Comparing external and internal validation methods in correcting outcome misclassification bias in logistic regression: a simulation study and application to the case of postsurgical venous thromboembolism following total hip and knee arthroplasty. Pharmacoepidemiology and Drug Safety.Google Scholar
Figure 0

Table 1. Population classified into different categories of interest in VE studies

Figure 1

Fig. 1. Bias in VE estimates caused by misclassification for different combinations of parameter values. (a)–(c) Estimated VE plotted against sensitivity. (a) True case ratio (the true ratio between TD and ND cases included in the study) = 0.2 (b) 0.4 (c) 0.6. Each two sets of lines respectively correspond to different true VEs (80% and 40%, denoted by the dotted lines). (d)–(f) Estimated VE plotted against specificity. (d) True case ratio = 0.2 (e) 0.4 (f) 0.6.

Figure 2

Fig. 2. Biased VE estimates with varying case ratio and the observed proportion of positive patients. (a) The proportion of test-positive patients in TND studies from systematic reviews. The proportions were retrieved from three systematic reviews [21–23]. (b) Estimated VE plotted against case ratio. Two sets of lines respectively correspond to different true VEs (80% and 40%, denoted by the dotted lines). The histogram in Panel (a) is overlaid on the x-axis.

Figure 3

Table 2. Simulation settings

Figure 4

Fig. 3. Bias correction for simulated data in the univariate setting. The distributions of bias-corrected VE estimates (boxplots in blue) are compared with those of raw VE estimates without correction (red). Five hundred independent datasets were randomly generated for each set of parameter values, and the corrected and uncorrected VE estimates are compared with the true value (black solid line). See Table 2 for parameter settings in each scenario.

Figure 5

Fig. 4. Bias correction method applied to published VE estimates assuming various test sensitivity and specificity. Case count data were extracted from two systematic reviews [21, 23]. Each connected set of dots show how (crude) VE estimates reported in the review varies when imperfect sensitivity and specificity are assumed. Black dots on the grey diagonal line denote the original VEs reported in the reviews. This should correspond to the true value if sensitivity = specificity = 1. Coloured dots show the bias-corrected VE considering potential misclassification.

Figure 6

Fig. 5. Multiple imputation with parametric bootstrapping.

Figure 7

Fig. 6. Bias correction for simulated data in the multivariate setting. The distributions of bias-corrected (blue) and uncorrected (red) VE estimates from 500 simulations are compared. Dotted lines denote median and black solid lines denote the true VE. The parametric bootstrapping bias correction method was used for bias correction.

Figure 8

Fig. 7. Bias in raw VE estimates from simulated data in the presence of different numbers of confounding variables. The distributions in red, purple and blue correspond to uncorrected VE estimates in the presence of 2, 6 and 10 confounding variables in addition to the vaccination history.

Supplementary material: PDF

Endo et al. Supplementary Materials

Endo et al. Supplementary Materials

Download Endo et al. Supplementary Materials(PDF)
PDF 5.4 MB