G-computation#

Terminology#

Names for this method:

There has been some debate around terminology. As quoted from Vansteelandt and Keiding 2011:

‘The term standardization is revealing and rather well-known to epidemiologists and therefore, in our opinion, is the terminology of choice. The term G-computation has so far been mostly reserved to refer to standardization of the effects of time-varying exposures; potentially the term “G-standardization” as nomenclature for “standardization with respect to generalized exposure regimens” would have been more enlightening. Despite the essential equivalence of G-computation for point exposures and standardization with the total population as the reference, we believe that the developments from the causal inference literature add to the literature on standardization. They give a precise meaning to standardized effect measures in terms of counterfactuals, provide insight into the delicate differences between conditional and marginal epidemiologic effect measures, and suggest novel standardization techniques that combine precision with robustness against model misspecification and extrapolation.’

About the method#

G-computation involves using a statistical model (e.g. predict) to predict potential outcomes (counterfactuals - with and without exposure).[Igelström et al. 2022]

flowchart LR; X("Binary treatment (X)"):::white; Y("Outcome (Y)"):::white; W("Confounder (W)"):::white; X --> Y; W --> X; W --> Y; classDef white fill:#FFFFFF, stroke:#FFFFFF; classDef black fill:#FFFFFF, stroke:#000000; classDef empty width:0px,height:0px; classDef green fill:#DDF2D1, stroke: #FFFFFF;

Key steps:

  1. Fit model on observed data - regression model with Y as outcome and X and W as predictors (perhaps also with polynomials and/or interactions if there are multiple control variables)

  2. Make identical copy of observed data, but just replace outcome so all X=1.

  3. Create another where all to X=0.

  4. Use our fitted model to predict outcomes in the two counterfactual datasets [source] [Batten 2023] [source]

  5. Estimate the Average treatment effect (ATE) - this is the mean difference (+ 95% CI) in the predicted outcomes between the two groups.[Batten 2023] It essentially describes the average effect, at a population level, of moving an entire population from untreated to treated.[Chatton et al. 2020]

For average treatment effect on the treated (ATT):

  • This is the average effect of treatment on those subjects who ultimately received the treatement [Chatton et al. 2020]

  • To calculate ATT, use only the treated units in steps 2 and 3. ‘The control units are still used to fit the model in 1, but only the treated units are used to compute the predicted values.’ [source]

You could also compute average treatment effect on the untreated (ATU). [Wang et al. 2017]

To get standard errors, you can use bootstrapping or the delta method. [source]

How does this method enable us to deal with treatment-dependent confounding?#

See the example of treatment-dependent confounding below.

flowchart LR; ak("A<sub>K</sub>: Treatment"):::green; ak1("A<sub>K+1</sub>: Treatment"):::green; yk1("Y<sub>K+1</sub>: Outcome"):::green; yk2("Y<sub>K+2</sub>: Outcome"):::green; lk("L<sub>K</sub>: Confounder"):::white; lk1("L<sub>K+1</sub>: Confounder"):::white; u("U: Unmeasured<br>confounder"):::white; ak --> yk1; u --> yk1; u --> lk; lk --> ak; lk --> ak1; ak --> ak1; ak --> lk1; u --> lk1; lk1 --> ak1; ak1 --> yk2; u --> yk2; lk --> yk1; lk --> lk1; lk1 --> yk2; yk1 --> yk2; classDef white fill:#FFFFFF, stroke:#FFFFFF classDef black fill:#FFFFFF, stroke:#000000 classDef green fill:#DDF2D1, stroke: #FFFFFF;

If we regressed AK and AK+1 on YK+2 while adjusting for confounders should we adjust for the time-varying LK+1 and YK+1? We can see that:

  1. LK+1 is a collider on the non-causal path AK –> LK+1 <– U –> YK+2

flowchart LR; ak("A<sub>K</sub>: Treatment"):::red; ak1("A<sub>K+1</sub>: Treatment"):::white; yk1("Y<sub>K+1</sub>: Outcome"):::white; yk2("Y<sub>K+2</sub>: Outcome"):::red; lk("L<sub>K</sub>: Confounder"):::white; lk1("L<sub>K+1</sub>: Confounder"):::red; u("U: Unmeasured<br>confounder"):::red; ak --> yk1; u --> yk1; u --> lk; lk --> ak; lk --> ak1; ak --> ak1; ak --> lk1; u --> lk1; lk1 --> ak1; ak1 --> yk2; u --> yk2; lk --> yk1; lk --> lk1; lk1 --> yk2; yk1 --> yk2; classDef white fill:#FFFFFF, stroke:#FFFFFF classDef black fill:#FFFFFF, stroke:#000000 classDef red fill:#FFCCCB, stroke:red classDef green fill:#DDF2D1, stroke: #FFFFFF linkStyle 6,7,10 stroke:red;
  1. LK+1 is a mediator on the causal path AK –> LK+1 –> AK+1 –> YK+2

flowchart LR; ak("A<sub>K</sub>: Treatment"):::red; ak1("A<sub>K+1</sub>: Treatment"):::red; yk1("Y<sub>K+1</sub>: Outcome"):::white; yk2("Y<sub>K+2</sub>: Outcome"):::red; lk("L<sub>K</sub>: Confounder"):::white; lk1("L<sub>K+1</sub>: Confounder"):::red; u("U: Unmeasured<br>confounder"):::white; ak --> yk1; u --> yk1; u --> lk; lk --> ak; lk --> ak1; ak --> ak1; ak --> lk1; u --> lk1; lk1 --> ak1; ak1 --> yk2; u --> yk2; lk --> yk1; lk --> lk1; lk1 --> yk2; yk1 --> yk2; classDef white fill:#FFFFFF, stroke:#FFFFFF classDef black fill:#FFFFFF, stroke:#000000 classDef red fill:#FFCCCB, stroke:red classDef green fill:#DDF2D1, stroke: #FFFFFF linkStyle 6,8,9 stroke:red;
  1. LK+1 is a confounder on the non-causal path AK+1 <– LK+1 –> YK+2

flowchart LR; ak("A<sub>K</sub>: Treatment"):::white; ak1("A<sub>K+1</sub>: Treatment"):::red; yk1("Y<sub>K+1</sub>: Outcome"):::white; yk2("Y<sub>K+2</sub>: Outcome"):::red; lk("L<sub>K</sub>: Confounder"):::white; lk1("L<sub>K+1</sub>: Confounder"):::red; u("U: Unmeasured<br>confounder"):::white; ak --> yk1; u --> yk1; u --> lk; lk --> ak; lk --> ak1; ak --> ak1; ak --> lk1; u --> lk1; lk1 --> ak1; ak1 --> yk2; u --> yk2; lk --> yk1; lk --> lk1; lk1 --> yk2; yk1 --> yk2; classDef white fill:#FFFFFF, stroke:#FFFFFF classDef black fill:#FFFFFF, stroke:#000000 classDef red fill:#FFCCCB, stroke:red classDef green fill:#DDF2D1, stroke: #FFFFFF linkStyle 8,9,13 stroke:red;

If we adjusted for for LK+1 we would therefore:

  • Eliminate confounding bias

  • Introduce collider (stratification) bias

  • Introduce over-adjustment bias in the effect of AK (from controlling for mediator)

The same argument applies when adjusting for the intermediate outcome YK+1. In other words, its impossible as we need to simultaneously adjust for and not adjust for LK+1 and YK+1. ‘The g-formula resolves this problem by decoupling adjusting, and not adjusting, for treatment-dependent confounders’ like LK+1.

  1. G-formula step 1 involves fitting a model with outcome, treatment and confounders. This adjusts for LK+1 and YK+1, to avoid confounding bias.

  2. G-formula step 2 involves predicting outcomes where all patients are set to no treatment, or all to treatment, and comparing the outcomes to get the average treatment effect. This marginalises/averages over the counterfactual distribution of LK+1 and YK+1, so they are not adjusted for, to avoid collider and over-adjustment biases

Hence, it simultaneously does and not adjust for LK+1 and YK+1. [Loh et al. 2023]

Assumptions#

  • Sequential ignorability assumption i.e. no unmeasured confounding - ‘when there is no causal effect, the treatment and outcome are conditionally independent given a set of pre-treatment covariates’ [Loh et al. 2023]

G-computation v.s. IPTW#

‘Standardization models the outcome, whereas inverse probability weighting models the treatment’. If we were to do IPW and standardisation ‘without using any models (i.e. non-parametrically) then we would expect both methods to give the exact same result’. [Batten 2023]

However, we expect them to differ if we use models to estimate them since some degree of misspecification is inescapable in models, ‘but misspecification in the treatment model (IP weighting) and outcome model (standardisation) will not generally result in the same magnitude and direction of bias in the effect estimate’.

‘Both IP weighting and standardization are estimators of the g-formula, a general method for causal inference first described in 1986… We say that standardization is a plug-in g-formula estimator because it simply replaces the conditional mean outcome in the g-formula by its estimates. When those estimates come from parametric models, we refer to the method as the parametric g-formula. Because here we were only interested in the average causal effect, we estimated parametrically the conditional mean outcome.’

‘Often there is no need to choose between IP weighting and the parametric g-formula. When both methods can be used to estimate a causal effect, just use both methods. Also, whenever possible, use doubly robust methods that combine models for treatment and for outcome in the same estimator’. [Hernán and Robins 2024]