# Causal inference --- ## From internal validity to causal inference - Causal inference based on criteria - Causal pie or apportionment - Explore causation using DAGs - Counterfactual theories of causation --- ## Simpson's paradox: why we need to study causality ![](https://i.imgur.com/eSdkVad.png) --- ## Sir Austin Bradford-Hill (1965 Hill's Criteria) ![](https://i.imgur.com/WB8nyLZ.jpg) --- ## Bradford-Hill Criteria - 1 - Strength of Association - Consistency of findings - Specificity - Temporality --- ## Bradford-Hill Criteria - 2 - Biological gradient - Plausibility - Coherence - Experiment - Analogy --- ## Strength - Stronger an association, more likely it is due to cause and effect - A **stronger** association would mean confounding - Strong associations also mean high PAF% with identical prevalence of exposure - pE = 0.20,RR: 3.0, PAF: 28.57% - pE = 0.20, RR: 10, PAF: 64.28% --- ## Temporality - Cause must precede Effect --- ## Biological Gradient - As dose of exposure increases, - So does effect size --- ## Sufficient and Component Cause Model ![](https://i.imgur.com/ZiR7dvu.jpg) --- ## Directed Acyclic graphs for causal inference - Go to [http://dagitty.net/dags.html#](http://dagitty.net/dags.html#) - Take out your papers and pencils and follow along (you may work in groups) - You must close all open backdoor paths - You must not open any closed backdoor paths --- ## Backdoor paths - Backdoor paths are open if they have confounders or mediators - Backdoor paths are closed if they have colliders - Conditioning on colliders open closed backdoor paths --- ## Counterfactual causality concept 1 - Imagine A and Y are both binary, 1 and 0 - For A, we say a is counterfactual - a = 0 or a = 1, --- ## Counterfactual causality concept 2 - If everyone in the study were to - receive treatment or be exposed - simultaneously, and - what would the outcome? - P[Y_(a = 1) = 1] - Probability of the outcome Y under - a = 1 --- ## Counterfactual causality concept 3 - If everyone in the study were to - receive the control condition or non-exposed - simultaneously, - What would be the outcome? - P[Y_(a = 0) = 1] - Probability of the outcome Y under - a = 0 --- ## Causal Risk Ratio - P[Y_(a = 1)] / P[Y_(A = 0)] --- ## What do our observations show us? - But we do not get to see this, instead - We see P[Y = 1 | A = 1] - That is probability of the outcome given intervention or exposure - And, - P[Y = 1 | A = 0] - That is probability of outcome given control --- ## Associational Risk Ratio - P[Y = 1 | A = 1] / P[Y = 1 | A = 0] - if causal ratio = associational ratio, then - association == causation, otherwise not --- ## How do we measure counterfactuals? - Inverse probability weighting - Standardisation - g-methods - Instrumental variables --- ## Conclusions - Moving from internal validity to causality is complex - Criteria based - Counterfactual theories of causation - Cause can be conceptualised in sufficient and component causes - Next up: Study designs that best capture these relationships
{"metaMigratedAt":"2023-06-15T10:50:20.324Z","metaMigratedFrom":"YAML","title":"Causal inference","breaks":false,"contributors":"[{\"id\":\"2a200359-0c0b-4042-9234-d7df32d1a61b\",\"add\":3610,\"del\":436}]"}
    410 views