You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: 1RHC.Rmd
+6-2Lines changed: 6 additions & 2 deletions
Original file line number
Diff line number
Diff line change
@@ -188,7 +188,7 @@ Table 5 in @connors1996effectiveness showed that, after propensity score pair (1
188
188
We also conduct propensity score pair matching analysis, as follows.
189
189
190
190
```{block, type='rmdcomment'}
191
-
**Note**: In this workshop, we will not cover Propensity Score Matching (PSM) in this workshop. If you want to learn more about this, feel free to check out this other workshop: [Understanding Propensity Score Matching](https://ehsanx.github.io/psw/).
191
+
**Note**: In this workshop, we will not cover Propensity Score Matching (PSM). If you want to learn more about this, feel free to check out this other workshop: [Understanding Propensity Score Matching](https://ehsanx.github.io/psw/) and the [video recording](https://www.youtube.com/watch?v=u4Nl7gnDEAY) on youtube.
192
192
```
193
193
194
194
```{r ps16854, cache=TRUE, echo = TRUE}
@@ -236,6 +236,8 @@ The love plot suggests satisfactory propensity score matching (all SMD < 0.1).
Hence, our conclusion based on propensity score pair matched data ($p \lt 0.001$) is different than Table 5 in @connors1996effectiveness ($p = 0.14$). Variability in results for 1-to-1 matching is possible, and modelling choices may be different (we used caliper option here).
251
+
Our conclusion based on propensity score pair matched data ($p \lt 0.001$) is different than Table 5 in @connors1996effectiveness ($p = 0.14$). Variability in results for 1-to-1 matching is possible, and modelling choices may be different (we used caliper option here).
250
252
```
251
253
254
+
##### Treatment effect
255
+
252
256
- We can also estimate the effect of `RHC` on `length of stay` using propensity score-matched sample:
Copy file name to clipboardExpand all lines: 2gcomp2.Rmd
+5-2Lines changed: 5 additions & 2 deletions
Original file line number
Diff line number
Diff line change
@@ -289,7 +289,7 @@ Notice that the mean is very similar to the parametric G-computation method.
289
289
## G-comp using SuperLearner
290
290
291
291
```{block, type='rmdcomment'}
292
-
SuperLearner is an ensemble MLtechnique, that uses **cross-validation** to find a weighted combination of estimates provided by different **candidate learners** (that help predict).
292
+
SuperLearner is an ensemble ML technique, that uses **cross-validation** to find a weighted combination of estimates provided by different **candidate learners** (that help predict).
293
293
```
294
294
295
295
- There exists many candidate learners. Here we are using a combination of
- Assuming all covariates are measured, **parametric models** such as linear and logistic regressions are very efficient, but relies on strong assumptions. In real-world scenarios, it is often hard (if not impossible) to guess the correct specification of the right hand side of the regression equation.
82
82
- Machine learning (ML) methods are very helpful for prediction goals. They are also helpful in **identifying complex functions** (non-linearities and non-additive terms) of the covariates (again, assuming they are measured).
83
83
- There are many ML methods, but the procedures are very different, and they come with their own advantages and disadvantages. In a given real data, it is **hard to apriori predict which is the best ML algorithm** for a given problem.
84
-
- That's where super learner is helpful in **combining strength from various algorithms**, and producing 1 prediction column that has **optimal statistical properties**.
84
+
85
+
86
+
```{block, type='rmdcomment'}
87
+
Super learner is helpful in **combining strength from various algorithms**, and producing 1 prediction column that has **optimal statistical properties**.
- For causal inference goals (when we have a primary exposure of interest), machine learning methods are often misleading. This is primarily due to the fact that they usually do not have an inherent mechanism of focusing on **primary exposure** (RHC in this example); and treats the primary exposure as any other predictors.
125
-
- When using g-computation with ML methods, estimation of variance becomes a difficult problem. Generalized procedures such as **robust SE or bootstrap methods** are not supported by theory.
126
-
- That's where TMLE methods shine, with the help of it's important **statistical properties (double robustness, finite sample properties)**.
129
+
- When using g-computation with ML methods, estimation of variance becomes a difficult problem (with correct coverage). Generalized procedures such as **robust SE or bootstrap methods** are not supported by theory.
130
+
131
+
132
+
```{block, type='rmdcomment'}
133
+
TMLE method shine, with the help of it's important **statistical properties (double robustness, finite sample properties)**.
0 commit comments