Result validity despite low predictive performances #215
-
Hi, To address this question I tried to replicate your analysis by re-running it and evaluating the performances of the ML algorithms you used. You can find my attempt in a colab notebook in which I removed all the comments and I added the performance evaluation part before the CI study. Also in this case, the ML algorithms does not seem to have very good performances (R^2 max of 0.3), but you labeled your analysis as valid. It would be great to have a discussion about this topic, thanks a lot for your help. |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 7 replies
-
Hi Dario, Could you elaborate more on how you determine "bad" ML estimates? Of course it might lead to issues if the machine learning model is not able to learn the nuisance elements e.g. the propensity score |
Beta Was this translation helpful? Give feedback.
I am sorry for the late response.
I agree with most of the points.
evaluate_learners()
) as this returns a cross-fitted value of the metricI think the important distinction is that usually the relevant confounding variables have to be included (in most cases variables affecting both
This does not mean that the variables have to be important predictors of