Infection and Immunity spring meeting
Department of Data Science Methods, Julius Center, University Medical Center Utrecht
2024-04-18
What is artificial intelligence?
computers doing tasks that normally require intelligence 1
What is artificial general intelligence?
General purpose AI that performs a range of tasks in different domains like humans
data:
i | length | weight | sex |
---|---|---|---|
1 | 137 | 30 | boy |
2 | 122 | 24 | girl |
3 | 101 | 18 | girl |
… | … | … | … |
\[l_i,w_i,s_i \sim p(l,w,s)\]
use samples to learn model \(p_{\theta}\) for joint distribution \(p\) \[ l_j,w_j,s_j \sim p_{\theta}(l,w,s) \]
use samples to learn model for conditional distribution \(p\) \[ l_j,w_j \sim p_{\theta}(l,w|s=\text{boy}) \]
task | |
---|---|
generation | \(l_j,w_j,s_j \sim p_{\theta}(l,w,s)\) |
use samples to learn model for conditional distribution \(p\) of one variable \[ s_j \sim p_{\theta}(s|l=l',w=w') \]
task | |
---|---|
generation | \(l_j,w_j,s_j \sim p_{\theta}(l,w,s)\) |
conditional generation | \(l_j,w_j \sim p_{\theta}(l,w|s=\text{boy})\) |
call this one variable outcome and classify when expected value passes threshold (e.g. 0.5): \[ s_j = p_{\theta}(s|l=l',w=w') > 0.5 \]
task | |
---|---|
generation | \(l_j,w_j,s_j \sim p_{\theta}(l,w,s)\) |
conditional generation | \(l_j,w_j \sim p_{\theta}(l,w|s=\text{boy})\) |
discrimination | \(p_{\theta}(s|l=l_i,w=w_i) > 0.5\) |
- There is no fundamental difference between statistics and machine learning
- both optimize parameters to improve some criterion (loss / likelihood) that measures model fit to data
- models used in machine learning are more flexible
What is chatGPT?
a stochastic auto-regressive next-word predictor with a chatbot interface
\[\begin{align} \text{word}_1 &\sim p_{\text{chatGPT}}(\text{word}|\text{prompt})\\ \end{align}\]
\[\begin{align} \text{word}_1 &\sim p_{\text{chatGPT}}(\text{word}|\text{prompt})\\ \text{word}_2 &\sim p_{\text{chatGPT}}(\text{word}|\text{word}_1,\text{prompt}) \end{align}\]
\[\begin{align} \text{word}_1 &\sim p_{\text{chatGPT}}(\text{word}|\text{prompt})\\ \text{word}_2 &\sim p_{\text{chatGPT}}(\text{word}|\text{word}_1,\text{prompt}) \end{align}\]
\[\begin{align} \text{word}_1 &\sim p_{\text{chatGPT}}(\text{word}|\text{prompt})\\ \text{word}_2 &\sim p_{\text{chatGPT}}(\text{word}|\text{word}_1,\text{prompt})\\ \text{word}_n &\sim p_{\text{chatGPT}}(\text{word}|\text{word}_{n-1},\ldots,\text{word}_1,\text{prompt}) \end{align}\]
\[\begin{align} \text{word}_1 &\sim p_{\text{chatGPT}}(\text{word}|\text{prompt})\\ \text{word}_2 &\sim p_{\text{chatGPT}}(\text{word}|\text{word}_1,\text{prompt})\\ \text{word}_n &\sim p_{\text{chatGPT}}(\text{word}|\text{word}_{n-1},\ldots,\text{word}_1,\text{prompt})\\ \text{STOP} &\sim p_{\text{chatGPT}}(\text{word}|\text{word}_{n-1},\ldots,\text{word}_1,\text{prompt}) \end{align}\]
machine learning
statistics (e.g. GLMs)
- ML in medicine has been ‘hot’ since at least the 90s (Cooper et al. 1997)
- not much evidence that it outperforms regression on most tasks (Christodoulou et al. 2019)
- though many poorly performed studies (Dhiman et al. 2022)
Question 1
Question 2
Tip
building models for decision support without regards for the historic treatment policy is a bad idea
Note
The question is not “is my model accurate before / after deploytment”, but did deploying the model improve patient outcomes?
Tip
It may seem obvious that you should not ignore historical treatments in your prediction models, if you want to improve treatment decisions, but many of these models are published daily, and some guidelines even allow for implementing these models based on predictve performance only
- condition on given treatment and traits
- unobserved confounding (hat type) leads to wrong treatment decisions
because they do not evaluate the policy change
What to do?
What to do?
Evaluate policy change (cluster randomized controlled trial)
Build models that are likely to have value for decision making
Predict outcome under hypothetical intervention of giving certain treatment
always discuss:
1. what is effect on treatment policy?
2. what is effect on patient outcomes?
Don’t assume predicting well leads to good decisions
think about the policy change
From algorithms to action: improving patient care requires causality (Amsterdam, Jong, et al. 2024)
When accurate prediction models yield harmful sel-fulfilling prophecies (Amsterdam, Geloven, et al. 2024)
thank you
©Wouter van Amsterdam — WvanAmsterdam — wvanamsterdam.com/talks