BMS-Aned seminar
Department of Data Science Methods, Julius Center, University Medical Center Utrecht
2024-09-26
What is AI?
Artificial Intelligence is the branch of computer science that focuses on creating systems capable of performing tasks that typically require human intelligence. (Russell and Norvig 2020)
What is Machine Learning?
Machine Learning is a subset of artificial intelligence that involves the use of algorithms and statistical models to enable computers to perform tasks without explicit instructions. Instead, they rely on patterns and inference from data. (Samuel 1959)
What tasks can we perform with machine learning?
i | length | weight | sex |
---|---|---|---|
1 | 137 | 30 | boy |
2 | 122 | 24 | girl |
3 | 101 | 18 | girl |
… | … | … | … |
We typically assume these data are (i.i.d.) samples from some unknown distribution p(l,w,s):
li,wi,si∼p(l,w,s)
lj,wj,sj∼pθ(l,w,s)
task | |
---|---|
generation | lj,wj,sj∼pθ(l,w,s) |
use samples to learn model for conditional distribution p lj,wj∼pθ(l,w|s=boy)
task | |
---|---|
generation | lj,wj,sj∼pθ(l,w,s) |
conditional generation | lj,wj∼pθ(l,w|s=boy) |
use samples to learn model for conditional distribution p of one variable sj∼pθ(s|l=l′,w=w′)
task | |
---|---|
generation | lj,wj,sj∼pθ(l,w,s) |
conditional generation | lj,wj∼pθ(l,w|s=boy) |
call this one variable outcome and - classify when majority of generated samples are of a certain class - or: have a model that outputs expected values sj=pθ(s|l=l′,w=w′)>0.5
task | |
---|---|
generation | lj,wj,sj∼pθ(l,w,s) |
conditional generation | lj,wj∼pθ(l,w|s=boy) |
discrimination | pθ(s|l=li,w=wi)>0.5 |
y=5∑i=0xiβi
hi=w0i+w1ix1+…hi=g(hi)y=3∑i=1hiwi
L(θ)=n∑i=1ℓ(yi,f(xi;θ))
∇L(θ)≈1mm∑i=1∇ℓ(yi,f(xi;θ))
θt+1=θt−α∇L(θ)
https://www.kaggle.com/code/ryanholbrook/stochastic-gradient-descent
Parameter counting is a bad proxy for model complexity in neural networks
Whereas in regression models, model complexity is well-captured by the number of parameters, this is not the case for neural networks.
DOI: 10.1093/llc/fqy085
deep learning for cell counting (Moen et al. 2019)
word1∼pchatGPT(word|prompt)
Prompt=“Frank went to the bar and”
had∼pchatGPT(word|Frank went to the bar and)
Prompt=“Frank went to the bar and”
had∼pchatGPT(word|Frank went to the bar and)a∼pchatGPT(word|Frank went to the bar and had)
Prompt=“Frank went to the bar and”
had∼pchatGPT(word|Frank went to the bar and)a∼pchatGPT(word|Frank went to the bar and had)drink∼pchatGPT(word|Frank went to the bar and had a)
Prompt=“Frank went to the bar and”
had∼pchatGPT(word|Frank went to the bar and)a∼pchatGPT(word|Frank went to the bar and had)drink∼pchatGPT(word|Frank went to the bar and had a)STOP∼pchatGPT(word|Frank went to the bar and had a drink)
Prompt=“Frank went to the bar and”
met∼pchatGPT(word|Frank went to the bar and)
Prompt=“Frank went to the bar and”
met∼pchatGPT(word|Frank went to the bar and)a∼pchatGPT(word|Frank went to the bar and met)
Prompt=“Frank went to the bar and”
met∼pchatGPT(word|Frank went to the bar and)a∼pchatGPT(word|Frank went to the bar and met)friend∼pchatGPT(word|Frank went to the bar and met a)
Prompt=“Frank went to the bar and”
met∼pchatGPT(word|Frank went to the bar and)a∼pchatGPT(word|Frank went to the bar and met)friend∼pchatGPT(word|Frank went to the bar and met a)STOP∼pchatGPT(word|Frank went to the bar and met a friend)
- more compute resources
- bigger data
- bigger models (enabled by data and compute)
©Wouter van Amsterdam — WvanAmsterdam — wvanamsterdam.com/talks