AI 101
HMIA 2025

AI 101
HMIA 2025
Goals
Demystify some terms. Fill in some gaps. Not exactly "big picture" but "starting picture." Meet 30 people at 30 different places. Apologies to those who know way more than I do.
THERE ARE NO SILLY QUESTIONS.
I don't have all the answers.
AI 101
HMIA 2025
"Readings"
PRE-CLASS
CLASS
Outline
- Programing vs Machine Learning
- AI landscape
- How machines learn
- Training/Deployment; pretraining/finetuning; supervised/unsupervised; RL/RLHF; loss & optimization
- Perceptron, neuron, layers, forward, backward
- Accuracy, precision, reliability, validity
- Bias/fairness
- Alignment
HMIA 2025 AI Bootcamp
HMIA 2025 AI Bootcamp
HMIA 2025 AI Bootcamp
PRE-CLASS
HMIA 2025 AI Bootcamp
PRE-CLASS
HMIA 2025 AI Bootcamp
PRE-CLASS
HMIA 2025 AI Bootcamp
PRE-CLASS
Intro to Machine Learning (ML Zero to Hero - Part 1) [7m17s]
HMIA 2025 AI Bootcamp
PRE-CLASS
HMIA 2025 AI Bootcamp
PRE-CLASS
HMIA 2025 AI Bootcamp
Text
3Blue1Brown: Large Language Models Explained Briefly [7m57s]
PRE-CLASS
HMIA 2025 AI Bootcamp
PRE-CLASS
HMIA 2025 AI Bootcamp
PRE-CLASS
History
HMIA 2025 AI Bootcamp
CLASS

Models
HMIA 2025 AI Bootcamp
CLASS
Models
HMIA 2025 AI Bootcamp
CLASS
m
b
X→
→Y
a black box with two dials
prediction
data
parameters
error/loss
Brains Can Be Built of Biological Neurons
HMIA 2025 AI Bootcamp
CLASS
Inputs
Output
dendrites
axon
nucleus
Models Can Be Built of Artificial Neurons
HMIA 2025 AI Bootcamp
CLASS
Inputs
Output
1
0
Models as Functions
HMIA 2025 AI Bootcamp
CLASS

Input
Hidden Layers
Output
weights (that's what's "learned")
Deep (multiple hidden layers)
Neural (each node an artificial neuron)
Network
Why Linear Algebra?
HMIA 2025 AI Bootcamp
CLASS
Input
Hidden Layers
Cat Detector as Function
HMIA 2025 AI Bootcamp
CLASS


point in super high dimensional space
f
Reinforcement Learning
HMIA 2025 AI Bootcamp
CLASS

STOP+THINK
What else about learning to ride a bike strikes you as interesting? Do you think it's important that it's fun? Is it important that mom is there?
Living is One Fork in the Road After Another

https://devblogs.nvidia.com/deep-learning-nutshell-reinforcement-learning/ (Video courtesy of Mark Harris, who says he is “learning reinforcement” as a parent.)

Agent Models
Reinforcement Learning with Human Feedback
(RLHF)
Remember the other day when we talked about having "other people in our head"?
Train your model to produce outputs from inputs.
Let it run on a lot of different inputs and rate the outputs.
Build a model that predicts the rating of each output, given a particular input.
IMAGES

|



Try out everything over and over and over again and keep track of how everything works out each time.
The Plan
What can models do?
HMIA 2025 AI Bootcamp
CLASS
Classify: "look" at an item and put it in a category
Predict: "look" at some information and say what comes/happens next
Optimize: "look" at situation, objective, and resources and choose best solution
Generate: "look" at a description, create a text or image or sound
Translate: "look" at something in one form and transform it to another form
Networks of Neurons
HMIA 2025 AI Bootcamp
CLASS

Imagine you are a worm learning how to move around in the world. You are constantly bumping into things. Some things are sharp and threatening. You can defend yourself against sharp, threatening things by curling into a ball. But if you curl into a ball everything something touches you, all the other worms will laugh and you'll never get anywhere.
So you have this neuron that responds to touch
and drives the curl-up action in response. How does the worm train the neuron to respond to the proper levels of touch (only curl up when you encounter a sharp and threatening touch)?
Let's suppose the neuron works like this: the input is a measure of the touch; this is compared to a threshold and if it is higher, the neuron triggers a curl-up action.
Initially, the threshold is zero: the worm curls up at the slightest touch. This happens a few times, touches of level 1 and level 2 and level 9, and the worm notices that nothing happened, the touch did not cause any harm. So, it raises the threshold to 1. Now when it gets a level 1 touch it just keeps moving. But level 2 touches cause it to curl up. And then some 5s and 6s and it curls up. Again, it notices that nothing happened so it raises the threshold to level 2.
This continues for a bit and it has raised the threshold to level 6. Then
one day, it gets a level 5 poke that really hurts — a sharp thorn or a
hidden barb. So the worm lowers its threshold by 1 to level 5. Now it curls up when it feels a level 5 poke. For a while, it seems to be avoiding injury quite well so it lowers the threshold to 4. Then it notices that sometimes a level 5 poke is painful but sometimes not and so it leaves the threshold at 4. The worm has learned an approximate rule: curl up when things feel like level 5 or higher — but not always, and not perfectly.
HMIA 2025 AI Bootcamp
CLASS

We all know about the game of guessing the number of jelly beans in a jar. And if you have taken a statistics you have heard of the "wisdom of the crowd" : we can ask a bunch of people to estimate and then we take the average of their guesses.
Could we teach a machine to learn how much to
trust each person’s guess — and adjust based
on feedback?
ROUND 1. Present the jar, record the guesses,
compute the average, reveal the actual number of
jelly beans.
Remind ourselves that computing the average is
the same as computing a weighted sum where the weights are all 1/n where n is the number of people guessing. We count the number of folks in the room
and compute 1/n. "This is your initial weight."
ROUND 2. We ask for show of hands: how many had a guess that was high? Low? If your guess last round was higher than the actual number, subtract 0.05 from your weight. If it was lower, add 0.05.
If you were spot-on, leave it as is.
Now we introduce a new jar and ask folks to write down their guess.
We have folks do a little math - multiply their
guess times their weight - and we add up all the results.
Then we reveal the actual number of jelly beans.
ROUND 3. We repeat the process.
STOP+THINK. What's going to happen to folks who consistently guess too low?
We can introduce some limits like "your weight can never go above 1 or below 0. Or, after we increment or decrement everyone's weights we could adjust them all so they still add up to 1. This would preserve the math of the weighted average.
STOP+THINK. Would it be better to add/subtract more weight from people who were really far off, and less from those who were just a little off?
HMIA 2025 AI Bootcamp
CLASS

HMIA 2025 AI Bootcamp


low
high
high
low
low
CLASS
HMIA 2025 AI Bootcamp

CLASS
Models as Functions
HMIA 2025 AI Bootcamp
CLASS

Input
Hidden Layers
Output
Models
HMIA 2025 AI Bootcamp
CLASS
HMIA 2025 AI Bootcamp
CLASS

HMIA 2025
HMIA 2025 AI Bootcamp
By Dan Ryan
HMIA 2025 AI Bootcamp
- 75