Shawn Graham
Catch me on Scholar Social https://scholar.social/@electricarchaeo
Shawn Graham
Department of History, Carleton U
https://scholar.social/@electricarchaeo
follow along at:
https://bit.ly/sg-aug26
image via user 'photos_ frompasttofuture' unsplash
My first excavation involved a vampire...
Text
Colton Sturgeon, Unsplash
You're now going to get 10 weeks of class compressed into a single oracular presentation.
(by the way you can see the class at hist4805.netlify.app)
Out any day now.
Any day.
...soon.
any time...
here it comes...
(it's coming, I swear)
...any day now.
The Digital Press at the University of North Dakota
https://thedigitalpress.org/
A press open to experimentation and Open Access publishing
Łukasz Łada, Unsplash
Nuggan
Clacks
Where do the shadows come from?
Networks can represent the past
Networks are present in archaeological materials
Networks can be a substrate for further simulation
Networks can problem solve
Dash Khatami via unsplash
Good ol' Wikipedia! https://en.wikipedia.org/wiki/Neural_network https://en.wikipedia.org/wiki/Perceptron
https://en.wikipedia.org/wiki/Perceptron; Charles Wightman adjusting the machine
Workable ways of (re)training neural networks in a reasonable amount of time, with a reasonable amount of data*
*terms and conditions apply
Machine learning - learning of patterns - is enormously useful in archaeology.
...and what used to be difficult is now something you can do in a webpage...
(and don't get me started about 'sharing' data via a PDF)
DH: analyzes an image to say this is a 'red figure kantharos'
GenAI says: here are the conventions that align with the idea of a kantharos rendered as pixel data
Generative AI is digital humanities run in reverse.... AI becomes a system for producing approximations of human media that align with all the data swept together to describe that media.
Salvaggio, Toward a Multi Modal Media Theory
Sam Barber via unsplash
h/t to Steven Johnson for a piece connecting Molaison & Context Length
"The computational cost scales quadratically with the length of the context window, meaning a model with a context length of 4096 tokens requires 64 times more computational resources than a model with a context length of 1024 tokens."
and we're not even talking
here about the cost(s) of siphoning
all the content in the world to get the
necessary data
...the materiality of digital archaeology...
Isi Parente Unsplash
https://shawngraham.github.io/homecooked-history/sentence-analyzer/
This toy lets you define a direction in semantic space and then visualize sentence fragments along that direction. Uses a language model but *not* a generative one
Give it a try on any AI pronouncement; define a spectrum capturing 'religiosity'. What do you spot?
see Drew Breunig: The 3 AI Use Cases
No gods, no interns. Only cogs.
(which would mean way less money/power for techbro oligarchs/klept)
God - Intern - Cog
necromancy for good instead of evil
yes, you've got homework
We'll take what Petrie wrote, and fill in the gaps with an LLM.... just like those scientists at Jurassic Park filled in the gaps in dino dna with amphibian DNA.
And that worked out ok.
Right? ... Right...?
(we're using GPT2, a 'completion' model that can have additional layers of training added to it. By playing with this we dispel some of the Eliza effect)
You'd think this would be difficult to do. Nope.
(and now there's a small industry of academic papers on 'digital necromancy' meaning all this)
Use the same link as before, exercise 2
what 'attractors' for 'archaeological excavation' do we see?
We're using a slight modification to Salvaggio's method for
Ghosts in Text Generation?
Same link as before, exercise 3
same principle - different scale
https://www.youtube.com/watch?v=bxXdGBSDCHQ
DIY run-of-stream hydro electricity rig
https://en.wikipedia.org/wiki/File:Hoover_dam_from_air.jpg
DO NOT teach 'prompts for academic writing' etc. THAT's what OpenAI wants: it leans into the idea that there is an intelligence who cares on the other end of the screen.
AS MANAGERS do NOT mandate AI use or enable un-opt-outable AI 'features': that is compelled religion
These are models of culture, NOT Data from Star Trek, not AGIMUS... but they could become Landru
https://github.com/shawngraham/personal-image-search-engine
LLM are no good for specific information retrieval; the ghosts push towards the mean, the average, so you get plausible text, not correct text.
This same property does permit searching by 'vibe' or 'similarity'
Last year, we spent several weeks dispelling hype, opening the hood, poking at the models, interrogating the ghosts, discussing the harms, and trying to figure out what good any of these things are. I am indebted to their goodwill and good humour.
Thanks, gang.
You are welcome to take, use, re-use, critique, expand, tear-apart, re-build, improve, sneer at, any and all code of mine that I've shared today.
By Shawn Graham
Catch me on Scholar Social https://scholar.social/@electricarchaeo