Shawn Graham
Catch me on Scholar Social https://scholar.social/@electricarchaeo
Shawn Graham
Department of History, Carleton U
https://scholar.social/@electricarchaeo
follow along at:
https://bit.ly/sg-necromancy
image via user 'photos_ frompasttofuture' unsplash
My first excavation involved a vampire...
Text
Colton Sturgeon, Unsplash
You're now going to get 10 weeks of class compressed into 40 minutes. Buckle Up.
(by the way you can see the class at hist4805.netlify.app)
Łukasz Łada, Unsplash
Nuggan
Clacks
Where do the shadows come from?
Networks can represent the past
Networks are present in archaeological materials
Networks can be a substrate for further simulation
Networks can problem solve
Dash Khatami via unsplash
Good ol' Wikipedia! https://en.wikipedia.org/wiki/Neural_network https://en.wikipedia.org/wiki/Perceptron
https://en.wikipedia.org/wiki/Perceptron; Charles Wightman adjusting the machine
Workable ways of (re)training neural networks in a reasonable amount of time, with a reasonable amount of data*
*terms and conditions apply
Machine learning - learning of patterns - is enormously useful in archaeology.
...and what used to be difficult is now something you can do in a webpage...
(and don't get me started about 'sharing' data via a PDF)
DH: analyzes an image to say this is a 'red figure kantharos'
GenAI says: here are the conventions that align with the idea of a kantharos rendered as pixel data
Generative AI is digital humanities run in reverse.... AI becomes a system for producing approximations of human media that align with all the data swept together to describe that media.
Salvaggio, Toward a Multi Modal Media Theory
Sam Barber via unsplash
h/t to Steven Johnson for a piece connecting Molaison & Context Length
"The computational cost scales quadratically with the length of the context window, meaning a model with a context length of 4096 tokens requires 64 times more computational resources than a model with a context length of 1024 tokens."
and we're not even talking
here about the cost(s) of siphoning
all the content in the world to get the
necessary data
...the materiality of digital archaeology...
Isi Parente Unsplash
see Drew Breunig: The 3 AI Use Cases
No gods, no interns. Only cogs.
(which would mean way less money/power for techbro oligarchs/klept)
God - Intern - Cog
necromancy for good instead of evil
yes, you've got homework
We'll take what Petrie wrote, and fill in the gaps with an LLM.... just like those scientists at Jurassic Park filled in the gaps in dino dna with amphibian DNA.
And that worked out ok.
Right? ... Right...?
(we're using GPT2, a 'completion' model that can have additional layers of training added to it. By playing with this we dispel some of the Eliza effect)
You'd think this would be difficult to do. Nope.
(and now there's a small industry of academic papers on 'digital necromancy' meaning all this)
notebook for images-as-infographics
what 'attractors' for 'archaeological excavation' do we see?
We're using a slight modification to Salvaggio's method for
Ghosts in Text Generation?
same principle - different scale
https://www.youtube.com/watch?v=bxXdGBSDCHQ
DIY run-of-stream hydro electricity rig
https://en.wikipedia.org/wiki/File:Hoover_dam_from_air.jpg
https://archaeo.social/@mrundkvist/112447233936751700
Martin Rundkvist
mrundkvist@archaeo.social
Back when #BigData was the fashionable buzz word, I repeatedly had to explain to enthusiasts that archaeological data are not just Big, they are Confused and Patchy and Hairy.
I can't really see how the current generative algorithms could make me obsolete or even speed up much of the work I do. Because I'm in this really niche activity with no commercial potential that demands constant engagement with wildly non-standardised data as well as creative writing about them.
He's right! all this is slow archaeology; this tech expands limited capabilities when you know which ghosts are haunting it. But that takes a lot to figure out.
https://github.com/shawngraham/personal-image-search-engine
LLM are no good for specific information retrieval; the ghosts push towards the mean, the average, so you get plausible text, not correct text.
This same property does permit searching by 'vibe' or 'similarity'
This term, we have spent several weeks dispelling hype, opening the hood, poking at the models, interrogating the ghosts, discussing the harms, and trying to figure out what good any of these things are. I am indebted to their goodwill and good humour.
Thanks, gang.
You are welcome to take, use, re-use, critique, expand, tear-apart, re-build, improve, sneer at, any and all code of mine that I've shared today.
By Shawn Graham
Catch me on Scholar Social https://scholar.social/@electricarchaeo