typing was never the hard part

DevOpsDays Austin 2026

Ian Littman / @ian@phpc.social / @ian.im / @iansltx

Slides at ian.im/doda26

Warning: we're gonna talk about LLMs (#aI)

  1. What I think LLMs are good for
  2. Recent history highlights (since last year's DevOpsDays Austin)
  3. What this seems to mean software dev workflows
  4. What this seems to mean for DevOps workflows

 

I'll include some implementation details on my setup. Find me after to chat more.

 

Slides are mine. LLMs didn't touch them.

What LLMs are good for

  1. Easy to describe acceptance criteria
  2. Toilsome to implement
  3. Following an existing pattern
  4. Straightforward (preferably automated) to verify
  5. Small units of work (this is no longer a constraint)

History: Better models && better harnesses

Late May 2025

  • Claude Sonnet 4 released by Anthropic
  • Claude Code went GA
  • I start using Sonnet ~1mo later, and then more in August
  • Opus existed but was $$$; Sonnet was a situational toil-reduction pick
  • Already easy to find a model that was better at jq/bash/regex than I was

september 2025

  • Claude Sonnet 4.5 released by Anthropic
  • Useful for a larger selection of toil
  • Used occasionally by me
  • Qwen3-Next, GLM 4.6 released

Late November 2025

  • Claude Opus 4.5 released by Anthropic
  • Step change in capabilities
  • Significantly less expensive
  • I started throwing it larger tasks ~year-end
  • GLM 4.7 released in December

February 2026

  • Sonnet 4.6 + Opus 4.6 released
  • GPT 5.3 Codex released
  • Qwen3.5 released (step change on local model ability)
  • GLM 5 released
  • Kimi K2.5 released in January (I used it a bit)

march 2026

  • Claude Code instability
  • Claude plan rate limit revisions
  • Claude code review is more widely available
  • GPT-5.4 released

April 2026

  • Releases (non-exhaustive list)
    • GPT-5.5
    • GLM 5.1
    • Qwen3.6
    • Kimi K2.6
    • DeepSeek V4
    • Opus 4.7
  • Rugpulls
    • GitHub Copilot
    • Claude Enterprise

Okay, that was a lot

This means that...

  • Using an LLM is increasingly likely to cost real money
  • Business models based on sustained LLM subsidies are time bombs
  • Doing less (token count) with less (cheaper/simpler models) matters
  • If deterministic code and LLMs can both do a thing,
    pick (or generate) the former

Software dev

  • Using an LLM is increasingly likely to cost real money
  • Business models based on sustained LLM subsidies are time bombs
  • Doing less (token count) with less (cheaper/simpler models) matters
  • If deterministic code and LLMs can both do a thing,
    pick (or generate) the former

let's talk about subsidies

  • LLM lab plans (OpenAI, Anthropic)
  • GitHub Copilot
  • OpenCode Go (less egregious?)
  • Compared to paying APIs per-token

Thanks!

Questions? Find me here / @ian@phpc.social / @ian.im / @iansltx

Slides: https://ian.im/doda26

Typing was never the hard part - DevOpsDays Austin 2026

By Ian Littman

Typing was never the hard part - DevOpsDays Austin 2026

It's 2026, and you can now outsource your toil to a GPU cluster...and sometimes to a chip in your laptop. If you're brave, you can outsource your decisions there too. Tens of billions of dollars (trillions if you look at market cap) later, this newish tool's direction to jour is turning out deterministic software (my day job). Suddenly the bottlenecks for value delivery aren't where they used to be. Drawing on tales from both my personal experience as a dev and others' on the infra-as-code side, we'll walk through where the bottlenecks now land...turns out, the hard part of building is still the hard part, it's just more obvious now. Then we'll figure out how to fit these new cybernetic exoskeletons in to deliver real, additional, value, end to end.

  • 16