Real World Agents: Case Studies from LlamaIndex

Jason Liu webinar 2025-10-08

What are we talking about?

  • What is LlamaIndex?
  • What are document agents?
  • Case studies
  • Takeaways

What is LlamaIndex?

 

LlamaIndex core = RAG

LlamaIndex workflows = Agents

 

developers.llamaindex.ai

LlamaCloud

RAG as a service

 

cloud.llamaindex.ai

LlamaParse

  • The best document parser in the world
  • Free for 10,000 pages/month

Why use LlamaIndex and LlamaCloud?

  • Solve real business problems
  • Avoid technical boilerplate
  • Focus on business value

What can you actually build?

AI agents

+

RAG systems

=

Document agents

RAG is essential to the enterprise

  • LLMs need your data to do anything useful
  • Context windows are limited
  • Speed and cost are important

RAG by itself

is insufficient

  • Semantic search has limits
  • Fails at complex, multi-part questions
  • Needs planning and comparison

RAG needs Agents, Agents need RAG

  • RAG needs agents to be good
  • Agents need RAG to be useful
  • = Document Agents

Case studies #1

  • Domain: Government construction RFPs
  • $7 billion market in Poland
  • 100+ page technical documents

What they were doing

 

  • Hours or days per RFP
  • Keyword search for discovery
  • False positives and missed opportunities

What they built

  • Agent workflow mimicking human process
  • 20-30 page standardized reports
  • Executive summaries, risk assessments, recommendations

Outcomes

 

  • Semantic search finds more relevant RFPs
  • 10 minutes average processing time
  • 3 → 20-30 tenders per employee per day

Bullseye use case

for LLMs

Key takeaway #1

Workflows already existed

Key takeaway #2

Case studies #2

Pursuit

  • Domain: Public sector business intelligence
  • 90,000+ government entities
  • Budgets, plans, transcripts

 

What they were doing

  • Manual tracking across thousands of sites
  • Impossible to stay current
  • Painstaking manual review
  • Most opportunities never discovered

 

What they built

  • 4 million pages parsed in one weekend
  • Extracts from tables, figures, scanned images
  • Searchable and filterable across all entities

 

Outcomes

  • 25-30% increase in accuracy
  • Previously invisible opportunities now visible
  • Entire public sector now searchable

 

Scale was outside human capability

Key takeaway #1

Accuracy across messy real-world documents

Key takeaway #2

Outcomes were well-defined

Key takeaway #3

Case studies #3

Scaleport AI

  • Domain: Travel insurance claims
  • Medical reports from around the world
  • Handwriting, notes, wild formatting

 

What they were doing

  • 20-40 minutes per claim
  • Manual extraction by adjusters
  • Hundreds of claims per month
  • No way to scale without hiring

 

What they built

  • LlamaParse for document extraction
  • Multiple automated checks
  • Structured analysis for adjuster review

Outcomes

  • 20-40 minutes → 10 minutes
  • 50-75% improvement
  • 2x processing capacity
  • Augmentation, not replacement

 

Parsing was critical

Key takeaway #1

Full-workflow system

Key takeaway #2

Human-AI collaboration

Key takeaway #3

Case studies #4

Arcee AI

  • Domain: Academic research papers
  • 4 million pages of NLP research
  • Dense PDFs with equations, tables, charts

 

What they were doing

  • Traditional OCR failed
  • Tables mangled, equations lost
  • Manual extraction impractical at scale

 

What they built

  • LlamaParse with parsing instructions
  • Iterative refinement through prompts
  • High-fidelity structured extraction

 

Outcomes

  • Minimal data loss
  • Tables, equations, context preserved
  • High-quality dataset in fraction of time
  • Better model performance

 

More than just text

Key takeaway #1

Parsing instructions

Key takeaway #2

Production scale

Key takeaway #3

Case studies #5

11x.ai

  • Domain: AI SDR onboarding
  • Training AI agents on product offerings
  • Multimodal content in multiple formats

 

What they were doing

  • Manual onboarding per product
  • Weeks or months for enterprise customers
  • Evaluated building in-house pipeline
  • Massive engineering effort, quality issues

 

What they built

  • LlamaParse for multimodal ingestion
  • Audio, PDFs, Word docs, web pages
  • Automated knowledge extraction
  • Campaign messaging generated at scale

 

Outcomes

  • Weeks → days onboarding time
  • Prototype → production in 3 days
  • Immediate adoption after launch

 

Buy instead of build

Key takeaway #1

TypeScript support!

Key takeaway #2

Fine-grained control

Key takeaway #3

Document agents are domain specific

  • Tons of unstructured data
  • LlamaIndex engineered for this
  • Not generic agents

 

Workflows are critical

  • Direct reflection of business processes
  • Know your business before automating
  • Hard-code processes for reliability

 

Parsing, parsing, parsing

  • Quality is paramount
  • Makes the difference in outcomes
  • Why we talk about LlamaParse all the time

Thank you!

You can follow me on BlueSky: @seldo.com

Real World Agents: Case Studies from LlamaIndex

By Laurie Voss

Real World Agents: Case Studies from LlamaIndex

  • 10