From datasets and feature engineering to context-based embeddings in the Transformer
| Task | Approach |
|---|---|
| Spam detection | Naive Bayes, SVMs |
| Image classification | HOG features + SVM |
| Recommendation | Matrix factorization |
| Speech | HMM + GMM |
Feature engineering was the art — and the limitation. A model was only as good as the human intuition behind its features.
"Instead of telling the model what to look for, we give it enough data and let it figure it out."
Input: "The bank by the river was ..."
↑
What does "bank" mean?
Context is spread across the sentence.
RNNs forget early context. Language needs global context awareness.
"The animal didn't cross the street because it was too tired"
↑
Attention links "it" → "animal"
Input Text
↓
[Tokenization + Positional Encoding]
↓
[Multi-Head Self-Attention Layer] × N
↓
[Feed-Forward Layer]
↓
[Layer Normalization + Residual Connections]
↓
Output Probabilities
1. Tokenization + Embeddings
2. Self-Attention (Q, K, V)
3. Positional Encoding
More parameters + more data + more compute = emergent capabilities
| Year | Model | Parameters |
|---|---|---|
| 2018 | BERT (Google) | 340M |
| 2020 | GPT-3 (OpenAI) | 175B |
| 2022 | ChatGPT | — (instruction tuned GPT-3.5) |
| 2023 | GPT-4, Claude, Llama | 100B–1T+ |
| 2024–25 | Claude 3.5/4, GPT-4o, Gemini | Multimodal, reasoning |
User Query
↓
[Embedding Model] → Vector Search → [Knowledge Base / Documents]
↓
Relevant Context Retrieved
↓
Context + Query → [LLM] → Grounded Answer
LLM (Claude, GPT...)
↓
[MCP Client]
↓
[MCP Server] ←→ [File System / GitHub / Slack / Database / APIs]
| Generation | Capability |
|---|---|
| LLM (2020) | Generate text responses |
| LLM + RAG (2022) | Answer with retrieved knowledge |
| LLM + Tools (2023) | Call APIs, run code, search web |
| Agentic AI (2024+) | Plan, act, self-correct, collaborate |
"An agent doesn't just answer — it acts."
Task → [LLM Step 1] → Output 1 → [LLM Step 2] → Output 2 → Final Result
Input → [Router LLM] → classify → [Specialist Agent A]
→ [Specialist Agent B]
→ [Specialist Agent C]
┌→ [Agent A: Research] ─┐
Input → [Splitter] ─┤→ [Agent B: Code] ├→ [Aggregator] → Result
└→ [Agent C: Verify] ─┘
[Orchestrator Agent]
↓ assigns sub-tasks
[Worker 1] [Worker 2] [Worker 3]
↓ results
[Orchestrator] → synthesizes → Final Output
[Generator Agent] → Output → [Evaluator Agent]
↑ ↓
└──── Feedback / Retry ─────┘
↓ (when quality threshold met)
Final Output
from langchain.agents import initialize_agent, Tool
from langchain.chat_models import ChatAnthropic
agent = initialize_agent(
tools=[search_tool, calculator_tool],
llm=ChatAnthropic(model="claude-sonnet-4-6"),
agent_type="zero-shot-react-description"
)
agent.run("Research the latest trends in AI and summarize them")
from crewai import Agent, Task, Crew
researcher = Agent(role="Researcher", goal="Find key AI trends")
writer = Agent(role="Writer", goal="Write a clear summary")
task = Task(description="Research and write about Agentic AI",
agent=writer)
crew = Crew(agents=[researcher, writer], tasks=[task])
crew.kickoff()
| Framework | Strength |
|---|---|
| AutoGen (Microsoft) | Multi-agent conversations |
| LlamaIndex | Advanced RAG & data pipelines |
| Semantic Kernel | Enterprise .NET/Python integration |
| Autogen Studio | Visual multi-agent builder |
| Person | Contribution |
|---|---|
| Geoffrey Hinton | Godfather of deep learning, neural nets |
| Yann LeCun | CNNs, Meta Chief AI Scientist |
| Sam Altman | Driving GPT/ChatGPT to mass adoption |
| Dario Amodei | Anthropic CEO, AI safety focus |
| Ilya Sutskever | Co-founder OpenAI, scaling laws |
| Harrison Chase | Creator of LangChain |
| Dimension | ML Era (2012) | Agentic AI Era (2025) |
|---|---|---|
| Input | Structured datasets | Natural language |
| Feature Engineering | Manual, domain expertise | Learned automatically |
| Model Role | Predict a label | Reason, plan, and act |
| Human involvement | Every step | High-level goal setting |
| Knowledge | Frozen in weights | Dynamic via RAG + tools |
| Output | Classification / number | Code, reports, decisions |
| Collaboration | Single model | Multi-agent systems |
2010: Feature Engineering + SVMs
↓
2012: Deep Learning — CNNs (AlexNet)
↓
2014: RNNs, LSTMs — Sequential Learning
↓
2017: Transformer — "Attention Is All You Need"
↓
2020: Large Language Models — GPT-3, BERT
↓
2022: Instruction Tuning + RLHF — ChatGPT
↓
2023: LLMs + Tools + RAG — Grounded AI
↓
2024: MCP + Agentic Patterns — Action-taking AI
↓
2025+: Multi-Agent Systems — Collaborative AI
"We went from teaching machines to recognize cats, to building systems that can think, plan, and act in the world."
Key Takeaways:
Presentation prepared February 2026