Dario Amodei — CEO, former VP of Research at OpenAI, deep expertise in AI scaling laws
Daniela Amodei — President, former VP of Operations at OpenAI, business and operations leadership
Chris Olah — Co-founder, pioneer in neural network interpretability
Tom Brown — Co-founder, lead author on GPT-3 paper
Jared Kaplan — Co-founder, scaling laws researcher
Sam McCandlish — Co-founder, key contributor to scaling research
Research Focus
AI Safety & Alignment
Interpretability (Mechanistic)
Constitutional AI
Scaling Laws
03
The Claude Model Family
Anthropic offers three tiers of Claude models, each balancing capability, speed, and cost.
Claude Haiku 4.5
FastAffordable
Fastest model in the family
Near-instant responses
Ideal for high-volume tasks
Classification, extraction, routing
Cost-effective at scale
Claude Sonnet 4.6
BalancedVersatile
Best balance of speed and intelligence
Strong coding capabilities
Excellent for most production tasks
Powers Claude.ai (default)
Great for agentic workflows
Claude Opus 4.6
PowerfulDeep
Most capable model
Best for complex reasoning
Extended thinking support
Research & analysis tasks
Highest accuracy
Feature
Haiku 4.5
Sonnet 4.6
Opus 4.6
Context Window
200K tokens
1M tokens
1M tokens
Max Output
8,192 tokens
16,384 tokens
32,768 tokens
Speed
Fastest
Moderate
Slower
Reasoning
Good
Strong
Best-in-class
Extended Thinking
No
Yes
Yes (advanced)
Vision
Yes
Yes
Yes
Tool Use
Yes
Yes
Yes
04
Constitutional AI
Claude is trained differently from most LLMs. Anthropic developed Constitutional AI (CAI) to align models with human values without relying solely on human feedback for safety.
How CAI Works
A set of principles (the "constitution") guides model behavior
The model critiques its own outputs against these principles
It then revises to better align with the constitution
This is combined with RLHF from human feedback
Why It Matters
Reduces reliance on human labelers for safety
More scalable alignment approach
Transparent — principles are published
Model learns to reason about values, not just memorize rules
05
Claude's Core Capabilities
Reasoning & Analysis
Multi-step logical reasoning
Complex problem decomposition
Nuanced judgment calls
Extended thinking for hard problems
Chain-of-thought processing
Coding & Development
Code generation in 20+ languages
Bug detection and fixing
Code review and refactoring
Architecture design
Test writing and documentation
Writing & Communication
Long-form content creation
Tone and style adaptation
Editing and proofreading
Summarization at any length
Translation and localization
Math & Science
Advanced mathematical reasoning
Statistical analysis
Scientific literature review
Formula derivation
Data interpretation
Vision & Multimodal
Image understanding and analysis
Chart and graph interpretation
Document and PDF processing
Handwriting recognition
Visual question answering
Multilingual
Fluent in 30+ languages
Cross-language tasks
Cultural context awareness
Code-switching support
Translation with nuance
06
Benchmark Performance
Claude Opus 4.6 achieves state-of-the-art or near-state-of-the-art results across major benchmarks. Numbers are approximate and from publicly available reports.
Benchmark
Claude Opus 4.6
GPT-4o
Gemini 2.5 Pro
MMLU (5-shot)
89.5%
88.7%
89.2%
GPQA (Diamond)
65.2%
53.6%
63.8%
HumanEval (Coding)
93.0%
90.2%
91.4%
MATH (Hard)
78.3%
76.6%
80.1%
GSM8K
97.8%
97.1%
96.8%
ARC-Challenge
96.7%
96.4%
96.1%
SWE-Bench Verified
72.5%
38.4%
63.8%
TAU-Bench (Airline)
68.0%
52.0%
58.0%
Strongest Areas
Claude particularly excels in agentic coding (SWE-bench), graduate-level reasoning (GPQA), and real-world tool use (TAU-bench).
Note on Benchmarks
Benchmark scores are a useful but imperfect measure. Real-world performance, safety, and user experience matter as much as leaderboard positions. Scores evolve rapidly.
07
Context Windows
Claude supports a 200K–1M token context window depending on model. Opus 4.6 and Sonnet 4.6 support 1M tokens at standard pricing, while Haiku 4.5 supports 200K.
How Context Works
Context = input tokens + output tokens
Everything in the conversation is "in context"
Claude attends to the entire window simultaneously
No sliding window or summarization — full attention
Extended Thinking
Sonnet & Opus can use extended thinking
Model reasons step-by-step before answering
Dramatically improves accuracy on hard problems
Thinking tokens are separate from output tokens
Configurable via effort parameter (recommended for 4.6 models) or budget_tokens
Prompt Caching
Reuse large context prefixes across calls for up to 90% cost reduction and faster time-to-first-token.
08
The Claude Ecosystem
09
Pricing & Plans
Claude is available across multiple tiers for individuals and organizations.
Plan
Price
Models
Features
Free
$0
Sonnet 4.6
Basic usage with rate limits, web & mobile access
Pro
$20/mo
Sonnet, Opus, Haiku
5x more usage, extended thinking, Projects, priority access
Max
$100–$200/mo
All models
20x more usage than Pro, extended thinking at highest limits, longer outputs
Team
$25/user/mo
All models
Higher limits, admin controls, team sharing, longer context
Enterprise
Custom
All models
SSO/SAML, SCIM, audit logs, custom retention, SLA, dedicated support
Free Tier Highlights
No credit card required
Access to Claude Sonnet
File uploads and vision
Limited daily messages
Pro Tier Value
Access to all models including Opus
Extended thinking for deep reasoning
Projects with custom instructions
Priority during high-traffic periods
10
API Pricing
Anthropic's API uses per-token pricing. Input tokens (your prompt) and output tokens (Claude's response) are billed separately.
Model
Input (per 1M tokens)
Output (per 1M tokens)
Prompt Caching (Input)
Haiku 4.5
$0.80
$4.00
$0.08 (90% saving)
Sonnet 4.6
$3.00
$15.00
$0.30 (90% saving)
Opus 4.6
$15.00
$75.00
$1.50 (90% saving)
Cost Optimization
Prompt caching — reuse context prefixes
Batch API — 50% discount for async jobs
Choose the right model tier for the task
What's a Token?
~4 characters = 1 token (English)
~0.75 words = 1 token
1M tokens ≈ 750K words
A typical page ≈ 300-400 tokens
Rate Limits
Tiered by usage level (1-4)
Requests per minute (RPM)
Tokens per minute (TPM)
Increase with spend history
11
Safety & Alignment
Safety is not a feature of Claude — it is the core mission of Anthropic. The company was founded specifically to pursue AI safety research.
The HHH Framework
Helpful — genuinely assist users with their tasks
Harmless — avoid causing harm, even when asked to
Honest — be truthful and transparent about limitations
Responsible Scaling Policy
AI Safety Levels (ASL) framework
Commitments tied to model capabilities
Independent safety evaluations
Red-teaming before deployment
Phased rollouts with monitoring
Interpretability Research
Mechanistic interpretability — understanding how models think
Feature identification inside neural networks
Published groundbreaking research on "features" in LLMs
Goal: make AI systems transparent and auditable
Trust & Safety Controls
Usage policies with clear guidelines
Content filtering and moderation layers
No training on customer API data
SOC 2 Type II compliance
HIPAA eligibility (Enterprise)
12
Claude vs Competitors
A balanced comparison of leading AI assistants as of early 2026. Each has unique strengths.
Dimension
Claude (Anthropic)
GPT-4o (OpenAI)
Gemini 2.5 (Google)
Context Window
200K–1M tokens
128K tokens
1M tokens
Safety Approach
Constitutional AI
RLHF + rules
RLHF + filters
Coding
Excellent (SWE-bench leader)
Strong
Strong
Reasoning
Excellent
Strong
Strong
Multimodal
Vision input
Vision + audio + image gen
Vision + audio + video
Long-form Writing
Best-in-class
Good
Good
Instruction Following
Excellent
Very Good
Good
Web Search
Limited (via tools)
Built-in
Built-in
Ecosystem
Growing
Mature (plugins, GPTs)
Google integrated
Open Source
No
No
Gemma (partial)
Key Takeaway
Claude excels in safety, coding, instruction following, and long-form writing. Competitors may lead in multimodal breadth (audio, video, image generation) and ecosystem maturity. The landscape evolves rapidly.
13
Use Cases Overview
Enterprise
Customer support automation
Internal knowledge Q&A
Document processing at scale
Report generation
Data extraction & classification
Software Development
Code generation & completion
Debugging & code review
Architecture consulting
Migration & refactoring
Documentation writing
Research & Analysis
Literature review
Data analysis & interpretation
Hypothesis generation
Summarizing papers
Research methodology review
Creative & Content
Marketing copy & ad writing
Blog posts & articles
Creative fiction & brainstorming
Social media content
Script and speech writing
Education
Personalized tutoring
Curriculum development
Assignment feedback
Concept explanation
Study guide creation
Legal & Finance
Contract review
Regulatory compliance
Financial report analysis
Risk assessment
Due diligence support
14
What Makes Claude Different
Personality & Tone
Thoughtful and measured responses
Willing to express uncertainty
Avoids sycophantic agreement
Genuine intellectual curiosity
Adapts tone to context naturally
Instruction Following
Precisely follows complex multi-step instructions
Respects constraints (word count, format, style)
System prompts are deeply respected
Handles edge cases in instructions gracefully
Nuance & Depth
Acknowledges complexity and trade-offs
Presents multiple perspectives
Does not oversimplify difficult topics
Provides caveats when appropriate
Comfortable with ambiguity
Long-Form Excellence
Maintains coherence across very long outputs
Consistent style and voice throughout
Strong structural organization
Does not "forget" earlier context or instructions
"Claude is designed to be an assistant that genuinely engages with ideas rather than just producing plausible text. It reasons about problems rather than pattern-matching to expected answers."
15
Limitations & Considerations
Like all LLMs, Claude has important limitations that users should understand.
Known Limitations
Hallucinations — can generate plausible but incorrect information, especially for obscure facts
Knowledge cutoff — training data has a cutoff date; Claude may not know about very recent events
No internet access — cannot browse the web or access live data (without tool use)
Math errors — can make arithmetic mistakes, especially with large numbers
No persistent memory — each conversation starts fresh (no cross-session memory by default)
Rate Limits
Free tier: limited daily messages
Pro tier: significantly higher but not unlimited
API: tiered rate limits based on usage history
High-demand periods may see temporary throttling
Best Practices
Verify important facts independently
Use Claude as a collaborator, not sole authority
Be specific in your instructions
Provide context and examples when possible
Break complex tasks into steps
16
The Claude Timeline
Rapid Evolution
In under 3 years, Claude went from initial release to a family of models competing at the frontier of AI capabilities, with each generation bringing major improvements.
Key Milestones
Notable jumps include the 200K context window (Claude 3), Sonnet's coding breakthrough (3.5), Computer Use, and the Claude 4 generation with extended thinking and Claude Code.
17
Getting the Most from Claude
A preview of effective prompting strategies — covered in depth in Presentation 3: Prompt Engineering.
Be Specific & Clear
Instead of "Tell me about Python", try:
"Explain Python's asyncio module, focusing on when to use async/await vs threading, with a code example for each approach."
Give Context
Explain your role and audience
Provide relevant background
Share examples of desired output
State the format you want
Use System Prompts
System prompts set persistent behavior. Define Claude's role, tone, constraints, and output format upfront for consistent results.
Structure Complex Tasks
Use XML tags to organize sections
Number your requirements
Use markdown formatting in prompts
Break large tasks into smaller steps
Common Mistakes
Too vague — "make this better"
Too many tasks in one prompt
No context about audience or purpose
Not specifying output format
Ignoring Claude's request for clarification
Iterate & Refine
Prompting is a conversation. If the first result is not right, provide feedback and ask Claude to adjust rather than starting over.
18
The Series Ahead
This is Presentation 1 of 10 in the Anthropic Claude Series. Here is what is coming next:
01Introduction to Claude
You are here — overview of Claude, Anthropic, and the ecosystem.
02 Claude's Architecture & Training
Deep dive into how Claude is built, trained, and fine-tuned.
03 Prompt Engineering
Master the art of effective prompting for optimal results.
04 The Claude API
Building applications with Claude — endpoints, SDKs, and patterns.
05 Tool Use & Function Calling
Extending Claude's capabilities with external tools and APIs.
06 Claude for Coding
Claude Code, IDE integrations, and agentic development workflows.
07 Enterprise Deployment
Scaling Claude in production — security, compliance, and architecture.
08 Advanced Techniques
Extended thinking, multi-turn strategies, and complex workflows.
09 Safety & Alignment Deep Dive
Constitutional AI, interpretability, and responsible deployment.
10 The Future of Claude
Roadmap, emerging capabilities, and the path to AGI safety.
19
Summary & Resources
Key Takeaways
Claude is Anthropic's family of AI models — Haiku, Sonnet, and Opus
Built with Constitutional AI — a novel safety-first alignment approach
Excels at reasoning, coding, writing, and analysis
200K–1M token context window (up to 1M on Sonnet & Opus 4.6)
Available via Claude.ai, API, and Claude Code
Competitive performance with strong safety guarantees
Flexible pricing from free to enterprise
Next Up
Presentation 2: Claude's Architecture & Training — understanding the technical foundations behind Claude's capabilities.