Claude Sonnet 4.6
Active44.8BasicAnthropic Claude Sonnet 4.6.
Intelligence Index
44.8/ 100
Basicweighted across 6 benchmarks
- Coding
- 72.7
- Factual grounding
- 63.6
- Instruction following
- 41.7
- Long context
- 28.2
- Medical
- 17.7
Computed as the mean of per-category averages across MMLU, GPQA, SWE-bench, HumanEval, MATH, GSM8K, AIME, Aider Polyglot and more. See each for methodology.
History
Claude Sonnet 4.6 became available via the Anthropic API on 2026-02-17.
Training & availability
Anthropic has not released the underlying model weights — access is via their hosted API only.
Capabilities
-
Context window: 1.0M tokens.
-
Max output: 64K tokens.
-
Input modalities: text, image.
Recommended for: vision, agentic, long-context, code.
Example interactions
Curated prompts showing the model's response style — not cherry-picked to impress, picked to show what typical output looks like.
Quick start
Minimal example using the anthropic API. Copy, paste, replace the key.
from anthropic import Anthropic
client = Anthropic(api_key="sk-ant-...")
resp = client.messages.create(
model="claude-sonnet-4-6",
max_tokens=1024,
messages=[{"role": "user", "content": "Explain quantum computing in one sentence."}],
)
print(resp.content[0].text)Benchmarks
| Benchmark | Score | Source |
|---|---|---|
| FACTS GroundingFactual grounding | 63.64% | Third-party llm-stats.com |
| HealthBenchMedical | 17.74% | Third-party llm-stats.com |
| LongBench v2Long context | 5.56% | Third-party llm-stats.com |
Integrations & tooling support
- Tool calling
- Supported
- Structured outputs
- Supported
Price vs quality
Solid benchmark performance. Pricing not publicly available — check the provider.
- Quality percentile
- 53.3%
- Effective price
- —
- Pricing breakdown
- — in
— out
Community ratings
Rate Claude Sonnet 4.6
Sign in to rate and review.
Comments
Sign in to leave a comment.