Price Per TokenPrice Per Token
Meta-llama
Meta-llama
vs
Meta-llama
Meta-llama

Llama 3.1 405B Instruct vs Llama 3.2 11B Vision Instruct

A detailed comparison of pricing, benchmarks, and capabilities

OpenClaw

Best LLMs for OpenClaw Vote for which model works best with OpenClaw

112 out of our 301 tracked models have had a price change in February.

Get our weekly newsletter on pricing changes, new releases, and tools.

Key Takeaways

Llama 3.1 405B Instruct wins:

  • Higher intelligence benchmark
  • Better at coding
  • Better at math

Llama 3.2 11B Vision Instruct wins:

  • Cheaper input tokens
  • Cheaper output tokens
  • Larger context window
  • Faster response time
  • Supports vision
Price Advantage
Llama 3.2 11B Vision Instruct
Benchmark Advantage
Llama 3.1 405B Instruct
Context Window
Llama 3.2 11B Vision Instruct
Speed
Llama 3.2 11B Vision Instruct

Pricing Comparison

Price Comparison

MetricLlama 3.1 405B InstructLlama 3.2 11B Vision InstructWinner
Input (per 1M tokens)$4.00$0.05 Llama 3.2 11B Vision Instruct
Output (per 1M tokens)$4.00$0.05 Llama 3.2 11B Vision Instruct
Using a 3:1 input/output ratio, Llama 3.2 11B Vision Instruct is 99% cheaper overall.

Llama 3.1 405B Instruct Providers

Hyperbolic $4.00 (Cheapest)
Google $5.00

Llama 3.2 11B Vision Instruct Providers

Cloudflare $0.05 (Cheapest)
DeepInfra $0.05 (Cheapest)
Novita $0.06
Together $0.18

Benchmark Comparison

8
Benchmarks Compared
7
Llama 3.1 405B Instruct Wins
0
Llama 3.2 11B Vision Instruct Wins

Benchmark Scores

BenchmarkLlama 3.1 405B InstructLlama 3.2 11B Vision InstructWinner
Intelligence Index
Overall intelligence score
14.210.9
Coding Index
Code generation & understanding
14.54.3
Math Index
Mathematical reasoning
3.01.7
MMLU Pro
Academic knowledge
73.246.4
GPQA
Graduate-level science
51.522.1
LiveCodeBench
Competitive programming
30.511.0
Aider
Real-world code editing
66.2--
AIME
Competition math
21.39.3
Llama 3.1 405B Instruct significantly outperforms in coding benchmarks.

Cost vs Quality

X-axis:
Y-axis:
Loading chart...
Llama 3.1 405B Instruct
Other models

Context & Performance

Context Window

Llama 3.1 405B Instruct
131,000
tokens
Llama 3.2 11B Vision Instruct
131,072
tokens
Max output: 16,384 tokens
Llama 3.2 11B Vision Instruct has a 0% larger context window.

Speed Performance

MetricLlama 3.1 405B InstructLlama 3.2 11B Vision InstructWinner
Tokens/second25.2 tok/s69.7 tok/s
Time to First Token0.79s0.41s
Llama 3.2 11B Vision Instruct responds 177% faster on average.

Capabilities

Feature Comparison

FeatureLlama 3.1 405B InstructLlama 3.2 11B Vision Instruct
Vision (Image Input)
Tool/Function Calls
Reasoning Mode
Audio Input
Audio Output
PDF Input
Prompt Caching
Web Search

License & Release

PropertyLlama 3.1 405B InstructLlama 3.2 11B Vision Instruct
LicenseOpen SourceOpen Source
AuthorMeta-llamaMeta-llama
ReleasedJul 2024Sep 2024

Llama 3.1 405B Instruct Modalities

Input
text
Output
text

Llama 3.2 11B Vision Instruct Modalities

Input
textimage
Output
text

Related Comparisons

Compare Llama 3.1 405B Instruct with:

Compare Llama 3.2 11B Vision Instruct with:

Frequently Asked Questions

Llama 3.2 11B Vision Instruct has cheaper input pricing at $0.05/M tokens. Llama 3.2 11B Vision Instruct has cheaper output pricing at $0.05/M tokens.
Llama 3.1 405B Instruct scores higher on coding benchmarks with a score of 14.5, compared to Llama 3.2 11B Vision Instruct's score of 4.3.
Llama 3.1 405B Instruct has a 131,000 token context window, while Llama 3.2 11B Vision Instruct has a 131,072 token context window.
Llama 3.1 405B Instruct does not support vision. Llama 3.2 11B Vision Instruct supports vision.