Price Per TokenPrice Per Token
Meta-llama
Meta-llama
vs
Meta-llama
Meta-llama

Llama 3.1 8B Instruct vs Llama 3.3 70B Instruct

A detailed comparison of pricing, benchmarks, and capabilities

OpenClaw

Best LLMs for OpenClaw Vote for which model works best with OpenClaw

112 out of our 301 tracked models have had a price change in February.

Get our weekly newsletter on pricing changes, new releases, and tools.

Key Takeaways

Llama 3.1 8B Instruct wins:

  • Cheaper input tokens
  • Cheaper output tokens
  • Faster response time

Llama 3.3 70B Instruct wins:

  • Larger context window
  • Higher intelligence benchmark
  • Better at coding
  • Better at math
Price Advantage
Llama 3.1 8B Instruct
Benchmark Advantage
Llama 3.3 70B Instruct
Context Window
Llama 3.3 70B Instruct
Speed
Llama 3.1 8B Instruct

Pricing Comparison

Price Comparison

MetricLlama 3.1 8B InstructLlama 3.3 70B InstructWinner
Input (per 1M tokens)$0.02$0.10 Llama 3.1 8B Instruct
Output (per 1M tokens)$0.05$0.32 Llama 3.1 8B Instruct
Using a 3:1 input/output ratio, Llama 3.1 8B Instruct is 82% cheaper overall.

Llama 3.1 8B Instruct Providers

Nebius $0.02 (Cheapest)
DeepInfra $0.02 (Cheapest)
Novita $0.02 (Cheapest)
Groq $0.05
SiliconFlow $0.06

Llama 3.3 70B Instruct Providers

DeepInfra $0.10 (Cheapest)
Novita $0.14
Parasail $0.22
Nebius $0.25
Crusoe $0.25

Benchmark Comparison

8
Benchmarks Compared
0
Llama 3.1 8B Instruct Wins
8
Llama 3.3 70B Instruct Wins

Benchmark Scores

BenchmarkLlama 3.1 8B InstructLlama 3.3 70B InstructWinner
Intelligence Index
Overall intelligence score
11.714.2
Coding Index
Code generation & understanding
4.910.7
Math Index
Mathematical reasoning
4.37.7
MMLU Pro
Academic knowledge
47.671.3
GPQA
Graduate-level science
25.949.8
LiveCodeBench
Competitive programming
11.628.8
Aider
Real-world code editing
37.659.4
AIME
Competition math
7.730.0
Llama 3.3 70B Instruct significantly outperforms in coding benchmarks.

Cost vs Quality

X-axis:
Y-axis:
Loading chart...
Llama 3.1 8B Instruct
Other models

Context & Performance

Context Window

Llama 3.1 8B Instruct
16,384
tokens
Max output: 16,384 tokens
Llama 3.3 70B Instruct
131,072
tokens
Max output: 16,384 tokens
Llama 3.3 70B Instruct has a 88% larger context window.

Speed Performance

MetricLlama 3.1 8B InstructLlama 3.3 70B InstructWinner
Tokens/second162.2 tok/s104.4 tok/s
Time to First Token0.33s0.49s
Llama 3.1 8B Instruct responds 55% faster on average.

Capabilities

Feature Comparison

FeatureLlama 3.1 8B InstructLlama 3.3 70B Instruct
Vision (Image Input)
Tool/Function Calls
Reasoning Mode
Audio Input
Audio Output
PDF Input
Prompt Caching
Web Search

License & Release

PropertyLlama 3.1 8B InstructLlama 3.3 70B Instruct
LicenseOpen SourceOpen Source
AuthorMeta-llamaMeta-llama
ReleasedJul 2024Dec 2024

Llama 3.1 8B Instruct Modalities

Input
text
Output
text

Llama 3.3 70B Instruct Modalities

Input
text
Output
text

Related Comparisons

Compare Llama 3.1 8B Instruct with:

Compare Llama 3.3 70B Instruct with:

Frequently Asked Questions

Llama 3.1 8B Instruct has cheaper input pricing at $0.02/M tokens. Llama 3.1 8B Instruct has cheaper output pricing at $0.05/M tokens.
Llama 3.3 70B Instruct scores higher on coding benchmarks with a score of 10.7, compared to Llama 3.1 8B Instruct's score of 4.9.
Llama 3.1 8B Instruct has a 16,384 token context window, while Llama 3.3 70B Instruct has a 131,072 token context window.
Llama 3.1 8B Instruct does not support vision. Llama 3.3 70B Instruct does not support vision.