Price Per TokenPrice Per Token
OpenAI
OpenAI
vs
Qwen
Qwen

GPT-OSS-120b vs Qwen2.5 VL 32B Instruct

A detailed comparison of pricing, benchmarks, and capabilities

OpenClaw

Best LLMs for OpenClaw Vote for which model works best with OpenClaw

112 out of our 301 tracked models have had a price change in February.

Get our weekly newsletter on pricing changes, new releases, and tools.

Key Takeaways

GPT-OSS-120b wins:

  • Cheaper input tokens
  • Cheaper output tokens
  • Larger context window
  • Faster response time
  • Higher intelligence benchmark
  • Better at coding
  • Better at math

Qwen2.5 VL 32B Instruct wins:

  • Supports vision
  • Supports tool calls
Price Advantage
GPT-OSS-120b
Benchmark Advantage
GPT-OSS-120b
Context Window
GPT-OSS-120b
Speed
GPT-OSS-120b

Pricing Comparison

Price Comparison

MetricGPT-OSS-120bQwen2.5 VL 32B InstructWinner
Input (per 1M tokens)$0.04$0.05 GPT-OSS-120b
Output (per 1M tokens)$0.19$0.22 GPT-OSS-120b
Cache Read (per 1M)N/A$25000.00 Qwen2.5 VL 32B Instruct
Using a 3:1 input/output ratio, GPT-OSS-120b is 17% cheaper overall.

GPT-OSS-120b Providers

Chutes $0.04 (Cheapest)
SiliconFlow $0.05
Novita $0.05
Clarifai $0.09
Google $0.09

Qwen2.5 VL 32B Instruct Providers

Chutes $0.05 (Cheapest)
DeepInfra $0.20

Benchmark Comparison

8
Benchmarks Compared
4
GPT-OSS-120b Wins
0
Qwen2.5 VL 32B Instruct Wins

Benchmark Scores

BenchmarkGPT-OSS-120bQwen2.5 VL 32B InstructWinner
Intelligence Index
Overall intelligence score
33.313.2
Coding Index
Code generation & understanding
28.6--
Math Index
Mathematical reasoning
93.4--
MMLU Pro
Academic knowledge
80.869.7
GPQA
Graduate-level science
78.246.6
LiveCodeBench
Competitive programming
87.824.8
Aider
Real-world code editing
41.8--
AIME
Competition math
-11.0-
GPT-OSS-120b significantly outperforms in coding benchmarks.

Cost vs Quality

X-axis:
Y-axis:
Loading chart...
GPT-OSS-120b
Other models

Context & Performance

Context Window

GPT-OSS-120b
131,072
tokens
Qwen2.5 VL 32B Instruct
16,384
tokens
Max output: 16,384 tokens
GPT-OSS-120b has a 88% larger context window.

Speed Performance

MetricGPT-OSS-120bQwen2.5 VL 32B InstructWinner
Tokens/second311.5 tok/s0.0 tok/s
Time to First Token0.47s0.00s
GPT-OSS-120b responds Infinity% faster on average.

Capabilities

Feature Comparison

FeatureGPT-OSS-120bQwen2.5 VL 32B Instruct
Vision (Image Input)
Tool/Function Calls
Reasoning Mode
Audio Input
Audio Output
PDF Input
Prompt Caching
Web Search

License & Release

PropertyGPT-OSS-120bQwen2.5 VL 32B Instruct
LicenseOpen SourceOpen Source
AuthorOpenAIQwen
ReleasedAug 2025Mar 2025

GPT-OSS-120b Modalities

Input
text
Output
text

Qwen2.5 VL 32B Instruct Modalities

Input
textimage
Output
text

Related Comparisons

Compare GPT-OSS-120b with:

Compare Qwen2.5 VL 32B Instruct with:

Frequently Asked Questions

GPT-OSS-120b has cheaper input pricing at $0.04/M tokens. GPT-OSS-120b has cheaper output pricing at $0.19/M tokens.
GPT-OSS-120b scores higher on coding benchmarks with a score of 28.6, compared to Qwen2.5 VL 32B Instruct's score of N/A.
GPT-OSS-120b has a 131,072 token context window, while Qwen2.5 VL 32B Instruct has a 16,384 token context window.
GPT-OSS-120b does not support vision. Qwen2.5 VL 32B Instruct supports vision.