Loading...
Loading...
Qwen2.5-Coder-7B-Instruct is a 7B parameter instruction-tuned language model optimized for code-related tasks such as code generation, reasoning, and bug fixing. Based on the Qwen2.5 architecture, it incorporates enhancements like RoPE, SwiGLU, RMSNorm, and GQA attention with support for up to 128K tokens using YaRN-based extrapolation. It is trained on a large corpus of source code, synthetic data, and text-code grounding, providing robust performance across programming languages and agentic coding workflows. This model is part of the Qwen2.5-Coder family and offers strong compatibility with tools like vLLM for efficient deployment. Released under the Apache 2.0 license.
Quality Index
10.0
364th of 442
Top 83%
Price/1M
$0.00
1st cheapest
100% below median
Top 27%
Speed
0 tok/s
TTFT
0.00s
Context Window
33K
291st largest
Top 91%
Input
$0.00
per 1M tokens
Output
$0.00
per 1M tokens
Blended
$0.00
per 1M tokens
Cheaper than 73% of models. Median price is $0.31/1M tokens.
Daily
$0.00
Monthly
$0.00
0
tokens/sec
Faster than 0% of models
0.00
seconds
Faster than 61% of models
0.00
seconds
Faster than 61% of models
Market Median
46 tok/s
100% slower
Median TTFT
0.42s
100% faster
Speed Comparison
Context Window
33K
tokens
Larger than 9% of models
2.5M
669
8-16 GB
RTX 4070 / M2 Pro