Loading...
Loading...
DeepSeek-V3.1 Terminus is an update to [DeepSeek V3.1](/deepseek/deepseek-chat-v3.1) that maintains the model's original capabilities while addressing issues reported by users, including language consistency and agent capabilities, further optimizing the model's performance in coding and search agents. It is a large hybrid reasoning model (671B parameters, 37B active) that supports both thinking and non-thinking modes. It extends the DeepSeek-V3 base with a two-phase long-context training process, reaching up to 128K tokens, and uses FP8 microscaling for efficient inference. Users can control the reasoning behaviour with the `reasoning` `enabled` boolean. [Learn more in our docs](https://openrouter.ai/docs/use-cases/reasoning-tokens#enable-reasoning-with-default-config) The model improves tool use, code generation, and reasoning efficiency, achieving performance comparable to DeepSeek-R1 on difficult benchmarks while responding more quickly. It supports structured tool calling, code agents, and search agents, making it suitable for research, coding, and agentic workflows.
Índice de Qualidade
28.5
110th de 442
Top 25%
Índice de Código
31.9
65th de 352
Top 19%
Índice de Matemática
53.7
132nd de 268
Top 50%
Preço/1M
$0.63
409th mais barato
102% acima da mediana
Top 60%
Velocidade
0 tok/s
TTFT
0.00s
Janela de Contexto
164K
135th maior
Top 41%
Entrada
$0.34
por 1M tokens
Saída
$1.50
por 1M tokens
Combinado
$0.63
por 1M tokens
Mais barato que 40% dos modelos. Preço mediano é $0.31/1M tokens.
Diário
$0.63
Mensal
$18.78
0
tokens/seg
Mais rápido que 0% dos modelos
0.00
segundos
Mais rápido que 61% dos modelos
0.00
segundos
Mais rápido que 61% dos modelos
Mediana do Mercado
46 tok/s
100% mais lento
TTFT Mediano
0.42s
100% mais rápido
Vazão/Dólar
0
tok/s por $/1M
Comparação de Velocidade
Janela de Contexto
164K
tokens
Maior que 59% dos modelos
Saída Máxima
66K
tokens
40% do contexto