Loading...
Loading...
A sophisticated text-based Mixture-of-Experts (MoE) model featuring 21B total parameters with 3B activated per token, delivering exceptional multimodal understanding and generation through heterogeneous MoE structures and modality-isolated routing. Supporting an extensive 131K token context length, the model achieves efficient inference via multi-expert parallel collaboration and quantization, while advanced post-training techniques including SFT, DPO, and UPO ensure optimized performance across diverse applications with specialized routing and balancing losses for superior task handling.
Price/1M
$0.12
241st cheapest
60% below median
Top 36%
Context Window
120K
267th largest
Top 76%
Input
$0.07
per 1M tokens
Output
$0.28
per 1M tokens
Blended
$0.12
per 1M tokens
Cheaper than 64% of models. Median price is $0.31/1M tokens.
Daily
$0.12
Monthly
$3.68
Context Window
120K
tokens
Larger than 24% of models
Max Output
8K
tokens
7% of context
Context Window Comparison