Loading...
Loading...
LFM2-24B-A2B is the largest model in the LFM2 family of hybrid architectures designed for efficient on-device deployment. Built as a 24B parameter Mixture-of-Experts model with only 2B active parameters per token, it delivers high-quality generation while maintaining low inference costs. The model fits within 32 GB of RAM, making it practical to run on consumer laptops and desktops without sacrificing capability.
Price/1M
$0.05
200th cheapest
83% below median
Top 29%
Context Window
33K
291st largest
Top 91%
Input
$0.03
per 1M tokens
Output
$0.12
per 1M tokens
Blended
$0.05
per 1M tokens
Cheaper than 71% of models. Median price is $0.31/1M tokens.
Daily
$0.05
Monthly
$1.57
Context Window
33K
tokens
Larger than 9% of models
Context Window Comparison