Description
LFM2-8B-A1B is an efficient on-device Mixture-of-Experts (MoE) model from Liquid AI’s LFM2 family, built for fast, high-quality inference on edge hardware. It uses 8.3B total parameters with only ~1.5B active per token, delivering strong performance while keeping compute and memory usage low—making it ideal for phones, tablets, and laptops.
API Usage Examples
OpenAI Compatible Endpoint
Use this endpoint with any OpenAI-compatible library. Model: LiquidAI: LFM2-8B-A1B (liquid/lfm2-8b-a1b)
curl https://api.ridvay.com/v1/chat/completions -H "Content-Type: application/json" -H "Authorization: Bearer YOUR_API_KEY" -d '{
"model": "liquid/lfm2-8b-a1b",
"messages": [
{
"role": "user",
"content": "Explain the capabilities of the LiquidAI: LFM2-8B-A1B model"
}
],
"temperature": 0.7,
"max_tokens": 1024
}'Supported Modalities
- Text
API Pricing
- Input: 0.01$ / 1M tokens
- Output: 0.02$ / 1M tokens
Token Limits
- Max Output: 32,768 tokens
- Max Context: 32,768 tokens
Subscription Tiers
- free
- pro
- ultimate