The fastest AI inference — custom LPU chips delivering 10x speed for open-source models.
+ Fastest inference speeds available
+ Generous free tier
+ OpenAI-compatible API
- Limited model selection
- No fine-tuning support
- Availability can be constrained
Generous free tier. Pay-per-token for higher limits. Very competitive pricing.
More in LLM Providers & APIs