Fireworks AI
AI Infrastructure✓ Editorial
Fast, cheap inference for open-source LLMs — Llama, Mixtral, Qwen, DeepSeek served at sub-second latencies.
small, mediumprofessional
Baseten
AI Infrastructure✓ Editorial
Production-grade model serving for custom and open-source models — autoscaling GPU inference.
small, mediumprofessional
Lambda Labs
AI Infrastructure✓ Editorial
GPU cloud for AI training and inference — H100, H200, B200 instances at competitive on-demand prices.
medium, largeenterprise
RunPod
AI Infrastructure✓ Editorial
GPU cloud with serverless inference — pay-per-second GPU access from $0.20/hr for community-tier hardware.
solo, smallstarter
Mem0
AI Infrastructure✓ Editorial
Memory layer for AI agents — long-term, structured memory that survives across sessions and conversations.
solo, smallstarter
Letta
AI Infrastructure✓ Editorial
Stateful agent framework (formerly MemGPT) — agents with long-term memory, sleep cycles, and self-editing context.
small, mediumstarter