290+ models · 40+ providers · real-time pricing

Choose the right inference
for your customer support

Compare LLM providers. Size GPUs for self-hosting.
Understand the economics behind every token.

Tools

Three ways to reason about AI costs

What you can do
Compare inference costs for your exact workload
Estimate VRAM requirements for inference and training
Compare accelerators — real memory and bandwidth
Understand trade-offs: context, batch size, precision
Connect model choices to cost and scalability
Philosophy

A technical planning tool for people who design, deploy, and scale LLM systems. Focused on how models actually consume memory and bandwidth.

No benchmarks for marketing. No abstract performance scores.
All assumptions are explicit and inspectable.

ML EngineersCTOsPlatform TeamsFounders