Best AI models for cost-sensitive tasks
Cheap-and-fast models for high-volume work where 80% quality at 5% cost is the right trade.
Triage, classification, lightweight transforms, and high-fanout fanouts (one prompt, many rows) all reward cheap models with predictable latency. Reserve the flagship LLMs for the few prompts where quality actually moves the metric.
Switchy's picks
- 1Anthropic: Claude Haiku 4.5
Claude Haiku 4.5 is Anthropic’s fastest and most efficient model, delivering near-frontier intelligence at a fraction of the cost and latency of larger Claude models. Matching Claude Sonnet 4’s performance...
anthropic200K context$1.00/Mtok in - 2
- 3OpenAI: GPT-4o-mini
GPT-4o mini is OpenAI's newest model after [GPT-4 Omni](/models/openai/gpt-4o), supporting both text and image inputs with text outputs. As their most advanced small model, it is many multiples more affordable...
openai128K context$0.15/Mtok in - 4OpenAI: GPT-4.1 Mini
GPT-4.1 Mini is a mid-sized model delivering performance competitive with GPT-4o at substantially lower latency and cost. It retains a 1 million token context window and scores 45.1% on hard...
openai1048K context$0.40/Mtok in
Other llm models
- AI21: Jamba Large 1.7ai21
- AAionLabs: Aion-1.0aion-labs
- AAionLabs: Aion-1.0-Miniaion-labs
- AAionLabs: Aion-2.0aion-labs
- AAionLabs: Aion-RP 1.0 (8B)aion-labs
- AAlfredPros: CodeLLaMa 7B Instruct Solidityalfredpros
- AllenAI: Olmo 3 32B Thinkallenai
- AllenAI: Olmo 3.1 32B Instructallenai
- Amazon: Nova 2 Liteamazon
- Amazon: Nova Lite 1.0amazon
- Amazon: Nova Micro 1.0amazon
- Amazon: Nova Premier 1.0amazon