Pricing & Benchmarks for
300+ AI Models
The professional kit for developers and enterprises to compare AI model costs, context windows, and performance benchmarks.
Start ComparingFeatured Models
View All →Anthropic: Claude Opus 4.7
Opus 4.7 is the next generation of Anthropic's Opus family, built for long-running, asynch...
Elephant
Elephant Alpha is a 100B-parameter text model focused on intelligence efficiency, deliveri...
Anthropic: Claude Opus 4.6 (Fast)
Fast-mode variant of [Opus 4.6](/anthropic/claude-opus-4.6) - identical capabilities with ...
Z.ai: GLM 5.1
GLM-5.1 delivers a major leap in coding capability, with particularly significant gains in...
Google: Gemma 4 26B A4B (free)
Gemma 4 26B A4B IT is an instruction-tuned Mixture-of-Experts (MoE) model from Google Deep...
Google: Gemma 4 26B A4B
Gemma 4 26B A4B IT is an instruction-tuned Mixture-of-Experts (MoE) model from Google Deep...
Google: Gemma 4 31B (free)
Gemma 4 31B Instruct is Google DeepMind's 30.7B dense multimodal model supporting text and...
Google: Gemma 4 31B
Gemma 4 31B Instruct is Google DeepMind's 30.7B dense multimodal model supporting text and...
Qwen: Qwen3.6 Plus
Qwen 3.6 Plus builds on a hybrid architecture that combines efficient linear attention wit...
Z.ai: GLM 5V Turbo
GLM-5V-Turbo is Z.ai’s first native multimodal agent foundation model, built for vision-...
Arcee AI: Trinity Large Thinking
Trinity Large Thinking is a powerful open source reasoning model from the team at Arcee AI...
xAI: Grok 4.20 Multi-Agent
Grok 4.20 Multi-Agent is a variant of xAI’s Grok 4.20 designed for collaborative, agent-...
Browse by Provider
Frequently Asked Questions
On this page:
What is "Writer: Palmyra X5" optimized for?
Palmyra X5 is Writer's most advanced model, purpose-built for building and scaling AI agents across the enterprise. It delivers industry-leading speed and efficiency on context windows up to 1 million tokens.
What are the input and output modalities of "OpenAI: GPT Audio"?
"OpenAI: GPT Audio" can process both text and audio as input, and can generate both text and audio as output.
What is the context length of "MiniMax: MiniMax M2-her"?
"MiniMax: MiniMax M2-her" has a context length of 32,768 tokens, making it suitable for multi-turn conversations.
Is there a free model for on-device AI?
Yes, "LiquidAI: LFM2.5-1.2B-Instruct (free)" is a compact, high-performance instruction-tuned model built for fast on-device AI and is free to use.