baidu

Baidu: ERNIE 4.5 21B A3B

A sophisticated text-based Mixture-of-Experts (MoE) model featuring 21B total parameters with 3B activated per token, delivering exceptional multimodal understanding and generation through heterogeneous MoE structures and modality-isolated routing. Supporting an extensive 131K token context length, the model achieves efficient inference via multi-expert parallel collaboration and quantization, while advanced post-training techniques including SFT, DPO, and UPO ensure optimized performance across diverse applications with specialized routing and balancing losses for superior task handling.

Input Cost
$0.07
per 1M tokens
Output Cost
$0.28
per 1M tokens
Context Window
120,000
tokens
Compare vs GPT-4o
Developer ID: baidu/ernie-4.5-21b-a3b

Related Models

baidu
$0.14/1M

Baidu: ERNIE 4.5 VL 28B A3B

A powerful multimodal Mixture-of-Experts chat model featuring 28B total parameters with 3B...

📝 30,000 ctx Compare →
baidu
$0.07/1M

Baidu: ERNIE 4.5 21B A3B Thinking

ERNIE-4.5-21B-A3B-Thinking is Baidu's upgraded lightweight MoE model, refined to boost rea...

📝 131,072 ctx Compare →
baidu
$0.28/1M

Baidu: ERNIE 4.5 300B A47B

ERNIE-4.5-300B-A47B is a 300B parameter Mixture-of-Experts (MoE) language model developed ...

📝 123,000 ctx Compare →