Artificial Intelligence

Baidu Drops ERNIE-Image: A Compact 8B Open-Source Text-to-Image Model That Tops the Charts

|Author: Viacheslav Vasipenok|3 min read| 11
Baidu Drops ERNIE-Image: A Compact 8B Open-Source Text-to-Image Model That Tops the Charts

Baidu has just released ERNIE-Image — a new open-weight text-to-image generator that is already turning heads in the AI community. With only 8 billion parameters, the model delivers state-of-the-art performance among all open models, competing head-to-head with significantly larger systems like Qwen-Image while outperforming Z-Image across multiple benchmarks.

The official announcement and weights landed on April 15, 2026, and are now available on Hugging Face under a fully permissive Apache 2.0 license — meaning developers can use, modify, and even commercialize the model with almost no restrictions.


Single-Stream MM-DiT Architecture: Simpler, Smaller, Surprisingly Strong

Baidu Drops ERNIE-Image: A Compact 8B Open-Source Text-to-Image Model That Tops the ChartsAt its core, ERNIE-Image uses a single-stream Diffusion Transformer (DiT) inside a latent diffusion (LDM) framework. Unlike models such as Flux that run separate text and image branches, ERNIE-Image feeds text tokens and image patches into the same transformer from the very first layer. All weights are shared, making the architecture cleaner, more parameter-efficient, and easier to run.

It’s conceptually similar to Z-Image but notably simpler and more compact — yet it still matches or beats the competition in quality.

One of the standout capabilities is text rendering. For an 8B model running at roughly 1-megapixel resolution, ERNIE-Image produces remarkably clean, readable, and layout-aware text in English, Chinese, and other languages. It handles dense paragraphs, posters, manga speech bubbles, and multi-panel compositions with impressive fidelity (ranking #2 on LongTextBench).


Prompt Enhancer + Turbo Variant = Maximum Flexibility

Baidu Drops ERNIE-Image: A Compact 8B Open-Source Text-to-Image Model That Tops the ChartsBaidu ships a 3B-parameter Prompt Enhancer (a fine-tuned Ministral 3B) that automatically expands short user prompts into rich, structured descriptions. It noticeably boosts output quality, but the model works perfectly fine without it if you prefer raw control.

There’s also ERNIE-Image-Turbo — a distilled version optimized with Distribution Matching Distillation (DMD) and reinforcement learning. It needs just 8 inference steps (instead of the usual 50) while delivering stronger aesthetics and faster generation. Perfect for rapid iteration.

Runs on Consumer Hardware

The entire model family fits comfortably in 24 GB VRAM, making it accessible on high-end consumer GPUs. One early tester reported generating images in just 11 seconds on an H200 using the Turbo version straight out of the box.


Benchmark Dominance

Baidu Drops ERNIE-Image: A Compact 8B Open-Source Text-to-Image Model That Tops the ChartsERNIE-Image currently ranks as the #1 open-weight text-to-image model across key evaluations:

  • GenEval (compositional generation): 0.8856 (w/o Prompt Enhancer) — beats Qwen-Image (0.8683) and Z-Image (0.8400);
  • OneIG-EN / OneIG-ZH (open-domain English & Chinese): top-3 overall, #1 among open models;
  • LongTextBench (text rendering fidelity): 0.9733 — second only to closed-source leaders;

It excels at complex multi-object scenes, precise attribute binding, and structured visuals like posters, anime storyboards, and cinematic compositions.

Also read:


Try It Now

If you’re looking for a lightweight, fully open, and surprisingly capable text-to-image model that doesn’t require a datacenter GPU, ERNIE-Image is worth trying right now. The combination of strong text rendering, efficient architecture, and Apache 2.0 licensing makes it one of the most exciting open releases of 2026 so far.

Share:
0