UNO-FLUX: Redefining Image Generation with Smarter, Simpler Control

Uno Image Creation

Uno Image Creation

Redefining the Future of AI Image Generation

Imagine a world where you can guide AI creativity with just a few images and a simple prompt — and the results don’t just approximate your vision; they perfectly reflect it.
That future is here with UNO-FLUX, the latest breakthrough from ByteDance Research.

Built upon the cutting-edge ideas presented in “Less-to-More Generalization: Unlocking More Controllability by In-Context Generation,” UNO-FLUX transforms how we approach subject-driven image generation — offering a new level of consistency, precision, and creative control.

🚀 Try it yourself on Hugging Face Spaces.


What Makes UNO-FLUX Different?

At its core, UNO-FLUX is not just another image generator — it’s a new paradigm.

Traditional models struggle when multiple subjects are involved. UNO-FLUX thrives in complexity, allowing you to:

  • Upload up to four reference images.
  • Input a natural language prompt.
  • Seamlessly blend both visual and textual intent into a coherent, striking output.

Whether you want to generate a surreal multi-character scene or a hyper-specific object portrait, UNO-FLUX delivers images that stay true to your references without losing creative vibrancy.


The Innovation Behind It

  • 🎯 Less-to-More Generalization:
    Instead of requiring endless fine-tuning, UNO-FLUX learns from fewer examples and generalizes better, putting more creative power in your hands.
  • 🔄 Multi-Modal Fusion:
    It doesn’t just “see” your pictures and “read” your prompt separately — it understands them together. Every pixel is infused with context.
  • 🧠 Progressive Cross-Modal Alignment:
    Ensures that the deeper the generation process goes, the tighter the bond between text and imagery becomes.
  • 🌀 Universal Rotary Position Embedding (URPE):
    Think of it as giving the model a natural sense of “where things belong,” improving spatial consistency across multiple subjects.

How to Experience It

  1. Visit the UNO-FLUX Demo.
  2. Enter a text description of the scene you want.
  3. Upload up to four images that the AI should use as references.
  4. Click generate and watch your vision come alive — faithfully, beautifully, and consistently.

Technical Details (For the Curious)

  • Model Backbone:
    Built on FLUX.1-dev, a highly flexible diffusion transformer.
  • Data Strategy:
    Powered by a synthetic paired data pipeline designed for multi-subject generation without manual intervention.
  • Licensing:
    • Code: Apache 2.0 License
    • Models: CC BY-NC 4.0 License, building upon FLUX.1-dev’s licensing foundation.

Explore More


Final Take

UNO-FLUX is not just about generating images — it’s about controlling creativity with precision.
It invites artists, developers, and researchers alike to rethink what’s possible when AI doesn’t just follow a prompt, but understands the context behind it.

Welcome to the next generation of creative AI. Welcome to UNO-FLUX.