Black Forest Labs released FLUX.2, their second-generation image creation and editing software. FLUX.2 is designed to target real-world creative workflows like marketing assets, photography of products, layouts for design, complex infographics and more. It supports editing up to four megapixels with strong controls over logos and typography.
The FLUX.2 family of products and the FLUX.2 [dev]
FLUX.2 includes both hosted APIs (open weights) and hosted APIs.
- FLUX.2 [pro] This is the managed API Tier. This tier aims to provide state-of-the art quality as compared with closed models.
- FLUX.2 [flex] Developers can adjust parameters like the number of steps or guidance scale to optimize latency and text rendering accuracy.
- FLUX.2 [dev] The open weight checkpoint is derived directly from the FLUX.2 base model. The model is called the most powerful image creation and editing open weight checkpoint. It combines text to image, multi-image editing, and 32 billion parameters in one checkpoint.
- FLUX.2 [klein] There is an open-source Apache 2.0 version that will be released soon, resized from the original model to fit smaller installations, but with the same features.
The model can be edited from multiple text references and images. This eliminates the need for separate checking points.
FLUX.2 and Architecture: Latent Flow, FLUX.2
FLUX.2 utilizes a latent-flow matching architecture. The core concept couples the performance of a Mistral-3 24B vision language model With a Reverse flow Transformer The latent image is used to operate the model. Vision language provides world knowledge while transformer learns about spatial structure and materials.
Text-driven synthesis is supported by the same architecture, which uses a model that has been trained to map image noise latents into text latents. The latents for editing are initially initialized using existing images and then updated within the same flow while maintaining structure.
New FLUX.2 (VAE) Definition of the latent area. The software is licensed under Apache 2.0 and released by Hugging Face. The autoencoder forms the basis of all FLUX.2 models, and it can be used in other generative system.
Features for workflows in production
The FLUX.2 Docs Integration highlights key capabilities.
- Multi-reference supportFLUX.2 combines up to 10 references images in order to preserve character identitiy, style, and product appearance across all outputs.
- Photoreal detail at 4MPModel can generate and edit images of up to 4 Megapixels. Improved textures are available for skin, clothing, lighting and hands.
- Text and Layout rendering robustIt can render infographics and memes as well as user interfaces in a small, readable font, which was a weakness of many older models.
- Space logic and world knowledgeThe model has been trained to reduce artifacts, synthetic appearance, and perspective.

What you need to know
- FLUX.2 32B is a matching latent-flow transformer which unifies the text to image conversion, image editing and multi-reference composition at a checkpoint.
- FLUX.2 [dev] Open weight is paired to the Apache 2.0 FLUX.2VAE while core weights use FLUX.2-dev Non Commercial License and mandatory safety filters.
- This system allows for up to four megapixels of generation and editing. Text and layout rendering is robust, with up to ten visual references to ensure consistency between characters, products and styles.
- Full precision inference needs more than 80GB RAM, but FLUX.2 is able to do it with 4 bit quantized pipes and FP8 pipelines. [dev] You can use it on GPUs from 18GB up to 24GB, and 8GB graphics cards as long as you have sufficient RAM.
Editorial Notes
FLUX.2 represents an important milestone in open-weight visual generation. It combines the 32B rectified flow transform, Mistral 3’s 24B language model and FLUX.2 Visual Analyzer Engine into one high-fidelity pipeline that can be used for editing and text to images. Clear VRAM profiles, quantized variations, and integrations with Diffusers and Cloudflare Workers, as well as strong integrations to Diffusers and ComfyUI make this a practical tool for actual workloads and not just benchmarks. Open image models are now closer to being used in production-grade creative infrastructure.
Click here to find out more Technical details, Model weight The following are some examples of how to get started: Repo. Check out our GitHub Page for Tutorials, Codes and Notebooks. Also, feel free to follow us on Twitter Don’t forget about our 100k+ ML SubReddit Subscribe now our Newsletter. Wait! Are you using Telegram? now you can join us on telegram as well.


