Black Forest Labs (BFL) is a German-American artificial intelligence company founded in 2024 by Robin Rombach, Andreas Blattmann, Patrick Esser, and Dominik Lorenz, all of whom were key researchers behind the latent diffusion technology that powered Stable Diffusion. The company is best known for creating the FLUX family of text-to-image models, which rapidly became some of the most widely used image generation models in the industry. Black Forest Labs has raised over $430 million in total funding, reaching a valuation of $3.25 billion as of its December 2025 Series B round [1][2].
The founding of Black Forest Labs represents a notable case of core technical innovators leaving a company (Stability AI) to build a new venture based on their own foundational research. The FLUX models have been adopted by major platforms including Elon Musk's Grok chatbot for image generation, and in September 2025, Adobe integrated FLUX.1 Kontext Pro into Photoshop's generative fill tool [3][4].
The story of Black Forest Labs begins with a research paper published in 2022: "High-Resolution Image Synthesis with Latent Diffusion Models," presented at CVPR 2022. The paper was authored by Robin Rombach, Andreas Blattmann, Dominik Lorenz, Patrick Esser, and Bjorn Ommer, all affiliated with Ludwig-Maximilians-Universitat (LMU) Munich and Heidelberg University [5].
This paper introduced the concept of latent diffusion, a technique that applies the diffusion process in a compressed latent space rather than directly on pixel-level images. By operating in this lower-dimensional space, latent diffusion models could generate high-quality images at a fraction of the computational cost of previous diffusion approaches. The paper's core innovation enabled the creation of Stable Diffusion, which became the most widely used open-source AI image generation system in the world [5].
Stable Diffusion was released in August 2022 by Stability AI, a company founded by Emad Mostaque. The model was built directly on the latent diffusion research of Rombach and his colleagues from the CompVis group at LMU Munich. Four of the five original latent diffusion paper authors (Rombach, Blattmann, Esser, and Lorenz) joined Stability AI to continue developing the technology commercially [5][6].
At Stability AI, the team developed subsequent versions of Stable Diffusion, including Stable Diffusion 2.0 and SDXL. However, Stability AI experienced significant organizational and financial turbulence in 2023 and 2024, including leadership changes, layoffs, and questions about the company's financial sustainability. CEO Emad Mostaque resigned in March 2024 amid growing pressure [6].
The departure of the core technical team from Stability AI to found Black Forest Labs reflected a broader pattern in the AI industry where the researchers who develop key technologies often move on to build new companies, taking their expertise (though not their former employer's proprietary work) with them.
| Founder | Role | Background |
|---|---|---|
| Robin Rombach | CEO | Lead author of the latent diffusion paper; PhD at LMU Munich/Heidelberg; studied physics at University of Heidelberg (2013-2020) |
| Andreas Blattmann | Co-founder | Co-author of latent diffusion paper; researcher at LMU Munich; contributed to video diffusion research |
| Patrick Esser | Co-founder | Co-author of latent diffusion paper; co-developed VQGAN; researcher at LMU Munich |
| Dominik Lorenz | Co-founder | Co-author of latent diffusion paper; researcher at LMU Munich |
All four founders had previously worked in the Computer Vision and Learning (CompVis) group at LMU Munich under the supervision of Bjorn Ommer. Their shared research background and years of collaboration provided a strong technical foundation for the new company [5].
Black Forest Labs has raised capital rapidly, reflecting strong investor confidence in the team's technical capabilities and the commercial potential of their image generation technology.
| Round | Date | Amount | Valuation | Lead Investors |
|---|---|---|---|---|
| Seed | August 2024 | $31 million | ~$150 million (post-money) | Andreessen Horowitz (a16z) |
| Series A | Late 2024 | ~$100 million | ~$1 billion | Andreessen Horowitz |
| Series B | December 2025 | $300 million | $3.25 billion | Salesforce Ventures, AMP (Anjney Midha) |
| Total | $430+ million | $3.25 billion |
The seed round of $31 million was announced alongside the company's public launch on August 1, 2024. In addition to a16z, the seed round included investments from notable figures including Brendan Iribe (Oculus co-founder), Michael Ovitz, and Garry Tan (Y Combinator CEO) [1].
By September 2024, reports emerged that Black Forest Labs was seeking to raise approximately $100 million at a $1 billion valuation, representing a dramatic jump from the $150 million post-money valuation of the seed round. The Series A was led by Andreessen Horowitz, with participation from BroadLight Capital, Creandum, Earlybird VC, General Catalyst, Northzone, and NVIDIA [3].
The December 2025 Series B of $300 million at a $3.25 billion valuation was co-led by Salesforce Ventures and AMP, with additional participation from a16z, NVIDIA, General Catalyst, and Temasek. The round reflected the strong commercial traction of the FLUX models and the growing demand for high-quality image generation capabilities across multiple industries [2].
The FLUX models are Black Forest Labs' primary products, representing a significant technical advancement over previous image generation models including Stable Diffusion.
FLUX.1 was released alongside the company's public launch on August 1, 2024. The model family initially comprised three variants, each targeting different use cases and licensing requirements [7].
| Variant | Parameters | License | Speed | Availability |
|---|---|---|---|---|
| FLUX.1 [schnell] | 12 billion | Apache 2.0 | Fastest (1-4 steps) | Open weights, local deployment |
| FLUX.1 [dev] | 12 billion | Non-commercial | Medium | Open weights (Hugging Face) |
| FLUX.1 [pro] | 12 billion | Proprietary | Slower (higher quality) | API only |
FLUX.1 [schnell] (German for "fast") is optimized for speed and local development, capable of generating images in as few as one to four inference steps. It is released under the Apache 2.0 license, making it freely available for personal, research, and commercial use [7].
FLUX.1 [dev] is an open-weight model released for non-commercial use. It offers higher quality than schnell while still being available for download and local deployment. Researchers and developers can use it for experimentation and prototyping [7].
FLUX.1 [pro] is the highest-quality variant, available exclusively through Black Forest Labs' API. It targets commercial applications where image quality is the primary concern [7].
Upon release, FLUX.1 quickly demonstrated state-of-the-art performance. The models outperformed Midjourney 6.1, DALL-E 3, and Stable Diffusion XL on multiple evaluation metrics, including visual quality, prompt adherence, and text rendering within images [7].
In October 2024, Black Forest Labs released FLUX.1.1 Pro, an upgraded version of the pro model. The key improvements included six times faster generation speed compared to FLUX.1 Pro (generating images in approximately 4.5 seconds), improved image quality and photorealism, better text rendering capabilities, and enhanced prompt adherence [8].
FLUX.1.1 Pro achieved the top ranking on Artificial Analysis, a leading benchmark platform for text-to-image models, outperforming all competitors including Midjourney 6.1 and Ideogram v2 in both visual fidelity and prompt accuracy [8].
Two additional modes were introduced in November 2024: Ultra (generating images at four times higher resolution, up to 4 megapixels, without affecting generation speed) and Raw (generating hyper-realistic images in the style of candid photography) [8].
FLUX.1 Kontext, announced on May 29, 2025, represented a new direction for the model family. Rather than purely text-to-image generation, Kontext enables in-context image generation and editing, allowing users to prompt with both text and images. The model can extract and modify visual concepts from input images to produce new, coherent renderings [4].
| Kontext Variant | Focus | Availability |
|---|---|---|
| Kontext [max] | Highest quality; iterative image modification | API |
| Kontext [pro] | Balanced quality and speed | API |
| Kontext [dev] | Non-commercial research | Open weights |
FLUX.1 Kontext Pro performs at a level described as eight times faster than comparable advanced models (such as GPT-Image) across six context image generation tasks. In September 2025, Adobe announced that Photoshop beta users could use FLUX.1 Kontext Pro as a model for its generative fill tool, marking a significant commercial milestone for Black Forest Labs [4].
On November 25, 2025, Black Forest Labs announced FLUX.2, the second major generation of the model family. The release included several variants [9].
| Model | Description | License |
|---|---|---|
| FLUX.2 Pro | Highest quality generation | Proprietary (API) |
| FLUX.2 Flex | Flexible generation and editing | Proprietary (API) |
| FLUX.2 Dev | Development and research | Non-commercial |
| FLUX.2 Klein | Small, efficient model | Apache 2.0 |
| FLUX.2 VAE | Variational autoencoder | Apache 2.0 |
The FLUX.2 variational autoencoder was released as open-source software under the Apache 2.0 license, allowing the community to build on the model's image encoding and decoding capabilities [9].
The FLUX models represent a significant architectural departure from the U-Net-based architecture used in Stable Diffusion. Instead, FLUX builds on two key technical innovations: the Diffusion Transformer (DiT) architecture and flow matching [7][10].
Traditional diffusion models for image generation, including all versions of Stable Diffusion, used U-Net architectures for the denoising network. The DiT approach replaces the U-Net with a transformer architecture, bringing the scalability advantages of transformers (which had proven so effective in language modeling) to image generation [10].
FLUX uses a novel multimodal DiT variant called MM-DiT (Multimodal Diffusion Transformer), which is specifically designed to handle the multimodal nature of text-to-image generation. The architecture processes text and image information through parallel streams that interact at multiple points during the generation process [10].
The architecture consists of two types of transformer blocks:
This hybrid approach allows the model to maintain modality-specific processing where needed while also enabling deep integration between text understanding and image generation [10].
FLUX uses rectified flow matching rather than the traditional DDPM (Denoising Diffusion Probabilistic Models) approach used in Stable Diffusion. Flow matching is a more general framework for training generative models that includes diffusion as a special case [10].
The key advantage of rectified flows is that they encourage linear denoising trajectories, meaning the model learns to transform noise into images along straighter paths in the generative process. This property enables more efficient sampling: the model can generate high-quality images in fewer steps because each step makes more progress along the generation trajectory. This is why FLUX.1 [schnell] can produce good results in as few as one to four steps, compared to the 20-50 steps typically needed by Stable Diffusion [10].
| Feature | Description | Benefit |
|---|---|---|
| Rotary Position Embeddings (RoPE) | Position encoding via rotation matrices | Better handling of varying image resolutions |
| Parallel Attention Layers | Concurrent computation of attention and feedforward | Improved hardware efficiency |
| Latent Space Operation | Generation in compressed representation | Lower compute requirements |
| 12 Billion Parameters | Large model capacity | Strong generation quality |
The combination of these architectural choices results in a model that generates higher-quality images, handles text rendering better (a longstanding weakness of diffusion models), and runs more efficiently than previous approaches [7][10].
The relationship between Black Forest Labs and Stable Diffusion is central to understanding the company's position in the AI image generation landscape.
The foundational latent diffusion research was conducted at LMU Munich and Heidelberg University, funded by academic grants and the German Research Foundation (DFG). When Stability AI licensed and commercialized this research as Stable Diffusion in 2022, the core researchers (Rombach, Blattmann, Esser, Lorenz) joined the company [5][6].
Stability AI funded further development of the technology, producing Stable Diffusion 2.0, SDXL, and Stable Diffusion 3. However, as Stability AI encountered financial difficulties and leadership turmoil in 2023-2024, the technical team departed to start Black Forest Labs. The FLUX models represent a clean break architecturally (using DiT and flow matching instead of U-Net and DDPM), but the intellectual lineage from latent diffusion through Stable Diffusion to FLUX is direct and continuous [6].
| Aspect | Stable Diffusion (1.x/2.x/XL) | FLUX |
|---|---|---|
| Architecture | U-Net based | Diffusion Transformer (DiT) |
| Training Method | DDPM | Rectified Flow Matching |
| Text Encoder | CLIP | T5 + CLIP |
| Developer | Stability AI | Black Forest Labs |
| Core Researchers | Rombach, Blattmann, Esser, Lorenz | Same (now at BFL) |
| Image Quality | Good (progressive improvement) | State-of-the-art at release |
| Text Rendering | Weak | Significantly improved |
| Generation Speed | 20-50 steps typical | 1-4 steps (schnell), 20-50 (pro) |
Stability AI continued to develop its own models after the departure of the founding researchers, releasing Stable Diffusion 3 and subsequent versions. However, FLUX models have generally been regarded as technically superior, leading to a situation where the spiritual successors to Stable Diffusion now compete against it [6].
Black Forest Labs has secured several high-profile commercial partnerships that have driven adoption of FLUX models.
xAI (Grok): Elon Musk's xAI uses FLUX.1 to power image generation capabilities in the Grok chatbot. This partnership brought significant visibility to Black Forest Labs and demonstrated the models' production readiness for high-traffic consumer applications [3].
Adobe: In September 2025, Adobe integrated FLUX.1 Kontext Pro into the generative fill tool in Photoshop (beta). This partnership is particularly significant because it places FLUX technology in one of the world's most widely used creative software suites, alongside Adobe's own Firefly models [4].
API Access: Black Forest Labs operates its own API (api.bfl.ai) for direct access to FLUX models. The API is also available through third-party platforms including Replicate, fal.ai, and Together AI, broadening the model's accessibility to developers [7].
The AI image generation market as of early 2026 includes several major players, each with distinct strengths.
| Model | Developer | Architecture | Open Weights | Text Quality | Speed | Key Strength |
|---|---|---|---|---|---|---|
| FLUX.2 | Black Forest Labs | DiT + Flow Matching | Partial (Klein/VAE) | Excellent | Fast | Technical quality, open variants |
| Midjourney v6 | Midjourney | Unknown (proprietary) | No | Good | Medium | Artistic aesthetics |
| DALL-E 3 | OpenAI | Unknown (proprietary) | No | Good | Medium | Integration with ChatGPT |
| Stable Diffusion 3 | Stability AI | DiT variant | Partial | Good | Medium | Community ecosystem |
| Imagen 3 | Diffusion + Transformer | No | Excellent | Medium | Photorealism | |
| Firefly | Adobe | Proprietary | No | Good | Fast | Commercial licensing clarity |
FLUX models have generally achieved top rankings on independent benchmarks for text-to-image quality, particularly excelling in prompt adherence, text rendering, and photorealism. The availability of open-weight variants (schnell under Apache 2.0, dev for non-commercial use) gives FLUX a significant advantage among developers and researchers who want to run models locally or fine-tune them for specific applications [7][8].
As of early 2026, Black Forest Labs is one of the leading companies in AI image generation. The FLUX model family has evolved through multiple generations and variants, establishing itself as a top-tier option for both open-source and commercial image generation.
The company's $3.25 billion valuation from its December 2025 Series B reflects strong investor confidence in the team's ability to maintain technical leadership in a rapidly evolving field. The combination of open-weight models (which drive community adoption and developer ecosystems), proprietary API services (which generate revenue), and high-profile partnerships (Adobe, xAI) provides a multi-pronged business model [2].
The broader AI image generation landscape continues to advance rapidly, with competition from Midjourney, OpenAI, Google, Adobe, and others. Black Forest Labs' core advantage remains the deep technical expertise of its founding team, the same researchers who created the latent diffusion approach that transformed the entire field.