Luma AI (operating as Luma Labs, Inc.) is an artificial intelligence company headquartered in Palo Alto, California, that develops generative AI models for video, 3D, and image creation. Founded in 2021 by Amit Jain, Alex Yu, and Alberto Taiuti, the company initially gained attention for its neural radiance field (NeRF) based 3D capture technology before expanding into AI video generation with its Dream Machine platform. Luma AI's flagship Ray series of video models powers text-to-video, image-to-video, and video-to-video generation through its consumer-facing Dream Machine application and developer API.
As of early 2026, Luma AI has raised over $1 billion in total funding, is valued at $4 billion, and serves more than 30 million registered users on its Dream Machine platform.
Luma AI was co-founded in August 2021 by Amit Jain, Alex Yu, and Alberto Taiuti. Jain and Taiuti both previously worked at Apple, where Jain led work on the Passthrough feature for the Apple Vision Pro and the integration of the first LiDAR sensors for the iPhone. Taiuti also worked at Apple as an AR/VR Engineer and previously served as a Senior Autonomy Software Engineer at Skydio. Alex Yu was a researcher at UC Berkeley, where he studied neural rendering under Professor Angjoo Kanazawa and authored papers on real-time neural rendering of 3D scenes.
The company's initial product focused on NeRF-based 3D capture, allowing users to create photorealistic 3D scenes from smartphone video. The technology used deep learning networks to process 5D coordinates (location and view direction), outputting volume density and RGB radiance for each point in the scene. Users could walk around a subject while recording video with a phone, and the AI would analyze camera movement and scene geometry to reconstruct a full 3D representation.
Luma AI raised a $4.3 million seed round on October 30, 2021, led by Amplify Partners.
In March 2023, Luma AI raised $20 million in a Series A round led by Amplify Partners, with participation from NVentures (NVIDIA's venture arm) and General Catalyst. The round valued the company at $100 million post-money.
During this period, the company released its iOS app for 3D capture, which allowed anyone with an iPhone 11 or newer to create lifelike 3D models without specialized equipment such as LiDAR scanners. The app supported exporting to industry-standard formats including USDZ, glTF, and OBJ, making captures compatible with tools like Blender, Unity, and Unreal Engine.
In April 2023, Luma AI released a plugin for Unreal Engine, allowing users to import and render photorealistic NeRF captures directly inside the game engine. The plugin included tools for scene cleanup, cropping, culling, and relighting via analytical or sun lights.
In June 2023, Luma AI hired Matthew Tancik, a co-author of the original NeRF paper and co-founder of the nerfstudio framework. Tancik joined to lead the applied research team. He had previously collaborated with co-founder Alex Yu on papers such as PlenOctrees and Plenoxels.
In November 2023, Luma AI launched Genie as a research preview, a text-to-3D generation tool. Genie attracted tens of thousands of users on its first day.
On January 9, 2024, Luma AI raised $43 million in a Series B round led exclusively by Andreessen Horowitz (a16z). Alongside the funding announcement, Luma released Genie 1.0 for general availability.
a16z's investment thesis cited Luma's position in generative 3D, noting that the field was "on the cusp of [its] own breakthrough into the mainstream" following inflection points in 2D image generation in 2022 and computer vision in 2012. The firm highlighted the strength of Luma's technical team, which by this point included Jiaming Song as Chief Scientist. Song, who holds a Ph.D. from Stanford University under advisor Stefano Ermon, invented the DDIM (Denoising Diffusion Implicit Models) algorithm, which reduced the sampling steps required for diffusion models from roughly 1,000 iterations to 20-50. DDIM was subsequently adopted in production systems including DALL-E 2, Imagen, and Stable Diffusion.
On June 12, 2024, Luma AI launched Dream Machine, a text-to-video generation platform powered by its first Ray model. The tool allowed anyone to type a descriptive prompt and generate a short high-definition video clip in minutes. Dream Machine gained one million users within four days of launch, achieved entirely through word-of-mouth without a marketing budget.
The launch attracted wide attention on social media. Users created video versions of images generated with Midjourney, as well as moving recreations of famous artworks such as "Girl with a Pearl Earring" and popular internet memes including Doge and the distracted boyfriend meme. Director Ellenor Argyropoulos posted a Pixar-style animation of a girl in ancient Egypt that went viral.
Liam Connell, lead AI engineer at Boston Consulting Group, noted at the time that Dream Machine's video quality and realism stood out from competing models, with fewer obvious inconsistencies than alternatives.
On November 25, 2024, Luma AI introduced Photon, a text-to-image foundation model. The Photon family included two variants: Photon (full quality) and Photon Flash (optimized for speed). Luma claimed the models were 800 percent faster and cheaper than comparable image generation services. Photon generated ultra-high-quality 1080p images at a cost of 1.5 cents per image, while Photon Flash cost 0.4 cents per image.
Photon introduced several capabilities new to image generation at the time, including a large context window for visual generative models and consistent character generation from a single reference image (in beta). Alongside Photon, Luma launched subscription tiers for Dream Machine: Lite ($9.99/month), Plus ($29.99/month), Professional ($99.99/month), and Enterprise.
On December 6, 2024, Luma AI raised $90 million in an initial Series C round with participation from new investors Amazon and AMD, alongside existing backers a16z, Amplify Partners, and Matrix Partners. This brought total funding to approximately $157 million.
On January 15, 2025, Luma AI released Ray2, its second-generation video model. Ray2 was built on a new multi-modal architecture with 10x the compute power of Ray1. The model was trained directly on video sequences, teaching it to understand motion as continuous flow rather than discrete snapshots.
Ray2 supported video generation at 540p, 720p, and 1080p resolution with 4K upscaling, producing clips of 5 to 10 seconds that could be extended to approximately 30 seconds. The model introduced structured prompting, seamless looping, camera movement controls (crane, tracking, dolly), keyframe control, style references, and color grading. Ray2 was available exclusively to paid Dream Machine subscribers, via the iOS app, and through the developer API.
Luma AI's Ray2 model was also made available through Amazon Bedrock, making AWS the first cloud provider to offer Luma's video models to enterprise customers. Under the partnership, AWS became the preferred compute partner for Luma AI, and Luma began training its foundation models on Amazon SageMaker HyperPod infrastructure.
On November 19, 2025, at the U.S.-Saudi Investment Forum during Crown Prince Mohammed bin Salman's visit to Washington, D.C., Luma AI announced a $900 million Series C investment. The round was led by HUMAIN, a company owned by Saudi Arabia's Public Investment Fund (PIF) that delivers global full-stack AI solutions. AMD Ventures also participated significantly, along with existing investors a16z, Amplify Partners, and Matrix Partners. The round valued Luma AI at over $4 billion, making it a unicorn.
As part of the deal, Luma AI became a customer of HUMAIN's Project Halo, a planned 2-gigawatt AI supercluster in Saudi Arabia that would be one of the world's largest compute infrastructure buildouts. Luma AI stated it was developing technology capable of training on "peta-scale multimodal data," handling 1,000 to 10,000 times more information than frontier large language models (LLMs). The partnership also included HUMAIN Create, an initiative to build sovereign AI models trained on Arabic and regional data, with Luma working on what it described as the world's first Arabic video model.
On September 18, 2025, Luma AI released Ray3, which it described as the world's first reasoning video model. Ray3 was more than twice the size of Ray2 and introduced several technical firsts. It generated physically accurate videos with improved fidelity, instruction following, and temporal coherence. Ray3 was the first video generation model capable of producing High Dynamic Range (HDR) video in ACES EXR format, a professional color encoding standard used in film and television production.
Ray3 included reasoning-driven generation, improved video-to-video workflows with character reference and keyframe support, a Draft Mode for rapid iteration, and an HDR pipeline. Adobe integrated Ray3 into Adobe Firefly and Firefly Boards, making Luma AI's first third-party distribution partner. Users on paid Firefly or Creative Cloud Pro plans received unlimited Ray3 generations through October 1, 2025. Content generated with Ray3 in Firefly could be synced to Creative Cloud accounts for editing in Premiere Pro and other Adobe applications.
On December 2, 2025, Luma AI opened its first international office in London, United Kingdom. The company appointed Jason Day, a former executive at Monks and WPP, as Head of EMEA to lead international expansion across creative, advertising, gaming, and entertainment sectors. Luma AI planned to hire approximately 200 employees in London by 2026, representing about 40% of its workforce. The company chose London in part because of proximity to institutions like DeepMind and the strength of the UK's AI research community.
In December 2025, Luma AI released Ray3 Modify, a model designed for hybrid workflows combining real-world footage with AI generation. Ray3 Modify introduced four capabilities: Keyframe Control, Character Reference, enhanced Modify Video, and Start-and-End Frame control. The model allowed creative teams to provide a start frame and end frame, then generate transitional footage between them. It could retain an actor's original motion, timing, eye line, and emotional delivery while transforming the surrounding scene, costumes, or environment.
On January 26, 2026, Luma AI launched Ray3.14, an update that delivered native 1080p generation, 4x faster generation speeds, and per-second pricing that was 3x cheaper than Ray3. The model applied the reasoning engine from Ray3 more powerfully to animation and professional video workflows, producing higher detail adherence and the strongest temporal stability Luma had achieved. Modify Video support was extended to 18-second clips.
On February 11, 2026, Luma AI announced plans to open a Riyadh, Saudi Arabia office to accelerate the HUMAIN Create initiative. The company also named Publicis Groupe Middle East as its preferred AI creative partner across the MENA region. Through the partnership, Publicis would integrate Luma AI's generative video and multimodal AI technologies into creative production workflows for brands across the Middle East and North Africa.
Dream Machine is Luma AI's consumer-facing web platform and iOS application for AI video generation. Launched in June 2024, it allows users to generate video from text prompts, images, or existing video clips. The platform supports text-to-video, image-to-video, video-to-video, and video extension workflows.
| Feature | Description |
|---|---|
| Text-to-video | Generate video clips from natural language descriptions |
| Image-to-video | Animate a static image into a video sequence |
| Video-to-video | Modify existing footage with AI-driven transformations |
| Start/End frame | Provide beginning and ending frames; the model generates transitional footage |
| Video extension | Extend an existing clip while maintaining motion coherence |
| Camera controls | Specify camera movements such as crane, tracking, dolly, and pan |
| Keyframe control | Set keyframes to guide the model's generation across a clip |
| Character reference | Provide a reference image to maintain character consistency |
| Style reference | Apply a visual style from a reference image |
| HDR output | Generate video in High Dynamic Range using ACES EXR format (Ray3+) |
| Looping | Create seamlessly looping video clips |
| 4K upscaling | Upscale generated video from native resolution to 4K |
As of early 2026, Dream Machine has over 30 million registered users and generated approximately $21.2 million in revenue during 2025.
| Plan | Monthly Price | Features |
|---|---|---|
| Free | $0 | Limited generations per day, watermarked output |
| Lite | $9.99 | Increased credits, commercial use rights |
| Plus | $29.99 | More credits, faster generation |
| Unlimited | $94.99 | Unlimited generations, priority access |
| Enterprise | Custom | Custom volume, dedicated support, SLAs |
The Ray models are Luma AI's core video generation models, powering the Dream Machine platform and API.
| Model | Release Date | Key Capabilities |
|---|---|---|
| Ray 1 | June 2024 | First-generation video model; text-to-video and image-to-video at launch |
| Ray 1.6 | 2024 | Improved visual quality and coherence over Ray 1 |
| Ray 2 | January 15, 2025 | New multi-modal architecture with 10x compute; 540p/720p/1080p with 4K upscale; extended clips up to 30 seconds; camera controls, looping, style reference |
| Ray 3 | September 18, 2025 | First reasoning video model; 2x+ size of Ray2; native 1080p; HDR in ACES EXR; character reference, keyframes, Draft Mode |
| Ray 3 Modify | December 2025 | Hybrid AI/live-action workflows; start/end frame control; actor performance preservation; scene transformation |
| Ray 3.14 | January 26, 2026 | Native 1080p; 4x faster; 3x cheaper per-second pricing; strongest temporal stability; 18s Modify Video support |
Ray2's transformer architecture processes text prompts by decomposing descriptions into scene components, motion dynamics, lighting conditions, and camera behavior. It then generates video by predicting the temporal evolution of the scene frame by frame, maintaining consistency with physical properties learned during training. The model was trained directly on video sequences rather than still images, which Luma credits with its ability to produce natural, continuous motion.
Genie is Luma AI's text-to-3D generation tool. It converts natural language prompts into 3D models in under 10 seconds. The system follows a two-stage process: first generating four low-resolution previews in approximately 10 seconds, then allowing users to refine their chosen variant into a high-quality model with PBR (Physically Based Rendering) textures and optimized mesh topology.
Genie generates quad meshes and materials at configurable polygon counts, exporting in standard formats compatible with professional 3D pipelines. It is accessible via web, iOS, and Discord. The tool targets developers, 3D artists, game designers, and VR/AR creators who need rapid 3D asset prototyping.
Genie was first released as a research preview in November 2023 and reached general availability with version 1.0 in January 2024.
Photon is Luma AI's text-to-image foundation model, released in November 2024. The model family includes Photon (full quality at 1.5 cents per 2MP 1080p image) and Photon Flash (speed-optimized at 0.4 cents per image). Luma claimed Photon was 800% faster and cheaper than comparable models at launch.
Photon supports a large context window for visual generative models and can generate consistent characters from a single input reference image. The model is available through the Dream Machine web interface and via API.
Luma AI's original product was a 3D capture tool built on neural radiance field technology. The iOS app allows users to capture 3D scenes by walking around a subject while recording video on an iPhone (model 11 or newer). The AI processes the video to reconstruct 3D geometry, lighting, and material properties.
Key capabilities of the 3D capture system include:
The Unreal Engine plugin (released April 2023) allows photorealistic NeRF and Gaussian Splatting captures to be imported, cropped, relit, and combined with traditional Unreal Engine VFX and real-time graphics.
Luma AI provides a developer API for integrating video and image generation into third-party applications. API credits are purchased separately from Dream Machine subscription credits.
| API Plan | Monthly Credits | Features |
|---|---|---|
| Free | Limited | Basic text-to-video |
| Creator | $10 free credits | Unlimited task speed, email support, file-to-URL conversion |
| Pro | $60 free credits | All task types, live support, custom invoicing |
| Enterprise | Custom | Dedicated infrastructure, SLAs, volume pricing |
The API supports text-to-video, image-to-video, video extension, end frame specification, and watermark removal. The Ray2 model is also available through Amazon Bedrock, allowing AWS customers to access Luma's models through a managed API.
| Name | Role | Background |
|---|---|---|
| Amit Jain | Co-founder, CEO | Former Apple Systems and ML Engineer; led Passthrough for Apple Vision Pro and LiDAR integration for iPhone; B.S. in Mathematics and Computer Science from Missouri Valley College |
| Alex Yu | Co-founder, CTO | UC Berkeley researcher; studied NeRF neural rendering under Professor Angjoo Kanazawa; turned down PhD offers from Stanford and MIT to start Luma; authored papers on real-time neural rendering and single-image 3D generation |
| Alberto Taiuti | Co-founder, CTO | Former Apple AR/VR Engineer (2 years); former Senior Autonomy Software Engineer at Skydio; Abertay University alumnus |
| Jiaming Song | Chief Scientist | Ph.D. in Computer Science from Stanford (advisor: Stefano Ermon); inventor of DDIM; former Principal Research Scientist at NVIDIA Deep Imagination group; joined Luma AI in July 2023 |
| Matthew Tancik | Applied Research Lead | Co-author of the original NeRF paper; co-founder of nerfstudio; joined from UC Berkeley in June 2023 |
| Jason Day | Head of EMEA | Former executive at Monks and WPP; appointed December 2025 to lead London office and international expansion |
As of February 2026, Luma AI has approximately 246 employees at its Palo Alto headquarters, with plans to grow to over 400 employees globally through London and Riyadh office expansions.
Luma AI has raised over $1.07 billion across six funding rounds.
| Round | Date | Amount | Lead Investor(s) | Notable Participants | Post-money Valuation |
|---|---|---|---|---|---|
| Seed | October 30, 2021 | $4.3M | Amplify Partners | - | - |
| Series A | March 20, 2023 | $20M | Amplify Partners | NVentures, General Catalyst | $100M |
| Series B | January 9, 2024 | $43M | Andreessen Horowitz (a16z) | - | - |
| Series C (initial) | December 6, 2024 | $90M | - | Amazon, AMD, a16z, Amplify Partners, Matrix Partners | - |
| Series C (extended) | November 19, 2025 | $900M | HUMAIN (PIF) | AMD Ventures, a16z, Amplify Partners, Matrix Partners | $4B+ |
The company has attracted a total of 17 investors, including Amazon, General Catalyst, NVentures, Amplify Partners, a16z, AMD Ventures, and HUMAIN.
Luma AI and AWS formed a strategic partnership in late 2024. AWS became the preferred compute partner for Luma AI, and Luma's Ray2 model was integrated into Amazon Bedrock. Luma trains its foundation models on Amazon SageMaker HyperPod and plans to bring models to AWS Trainium and Inferentia chips for lower-cost training and inference.
Adobe integrated Luma AI's Ray3 model into Adobe Firefly and Firefly Boards in September 2025. This was Luma's first third-party distribution partnership. Ray2 was also subsequently added to Firefly in April 2025. The integration allows users to generate video within Adobe's creative tools and sync content to Premiere Pro for editing.
Through the $900 million Series C investment in November 2025, Luma AI became a customer of HUMAIN's Project Halo, a 2-gigawatt AI supercluster in Saudi Arabia. The partnership includes development of Arabic-native foundation models under the HUMAIN Create initiative and Luma AI's plans to open a Riyadh office.
In February 2026, Publicis Groupe Middle East became Luma AI's preferred AI creative partner across the MENA region, integrating Luma's generative video technologies into advertising and brand production workflows.
Luma AI operates in the AI video generation market alongside several other companies.
| Company | Product | Strengths | Typical Use Cases |
|---|---|---|---|
| Luma AI | Dream Machine (Ray models) | Cinematic realism, natural motion, physics simulation, fast generation, HDR support | Motion-intensive content, cinematic storytelling, product videos |
| Runway | Gen-3 / Gen-4 | Motion Brush, Director Mode, Camera Control, fine-grained editing | Brand advertising, 4K master files, precise frame control |
| OpenAI | Sora | ChatGPT integration, synchronized dialogue and sound, long-form generation | Realistic scenes, imaginative storytelling, dialogue-driven content |
| Pika Labs | Pika | Fast rendering, low cost, stylistic expression, social-first | Social media content, meme animations, dance clips |
| Google DeepMind | Veo | High physical realism, long-form video, integration with Google tools | Professional video production |
| Kuaishou | Kling | Strong motion quality, competitive pricing | General video generation |
Luma AI holds an estimated 15-20% market share in AI video generation as of late 2025, positioning it between Pika's experimental tools and Runway's professional offerings. Dream Machine's lower entry price ($9.99 for the Lite plan) compared to competitors like Runway ($15 for Standard) has contributed to rapid user growth.
Luma AI's tools are used across several industries:
Luma AI's video models are built on a multi-modal transformer architecture. When processing a text prompt, the model decomposes the description into separate components: scene layout, motion dynamics, lighting conditions, and camera behavior. The model then generates video by predicting the temporal evolution of the scene, maintaining frame-to-frame consistency with physical properties learned during training.
Key technical characteristics of the Ray model family include:
For 3D capture, Luma AI uses both NeRF (neural radiance field) representations and Gaussian Splatting. NeRF represents scenes as continuous volumetric functions, while Gaussian Splatting uses explicit 3D Gaussian primitives for faster rendering. Both methods can be exported for use in game engines and 3D software.