AiPhreaks ← Back to News Feed

PRX Part 3 — Training a Text-to-Image Model in 24h!

By Jakub Antkiewicz

2026-03-04T08:39:24Z

The research team at Photoroom has demonstrated the ability to train a competent, high-resolution text-to-image model in just 24 hours with a compute budget of approximately $1,500. This achievement marks a significant reduction in the resources typically associated with developing foundation models, suggesting that the barrier to entry for creating competitive generative systems is becoming lower than previously thought. By open-sourcing the underlying code, the team provides a tangible recipe for how engineering efficiency can substitute for massive capital investment in compute.

The accelerated training run was conducted on a cluster of 32 H200 GPUs and relied on a carefully curated stack of modern techniques. Instead of using a VAE, the model was trained directly in pixel space, a method made computationally feasible by controlled token counts. The team integrated perceptual losses using LPIPS and DINOv2 to improve convergence and visual quality. To manage costs, they employed TREAD token routing, which selectively bypasses transformer blocks for a portion of tokens, and used the Muon optimizer for key parameters, citing improved performance over standard Adam.

This result signals a potential shift in the AI ecosystem, where the competitive landscape may become less defined by access to massive GPU clusters and more by technical expertise in model architecture and training optimization. The ability to create a functional model so affordably and quickly empowers smaller companies and research labs to pursue development of specialized, in-house models. This could foster a broader range of AI applications tailored to specific needs, reducing reliance on a handful of large, general-purpose models from major providers.

The Photoroom speedrun indicates that the primary moat in generative model development is shifting from raw compute access to the sophisticated integration of optimized training techniques. Competitive advantage increasingly lies not just in capital expenditure, but in the engineering acumen required to efficiently combine methods like pixel-space training, perceptual guidance, and efficient token routing.