Google’s Masked Generative Transformers Achieve SOTA Text-To-Image Performance With Improved Efficiency | Synced

In the new paper Muse: Text-To-Image Generation via Masked Generative Transformers, a Google Research team introduces Muse, a transformer-based text-to-image synthesis model that leverages masked i...

By · · 1 min read

Source: Synced | AI Technology & Industry Review

In the new paper Muse: Text-To-Image Generation via Masked Generative Transformers, a Google Research team introduces Muse, a transformer-based text-to-image synthesis model that leverages masked image modelling to achieve state-of-the-art performance while being significantly faster than diffusion or autoregressive models.