Revolutionizing image generation by AI: Turning text into images

LMU München

Image generated from the text "Happy vegetables waiting for supper".

  • The new Stable Diffusion AI model does not require a supercomputer and is available free of charge to all users.
  • The AI model distils the essence of billions of training images into a few gigabytes.

Creating images from text in seconds - and doing so with a conventional graphics card and without supercomputers? As fanciful as it may sound, this is made possible by the new Stable Diffusion AI model. The underlying algorithm was developed by the Machine Vision & Learning Group led by Prof. Björn Ommer (LMU Munich).

"Even for laypeople not blessed with artistic talent and without special computing know-how and computer hardware, the new model is an effective tool that enables computers to generate images on command. As such, the model removes a barrier to ordinary people expressing their creativity," says Ommer. But there are benefits for seasoned artists as well, who can use Stable Diffusion to quickly convert new ideas into a variety of graphic drafts. The researchers are convinced that such AI-based tools will be able to expand the possibilities of creative image generation with paintbrush and Photoshop as fundamentally as computer-based word processing revolutionized writing with pens and typewriters.

In their project, the LMU scientists had the support of the start-up Stability.Ai, on whose servers the AI model was trained. "This additional computing power and the extra training examples turned our AI model into one of the most powerful image synthesis algorithms," says the computer scientist with a smile.

/Courtesy of LMU München. This material from the originating organization/author(s) might be of the point-in-time nature, and edited for clarity, style and length. Mirage.News does not take institutional positions or sides, and all views, positions, and conclusions expressed herein are solely those of the author(s).