MENU

KAIST develops AI that generates original designs

KAIST develops AI that generates original designs

Technology News |
By Jean-Pierre Joosting



KAIST researchers have developed a technology that can enhance the creativity of text-based AI image generation models, such as Stable Diffusion, without requiring additional training.

A research team led by Professor Jaesik Choi at the KAIST Kim Jaechul Graduate School of AI, in collaboration with NAVER AI Lab, has developed this technology to enhance the creative generation of AI generative models without the need for additional training.

The AI technology enhances creative generation by amplifying the internal feature maps of text-based image generation models. The researchers also discovered that shallow blocks within the model play a crucial role in creative generation. They confirmed that amplifying values in the high-frequency region after converting feature maps to the frequency domain can lead to noise or fragmented colour patterns. The team demonstrated that amplifying the low-frequency region of shallow blocks can effectively enhance creative generation.

Considering originality and usefulness as two key elements defining creativity, the research team proposed an algorithm that automatically selects the optimal amplification value for each block within the generative model.

Through the algorithm, the appropriate amplification of the internal feature maps of a pre-trained Stable Diffusion model enhanced creative generation without requiring additional classification data or training.

The research team quantitatively proved, using various metrics, that their developed AI algorithm can generate images that are more novel than those from existing models, without significantly compromising utility.

In particular, they confirmed an increase in image diversity by mitigating the mode collapse problem that occurs in the SDXL-Turbo model, which was developed to significantly improve the image generation speed of the Stable Diffusion XL (SDXL) model. Furthermore, user studies found that human evaluation also confirmed a significant improvement in novelty relative to utility compared to existing methods.

Jiyeon Han and Dahee Kwon, Ph.D. candidates at KAIST and co-first authors of the paper, stated, “This is the first methodology to enhance the creative generation of generative models without new training or fine-tuning. We have shown that the latent creativity within trained AI generative models can be enhanced through feature map manipulation.”

They added, “This research makes it easy to generate creative images using only text from existing trained models. It is expected to provide new inspiration in various fields, such as creative product design, and contribute to the practical and useful application of AI models in the creative ecosystem.”

Image: Professor Jaesik Choi, KAIST Kim Jaechul Graduate School of AI. Credit: Statistical Artificial Intelligence Lab, KAIST.

Paper: https://doi.org/10.48550/arXiv.2503.23538

If you enjoyed this article, you will like the following ones: don't miss them by subscribing to :    eeNews on Google News

Share:

Linked Articles
10s