Compositional Conditional Diffusion Model

Shih Lun Lai, Pin Chuan Chen, Ching Wen Ma

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

Generative model often struggle to produce data beyond the training distribution. To address this, we explored various methods for unseen data generation, eventually focusing on compositional zero-shot image generation. By guiding the generation model with compositional class labels, we achieved better control over the generation process. Our model can generate unseen images whose compositional labels are not appear in the training set. While large language models like GPT-4 and image generation models like DALL-E offer similar zero-shot generation capabilities, our research emphasizes domain-specific zero-shot generation using smaller models. Through a series of tasks, we demonstrated the effectiveness of compositional zero-shot image generation across various complexities, showcasing its potential in contemporary machine learning.

Original languageEnglish
Title of host publicationProceedings of the 2024 10th International Conference on Applied System Innovation, ICASI 2024
EditorsShoou-Jinn Chang, Sheng-Joue Young, Artde Donald Kin-Tak Lam, Liang-Wen Ji, Stephen D. Prior
PublisherInstitute of Electrical and Electronics Engineers Inc.
Pages377-379
Number of pages3
ISBN (Electronic)9798350394924
DOIs
StatePublished - 2024
Event10th International Conference on Applied System Innovation, ICASI 2024 - Kyoto, Japan
Duration: 17 Apr 202421 Apr 2024

Publication series

NameProceedings of the 2024 10th International Conference on Applied System Innovation, ICASI 2024

Conference

Conference10th International Conference on Applied System Innovation, ICASI 2024
Country/TerritoryJapan
CityKyoto
Period17/04/2421/04/24

Keywords

  • Compositional Zero-shot Learning
  • Diffusion Model
  • Image Synthesis

Fingerprint

Dive into the research topics of 'Compositional Conditional Diffusion Model'. Together they form a unique fingerprint.

Cite this