Text-to-image synthesis with self-supervised bi-stage generative adversarial network

Yong Xuan Tan, Chin Poo Lee, Mai Neo, Kian Ming Lim, Jit Yan Lim

Research output: Journal PublicationArticlepeer-review

5 Citations (Scopus)

Abstract

Text-to-image synthesis is challenging as generating images that are visually realistic and semantically consistent with the given text description involves multi-modal learning with text and image. To address the challenges, this paper presents a text-to-image synthesis model that utilizes self-supervision and bi-stage image distribution architecture, referred to as the Self-Supervised Bi-Stage Generative Adversarial Network (SSBi-GAN). The self-supervision diversifies the learned representation thus improving the quality of the synthesized images. Besides that, the bi-stage architecture with Residual network enables the generation of larger images with finer visual contents. Not only that, some enhancements including L1 distance, one-sided smoothing and feature matching are incorporated to enhance the visual realism and semantic consistency of the images as well as the training stability of the model. The empirical results on Oxford-102 and CUB datasets corroborate the ability of the proposed SSBi-GAN in generating visually realistic and semantically consistent images.

Original languageEnglish
Pages (from-to)43-49
Number of pages7
JournalPattern Recognition Letters
Volume169
DOIs
Publication statusPublished - May 2023
Externally publishedYes

Keywords

  • GAN
  • Generative adversarial network
  • Self-supervised learning
  • Text-to-image-synthesis

ASJC Scopus subject areas

  • Software
  • Signal Processing
  • Computer Vision and Pattern Recognition
  • Artificial Intelligence

Fingerprint

Dive into the research topics of 'Text-to-image synthesis with self-supervised bi-stage generative adversarial network'. Together they form a unique fingerprint.

Cite this