MCGAN mask controlled generative adversarial network for image retargeting

Jilyan Bianca Dy, John Jethro Virtusio, Daniel Stanley Tan, Yong Xiang Lin, Joel Ilao, Yung Yao Chen, Kai Lung Hua*

*Corresponding author for this work

Research output: Contribution to journalArticleAcademicpeer-review

Abstract

Image retargeting aims to resize a photograph without distorting its content. Recent solutions leverage generative adversarial networks (GANs) to learn an image’s internal distribution. Patches are then replicated and generated to fill the target aspect ratio to preserve the internal distribution. Although these approaches are somewhat successful, they do not have a sense of image semantics or the image’s contents, causing semantic errors during generation (e.g., a person with two heads). Our model addresses this by allowing user intervention. The users can preserve the objects they desire through user-defined masks. Our model enforces the masked object to appear at the user-defined location. It also utilizes a de-association regularizer to loosen the association of the selected object with its surroundings. Our method prevents object distortion and enables the user to remove or relocate the object from the input image while allowing the user to set its target size.

Original languageEnglish
Pages (from-to)10497-10509
Number of pages13
JournalNeural Computing and Applications
Volume35
Issue number14
DOIs
Publication statusPublished - May 2023

Keywords

  • Computer vision
  • Conditional GAN
  • Controlled GAN
  • Image retargeting

Fingerprint

Dive into the research topics of 'MCGAN mask controlled generative adversarial network for image retargeting'. Together they form a unique fingerprint.

Cite this