11 December 2024 Adversarial capsule autoencoder with style vectors for image generation
Xiufeng Liu, Yi Yang, Zhongqiu Zhao, Zhao Zhang
Author Affiliations +
Abstract

Capsule networks get achievements in many computer vision tasks. However, in the field of image generation, they have huge room for improvement compared with the mainstream models. This is because capsule networks cannot fully parse useful features and have limited capabilities of modeling the hierarchical and geometrical structure of the object in background noise. To tackle these issues, we propose a capsule autoencoder that can learn the part–object spatial hierarchical features, and we dub it an adversarial capsule autoencoder with style vector (StyleACAE). Specifically, StyleACAE decomposes the object into a set of semantic-consistent part-level descriptions and then assembles them into object-level descriptions to learn the spatial hierarchy information. Furthermore, we effectively apply the modified generator structure that contains novel style modulation and demodulation, which extracts the valid information of the capsules and eliminates the scaled effect of styles. The experimental results show that StyleACAE can generate high-quality images and has competitive performance to the state-of-the-art generative models.

© 2024 SPIE and IS&T
Xiufeng Liu, Yi Yang, Zhongqiu Zhao, and Zhao Zhang "Adversarial capsule autoencoder with style vectors for image generation," Journal of Electronic Imaging 33(6), 063044 (11 December 2024). https://doi.org/10.1117/1.JEI.33.6.063044
Received: 15 May 2024; Accepted: 28 October 2024; Published: 11 December 2024
Advertisement
Advertisement
RIGHTS & PERMISSIONS
Get copyright permission  Get copyright permission on Copyright Marketplace
KEYWORDS
Modulation

Demodulation

RGB color model

Gallium nitride

Data modeling

Adversarial training

Performance modeling

Back to Top