The University of Southampton
University of Southampton Institutional Repository

Generative flows with invertible attentions

Generative flows with invertible attentions
Generative flows with invertible attentions
Flow-based generative models have shown an excellent ability to explicitly learn the probability density function of data via a sequence of invertible transformations. Yet, learning attentions in generative flows remains understudied, while it has made breakthroughs in other domains. To fill the gap, this paper introduces two types of invertible attention mechanisms, i.e., map-based and transformer-based attentions, for both unconditional and conditional generative flows. The key idea is to exploit a masked scheme of these two attentions to learn long-range data dependencies in the context of generative flows. The masked scheme allows for invertible attention modules with tractable Jacobian determinants, enabling its seamless integration at any positions of the flow-based models. The proposed attention mechanisms lead to more efficient generative flows, due to their capability of modeling the long-term data dependencies. Evaluation on multiple image synthesis tasks shows that the proposed attention flows result in efficient models and compare favorably against the state-of-the-art unconditional and conditional generative flows.
11224-11233
Sukthanker, Rhea Sanjay
efad96cb-52a9-4a32-bae2-1df2d009e323
Huang, Zhiwu
84f477cd-9097-44dd-a33e-ff71f253d36b
Kumar, Suryansh
048d50f7-9b57-4d05-940e-fe7e43626d46
Timofte, Radu
c6b34cfa-2aab-4e6d-9882-a65e24b4a4d0
Van Gool, Luc
ca42ab81-b646-4992-8c7c-c72a28247b90
Sukthanker, Rhea Sanjay
efad96cb-52a9-4a32-bae2-1df2d009e323
Huang, Zhiwu
84f477cd-9097-44dd-a33e-ff71f253d36b
Kumar, Suryansh
048d50f7-9b57-4d05-940e-fe7e43626d46
Timofte, Radu
c6b34cfa-2aab-4e6d-9882-a65e24b4a4d0
Van Gool, Luc
ca42ab81-b646-4992-8c7c-c72a28247b90

Sukthanker, Rhea Sanjay, Huang, Zhiwu, Kumar, Suryansh, Timofte, Radu and Van Gool, Luc (2022) Generative flows with invertible attentions. In The IEEE / CVF Computer Vision and Pattern Recognition Conference. pp. 11224-11233 . (doi:10.1109/CVPR52688.2022.01095).

Record type: Conference or Workshop Item (Paper)

Abstract

Flow-based generative models have shown an excellent ability to explicitly learn the probability density function of data via a sequence of invertible transformations. Yet, learning attentions in generative flows remains understudied, while it has made breakthroughs in other domains. To fill the gap, this paper introduces two types of invertible attention mechanisms, i.e., map-based and transformer-based attentions, for both unconditional and conditional generative flows. The key idea is to exploit a masked scheme of these two attentions to learn long-range data dependencies in the context of generative flows. The masked scheme allows for invertible attention modules with tractable Jacobian determinants, enabling its seamless integration at any positions of the flow-based models. The proposed attention mechanisms lead to more efficient generative flows, due to their capability of modeling the long-term data dependencies. Evaluation on multiple image synthesis tasks shows that the proposed attention flows result in efficient models and compare favorably against the state-of-the-art unconditional and conditional generative flows.

This record has no associated files available for download.

More information

Published date: 21 June 2022
Venue - Dates: The IEEE / CVF Computer Vision and Pattern Recognition Conference (CVPR), , Louisiana, United States, 2022-06-19

Identifiers

Local EPrints ID: 501680
URI: http://eprints.soton.ac.uk/id/eprint/501680
PURE UUID: 62e05ee6-f498-419b-aa67-101acc0c32db
ORCID for Zhiwu Huang: ORCID iD orcid.org/0000-0002-7385-079X

Catalogue record

Date deposited: 05 Jun 2025 16:57
Last modified: 06 Jun 2025 02:06

Export record

Altmetrics

Contributors

Author: Rhea Sanjay Sukthanker
Author: Zhiwu Huang ORCID iD
Author: Suryansh Kumar
Author: Radu Timofte
Author: Luc Van Gool

Download statistics

Downloads from ePrints over the past year. Other digital versions may also be available to download e.g. from the publisher's website.

View more statistics

Atom RSS 1.0 RSS 2.0

Contact ePrints Soton: eprints@soton.ac.uk

ePrints Soton supports OAI 2.0 with a base URL of http://eprints.soton.ac.uk/cgi/oai2

This repository has been built using EPrints software, developed at the University of Southampton, but available to everyone to use.

We use cookies to ensure that we give you the best experience on our website. If you continue without changing your settings, we will assume that you are happy to receive cookies on the University of Southampton website.

×