Generative flows with invertible attentions
Generative flows with invertible attentions
Flow-based generative models have shown an excellent ability to explicitly learn the probability density function of data via a sequence of invertible transformations. Yet, learning attentions in generative flows remains understudied, while it has made breakthroughs in other domains. To fill the gap, this paper introduces two types of invertible attention mechanisms, i.e., map-based and transformer-based attentions, for both unconditional and conditional generative flows. The key idea is to exploit a masked scheme of these two attentions to learn long-range data dependencies in the context of generative flows. The masked scheme allows for invertible attention modules with tractable Jacobian determinants, enabling its seamless integration at any positions of the flow-based models. The proposed attention mechanisms lead to more efficient generative flows, due to their capability of modeling the long-term data dependencies. Evaluation on multiple image synthesis tasks shows that the proposed attention flows result in efficient models and compare favorably against the state-of-the-art unconditional and conditional generative flows.
11224-11233
Sukthanker, Rhea Sanjay
efad96cb-52a9-4a32-bae2-1df2d009e323
Huang, Zhiwu
84f477cd-9097-44dd-a33e-ff71f253d36b
Kumar, Suryansh
048d50f7-9b57-4d05-940e-fe7e43626d46
Timofte, Radu
c6b34cfa-2aab-4e6d-9882-a65e24b4a4d0
Van Gool, Luc
ca42ab81-b646-4992-8c7c-c72a28247b90
21 June 2022
Sukthanker, Rhea Sanjay
efad96cb-52a9-4a32-bae2-1df2d009e323
Huang, Zhiwu
84f477cd-9097-44dd-a33e-ff71f253d36b
Kumar, Suryansh
048d50f7-9b57-4d05-940e-fe7e43626d46
Timofte, Radu
c6b34cfa-2aab-4e6d-9882-a65e24b4a4d0
Van Gool, Luc
ca42ab81-b646-4992-8c7c-c72a28247b90
Sukthanker, Rhea Sanjay, Huang, Zhiwu, Kumar, Suryansh, Timofte, Radu and Van Gool, Luc
(2022)
Generative flows with invertible attentions.
In The IEEE / CVF Computer Vision and Pattern Recognition Conference.
.
(doi:10.1109/CVPR52688.2022.01095).
Record type:
Conference or Workshop Item
(Paper)
Abstract
Flow-based generative models have shown an excellent ability to explicitly learn the probability density function of data via a sequence of invertible transformations. Yet, learning attentions in generative flows remains understudied, while it has made breakthroughs in other domains. To fill the gap, this paper introduces two types of invertible attention mechanisms, i.e., map-based and transformer-based attentions, for both unconditional and conditional generative flows. The key idea is to exploit a masked scheme of these two attentions to learn long-range data dependencies in the context of generative flows. The masked scheme allows for invertible attention modules with tractable Jacobian determinants, enabling its seamless integration at any positions of the flow-based models. The proposed attention mechanisms lead to more efficient generative flows, due to their capability of modeling the long-term data dependencies. Evaluation on multiple image synthesis tasks shows that the proposed attention flows result in efficient models and compare favorably against the state-of-the-art unconditional and conditional generative flows.
This record has no associated files available for download.
More information
Published date: 21 June 2022
Venue - Dates:
The IEEE / CVF Computer Vision and Pattern Recognition Conference (CVPR), , Louisiana, United States, 2022-06-19
Identifiers
Local EPrints ID: 501680
URI: http://eprints.soton.ac.uk/id/eprint/501680
PURE UUID: 62e05ee6-f498-419b-aa67-101acc0c32db
Catalogue record
Date deposited: 05 Jun 2025 16:57
Last modified: 06 Jun 2025 02:06
Export record
Altmetrics
Contributors
Author:
Rhea Sanjay Sukthanker
Author:
Zhiwu Huang
Author:
Suryansh Kumar
Author:
Radu Timofte
Author:
Luc Van Gool
Download statistics
Downloads from ePrints over the past year. Other digital versions may also be available to download e.g. from the publisher's website.
View more statistics