Binno: a 1st-order method for Bi-level nonconvex nonsmooth optimization for matrix factorizations
Binno: a 1st-order method for Bi-level nonconvex nonsmooth optimization for matrix factorizations
In this work, we develop a method for nonconvex, nonsmooth bi-level optimization and we introduce Binno, a first order method that leverages proximal constructions together with carefully designed descent conditions and variational analysis. Within this framework, Binno provably enforces a descent property for the overall objective surrogate associated with the bi-level problem. Each iteration performs blockwise proximal-gradient updates for the upper and the lower problems separately and then forms a calibrated, block-diagonal convex combination of the two tentative iterates. A linesearch selects the combination weights to enforce simultaneous descent of both level-wise objectives, and we establish conditions guaranteeing the existence of such weights together with descent directions induced by the associated proximal-gradient maps. We also apply Binno in the context of sparse low-rank factorization, where the upper level uses elementwise $\ell_1$ penalties and the lower level uses nuclear norms, coupled via a Frobenius data term. We test Binno on synthetic matrix and a real traffic-video dataset, attaining lower relative reconstruction error and higher peak signal-to-noise ratio than some standard methods.
math.OC, 65K10, 90C26, 90C30
Selicato, Laura
77656e85-1d36-4974-b7fb-8fcf55f7a178
Esposito, Flavia
8dc4f35b-400e-4260-82ae-4cd8fbe2e680
Ang, Andersen
ed509ecd-39a3-4887-a709-339fdaded867
24 October 2025
Selicato, Laura
77656e85-1d36-4974-b7fb-8fcf55f7a178
Esposito, Flavia
8dc4f35b-400e-4260-82ae-4cd8fbe2e680
Ang, Andersen
ed509ecd-39a3-4887-a709-339fdaded867
[Unknown type: UNSPECIFIED]
Abstract
In this work, we develop a method for nonconvex, nonsmooth bi-level optimization and we introduce Binno, a first order method that leverages proximal constructions together with carefully designed descent conditions and variational analysis. Within this framework, Binno provably enforces a descent property for the overall objective surrogate associated with the bi-level problem. Each iteration performs blockwise proximal-gradient updates for the upper and the lower problems separately and then forms a calibrated, block-diagonal convex combination of the two tentative iterates. A linesearch selects the combination weights to enforce simultaneous descent of both level-wise objectives, and we establish conditions guaranteeing the existence of such weights together with descent directions induced by the associated proximal-gradient maps. We also apply Binno in the context of sparse low-rank factorization, where the upper level uses elementwise $\ell_1$ penalties and the lower level uses nuclear norms, coupled via a Frobenius data term. We test Binno on synthetic matrix and a real traffic-video dataset, attaining lower relative reconstruction error and higher peak signal-to-noise ratio than some standard methods.
Text
2510.21390v1
- Author's Original
More information
Published date: 24 October 2025
Additional Information:
29 pages, 5 figures
Keywords:
math.OC, 65K10, 90C26, 90C30
Identifiers
Local EPrints ID: 507059
URI: http://eprints.soton.ac.uk/id/eprint/507059
PURE UUID: 7ccc1376-5ee1-4ede-9d0e-8f5f3cc3f703
Catalogue record
Date deposited: 26 Nov 2025 17:36
Last modified: 27 Nov 2025 03:05
Export record
Altmetrics
Contributors
Author:
Laura Selicato
Author:
Flavia Esposito
Author:
Andersen Ang
Download statistics
Downloads from ePrints over the past year. Other digital versions may also be available to download e.g. from the publisher's website.
View more statistics