Convex optimization learning of faithful Euclidean distance representations in nonlinear dimensionality reduction
Convex optimization learning of faithful Euclidean distance representations in nonlinear dimensionality reduction
Classical multidimensional scaling only works well when the noisy distances observed in a high dimensional space can be faithfully represented by Euclidean distances in a low dimensional space. Advanced models such as Maximum Variance Unfolding (MVU) and Minimum Volume Embedding (MVE) use Semi-Definite Programming (SDP) to reconstruct such faithful representations. While those SDP models are capable of producing high quality configuration numerically, they suffer two major drawbacks. One is that there exist no theoretically guaranteed bounds on the quality of the configuration. The other is that they are slow in computation when the data points are beyond moderate size. In this paper, we propose a convex optimization model of Euclidean distance matrices. We establish a non-asymptotic error bound for the random graph model with sub-Gaussian noise, and prove that our model produces a matrix estimator of high accuracy when the order of the uniform sample size is roughly the degree of freedom of a low-rank matrix up to a logarithmic factor. Our results partially explain why MVU and MVE often work well. Moreover, we develop a fast inexact accelerated proximal gradient method. Numerical experiments show that the model can produce configurations of high quality on large data points that the SDP approach would struggle to cope with.
341-381
Ding, Chao
67300df2-ae66-48d2-89a4-c488d148aa4a
Qi, Hou-Duo
e9789eb9-c2bc-4b63-9acb-c7e753cc9a85
July 2017
Ding, Chao
67300df2-ae66-48d2-89a4-c488d148aa4a
Qi, Hou-Duo
e9789eb9-c2bc-4b63-9acb-c7e753cc9a85
Ding, Chao and Qi, Hou-Duo
(2017)
Convex optimization learning of faithful Euclidean distance representations in nonlinear dimensionality reduction.
Mathematical Programming, 164 (1-2), .
(doi:10.1007/s10107-016-1090-7).
Abstract
Classical multidimensional scaling only works well when the noisy distances observed in a high dimensional space can be faithfully represented by Euclidean distances in a low dimensional space. Advanced models such as Maximum Variance Unfolding (MVU) and Minimum Volume Embedding (MVE) use Semi-Definite Programming (SDP) to reconstruct such faithful representations. While those SDP models are capable of producing high quality configuration numerically, they suffer two major drawbacks. One is that there exist no theoretically guaranteed bounds on the quality of the configuration. The other is that they are slow in computation when the data points are beyond moderate size. In this paper, we propose a convex optimization model of Euclidean distance matrices. We establish a non-asymptotic error bound for the random graph model with sub-Gaussian noise, and prove that our model produces a matrix estimator of high accuracy when the order of the uniform sample size is roughly the degree of freedom of a low-rank matrix up to a logarithmic factor. Our results partially explain why MVU and MVE often work well. Moreover, we develop a fast inexact accelerated proximal gradient method. Numerical experiments show that the model can produce configurations of high quality on large data points that the SDP approach would struggle to cope with.
Text
EDME_Revised_MP.pdf
- Accepted Manuscript
Text
art_10.1007_s10107-016-1090-7
- Version of Record
More information
Accepted/In Press date: 21 October 2016
e-pub ahead of print date: 11 November 2016
Published date: July 2017
Organisations:
Operational Research
Identifiers
Local EPrints ID: 391500
URI: http://eprints.soton.ac.uk/id/eprint/391500
ISSN: 0025-5610
PURE UUID: 6b547aa3-1a20-4480-b7da-0399d64dae19
Catalogue record
Date deposited: 13 May 2016 14:38
Last modified: 15 Mar 2024 05:29
Export record
Altmetrics
Contributors
Author:
Chao Ding
Download statistics
Downloads from ePrints over the past year. Other digital versions may also be available to download e.g. from the publisher's website.
View more statistics