How deep do we need: accelerating training and inference of neural ODEs via control perspective
How deep do we need: accelerating training and inference of neural ODEs via control perspective
Neural Ordinary Differential Equations (ODEs) have shown promise in learning continuous dynamics. However, their slow training and inference speed hinder wider applications. In this paper, we propose to optimize Neural ODEs from a spatial and temporal perspective, drawing inspiration from control theory. We aim to find a reasonable depth of the network, accelerating both training and inference while maintaining network performance. Two approaches are proposed. One reformulates training as a minimum-time optimal control problem directly in a single stage to search for the terminal time and network weights. The second approach uses pre-training coupled with a Lyapunov method in an initial stage, and then at a secondary stage introduces a safe terminal time updating mechanism in the forward direction. Experimental results demonstrate the effectiveness of speeding up Neural ODEs.
35528-35545
Miao, Keyan
bb159bd5-cc62-487c-bd79-bb9fcd6d5049
Gatsis, Konstantinos
f808d11b-38f1-4a44-ba56-3364d63558d7
2024
Miao, Keyan
bb159bd5-cc62-487c-bd79-bb9fcd6d5049
Gatsis, Konstantinos
f808d11b-38f1-4a44-ba56-3364d63558d7
Miao, Keyan and Gatsis, Konstantinos
(2024)
How deep do we need: accelerating training and inference of neural ODEs via control perspective.
Salakhutdinov, R., Kolter, Z., Heller, K., Weller, A., Oliver, N., Scarlett, J. and Berkenkamp, F.
(eds.)
In 41st International Conference on Machine Learning, ICML 2024.
vol. 235,
ML Research Press.
.
Record type:
Conference or Workshop Item
(Paper)
Abstract
Neural Ordinary Differential Equations (ODEs) have shown promise in learning continuous dynamics. However, their slow training and inference speed hinder wider applications. In this paper, we propose to optimize Neural ODEs from a spatial and temporal perspective, drawing inspiration from control theory. We aim to find a reasonable depth of the network, accelerating both training and inference while maintaining network performance. Two approaches are proposed. One reformulates training as a minimum-time optimal control problem directly in a single stage to search for the terminal time and network weights. The second approach uses pre-training coupled with a Lyapunov method in an initial stage, and then at a secondary stage introduces a safe terminal time updating mechanism in the forward direction. Experimental results demonstrate the effectiveness of speeding up Neural ODEs.
Text
8996_How_Deep_Do_We_Need_Accel
- Version of Record
More information
Published date: 2024
Venue - Dates:
41st International Conference on Machine Learning, ICML 2024, , Vienna, Austria, 2024-07-21 - 2024-07-27
Identifiers
Local EPrints ID: 494683
URI: http://eprints.soton.ac.uk/id/eprint/494683
ISSN: 2640-3498
PURE UUID: f77f4723-5dd1-4223-9b10-010665d4be2d
Catalogue record
Date deposited: 14 Oct 2024 16:35
Last modified: 15 Oct 2024 02:09
Export record
Contributors
Author:
Keyan Miao
Author:
Konstantinos Gatsis
Editor:
R. Salakhutdinov
Editor:
Z. Kolter
Editor:
K. Heller
Editor:
A. Weller
Editor:
N. Oliver
Editor:
J. Scarlett
Editor:
F. Berkenkamp
Download statistics
Downloads from ePrints over the past year. Other digital versions may also be available to download e.g. from the publisher's website.
View more statistics