no code implementations • 21 Feb 2024 • Yongcun Song, Ziqi Wang, Enrique Zuazua
First, we design an inexactness criterion for the clients' local updates to eliminate the need for empirically setting the local training accuracy.
no code implementations • 18 Jan 2024 • Antonio Álvarez-López, Arselane Hadj Slimane, Enrique Zuazua
Our findings reveal a balancing trade-off between $p$ and $L$, with $L$ scaling as $O(1+N/p)$ for dataset interpolation, and $L=O\left(1+(p\varepsilon^d)^{-1}\right)$ for measure interpolation.
no code implementations • 21 Dec 2023 • Antonio Álvarez-López, Rafael Orive-Illera, Enrique Zuazua
Classification of $N$ points becomes a simultaneous control problem when viewed through the lens of neural ordinary differential equations (neural ODEs), which represent the time-continuous limit of residual networks.
no code implementations • 13 Aug 2023 • Yongcun Song, Ziqi Wang, Enrique Zuazua
Federated Learning (FL) is a distributed learning paradigm that enables multiple clients to collaborate on building a machine learning model without sharing their private data.
no code implementations • 8 Feb 2022 • Borjan Geshkovski, Enrique Zuazua
The \emph{turnpike property} in contemporary macroeconomics asserts that if an economic planner seeks to move an economy from one level of capital to another, then the most efficient path, as long as the planner has enough time, is to rapidly move stock to a level close to the optimal stationary or constant path, then allow for capital to develop along that path until the desired term is nearly reached, at which point the stock ought to be moved to the final target.
no code implementations • 2 Jan 2021 • Yesim Sarac, Enrique Zuazua
We analyze the sidewise controllability for the variable coefficients one-dimensional wave equation.
Optimization and Control
1 code implementation • 6 Aug 2020 • Carlos Esteve, Borjan Geshkovski, Dario Pighin, Enrique Zuazua
We consider the neural ODE perspective of supervised learning and study the impact of the final time $T$ (which may indicate the depth of a corresponding ResNet) in training.