1. Gradient Networks
- Author
-
Chaudhari, Shreyas, Pranav, Srinivasa, and Moura, Jose M.F.
- Abstract
Directly parameterizing and learning gradients of functions has widespread significance, with specific applications in inverse problems, generative modeling, and optimal transport. This paper introduces gradient networks (
GradNets ): novel neural network architectures that parameterize gradients of various function classes.GradNets exhibit specialized architectural constraints that ensure correspondence to gradient functions. We provide a comprehensiveGradNet design framework that includes methods for transformingGradNets into monotone gradient networks (mGradNets ), which are guaranteed to represent gradients of convex functions. Our results establish that our proposedGradNet (andmGradNet ) universally approximate the gradients of (convex) functions. Furthermore, these networks can be customized to correspond to specific spaces of potential functions, including transformed sums of (convex) ridge functions. Our analysis leads to two distinctGradNet architectures,GradNet-C andGradNet-M , and we describe the corresponding monotone versions,mGradNet-C andmGradNet-M . Our empirical results demonstrate that these architectures provide efficient parameterizations and outperform existing methods by up to 15 dB in gradient field tasks and by up to 11 dB in Hamiltonian dynamics learning tasks.- Published
- 2025
- Full Text
- View/download PDF