Nonlinear Optimization with GPU-Accelerated Neural Network Constraints

Generative AI & LLMs
Published: arXiv: 2509.22462v1
Authors

Robert Parker Oscar Dowson Nicole LoGiudice Manuel Garcia Russell Bent

Abstract

We propose a reduced-space formulation for optimizing over trained neural networks where the network's outputs and derivatives are evaluated on a GPU. To do this, we treat the neural network as a "gray box" where intermediate variables and constraints are not exposed to the optimization solver. Compared to the full-space formulation, in which intermediate variables and constraints are exposed to the optimization solver, the reduced-space formulation leads to faster solves and fewer iterations in an interior point method. We demonstrate the benefits of this method on two optimization problems: Adversarial generation for a classifier trained on MNIST images and security-constrained optimal power flow with transient feasibility enforced using a neural network surrogate.

Paper Summary

Problem
The main challenge this paper addresses is the scalability issue of solving optimization problems that involve large neural networks. Current methods for optimizing over trained machine learning models are limited to small neural network models, and it's difficult to apply them to larger models due to the complexity and computational cost.
Key Innovation
The key innovation of this paper is a reduced-space formulation for optimizing over trained neural networks, which exploits the efficiency of automatic differentiation and GPU acceleration. This method treats the neural network as a "gray box" where intermediate variables and constraints are not exposed to the optimization solver, leading to faster solves and fewer iterations.
Practical Impact
This research has significant practical implications for various applications, such as: * Generating adversarial examples for image classification models * Security-constrained optimal power flow in power grids * Optimization-based design and control of complex systems By enabling the efficient optimization of large neural networks, this work can improve the performance and robustness of these applications.
Analogy / Intuitive Explanation
Think of a neural network as a complex mathematical function that takes inputs and produces outputs. The optimization problem is like trying to find the optimal settings for this function to achieve a desired output. The reduced-space formulation is like using a shortcut to calculate the function's output, bypassing the need to explicitly solve for the intermediate variables and constraints. This shortcut allows for faster and more efficient optimization, making it possible to solve complex problems that were previously intractable.
Paper Information
Categories:
cs.LG
Published Date:

arXiv ID:

2509.22462v1

Quick Actions