Solving Constrained Optimization Problems by LSTM Networks

Abstract Many technical issues encountered by coalition forces, including SDC resource sharing, network management, distributed analytics and machine learning, can be formulated as optimization problems. Gradient-based iterative algorithms have been widely utilized to solve these problems. The core difference among various algorithms is the way to update the solution based on the current gradient values in each iteration. Instead of designing a new method by exploiting the structures of the involved functions, employing machine-learning techniques to learn how to design an efficient solution framework for optimization problems is of interest. We propose here a learning approach to solve non-convex, constrained stochastic optimization problems with user-defined objective and constraint functions. Two coupled Long Short-Term Memory (LSTM) networks are used to find the optimal solution. The new approach is validated by re-producing the solution to the non-convex problem for in-network data processing [1]. Efforts are in progress to extend the new technique for outstanding stochastic optimization problems and systems with temporal dynamics (e.g., SDC). Our latest results confirm the feasibility of the LSTMs to include temporal system evolution, while providing the optimal control decisions. Furthermore, it is also shown to solve some forms of stochastic optimization problems.
Authors
  • Zheyu (Joe) Chen (Imperial)
  • Kin Leung (Imperial)
  • Shiqiang Wang (IBM US)
  • Leandros Tassiulas (Yale)
  • Kevin Chan (ARL)
Date Sep-2020
Venue 4th Annual Fall Meeting of the DAIS ITA, 2020