A Dataless Reinforcement Learning Approach to Rounding Hyperplane Optimization for Max-Cut
Maliakal, Gabriel, Alkhouri, Ismail, Velasquez, Alvaro, Alessio, Adam M, Ravishankar, Saiprasad
The Maximum Cut (MaxCut) problem is NP-Complete, and obtaining its optimal solution is NP-hard in the worst case. As a result, heuristic-based algorithms are commonly used, though their design often requires significant domain expertise. More recently, learning-based methods trained on large (un)labeled datasets have been proposed; however, these approaches often struggle with generalizability and scalability. A well-known approximation algorithm for MaxCut is the Goemans-Williamson (GW) algorithm, which relaxes the Quadratic Unconstrained Binary Optimization (QUBO) formulation into a semidefinite program (SDP). The GW algorithm then applies hyperplane rounding by uniformly sampling a random hyperplane to convert the SDP solution into binary node assignments. In this paper, we propose a training-data-free approach based on a non-episodic reinforcement learning formulation, in which an agent learns to select improved rounding hyperplanes that yield better cuts than those produced by the GW algorithm. By optimizing over a Markov Decision Process (MDP), our method consistently achieves better cuts across large-scale graphs with varying densities and degree distributions.
Jun-17-2025
- Country:
- North America > United States
- California > Los Angeles County
- Pasadena (0.04)
- Colorado > Boulder County
- Boulder (0.04)
- Michigan > Washtenaw County
- Ann Arbor (0.04)
- California > Los Angeles County
- North America > United States
- Genre:
- Research Report (0.40)