Table of Links
-
Convex Relaxation Techniques for Hyperbolic SVMs
B. Solution Extraction in Relaxed Formulation
C. On Moment Sum-of-Squares Relaxation Hierarchy
E. Detailed Experimental Results
F. Robust Hyperbolic Support Vector Machine
C On Moment Sum-of-Squares Relaxation Hierarchy
In this section, we provide necessary background on moment-sum-of-squares hierarchy. We start by considering a general Polynomial Optimization Problem (POP) and introduce the sparse version. This section borrows substantially from the course note [4].
C.1 Polynomial Optimization and Dual Cones
Polynomial optimization problem (POP) in the most generic form can be presented as
where π(π₯) is our polynomial objective and βπ(π₯), ππ(π₯) are our polynomial equality and inequality constraints respectively. However, in general, solving such POP to global optimality is NP-hard [9, 29]. To address this challenge, we leverage methods from algebraic geometry [9, 21], allowing us to approximate global solutions using convex optimization methods.
To start with, we define sum-of-squares (SOS) polynomials as polynomials that could be expressed as a sum of squares of some other polynomials, and we define Ξ£[π₯] to be the collection of SOS polynomials. More formally, we have
where R[π₯] denotes the polynomial ring over R.
Next, we recall the definitions of quadratic module and its dual. Given a set of polynomials g = [π1, π2, ..., ππ], the quadratic module generated by g is defined as
and its degree 2π-truncation is defined as,
Similarly, given a set of polynomials h = [β1, β2, ..., βπ], the ideal generated by h is defined as,
and its degree 2π-truncation is defined as,
where ππ βs are also called polynomial multipliers. Interestingly, it is shown that we can perfectly characterize the dual of the sum of ideal and quadratic module,
(Ideal[h]2π + Qmodule[g]2π) β = π΅[β]2π β© β³[π]2π ,
With these notions setup, we can reformulate the POP above into the following SOS program for arbitrary π β N as the relaxataion order,
C.2 Sparse Polynomial Optimization
Definition 1. Correlated Sparsity for an objective π β R[π₯] and associated set of constraints means
With correlated sparsity and data regularity (Putinarβs Positivestellentz outlined in Nie [9]), we are able to decompose the Qmodule generated by the entire set of decision variables into the Minkowski sum of Qmodules generated by each sparsity group of variables, effectively reducing the number of decision variables in the implementations. For a problem with only inequality constraints, which is our case for HSVM, the sparse POP for our problem reads as
and we could derive its dual accordingly and present the SDP form for implementation in Equation (14).
Authors:
(1) Sheng Yang, John A. Paulson School of Engineering and Applied Sciences, Harvard University, Cambridge, MA ([email protected]);
(2) Peihan Liu, John A. Paulson School of Engineering and Applied Sciences, Harvard University, Cambridge, MA ([email protected]);
(3) Cengiz Pehlevan, John A. Paulson School of Engineering and Applied Sciences, Harvard University, Cambridge, MA, Center for Brain Science, Harvard University, Cambridge, MA, and Kempner Institute for the Study of Natural and Artificial Intelligence, Harvard University, Cambridge, MA ([email protected]).
This paper is
[4] Chapter 5 Moment Relaxation: https://hankyang.seas.harvard.edu/Semidefinite/Moment.html
