Just noticed the following recent solver. From the main page:
DECOPT - DEcomposition Convex OPTimization
Overview
DECOPT is a MATLAB software package for solving the following generic constrained convex optimization problem:
minx∈Rp,r∈Rn{f(x)+g(r):Ax−r=b, l≤x≤u}, (CP) wheref andg are two proper, closed and convex functions,A∈Rn×p ,b∈Rn andl,u∈Rp are the lower and upper bounds ofx .Here, we assume thatf andg are proximally tractable. By proximal tractability, we mean that the proximal operatorproxφ of a given proper, closed and convex functionφ :proxφ(x):=argminy∈Rp{φ(y)+(1/2)∥y−x∥22} is efficient to compute (e.g., by a closed form solution or by polynomial time algorithms).DECOPT is implemented by Quoc Tran-Dinh at the Laboratory for Information and Inference Systems (LIONS), EPFL, Lausanne, Switzerland. This is a joint work with Volkan Cevher at LIONS, EPFL.DECOPT aims at solving (CP) for any convex functionsf andg , where their proximal operator is provided. The following special cases have been customized in DECOPT:
- Basis pursuit:
minx{∥x∥1:Ax=b,l≤x≤u}.
ℓ1/ℓ2 -unconstrained LASSO problem:
minx12∥Ax−b∥22+λ∥x∥1
ℓ1/ℓ1 -convex problem:
minx∥Ax−b∥1+λ∥x∥1.
- Square-root
ℓ1/ℓ2 LASSO problem:
minx∥Ax−b∥2+λ∥x∥1.
ℓ1/ℓ2 - the basis denosing (BPDN) problem:
minx{∥x∥1:∥Ax−b∥2≤δ}.
ℓ2/ℓ1 - theℓ1 -constrained LASSO problem:minx{(1/2)∥Ax−b∥22:∥x∥1≤δ}. Here,λ>0 is a penalty parameter,δ>0 is a given level parameter.
And the references:
References
The full algorithms and theory of DECOPT can be found in the following references:
Available at: http://infoscience.epfl.ch/record/201842
- Q. Tran-Dinh and V. Cevher, "Constrained convex minimization via model-based excessive gap". Proceedings of the annual conference on Neural Information Processing Systems Foundation (NIPS), Montreal, Canada, (2014).
Abstract: Abstract We introduce a model-based excessive gap technique to analyze first-order primal- dual methods for constrained convex minimization. As a result, we construct first-order primal-dual methods with optimal convergence rates on the primal objec-tive residual and the primal feasibility gap of their iterates separately. Through a dual smoothing and prox- center selection strategy, our framework subsumes the augmented Lagrangian, alternating direction, and dual fast-gradient methods as special cases, where our rates apply.
Available at: http://infoscience.epfl.ch/record/199844
- Q. Tran-Dinh, and V. Cevher. "A Primal-Dual Algorithmic Framework for Constrained Convex Minimization", LIONS Tech. Report EPFL-REPORT-199844, (2014).
Abstract: We present a primal-dual algorithmic framework to obtain approximate solutions to a prototypical constrained convex optimization problem, and rigorously characterize how common structural assumptions affect the numerical efficiency. Our main analysis technique provides a fresh perspective on Nesterov's excessive gap technique in a structured fashion and unifies it with smoothing and primal-dual methods. For instance, through the choices of a dual smoothing strategy and a center point, our framework subsumes decomposition algorithms, augmented Lagrangian as well as the alternating direction method-of-multipliers methods as its special cases, and provides optimal convergence rates on the primal objective residual as well as the primal feasibility gap of the iterates for all.
No comments:
Post a Comment