This article reports an algorithm for multiagent distributed optimization problems with a common decision variable, local linear equality, and inequality, constraints and set constraints with convergence rate guarantees. The algorithm accrues all the benefits of the alternating direction method of multipliers (ADMM) approach. It also overcomes the limitations of existing methods on convex optimization problems with linear inequality, equality, and set constraints by allowing directed communication topologies. Moreover, the algorithm can be synthesized distributively. The developed algorithm has: first, a O(1/k) rate of convergence, where k is the iteration counter, when individual functions are convex but not-necessarily differentiable, and second, a geometric rate of convergence to any arbitrary small neighborhood of the optimal solution, when the objective functions are smooth and restricted strongly convex at the optimal solution. The efficacy of the algorithm is evaluated by a comparison with state-of-The-Art constrained optimization algorithms in solving a constrained distributed-regularized logistic regression problem, and unconstrained optimization algorithms in solving a-regularized Huber loss minimization problem. Additionally, a comparison of the algorithm's performance with other algorithms in the literature that utilize multiple communication steps is provided.
Bibliographical notePublisher Copyright:
- Alternating direction method of multipliers (ADMM)
- constrained optimization
- directed graphs
- distributed optimization
- finite-Time consensus
- multiagent networks