## Abstract

This paper is concerned with the alternating minimization (AM) method for solving convex minimization problems where the decision variables vector is split into two blocks. The objective function is a sum of a differentiable convex function and a separable (possibly) nonsmooth extended real-valued convex function, and consequently constraints can be incorporated. We analyze the convergence rate of the method and establish a nonasymptotic sublinear rate of convergence where the multiplicative constant depends on the minimal block Lipschitz constant. We then analyze the iteratively reweighted least squares (IRLS) method for solving convex problems involving sums of norms. Based on the results derived for the AM method, we establish a nonasymptotic sublinear rate of convergence of the IRLS method. In addition, we show an asymptotic rate of convergence whose efficiency estimate does not depend on the data of the problem. Finally, we study the convergence properties of a decomposition-based approach designed to solve a composite convex model.

Original language | English |
---|---|

Pages (from-to) | 185-209 |

Number of pages | 25 |

Journal | SIAM Journal on Optimization |

Volume | 25 |

Issue number | 1 |

DOIs | |

State | Published - 2015 |

Externally published | Yes |

## Keywords

- Alternating minimization
- Convex optimization
- Iteratively reweighted least squares
- Rate of convergence