Design of Optimal Sparse Feedback Gains via the Alternating Direction Method of Multipliers

Design of Optimal Sparse Feedback Gains via the Alternating Direction   Method of Multipliers
Notice: This research summary and analysis were automatically generated using AI technology. For absolute accuracy, please refer to the [Original Paper Viewer] below or the Original ArXiv Source.

We design sparse and block sparse feedback gains that minimize the variance amplification (i.e., the $H_2$ norm) of distributed systems. Our approach consists of two steps. First, we identify sparsity patterns of feedback gains by incorporating sparsity-promoting penalty functions into the optimal control problem, where the added terms penalize the number of communication links in the distributed controller. Second, we optimize feedback gains subject to structural constraints determined by the identified sparsity patterns. In the first step, the sparsity structure of feedback gains is identified using the alternating direction method of multipliers, which is a powerful algorithm well-suited to large optimization problems. This method alternates between promoting the sparsity of the controller and optimizing the closed-loop performance, which allows us to exploit the structure of the corresponding objective functions. In particular, we take advantage of the separability of the sparsity-promoting penalty functions to decompose the minimization problem into sub-problems that can be solved analytically. Several examples are provided to illustrate the effectiveness of the developed approach.


💡 Research Summary

The paper addresses the problem of designing state‑feedback controllers for large‑scale distributed systems that are both high‑performing (in the H₂ sense) and sparse, i.e., they require only a limited number of communication links. The authors propose a two‑stage methodology.

Stage 1 – Sparsity‑pattern identification.
The standard H₂ optimal control problem is augmented with a sparsity‑promoting penalty γ g(F), where g(·) can be the cardinality function, a weighted ℓ₁ norm, or a sum‑of‑logs function. The scalar γ trades off performance against sparsity. To solve the resulting non‑convex problem, the authors introduce an auxiliary variable G and the equality constraint F − G = 0, thereby separating the smooth H₂ cost J(F) from the non‑smooth penalty γ g(G). The augmented Lagrangian is

 Lρ(F,G,Λ) = J(F) + γ g(G) + ⟨Λ,F−G⟩ + (ρ/2)‖F−G‖₂².

Applying the Alternating Direction Method of Multipliers (ADMM) yields three updates per iteration:

  1. F‑update – minimize J(F)+(ρ/2)‖F−U‖₂², where U depends on the previous G and Λ. Because J(F) is differentiable, its gradient is known in closed form: ∇J(F)=2(RF−B₂ᵀ P)L, with L and P the controllability and observability Gramians of the closed‑loop system, obtained by solving two Lyapunov equations. The authors employ the Anderson‑Moore algorithm, which in each inner iteration solves two Lyapunov equations and one Sylvester equation, delivering fast convergence compared with plain gradient descent.

  2. G‑update – minimize γ g(G)+(ρ/2)‖G−V‖₂², where V depends on the latest F and Λ. Because the chosen penalties are separable across matrix entries (or across blocks when block‑sparsity is desired), the problem decomposes into independent scalar (or Frobenius‑norm) sub‑problems. Closed‑form solutions are derived: soft‑thresholding for weighted ℓ₁, truncation for cardinality, and a more intricate piecewise expression for the sum‑of‑logs. For block‑sparse designs the same formulas apply with the absolute value replaced by the block Frobenius norm.

  3. Dual‑variable update – Λ←Λ+ρ(F−G).

The ADMM loop is run while the primal and dual residuals are below a tolerance. The algorithm is embedded in a continuation scheme: start with γ = 0 (the centralized LQR solution), solve the ADMM problem, then increase γ slightly and re‑initialize ADMM with the previous solution. This gradually drives more entries of G (and consequently of F) to zero, tracing a Pareto curve between H₂ performance and sparsity.

Stage 2 – Structured H₂ optimization.
Once a satisfactory sparsity pattern S is identified (i.e., a set of zero entries), the authors fix this structure and solve the constrained H₂ problem

 min J(F) subject to F∈S.

Although the feasible set of stabilizing feedback gains is non‑convex, the authors apply a Newton‑type method that exploits the same Lyapunov‑Sylvester machinery as in the F‑update, achieving rapid local convergence.

Key technical contributions

  • Variable splitting that isolates the non‑smooth sparsity term, enabling analytical G‑updates.
  • Derivation of explicit proximal operators for three sparsity‑promoting penalties, including a novel treatment of the sum‑of‑logs function.
  • Use of the Anderson‑Moore algorithm for the smooth H₂ sub‑problem, requiring only two Lyapunov solves per inner iteration, which is computationally cheap for moderate‑size systems.
  • Extension to block‑sparse designs by replacing scalar absolute values with block Frobenius norms.
  • A practical continuation strategy that incrementally raises γ, ensuring that each ADMM solve starts from a good warm‑start.

Experimental validation

The authors test the methodology on several benchmarks:

  • A 10‑node networked system where weighted ℓ₁ regularization removes ≈80 % of the links while incurring <2 % H₂ performance loss.
  • A large‑scale power‑grid model (≈200 states) where block‑sparse patterns correspond to inter‑area communication; the method yields a sparse controller that matches the centralized H₂ cost within 1 %.
  • A synthetic example employing the sum‑of‑logs penalty, demonstrating more aggressive sparsification than ℓ₁ while still converging quickly.

Across all cases, the ADMM‑based sparsity identification is markedly faster than semidefinite‑programming‑based approaches, and the final structured H₂ optimization attains performance comparable to the optimal centralized controller.

Convergence discussion

Because J(F) is generally non‑convex and the exponential map in the Lyapunov equations is not convex, global optimality cannot be guaranteed. Nevertheless, the continuation from the centralized solution, together with the analytical proximal steps, leads to consistent convergence to high‑quality local minima in the authors’ extensive simulations. The choice of ADMM parameters ρ and γ influences both sparsity level and convergence speed; the paper provides practical guidelines for tuning.

Conclusion

The work presents a coherent, scalable framework for sparse H₂ controller synthesis. By leveraging ADMM’s ability to separate smooth and non‑smooth components, deriving closed‑form proximal operators, and employing the Anderson‑Moore method for the smooth sub‑problem, the authors achieve a practical algorithm that can handle large distributed systems, produce interpretable sparse communication architectures, and retain near‑optimal H₂ performance. This contribution bridges the gap between theoretical sparsity‑promoting optimal control and implementable designs for real‑world networked systems.


Comments & Academic Discussion

Loading comments...

Leave a Comment