Constrained gradient? (optimization theory)

  • Context: Graduate 
  • Thread starter Thread starter brydustin
  • Start date Start date
  • Tags Tags
    Gradient Theory
Click For Summary

Discussion Overview

The discussion revolves around the application of the Karush-Kuhn-Tucker (KKT) conditions in constrained optimization, particularly the implications of including additional constraints on the gradient of the objective function within the Lagrangian. Participants explore whether imposing a requirement that the gradient of the objective function is zero can lead to better solutions while considering existing equality and inequality constraints.

Discussion Character

  • Debate/contested
  • Technical explanation
  • Mathematical reasoning

Main Points Raised

  • One participant questions the practicality of including a constraint that the gradient of the objective function is zero in the Lagrangian, suggesting it may help find better solutions under certain conditions.
  • Another participant argues that if an extremal point is not on the boundary, the KKT conditions require the gradient to be zero, but imposing this condition on boundary points may exclude valid solutions.
  • A different participant clarifies that the KKT requirement states the gradient of the Lagrangian must be zero, not necessarily the gradient of the objective function, prompting a reconsideration of the initial proposal.
  • One participant emphasizes that in an unconstrained region, certain multipliers are zero, which may affect the validity of the proposed additional constraint on the gradient.
  • Another participant reiterates that the discussion is about constrained optimization and questions the feasibility of satisfying all constraints simultaneously if the gradient condition is added.

Areas of Agreement / Disagreement

Participants express differing views on the implications of adding a gradient constraint to the Lagrangian. There is no consensus on whether this approach is beneficial or practical, and the discussion remains unresolved regarding the potential effects on finding solutions.

Contextual Notes

Participants acknowledge the complexity introduced by additional constraints and the potential for conflicting requirements among existing equality and inequality constraints. The implications of KKT multipliers and their geometric interpretations are also noted as relevant considerations.

brydustin
Messages
201
Reaction score
0
The first order KKT condition for constrained optimization requires that the gradient for the Lagrangian is equal to zero, however this does not necessarily imply that the objective function's gradient is equal to zero. Is it absurd to include in one's Lagrangian the requirement that the entry wise components of the objective gradient are themselves zero? I realize that if one already has other constraints then this requirement may become infeasible if in fact both sets of constraints are mutually disjoint... however, is it a practical method for finding "better" solutions (ones that satisfy the KKT for constrained optimization as well as satisfying the unconstrained optimization)... My idea is to direct a constraint to an area where, had one started close to such a solution, any constraint would have effectively been unnecessary.

In one dimension the idea is simple (I use ± because different texts write the Lagrangian differently):

f(x) = x^2, f ' (x) = 2x. Define the Lagrangian: L = f ± λ (f ' (x)) : such that f ' (x) =0.
Then x must equal zero. Therefore, L = 0 ± λ*0.

The purpose is for (multidimensional, nonlinear) numerical optimization, so if I do this, then when I compute the first derivative I would effectively have to have the second derivative (hessian) and when computing the second derivative I would have to compute the sum of the parts of the third derivative. Seems messy
 
Physics news on Phys.org
If there is an extremal point which is not contained on the boundary, the KKT constraints ARE that the gradient must be zero. If the extremal point is on the boundary, requiring that the gradient must be zero is going to cause you to not find that point. So all you're doing is throwing away possible solutions on the boundary of the feasible set
 
The KKT requirement is that:

∇ f(x) + \sum_{i=1}^m μ_i *∇ g_i(x) + \sum_{j=1}^l λ_j ∇ h_j(x^*) = 0,

NOT that ∇ f(x) = 0.

Merely that the gradient of the lagrangian is zero, NOT the gradient of the objective function.
Does that change your opinion?
 
Assuming you are using the notation of wikipedia
http://en.wikipedia.org/wiki/Karush–Kuhn–Tucker_conditions

If you're in an unconstrained region, it means you have no hjs since those are equality constraints that must be satisfied. Furthermore, all the μ_i s are zero because μ_i gi(x) = 0 at the critical point, and if the g constraint isn't satisfed gi must be strictly smaller than zero. So my opinion is unchanged and I suggest you really think over the geometric implications of the KKT multipliers some more
 
However, this is not unconstrained optimization. I already have equality and inequality constraints and I'm asking if adding the constraint for the gradient is a good idea? It seems like you are saying that it may be impossible to satisfy all constraints simultaneously (fyi, I already know about why the multipliers must have a certain sign or be zero).
 

Similar threads

  • · Replies 18 ·
Replies
18
Views
3K
  • · Replies 2 ·
Replies
2
Views
1K
  • · Replies 8 ·
Replies
8
Views
2K
Replies
4
Views
2K
  • · Replies 10 ·
Replies
10
Views
7K
  • · Replies 6 ·
Replies
6
Views
2K
  • · Replies 2 ·
Replies
2
Views
2K
  • · Replies 0 ·
Replies
0
Views
2K
  • · Replies 7 ·
Replies
7
Views
2K
  • · Replies 3 ·
Replies
3
Views
2K