Line Search Methods in Optimization#

So far, we’ve discussed gradient descent, but we haven’t covered how to choose the step size \(\eta_t\).

Line search methods are a family of optimization techniques used to determine how far to move along a given descent direction during iterative optimization like gradient descent.

While the descent direction tells where to go, line search methods answer:

“How far should I go in that direction?”

They are essential for balancing progress (fast convergence) and stability (avoiding overshooting or divergence).


Setup#

Suppose you’re minimizing a differentiable function \(f : \mathbb{R}^n \to \mathbb{R}\), and you’re at a point \(\mathbf{x}_t\) with a descent direction \(\mathbf{d}_t\) (e.g., \(-\nabla f(\mathbf{x}_t)\)).

The line search chooses a step size \(\eta_t > 0\) such that:

\[ \mathbf{x}_{t+1} = \mathbf{x}_t + \eta_t \mathbf{d}_t \]

The goal is to choose \(\eta_t\) so that it sufficiently decreases the function.


🧠 Types of Line Search Methods#

Line search methods can be broadly categorized into exact and inexact methods.