paint-brush
Globalization Strategies: What Do They Allow Us to Do?โ€‚by@linearization

Globalization Strategies: What Do They Allow Us to Do?

by Linearization Technology4mMarch 26th, 2025
Read on Terminal Reader
Read this story w/o Javascript
tldt arrow

Too Long; Didn't Read

Globalization Strategies allow us to enhance the local algorithms described in Subsection 2.1 and ensure global convergence under suitable assumptions
featured image - Globalization Strategies: What Do They Allow Us to Do?
Linearization Technology HackerNoon profile picture
0-item

Abstract and 1. Introduction

2. Mathematical Description and 2.1. Numerical Algorithms for Nonlinear Equations

2.2. Globalization Strategies

2.3. Sensitivity Analysis

2.4. Matrix Coloring & Sparse Automatic Differentiation

3. Special Capabilities

3.1. Composable Building Blocks

3.2. Smart PolyAlgortihm Defaults

3.3. Non-Allocating Static Algorithms inside GPU Kernels

3.4. Automatic Sparsity Exploitation

3.5. Generalized Jacobian-Free Nonlinear Solvers using Krylov Methods

4. Results and 4.1. Robustness on 23 Test Problems

4.2. Initializing the Doyle-Fuller-Newman (DFN) Battery Model

4.3. Large Ill-Conditioned Nonlinear Brusselator System

5. Conclusion and References

2.2. Globalization Strategies.

Globalization Strategies allow us to enhance the local algorithms described in Subsection 2.1 and ensure global convergence under suitable assumptions [23, Theorem 3.2, 4.5, 4.6].


2.2.1. Line Search. Consider the ๐‘-Dimensional Generalized Rosenbrock Function, which is defined as:



If we initialize the problem with (๐‘ข0)1 = โˆ’1.2, Newton-Raphson fails to converge to the solution [Figure 1]. Line search methods are globalization strategies that avoid such failures by determining a positive step length ๐›ผ๐‘˜ given the current iterate ๐‘ข๐‘˜ and the search direction ๐›ฟ๐‘ขk



The direction is often given by the Newton Descent Direction, Steepest Descent, or one of the Multi-Step Schemes described in Subsection 2.1. The optimal choice for the step length is given by:



However, solving a global optimization problem on each step of the iterative nonlinear solver is prohibitively expensive. Instead, practical line search methods rely on selecting a set of candidate step sizes and terminating the search based on certain conditions:


Armijo Rule and Curvature Criteria: The Armijo Rule or Sufficient Decrease criteria states that ๐›ผ๐‘˜ is acceptable only if there is a sufficient decrease in the objective function:



Additionally, we use the curvature condition to filter out values for ๐›ผ๐‘˜ that satisfy sufficient decrease but are very close to the current iterate. This ensures that the algorithm makes reasonable progress at every step.



where ๐‘2 โˆˆ (๐‘1, 1) and ๐‘1 โˆˆ (0, 1). These two conditions are collectively known as the Wolfe Conditions [26, 27].


Strong Wolfe Condition: Strong Wolfe conditions additionally restrict the curvature criteria to restrict the derivative from being too positive:



Backtracking Line Search: In this method, we start with an initial step length (typically 1 for most root finding algorithms) and shrink the step length by a (adaptive) factor ๐œŒ โˆˆ (0, 1). In this case, we automatically ensure that the algorithm is making reasonable progress; hence, we only need to check for the sufficient decrease criteria [Equation (2.16)] [23, Section 3.1, Sufficient Decrease and Backtracking].


NonlinearSolve.jl supports line search routines for its solvers using the LineSearches.jl package [20]. In Figure 1, we demonstrate how NonlinearSolve.jl provides a uniform API to switch between different line search routines like BackTracking [23], HagerZhang [28], MoreThuente [29], etc. Using line search methods, we successfully solve the generalized Rosenbrock function [Equation (2.12)].


2.2.2. Trust Region Methods. As an alternative to using line search as a globalization strategy, trust-region methods restrict the next iterate to lie in a local


Fig. 2: Work-Precision Diagrams for Two Instances of the 23 Test Problems: Trust Region Radius Update Schemes directly affect the convergence of the algorithm, NonlinearSolve.jl implements various update schemes and allows switching them with a unified API.


region around the current iterate. These methods estimate a quadratic model for the objective function:



and attempt to solve the following sub-problem:



where ฮ”๐‘˜ is the trust region radius at the current iterate ๐‘˜. To perform an update to the current iterate and the trust region radius, we compute the ratio of the actual reduction to the predicted reduction (๐œŒ๐‘˜):



If ๐œŒ๐‘˜ is greater than a threshold, i.e., there is a strong agreement between the predicted model and true system, we accept the step and increase the trust-region radius ฮ”๐‘˜. If 0 < ๐œŒ๐‘˜ << 1, we accept the step but keep the trust-region radius unaffected. Otherwise, the trust region is shrunk, and the step is rejected. The exact specification for thresholds for these choices depends on the underlying algorithms being used. NonlinearSolve.jl implements additional sophisticated algorithms to update the trust region โ€“ Hei [30], Yuan [31], Fan [32], and Bastin [33]. In Figure 2 and Section 4, we demonstrate how different radius update schemes achieve a compromise between the computational cost and solution accuracy.



This paper is available on arxiv under CC BY 4.0 DEED license.

Authors:

(1) AVIK PAL, CSAIL MIT, Cambridge, MA;

(2) FLEMMING HOLTORF;

(3) AXEL LARSSON;

(4) TORKEL LOMAN;

(5) UTKARSH;

(6) FRANK SCHร„FER;

(7) QINGYU QU;

(8) ALAN EDELMAN;

(9) CHRIS RACKAUCKAS, CSAIL MIT, Cambridge, MA.