A global optimization algorithm, αBB, for twice-differentiable NLPs is presented. It operates within a branch-and-bound framework and requires the construction of a convex lower bound-ing problem. A technique to generate such a valid convex underestimator for arbitrary twice-differentiable functions is described. The αBB has been applied to a variety of problems and a summary of the results obtained is provided.
All Science Journal Classification (ASJC) codes
- Chemical Engineering(all)
- Computer Science Applications