Slope Intervals, Generalized Gradients, Semigradients, Slant Derivatives, and Csets Many practical optimization problems are nonsmooth, and derivative-type methods cannot be applied. To overcome this difficulty, there are different concepts to replace the derivative of a function f : $$\mathbb{R}^n \to \mathbb{R}$$ : interval slopes, semigradients, generalized gradients, and slant derivatives are some examples. These approaches generalize the success of convex analysis, and are effective in optimization. However, with the exception of interval slopes, it is not clear how to automatically compute these; having a general analogue to the chain rule, interval slopes can be computed with automatic differentiation techniques. In this paper we study the relationships among these approaches for nonsmooth Lipschitz optimization problems in finite dimensional Euclidean spaces. Inclusion theorems concerning the equivalence of these concepts when there exist one sided derivatives in one dimension and in multidimensional cases are proved separately. Valid enclosures are produced. Under containment set (cset) theory, for instance, the cset of the gradient of a locally Lipschitz function f near x is included in its generalized gradient. http://www.deepdyve.com/assets/images/DeepDyve-Logo-lg.png Reliable Computing Springer Journals

, Volume 10 (3) – Oct 18, 2004
31 pages

Publisher
Subject
Mathematics; Numeric Computing; Approximations and Expansions; Computational Mathematics and Numerical Analysis; Mathematical Modeling and Industrial Mathematics
ISSN
1385-3139
eISSN
1573-1340
D.O.I.
10.1023/B:REOM.0000032107.85627.45
Publisher site
See Article on Publisher Site

Abstract

Many practical optimization problems are nonsmooth, and derivative-type methods cannot be applied. To overcome this difficulty, there are different concepts to replace the derivative of a function f : $$\mathbb{R}^n \to \mathbb{R}$$ : interval slopes, semigradients, generalized gradients, and slant derivatives are some examples. These approaches generalize the success of convex analysis, and are effective in optimization. However, with the exception of interval slopes, it is not clear how to automatically compute these; having a general analogue to the chain rule, interval slopes can be computed with automatic differentiation techniques. In this paper we study the relationships among these approaches for nonsmooth Lipschitz optimization problems in finite dimensional Euclidean spaces. Inclusion theorems concerning the equivalence of these concepts when there exist one sided derivatives in one dimension and in multidimensional cases are proved separately. Valid enclosures are produced. Under containment set (cset) theory, for instance, the cset of the gradient of a locally Lipschitz function f near x is included in its generalized gradient.

Journal

Reliable ComputingSpringer Journals

Published: Oct 18, 2004

DeepDyve is your personal research library

It’s your single place to instantly
that matters to you.

over 18 million articles from more than
15,000 peer-reviewed journals.

All for just $49/month Explore the DeepDyve Library Search Query the DeepDyve database, plus search all of PubMed and Google Scholar seamlessly Organize Save any article or search result from DeepDyve, PubMed, and Google Scholar... all in one place. Access Get unlimited, online access to over 18 million full-text articles from more than 15,000 scientific journals. Your journals are on DeepDyve Read from thousands of the leading scholarly journals from SpringerNature, Elsevier, Wiley-Blackwell, Oxford University Press and more. All the latest content is available, no embargo periods. DeepDyve Freelancer DeepDyve Pro Price FREE$49/month
\$360/year

Save searches from
PubMed

Create lists to

Export lists, citations