The mean value inequality and equalities #
In this file we prove the following facts:
- 
convex.norm_image_sub_le_of_norm_deriv_le: iffis differentiable on a convex setsand the norm of its derivative is bounded byC, thenfis Lipschitz continuous onswith constantC; also a variant in which what is bounded byCis the norm of the difference of the derivative from a fixed linear map. This lemma and its versions are formulated usingis_R_or_C, so they work both for real and complex derivatives.
- 
image_le_of*,image_norm_le_of_*: several similar lemmas deducingf x ≤ B xor∥f x∥ ≤ B xfrom upper estimates onf'or∥f'∥, respectively. These lemmas differ by their assumptions:- of_liminf_*lemmas assume that limit inferior of some ratio is less than- B' x;
- of_deriv_right_*,- of_norm_deriv_right_*lemmas assume that the right derivative or its norm is less than- B' x;
- of_*_lt_*lemmas assume a strict inequality whenever- f x = B xor- ∥f x∥ = B x;
- of_*_le_*lemmas assume a non-strict inequality everywhere on- [a, b);
- name of a lemma ends with 'if (1) it assumes thatBis continuous on[a, b]and has a right derivative at every point of[a, b), and (2) the lemma has a counterpart assuming thatBis differentiable everywhere onℝ
 
- 
norm_image_sub_le_*_segment: if derivative offon[a, b]is bounded above by a constantC, then∥f x - f a∥ ≤ C * ∥x - a∥; several versions deal with right derivative and derivative within[a, b](has_deriv_within_atorderiv_within).
- 
convex.is_const_of_fderiv_within_eq_zero: if a function has derivative0on a convex sets, then it is a constant ons.
- 
exists_ratio_has_deriv_at_eq_ratio_slopeandexists_ratio_deriv_eq_ratio_slope: Cauchy's Mean Value Theorem.
- 
exists_has_deriv_at_eq_slopeandexists_deriv_eq_slope: Lagrange's Mean Value Theorem.
- 
domain_mvt: Lagrange's Mean Value Theorem, applied to a segment in a convex domain.
- 
convex.image_sub_lt_mul_sub_of_deriv_lt,convex.mul_sub_lt_image_sub_of_lt_deriv,convex.image_sub_le_mul_sub_of_deriv_le,convex.mul_sub_le_image_sub_of_le_deriv, if∀ x, C (</≤/>/≥) (f' x), thenC * (y - x) (</≤/>/≥) (f y - f x)wheneverx < y.
- 
convex.monotone_on_of_deriv_nonneg,convex.antitone_on_of_deriv_nonpos,convex.strict_mono_of_deriv_pos,convex.strict_anti_of_deriv_neg: if the derivative of a function is non-negative/non-positive/positive/negative, then the function is monotone/antitone/strictly monotone/strictly monotonically decreasing.
- 
convex_on_of_deriv_monotone_on,convex_on_of_deriv2_nonneg: if the derivative of a function is increasing or its second derivative is nonnegative, then the original function is convex.
- 
strict_fderiv_of_cont_diff: a C^1 function over the reals is strictly differentiable. (This is a corollary of the mean value inequality.)
One-dimensional fencing inequalities #
General fencing theorem for continuous functions with an estimate on the derivative.
Let f and B be continuous functions on [a, b] such that
- f a ≤ B a;
- Bhas right derivative- B'at every point of- [a, b);
- for each x ∈ [a, b)the right-side limit inferior of(f z - f x) / (z - x)is bounded above by a functionf';
- we have f' x < B' xwheneverf x = B x.
Then f x ≤ B x everywhere on [a, b].
General fencing theorem for continuous functions with an estimate on the derivative.
Let f and B be continuous functions on [a, b] such that
- f a ≤ B a;
- Bhas derivative- B'everywhere on- ℝ;
- for each x ∈ [a, b)the right-side limit inferior of(f z - f x) / (z - x)is bounded above by a functionf';
- we have f' x < B' xwheneverf x = B x.
Then f x ≤ B x everywhere on [a, b].
General fencing theorem for continuous functions with an estimate on the derivative.
Let f and B be continuous functions on [a, b] such that
- f a ≤ B a;
- Bhas right derivative- B'at every point of- [a, b);
- for each x ∈ [a, b)the right-side limit inferior of(f z - f x) / (z - x)is bounded above byB'.
Then f x ≤ B x everywhere on [a, b].
General fencing theorem for continuous functions with an estimate on the derivative.
Let f and B be continuous functions on [a, b] such that
- f a ≤ B a;
- Bhas right derivative- B'at every point of- [a, b);
- fhas right derivative- f'at every point of- [a, b);
- we have f' x < B' xwheneverf x = B x.
Then f x ≤ B x everywhere on [a, b].
General fencing theorem for continuous functions with an estimate on the derivative.
Let f and B be continuous functions on [a, b] such that
- f a ≤ B a;
- Bhas derivative- B'everywhere on- ℝ;
- fhas right derivative- f'at every point of- [a, b);
- we have f' x < B' xwheneverf x = B x.
Then f x ≤ B x everywhere on [a, b].
General fencing theorem for continuous functions with an estimate on the derivative.
Let f and B be continuous functions on [a, b] such that
- f a ≤ B a;
- Bhas derivative- B'everywhere on- ℝ;
- fhas right derivative- f'at every point of- [a, b);
- we have f' x ≤ B' xon[a, b).
Then f x ≤ B x everywhere on [a, b].
Vector-valued functions f : ℝ → E #
        General fencing theorem for continuous functions with an estimate on the derivative.
Let f and B be continuous functions on [a, b] such that
- ∥f a∥ ≤ B a;
- Bhas right derivative at every point of- [a, b);
- for each x ∈ [a, b)the right-side limit inferior of(∥f z∥ - ∥f x∥) / (z - x)is bounded above by a functionf';
- we have f' x < B' xwhenever∥f x∥ = B x.
Then ∥f x∥ ≤ B x everywhere on [a, b].
General fencing theorem for continuous functions with an estimate on the norm of the derivative.
Let f and B be continuous functions on [a, b] such that
- ∥f a∥ ≤ B a;
- fand- Bhave right derivatives- f'and- B'respectively at every point of- [a, b);
- the norm of f'is strictly less thanB'whenever∥f x∥ = B x.
Then ∥f x∥ ≤ B x everywhere on [a, b]. We use one-sided derivatives in the assumptions
to make this theorem work for piecewise differentiable functions.
General fencing theorem for continuous functions with an estimate on the norm of the derivative.
Let f and B be continuous functions on [a, b] such that
- ∥f a∥ ≤ B a;
- fhas right derivative- f'at every point of- [a, b);
- Bhas derivative- B'everywhere on- ℝ;
- the norm of f'is strictly less thanB'whenever∥f x∥ = B x.
Then ∥f x∥ ≤ B x everywhere on [a, b]. We use one-sided derivatives in the assumptions
to make this theorem work for piecewise differentiable functions.
General fencing theorem for continuous functions with an estimate on the norm of the derivative.
Let f and B be continuous functions on [a, b] such that
- ∥f a∥ ≤ B a;
- fand- Bhave right derivatives- f'and- B'respectively at every point of- [a, b);
- we have ∥f' x∥ ≤ B xeverywhere on[a, b).
Then ∥f x∥ ≤ B x everywhere on [a, b]. We use one-sided derivatives in the assumptions
to make this theorem work for piecewise differentiable functions.
General fencing theorem for continuous functions with an estimate on the norm of the derivative.
Let f and B be continuous functions on [a, b] such that
- ∥f a∥ ≤ B a;
- fhas right derivative- f'at every point of- [a, b);
- Bhas derivative- B'everywhere on- ℝ;
- we have ∥f' x∥ ≤ B xeverywhere on[a, b).
Then ∥f x∥ ≤ B x everywhere on [a, b]. We use one-sided derivatives in the assumptions
to make this theorem work for piecewise differentiable functions.
A function on [a, b] with the norm of the right derivative bounded by C
satisfies ∥f x - f a∥ ≤ C * (x - a).
A function on [a, b] with the norm of the derivative within [a, b]
bounded by C satisfies ∥f x - f a∥ ≤ C * (x - a), has_deriv_within_at
version.
A function on [a, b] with the norm of the derivative within [a, b]
bounded by C satisfies ∥f x - f a∥ ≤ C * (x - a), deriv_within
version.
A function on [0, 1] with the norm of the derivative within [0, 1]
bounded by C satisfies ∥f 1 - f 0∥ ≤ C, has_deriv_within_at
version.
A function on [0, 1] with the norm of the derivative within [0, 1]
bounded by C satisfies ∥f 1 - f 0∥ ≤ C, deriv_within version.
If two continuous functions on [a, b] have the same right derivative and are equal at a,
then they are equal everywhere on [a, b].
If two differentiable functions on [a, b] have the same derivative within [a, b] everywhere
on [a, b) and are equal at a, then they are equal everywhere on [a, b].
Vector-valued functions f : E → G #
Theorems in this section work both for real and complex differentiable functions. We use assumptions
[is_R_or_C 𝕜] [normed_space 𝕜 E] [normed_space 𝕜 G] to achieve this result. For the domain E we
also assume [normed_space ℝ E] to have a notion of a convex set.
The mean value theorem on a convex set: if the derivative of a function is bounded by C, then
the function is C-Lipschitz. Version with has_fderiv_within.
The mean value theorem on a convex set: if the derivative of a function is bounded by C on
s, then the function is C-Lipschitz on s. Version with has_fderiv_within and
lipschitz_on_with.
Let s be a convex set in a real normed vector space E, let f : E → G be a function
differentiable within s in a neighborhood of x : E with derivative f'. Suppose that f' is
continuous within s at x. Then for any number K : ℝ≥0 larger than ∥f' x∥₊, f is
K-Lipschitz on some neighborhood of x within s. See also
convex.exists_nhds_within_lipschitz_on_with_of_has_fderiv_within_at for a version that claims
existence of K instead of an explicit estimate.
Let s be a convex set in a real normed vector space E, let f : E → G be a function
differentiable within s in a neighborhood of x : E with derivative f'. Suppose that f' is
continuous within s at x. Then for any number K : ℝ≥0 larger than ∥f' x∥₊, f is Lipschitz
on some neighborhood of x within s. See also
convex.exists_nhds_within_lipschitz_on_with_of_has_fderiv_within_at_of_nnnorm_lt for a version
with an explicit estimate on the Lipschitz constant.
The mean value theorem on a convex set: if the derivative of a function within this set is
bounded by C, then the function is C-Lipschitz. Version with fderiv_within.
The mean value theorem on a convex set: if the derivative of a function is bounded by C on
s, then the function is C-Lipschitz on s. Version with fderiv_within and
lipschitz_on_with.
The mean value theorem on a convex set: if the derivative of a function is bounded by C,
then the function is C-Lipschitz. Version with fderiv.
The mean value theorem on a convex set: if the derivative of a function is bounded by C on
s, then the function is C-Lipschitz on s. Version with fderiv and lipschitz_on_with.
Variant of the mean value inequality on a convex set, using a bound on the difference between
the derivative and a fixed linear map, rather than a bound on the derivative itself. Version with
has_fderiv_within.
Variant of the mean value inequality on a convex set. Version with fderiv_within.
Variant of the mean value inequality on a convex set. Version with fderiv.
If a function has zero Fréchet derivative at every point of a convex set, then it is a constant on this set.
The mean value theorem on a convex set in dimension 1: if the derivative of a function is
bounded by C, then the function is C-Lipschitz. Version with has_deriv_within.
The mean value theorem on a convex set in dimension 1: if the derivative of a function is
bounded by C on s, then the function is C-Lipschitz on s.
Version with has_deriv_within and lipschitz_on_with.
The mean value theorem on a convex set in dimension 1: if the derivative of a function within
this set is bounded by C, then the function is C-Lipschitz. Version with deriv_within
The mean value theorem on a convex set in dimension 1: if the derivative of a function is
bounded by C on s, then the function is C-Lipschitz on s.
Version with deriv_within and lipschitz_on_with.
The mean value theorem on a convex set in dimension 1: if the derivative of a function is
bounded by C, then the function is C-Lipschitz. Version with deriv.
The mean value theorem on a convex set in dimension 1: if the derivative of a function is
bounded by C on s, then the function is C-Lipschitz on s.
Version with deriv and lipschitz_on_with.
The mean value theorem set in dimension 1: if the derivative of a function is bounded by C,
then the function is C-Lipschitz.  Version with deriv and lipschitz_with.
If f : 𝕜 → G, 𝕜 = R or 𝕜 = ℂ, is differentiable everywhere and its derivative equal zero,
then it is a constant function.
Functions [a, b] → ℝ. #
        Cauchy's Mean Value Theorem, has_deriv_at version.
Cauchy's Mean Value Theorem, extended has_deriv_at version.
Lagrange's Mean Value Theorem, has_deriv_at version
Cauchy's Mean Value Theorem, deriv version.
Cauchy's Mean Value Theorem, extended deriv version.
Let f be a function continuous on a convex (or, equivalently, connected) subset D
of the real line. If f is differentiable on the interior of D and C < f', then
f grows faster than C * x on D, i.e., C * (y - x) < f y - f x whenever x, y ∈ D,
x < y.
Let f be a function continuous on a convex (or, equivalently, connected) subset D
of the real line. If f is differentiable on the interior of D and C ≤ f', then
f grows at least as fast as C * x on D, i.e., C * (y - x) ≤ f y - f x whenever x, y ∈ D,
x ≤ y.
Let f be a function continuous on a convex (or, equivalently, connected) subset D
of the real line. If f is differentiable on the interior of D and f' < C, then
f grows slower than C * x on D, i.e., f y - f x < C * (y - x) whenever x, y ∈ D,
x < y.
Let f be a function continuous on a convex (or, equivalently, connected) subset D
of the real line. If f is differentiable on the interior of D and f' ≤ C, then
f grows at most as fast as C * x on D, i.e., f y - f x ≤ C * (y - x) whenever x, y ∈ D,
x ≤ y.
Let f be a function continuous on a convex (or, equivalently, connected) subset D
of the real line. If f is differentiable on the interior of D and f' is positive, then
f is a strictly monotone function on D.
Note that we don't require differentiability explicitly as it already implied by the derivative
being strictly positive.
Let f : ℝ → ℝ be a differentiable function. If f' is positive, then
f is a strictly monotone function.
Note that we don't require differentiability explicitly as it already implied by the derivative
being strictly positive.
Let f be a function continuous on a convex (or, equivalently, connected) subset D
of the real line. If f is differentiable on the interior of D and f' is nonnegative, then
f is a monotone function on D.
Let f : ℝ → ℝ be a differentiable function. If f' is nonnegative, then
f is a monotone function.
Let f be a function continuous on a convex (or, equivalently, connected) subset D
of the real line. If f is differentiable on the interior of D and f' is negative, then
f is a strictly antitone function on D.
Let f : ℝ → ℝ be a differentiable function. If f' is negative, then
f is a strictly antitone function.
Note that we don't require differentiability explicitly as it already implied by the derivative
being strictly negative.
Let f be a function continuous on a convex (or, equivalently, connected) subset D
of the real line. If f is differentiable on the interior of D and f' is nonpositive, then
f is an antitone function on D.
Let f : ℝ → ℝ be a differentiable function. If f' is nonpositive, then
f is an antitone function.
If a function f is continuous on a convex set D ⊆ ℝ, is differentiable on its interior,
and f' is monotone on the interior, then f is convex on D.
If a function f is continuous on a convex set D ⊆ ℝ, is differentiable on its interior,
and f' is antitone on the interior, then f is concave on D.
If a function f is continuous on a convex set D ⊆ ℝ, and f' is strictly monotone on the
interior, then f is strictly convex on D.
Note that we don't require differentiability, since it is guaranteed at all but at most
one point by the strict monotonicity of f'.
If a function f is continuous on a convex set D ⊆ ℝ and f' is strictly antitone on the
interior, then f is strictly concave on D.
Note that we don't require differentiability, since it is guaranteed at all but at most
one point by the strict antitonicity of f'.
If a function f is differentiable and f' is antitone on ℝ then f is concave.
If a function f is continuous and f' is strictly monotone on ℝ then f is strictly
convex. Note that we don't require differentiability, since it is guaranteed at all but at most
one point by the strict monotonicity of f'.
If a function f is continuous and f' is strictly antitone on ℝ then f is strictly
concave. Note that we don't require differentiability, since it is guaranteed at all but at most
one point by the strict antitonicity of f'.
If a function f is continuous on a convex set D ⊆ ℝ, is twice differentiable on its
interior, and f'' is nonnegative on the interior, then f is convex on D.
If a function f is continuous on a convex set D ⊆ ℝ, is twice differentiable on its
interior, and f'' is nonpositive on the interior, then f is concave on D.
If a function f is continuous on a convex set D ⊆ ℝ and f'' is strictly positive on the
interior, then f is strictly convex on D.
Note that we don't require twice differentiability explicitly as it is already implied by the second
derivative being strictly positive, except at at most one point.
If a function f is continuous on a convex set D ⊆ ℝ and f'' is strictly negative on the
interior, then f is strictly concave on D.
Note that we don't require twice differentiability explicitly as it already implied by the second
derivative being strictly negative, except at at most one point.
If a function f is twice differentiable on a open convex set D ⊆ ℝ and
f'' is nonnegative on D, then f is convex on D.
If a function f is twice differentiable on an open convex set D ⊆ ℝ and
f'' is nonpositive on D, then f is concave on D.
If a function f is continuous on a convex set D ⊆ ℝ and f'' is strictly positive on D,
then f is strictly convex on D.
Note that we don't require twice differentiability explicitly as it is already implied by the second
derivative being strictly positive, except at at most one point.
If a function f is continuous on a convex set D ⊆ ℝ and f'' is strictly negative on D,
then f is strictly concave on D.
Note that we don't require twice differentiability explicitly as it is already implied by the second
derivative being strictly negative, except at at most one point.
If a function f is twice differentiable on ℝ, and f'' is nonpositive on ℝ,
then f is concave on ℝ.
If a function f is continuous on ℝ, and f'' is strictly positive on ℝ,
then f is strictly convex on ℝ.
Note that we don't require twice differentiability explicitly as it is already implied by the second
derivative being strictly positive, except at at most one point.
If a function f is continuous on ℝ, and f'' is strictly negative on ℝ,
then f is strictly concave on ℝ.
Note that we don't require twice differentiability explicitly as it is already implied by the second
derivative being strictly negative, except at at most one point.
Functions f : E → ℝ #
        Lagrange's Mean Value Theorem, applied to convex domains.
Vector-valued functions f : E → F.  Strict differentiability. #
A C^1 function is strictly differentiable, when the field is ℝ or ℂ. This follows from the
mean value inequality on balls, which is a particular case of the above results after restricting
the scalars to ℝ. Note that it does not make sense to talk of a convex set over ℂ, but balls
make sense and are enough. Many formulations of the mean value inequality could be generalized to
balls over ℝ or ℂ. For now, we only include the ones that we need.
Over the reals or the complexes, a continuously differentiable function is strictly differentiable.
Over the reals or the complexes, a continuously differentiable function is strictly differentiable.