Differential (infinitesimal)

From Wikipedia, the free encyclopedia

In differential calculus, a differential is traditionally an infinitesimally small change in a variable. For example, if x is a variable, then a change in the value of x is often denoted Δx (or δx when this change is considered to be small). The differential dx represents such a change, but is infinitely small. Although this is not a rigorous mathematical concept, it is extremely useful intuitively, and there are a number of ways to make the notion mathematically precise.

The key property of the differential is that if y is a function of x, then the differential dy of y is related to dx by the formula

\mathrm d y = \frac{\mathrm d y}{\mathrm d x} \mathrm d x,

where dy/dx denotes the derivative of y with respect to x. This formula summarizes the intuitive idea that the derivative of y with respect to x is the limit of the ratio Δyx of differences as Δx becomes infinitesimally small.

There are several approaches for making the notion of differentials mathematically precise.

  1. Differentials as linear maps. This approach underlies the definition of the derivative and the exterior derivative in differential geometry.
  2. Differentials as nilpotent elements of commutative rings. This approach is popular in algebraic geometry.
  3. Differentials in smooth models of set theory. This approach is known as synthetic differential geometry or smooth infinitesimal analysis and is closely related to the algebraic geometric approach, except that ideas from topos theory are used to hide the mechanisms by which nilpotent infinitesimals are introduced.
  4. Differentials as infinitesimals in hyperreal number systems, which are extensions of the real numbers which contain invertible infinitesimals and infinitely large numbers. This is the approach of nonstandard analysis pioneered by Abraham Robinson.

These approaches are very different from each other, but they have in common the idea to be quantitative, i.e., to say not just that a differential is infinitesimally small, but how small it is.

Contents

[edit] History and usage

Topics in calculus

Fundamental theorem
Limits of functions
Continuity
Vector calculus
Matrix calculus
Mean value theorem

Differentiation

Product rule
Quotient rule
Chain rule
Implicit differentiation
Taylor's theorem
Related rates
List of differentiation identities

Integration

Lists of integrals
Improper integrals
Integration by:
parts, disks, cylindrical
shells
, substitution,
trigonometric substitution,
partial fractions, changing order

Infinitesimal quantities played a significant role in the development of calculus. Archimedes used them, even though he didn't believe that arguments involving infinitesimals were rigorous. Newton referred to them as fluxions. However, it was Leibniz who coined the term differentials for infinitesimal quantities, and introduced the notation for them which is still used today.

In Leibniz's notation, if x is a variable quantity, then dx denotes an infinitesimally small change in the variable x. Thus, if y is a function of x, then the derivative of y with respect to x is often denoted \frac{\mathrm dy}{\mathrm dx}, which would otherwise be denoted (in the notation of Newton or Lagrange) {\dot y}(x) or y'(x). The use of differentials in this form attracted much criticism, for instance in the famous pamphlet The Analyst by Bishop Berkeley. Nevertheless the notation has remained popular because it suggests strongly the idea that the derivative of a function y(x) is its slope, which may be obtained by taking the limit of the ratio \frac{\Delta\,y}{\Delta\,x} of the change in y over the change in x, as the change in x becomes arbitrarily small. Differentials are also compatible with dimensional analysis, where a differential such as dx has the same dimensions as the variable x.

Differentials are also used in the notation for integrals because an integral can be regarded as an infinite sum of infinitesimally small quantities: the area under a graph is obtained by subdividing the graph into infinitesimally thin strips and summing their areas. In an expression such as

\int f(x) \, {\mathrm d}x,

the integral sign (which is a modified S) denotes the infinite sum, whereas the differential dx denotes the infinitesimally thin strips.

For more information, see the history of calculus.

[edit] Differentials as linear maps

There is a simple way to make precise sense of differentials by regarding them as linear maps. One way to explain this point of view is to regard the variable x in an expression such as f(x) as a function on the real line, the standard coordinate or identity map which takes a real number p to itself (x(p) = p): then f(x) denotes the composite f\circ x of f with x, whose value at p is f(x(p)). The differential df is then a function on the real line whose value at p (usually denoted dfp) is not a number, but a linear map from \mathbb{R} to \mathbb{R}. Since a linear map from \mathbb{R} to \mathbb{R} is given by a 1×1 matrix, it is essentially the same thing as a number, but the change in the point of view allows us to think of dfp as an infinitesimal and compare it with the standard infinitesimal dxp which is again just the identity map from \mathbb{R} to \mathbb{R} (a 1×1 matrix with entry 1). It may seem fanciful to regard the identity map as an infinitesimal, but it does at least have the property that if ε is very small, then dxp(ε) is very small. The differential dfp has the same property, because it is just a multiple of dxp, and this multiple is the derivative f'(p)\, by definition. We therefore obtain that \mathrm{d}f_p = f'(p)\, \mathrm{d}x_p, and hence \mathrm{d}f = f'\, \mathrm{d}x. Thus we recover the idea that f'\, is the ratio of the differentials df and dx.

This would just be a trick were it not for the fact that:

  1. it captures the idea of the derivative of f at p as the best linear approximation to f at p;
  2. it has many generalizations.

For instance if f is a function from \mathbb{R}^n to \mathbb{R} then we say f is differentiable at p \in \mathbb{R}^n if there is a linear map dfp from \mathbb{R}^n to \mathbb{R} such that for any ε > 0, there is a neighbourhood N(p) of p such that for x \in N(p):

\left|f(x)-f(p)-\mathrm d f_p(x-p)\right|<\varepsilon\left|x-p\right|.

(See, for example, Apostol's book[1].)

We can now use the same trick as in the one dimensional case, and think of the expression f(x^1, x^2, \dots, x^n) as the composite of f with the standard coordinates (x^1, x^2, \dots, x^n) on \mathbb{R}^n (so that xj(p) is the j-th component of p \in \mathbb{R}^n). Then the differentials (dx1)p, (dx2)p, (dxn)p (at a point p) form a basis for the vector space of linear maps from \mathbb{R}^n to \mathbb{R} and therefore, if f is differentiable at p, we can write dfp as a linear combination of these basis elements:

 {\mathrm d}f_p = \sum_{j=1}^n D_j f(p)\, ({\mathrm d}x^{j})_p.

The coefficients Djf(p) are (by definition) the partial derivatives of f at p with respect to x1, x2, …, xn. Hence, if f is differentiable on all of \mathbb{R}^n, we can write, more concisely:

 \mathrm df = \frac{\partial f}{\partial x^1}\, \mathrm dx^1 + \frac{\partial f}{\partial x^2}\, \mathrm dx^2 + \cdots +\frac{\partial f}{\partial x^n}\, \mathrm dx^n.

In the one-dimensional case this becomes

\mathrm df = \frac{\mathrm df}{\mathrm dx}\mathrm dx

as before.

This idea generalizes straightforwardly to functions from \mathbb{R}^n to \mathbb{R}^m. Furthermore, it has the decisive advantage over other definitions of the derivative that it is invariant under changes of coordinates. This means that the same idea can be used to define the differential of smooth maps between smooth manifolds.

Aside: Note that the existence of all the partial derivatives of f(x) at x is a necessary condition for the existence of a differential at x. However it is not a sufficient condition. For counterexamples, see Gateaux derivative.

[edit] The algebraic geometry approach

In algebraic geometry, differentials and other infinitesimal notions are handled in a very explicit way by accepting that the coordinate ring or sheaf of a space may contain nilpotent elements. The simplest example is the ring of dual numbers R[ε], where ε2 = 0.

This can be motivated by the algebro-geometric point of view on the derivative of a function f from R to R at a point p. For this, note first that ff(p)1 (where 1 is the identity function) belongs to the ideal Ip of functions on R which vanish at p. If the derivative f vanishes at p, then ff(p)1 belongs to the square Ip2 of this ideal. Hence the derivative of f at p may be captured by the equivalence class [ff(p)1] in the quotient space Ip/Ip2, and the 1-jet of f (which encodes its value and its first derivative) is the equivalence class of f in the space of all functions modulo Ip2. Algebraic geometers regard this equivalence class as the restriction of f to a thickened version of the point p whose coordinate ring is not R (which is the quotient space of functions on R modulo Ip) but R[ε] which is the quotient space of functions on R modulo Ip2. Such a thickened point is a simple example of a scheme.

[edit] Synthetic differential geometry

A third approach to infinitesimals is the method of synthetic differential geometry or smooth infinitesimal analysis. This is closely related to the algebraic geometric approach, except that the infinitesimals are more implicit and intuitive. The main idea of this approach is to replace the category of sets with another category of smoothly varying sets which is a topos. In this category, one can define the real numbers, smooth functions, and so on, but the real numbers automatically contain nilpotent infinitesimals, so these do not need to be introduced by hand as in the algebraic geometric approach. However the logic in this new category is not identical to the familiar logic of the category of sets: in particular, the law of the excluded middle does not hold. This means that set-theoretic mathematical arguments only extend to smooth infinitesimal analysis if they are constructive (e.g., do not use proof by contradiction). Some regard this disadvantage as a positive thing, since it forces one to find constructive arguments wherever they are available.

[edit] Non-standard analysis

The final approach to infinitesimals again involves extending the real numbers, but in a much less drastic way. In this approach there are no nilpotent infinitesimals, only invertible ones, which may be viewed as the reciprocals of infinitely large numbers. Such extensions of the real numbers may be constructed explicitly using equivalence classes of sequences of rational numbers, so that, for example, the sequence (1,1/2,1/3,...1/n,...) represents an infinitesimal. The first-order logic of this new set of hyperreal numbers is the same as the logic for the usual real numbers, but the completeness axiom (which involves second-order logic) does not hold. Nevertheless, this suffices to develop an elementary and quite intuitive approach to calculus using infinitesimals.

[edit] References

  1. ^ Tom M Apostol (1967). Calculus, 2nd Ed. Wiley. ISBN 0-471-00005-1 and ISBN 0-471-00007-8.

[edit] See also