224
Step function In mathematics , a function on the real numbers is called a step function (or staircase function) if it can be written as a finite linear combination of indicator functions of intervals . Informally speaking, a step function is a piecewise constant function having only finitely many pieces. Example of a step function (the red graph). This particular step function is right-continuous . [edit ] Definition and first consequences A function is called a step function if it can be written as for all real numbers x where α i are real numbers, A i are intervals, and is the indicator function of A:

Elementary Functions

  • View
    184

  • Download
    3

Embed Size (px)

Citation preview

Page 1: Elementary Functions

Step functionIn mathematics, a function on the real numbers is called a step function (or staircase function) if it can be written as a finite linear combination of indicator functions of intervals. Informally speaking, a step function is a piecewise constant function having only finitely many pieces.

Example of a step function (the red graph). This particular step function is right-continuous.

[edit] Definition and first consequences

A function is called a step function if it can be written as

for all real numbers x

where αi are real numbers, Ai are intervals, and is the indicator function of A:

In this definition, the intervals Ai can be assumed to have the following two properties:

1. The intervals are disjoint, for

2. The union of the intervals is the entire real line,

Page 2: Elementary Functions

Indeed, if that is not the case to start with, a different set of intervals can be picked for which these assumptions hold. For example, the step function

can be written as

[edit] Examples

The Heaviside step function is an often used step function.

A constant function is a trivial example of a step function. Then there is only one

interval, The Heaviside function H(x) is an important step function. It is the mathematical

concept behind some test signals, such as those used to determine the step response of a dynamical system.

The rectangular function, the next simplest step function.

Page 3: Elementary Functions

The rectangular function, the normalized boxcar function, is the next simplest step function, and is used to model a unit pulse.

[edit] Non-examples

The integer part function is not a step function according to the definition of this article, since it has an infinite number of intervals. However, some authors define step functions also with an infinite number of intervals.[1]

[edit] Properties

The sum and product of two step functions is again a step function. The product of a step function with a number is also a step function. As such, the step functions form an algebra over the real numbers.

A step function takes only a finite number of values. If the intervals Ai, in the above definition of the step function are disjoint and their

union is the real line, then for all

The Lebesgue integral of a step function is

where is the length of the interval A, and it is

assumed here that all intervals Ai have finite length. In fact, this equality (viewed as a definition) can be the first step in constructing the Lebesgue integral.[2]

Heaviside step function

The Heaviside step function, using the half-maximum convention

The Heaviside step function, or the unit step function, usually denoted by H (but sometimes u or θ), is a discontinuous function whose value is zero for negative argument and one for positive argument. It seldom matters what value is used for H(0), since H is mostly used as a distribution. Some common choices can be seen below.

Page 4: Elementary Functions

The function is used in the mathematics of control theory and signal processing to represent a signal that switches on at a specified time and stays switched on indefinitely. It is also used in structural mechanics together with the Dirac delta function to describe different types of structural loads. It was named after the English polymath Oliver Heaviside.

It is the cumulative distribution function of a random variable which is almost surely 0. (See constant random variable.)

The Heaviside function is the integral of the Dirac delta function: H′ = δ. This is sometimes written as

although this expansion may not hold (or even make sense) for x = 0, depending on which formalism one uses to give meaning to integrals

involving δ.

[edit] Discrete form

An alternative form of the unit step, as a function of a discrete variable n:

where n is an integer. Unlike the usual (not discrete) case, the definition of H[0] is significant.

The discrete-time unit impulse is the first difference of the discrete-time step

This function is the cumulative summation of the Kronecker delta:

where

is the discrete unit impulse function.

Page 5: Elementary Functions

[edit] Analytic approximations

For a smooth approximation to the step function, one can use the logistic function

where a larger k corresponds to a sharper transition at x = 0. If we take H(0) = ½, equality holds in the limit:

There are many other smooth, analytic approximations to the step function.[1] Among the possibilities are:

These limits hold pointwise and in the sense of distributions. In general, however, pointwise convergence need not imply distributional convergence, and vice-versa distributional convergence need not imply pointwise convergence.

In general, any cumulative distribution function (c.d.f.) of a continuous probability distribution that is peaked around zero and has a parameter that controls for variance can serve as an approximation, in the limit as the variance approaches zero. For example, all three of the above approximations are c.d.f.s of common probability distributions: The logistic, Cauchy and normal distributions, respectively.

[edit] Integral representations

Often an integral representation of the Heaviside step function is useful:

[edit] Zero argument

Page 6: Elementary Functions

Since H is usually used in integration, and the value of a function at a single point does not affect its integral, it rarely matters what particular value is chosen of H(0). Indeed when H is considered as a distribution or an element of (see Lp space) it does not even make sense to talk of a value at zero, since such objects are only defined almost everywhere. If using some analytic approximation (as in the examples above) then often whatever happens to be the relevant limit at zero is used.

There exist, however, reasons for choosing a particular value.

H(0) = ½ is often used since the graph then has rotational symmetry; put another way, H-½ is then an odd function. In this case the following relation with the sign function holds for all x:

H(0) = 1 is used when H needs to be right-continuous. For instance cumulative distribution functions are usually taken to be right continuous, as are functions integrated against in Lebesgue–Stieltjes integration. In this case H is the indicator function of a closed semi-infinite interval:

H(0) = 0 is used when H needs to be left-continuous. In this case H is an indicator function of an open semi-infinite interval:

[edit] Antiderivative and derivative

The ramp function is the antiderivative of the Heaviside step function:

The distributional derivative of the Heaviside step function is the Dirac delta function: dH(x) / dx = δ(x).

[edit] Fourier transform

The Fourier transform of the Heaviside step function is a distribution. Using one choice of constants for the definition of the Fourier transform we have

Page 7: Elementary Functions

Here is the distribution that takes a test function ϕ to the Cauchy

principal value of The limit appearing in the integral is also taken in the sense of (tempered) distributions.

[edit] Algebraic representation

If n is a decimal number with no more than d decimal digits, the Heaviside step function can be represented by means of the following algebraic expression:

where p and q are arbitrary integers that satisfy , and δn0 is a Kronecker delta function.

For instance, if n is integer, the simplest choice is: p = 2, q = 1. On the other hand, if n belongs to a set of decimal numbers with d decimal digits, the simplest choice is: p = 10d + 1, q = 1.[citation needed]

[edit] Hyperfunction representation

This can be represented as a hyperfunction as

Unit doubletIn mathematics, the unit doublet is the derivative of the Dirac delta function. It can be used to differentiate signals in electrical engineering:[1] If u1 is the unit doublet, then

Page 8: Elementary Functions

where * refers to the convolution operator.

The function is zero for all values except zero, where its behaviour is interesting. Its integral over any interval enclosing zero is zero. However, the integral of its absolute value over any region enclosing zero goes to infinity. The function can be thought of as the limiting case of two rectangles, one in the second quadrant, and the other in the fourth. The length of each rectangle is k, whereas their breadth is 1/k2, where k tends to zero.

Triangular function

Triangular function

The triangular function (also known as the triangle function, hat function, or tent function) is defined either as:

or, equivalently, as the convolution of two identical unit rectangular functions:

The triangular function can also be represented as the product of the rectangular and absolute value functions:

Page 9: Elementary Functions

The function is useful in signal processing and communication systems engineering as a representation of an idealized signal, and as a prototype or kernel from which more realistic signals can be derived. It also has applications in pulse code modulation as a pulse shape for transmitting digital signals and as a matched filter for receiving the signals. It is also equivalent to the triangular window sometimes called the Bartlett window.

[edit] Scaling

For any parameter, :

[edit] Fourier transform

The transform is easily determined using the convolution property of Fourier transforms and the Fourier transform of the rectangular function:

.

Rectangular function

Page 10: Elementary Functions

Rectangular function

The rectangular function (also known as the rectangle function, rect function, Pi function, gate function, unit pulse, or the normalized boxcar function) is defined as:[1]

Alternate definitions of the function define to be 0, 1, or undefined.

[edit] Relation to the Step Function

The rectangular function may be expressed in terms of the Heaviside step function as [1]

or

More generally:

Page 11: Elementary Functions

or

[edit] Relation to the Boxcar Function

The rectangular function is a special case of the more general boxcar function:

Where the function is centred at X and has duration Y.

Boxcar function

A graphical representation of a boxcar function.

In mathematics, a boxcar function is any function which is zero over the entire real line except for a single interval where it is equal to a constant, A; it is a simple step function. The boxcar function can be expressed in terms of the uniform distribution as

where f(a,b;x) is the uniform distribution of x for the interval [a, b]. As with most such discontinuous functions, there is a question of the value at the transition points. These values are probably best chosen for each individual application.

When a boxcar function is selected as the impulse response of a filter, the result is a moving average filter.

Page 12: Elementary Functions

[edit] Fourier transform of the rectangular function

The unitary Fourier transforms of the rectangular function are:[1]

and:

where sinc is the normalized form.

Note that as long as the definition of the pulse function is only motivated by the time-domain experience of it, there is no reason to believe that the oscillatory interpretation (i.e. the Fourier transform function) should be intuitive, or directly understood by humans. However, some aspects of the theoretical result may be understood intuitively, such as the infinite bandwidth requirement incurred by the indefinitely-sharp edges in the time-domain definition.

[edit] Relation to the Triangular Function

We can define the triangular function as the convolution of two rectangular functions:

[edit] Use in probability

Viewing the rectangular function as a probability density function, it is a special case of the continuous uniform distribution with

. The characteristic function is:

Page 13: Elementary Functions

and its moment generating function is:

where sinh(t) is the hyperbolic sine function.

[edit] Rational approximation

The pulse function may also be expressed as a limit of a rational function:

Demonstration of validity

First, we consider the case where . Notice that the term (2t)2n is always positive for integer n. However, 2t < 1 and hence (2t)^{2n} approaches zero for large n.

It follows that:

Second, we consider the case where . Notice that the term (2t)2n is always positive for integer n. However, 2t > 1 and hence (2t)^{2n} grows very large for large n.

It follows that:

Third, we consider the case where . We may simply substitute in our equation:

Page 14: Elementary Functions

We see that it satisfies the definition of the pulse function.

Sign function

Signum function y = sgn(x)

In mathematics, the sign function is an odd mathematical function that extracts the sign of a real number. To avoid confusion with the sine function, this function is often called the signum function (from signum, Latin for "sign").

In mathematical expressions the sign function is often represented as sgn.

[edit] Definition

The signum function of a real number x is defined as follows:

[edit] Properties

Page 15: Elementary Functions

Any real number can be expressed as the product of its absolute value and its sign function:

From equation (1) it follows that whenever x is not equal to 0 we have

The signum function is the derivative of the absolute value function (up to the indeterminacy at zero): Note, the resultant power of x is 0, similar to the ordinary derivative of x. The numbers cancel and all we are left with is the sign of x.

.

The signum function is differentiable with derivative 0 everywhere except at 0. It is not differentiable at 0 in the ordinary sense, but under the generalised notion of differentiation in distribution theory, the derivative of the signum function is two times the Dirac delta function,

The signum function is related to the Heaviside step function H1/2(x) thus:

where the 1/2 subscript of the step function means that H1/2(0) = 1/2. The signum can also be written using the Iverson bracket notation:

For , a smooth approximation of the sign function is

Another approximation is

which gets sharper as , note that it's the derivative of . This is inspired from the fact that the above is exactly equal for all nonzero x if

Page 16: Elementary Functions

, and has the advantage of simple generalization to higher dimensional analogues of the sign function

(for example, the partial derivatives of ).

[edit] Complex signum

The signum function can be generalized to complex numbers as

for any z ∈ except z = 0. The signum of a given complex number z is the point on the unit circle of the complex plane that is nearest to z. Then, for z ≠ 0,

where arg is the complex argument function. For reasons of symmetry, and to keep this a proper generalization of the signum function on the reals, also in the complex domain one usually defines sgn 0 = 0.

Another generalization of the sign function for real and complex expressions is csgn,[1] which is defined as:

where is the real part of z, is the imaginary part of z.

We then have (except for z = 0):

[edit] Generalized signum function

Page 17: Elementary Functions

At real values of , it is possible to define a generalized function–version

of the signum function, ε(x), such that everywhere, including at

the point (unlike , for which ). This generalized signum allows construction of the algebra of generalized functions, but the price of such generalization is the loss of commutativity. In particular, the generalized signum anticommutes with the delta-function,[2]

in addition, cannot be evaluated at ; and the special name, ε is necessary to distinguish it from the function . (ε(0) is not defined, but

.)

[edit] Algebraic representation

If n is a decimal number with no more than d decimal digits, the signum function can be represented by means of the following algebraic expression:

Page 18: Elementary Functions

where p and q are arbitrary integers that

satisfy , and δn0 is a Kronecker delta function.

For instance, if n is integer, the simplest choice is: p = 2, q = 1. On the other hand, if n belongs to a set of decimal numbers with d decimal digits, the simplest choice is: p = 10d + 1, q = 1.

Sigmoid functionFrom Wikipedia, the free encyclopediaJump to: navigation, search

This article needs additional citations for verification. Please help improve this article by adding citations to reliable sources. Unsourced material may be challenged and removed. (May 2008)

The logistic curve

Page 19: Elementary Functions

Plot of the error function

Many natural processes, including those of complex system learning curves, exhibit a progression from small beginnings that accelerates and approaches a climax over time. When a detailed description is lacking, a sigmoid function is often used. A sigmoid curve is produced by a mathematical function having an "S" shape. Often, sigmoid function refers to the special case of the logistic function shown at right and defined by the formula

Another example is the Gompertz curve. It is used in modeling systems that saturate at large values of t. Another example is the ogee curve as used in the spillway of some dams. A wide variety of sigmoid functions have been used as the activation function of artificial neurons, including the logistic function and tanh(x).

Contents

[hide]

1 Properties 2 Examples 3 See also 4

References

[edit] Properties

In general, a sigmoid function is real-valued and differentiable, having either a non-negative or non-positive first derivative which is bell

Page 20: Elementary Functions

shaped. There are also a pair of horizontal asymptotes as . The logistic functions are sigmoidal and are characterized as the solutions of the differential equation[1]

[edit] Examples

Some sigmoid functions compared. In the drawing all functions are normalized in such a way that their slope at 0 is 1.

Besides the logistic function, sigmoid functions include the ordinary arctangent, the hyperbolic tangent, and the error function, but also the generalised logistic function and algebraic functions like

.

The integral of any smooth, positive, "bump-shaped" function will be sigmoidal, thus the cumulative distribution functions for many common probability distributions are sigmoidal. The most famous such example is the error function.

Logistic functionFrom Wikipedia, the free encyclopedia

Jump to: navigation, search

For the recurrence relation, see Logistic map.

Page 21: Elementary Functions

Standard logistic sigmoid function

A logistic function or logistic curve is a common sigmoid curve, given its name in 1844 or 1845 by Pierre François Verhulst who studied it in relation to population growth. It can model the "S-shaped" curve (abbreviated S-curve) of growth of some population P. The initial stage of growth is approximately exponential; then, as saturation begins, the growth slows, and at maturity, growth stops.

A simple logistic function may be defined by the formula

where the variable P might be considered to denote a population and the variable t might be thought of as time.[1] For values of t in the range of real numbers from −∞ to +∞, the S-curve shown is obtained. In practice, due to the nature of the exponential function e−t, it is sufficient to compute t over a small range of real numbers such as [−6, +6].

The logistic function finds applications in a range of fields, including artificial neural networks, biology, biomathematics, demography, economics, chemistry, mathematical psychology, probability, sociology, political science, and statistics. It has an easily calculated derivative:

It also has the property that

Page 22: Elementary Functions

In other words, the function P − 1/2 is odd.

[edit] Logistic differential equation

The logistic function is the solution of the simple first-order non-linear differential equation

where P is a variable with respect to time t and with boundary condition P(0) = 1/2. This equation is the continuous version of the logistic map.

The qualitative behavior is easily understood in terms of the phase line: the derivative is 0 at P = 0 or 1 and the derivative is positive for P between 0 and 1, and negative for P above 1 or less than 0 (though negative populations do not generally accord with a physical model). This yields an unstable equilibrium at 0, and a stable equilibrium at 1, and thus for any value of P greater than 0 and less than 1, P grows to 1.

One may readily find the (symbolic) solution to be

Choosing the constant of integration ec = 1 gives the other well-known form of the definition of the logistic curve

More quantitatively, as can be seen from the analytical solution, the logistic curve shows early exponential growth for negative t, which slows to linear growth of slope 1/4 near t = 0, then approaches y = 1 with an exponentially decaying gap.

The logistic function is the inverse of the natural logit function and so can be used to convert the logarithm of odds into a probability; the conversion from the log-likelihood ratio of two alternatives also takes the form of a logistic curve.

Page 23: Elementary Functions

The logistic sigmoid function is related to the hyperbolic tangent, A.p. by

[edit] In ecology: modeling population growth

Pierre-François Verhulst (1804–1849)

A typical application of the logistic equation is a common model of population growth, originally due to Pierre-François Verhulst in 1838, where the rate of reproduction is proportional to both the existing population and the amount of available resources, all else being equal. The Verhulst equation was published after Verhulst had read Thomas Malthus' An Essay on the Principle of Population. Verhulst derived his logistic equation to describe the self-limiting growth of a biological population. The equation is also sometimes called the Verhulst-Pearl equation following its rediscovery in 1920. Alfred J. Lotka derived the equation again in 1925, calling it the law of population growth.

Letting P represent population size (N is often used in ecology instead) and t represent time, this model is formalized by the differential equation:

Page 24: Elementary Functions

where the constant r defines the growth rate and K is the carrying capacity.

In the equation, the early, unimpeded growth rate is modeled by the first term +rP. The value of the rate r represents the proportional increase of the population P in one unit of time. Later, as the population grows, the second term, which multiplied out is −rP2/K, becomes larger than the first as some members of the population P interfere with each other by competing for some critical resource, such as food or living space. This antagonistic effect is called the bottleneck, and is modeled by the value of the parameter K. The competition diminishes the combined growth rate, until the value of P ceases to grow (this is called maturity of the population).

Dividing both sides of the equation by K gives

Now setting x = P / K gives the differential equation

For r = 1 we have the particular case with which we started.

In ecology, species are sometimes referred to as r-strategist or K-strategist depending upon the selective processes that have shaped their life history strategies. The solution to the equation (with P0 being the initial population) is

where

Page 25: Elementary Functions

Which is to say that K is the limiting value of P: the highest value that the population can reach given infinite time (or come close to reaching in finite time). It is important to stress that the carrying capacity is asymptotically reached independently of the initial value P(0) > 0, also in case that P(0) > K.

[edit] Time-varying carrying capacity

Since the environmental conditions influence the carrying capacity, as a consequence it can be time-varying: K(t) > 0, leading to the following mathematical model:

A particularly important case is that of carrying capacity that varies periodically with period T:

It can be shown that in such a case, independently from the initial value P(0) > 0, P(t) will tend to a unique periodic solution P*(t), whose period is T.

A typical value of T is one year: in such case K(t) reflects periodical variations of weather conditions.

Another interesting generalization is to consider that the carrying capacity K(t) is a function of the population at an earlier time, capturing a delay in the way population modifies its environment. This leads to a logistic delay

Page 26: Elementary Functions

equation,[2] which has a very rich behavior, with bistability in some parameter range, as well as a monotonic decay to zero, smooth exponential growth, punctuated unlimited growth (i.e., multiple S-shapes), punctuated growth or alternation to a stationary level, oscillatory approach to a stationary level, sustainable oscillations, finite-time singularities as well as finite-time death.

[edit] In neural networks

Logistic functions are often used in neural networks to introduce nonlinearity in the model and/or to clamp signals to within a specified range. A popular neural net element computes a linear combination of its input signals, and applies a bounded logistic function to the result; this model can be seen as a "smoothed" variant of the classical threshold neuron.

A common choice for the activation or "squashing" functions, used to clip for large magnitudes to keep the response of the neural network bounded[3] is

which we recognize to be of the form of the logistic function. These relationships result in

Page 27: Elementary Functions

simplified implementations of artificial neural networks with artificial neurons. Practitioners caution that sigmoidal functions which are antisymmetric about the origin (e.g. the hyperbolic tangent) lead to faster convergence when training networks with backpropagation.[4]

[edit] In statistics

Logistic functions are used in several roles in statistics. Firstly, they are the cumulative distribution function of the logistic family of distributions. Secondly they are used in logistic regression to model how the probability p of an event may be affected by one or more explanatory variables: an example would be to have the model

where x is the explanatory variable and a and b are model parameters to be fitted.

An important application of the logistic function is in the Rasch model, used in item response theory. In particular, the Rasch model forms a basis for maximum likelihood estimation of

Page 28: Elementary Functions

the locations of objects or persons on a continuum, based on collections of categorical data, for example the abilities of persons on a continuum based on responses that have been categorized as correct and incorrect.

[edit] In medicine: modeling of growth of tumors

Another application of logistic curve is in medicine, where the logistic differential equation is used to model the growth of tumors. This application can be considered an extension of the above mentioned use in the framework of ecology. Denoting with X(t) the size of the tumor at time t, its dynamics are governed by:

which is of the type:

where F(X) is the proliferation rate of the tumor.

Page 29: Elementary Functions

If a chemotherapy is started with a log-kill effect, the equation may be revised to be

where c(t) is the therapy-induced death rate. In the idealized case of very long therapy, c(t) can be modeled as a periodic function (of period T) or (in case of continuous infusion therapy) as a constant function, and one has that

i.e. if the average therapy-induced death rate is greater than the baseline proliferation rate then there is the eradication of the disease. Of

Page 30: Elementary Functions

course, this is an over-simplified model of both the growth and the therapy (e.g. it does not take into account the phenomenon of clonal resistance).

[edit] In chemistry: reaction models

The concentration of reactants and products in autocatalytic reactions follow the logistic function.

Page 31: Elementary Functions

[edit] In physics: Fermi distribution

The logistic function determines the statistical distribution of fermions over the energy states of a system in thermal equilibrium. In particular, it is the distribution of the probabilities that each possible energy level is occupied by a fermion, according to Fermi–Dirac statistics.

[edit] In

Page 32: Elementary Functions

linguistics: language change

In linguistics, the logistic function can be used to model language change [5] : an innovation that is at first marginal begins to spread more quickly with time, and then more slowly as it becomes more universally adopted.

[edit] In economics: diffusion of

Page 33: Elementary Functions

innovations

The logistic function can be used to illustrate the progress of the diffusion of an innovation through its life cycle. This method was used in papers by several researchers at the International Institute of Applied Systems Analysis (IIASA). These papers deal with the diffusion of various innovations, infrastructures and energy source substitutions and the role of work in the

Page 34: Elementary Functions

economy as well as with the long economic cycle. Long economic cycles were investigated by Robert Ayres (1989).[6] Cesare Marchetti published on long economic cycles and on diffusion of innovations.[7][8] Arnulf Grübler’s book (1990) gives a detailed account of the diffusion of infrastructures including canals, railroads, highways and airlines , showing that their diffusion followed logistic

Page 35: Elementary Functions

shaped curves.[9]

Carlota Perez used a logistic curve to illustrate the long (Kondratiev) business cycle with the following labels: beginning of a technological era as irruption, the ascent as frenzy, the rapid build out as synergy and the completion as maturity.[10]

[edit] Double logistic function

Page 36: Elementary Functions

Double logistic sigmoid curve

The double logistic is a function similar to the logistic function with numerous applications. Its general formula is:

where d is its centre and s is the steepness factor. Here "sgn" represents the sign function.

Page 37: Elementary Functions

It is based on the Gaussian curve and graphically it is similar to two identical logistic sigmoids bonded together at the point x = d.

One of its applications is non-linear normalization of a sample, as it has the property of eliminating outliers.

Ramp function

Page 38: Elementary Functions

From Wikipedia, the free encyclopedia

Jump to: navigation, search

The ramp function is an elementary unary real function, easily computable as the mean of its independent variable and its absolute value.

This function is applied in engineering (e.g., in the theory of DSP). The name ramp function can be derived by the look of its graph.

Contents

[hide]

1 Definitions 2 Analytic properties

o 2.1 Non-negativity o 2.2 Derivative o 2.3 Fourier transform o 2.4 Laplace transform

3 Algebraic properties o 3.1 Iteration invariance

4 References

[edit] Definitions

Graph of the ramp function

The ramp function ( ) may be defined analytically in several ways. Possible definitions are:

The mean of a straight line with unity gradient and its modulus:

Page 39: Elementary Functions

this can be derived by noting the following definition of ,

for which a = x and b = 0

The Heaviside step function multiplied by a straight line with unity gradient:

The convolution of the Heaviside step function with itself:

The integral of the Heaviside step function:

[edit] Analytic properties

[edit] Non-negativity

In the whole domain the function is non-negative, so its absolute value is itself, i.e.

and

Proof: by the mean of definition [2] it is non-negative in the I. quarter, and zero in the II.; so everywhere it is non-negative.

[edit] Derivative

Its derivative is the Heaviside function:

Page 40: Elementary Functions

From this property definition [5]. goes.

[edit] Fourier transform

= =

Where δ(x) is the Dirac delta (in this formula, its derivative appears).

[edit] Laplace transform

The single-sided Laplace transform of R(x) is given as follows,

[edit] Algebraic properties

[edit] Iteration invariance

Every iterated function of the ramp mapping is itself, as

.

Proof: =

= .

We applied the non-negative property.

Dirac delta functionFrom Wikipedia, the free encyclopedia

Jump to: navigation, search

Page 41: Elementary Functions

Schematic representation of the Dirac delta function by a line surmounted by an arrow. The height of the arrow is usually used to specify the value of any multiplicative constant, which will give the area under the function. The other convention is to write the area next to the arrowhead.

The Dirac delta function as the limit (in the sense of distributions) of the sequence of

Gaussians as

The Dirac delta function, or δ function, is (informally) a generalized function depending on a real parameter such that it is zero for all values of the parameter except when the parameter is zero, and its integral over the parameter from −∞ to ∞ is equal to one.[1][2] It was introduced by theoretical physicist Paul Dirac. In the context of signal processing it is

Page 42: Elementary Functions

often referred to as the unit impulse function. It is a continuous analog of the Kronecker delta function which is usually defined on a finite domain, and takes values 0 and 1.

From a purely mathematical viewpoint, the Dirac delta is not strictly a function, because any extended-real function that is equal to zero everywhere but a single point must have total integral zero.[3] While for many purposes the Dirac delta can be manipulated as a function, formally it can be defined as a distribution that is also a measure. In many applications, the Dirac delta is regarded as a kind of limit (a weak limit) of a sequence of functions having a tall spike at the origin. The approximating functions of the sequence are thus "approximate" or "nascent" delta functions.

Contents

[hide]

1 Overview 2 Definitions

o 2.1 As a measure o 2.2 As a distribution o 2.3 Generalizations

3 Properties o 3.1 Scaling and symmetry o 3.2 Algebraic properties o 3.3 Translation o 3.4 Composition with a function o 3.5 Properties in n dimensions

4 Fourier transform 5 Distributional derivatives

o 5.1 Higher dimensions 6 Representations of the delta function

o 6.1 Approximations to the identity o 6.2 Probabilistic considerations o 6.3 Semigroups o 6.4 Oscillatory integrals o 6.5 Plane wave decomposition o 6.6 Fourier kernels o 6.7 Hilbert space theory o 6.8 Infinitesimal delta functions

7 Dirac comb 8 Sokhatsky–Weierstrass theorem 9 Relationship to the Kronecker delta 10 Applications to probability theory 11 Application to quantum mechanics 12 Application to structural mechanics 13 See also 14 Notes 15 References 16 External links

Page 43: Elementary Functions

[edit] Overview

The graph of the delta function is usually thought of as following the whole x-axis and the positive y-axis. (This informal picture can sometimes be misleading, for example in the limiting case of the sinc function.)

Despite its name, the delta function is not truly a function, at least not a usual one with domain in reals. For example, the objects f(x) = δ(x) and g(x) = 0 are equal everywhere except at x = 0 yet have integrals that are different. According to Lebesgue integration theory, if f and g are functions such that f = g almost everywhere, then f is integrable if and only if g is integrable and the integrals of f and g are identical. Rigorous treatment of the Dirac delta requires measure theory, the theory of distributions, or a hyperreal framework.

The Dirac delta is used to model a tall narrow spike function (an impulse), and other similar abstractions such as a point charge, point mass or electron point. For example, to calculate the dynamics of a baseball being hit by a bat, one can approximate the force of the bat hitting the baseball by a delta function. In doing so, one not only simplifies the equations, but one also is able to calculate the motion of the baseball by only considering the total impulse of the bat against the ball rather than requiring knowledge of the details of how the bat transferred energy to the ball.

In applied mathematics, the delta function is often manipulated as a kind of limit (a weak limit) of a sequence of functions, each member of which has a tall spike at the origin: for example, a sequence of Gaussian distributions centered at the origin with variance tending to zero.

An infinitesimal formula for an infinitely tall, unit impulse delta function (infinitesimal version of Cauchy distribution) explicitly appears in an 1827 text of Augustin Louis Cauchy.[4] Siméon Denis Poisson considered the issue in connection with the study of wave propagation as did Gustav Kirchhoff somewhat later. Kirchhoff and Hermann von Helmholtz also introduced the unit impulse as a limit of Gaussians, which also corresponded to Lord Kelvin's notion of a point heat source. At the end of the 19th century, Oliver Heaviside used formal Fourier series to manipulate the unit impulse.[5] The Dirac delta function as such was introduced as a "convenient notation" by Paul Dirac in his influential 1927 book Principles of Quantum Mechanics.[6] He called it the "delta function" since he used it as a continuous analogue of the discrete Kronecker delta.

[edit] Definitions

The Dirac delta can be loosely thought of as a function on the real line which is zero everywhere except at the origin, where it is infinite,

Page 44: Elementary Functions

and which is also constrained to satisfy the identity

[7]

This is merely a heuristic definition. The Dirac delta is not a true function, as no function has the above properties.[6] Moreover there exist descriptions of the delta function which differ from the above conceptualization. For example, sinc(x/a)/a becomes the delta function in the limit as a → 0,[8] yet this function does not approach zero for values of x outside the origin, rather it oscillates between 1/x and −1/x more and more rapidly as a approaches zero.

The Dirac delta function can be rigorously defined either as a distribution or as a measure.

[edit] As a measure

One way to rigorously define the delta function is as a measure, which accepts as an argument a subset A of the real line R, and returns δ(A) = 1 if 0 ∈ A, and δ(A) = 0 otherwise.[9] If the delta function is conceptualized as modeling an idealized point mass at 0, then δ(A) represents the mass contained in the set A. One may then define the integral against δ as the integral of a function against this mass distribution. Formally, the Lebesgue integral provides the necessary analytic device. The Lebesgue integral with respect to the measure δ satisfies

for all continuous compactly supported functions ƒ. The measure δ is not absolutely continuous with respect to the Lebesgue measure — in fact, it is a singular measure. Consequently, the delta measure has no Radon–Nikodym derivative — no true function for which the property

holds.[10] As a result, the latter notation is a convenient abuse of notation, and not a standard (Riemann or Lebesgue) integral.

Page 45: Elementary Functions

As a probability measure on R, the delta measure is characterized by its cumulative distribution function, which is the unit step function [11]

This means that H(x) is the integral of the cumulative indicator function 1(−∞, x] with respect to the measure δ; to wit,

Thus in particular the integral of the delta function against a continuous function can be properly understood as a Stieltjes integral:[12]

All higher moments of δ are zero. In particular, characteristic function and moment generating function are both equal to one.

[edit] As a distribution

In the theory of distributions a generalized function is thought of not as a function itself, but only in relation to how it affects other functions when it is "integrated" against them. In keeping with this philosophy, to define the delta function properly, it is enough to say what the "integral" of the delta function against a sufficiently "good" test function is. If the delta function is already understood as a measure, then the Lebesgue integral of a test function against that measure supplies the necessary integral.

A typical space of test functions consists of all smooth functions on R with compact support. As a distribution, the Dirac delta is a linear functional on the space of test functions and is defined by[13]

(1)

Page 46: Elementary Functions

for every test function φ.

For δ to be properly a distribution, it must be "continuous" in a suitable sense. In general, for a linear functional S on the space of test functions to define a distribution, it is necessary and sufficient that, for every positive integer N there is an integer MN and a constant CN such that for every test function φ, one has the inequality[14]

With the δ distribution, one has such an inequality (with CN = 1) with MN = 0 for all N. Thus δ is a distribution of order zero. It is, furthermore, a distribution with compact support (the support being {0}).

The delta distribution can also be defined in a number of equivalent ways. For instance, it is the distributional derivative of the Heaviside step function. This means that, for every test function φ, one has

Intuitively, if integration by parts were permitted, then the latter integral should simplify to

and indeed, a form of integration by parts is permitted for the Stieltjes integral, and in that case one does have

Page 47: Elementary Functions

In the context of measure theory, the Dirac measure gives rise to a distribution by integration. Conversely, equation (1) defines a Daniell integral on the space of all compactly supported continuous functions φ which, by the Riesz representation theorem, can be represented as the Lebesgue integral of φ with respect to some Radon measure.

[edit] Generalizations

The delta function can be defined in n-dimensional Euclidean space Rn as the measure such that

for every compactly supported continuous function ƒ. As a measure, the n-dimensional delta function is the product measure of the 1-dimensional delta functions in each variable separately. Thus, formally, with x = (x1,x2,...,xn), one has[15]

(2)

The delta function can also be defined in the sense of distributions exactly as above in the one-dimensional case.[16] However, despite widespread use in engineering contexts, (2) should be manipulated with care, since the product of distributions can only be defined under quite narrow circumstances.[17]

Page 48: Elementary Functions

The notion of a Dirac measure makes sense on any set whatsoever.[9] Thus if X is a set, x0 ∈ X is a marked point, and Σ is any sigma algebra of subsets of X, then the measure defined on sets A ∈ Σ by

is the delta measure or unit mass concentrated at x0.

Another common generalization of the delta function is to a differentiable manifold where most of its properties as a distribution can also be exploited because of the differentiable structure. The delta function on a manifold M centered at the point x0 ∈ M is defined as the following distribution:

(3)

for all compactly supported smooth real-valued functions φ on M.[18] A common special case of this construction is when M is an open set in the Euclidean space Rn.

On a locally compact Hausdorff space X, the

Page 49: Elementary Functions

Dirac delta measure concentrated at a point x is the Radon measure associated with the Daniell integral (3) on compactly supported continuous functions φ. At this level of generality, calculus as such is no longer possible, however a variety of techniques from abstract analysis are available. For instance, the mapping

is a continuous embedding of X into the space of finite Radon measures on X, equipped with its vague topology. Moreover, the convex hull of the image of X under this embedding is dense in the space of probability measures on X.[19]

[edit] Properties

[edit] Scaling and symmetry

The delta function satisfies the following scaling property for a non-zero scalar α:[20]

and so

(4)

Page 50: Elementary Functions

In particular, the delta function is an even distribution, in the sense that

δ( − x) = δ(x)

which is homogeneous of degree −1.

[edit] Algebraic properties

The distributional product of δ with x is equal to zero:

xδ(x) = 0.

Conversely, if xf(x) = xg(x), where f and g are distributions, then

f(x) = g(x) + cδ(x)

for some constant c.

[edit] Translation

The integral of the time-delayed Dirac delta is

Page 51: Elementary Functions

given by:

This is sometimes referred to as the sifting property[

21] or the sampling property. The delta function is said to "sift out" the value at

.

It follows that

Page 52: Elementary Functions

the effect of convolving a function ƒ(t) with the time-delayed Dirac delta is to time-delay ƒ(t) by the same amount:

(using (4): δ( − x) = δ(x))

This

Page 55: Elementary Functions

special case, for instance, we have the identity (underst

Page 56: Elementary Functions

ood in the distribution sense)

[edit] Composition with a function

More generally, the delta di

Page 58: Elementary Functions

that the familiar change of variables formula holds, that

provided that g is a

Page 59: Elementary Functions

continuously differentiable function with g′ nowhere zero.[22] That is, there is a unique way to assign meaning to the distribution

so that this identity holds for all compactly supported test

Page 60: Elementary Functions

functions ƒ. This distribution satisfies δ(g(x)) = 0 if g is nowhere zero, and otherwise if g has a real root at x0, then

It is natural therefore to define the composition δ(g(x)) for continuously differentiable functions g by

Page 61: Elementary Functions

where the sum extends over all roots of g(x), which are assumed to be simple.[22] Thus, for example

In the integral form the generalized scaling property may be written as

[edit] Properties in n dimensions

The delta distribution in an n-dimensional space satisfies the following scaling property instead:

so that δ is a homogeneousdistribution of degree −n. Under any reflection

Page 62: Elementary Functions

rotation ρ, the delta function is invariant:

As in the one-variable case, it is possible to define the composition of δ with a bi-Lipschitz function [23] guniquely so that the identity

for all compactly supported functions ƒ.

Using the coarea formulafrom geometric measure theory, one can also define the composition of the delta function with a submersion from one Euclidean space to another one of different dimension; the result is a type of current. In the special case of a continuously differentiable function Rn → R such that the gradient of gzero, the following identity holds[24]

where the integral on the right is over dimensional surface defined by g(x) = 0 with respect to the Minkowski contentmeasure. This is known as a simple layer

Page 63: Elementary Functions

[edit] Fourier transform

The delta function is a tempered distributiontherefore it has a well-defined Fourier transformone finds[25]

Properly speaking, the Fourier transform of a distribution is defined by imposing adjointness of the Fourier transform under the duality

pairing of tempered distributions with

functions. Thus unique tempered distribution satisfying

for all Schwartz functions φ. And indeed it follows from this that

As a result of this identity, the convolution of the delta function with any other tempered distribution S is simply S

That is to say that δ is an element for the convolution on tempered distributions, and in fact the space of compactly supported distributions under convolution is an associative algebradelta function. This property is fundamental in convolution with a tempered distribution is a

Page 64: Elementary Functions

system, and applying the linear time-invariant system measures its response. The impulse response can be computed to any desired degree of accuracy by choosing a suitable approximation for δ, and once it is known, it characterizes the system completely. See theory:Impulse response and convolution.

The inverse Fourier transform of the tempered distribution delta function. Formally, this is expressed

and more rigorously, it follows since

for all Schwartz functions

In these terms, the delta function provides a suggestive statement of the orthogonality property of the Fourier kernel on one has

This is, of course, shorthand for the assertion that the Fourier transform of the tempered distribution

is

which again follows by imposing self-adjointness of the Fourier transform.

Page 65: Elementary Functions

By analytic continuationLaplace transform

[edit] Distributional derivatives

The distributional derivative of the Dirac delta distribution is the distribution δ′ defined on compactly supported smooth test functions φ by

δ'[φ] = − δ[φ'] = − φ'(0).

The first equality here is a kind of integration by parts, for if δ were a true function then

The kth derivative of δ is defined similarly as the distribution given on test functions by

δ(k)[φ] = ( − 1)kφ(k)(0).

In particular δ is an infinitely differentiable distribution.

The first derivative of the delta function is the distributional limit of the difference quotients:

More properly, one has

where τh is the translation operator, defined on functions by τφ(x+h), and on a distribution

(τhS)[φ] = S[τ − hφ].

In the theory of represents a point magnetic referred to as a dipole or the

Page 66: Elementary Functions

The derivative of the delta function satisfies a number of basic properties, including:

xδ'(x) = − δ(x).

Furthermore, the convolution of δ' with a compactly supported smooth function

δ' * f = δ * f' = f',

which follows from the properties of the distributional derivative of a convolution.

[edit] Higher dimensions

More generally, on an delta distribution centered at a point

δa[ϕ] = ϕ(a)

for all φ ∈ S(αn) is any multi-indexthen the αth derivative ∂

That is, the αderivative of

The first partial derivatives of the delta function are thought of as coordinate planes. More generally, the surface is a double layer supported on that surface, and represents a laminar magnetic monopole. Higher derivatives of the delta function are known in physics as

Higher derivatives enter into mathematics naturally as the building blocks for the complete structure of distributions with point support. If {a} consisting of a single point, then there is an integer

[edit] Representations of the delta function

The delta function can be viewed as the limit of a sequence of functions

Page 67: Elementary Functions

where ηε(x) is sometimes called a either that

(5)

for all continuouswith compact support. The difference between these two slightly different modes of weak convergence is often subtle: the former is convergence in the convergence in the sense of

[edit] Approximations to the identity

Typically a nascent delta function ηintegrable function on

In n dimensions, one uses instead the scaling

Then a simple change of variables shows that ηcontinuous compactly supported functions ƒ, and so ηinitial η = η1

The ηε constructed in this way are known as an because the space functions: ƒ∗product: no element identity in the sense that

This limit holds in the sense of that it be a mollifier associated to a compactly supported function,convergence

The standard mollifier is given by Ψ(

Page 68: Elementary Functions

where

In some situations such as obtained by taking η

ηε(x) = ε − 1max(1 − | x / ε | ,0)

which are all continuous and compactly supported, although not smooth and so not a mollifier.

[edit] Probabilistic considerations

In the context of the identity should be positive, as such a function then represents a distribution is sometimes favorable because it does not result in combinationany probability distribution at all, and letting general this converges more rapidly to a delta function if, in addition,

instance, if ηwith n a new parameter:

Another example is with the

This is continuous and compactly supported, but not a mollifier because it is not smooth.

[edit] Semigroups

Nascent delta functions often arise as convolution with ηδ must satisfy

ηε * ηδ = ηε + δ

for all ε,δ > 0. Convolution semigroups in Labove sense, however the semigroup condition is quite a strong restriction.

Page 69: Elementary Functions

In practice, semigroups approximating the delta function arise as motivated ellipticof a linear time-invariant systemby solving the

in which the limit is as usual understood in the weak sense. Setting η

Some examples of physically important convolution semigroups arising from such a fundamental solution include the following.

The heat kernel

The heat kernel

represents the temperature in an infinite wire at time semigroup evolves according to the one-dimensional

In probability theoryof a particle starting at the origin following a standard Markov property

In higher dimensional Euclidean space

and has the same physical interpretation, distribution sense as ε → 0.

The Poisson kernel

The Poisson kernel

Page 70: Elementary Functions

is the fundamental solution of the potential along the edge is held at fixed at the delta function. The Poisson kernel is also closely related to the evolves according to the equation

where the operator is rigorously defined as the

[edit] Oscillatory integrals

In areas of physics such as result, the nascent delta functions that arise as fundamental solutions of the associated which comes from a solution of the

Although using the Fourier transform, it is easy to see that this generates a semigroup in some sense, it is not absolutely integrable and so cannot define a semigroup in the above strong sense. Many nascent delta functions constructed as oscillatory integrals only converge in the sense of distributions (an example is the

Another example is the Cauchy problem for the

The solution

Other approximations to the identity of this kind include the

and the Bessel function

[edit] Plane wave decomposition

One approach to the study of a linear partial differential equation

Page 71: Elementary Functions

where L is a

When L is particularly simple, this problem can often be resolved using the Fourier transform directly (as in the case of the Poisson kernel and heat kernel already mentioned). For more complicated operators, it is sometimes easier first to consider an equation of the form

where h is a

for some vector ξ. Such an equation can be resolved (if the coefficients of quadrature. So, if the delta function can be decomposed into plane waves, then one can in principle solve linear partial differential equations.

Such a decomposition of the delta function into plane waves was part of a general technique first introduced essentially by John (1955).

Then δ is obtained by applying a power of the

The Laplacian here is interpreted as a weak derivative, so that this equation is taken to mean that, for any test function φ,

The result follows from the formula for the because it recovers the value of

Page 72: Elementary Functions

where Rφ(ξ,p

An alternative equivalent expression of the plane wave decomposition, from

[edit] Fourier kernels

of period 2π is defined by convolution (on the interval [−π,π]) with the Dirichlet kernel:

Page 75: Elementary Functions

, depending on the properties of the basis φn. In most cases of practical interest, the orthonormal basis comes from an

[edit] Infinitesimal delta functions

Cauchy defined an infinitesimal in Cours d'Analyse (1827) in terms of a sequence tending to zero.

construction, where a null sequence becomes an infinitesimal in the sense of an equivalence class modulo a relation defined in terms of a suitable ultrafilter. The article by Yamashita (2007)

(DSP) and discrete time signal analysis. The Dirac comb is given as the infinite sum, whose

is given by the convolution

Page 77: Elementary Functions

distribution, using a probability density function (which is normally used to represent fully

, with corresponding probabilities , can be written as

, and 4/10 of the time returns exactly the value 3.5 (i.e. a partly continuous, partly discrete mixture distribution). The density function of

). The local time of a stochastic process B(t) is given by

in the range of the process. More precisely, in one dimension this integral can be written

of a particle gives the probability amplitude of finding a particle within a given region of space. Wave functions are assumed to be , and the total probability of finding a particle within a given interval is the integral of the magnitude of the wave function squared over the interval. A set {φn} of wave

where δ here refers to the Kronecker delta. A set of orthonormal wave functions is complete in the space of square-integrable functions if any wave function ψ can be expressed as a combination of the φn:

bound system) in quantum mechanics that measures the energy levels, which are called the , this equality implies the resolution of the identity:

Page 78: Elementary Functions

may be continuous rather than discrete. An example is the position observable, Qψ(x) = xψ(x). The spectrum of the position (in one . However, unlike the Hamiltonian, the position operator lacks proper eigenfunctions. The conventional way to overcome this shortcoming is to widen the

rigged Hilbert space.[50] In this context, the position operator has a complete set of

on the Hilbert space, provided the spectrum of P is continuous and there are no degenerate of distributions indexed by the elements of Ω, such that

has both continuous and discrete parts, then the resolution of the identity involves a summation over the

models for a single and double potential well.

The delta function can be used in structural mechanics to describe transient loads or point loads acting on structures. The governing equation of a simple mass-spring system excited by a sudden force

Page 79: Elementary Functions

Euler-Bernoulli theory,

the load distribution. If a beam is loaded by a point force F at x = x0, the load distribution is written

, it follows that the static deflection of a slender beam subject to multiple point loads is described by a set of piecewise

acting on a beam can be described by delta functions. Consider two opposing point forces F at a distance d apart. They then produce a moment M = Fd acting on the beam. is kept constant. The load distribution, assuming a clockwise moment acting at x = 0, is written

of the delta function. Integration of the beam equation again results in piecewise polynomial deflection.

Unit functionFrom Wikipedia, the free encyclopedia

Jump to: navigation, search

In number theory, the unit function is a completely multiplicative function on the positive integers defined as:

It is called the unit function because it is the identity element for Dirichlet convolution.

Page 80: Elementary Functions

It may be described as the "indicator function of 1" within the set of positive integers. It is also written as u(n) (not to be confused with μ(n)).

Multiplicative functionFrom Wikipedia, the free encyclopedia

Jump to: navigation, search

Outside number theory, the term multiplicative function is usually used for completely multiplicative functions. This article discusses number theoretic multiplicative functions.

In number theory, a multiplicative function is an arithmetic function f(n) of the positive integer n with the property that f(1) = 1 and whenever a and b are coprime, then

f(ab) = f(a) f(b).

An arithmetic function f(n) is said to be completely multiplicative (or totally multiplicative) if f(1) = 1 and f(ab) = f(a) f(b) holds for all positive integers a and b, even when they are not coprime.

Contents

[hide]

1 Examples 2 Properties 3 Convolution

o 3.1 Dirichlet series for some multiplicative functions

4 See also 5 References 6 External links

[edit] Examples

Examples of multiplicative functions include many functions of importance in number theory, such as:

ϕ(n): Euler's totient function ϕ, counting the positive integers coprime to (but not bigger than) n

μ(n): the Möbius function, related to the number of prime factors of square-free numbers

gcd(n,k): the greatest common divisor of n and k, where k is a fixed integer. d(n): the number of positive divisors of n,

Page 81: Elementary Functions

σ(n): the sum of all the positive divisors of n,

σk(n): the divisor function, which is the sum of the k-th powers of all the positive divisors of n (where k may be any complex number). In special cases we have

o σ0(n) = d(n) and

o σ1(n) = σ(n),

a(n): the number of non-isomorphic abelian groups of order n. 1(n): the constant function, defined by 1(n) = 1 (completely multiplicative)

1C(n) the indicator function of the set C of squares (or cubes, or fourth powers, etc.)

Id(n): identity function, defined by Id(n) = n (completely multiplicative) Idk(n): the power functions, defined by Idk(n) = nk for any natural (or even

complex) number k (completely multiplicative). As special cases we haveo Id0(n) = 1(n) ando Id1(n) = Id(n),

(n): the function defined by (n) = 1 if n = 1 and = 0 otherwise, sometimes called multiplication unit for Dirichlet convolution or simply the unit function;

sometimes written as u(n), not to be confused with μ(n) (completely multiplicative).

(n/p), the Legendre symbol, where p is a fixed prime number (completely multiplicative).

λ(n): the Liouville function, related to the number of prime factors dividing n (completely multiplicative).

γ(n), defined by γ(n)=(-1)ω(n), where the additive function ω(n) is the number of distinct primes dividing n.

All Dirichlet characters are completely multiplicative functions.

An example of a non-multiplicative function is the arithmetic function r2(n) - the number of representations of n as a sum of squares of two integers, positive, negative, or zero, where in counting the number of ways, reversal of order is allowed. For example:

1 = 12 + 02 = (-1)2 + 02 = 02 + 12 = 02 + (-1)2

and therefore r2(1) = 4 ≠ 1. This shows that the function is not multiplicative. However, r2(n)/4 is multiplicative.

In the On-Line Encyclopedia of Integer Sequences, sequences of values of a multiplicative function have the keyword "mult".

See arithmetic function for some other examples of non-multiplicative functions.

[edit] Properties

A multiplicative function is completely determined by its values at the powers of prime numbers, a consequence of the

Page 82: Elementary Functions

fundamental theorem of arithmetic. Thus, if n is a product of powers of distinct primes, say n = pa qb ..., then f(n) = f(pa) f(qb) ...

This property of multiplicative functions significantly reduces the need for computation, as in the following examples for n = 144 = 24 · 32:

d(144) = σ0(144) = σ0(24)σ0(32) = (10 + 20 + 40 + 80 + 160)(10 + 30 + 90) = 5 · 3 = 15,

σ(144) = σ1(144) = σ1(24)σ1(32) = (11 + 21 + 41 + 81 + 161)(11 + 31 + 91) = 31 · 13 = 403,

σ*(144) = σ*(24)σ*(32) = (11 + 161)(11 + 91) = 17 · 10 = 170.

Similarly, we have:

ϕ(144)=ϕ(24)ϕ(32) = 8 · 6 = 48

In general, if f(n) is a multiplicative function and a, b are any two positive integers, then

f(a) · f(b) = f(gcd(a,b)) · f(lcm(a,b)).

Every completely multiplicative function is a homomorphism of monoids and is completely determined by its restriction to the prime numbers.

[edit] Convolution

If f and g are two multiplicative functions, one defines a new multiplicative function f * g, the Dirichlet convolution of f and g, by

where the sum extends over all positive divisors d of n. With this operation, the set of all multiplicative functions turns into an abelian group; the identity element is .

Relations among the multiplicative functions discussed above include:

Page 83: Elementary Functions

μ * 1 = (the Möbius inversion formula)

(μ * Idk) * Idk = (generalized Möbius inversion)

ϕ * 1 = Id d = 1 * 1

σ = Id * 1 = ϕ * d

σk = Idk * 1

Id = ϕ * 1 = σ * μ Idk = σk * μ

The Dirichlet convolution can be defined for general arithmetic functions, and yields a ring structure, the Dirichlet ring.

[edit] Dirichlet series for some multiplicative functions

More examples are shown in the article on Dirichlet series.

Indicator functionFrom Wikipedia, the free encyclopediaJump to: navigation, search

The graph of the indicator function of a two-dimensional subset of a square.

Page 84: Elementary Functions

In mathematics, an indicator function or a characteristic function is a function defined on a set X that indicates membership of an element in a subset A of X, having the value 1 for all elements of A and the value 0 for all elements of X not in A.

Contents

[hide]

1 Definition 2 Remark on notation and terminology 3 Basic properties 4 Mean, variance and covariance 5 Characteristic function in recursion theory, Gödel's and Kleene's

representing function 6 Characteristic function in fuzzy set theory 7 See also 8 References

[edit] Definition

The indicator function of a subset A of a set X is a function

defined as

The Iverson bracket allows the equivalent notation, , to be

used instead of

The function is sometimes denoted or or even just . (The Greek letter χ appears because it is the initial letter of the Greek word characteristic.)

[edit] Remark on notation and terminology

The notation may signify the identity function. The notation may signify the characteristic function in convex

analysis.

Page 85: Elementary Functions

A related concept in statistics is that of a dummy variable (this must not be confused with "dummy variables" as that term is usually used in mathematics, also called a bound variable).

The term "characteristic function" has an unrelated meaning in probability theory. For this reason, probabilists use the term indicator function for the function defined here almost exclusively, while mathematicians in other fields are more likely to use the term characteristic function to describe the function which indicates membership in a set.

[edit] Basic properties

The indicator or characteristic function of a subset A of some set X, maps elements of X to the range {0,1}.

This mapping is surjective only when A is a proper subset of X. If , then . By a similar argument, if then .

In the following, the dot represents multiplication, 1·1 = 1, 1·0 = 0 etc. "+" and "−" represent addition and subtraction. " " and " " is intersection and union, respectively.

If A and B are two subsets of X, then

and the indicator function of the complement of A i.e. AC is:

More generally, suppose is a collection of subsets of X. For any ,

is clearly a product of 0s and 1s. This product has the value 1 at precisely those which belong to none of the sets Ak and is 0 otherwise. That is

Page 86: Elementary Functions

Expanding the product on the left hand side,

where | F | is the cardinality of F. This is one form of the principle of inclusion-exclusion.

As suggested by the previous example, the indicator function is a useful notational device in combinatorics. The notation is used in other places as well, for instance in probability theory: if X is a probability space with probability measure and A is a measurable set, then becomes a random variable whose expected value is equal to the probability of A:

This identity is used in a simple proof of Markov's inequality.

In many cases, such as order theory, the inverse of the indicator function may be defined. This is commonly called the generalized Möbius function, as a generalization of the inverse of the indicator function in elementary number theory, the Möbius function. (See paragraph below about the use of the inverse in classical recursion theory.)

[edit] Mean, variance and covariance

Given a probability space with , the indicator random variable

is defined by if

otherwise

Mean:

Variance:

Covariance:

Page 87: Elementary Functions

[edit] Characteristic function in recursion theory, Gödel's and Kleene's representing function

Kurt Gödel described the representing function in his 1934 paper "On Undecidable Propositions of Formal Mathematical Systems". (The paper appears on pp. 41-74 in Martin Davis ed. The Undecidable):

"There shall correspond to each class or relation R a representing function φ(x1, . . ., xn) = 0 if R(x1, . . ., xn) and φ(x1, . . ., xn)=1 if ~R(x1, . . ., xn)." (p. 42; the "~" indicates logical inversion i.e. "NOT")

Stephen Kleene (1952) (p. 227) offers up the same definition in the context of the primitive recursive functions as a function φ of a predicate P, takes on values 0 if the predicate is true and 1 if the predicate is false.

For example, because the product of characteristic functions φ1*φ2* . . . *φn = 0 whenever any one of the functions equals 0, it plays the role of logical OR: IF φ1=0 OR φ2=0 OR . . . OR φn=0 THEN their product is 0. What appears to the modern reader as the representing function's logical-inversion, i.e. the representing function is 0 when the function R is "true" or satisfied", plays a useful role in Kleene's definition of the logical functions OR, AND, and IMPLY (p. 228), the bounded- (p. 228) and unbounded- (p. 279ff) mu operators (Kleene

Page 88: Elementary Functions

(1952)) and the CASE function (p. 229).

[edit] Characteristic function in fuzzy set theory

In classical mathematics, characteristic functions of sets only take values 1 (members) or 0 (non-members). In fuzzy set theory, characteristic functions are generalized to take value in the real unit interval [0, 1], or more generally, in some algebra or structure (usually required to be at least a poset or lattice). Such generalized characteristic functions are more usually called membership functions, and the corresponding "sets" are called fuzzy sets. Fuzzy sets model the gradual change in the membership degree seen in many real-world predicates like "tall", "warm", etc.

Simple functionFrom Wikipedia, the free encyclopediaJump to: navigation, search

In the mathematical field of real analysis, a simple function is a (sufficiently 'nice' - see below for the formal definition) real-valued function over a subset of the real line which attains only a finite number of values. Some authors also require simple functions to be measurable; as used in practice, they invariably are.

A basic example of a simple function is the floor function over the half-open interval [1,9), whose only values are {1,2,3,4,5,6,7,8}. A more advanced example is the Dirichlet function over the real line, which takes the value 1 if x is rational and 0 otherwise. (Thus the "simple" of "simple

Page 89: Elementary Functions

function" has a technical meaning somewhat at odds with common language.) Note also that all step functions are simple.

Simple functions are used as a first stage in the development of theories of integration, such as the Lebesgue integral, because it is very easy to create a definition of an integral for a simple function, and also, it is straightforward to approximate more general functions by sequences of simple functions.

Contents

[hide]

1 Definition 2 Properties of simple functions 3 Integration of simple functions 4 Relation to Lebesgue

integration 5 References

[edit] Definition

Formally, a simple function is a finite linear combination of indicator functions of measurable sets. More precisely, let (X, Σ) be a measurable space. Let A1, ..., An ∈ Σ be a sequence of measurable sets, and let a1, ..., an be a sequence of real or complex numbers. A simple function is a function of the form

where is the indicator function of the set A.

[edit] Properties of simple functions

By definition, the sum, difference, and product of two simple functions are again simple functions, and multiplication by constant keeps a simple function simple; hence it follows that the collection of all simple functions on a given measurable space forms a commutative algebra over .

[edit] Integration of simple functions

Page 90: Elementary Functions

If a measure μ is defined on the space (X,Σ), the integral of f with respect to μ is

if all summands are finite.

[edit] Relation to Lebesgue integration

Any non-negative measurable function is the pointwise limit of a monotonic increasing sequence of non-negative simple functions. Indeed, let f be a non-negative measurable function defined over the measure space (X,Σ,μ) as before. For each , subdivide the range of f into 22n + 1 intervals, 22n of which have length 2 − n. For each n, set

for , and .

(Note that, for fixed n, the sets In,k are disjoint and cover the non-negative real line.)

Now define the measurable sets

for .

Then the increasing sequence of simple functions

converges pointwise to f as . Note that, when f is bounded, the convergence is uniform. This approximation of f by simple functions (which are easily integrable) allows us to define an integral f itself; see the article on Lebesgue integration for more details.

Page 91: Elementary Functions

Dirac combFrom Wikipedia, the free encyclopedia(Redirected from Sampling function)Jump to: navigation, search

A Dirac comb is an infinite series of Dirac delta functions spaced at intervals of T

In mathematics, a Dirac comb (also known as an impulse train and sampling function in electrical engineering) is a periodic Schwartz distribution constructed from Dirac delta functions

for some given period T. Some authors, notably Bracewell as well as some textbook authors in electrical engineering and circuit theory, refer to it as the Shah function (possibly because its graph resembles the shape of the Cyrillic letter sha Ш). Because the Dirac comb function is periodic, it can be represented as a Fourier series:

Contents

[hide]

1 Scaling property 2 Fourier series 3 Fourier transform 4 Sampling and aliasing 5 Use in directional

statistics

Page 92: Elementary Functions

6 See also 7 References

[edit] Scaling property

The scaling property follows directly from the properties of the Dirac delta function

[edit] Fourier series

It is clear that ΔT(t) is periodic with period T. That is

for all t. The complex Fourier series for such a periodic function is

where the Fourier coefficients, cn are

All Fourier coefficients are 1/T resulting in

[edit] Fourier transform

Page 93: Elementary Functions

The Fourier transform of a Dirac comb is also a Dirac comb.

Unitary transform to ordinary frequency domain (Hz):

Unitary transform to angular frequency domain (radian/s):

[edit] Sampling and aliasing

Main article: Nyquist–Shannon sampling theorem

Reconstruction of a continuous signal from samples taken at sampling interval T is done by some sort of interpolation, such as the Whittaker–Shannon interpolation formula. Mathematically, that process is often modelled as the output of a lowpass filter whose input is a Dirac comb whose teeth have been weighted by the sample values. Such a comb is equivalent to the product of a comb and the original continuous signal. That mathematical abstraction is often described as "sampling" for purposes of introducing the subjects of aliasing and the Nyquist-Shannon sampling theorem.

[edit] Use in directional statistics

In directional statistics, the Dirac comb of period 2π is equivalent to a wrapped Dirac delta function, and is the analog of the Dirac delta function in linear statistics.

Page 94: Elementary Functions

In linear statistics, the random variable (x) is usually distributed over the real number line, or some subset thereof, and the probability density of x is a function whose domain is the set real numbers, and whose integral from

to is unity. In directional statistics, the random variable (θ) is distributed over the unit circle and the probability density of θ is a function whose domain is some interval of the real numbers of length 2π and whose integral over that interval is unity. Just as the integral of the product of a Dirac delta function with an arbitrary function over the real number line yields the value of that function at zero, so the integral of the product of a Dirac comb of period 2π with an arbitrary function of period 2π over the unit circle yields the value of that function at zero.

Frequency combFrom Wikipedia, the free encyclopediaJump to: navigation, search

A frequency comb is the graphic representation of the spectrum of a mode locked laser. An octave spanning comb can be used for mapping radio frequencies into the optical frequency range or it can be used to steer a piezoelectric mirror within a carrier envelope phase correcting feedback loop. (It should not be confused with mono-mode laser frequency stabilization as mode-locking requires multi-mode lasers.)

Page 95: Elementary Functions

An ultrashort pulse of light in the time domain. In this figure, the amplitude and intensity are Gaussian functions. Note how the author chooses to set the maximum of the function into the maximum of the envelope.

A Dirac comb is an infinite series of Dirac delta functions spaced at intervals of T.

Contents

[hide]

1 Frequency comb generation 2 Frequency comb widening to one

octave 3 Carrier-envelope offset measurement 4 Carrier-envelope offset control 5 Applications 6 History 7 References 8 See also 9 Further reading 10 External links

[edit] Frequency comb generation

Modelocked lasers produce a series of optical pulses separated in time by the round-trip time of the laser cavity. The spectrum of such a pulse train is a series of Dirac delta functions separated by the repetition rate (the inverse of the round trip time) of the laser. This series of sharp spectral lines is called a frequency comb or a frequency Dirac comb.

A purely electronic device, which generates a series of pulses, also generates a frequency comb. These are produced for electronic sampling oscilloscopes, but also used for frequency comparison of microwaves,

Page 96: Elementary Functions

because they reach up to 1 THz. Since they include 0 Hz they do not need the tricks which make up the rest of this article.

[edit] Frequency comb widening to one octave

This requires broadening of the laser spectrum so that it spans an octave. This is typically achieved using supercontinuum generation by strong self-phase modulation in nonlinear photonic crystal fiber. However, it has been shown that an octave-spanning spectrum can be generated directly from a Ti:sapphire laser using intracavity self-phase modulation. Or the second harmonic can be generated in a long crystal so that by consecutive sum frequency generation and difference frequency generation the spectrum of first and second harmonic widens until they overlap.

[edit] Carrier-envelope offset measurement

Each line is displaced from a harmonic of the repetition rate by the carrier-envelope offset frequency. The carrier-envelope offset frequency is the rate at which the peak of the carrier frequency slips from the peak of the pulse envelope on a pulse-to-pulse basis.

Measurement of the carrier-envelope offset frequency is usually done with a self-referencing technique, in which the phase of one part of the spectrum is compared to its harmonic.

In the 'frequency − 2 × frequency' technique, light at the lower energy side of the broadened spectrum is doubled using second harmonic generation in a nonlinear crystal and a heterodyne beat is generated between that and light at the same wavelength on the upper energy side of the spectrum. This beat frequency, detectable with a photodiode, is the carrier-envelope offset frequency.

Alternatively, from light at the higher energy side of the broadened spectrum the frequency at the peak of the spectrum is subtracted in a nonlinear crystal and a heterodyne beat is generated between that and light at the same wavelength on the lower energy side of the spectrum. This beat frequency, detectable with a photodiode, is the carrier-envelope offset frequency.

Because the phase is measured directly and not the frequency, it is possible to set the frequency to zero and additionally lock the phase, but because the intensity of the laser and this detector is not very stable, and because the whole spectrum beats in phase source, one has to lock the phase on a fraction of the repetition rate.

Page 97: Elementary Functions

[edit] Carrier-envelope offset control

In the absence of active stabilization, the repetition rate and carrier-envelope offset frequency would be free to drift. They vary with changes in the cavity length, refractive index of laser optics, and nonlinear effects such as the Kerr effect. The repetition rate can be stabilized using a piezoelectric transducer, which moves a mirror to change the cavity length.

In Ti:sapphire lasers using prisms for dispersion control, the carrier-envelope offset frequency can be controlled by tilting the high reflector mirror at the end of the prism pair. This can be done using piezoelectric transducers.

In high repetition rate Ti:sapphire ring lasers, which often use double-chirped mirrors to control dispersion, modulation of the pump power using an acousto-optic modulator is often used to control the offset frequency. The phase slip depends strongly on the Kerr effect, and by changing the pump power one changes the peak intensity of the laser pulse and thus the size of the Kerr phase shift. This shift is far smaller than 6 rad, so an additional device for coarse adjustment is needed. See also: phase-locked loop

The breakthrough which led to a practical frequency comb was the development of technology for stabilizing the carrier-envelope offset frequency.

[edit] Applications

A frequency comb allows a direct link from radio frequency standards to optical frequencies. Current frequency standards such as atomic clocks operate in the microwave region of the spectrum, and the frequency comb brings the accuracy of such clocks into the optical part of the electromagnetic spectrum. A simple electronic feedback loop can lock the repetition rate to a frequency standard.

There are two distinct applications of this technique. One is the optical clock where an optical frequency is overlapped with a single tooth of the comb on a photodiode and a radio frequency is compared to the beat signal, the repetition rate, and the CEO-frequency. Applications for the frequency comb technique include optical metrology, frequency chain generation, optical atomic clocks, high precision spectroscopy, and more precise GPS technology.[1].

The other is doing experiments with few cycle pulses, like above threshold ionization, attosecond pulses, highly efficient nonlinear optics or high harmonics generation. This can be single pulses so that no comb exists

Page 98: Elementary Functions

and therefore it is not possible to define a carrier envelope offset frequency, rather the carrier envelope offset phase is important. A second photodiode can be added to the setup to gather phase and amplitude in a single shot, or difference frequency generation can be used to even lock the offset on a single shot basis albeit with low power efficiency.

Without an actual comb one can look at the phase vs frequency. Without a carrier envelope offset all frequencies are cosines. That means all frequencies have the phase zero. The time origin is arbitrary. If a pulse comes at later times, the phase increases linearly with frequency, but still the zero frequency phase is zero. This phase at zero frequency is the carrier envelope offset. The second harmonic not only has twice the frequency but also twice the phase. That means for a pulse with zero offset the second harmonic of the low frequency tail is in phase with the fundamental of the high frequency tail and otherwise it is not. Spectral phase interferometry for direct electric-field reconstruction (SPIDER) measures how the phase increases with frequency, but it cannot determine the offset, so the name “electric field reconstruction” is a bit misleading.

[edit] History

Theodor W. Hänsch and John L. Hall shared half of the 2005 Nobel Prize in Physics for contributions to the development of laser-based precision spectroscopy, including the optical frequency comb technique. The other half of the prize was awarded to Roy Glauber.

The femtosecond comb technique has, in 2006, been extended to the extreme ultraviolet range, which enables frequency metrology to that region of the spectrum.[2]

Green's functionFrom Wikipedia, the free encyclopedia

Jump to: navigation, search

This article is about the classical approach to Green's functions. For a modern discussion, see fundamental solution.

Page 99: Elementary Functions

In mathematics, a Green's function is a type of function used to solve inhomogeneous differential equations subject to specific initial conditions or boundary conditions. Under many-body theory, the term is also used in physics, specifically in quantum field theory, electrodynamics and statistical field theory, to refer to various types of correlation functions, even those that do not fit the mathematical definition.

Green's functions are named after the British mathematician George Green, who first developed the concept in the 1830s. In the modern study of linear partial differential equations, Green's functions are studied largely from the point of view of fundamental solutions instead.

Contents

[hide]

1 Definition and uses 2 Motivation 3 Green's functions for solving inhomogeneous boundary value

problemso 3.1 Framework o 3.2 Theorem

4 Finding Green's functions o 4.1 Eigenvalue expansions

5 Green's functions for the Laplacian 6 Example 7 Further examples 8 See also 9 References 10 External links

[edit] Definition and uses

A Green's function, G(x, s), of a linear differential operator L = L(x) acting on distributions over a subset of the Euclidean space Rn, at a point s, is any solution of

LG(x,s) = δ(x − s)

(1)

where δ is the Dirac delta function. This property of a Green's function can be exploited to solve differential equations of the form

Lu(x) = f(x)

(2)

Page 100: Elementary Functions

If the kernel of L is non-trivial, then the Green's function is not unique. However, in practice, some combination of symmetry, boundary conditions and/or other externally imposed criteria will give a unique Green's function. Also, Green's functions in general are distributions, not necessarily proper functions.

Green's functions are also a useful tool in solving wave equations, diffusion equations, and in quantum mechanics, where the Green's function of the Hamiltonian is a key concept, with important links to the concept of density of states. As a side note, the Green's function as used in physics is usually defined with the opposite sign; that is,

LG(x,s) = − δ(x − s).

This definition does not significantly change any of the properties of the Green's function.

If the operator is translation invariant, that is when L has constant coefficients with respect to x, then the Green's function can be taken to be a convolution operator, that is,

G(x,s) = G(x − s).

In this case, the Green's function is the same as the impulse response of linear time-invariant system theory.

[edit] Motivation

See also: Spectral theory

Loosely speaking, if such a function G can be found for the operator L, then if we multiply the equation (1) for the Green's function by f(s), and then perform an integration in the s variable, we obtain;

The right hand side is now given by the equation (2) to be equal to L u(x), thus:

Page 101: Elementary Functions

Because the operator L = L(x) is linear and acts on the variable x alone (not on the variable of integration s), we can take the operator L outside of the integration on the right hand side, obtaining;

And this suggests;

(3)

Thus, we can obtain the function u(x) through knowledge of the Green's function in equation (1), and the source term on the right hand side in equation (2). This process relies upon the linearity of the operator L.

In other words, the solution of equation (2), u(x), can be determined by the integration given in equation (3). Although f(x) is known, this integration cannot be performed unless G is also known. The problem now lies in finding the Green's function G that satisfies equation (1). For this reason, the Green's function is also sometimes called the fundamental solution associated to the operator L.

Not every operator L admits a Green's function. A Green's function can also be thought of as a right inverse of L. Aside from the difficulties of finding a Green's function for a particular operator, the integral in equation (3), may be quite difficult to evaluate. However the method gives a theoretically exact result.

This can be thought of as an expansion of f according to a Dirac delta function basis (projecting f over δ(x − s)) and a superposition of the solution on each projection. Such an integral equation is known as a Fredholm integral equation, the study of which constitutes Fredholm theory.

Page 102: Elementary Functions

[edit] Green's functions for solving inhomogeneous boundary value problems

The primary use of Green's functions in mathematics is to solve non-homogeneous boundary value problems. In modern theoretical physics, Green's functions are also usually used as propagators in Feynman diagrams (and the phrase Green's function is often used for any correlation function).

[edit] Framework

Let L be the Sturm–Liouville operator, a linear differential operator of the form

and let D be the boundary conditions operator

Let f(x) be a continuous function in [0,l]. We shall also suppose that the problem

is regular (i.e., only the trivial solution exists for the homogeneous problem).

[edit] Theorem

There is one and only one solution u(x) which satisfies

and it is given by

Page 103: Elementary Functions

where G(x,s) is a Green's function satisfying the following conditions:

1. G(x,s) is continuous in x and x, s)=0</math>

2. For , DG(x,s) = 03. Derivative "jump": G'(s + 0,s) −

G'(s − 0,s) = 1 / p(s)4. Symmetry: G(x, s) = G(s, x)

[edit] Finding Green's functions

[edit] Eigenvalue expansions

If a differential operator L admits a set of eigenvectors Ψn(x) (i.e., a set of functions Ψn and scalars λn such that LΨn = λnΨn)) that is complete, then it is possible to construct a Green's function from these eigenvectors and eigenvalues.

Complete means that the set of

functions satisfies the following completeness relation:

Then the following holds:

where represents complex conjugation.

Page 104: Elementary Functions

Applying the operator L to each side of this equation results in the completeness relation, which was assumed true.

The general study of the Green's function written in the above form, and its relationship to the function spaces formed by the eigenvectors, is known as Fredholm theory.

[edit] Green's functions for the Laplacian

Green's functions for linear differential operators involving the Laplacian may be readily put to use using the second of Green's identities.

To derive Green's theorem, begin with the divergence theorem (otherwise known as Gauss's theorem):

Let and substitute into Gauss' law. Compute

and apply the chain rule for the operator:

Page 105: Elementary Functions

Plugging this into the divergence theorem produces Green's theorem:

Suppose that the linear differential operator L is the Laplacian, , and that there is a Green's function G for the Laplacian. The defining property of the Green's function still holds:

Let ψ = G in Green's theorem. Then:

Using this expression, it is possible to solve Laplace's equation

or Poisson's equation

Page 106: Elementary Functions

, subject to either Neumann or Dirichlet boundary conditions. In other words, we can solve for ϕ(x) everywhere inside a volume where either (1) the value of ϕ(x) is specified on the bounding surface of the volume (Dirichlet boundary conditions), or (2) the normal derivative of ϕ(x) is specified on the bounding surface (Neumann boundary conditions).

Suppose the problem is to solve for ϕ(x) inside the region. Then the integral

Page 107: Elementary Functions

reduces to simply ϕ(x) due to the defining property of the Dirac delta function and we have:

This form expresses the well-known property of harmonic functions that if the value or nor

Page 108: Elementary Functions

mal derivative is known on a bounding surface, then the value of the function inside the volume is known everywhere.

In electrostatics, ϕ(x) is interpreted as the electric potential,

Page 109: Elementary Functions

ρ(x) as electric charge density, and the normal derivative

as the normal component of the electric field.

If the problem is to solve a Dirichlet boundary value problem, the Gre

Page 110: Elementary Functions

en's function should be chosen such that G(x,x') vanishes when either x or x' is on the bounding surface.Thus only one of the two terms in the surface integral remains. If the problem is to solve a

Page 111: Elementary Functions

Neumann boundary value problem, the Green's function is chosen such that its normal derivative vanishes on the bounding surface, as it would seems to be the most logical choi

Page 112: Elementary Functions

ce. (See Jackson J.D. classical electrodynamics, page 39). However, application of Gauss's theorem to the differential equation defining the Green's function yields

Page 113: Elementary Functions

meaning the normal derivative of G(x,x') cannot vanish on th

Page 114: Elementary Functions

e surface, because it must integrate to 1 on the surface. (Aga

Page 115: Elementary Functions

in, see Jackson J.D. classical electrodynamics, page 39 for

Page 116: Elementary Functions

this and the following argument). The simplest form the norm

Page 117: Elementary Functions

al derivative can take is that of a constant, namely 1 / S, where

Page 118: Elementary Functions

S is the surface area of the surface. The surface term in the so

Page 119: Elementary Functions

lution becomes

where

is the average value of the potential

Page 120: Elementary Functions

on the surface. This number is not known in general, but is often unim

Page 121: Elementary Functions

portant, as the goal is often to obtain the electric field given by the gr

Page 122: Elementary Functions

adient of the potential, rather than the potential itself.

With no b

Page 125: Elementary Functions

bounding surface goes out to infinity, and plugging in this expression for the Green's function, this gives the familiar expression for electric potential in terms of electric charge density (in the CGS unit syst

Page 126: Elementary Functions

em) as

[edit] Example

Given the problem

Find the Green's function.

First step: The Green's function for the linear operator at hand is defined as the solution to

If , then the delta function gives zero, and the general solution is

g(x,s) = c1cos x + c2sin x.

Page 127: Elementary Functions

For x < s, the boundary condition at implies

The equation of

skipped because

if

For x > s, the boundary condition at x = π / 2 implies

The equation of

is skipped for similar reasons.

To summarize the results thus far:

Second step:task is to determine c3.

Ensuring continuity in the Green's function at implies

One can also ensure proper discontinuity in the first

Page 128: Elementary Functions

derivative by integrating the defining differential equation from and taking the limit as to zero:

The two (dis)continuity equations can be solved for obtain

So the Green's function for this problem is:

[edit] Further examples

Let n = 1 and let the subset be all of be d/dx. Then, the H(x − x0) is a Green's function of L at

Let n = 2 and let the subset be the quarter-plane { (x, Laplacian. Also, assume a boundary conditiona Neumann boundary conditionat y = 0. Then the Green's function is

Page 129: Elementary Functions

Sinc functionFrom Wikipedia, the free encyclopedia(Redirected from Sinc)Jump to: navigation, search"Sinc" redirects here. For the designation used in the United Kingdom for areas of wildlife interest, see Site of Importance for Nature Conservation.

In mathematics, the sinc function, denoted by sinc(x) has two nearly equivalent definitions[1]. In digital signal processing and information theory, the normalized sinc function is commonly defined by

The normalized sinc (blue) and unnormalized sinc function (red) shown on the same scale.

Page 130: Elementary Functions

It is qualified as normalized because its integral over all x is 1. All of the zeros of the normalized sinc function are integer values of x. The Fourier transform of the normalized sinc function is the rectangular function with no scaling. This function is fundamental in the concept of reconstructing the original continuous bandlimited signal from uniformly spaced samples of that signal.

In mathematics, the historical unnormalized sinc function is defined by

The only difference between the two definitions is in the scaling of the independent variable (the x-axis) by a factor of π. In both cases, the value of the function at the removable singularity at zero is understood to be the limit value 1. The sinc function is analytic everywhere.

The term "sinc" (English pronunciation: /ˈsɪŋk/) is a contraction of the function's full Latin name, the sinus cardinalis (cardinal sine), first introduced by Phillip M. Woodward in 1953.[2][3][4]

Contents

[hide]

1 Properties 2 Relationship to the Dirac delta

distribution 3 See also 4 References 5 External links

[edit] Properties

Page 131: Elementary Functions

The local maxima and minima (small white dots) of the unnormalized, red sincfunction correspond to its intersections with the blue cosine function.

The zero crossings of the unnormalized sinc are at nonzero multiples of π, while zero crossings of the normalized sinc occur at nonzero integer values.

The local maxima and minima of the unnormalized sinc correspond to its intersections with the cosine function. That is, sin(ξ)/ξ = cos(ξ) for all points ξ where the derivative of sin(x)/x is zero and thus a local extremum is reached.

A good approximation of the x-coordinate of the n-th extremum with positive x-coordinate is

where odd n lead to a local minimum and even n to a local maximum. Besides the extrema at xn, the curve has an absolute maximum at ξ0 = (0,1) and because of its symmetry to the y-axis extrema with x-coordinates −xn.

The normalized sinc function has a simple representation as the infinite product

and is related to the gamma function Γ(x) by Euler's reflection formula:

Page 132: Elementary Functions

Euler discovered that

The continuous Fourier transform of the normalized sinc (to ordinary frequency) is rect(f),

where the rectangular function is 1 for argument between −1/2 and 1/2, and zero otherwise. This corresponds to the fact that the sinc filter is the ideal (brick-wall, meaning rectangular frequency response) low-pass filter. This Fourier integral, including the special case

is an improper integral and not a convergent Lebesgue integral, as

The normalized sinc function has properties that make it ideal in relationship to interpolation of sampled bandlimited functions:

It is an interpolating function, i.e., sinc(0) = 1, and sinc(k) = 0 for nonzero integer k.

The functions xk(t) = sinc(t−k) (k integer) form an orthonormal basis for bandlimited functions in the function space L2(R), with highest angular frequency ωH = π (that is, highest cycle frequency ƒH = 1/2).

Other properties of the two sinc functions include:

Page 133: Elementary Functions

The unnormalized sinc is the zeroth order spherical Bessel function of the first kind,

. The normalized sinc is

where Si(x) is the sine integral.

λ sinc(λ x) (not normalized) is one of two linearly independent solutions to the linear ordinary differential equation

The other is cos(λ x)/x, which is not bounded at x = 0, unlike its sinc function counterpart.

where the normalized sinc is meant.

[edit] Relationship to the Dirac delta distribution

The normalized sinc function can be used as a nascent delta function, meaning that the following weak limit holds:

This is not an ordinary limit, since the left side does not

Page 134: Elementary Functions

converge. Rather, it means that

for any smooth function with compact support.

In the above expression, as a approaches zero, the number of oscillations per unit length of the sinc function approaches infinity. Nevertheless, the expression always oscillates inside an envelope of ±1/(π a x), and approaches zero for any nonzero value of x. This complicates the informal picture of δ(x) as being zero for all x except at the point x = 0 and illustrates the problem of thinking of the delta function as a function rather than as a distribution. A similar situation is found in the Gibbs phenomenon.

Trigonometric integralFrom Wikipedia, the free encyclopedia

(Redirected from Sine integral)

Jump to: navigation, search

Page 135: Elementary Functions

Si(x) (blue) and Ci(x) (green) plotted on the same plot.

In mathematics, the trigonometric integrals are a family of integrals which involve trigonometric functions. A number of the basic trigonometric integrals are discussed at the list of integrals of trigonometric functions.

Contents

[hide]

1 Sine integral 2 Cosine integral 3 Hyperbolic sine integral 4 Hyperbolic cosine integral 5 Nielsen's spiral 6 Expansion

o 6.1 Asymptotic series (for large argument) o 6.2 Convergent series

7 Relation with the exponential integral of imaginary argument

8 See also o 8.1 Signal processing

9 References

[edit] Sine integral

Page 136: Elementary Functions

Plot of Si(x) for 0 ≤ x ≤ 8π.

The different sine integral definitions are:

Si(x) is the primitive of sinx / x which is zero for x = 0; si(x) is the primitive of sinx / x which is zero for .

Note that is the sinc function and also the zeroth spherical Bessel function.

When , this is known as the Dirichlet integral.

In signal processing, the oscillations of the Sine integral cause overshoot and ringing artifacts when using the sinc filter, and frequency domain ringing if using a truncated sinc filter as a low-pass filter.

The Gibbs phenomenon is a related phenomenon: thinking of sinc as a low-pass filter, it corresponds to truncating the Fourier series, which causes the Gibbs phenomenon.

[edit] Cosine integral

Plot of Ci(x) for 0 < x ≤ 8π.

The different cosine integral definitions are:

Page 138: Elementary Functions

Nielsen's spiral.

The spiral formed by parametric plot of si,ci is known as Nielsen's spiral. It is also referred to as the Euler spiral, the Cornu spiral, a clothoid, or as a linear-curvature polynomial spiral. The spiral is also closely related to the Fresnel integrals. This spiral has applications in vision processing, road and track construction and other areas.

[edit] Expansion

Various expansions can be used for evaluation of Trigonometric integrals, depending on the range of the argument.

[edit] Asymptotic series (for large argument)

These series are divergent[clarification needed], although can be used for estimates and

even precise evaluation at .

[edit] Convergent series

These series are convergent at any complex , although for

the series will converge slowly initially, requiring many terms for high precision.

Page 139: Elementary Functions

[edit] Relation with the exponential integral of imaginary argument

Function

is called exponential integral. It is closely related with Si and Ci:

As each involved function is analytic except the cut at negative values of the argument, the area of validity of the relation should be extended to Re(x) > 0. (Out of this range, additional terms which are integer factors of π appear in the expression).

Cases of imaginary argument of the generalized integro-exponential function are

which is the real part of

Similarly

Borwein integral

Page 140: Elementary Functions

From Wikipedia, the free encyclopediaJump to: navigation, search

This article may need to be wikified to meet Wikipedia's quality standards. Please help by adding relevant internal links , or by improving the article's layout. (July 2011)Click [show] on right for more details.[show]

In mathematics, a Borwein integral is an integral studied by Borwein & Borwein (2001) involving products of sinc(ax), where the sinc function is given by sinc(x) = sin(x)/x. These integrals are notorious for exhibiting apparent patterns that eventually break down. An example they give is

This pattern continues up to

However at the next step the obvious pattern fails:

In general similar integrals have value π/2 whenever the numbers 3, 5, ... are replaced by positive real numbers such that the sum of their reciprocals is less than 1. In the example above, 1/3 + 1/5 + ... + 1/13 < 1, but 1/3 + 1/5 + ... + 1/15 > 1.

Dirichlet integralFrom Wikipedia, the free encyclopedia

Page 141: Elementary Functions

Jump to: navigation, search

This article needs additional citations for verification. Please help improve this article by adding citations to reliable sources. Unsourced material may be challenged and removed. (June 2011)

In mathematics, there are several integrals known as the Dirichlet integral, after the German mathematician Peter Gustav Lejeune Dirichlet.

One of those is

This can be derived from attempts to evaluate a double improper integral two different ways. It can also be derived using differentiation under the integral sign.

Contents

[hide]

1 Evaluation o 1.1 Double Improper Integral Method o 1.2 Differentiation under the integral sign o 1.3 Complex integration

2 Notes 3 See also 4 External links

Evaluation

Double Improper Integral Method

Pre-knowledge of properties of Laplace transforms allows us to evaluate this Dirichlet integral succinctly in the following manner:

This is equivalent to attempting to evaluate the same double definite integral in two different ways, by reversal of the order of integration, viz.,

Page 142: Elementary Functions

Differentiation under the integral sign

First rewrite the integral as a function of variable . Let

then we need to find .

Differentiate with respect to and apply the Leibniz Integral Rule to obtain:

This integral was evaluated without proof, above, based on Laplace trasform tables; we derive it this time. It is made much simpler by recalling Euler's formula,

then,

where represents the imaginary part.

Integrating with respect to :

where is a constant to be determined. As,

Page 143: Elementary Functions

for some integers m & n. It is easy to show that has to be zero, by analyzing easily observed bounds for this integral:

End of proof.

Extending this result further, with the introduction of another

variable, first noting that is an even function and therefore

then:

Complex integration

The same result can be obtained via complex integration. Let's consider

As a function of the complex variable z, it has a simple pole at the origin, which prevents the application of Jordan's lemma, whose other hypotheses are satisfied. We shall then define a new function[1] g(z) as follows

Page 144: Elementary Functions

The pole has been moved away from the real axis, so g(z) can be integrated along the semicircle of radius R centered at z=0 and closed on the real axis, then the limit should be taken.

The complex integral is zero by the residue theorem, as there are no poles inside the integration path

The second therm vanishes as R goes to infinity; for arbitrarily small ε, the Sokhatsky–Weierstrass theorem applied to the first one yelds

Where P.V. indicates Cauchy Principal Value. By taking the imaginary part on both

Page 145: Elementary Functions

sides and noting that sinc(x) is even and by definition sinc(0) = 1, we get the desired result

Exponential integralFrom Wikipedia, the free encyclopedia

Jump to: navigation, search

Plot of E1 function (top) and Ei function (bottom).

In mathematics, the exponential integral is a special function defined on the complex plane given the symbol Ei.

Contents

[hide]

1 Definitions 2 Properties

o 2.1 Convergent series

Page 146: Elementary Functions

o 2.2 Asymptotic (divergent) series o 2.3 Exponential and logarithmic behavior:

bracketingo 2.4 Definition by Ein o 2.5 Relation with other functions o 2.6 Derivatives o 2.7 Exponential integral of imaginary argument

3 Applications 4 Notes 5 References 6 External links

[edit] Definitions

For real, nonzero values of x, the exponential integral Ei(x) can be defined as

The function is given as a special function because is not an elementary function, a fact which can be proven using the Risch Algorithm. The definition above can be used for positive values of x, but the integral has to be understood in terms of the Cauchy principal value, due to the singularity in the integrand at zero. For complex values of the argument, the definition becomes ambiguous due to branch points at 0 and .[1] In general, a branch cut is taken on the negative real axis and Ei can be defined by analytic continuation elsewhere on the complex plane.

The following notation is used,[2]

For positive values of the real part of z, this can be written[3]

The behaviour of E1 near the branch cut can be seen by the following relation[4]:

Page 147: Elementary Functions

[edit] Properties

Several properties of the exponential integral below, in certain cases, allow to avoid its explicit evaluation through the definition above.

[edit] Convergent series

Integrating the Taylor series for e − t / t, and extracting the logarithmic singularity, we can derive the following series representation for E1(x) for real x[5]:

For complex arguments off the negative real axis, this generalises to[6]

where γ is the Euler–Mascheroni constant. The sum converges for all complex z, and we take the usual value of the complex logarithm having a branch cut along the negative real axis.

[edit] Asymptotic (divergent) series

Relative error of the asymptotic approximation for different

number of term in the truncated sum

Page 148: Elementary Functions

Unfortunately, the convergence of the series above is slow for arguments of larger modulus. For example, for x=10 more than 40 terms are required to get an answer correct to three significant figures.[7] However, there is a divergent series approximation that can be obtained by integrating zezE1(z) by parts[8]:

which has error of order O(N!z − N) and is valid for large values of Re(z). The relative error of the approximation above is plotted on the figure to the right for various values of N (N = 1 in red, N = 5 in pink). When x > 40, the approximation above with N = 40 is correct to within 64 bit double precision.

[edit] Exponential and logarithmic behavior: bracketing

Bracketing of E1 by elementary functions

From the two series suggested in previous subsections, it follows that E1 behaves like a negative exponential for large values of the argument and like a logarithm for small values. For positive real values of the argument, E1 can be bracketed by elementary functions as follows[9]:

Page 149: Elementary Functions

The left-hand side of this inequality is shown in the graph to the left in blue; the central part E1(x) is shown in black and the right-hand side is shown in red.

[edit] Definition by Ein

Both Ei and E1 can be written more simply using the entire function Ein[10] defined as

(note that this is just the alternating series in the above definition of E1). Then we have

[edit] Relation with other functions

The exponential integral is closely related to the logarithmic integral function li(x) by the formula

for positive real values of x

The exponential integral may also be generalized to

which can be written as a special case of the incomplete gamma function [11] :

The generalized form is sometimes called the Misra function[12] φm(x), defined as

Page 150: Elementary Functions

Including a logarithm defines the generalized integro-exponential function[13]

.

[edit] Derivatives

The derivatives of the generalised functions En can be calculated by means of the formula [14]

Note that the function E0 is easy to evaluate (making this recursion useful), since it is just e − z / z.[15]

[edit] Exponential integral of imaginary argument

E1(ix) against x; real part black, imaginary part red.

Page 151: Elementary Functions

If z is imaginary, it has a nonnegative real part, so we can use the formula

to get a relation with the trigonometric integrals Si and Ci:

The real and imaginary parts of E1(x) are plotted in the figure to the right with black and red curves.

[edit] Applications

Time-dependent heat transfer

Nonequilibrium groundwater flow in the Theis solution (called a well function)

Radiative transfer in stellar atmospheres

Radial Diffusivity Equation for transient or

Page 152: Elementary Functions

unsteady state flow with line sources and sinks

Logarithmic integral functionFrom Wikipedia, the free encyclopedia(Redirected from Logarithmic integral)Jump to: navigation, search

In mathematics, the logarithmic integral function or integral logarithm li(x) is a special function. It occurs in problems of physics and has number theoretic significance, occurring in the prime number theorem as an estimate of the number of prime numbers less than a given value.

Logarithmic integral function plot

Contents

[hide]

1 Integral representation 2 Offset logarithmic integral 3 Series representation 4 Special values 5 Asymptotic expansion 6 Infinite logarithmic integral 7 Number theoretic

significance 8 See also 9 References

[edit] Integral representation

The logarithmic integral has an integral representation defined for all

positive real numbers by the definite integral:

Page 153: Elementary Functions

Here, ln denotes the natural logarithm. The function 1 / ln(t) has a singularity at t = 1, and the integral for x > 1 has to be interpreted as a Cauchy principal value:

[edit] Offset logarithmic integral

The offset logarithmic integral or Eulerian logarithmic integral is defined as

or

As such, the integral representation has the advantage of avoiding the singularity in the domain of integration.

This function is a very good approximation to the number of prime numbers less than x.

[edit] Series representation

The function li(x) is related to the exponential integral Ei(x) via the equation

which is valid for x > 1. This identity provides a series representation of li(x) as

where γ ≈ 0.57721 56649 01532 ... is the Euler–Mascheroni gamma constant. A more rapidly convergent series due to Ramanujan [1] is

Page 154: Elementary Functions

[edit] Special values

The function li(x) has a single positive zero; it occurs at x ≈ 1.45136 92348 ...; this number is known as the Ramanujan–Soldner constant.

li(2) ≈ 1.045163 780117 492784 844588 889194 613136 522615 578151…

This is where is the incomplete gamma function. It must be understood as the Cauchy principal value of the function.

[edit] Asymptotic expansion

The asymptotic behavior for x → ∞ is

where O is the big O notation. The full asymptotic expansion is

or

Note that, as an asymptotic expansion, this series is not convergent: it is a reasonable approximation only if the series is truncated at a finite number of terms, and only large values of x are employed. This expansion follows directly from the asymptotic expansion for the exponential integral.

Page 155: Elementary Functions

[edit] Infinite logarithmic integral

[clarification needed]:

and discussed in Paul Koosis, The Logarithmic Integral, volumes I and II, Cambridge University Press, second edition, 1998.

[edit] Number theoretic significance

The logarithmic integral is important in number theory, appearing in estimates of the number of prime numbers less than a given value. For example, the prime number theorem states that:

π(x)~li(x)

where π(x) denotes the number of primes smaller than or equal to x.

Gibbs phenomenonFrom Wikipedia, the free encyclopedia

Jump to: navigation, search

In mathematics, the Gibbs phenomenon, named after the American physicist J. Willard Gibbs, is the peculiar manner in which the Fourier series of a piecewise continuously differentiable periodic function behaves at a jump discontinuity: the nth partial sum of the Fourier series has large oscillations near the jump, which might increase the maximum of the partial sum above that of the function itself. The overshoot does not die out as the frequency increases, but approaches a finite limit.[1]

These are one cause of ringing artifacts in signal processing.

Contents

Page 156: Elementary Functions

[hide]

1 Description o 1.1 History o 1.2 Explanation o 1.3 Solutions

2 Formal mathematical description of the phenomenon

3 Signal processing explanation 4 The square wave example 5 Consequences 6 See also 7 Notes 8 References 9 External links and references

[edit] Description

Functional approximation of square wave using 5 harmonics

Functional approximation of square wave using 25 harmonics

Functional approximation of square wave using 125 harmonics

Page 157: Elementary Functions

The Gibbs phenomenon involves both the fact that Fourier sums overshoot at a jump discontinuity, and that this overshoot does not die out as the frequency increases.

The three pictures on the right demonstrate the phenomenon for a square wave whose Fourier expansion is

More precisely, this is the function f which equals π / 4 between 2nπ and (2n + 1)π and − π / 4 between (2n + 1)π and (2n + 2)π for every integer n; thus this square wave has a jump discontinuity of height π / 2 at every integer multiple of π.

As can be seen, as the number of terms rises, the error of the approximation is reduced in width and energy, but converges to a fixed height. A calculation for the square wave (see Zygmund, chap. 8.5., or the computations at the end of this article) gives an explicit formula for the limit of the height of the error. It turns out that the Fourier series exceeds the height π / 4 of the square wave by

or about 9 percent. More generally, at any jump point of a piecewise continuously differentiable function with a jump of a, the nth partial Fourier series will (for n very large) overshoot this jump

by approximately at one end and undershoot it by the same amount at the other end; thus the "jump" in the partial Fourier series will be about 18% larger than the jump in the original function. At the location of the discontinuity itself, the partial Fourier series will converge to the midpoint of the jump (regardless of what the actual value of the original function is at this point). The quantity

is sometimes known as the Wilbraham–Gibbs constant.

[edit] History

The Gibbs phenomenon was first noticed and analyzed by the obscure Henry Wilbraham.[2] He published a paper on it in 1848 that went unnoticed by the mathematical world. It was not until

Page 158: Elementary Functions

Albert Michelson observed the phenomenon via a mechanical graphing machine that interest arose. Michelson developed a device in 1898 that could compute and re-synthesize the Fourier series. When the Fourier coefficients for a square wave were input to the machine, the graph would oscillate at the discontinuities. This would continue to occur even as the number of Fourier coefficients increased.

Because it was a physical device subject to manufacturing flaws, Michelson was convinced that the overshoot was caused by errors in the machine. In 1898 J. Willard Gibbs published a paper on Fourier series in which he discussed the example of what today would be called a sawtooth wave, and described the graph obtained as a limit of the graphs of the partial sums of the Fourier series. Interestingly in this paper he failed to notice the phenomenon that bears his name, and the limit he described was incorrect. In 1899 he published a correction to his paper in which he describes the phenomenon and points out the important distinction between the limit of the graphs and the graph of the function that is the limit of the partial sums of the Fourier series.[citation needed] Maxime Bôcher gave a detailed mathematical analysis of the phenomenon in 1906 and named it the Gibbs phenomenon.

[edit] Explanation

Informally, it reflects the difficulty inherent in approximating a discontinuous function by a finite series of continuous sine and cosine waves. It is important to put emphasis on the word finite because even though every partial sum of the Fourier series overshoots the function it is approximating, the limit of the partial sums does not. The value of x where the maximum overshoot is achieved moves closer and closer to the discontinuity as the number of terms summed increases so, again informally, once the overshoot has passed by a particular x, convergence at the value of x is possible.

There is no contradiction in the overshoot converging to a non-zero amount, but the limit of the partial sums having no overshoot, because where that overshoot happens moves. We have pointwise convergence, but not uniform convergence. For a piecewise C1 function the Fourier series converges to the function at every point except at the jump discontinuities. At the jump discontinuities themselves the limit will converge to the average of the values of the function on either side of the jump. This is a consequence of the Dirichlet theorem.[3]

Page 159: Elementary Functions

The Gibbs phenomenon is also closely related to the principle that the decay of the Fourier coefficients of a function at infinity is controlled by the smoothness of that function; very smooth functions will have very rapidly decaying Fourier coefficients (resulting in the rapid convergence of the Fourier series), whereas discontinuous functions will have very slowly decaying Fourier coefficients (causing the Fourier series to converge very slowly). Note for instance that the Fourier coefficients 1, −1/3, 1/5, ... of the discontinuous square wave described above decay only as fast as the harmonic series, which is not absolutely convergent; indeed, the above Fourier series turns out to be only conditionally convergent for almost every value of x. This provides a partial explanation of the Gibbs phenomenon, since Fourier series with absolutely convergent Fourier coefficients would be uniformly convergent by the Weierstrass M-test and would thus be unable to exhibit the above oscillatory behavior. By the same token, it is impossible for a discontinuous function to have absolutely convergent Fourier coefficients, since the function would thus be the uniform limit of continuous functions and therefore be continuous, a contradiction. See more about absolute convergence of Fourier series.

[edit] Solutions

In practice, the difficulties associated with the Gibbs phenomenon can be ameliorated by using a smoother method of Fourier series summation, such as Fejér summation or Riesz summation, or by using sigma-approximation. Using a wavelet transform with Haar basis functions, the Gibbs phenomenon does not occur.

[edit] Formal mathematical description of the phenomenon

Let be a piecewise continuously differentiable function which is periodic with some period L > 0. Suppose that at some

point x0, the left limit and right limit of the function f differ by a non-zero gap a:

For each positive integer N ≥ 1, let SN f be the Nth partial Fourier series

Page 160: Elementary Functions

where the Fourier coefficients are given by the usual formulae

Then we have

and

but

More generally, if xN is any sequence of real numbers which converges to x0 as

, and if the gap a is positive then

and

If instead the gap a is negative, one needs to interchange limit

Page 161: Elementary Functions

superior with limit inferior, and also interchange the ≤ and ≥ signs, in the above two inequalities.

[edit] Signal processing explanation

For more details on this topic, see Ringing artifacts.

The sinc function, the impulse response of an ideal low-pass filter. Scaling narrows the function, and correspondingly increases magnitude (which is not shown here), but does not reduce the magnitude of the undershoot, which is the integral of the tail.

From the point of view of signal processing, the Gibbs phenomenon is the step response of a low-pass filter, and the oscillations are called ringing or ringing artifacts. Truncating the Fourier transform of a signal on the real line, or the Fourier series of a periodic signal (equivalently, a signal on the circle) corresponds to filtering out the higher frequencies by an ideal (brick-wall) low-pass/high-cut filter. This can be represented as convolution of

Page 162: Elementary Functions

the original signal with the impulse response of the filter (also known as the kernel), which is the sinc function. Thus the Gibbs phenomenon can be seen as the result of convolving a Heaviside step function (if periodicity is not required) or a square wave (if periodic) with a sinc function: the oscillations in the sinc function cause the ripples in the output.

The sine integral, exhibiting the Gibbs phenomenon for a step function on the real line.

In the case of convolving with a Heaviside step function, the resulting function is exactly the integral of the sinc function, the sine integral; for a square wave the description is not as simply stated. For the step function, the magnitude of the undershoot is thus exactly the integral of the (left) tail, integrating to the first negative zero: for the normalized sinc of unit sampling period, this

is The overshoot is accordingly of the same magnitude: the integral of the right tail, or, which amounts to the same thing, the difference between the integral from

Page 163: Elementary Functions

negative infinity to the first positive zero, minus 1 (the non-overshooting value).

The overshoot and undershoot can be understood thus: kernels are generally normalized to have integral 1, so they send constant functions to constant functions – otherwise they have gain. The value of a convolution at a point is a linear combination of the input signal, with coefficients (weights) the values of the kernel. If a kernel is non-negative, such as for a Gaussian kernel, then the value of the filtered signal will be a convex combination of the input values (the coefficients (the kernel) integrate to 1, and are non-negative), and will thus fall between the minimum and maximum of the input signal – it will not undershoot or overshoot. If, on the other hand, the kernel assumes negative values, such as the sinc function, then the value of the filtered signal will instead be an affine combination of the input values, and may fall outside of the minimum and maximum of the input signal, resulting in undershoot and overshoot, as in the Gibbs phenomenon.

Taking a longer expansion – cutting at a higher frequency – corresponds in the frequency domain to widening the brick-wall, which in the time domain corresponds to narrowing the sinc function and increasing its height by the same factor, leaving the integrals between corresponding points unchanged. This is a general feature of the

Page 164: Elementary Functions

Fourier transform: widening in one domain corresponds to narrowing and increasing height in the other. This results in the oscillations in sinc being narrower and taller and, in the filtered function (after convolution), yields oscillations that are narrower and thus have less area, but does not reduce the magnitude: cutting off at any finite frequency results in a sinc function, however narrow, with the same tail integrals. This explains the persistence of the overshoot and undershoot.

Oscillations can be interpreted as convolution with a sinc.

Higher cutoff makes the sinc narrower but taller, with the same magnitude tail integrals, yielding higher frequency oscillations, but whose magnitude does not vanish.

Page 165: Elementary Functions

Thus the features of the Gibbs phenomenon are interpreted as follows:

the undershoot is due to the impulse response having a negative tail integral, which is possible because the function takes negative values;

the overshoot offsets this, by symmetry (the overall integral does not change under filtering);

the persistence of the oscillations is because increasing the cutoff narrows the impulse response, but does not reduce its integral – the oscillations thus move towards the discontinuity, but do not decrease in magnitude.

[edit] The square wave example

Animation of the additive synthesis of a square wave with an increasing number of harmonics. The Gibbs phenomenon is visible especially when the number of harmonics is large.

Page 166: Elementary Functions

We now illustrate the above Gibbs phenomenon in the case of the square wave described earlier. In this case the period L is 2π, the discontinuity x0 is at zero, and the jump a is equal to π / 2. For simplicity let us just deal with the case when N is even (the case of odd N is very similar). Then we have

Substituting x = 0, we obtain

as claimed above. Next, we compute

If we introduce the normalized sinc

function, , we can rewrite this as

But the expression in square brackets is a numerical integration approximation to the integral

(more precisely, it is a midpoint rule approximation with

Page 167: Elementary Functions

spacing 2 / N). Since the sinc function is continuous, this approximation converges to the actual integral as

. Thus we have

which was what was claimed in the previous section. A similar computation shows

[edit] Consequences

In signal processing, the Gibbs phenomenon is undesirable because it causes artifacts, namely clipping from the overshoot and

Page 168: Elementary Functions

undershoot, and ringing artifacts from the oscillations. In the case of low-pass filtering, these can be reduced or eliminated by using different low-pass filters.

In MRI, the Gibbs phenomenon causes artifacts in the presence of adjacent regions of markedly differing signal intensity. This is most commonly encountered in spinal MR imaging, where the Gibbs phenomenon may simulate the appearance of syringomyelia.

Runge's phenomenonFrom Wikipedia, the free encyclopedia

Jump to: navigation, search

Page 169: Elementary Functions

The red curve is the Runge function.The blue curve is a 5th-order interpolating polynomial (using six equally-spaced interpolating points).The green curve is a 9th-order interpolating polynomial (using ten equally-spaced interpolating points).At the interpolating points, the error between the function and the interpolating polynomial is (by definition) zero. Between the interpolating points (especially in the region close to the endpoints 1 and −1), the error between the function and the interpolating polynomial gets worse for higher-order polynomials.

In the mathematical field of numerical analysis, Runge's phenomenon is a problem of oscillation at the edges of an interval that occurs when using polynomial interpolation with polynomials of high degree. It was discovered by Carl David Tolmé Runge when exploring the behavior of errors when using polynomial interpolation to approximate certain functions.[1] The discovery was important because it shows that going to higher degrees does not always improve accuracy. The phenomenon is similar to the Gibbs phenomenon in Fourier series approximations.

Contents

[hide]

1 Problem o 1.1 Reason

2 Mitigations to the problem o 2.1 Change of interpolation points o 2.2 Use of piecewise polynomials o 2.3 Constrained minimization o 2.4 Least squares fitting

3 See also

Page 170: Elementary Functions

4 References

[edit] Problem

Consider the function:

Runge found that if this function is interpolated at equidistant points xi between −1 and 1 such that:

with a polynomial Pn(x) of degree ≤ n, the resulting interpolation oscillates toward the end of the interval, i.e. close to −1 and 1. It can even be proven that the interpolation error tends toward infinity when the degree of the polynomial increases:

However, the Weierstrass approximation theorem states that there is some sequence of approximating polynomials for which the error goes to zero. This shows that high-degree polynomial interpolation at equidistant points can be troublesome.

[edit] Reason

The error between the generating function and the interpolating polynomial of order N is given by

for some ξ in (−1, 1).

For the case of the Runge function (also called the Cauchy–Lorentz function) shown above,

the first two derivatives are

Page 171: Elementary Functions

The magnitude of higher order derivatives of the Runge function get even larger. Therefore, the bound for the error (between the interpolating points) when using higher-order interpolating polynomials becomes larger.

[edit] Mitigations to the problem

[edit] Change of interpolation points

The oscillation can be minimized by using nodes that are distributed more densely towards the edges of the interval, specifically, with asymptotic density (on the

interval [−1,1]) given by the formula[2] . A standard example of such a set of nodes is Chebyshev nodes, for which the maximum error is guaranteed to diminish with increasing polynomial order. The phenomenon demonstrates that high degree polynomials are generally unsuitable for interpolation with equidistant nodes.

[edit] Use of piecewise polynomials

The problem can be avoided by using spline curves which are piecewise polynomials. When trying to decrease the interpolation error one can increase the number of polynomial pieces which are used to construct the spline instead of increasing the degree of the polynomials used.

[edit] Constrained minimization

One can also fit a polynomial of higher degree (for instance 2n instead of n + 1), and fit an interpolating polynomial whose first (or second) derivative has minimal L2 norm.

Page 172: Elementary Functions

[edit] Least squares fitting

Another method is fitting a polynomial of lower degree using the method of least squares. Generally, when

using m equidistant points, if then least squares approximation PN(x) is well-conditioned.[3]

Sigma approximationFrom Wikipedia, the free encyclopedia

Jump to: navigation, search

In mathematics, σ-approximation adjusts a Fourier summation to eliminate the Gibbs phenomenon which would otherwise occur at discontinuities.

A σ-approximated summation for a series of period T can be written as follows:

in terms of the normalized sinc function

Here, the term

is the Lanczos σ factor, which is responsible for eliminating most of the Gibbs phenomenon. It does not do so entirely, however, but one can square or even cube the expression to serially attenuate Gibbs Phenomenon in the most extreme cases.

Page 173: Elementary Functions

Square waveFrom Wikipedia, the free encyclopediaJump to: navigation, search

This article does not cite any references or sources. Please help improve this article by adding citations to reliable sources. Unsourced material may be challenged and removed. (September 2009)

Sine, square, triangle, and sawtooth waveforms

A square wave is a kind of non-sinusoidal waveform, most typically encountered in electronics and signal processing. An ideal square wave alternates regularly and instantaneously between two levels. Its stochastic counterpart is a two-state trajectory.

Contents

[hide]

1 Origins and uses 2 Examining the square wave 3 Characteristics of imperfect square

waves 4 Other definitions 5 See also

[edit] Origins and uses

Page 174: Elementary Functions

Square waves are universally encountered in digital switching circuits and are naturally generated by binary (two-level) logic devices. They are used as timing references or "clock signals", because their fast transitions are suitable for triggering synchronous logic circuits at precisely determined intervals. However, as the frequency-domain graph shows, square waves contain a wide range of harmonics; these can generate electromagnetic radiation or pulses of current that interfere with other nearby circuits, causing noise or errors. To avoid this problem in very sensitive circuits such as precision analog-to-digital converters, sine waves are used instead of square waves as timing references.

In musical terms, they are often described as sounding hollow, and are therefore used as the basis for wind instrument sounds created using subtractive synthesis. Additionally, the distortion effect used on electric guitars clip the outermost regions of the waveform, causing it to increasingly resemble a square wave as more distortion is applied.

Simple two-level Rademacher functions are square waves.

[edit] Examining the square wave

In contrast to the sawtooth wave, which contains all integer harmonics, the square wave contains only odd integer harmonics.

Using Fourier series we can write an ideal square wave as an infinite series of the form

A curiosity of the convergence of the Fourier series representation of the square wave is the Gibbs phenomenon. Ringing artifacts in non-ideal square waves can be shown to be related to this phenomenon. The Gibbs phenomenon can be prevented by the use of σ-approximation, which uses the Lanczos sigma factors to help the sequence converge more smoothly.

An ideal square wave requires that the signal changes from the high to the low state cleanly and instantaneously. This is impossible to achieve in real-world systems, as it would require infinite bandwidth.

Page 175: Elementary Functions

Animation of the additive synthesis of a square wave with an increasing number of harmonics

Real-world square-waves have only finite bandwidth, and often exhibit ringing effects similar to those of the Gibbs phenomenon, or ripple effects similar to those of the σ-approximation.

For a reasonable approximation to the square-wave shape, at least the fundamental and third harmonic need to be present, with the fifth harmonic being desirable. These bandwidth requirements are important in digital electronics, where finite-bandwidth analog approximations to square-wave-like waveforms are used. (The ringing transients are an important electronic consideration here, as they may go beyond the electrical rating limits of a circuit or cause a badly positioned threshold to be crossed multiple times.)

The ratio of the high period to the total period of a square wave is called the duty cycle. A true square wave has a 50% duty cycle - equal high and low periods. The average level of a square wave is also given by the duty cycle, so by varying the on and off periods and then averaging, it is possible to represent any value between the two limiting levels. This is the basis of pulse width modulation.

Square wave sound sample

5 seconds of square wave at 1 kHz

Problems listening to this file? See media help.

Page 176: Elementary Functions

[edit] Characteristics of imperfect square waves

As already mentioned, an ideal square wave has instantaneous transitions between the high and low levels. In practice, this is never achieved because of physical limitations of the system that generates the waveform. The times taken for the signal to rise from the low level to the high level and back again are called the rise time and the fall time respectively.

If the system is overdamped, then the waveform may never actually reach the theoretical high and low levels, and if the system is underdamped, it will oscillate about the high and low levels before settling down. In these cases, the rise and fall times are measured between specified intermediate levels, such as 5% and 95%, or 10% and 90%. Formulas exist that can determine the approximate bandwidth of a system given the rise and fall times of the waveform.

[edit] Other definitions

The square wave has many definitions, which are equivalent except at the discontinuities:

It can be defined as simply the sign function of a sinusoid:

which will be 1 when the sinusoid is positive, −1 when the sinusoid is negative, and 0 at the discontinuities. It can also be defined with respect to the Heaviside step function u(t) or the rectangular function ⊓(t):

T is 2 for a 50% duty cycle. It can also be defined in a piecewise way:

when

Page 177: Elementary Functions

Pulse waveFrom Wikipedia, the free encyclopedia

Jump to: navigation, search

This article is about a pulse wave form. For a heart beat, see Pulse.

This article is about a rectangular pulse train. For a Dirac pulse train, see Sampling function.

For other uses, see Pulse (disambiguation).

This article does not cite any references or sources. Please help improve this article by adding citations to reliable sources. Unsourced material may be challenged and removed. (December 2009)

The shape of the pulse wave is defined by its duty cycle D, which is the ratio between the

pulse duration (τ) and the period (T)

A pulse wave or pulse train is a kind of non-sinusoidal waveform that is similar to a square wave, but does not have the symmetrical shape associated with a perfect square wave. It is a term common to synthesizer programming, and is a typical waveform available on many synths. The exact shape of the wave is determined by the duty cycle of the oscillator. In many synthesizers, the duty cycle can be modulated (sometimes called pulse-width modulation) for a more dynamic timbre. The pulse wave is also known as the rectangular wave, the periodic version of the rectangular function.

Sine waveFrom Wikipedia, the free encyclopediaJump to: navigation, search"Sinusoid" redirects here. For the blood vessel, see Sinusoid (blood vessel).

Page 178: Elementary Functions

The graphs of the sine and cosine functions are sinusoids of different phases.

The sine wave or sinusoid is a mathematical function that describes a smooth repetitive oscillation. It occurs often in pure mathematics, as well as physics, signal processing, electrical engineering and many other fields. Its most basic form as a function of time (t) is:

where:

A, the amplitude, is the peak deviation of the function from its center position.

ω, the angular frequency, specifies how many oscillations occur in a unit time interval, in radians per second

φ, the phase, specifies where in its cycle the oscillation begins at t = 0.o When the phase is non-zero, the entire waveform appears to be

shifted in time by the amount φ/ω seconds. A negative value represents a delay, and a positive value represents an advance.

Sine wave

5 seconds of a 220 Hz sine wave

Problems listening to this file? See media help.

Page 179: Elementary Functions

The oscillation of an undamped spring-mass system around the equilibrium is a sine wave.

The sine wave is important in physics because it retains its waveshape when added to another sine wave of the same frequency and arbitrary phase. It is the only periodic waveform that has this property. This property leads to its importance in Fourier analysis and makes it acoustically unique.

Contents

[hide]

1 General form 2 Occurrences 3 Fourier

series 4 See also 5 References

[edit] General form

In general, the function may also have:

a spatial dimension, x (aka position), with frequency k (also called wavenumber)

a non-zero center amplitude, D

Page 180: Elementary Functions

which looks like this:

The wavenumber is related to the angular frequency by:.

where λ is the wavelength, f is the frequency, and c is the speed of propagation.

This equation gives a sine wave for a single dimension, thus the generalized equation given above gives the amplitude of the wave at a position x at time t along a single line. This could, for example, be considered the value of a wave along a wire.

In two or three spatial dimensions, the same equation describes a travelling plane wave if position x and wavenumber k are interpreted as vectors, and their product as a dot product. For more complex waves such as the height of a water wave in a pond after a stone has been dropped in, more complex equations are needed.

[edit] Occurrences

Illustrating the cosine wave's fundamental relationship to the circle.

This wave pattern occurs often in nature, including ocean waves, sound waves, and light waves.

Page 181: Elementary Functions

A cosine wave is said to be "sinusoidal", because cos(x) = sin(x + π / 2), which is also a sine wave with a phase-shift of π/2. Because of this "head start", it is often said that the cosine function leads the sine function or the sine lags the cosine.

The human ear can recognize single sine waves as sounding clear because sine waves are representations of a single frequency with no harmonics; some sounds that approximate a pure sine wave are whistling, a crystal glass set to vibrate by running a wet finger around its rim, and the sound made by a tuning fork.

To the human ear, a sound that is made up of more than one sine wave will either sound "noisy" or will have detectable harmonics; this may be described as a different timbre.

[edit] Fourier series

Sine, square, triangle, and sawtooth waveformsMain article: Fourier analysis

In 1822, Joseph Fourier, a French mathematician, discovered that sinusoidal waves can be used as simple building blocks to describe and approximate any periodic waveform including square waves. Fourier used it as an analytical tool in the study of waves and heat flow. It is frequently used in signal processing and the statistical analysis of time series.

Page 182: Elementary Functions

Triangle waveFrom Wikipedia, the free encyclopedia

Jump to: navigation, search

A triangle wave is a non-sinusoidal waveform named for its triangular shape.

A bandlimited triangle wave pictured in the time domain (top) and frequency domain (bottom). The fundamental is at 220 Hz (A3).

Like a square wave, the triangle wave contains only odd harmonics. However, the higher harmonics roll off much faster than in a square wave (proportional to the inverse square of the harmonic number as opposed to just the inverse).

It is possible to approximate a triangle wave with additive synthesis by adding odd harmonics of the fundamental, multiplying every (4n−1)th harmonic by −1 (or changing its phase by π), and rolling off the harmonics by the inverse square of their relative frequency to the fundamental.

This infinite Fourier series converges to the triangle wave:

Page 183: Elementary Functions

where is the angular frequency.

Animation of the additive synthesis of a triangle wave with an increasing number of harmonics. See Fourier Analysis for a mathematical description.

Triangle wave sound sample

5 seconds of anti-aliased triangle wave at 220 Hz

Problems listening to this file? See media help.

Another definition of the triangle wave, with range from -1 to 1 and period 2a is:

where the symbol represent the floor function of n.

Also, the triangle wave can be the absolute value of the sawtooth wave:

Page 184: Elementary Functions

The triangle wave can also be expressed as the integral of the square wave:

Sawtooth waveFrom Wikipedia, the free encyclopediaJump to: navigation, search

A bandlimited sawtooth wave pictured in the time domain (top) and frequency domain (bottom). The fundamental is at 220 Hz (A3).

The sawtooth wave (or saw wave) is a kind of non-sinusoidal waveform. It is named a sawtooth based on its resemblance to the teeth on the blade of a saw.

The convention is that a sawtooth wave ramps upward and then sharply drops. However, there are also sawtooth waves in which the wave ramps downward and then sharply rises. The latter type of sawtooth wave is called a "reverse sawtooth wave" or "inverse sawtooth wave".

The piecewise linear function

Page 185: Elementary Functions

based on the floor function of time t is an example of a sawtooth wave with period 1.

A more general form, in the range −1 to 1, and with period a, is

This sawtooth function has the same phase as the sine function.

A sawtooth wave's sound is harsh and clear and its spectrum contains both even and odd harmonics of the fundamental frequency. Because it contains all the integer harmonics, it is one of the best waveforms to use for synthesizing musical sounds, particularly bowed string instruments like violins and cellos, using subtractive synthesis.

A sawtooth can be constructed using additive synthesis. The infinite Fourier series

converges to an inverse sawtooth wave. A conventional sawtooth can be constructed using

In digital synthesis, these series are only summed over k such that the highest harmonic, Nmax, is less than the Nyquist frequency (half the sampling frequency). This summation can generally be more efficiently calculated with a fast Fourier transform. If the waveform is digitally created directly in the time domain using a non-bandlimited form, such as y = x - floor(x), infinite harmonics are sampled and the resulting tone contains aliasing distortion.

Page 186: Elementary Functions

Animation of the additive synthesis of a sawtooth wave with an increasing number of harmonics

An audio demonstration of a sawtooth played at 440 Hz (A4) and 880 Hz (A5) and 1760 Hz (A6) is available below. Both bandlimited (non-aliased) and aliased tones are presented.

Sawtooth aliasing demo

Sawtooth waves played bandlimited and aliased at 440 Hz, 880 Hz, and 1760 Hz

Problems listening to this file? See media help.

[edit] Applications

Sawtooth waves are perhaps best known for their use in music. The sawtooth and square waves are the most common starting points used to create sounds with subtractive analog and virtual analog music synthesizers.

The sawtooth wave is the form of the vertical and horizontal deflection signals used to generate a raster on CRT-based television or monitor screens. Oscilloscopes also use a sawtooth wave for their horizontal deflection, though they typically use electrostatic deflection.o On the wave's "ramp", the magnetic field produced

by the deflection yoke drags the electron beam across the face of the CRT, creating a scan line.

o On the wave's "cliff", the magnetic field suddenly collapses, causing the electron beam to return to its resting position as quickly as possible.

o The voltage applied to the deflection yoke is adjusted by various means (transformers, capacitors, center-tapped windings) so that the half-way voltage on the sawtooth's cliff is at the zero

Page 187: Elementary Functions

mark, meaning that a negative voltage will cause deflection in one direction, and a positive voltage deflection in the other; thus, a center-mounted deflection yoke can use the whole screen area to depict a trace. Frequency is 15.734 kHz on NTSC, 15.625 kHz for PAL and SECAM).

o The vertical deflection system operates the same way as the horizontal, though at a much lower frequency (59.94 Hz on NTSC, 50 Hz for PAL and SECAM).

o The ramp portion of the wave must appear as a straight line. If otherwise, it indicates that the voltage isn't increasing linearly, and therefore that the magnetic field produced by the deflection yoke is not linear. As a result, the electron beam will accelerate during the non-linear portions. This would result in a television image "squished" in the direction of the non-linearity. Extreme cases will show marked brightness increases, since the electron beam spends more time on that side of the picture.

o The first television receivers had controls allowing users to adjust the picture's vertical or horizontal linearity. Such controls were not present on later sets as the stability of electronic components had improved.