Differential and Integral Calculus

Site: MyCourses
Course: MS-A0111 - Differential and Integral Calculus 1, Lecture, 13.9.2021-27.10.2021
Book: Differential and Integral Calculus
Printed by: Guest user
Date: Friday, 22 November 2024, 6:14 PM

1. Sequences

Basics of sequences


This section contains the most important definitions about sequences. Through these definitions the general notion of sequences will be explained, but then restricted to real number sequences.

Definition: Sequence

Let M be a non-empty set. A sequence is a function:

f:\mathbb{N}\rightarrow M.

Occasionally we speak about a sequence in M.


Note. Characteristics of the set \mathbb{N} give certain characteristics to the sequence. Because \mathbb{N} is ordered, the terms of the sequence are ordered.

Definition: Terms and Indices

A sequence can be denoted denoted as

(a_{1}, a_{2}, a_{3}, \ldots) = (a_{n})_{n\in\mathbb{N}} = (a_{n})_{n=1}^{\infty} = (a_{n})_{n}

instead of f(n). The numbers a_{1},a_{2},a_{3},\ldots\in M are called the terms of the sequence.


Because of the mapping \begin{aligned} f:\mathbb{N} \rightarrow & M \\
		n \mapsto & a_{n}\end{aligned} we can assign a unique number n\in\mathbb{N} to each term. We write this number as a subscript and define it as the index; it follows that we can identify any term of the sequence by its index.

n 1 2 3 4 5 6 7 8 9 \ldots
\downarrow \downarrow \downarrow \downarrow \downarrow \downarrow \downarrow \downarrow \downarrow
a_{n}

a_{1} a_{2} a_{3} a_{4} a_{5} a_{6} a_{7} a_{8} a_{9} \ldots

A few easy examples

Example 1: The sequence of natural numbers

The sequence (a_{n})_{n} defined by a_{n}:=n,\,n\in \mathbb{N} is called the sequence of natural numbers. Its first few terms are: a_1=1,\, a_2=2,\, a_3=3, \ldots This special sequence has the property that every term is the same as its index.


Example 2: The sequence of triangular numbers

Triangular numbers get their name due to the following geometric visualization: Stacking coins to form a triangular shape gives the following diagram:


To the first coin in the first layer we add two coins in a second layer to form the second picture a_2. In turn, adding three coins to a_2 forms a_3. From a mathematical point of view, this sequence is the result of summing natural numbers. To calculate the 10th triangular number we need to add the first 10 natural numbers: D_{10} = 1+2+3+\ldots+9+10 In general form the sequence is defined as: D_{n} = 1+2+3+\ldots+(n-1)+n.


This motivates the following definition:

Notation and Definition: Sum sequence

Let (a_n)_n, a_n: \mathbb{N}\to M be a sequence with terms a_n, the sum is written: a_1 + a_2 + a_3 + \ldots + a_{n-1} + a_n =: \sum_{k=1}^n a_k The sign \sum is called sigma. Here, the index k increases from 1 to n.

Sum sequences are sequences whose terms are formed by summation of previous terms.

Thus the nth triangular number can be written as: D_n = \sum_{k=1}^n k

Example 3: Sequence of square numbers

The sequence of square numbers (q_n)_n is defined by: q_n=n^2. The terms of this sequence can also be illustrated by the addition of coins.

Interestingly, the sum of two consecutive triangular numbers is a square number. So, for example, we have: 3+1=4 and 6+3=9. In general this gives the relationship:

q_n=D_n + D_{n-1}


Example 4: Sequence of cube numbers

Analogously to the sequence of square number, we give the definition of cube numbers as a_n := n^3. The first terms of the sequence are: (1,8,27,64,125,\ldots).


Example 5.

Let (q_n)_n with q_n := n^2 be the sequence of square numbers \begin{aligned}(1,4,9,16,25,36,49,64,81,100 \ldots)\end{aligned} and define the function \varphi(n) = 2n. The composition (q_{2n})_n yields: \begin{aligned}(q_{2n})_n &= (q_2,q_4,q_6,q_8,q_{10},\ldots) \\
        &= (4,16,36,64,100,\ldots).\end{aligned}

Definition: Sequence of differences

Given a sequence (a_{n})_{n}=a_{1},\, a_{2},\, a_{3},\ldots,\, a_{n},\ldots; then (a_{n+1}-a_{n})_{n}:=a_{2}-a_{1}, a_{3}-a_{2},\dots is called the 1st difference sequence of (a_{n})_{n}

The 1st difference sequence of the 1st difference sequence is called the 2nd difference sequence. Analogously the nth difference< sequence is defined.

Example 6.

Given the sequence (a_n)_n with a_n := \frac{n^2+n}{2}, i.e. \begin{aligned}(a_n)_n &= (1,3,6,10,15,21,28,36,\ldots)\end{aligned} Let (b_n)_n be its 1st difference sequence. Then it follows that \begin{aligned}(b_n)_n &= (a_2-a_1, a_3-a_2, a_4-a_3,\ldots) \\
		&= (2,3,4,5,6,7,8,9)\end{aligned} A term of (b_n)_n has the general form \begin{aligned}b_n &= a_{n+1}-a_{n} \\
		&= \frac{(n+1)^2+(n+1)}{2} - \frac{n^2+n)}{2} \\
		&= \frac{(n+1)^2+(n+1)-n^2 - n }{2} \\
		&= \frac{(n^2+2n+1)+1-n^2}{2} \\
		&= \frac{2n+2}{2} \\
		&= n + 1.\end{aligned}

Some important sequences


There are a number of sequences that can be regarded as the basis of many ideas in mathematics, but also can be used in other areas (e.g. physics, biology, or financial calculations) to model real situations. We will consider three of these sequences: the arithmetic sequence, the geometric sequence, and Fibonacci sequence, i.e. the sequence of Fibonacci numbers.

The arithmetic sequence

There are many definitions of the arithmetic sequence:

Definition A: Arithmetic sequence

A sequence (a_{n})_{n} is called the arithmetic sequence, when the difference d \in \mathbb{R} between two consecutive terms is constant, thus: a_{n+1}-a_{n}=d \text{ with } d=const.

Note: The explicit rule of formation follows directly from definition A: a_{n}=a_{1}+(n-1)\cdot d For the nth term of an arithmetic sequence we also have the recursive formation rule: a_{n+1}=a_n + d.

Definition B: Arithmetic sequence

A non-constant sequence (a_{n})_{n} is called an arithmetic sequence (1st order) when its 1st difference sequence is a sequence of constant value.

This rule of formation gives the arithmetic sequence its name: The middle term of any three consecutive terms is the arithmetic mean of the other two, for example:

a_2 = \frac{a_1+a_3}{2}.

Example 1.

The sequence of natural numbers (a_n)_n = (1,2,3,4,5,6,7,8,9,\ldots) is an arithmetic sequence, because the difference, d, between two consecutive terms is always given as d=1.

The geometric sequence

The geometric sequence has multiple definitions:

Definition: Geometric sequence

A sequence (a_{n})_{n} is called a geometric sequence when the ratio of any two consecutive terms is always constant q\in\mathbb{R}, thus \frac{a_{n+1}}{a_{n}}=q \text{ for all } n\in\mathbb{N}.

Note.The recursive relationship a_{n+1} = q\cdot a_n of the terms of the geometric sequence and the explicit formula for the calculation of the n th term of a geometric sequence a_n=a_1\cdot q^{n-1} follows directly from the definition.

Again the name and the rule of formation of this sequence are connected: Here, the middle term of three consecutive terms is the geometric mean of the other two, e.g.: a_2 = \sqrt{a_1\cdot a_3}.

Example 2.

Let a and q be fixed positive numbers. The sequence (a_n)_n with a_n := aq^{n-1}, i.e. \left( a_1, a_2, a_3, a_4,\ldots \right) = \left( a, aq, aq^2, aq^3,\ldots \right) is a geometric sequence. If q\geq1 the sequence is monotonically increasing. If q it is strictly decreasing. The corresponding range {a,aq,aq^2, aq^3} is finite in the case q=1 (namely, a singleton), otherwise it is infinite.

The Fibonacci sequence

The Fibonacci sequence is famous because it plays a role in many biological processes, for instance in plant growth, and is frequently found in nature. The recursive definition is:

Definition: Fibonacci sequence

Let a_0 = a_1 = 1 and let a_n := a_{n-2}+a_{n-1} for n\geq2. The sequence (a_n)_n is then called the Fibonacci sequence. The terms of the sequence are called the Fibonacci numbers.

The sequence is named after the Italian mathematician Leonardo of Pisa (ca. 1200 AD), also known as Fibonacci (son of Bonacci). He considered the size of a rabbit population and discovered the number sequence: (1,1,2,3,5,8,13,21,34,55,\ldots),


Example 3.

The structure of sunflower heads can be described by a system of two spirals, which radiate out symmetrically but contra rotating from the centre; there are 55 spirals which run clockwise and 34 which run counter-clockwise.

Pineapples behave very similarly. There we have 21 spirals running in one direction and 34 running in the other. Cauliflower, cacti, and fir cones are also constructed in this manner.


Convergence, divergence and limits


The following chapter deals with the convergence of sequences. We will first introduce the idea of zero sequences. After that we will define the concept of general convergence.

Preliminary remark: Absolute value in \mathbb{R}

The absolute value function x \mapsto |x| is fundamental in the study of convergence of real number sequences. Therefore we should summarise again some of the main characteristics of the absolute value function:

Definition: Absolute Value

For any given number x\in\mathbb{R} its absolute value |x| is defined by \begin{aligned}|x|:=\begin{cases}x & \text{for }x\geq0,\\
		-x & \text{for }x

Graph of the absolute value function


Theorem: Calculation Rule for the Absolute Value

For x,y\in\mathbb{R} the following is always true:

  1. |x|\geq0,

  2. |x|=0 if and only if x=0.

  3. |x\cdot y|=|x|\cdot|y| (Multiplicativity)

  4. |x+y|\leq|x|+|y| (Triangle Inequality)

Proof.

Parts 1.-3. Results follow directly from the definition and by dividing it up into separate cases of the different signs of x and y

Part 4. Here we divide the triangle inequality into different cases.
Case 1.

First let x,y \geq 0. Then it follows that \begin{aligned}|x+y|=x+y=|x|+|y|\end{aligned} and the desired inequality is shown.

Case 2.

Next let x,y < 0. Then: \begin{aligned}|x+y|=-(x+y)=(-x)+ (-y)=|x|+|y|\end{aligned}

Case 3.

Finally we consider the case x\geq 0 and y. Here we have two subcases:

  • For x \geq -y we have x+y\geq 0 and thus |x+y|=x+y from the definition of absolute value. Because y then y and therefore also x+y < x-y. Overall we have: \begin{aligned}|x+y| = x+y < x-y = |x|+|y|\end{aligned}

  • For x < -y then x+y. We have |x+y|=-(x+y)=-x-y. Because x\geq0, we have -x < x and thus -x-y\leq x-y. Overall we have: \begin{aligned}|x+y| = -x-y \leq x-y = |x|+|y|\end{aligned}

Case 4.

The case x and y\geq0 we prove it analogously to the case 3, in which x and y are exchanged.

\square

Zero sequences

Definition: Zero sequence

A sequence (a_{n})_{n} s called a zero sequence, if for every \varepsilon>0, there exists an index n_{0}\in\mathbb{N} such that |a_{n}| < \varepsilon for every n\geq n_{0},\, n\in\mathbb{N}. In this case we also say that the sequence converges to zero.

Informally: We have a zero sequence, if the terms of the sequence with high enough indices are arbitrarily close to zero.

Example 1.

The sequence (a_n)_n defined by a_{n}:=\frac{1}{n}, i.e. \left(a_{1},a_{2},a_{3},a_{4},\ldots\right):=\left(\frac{1}{1},\frac{1}{2},\frac{1}{3},\frac{1}{4},\ldots\right) is called the harmonic sequence. Clearly, it is positive for all n\in\mathbb{N}, however as n increases the absolute value of each term decreases getting closer and closer to zero.
Take for example \varepsilon := \frac{1}{5000}, then choosing the index n_0 = 5000, it follows that a_n, for all n\geq n_0.

The harmonic sequence converges to zero

Example 2.

Consider the sequence (a_n)_n \text{ where } a_n:=\frac{1}{\sqrt{n}}. Let \varepsilon := \frac{1}{1000}.We then obtain the index n_0=1000000 in this manner that for all terms a_n where n\geq n_0 a_n < \frac{1}{1000}=\varepsilon.

Note. To check whether a sequence is a zero sequence, you must choose an (arbitrary) \varepsilon \in \mathbb{R} where \varepsilon > 0. Then search for a index n_0, after which all terms n are smaller then said \varepsilon.

Example 3.

We consider the sequence (a_n)_n, defined by a_n := \left( -1 \right)^n \cdot \frac{1}{n^2}.

Because of the factors (-1)^n two consecutive terms have different signs; we call a sequence whose signs change in this way an alternating sequence.

We want to show that this sequence is a zero sequence. According to the definition we have to show that for every \varepsilon > 0 there exist n_0 \in \mathbb{N}, such that we have the inequality: |a_n|< \varepsilon for every term a_n where n\geq n_0.

Proof.

Firstly we let \varepsilon > 0 be an arbitrary constant. Because the inequality  |a_n|< \varepsilon must hold true for an arbitrary \varepsilon we must find the index n_0 which depends on each \varepsilon. More exactly: The inequality |a_{n_0}|=\left| \frac{1}{{n_0}^2} \right|= \frac{1}{{n_0}^2} must be true for the index n_0. Solve for n_0: n_0 > \frac{1}{\sqrt{\varepsilon}}, this index n_0 gives our desired characteristic for every \varepsilon.

Negative examples

The following are examples of non-convergent alternating sequences:

  • a_n = (-1)^n

  • a_n = (-1)^n \cdot n

Theorem: Characteristics of Zero sequences

Let (a_n)_n and (b_n)_n be two sequences. Then:

  1. Let (a_n)_n be a zero sequence, if b_n = a_n or b_n = -a_n for all n\in\mathbb{N} then (b_n)_n is also a zero sequence.

  2. Let (a_n)_n be a zero sequence, if -a_n\leq b_n \leq a_n for all n\in\mathbb{N} then (b_n)_n is also a zero sequence.

  3. Let (a_n)_n be a zero sequence, then (c\cdot a_n)_n where c \in \mathbb{R} is also a zero sequence.

  4. If (a_n)_n and (b_n)_n are zero sequences, then (a_n + b_n)_n is also a zero sequence.

Proof.

Parts 1 and 2. If (a_n)_n is a zero sequence, then according to the definition there is an index n_0 \in \mathbb{N}, such that |a_n| for every n\geq n_0 and an arbitrary \varepsilon\in\mathbb{R}. But then we have |b_n|\leq|a_n|; this proves parts 1 and 2 are correct.

Part 3. If c=0, then the result is trivial. Let c\neq0 and choose \varepsilon > 0 such that \begin{aligned}|a_n| for all n\geq n_0. Rearranging we get: \begin{aligned} |c|\cdot|a_n|=|c\cdot a_n|

Part 4.

Because (a_n)_n is a zero sequence, by the definition we have |a_n| for all n\geq n_0. Analogously, for the zero sequence (b_n)_n there is a m_0 \in \mathbb{N} with |b_n| for all n\geq m_0.

Then for all n > \max(n_0,m_0) it follows (using the triangle inequality) that: \begin{aligned}|a_n + b_n|\leq|a_n|+|b_n|

\square

Convergence, divergence

The concept of zero sequences can be expanded to give us the convergence of general sequences:

Definition: Convergence and Divergence

A sequence (a_{n})_{n} is called convergent to a\in\mathbb{R}, if for every \varepsilon>0 there exists a n_{0} such that: |a_{n}-a| \lt \varepsilon \text{ for all }n\in\mathbb{N}_{0},\text{ where }n\geq n_{0}

An equivalent definition can be defined by:

A sequence (a_{n})_{n} is called convergent to a\in\mathbb{R}, if (a_{n}-a)_{n} is a zero sequence.

Example 4.

We consider the sequence (a_n)_n where a_n=\frac{2n^2+1}{n^2+1}. By plugging in large values of n, we can see that for n\to\infty a_n \to 2 and therefore we can postulate that the limit is a=2.

Proof.

For a vigorous proof, we show that for every \varepsilon > 0 there exists an index n_0\in\mathbb{N}, such that for every term a_n with n>n_0 the following relationship holds: \left| \frac{2n^2+1}{n^2+1} - 2\right| < \varepsilon.

Firstly we estimate the inequality: \begin{aligned}\left|\frac{2n^2+1}{n^2+1}-2\right| =&\left|\frac{2n^2+1-2\cdot\left(n^2+1\right)}{n^2+1}\right| \\
			=&\left|\frac{2n^2+1-2n^2-2}{n^2+1}\right| \\
			=&\left|-\frac{1}{n^2+1}\right| \\
			=&\left|\frac{1}{n^2+1}\right| \\

Now, let \varepsilon > 0 be an arbitrary constant. We then choose the index n_0\in\mathbb{N}, such that n_0 > \frac{1}{\varepsilon} \text{, or equivalently, }  \frac{1}{n_0} < \varepsilon. Finally from the above inequality we have: \left|\frac{2n^2+1}{n^2+1}-2\right| < \frac{1}{n} < \frac{1}{n_0} < \varepsilon, Thus we have proven the claim and so by definition a=2 is the limit of the sequence.

\square

If a sequence is convergent, then there is exactly one number which is the limit. This characteristic is called the uniqueness of convergence.

Theorem: Uniqueness of Convergence

Let (a_{n})_{n} be a sequence that converges to a\in\mathbb{R} and to b\in\mathbb{R}. This implies a=b.

Proof.

Assume a\ne b; choose \varepsilon\in\mathbb{R} with \varepsilon:=\frac{1}{3}|a-b|. Then in particular [a-\varepsilon,a+\varepsilon]\cap[b-\varepsilon,b+\varepsilon]=\emptyset.

Because (a_{n})_{n} converges to a, there is, according to the definition of convergence, a index n_{0}\in\mathbb{N} with |a_{n}-a|< \varepsilon for n\geq n_{0}. Furthermore, because (a_{n})_{n} converges to b there is also a \widetilde{n_{0}}\in\mathbb{N} with |a_{n}-b|< \varepsilon for n\geq\widetilde{n_{0}}. For n\geq\max\{n_{0},\widetilde{n_{0}}\} we have: \begin{aligned}\varepsilon\ = &\ \frac{1}{3}|a-b| \Rightarrow\\
			3\varepsilon\ = &\ |a-b|\\
			= &\ |(a-a_{n})+(a_{n}-b)|\\
			\leq &\ |a_{n}-a|+|a_{b}-b|\\
			< &\ \varepsilon+\varepsilon=2\varepsilon,\end{aligned} Consequently we have obtained 3\varepsilon\leq2\varepsilon, which is a contradiction as \varepsilon>0. Therefore the assumption must be wrong, so a=b.

\square


Definition: Divergent, Limit

If provided that a sequence (a_{n})_{n} and an a\in\mathbb{R} exist, to which the sequence converges, then the sequence is called convergent and a is called the limit of the sequence, otherwise it is called divergent.

Notation. (a_{n})_{n} is convergent to a is also written: a_{n}\rightarrow a,\text{ or }\lim_{n\rightarrow\infty}a_{n}=a. Such notation is allowed, as the limit of a sequence is always unique by the above Theorem (provided it exists).


Theorem: Bounded Sequences

A convergent sequence (a_n)_n is bounded i.e. there exists a constant r\in\mathbb{R} such that: |a_n| \lt r for all n\in\mathbb{N}.

Proof.

We assume that the sequence (a_n)_n has the limit a. By the definition of convergence, we have that |a_n - a| for all \varepsilon \in \mathbb{R} and n\geq n_0. Choosing \varepsilon = 1 gives:
\begin{aligned}|a_n|-|a|&\ \leq |a_n -a| \\
			&\ < 1,\end{aligned} And therefore also |a_n|\leq |a|+1.

Thus for all n\in \mathbb{N}: |a_n|\leq \max \left\{ |a_1|,|a_2|,\ldots,|a_{n_0}|,|a|+1 \right\}=:r

\square

Rules for convergent sequences

Theorem: Subsequences

Let (a_{n})_{n} be a sequence such that a_{n}\rightarrow a and let (a_{\varphi(n)})_{n} be a subsequence of (a_{n})_{n}. Then it follows that (a_{\varphi(n)})_{n}\rightarrow a.

Informally: If a sequence is convergent then all of its subsequences are also convergent and in fact converge to the same limit as the original.

Proof.

By the definition of a subsequence \varphi(n)\geq n. Because a_{n}\rightarrow a it is implicated that |a_{n}-a| for n\geq n_{0}, therefore |a_{\varphi(n)}-a| for these indices n.

\square


Theorem: Rules

Let (a_{n})_{n} and (b_{n})_{n} be sequences with a_{n}\rightarrow a and b_{n}\rightarrow b. Then for \lambda, \mu \in \mathbb{R} it follows that:

  1. \lambda \cdot (a_n)+\mu \cdot (b_n) \to \lambda \cdot a + \mu \cdot b

  2. (a_n)\cdot (b_n) \to a\cdot b

Informally: Sums, differences and products of convergent sequences are convergent.

Proof.

Part 1. Let \varepsilon > 0. We must show, that for all n \geq n_0 it follows that: |\lambda \cdot a_n + \mu \cdot b_n - \lambda \cdot a - \mu  \cdot b| < \varepsilon. The left hand side we estimate using: |\lambda (a_n-a)+\mu (b_n - b)| \leq |\lambda|\cdot|a_n-a|+|\mu|\cdot|b_n-b|.

Because (a_n)_n and (b_n)_n converge, for each given \varepsilon > 0 it holds true that: \begin{aligned}|a_n - a|

Therefore \begin{aligned}|\lambda|\cdot|a_n-a|+|\mu|\cdot|b_n-b| < &\ |\lambda|\varepsilon_1 + |\mu|\varepsilon_2 \\
			= &\ \textstyle{ \frac{\varepsilon}{2} + \frac{\varepsilon}{2} } = \varepsilon\end{aligned} for all numbers n \geq \max \{n_0,n_1\}. Therefore the sequence \left( \lambda \left( a_n - a \right) + \mu \left( b_n - b \right) \right)_n is a zero sequence and the desired inequality is shown.

Part 2. Let \varepsilon > 0. We have to show, that for all n > n_0 |a_n b_n - a b| < \varepsilon. Furthermore an estimation of the left hand side follows: \begin{aligned}
			|a_n b_n - a b| =&\ |a_n b_n - a b_n + a b_n - ab| \\
			\leq &\ |b_n|\cdot|a_n-a| + |a|\cdot|b_n - b|.\end{aligned} We choose a number B, such that |b_n| \lt b for all n and |a| \lt b. Such a value of B exists by the Theorem of convergent sequences being bounded. We can then use the estimation: \begin{aligned}|b_n|\cdot|a_n-a| + |a|\cdot|b_n - b| For all n>n_0 we have |a_n - a| and |b_n - b|, and - putting everything together - the desired inequality it shown.

\square

'; }], { useMathJax : true, fixed : true, strokeOpacity : 0.6 }); bars.push(newbar(k)); } board.fullUpdate(); })(); /* geometric sequence */ (function() { var board = JXG.JSXGraph.initBoard('jxgbox02', { boundingbox : [-1, 25, 11, -5 ], axis : false, shownavigation : false, showcopyright : false }); var xaxis = board.create('axis', [[0, 0], [1, 0]], { straightFirst : false, highlight : false, drawZero : true, scaleSymbol: 'a', ticks : { drawLabels : false, minorTicks : 0, majorHeight : 15, label : { highlight : false, offset : [0, -15] } } }); var yaxis = board.create('axis', [[0, 0], [0, 1]], { straightFirst : false, highlight : false, ticks : { minorTicks : 0, majorHeight : 15, label : { offset : [-15, 0 ], position : 'lrt', highlight : false } } }); xaxis.defaultTicks.ticksFunction = function() { return 1; }; yaxis.defaultTicks.ticksFunction = function() { return 5; }; var a = .5, q = 3/2; var newbar = function(k) { var height = a*Math.pow(q,k-1); return board.create('polygon', [[k-1/4, 0], [k+1/4, 0], [k+1/4, height], [k-1/4, height]] , { vertices : { visible : false }, borders : { strokeColor : 'black', strokeOpacity : .6, highlight : false }, fillColor : '#b2caeb', fixed : true, highlight : false }); } var bars = []; for(var i=0; i < 10; i++) { const k = i+1; board.create('text', [k-.1, -.6, function() { return 'a_{' + k + '}'; }], { useMathJax : true, fixed : true, strokeOpacity : 0.6 }); bars.push(newbar(k)); } board.fullUpdate(); })(); /* fibonacci sequence */ (function() { var board = JXG.JSXGraph.initBoard('jxgbox03', { boundingbox : [-1, 40, 10, -5 ], axis : false, shownavigation : false, showcopyright : false }); var xaxis = board.create('axis', [[0, 0], [1, 0]], { straightFirst : false, highlight : false, drawZero : true, ticks : { drawLabels : false, minorTicks : 0, majorHeight : 15, label : { highlight : false, offset : [0, -15] } } }); var yaxis = board.create('axis', [[0, 0], [0, 1]], { straightFirst : false, highlight : false, ticks : { minorTicks : 0, majorHeight : 15, label : { offset : [-15, 0 ], position : 'lrt', highlight : false } } }); xaxis.defaultTicks.ticksFunction = function() { return 1; }; yaxis.defaultTicks.ticksFunction = function() { return 5; }; var a_k = [1, 1]; var newbar = function(k) { var y; if(k-1 > a_k.length-1) { y = a_k[k-3] + a_k[k-2]; a_k.push(y); } else { y = a_k[k-1]; } return board.create('polygon', [[k-1/4, 0], [k+1/4, 0], [k+1/4, y], [k-1/4, y]] , { vertices : { visible : false }, borders : { strokeColor : 'black', strokeOpacity : .6, highlight : false }, fillColor : '#b2caeb', fixed : true, highlight : false }); } var bars = []; for(var i=0; i < 9; i++) { const k = i+1; board.create('text', [k-.1, -.6, function() { return 'a_{' + k + '}'; }], { useMathJax : true, fixed : true, strokeOpacity : 0.6 }); bars.push(newbar(k)); } board.fullUpdate(); })(); '; }], { useMathJax : true, strokeColor : '#2183de', fontSize : 13, fixed : true, highlight : false }); board.fullUpdate(); })(); '; }], { useMathJax : true, fixed : true, strokeOpacity : 0.6 }); board.create('point', [i, 1/i], { size : 4, name : '', strokeWidth : .5, strokeColor : 'black', face : 'diamond', fillColor : '#cf4490', fixed : true }); } board.fullUpdate(); })();

2. Series

Convergence


Convergence

If the sequence of partial sums (s_n) has a limit s\in \mathbb{R}, then the series of the sequence (a_k) converges and its sum is s. This is denoted by 
		a_1+a_2+\dots =\sum_{k=1}^{\infty} a_k = \lim_{n\to\infty}\underbrace{\sum_{k=1}^{n} a_k}_{=s_{n}} = s.

Indexing

The partial sums should be indexed in the same way as the sequence (a_k); e.g. the partial sums of a sequence (a_k)_{k=0}^{\infty} are s_0= a_0, s_1=a_0+a_1 etc.

The indexing of a series can be shifted without altering the series: \sum_{k=1}^{\infty} a_k =\sum_{k=0}^{\infty} a_{k+1} = \sum_{k=2}^{\infty} a_{k-1}.

In a concrete way: \sum_{k=1}^{\infty} \frac{1}{k^2}=1+\frac{1}{4}+\frac{1}{9}+\dots= \sum_{k=0}^{\infty} \frac{1}{(k+1)^2}

Interactivity.

Compute partial sums of the series \displaystyle\sum_{k=0}^{\infty}a_{k}

kth-element of the series: , start summation at

Divergence of a series

A series that does ot converge is divergent. This can happen in three different ways:

  1. the partial sums tend to infinity
  2. the partial sums tend to minus infinity
  3. the sequence of partial sums oscillates so that there is no limit.

In the case of a divergent series the symbol \displaystyle\sum_{k=1}^{\infty} a_k does not really mean anything (it isn't a number). We can then interpret it as the sequence of partial sums, which is always well-defined.

Basic results


Geometric series

A geometric series \sum_{k=0}^{\infty} aq^k converges if |q| (or a=0), and then its sum is \frac{a}{1-q}. If |q|\ge 1, then the series diverges.

Proof. The partial sums satisfy \sum_{k=0}^{n} aq^k =\frac{a(1-q^{n+1})}{1-q}, from which the claim follows.
\square

More generally \sum_{k=i}^{\infty} aq^k = \frac{aq^i}{1-q} = \frac{\text{1st term of the series}}{1-q},\text{ for } |q|

Example 1.

Calculate the sum of the series \sum_{k=1}^{\infty}\frac{3}{4^{k+1}}.

Solution. Since \frac{3}{4^{k+1}}  = \frac{3}{4}\cdot \left( \frac{1}{4}\right)^k, this is a geometric series. The sum is \frac{3}{4}\cdot \frac{1/4}{1-1/4} = \frac{1}{4}.

Rules of summation

Properties of convergent series:
  • \displaystyle{\sum_{k=1}^{\infty} (a_k+b_k) = \sum_{k=1}^{\infty} a_k + \sum_{k=1}^{\infty} b_k}
  • \displaystyle{\sum_{k=1}^{\infty} (c\, a_k) = c\sum_{k=1}^{\infty} a_k}, where c\in \mathbb{R} is a constant

Proof. These follow from the corresponding properties for limits of a sequence.
\square


Note: Compared to limits, there is no similar product-rule for series, because even for sums of two elements we have (a_1+a_2)(b_1+b_2) \neq a_1b_1 +a_2b_2. The correct generalization is the Cauchy product of two series, where also the cross terms are taken into account.

See e.g. https://en.wikipedia.org/wiki/Cauchy_product

Theorem 1.

If the series \displaystyle{\sum_{k=1}^{\infty} a_k} converges, then \displaystyle{\lim_{k\to \infty} a_k =0}.

Conversely: If \displaystyle{\lim_{k\to \infty} a_k \neq 0}, then the series \displaystyle{\sum_{k=1}^{\infty} a_k} diverges.

Proof.

If the sum of the series is s, then a_k=s_k-s_{k-1}\to s-s=0.
\square


Note: The property \lim_{k\to \infty} a_k = 0 cannot be used to justify the convergence of a series; cf. the following examples. This is one of the most common elementary mistakes many people do when studying series!

Example

Explore the convergence of the series \sum_{k=1}^{\infty} \frac{k}{k+1} = \frac{1}{2}+\frac{2}{3}+\frac{3}{4}+\dots

Solution. The limit of the general term of the series is \lim_{k\to\infty}\frac{k}{k+1} = 1. As this is different from zero, the series diverges.

Harmonic series

The harmonic series \sum_{k=1}^{\infty} \frac{1}{k} = 1+\frac{1}{2}+\frac{1}{3}+\dots diverges, although the limit of the general term a_k=1/k equals zero.

Proof.

This is a classical result first proven in the 14th century by Nicole Oresme after which a number of proofs using different approaches have been published. Here we present two different approaches for comparison.

i) An elementary proof by contradiction. Suppose, for the sake of contradiction, that the harmonic series converges i.e. there exists s\in\mathbb{R} such that s = \sum_{k=1}^{\infty}1/k. In this case 
			s = \left(\color{#4334eb}{1} + \color{#eb7134}{\frac{1}{2}}\right) + \left(\color{#4334eb}{\frac{1}{3}} + 
			\color{#eb7134}{\frac{1}{4}}\right) + \left(\color{#4334eb}{\frac{1}{5}} + \color{#eb7134}{\frac{1}{6}}\right) + \dots
			= \sum_{k=1}^{\infty}\left(\color{#4334eb}{\frac{1}{2k-1}} + \color{#eb7134}{\frac{1}{2k}}\right).
			Now, by direct comparison we get 
			\color{#4334eb}{\frac{1}{2k-1}} > \color{#eb7134}{\frac{1}{2k}} > 0, \text{ for all }k\ge 1~\Rightarrow~\sum_{k=1}^{\infty}\color{#4334eb}{\frac{1}{2k-1}} > \sum_{k=1}^{\infty}\color{#eb7134}{\frac{1}{2k}} = \frac{s}{2}
			hence following from the Properties of summation it follows that 
			s = \sum_{k=1}^{\infty}\color{#4334eb}{\frac{1}{2k-1}} + \sum_{k=1}^{\infty}\color{#eb7134}{\frac{1}{2k}} = \sum_{k=1}^{\infty}\color{#4334eb}{\frac{1}{2k-1}} + \frac{1}{2}\underbrace{\sum_{k=1}^{\infty}\frac{1}{k}}_{=s}.
			
			= \sum_{k=1}^{\infty}\color{#4334eb}{\frac{1}{2k-1}} + \frac{s}{2} > \sum_{k=1}^{\infty}\color{#eb7134}{\frac{1}{2k}} + \frac{s}{2} = \frac{s}{2} + \frac{s}{2} = s.
			But this implies that s>s, a contradiction. Therefore, the initial assumption that the harmonic series converges must be false and thus the series diverges.

\square


ii) Proof using integral: Below a histogram with heights 1/k lies the graph of the function f(x)=1/(x+1), so comparing areas we have \sum_{k=1}^{n} \frac{1}{k} \ge \int_0^n\frac{dx}{x+1} =\ln(n+1)\to\infty, as n\to\infty.
\square

Positive series

Summing a series is often difficult or even impossible in closed form, sometimes only a numerical approximation can be calculated. The first goal then is to find out whether a series is convergent or divergent.

A series \displaystyle{\sum_{k=1}^{\infty} p_k} is positive, if p_k > 0 for all k.

Convergence of positive series is quite straightforward:

Theorem 2.

A positive series converges if and only if the sequence of partial sums is bounded from above.

Why? Because the partial sums form an increasing sequence.

Example

Show that the partial sums of a superharmonic series \sum_{k=1}^{\infty}\frac{1}{k^2} satisfy s_n for all n, so the series converges.

Solution. This is based on the formula \frac{1}{k^2} < \frac{1}{k(k-1)} = \frac{1}{k-1}-\frac{1}{k}, for k\ge 2, as it implies that \sum_{k=1}^n\frac{1}{k^2} < 1+ \sum_{k=2}^n\frac{1}{k(k-1)} =2-\frac{1}{n}< 2 for all n\ge 2.

This can also be proven with integrals.


Leonhard Euler found out in 1735 that the sum is actually \pi^2/6. His proof was based on comparison of the series and product expansion of the sine function.

Absolute convergence


Definition

A series \displaystyle{\sum_{k=1}^{\infty} a_k} converges absolutely if the positive series \sum_{k=1}^{\infty} |a_k| converges.


Theorem 3.

An absolutely convergent series converges (in the usual sense) and \left| \sum_{k=1}^{\infty} a_k \right| \le \sum_{k=1}^{\infty} |a_k|.

This is a special case of the Comparison principle, see later.

Proof.

Suppose that \sum_k |a_k| converges. We study separately the positive and negative parts of \sum_k a_k: Let b_k=\max (a_k,0)\ge 0 \text{ and } c_k=-\min (a_k,0)\ge 0. Since b_k,c_k\le |a_k|, the positive series \sum b_k and \sum c_k converge by Theorem 2. Also, a_k=b_k-c_k, so \sum a_k converges as a difference of two convergent series.
\square

Example

Study the convergence of the alternating (= the signs alternate) series \sum_{k=1}^{\infty}\frac{(-1)^{k+1}}{k^2}=1-\frac{1}{4}+\frac{1}{9}-\dots

Solution. Since \displaystyle{\left| \frac{(-1)^{k+1}}{k^2}\right| =\frac{1}{k^2}} and the superharmonic series \sum_{k=1}^{\infty}\frac{1}{k^2} converges, then the original series is absolutely convergent. Therefore it also converges in the usual sense.

Alternating harmonic series

The usual convergence and absolute convergence are, however, different concepts:

Example

The alternating harmonic series \sum_{k=1}^{\infty}\frac{(-1)^{k+1}}{k} = 1-\frac{1}{2}+\frac{1}{3}-\frac{1}{4}+\dots converges, but not absolutely.

(Idea) Draw a graph of the partial sums (s_n) to get the idea that even and odd index partial sums s_{2n} and s_{2n+1} are monotone and converge to the same limit.


The sum of this series is \ln 2, which can be derived by integrating the formula of a geometric series.

First 100 partial sums of the alternatig harmonic series;
points are joined by line segments for visualization purposes

Convergence tests


Comparison test

The preceeding results generalize to the following:

Theorem 4.

(Majorant) If |a_k|\le p_k for all k and \sum_{k=1}^{\infty} p_k converges, then also \sum_{k=1}^{\infty} a_k converges.

(Minorant) If 0\le p_k \le a_k for all k and \sum p_k diverges, then also \sum a_k diverges.

Proof.

Proof for Majorant. Since a_k=|a_k|-(|a_k|-a_k) and 0\le |a_k|-a_k \le 2|a_k|, then \sum a_k is convergent as a difference of two convergent positive series. Here we use the elementary convergence property (Theorem 2.) for positive series; this is not a circular reasoning!

Proof for Minorant. It follows from the assumptions that the partial sums of \sum a_k tend to infinity, and the series is divergent.
\square

Example

Study the convergence of 
		\sum_{k=1}^{\infty} \frac{1}{1+k^3} \ \text{ ja }\
		\sum_{k=1}^{\infty} \frac{1}{\sqrt{k}}.

Solution. Since 0 for all k\in \mathbb{N}, the first series is convergent by the majorant principle.

On the other hand, \displaystyle{\frac{1}{\sqrt{k}}\ge \frac{1}{k}} for all k\in\mathbb{N}, so the second series has a divergent harmonic series as a minorant. The latter series is thus divergent.

Ratio test

In practice, one of the best ways to study convergence/divergence of a series is the so-called ratio test, where the terms of the sequence are compared to a suitable geometric series:

Theorem 5a.

Suppose that there is a constant 0< Q < 1 so that 		\left| \frac{a_{k+1}}{a_k} \right| \le Q starting from some index k\ge k_0.

Then the series \sum a_k converges (and the rate of convergence is comparable to the geometric series \sum Q^k, or is even higher).

Proof.

We may assume that the inequality is valid for all indices k, because the initial part has no effect on the convergence (although it has an effect to the sum!).

This now implies that |a_{k}|\le Q|a_{k-1}|\le Q^2|a_{k-2}|\le \dots\le Q^k|a_0|, so the series has a convergent geometric majorant.
\square

Limit form of ratio test

Theorem 5b.

Suppose that the limit \lim_{k\to \infty}  \left| \frac{a_{k+1}}{a_k} \right| = q exists. Then the series \sum a_k 
		\begin{cases}\text{converges} & \text{ if } 0\le q< 1,\\
		\text{diverges} & \text{ if } q > 1,\\
		\text{nay be convergent or divergent} & \text{ if } q=1.
		\end{cases}


(Idea) For a geometric series the ratio of two consecutive terms is exactly q. According to the ratio test, the convergence of some other series can also be investigated in a similar way, when the exact ratio q is replaced by the above limit.

Proof.

In the formal definition of a limit \varepsilon =(1-q)/2>0. Thus starting from some index k\ge k_{\varepsilon} we have  |a_{k+1}/a_k| < q + \varepsilon = (q+1)/2 = Q < 1, and the claim follows from Theorem 4.


In the case q>1 the general term of the series does not go to zero, so the series diverges.


The last case q=1 does not give any information.

This case occurs for the harmonic series (a_k=1/k, divergent!) and superharmonic (a_k=1/k^2, convergent!) series. In these cases the convergence or divergence must be settled in some other way, as we did before.
\square

Example

Is the series \sum_{k=1}^{\infty}\frac{(-1)^{k+1}k}{2^k}= \frac{1}{2}-\frac{2}{4}+\frac{3}{8}-\dots convergent?

Solution. Here a_k=(-1)^{k+1}k/2^k, so 
		\left| \frac{a_{k+1}}{a_k}\right| 
		= \left| \frac{(-1)^{k+2}(k+1)/2^{k+1}}{(-1)^{k+1}k/2^k}\right|
		=\frac{k+1}{2k} =\frac{1}{2}+\frac{1}{2k}\to \frac{1}{2} < 1,
		when k\to\infty. By the ratio test the series is convergent.

'); }],{ useMathJax : true, fontSize : 14, visible : false }); /*var txt2 = board.create('text', [15, 1.8, function() { return 'k=' + (series.dataX.length-1) + ': value = ' + series.dataY[series.dataY.length - 1]; }], {strokeColor: 'black'});*/ var TO; var approx = function() { series_add(); board.update(); if (series.dataX.length <= 50) { TO = setTimeout(approx, 500); } }; var a_k; var start_approx = function() { txt.setAttribute( { visible : true }); board.removeObject(dataPoints); dataPoints = []; series.dataX = []; series.dataY = []; var txtraw = document.getElementById('input').value; a_k = board.jc.snippet(txtraw, true, 'k', true); k = parseInt(document.getElementById('startval').value); start = k; approx(); } var clear_all = function() { clearTimeout(TO); }; document.getElementById('startSeq').onclick = start_approx; document.getElementById('stopSeq').onclick = clear_all; })();

3. Continuity

In this section we define a limit of a function f\colon S\to \mathbb{R} at a point x_0. It is assumed that the reader is already familiar with limit of a sequence, the real line and the general concept of a function of one real variable.

Limit of a function


For a subset of real numbers, denoted by S, assume that x_0 is such point that there is a sequence of points (x_k)\in S such that x_k\to x_0 as k\to \infty. Here the set S is often the set of all real numbers, but sometimes an interval (open or closed).

Example 1.

Note that it is not necessary for x_0 to be in S. For example, the sequence x_k = 1/k\to 0 as k\to \infty in S=]0,2[, and x_k\in S for all k=1,2,\ldots but 0 is not in S.

Limit of a function

We consider a function f defined in the set S. Then we define the limit of the function f\colon S\to \mathbb{R} at x_0 as follows.

Definition 1: Limit of a function

Suppose that S\subset \mathbb{R} and f\colon S\to \mathbb{R} is a function. Then we say that f has a limit y_{0} at x_{0}, and write \lim_{x \to x_{0}}f(x)=y_{0}, if, f(x_{k})\to y_{0} as k\to \infty for every sequence (x_{k}) in S\setminus\{x_0\}, such that x_{k}\to x_{0} as k\to \infty.

Example 2.

The function f\colon \mathbb{R} \to \mathbb{R} defined by f(x)=x^2 has a limit 0 at the point x=0.

Function y=x^2.

Example 3.

The function g\colon\mathbb{R}\to \mathbb{R} defined by g(x)= \left\{\begin{array}{rl}0 & \text{ for }x does not have a limit at the point x=0. To formally prove this, take sequences (x_k), (y_k) defined by x_k=1/k and y_k=-1/k for k=1,2,\ldots. Then the both sequences are in S=\mathbb{R}, but f(x_k)=1 and f(y_k)=0 for any k.

Function g(x)= \left\{\begin{array}{rl}0 & \text{ for }x

Example 4.

The function f(x)=x \sin(1/x), x>0 does have the limit 0 at 0.

Function y=x\sin(1/x) for x>0.

Example 5.

The function g(x)= \sin(1/x), x>0 does not have a limit at 0.

Function y=\sin(1/x) for x>0.

One-sided limits

An important property of limits is that they are always unique. That is, if \lim_{x\to x_0} f(x)=a and \lim_{x\to x_0} f(x)=b, then a=b. Although a function may have only one limit at a given point, it is sometimes useful to study the behavior of the function when x_k approaches the point x_0 from the left or the right side. These limits are called the left and the right limit of the function f at x_0, respectively.

Definition 2: One-sided limits

Suppose S is a set in \mathbb{R} and f is a function defined on the set S\setminus\{x_0\}. Then we say that f has a left limit y_{0} at x_{0}, and write \lim_{x \to x_{0}-}f(x)=y_{0}, if, f(x_{k})\to y_{0} as k\to \infty for every sequence (x_{k}) in the set S\cap ]-\infty,x_0[ =\{ x\in S : x < x_0 \}, such that x_{k}\to x_{0} as k\to \infty.

Similarly, we say that f has a right limit y_{0} at x_{0}, and write \lim_{x \to x_{0}+}f(x)=y_{0}, if, f(x_{k})\to y_{0} as k\to \infty for every sequence (x_{k}) in the set S\cap ]x_0,\infty[ =\{ x\in S : x_0 < x \}, such that x_{k}\to x_{0} as k\to \infty.

Theorem 1: Limit of a function

A function f\colon S\to \mathbb{R} has a limit y_0 at the point x_0 if and only if \lim_{x \to x_{0}-}f(x)= \lim_{x \to x_{0}+}f(x)=y_{0}.

Example 6.

The sign function \mathrm{sgn}(x)= \frac{x}{|x|} is defined on S= \mathbb{R}\setminus 0. Its left and right limits at 0 are \lim_{x\to 0-} \mathrm{sgn}(x)= -1,\qquad
	\lim_{x\to 0+} \mathrm{sgn}(x)= 1. However, the function \mathrm{sgn}(x) does not have a limit at 0.

Function y = \frac{x}{|x|}.

Example 7.

Function f: \mathbb{R}\setminus 0 \to \mathbb{R} f(x) = \frac{1}{x} does not have one-sided limits at 0.

Limit rules

The following limit rules are immediately obtained from the definition and basic algebra of real numbers.

Theorem 2: Limit rules

Let c\in \mathbb{R}, \lim_{x\to x_{0}} f(x)=a and \lim_{x\to x_{0}} g(x)=b. Then

  1. \lim_{x\to x_{0}} (cf)(x)=ca,
  2. \lim_{x\to x_{0}} (f+g)(x)=a+b,
  3. \lim_{x\to x_{0}} (fg)(x)=ab,
  4. \lim_{x\to x_{0}} (f/g)(x)=a/b \ (\text{if} \ b \neq 0).
Example 8.

Finding limits by calculating f(x_0):

a) \lim_{x\to 2}(5x-3)=10-3=7.

b) \lim_{x\to -2}\frac{3x+2}{x+5} = \frac{-6+2}{-2+5}=-\frac{4}{3}.

c) \lim_{x\to 2} \frac{x^2-4}{x-2} = \lim_{x\to 2} \frac{(x+2)(x-2)}{x-2} = \lim_{x\to 2}(x+2) = 4.

Limits and continuity


In this section, we define continuity of the function. The intutive idea behind continuity is that the graph of a continuous function is a connected curve. However, this is not sufficient as a mathematical definition for several reasons. For example, by using this definition, one cannot easily decide if \tan(x) is a continuous function or not.

For continuity of a function f at a given point x_0, it is required that:

  1. f(x_0) is defined,

  2. \lim_{x \to x_0} f(x) exists (and is finite),

  3. \lim_{x \to x_0} f(x) = f(x_0).

In other words:

Definition 2: Continuity

A function f\colon S\to \mathbb{R} is continuous at a point x_{0}\in S, if \lim_{x\to x_{0}}f(x)=f(x_{0}). A function f\colon S\to \mathbb{R} is continuous, if it is continuous at every point x_{0}\in S.

Example 1.

Let c\in \mathbb{R}. Functions f,g,h defined by f(x)=c, g(x)=x, h(x)=|x| are continuous at every point x\in \mathbb{R}.

Why? If x_{k}\to x_{0}, then f(x_{k})=c and \lim_{k\to \infty}f(x_k)= c=f(x_{0}). For g, we have g(x_{k})=x_{k} and hence, \lim_{k\to\infty} g(x_k)=x_{0}=g(x_{0}). Similarly, h(x_{k})=|x_{k}| and \lim_{k\to\infty}h(x_k)= |x_{0}|=h(x_{0}).

Continuous functions y=c, y=x and y=|x|.

Example 2.

Let x_{0}\in \mathbb{R}. We define a function f\colon\mathbb{R}\to \mathbb{R} by f(x)= \left\{\begin{array}{rl}2 & \text{ for }x \lt x_{0}, \\
		3 & \text{ for }x\geq x_{0}.\end{array}\right. Then \lim_{x \to x_{0}^{-}}f(x)=2,\text{ and } \lim_{x \to x_{0}^{+}}f(x)=3. Therefore f is not continuous at the point x_{0}.

Some basic properties of continuous functions of one real variable are given next. From the limit rules (Theorem 2) we obtain:

Theorem 3.

The sum, the product and the difference of continuous functions are continuous. Then, in particular, polynomials are continuous functions. If f and g are polynomials and g(x_{0})\neq 0, then f/g is continuous at a point x_{0}.

A composition of continuous functions is continuous if it is defined:

Theorem 4.

Let f\colon \mathbb{R}\to\mathbb{R} and g\colon \mathbb{R}\to \mathbb{R}. Suppose that f is continuous at a point x_{0} and g is continuous at f(x_{0}). Then g\circ f\colon \mathbb{R}\to \mathbb{R} is continuous at a point x_{0}.

Proof.

Note. If f is continuous, then |f| is continuous.

Why?

Write g(x):=|x|. Then (g\circ f)(x)=|f(x)|.

Note. If f and g are continuous, then \max (f,g) and \min (f,g) are continuous. (Here \max (f,g)(x):=\max \{f(x),g(x)\}.)

Why?

Write \begin{cases}(a+b)+|a-b|=2\max(a,b), \\
		(a+b)-|a-b|=2\min(a,b). \end{cases}

\text{Function }f(x)= \left\{\begin{array}{rl}2 & \text{ for }x\lt x_{0}, \\
		3 & \text{ for }x\geq x_{0}. \end{array}\right.

'}], { strokeColor : colors[0], fontSize : 16, visible : true }); l[1] = board.create('text', [-4, 3, function() { return 'y=x'}], { strokeColor : colors[1], fontSize : 16, visible : false }); l[2] = board.create('text', [-4, 3, function() { return 'y=|x|'}], { strokeColor : colors[2], fontSize : 16, visible : false }); g[0] = board.create('functiongraph', [f0, -6, 6], { visible : true, strokeWidth : 1.5, strokeColor : colors[0], highlight : false }); g[1] = board.create('functiongraph', [f1, -6, 6], { visible : false, strokeWidth : 1.5, strokeColor : colors[1], highlight : false }); g[2] = board.create('functiongraph', [f2, -6, 6], { visible : false, strokeWidth : 1.5, strokeColor : colors[2], highlight : false }); var currentGraph = g[0]; var currentLabel = l[0]; select.on('drag', function() { currentGraph.setAttribute({ visible : false }); currentLabel.setAttribute({ visible : false }); currentGraph = g[select.Value()]; currentLabel = l[select.Value()]; currentGraph.setAttribute({ visible : true }); currentLabel.setAttribute({ visible : true, useMathJax : true }); select.setAttribute({ fillColor : colors[select.Value()] }); board.update(); }); board.fullUpdate(); })(); /* Example 2. */ (function() { var board = JXG.JSXGraph.initBoard('jxgbox16', { boundingbox : [-3.5, 4.5, 3.5, -1.25], showcopyright : false, shownavigation : false}); var xaxis = board.create('axis', [[0, 0], [1, 0]]); xaxis.removeAllTicks(); var yaxis = board.create('axis', [[0, 0], [0, 1]], { drawZero : true, ticks : { majorHeight : 5, minorTicks : 0, ticksDistance : 1.0 } }); yaxis.defaultTicks.ticksFunction = function() { return 1; }; var xtick = board.create('segment', [[1, .05],[1, -.1]], { strokeWidth : 1, strokeColor : 'black', strokeOpacity : .4, highlight : false }); var f = function(x) { return 2; } var g = function(x) { return 3; } board.create('functiongraph', [f, -3.5, 1], { strokeColor : 'black', strokeWidth : 2, highlight : false }); board.create('functiongraph', [g, 1, 3.5], { strokeColor : 'black', strokeWidth : 2, highlight : false }); board.create('point', [1, f(1)], { name : '', fillColor : 'white', strokeColor : 'black', strokeWidth : .5, size : 2, fixed : true, showInfobox : false }); board.create('point', [1, g(1)], { name : '', fillColor : 'black', strokeColor : 'black', strokeWidth : .5, size : 2, fixed : true, showInfobox : false }); var x0 = board.create('text', [1, 0, function() { return 'x_{0}'; }], { useMathJax : true }); board.fullUpdate(); })();

Delta-epsilon definition

The so-called (\varepsilon,\delta)-definition for continuity is given next. The basic idea behind this test is that, for a function f continuous at x_0, the values of f(x) should get closer to f(x_0) as x gets closer to x_0.

This is the standard definition of continuity in mathematics, because it also works for more general classes of functions than ones on this course, but it not used in high-school mathematics. This important definition will be studied in-depth in Analysis 1 / Mathematics 1.

(\varepsilon,\delta)-test:

Theorem 5: (\varepsilon,\delta)-definition

Let f: S\to \mathbb{R}. Then the following conditions are equivalent:

  1. \lim_{x\to x_0} f(x)= y_0,
  2. For all \varepsilon> 0 there exists \delta >0 such that if 0 < |x-x_0| < \delta, then |f(x) - y_0| for all x\in S.

Proof.

Example 3.

From Theorem 3 we already know that the function f: \mathbb{R} \to \mathbb{R} defined by f(x) = 4x is continuous. We can also use the (\varepsilon,\delta)-definition to prove this.

Proof. Let x_0 \in \mathbb{R} and \varepsilon > 0. Now |f(x) - f(x_0)| = |4x - 4x_0| = 4|x - x_0| < \varepsilon, when |x - x_0| < \delta \text{, where } \delta = \frac{\varepsilon}{4}.

So for all \varepsilon > 0 there exists \delta > 0 such that if |x - x_0| < \delta, then |f(x) - f(x_0)| < \varepsilon for all x \in \mathbb{R}. Thus by Theorem 5 \lim_{x \to x_0} f(x) = f(x_0) for all x_0 \in \mathbb{R} and by definition this means that the function f: \mathbb{R} \to \mathbb{R} is continuous.
\square

Interactivity. (\varepsilon, \delta) in example 3.

'} ], { strokeColor : 'black', fontSize : 12, fixed : true }) var s = board.create('slider', [[-3, 1], [-1.25, 1], [0.05, 2, 2]], { name : '\\varepsilon', unitLabel : '', fillColor : '#bd4444', snapWidth : 0.01, label : { useMathJax : true, strokeColor : '#bd4444', offset : [0, 25] }}); var x1 = board.create('glider', [1/2, 0, xaxis], { name: 'x_{0}', strokeColor : 'black', strokeWidth : .5, fillColor : '#446abd', showinfobox : false, label : { useMathJax : true, strokeColor : '#446abd', offset : [5, 25] } }); /*board.suspendUpdate();*/ var y1 = board.create('point', [0, function(){return f(x1.X());}], { name : 'f(x_{0})', size : 2, strokeColor : 'black', strokeWidth : .5, fillColor : '#bd4444', showinfobox : false, highlight : false, label : { useMathJax : true, strokeColor : '#bd4444', offset : [5, 25] }}); var y2 = board.create('point', [0, function(){return f(x1.X())-s.Value();}], { visible : false }); var y3 = board.create('point', [0, function(){return f(x1.X())+s.Value();}], { visible : false }); var z1 = board.create('point', [function(){return y1.Y()/4;}, function(){return y1.Y();}], { visible : false }); var z2 = board.create('point', [function(){return y2.Y()/4;}, function(){return y2.Y();}], { visible : false }); var z3 = board.create('point', [function(){return y3.Y()/4;}, function(){return y3.Y();}], { visible : false }); var v1 = board.create('segment', [z1, y1], { strokeColor : '#bd4444', strokeWidth : 1, highlight : false }); var v2 = board.create('line', [z2, y2], { strokeColor : '#bd4444', dash : 2, strokeWidth : 1, highlight : false }); var v3 = board.create('line', [z3, y3], { strokeColor: '#bd4444', dash : 2, strokeWidth : 1, highlight : false }); var epsilon = board.create('polygon', [function() { return [-.5, y2.Y()]; }, function() { return [2.5, y2.Y()]; }, function() { return [2.5, y3.Y()]; }, function() { return [-.5, y3.Y()]; }], { fillColor : '#bd4444', fillOpacity : .3, highlight : false, vertices : { visible : false }, borders : { visible : false }}); var h1 = board.create('segment', [function() { return x1; }, function() { return z1; }], { strokeColor : '#446abd', strokeWidth : 1, highlight : false }); var h2 = board.create('segment', [function() { return [z2.X(), 0]; }, function() { return [z2.X(), 8]}], { strokeColor : '#446abd', dash : 2, strokeWidth : 1, highlight : false }); var h3 = board.create('segment', [function() { return [z3.X(), 0]; }, function() { return [z3.X(), 8]}], { strokeColor : '#446abd', dash : 2, strokeWidth : 1, highlight : false }); var delta = board.create('polygon', [h2.point1, h2.point2, h3.point2, h3.point1], { fillColor : '#446abd', fillOpacity : .3, highlight : false, vertices : { visible : false }, borders : { visible : false }}); var txt = board.create('text', [-2.5, .7, function() { return '\\delta = \\epsilon/4 = ' + (s.Value()/4).toFixed(3) + ''; }], { strokeColor : '#446abd', useMathJax : true, fixed : true }); /*board.unsuspendUpdate();*/ board.fullUpdate(); })();
Example 4.

Let x_{0}\in \mathbb{R}. We define a function f\colon\mathbb{R}\to \mathbb{R} by f(x)= \left\{\begin{array}{rl}2 & \text{ for }x \lt x_{0}, \\
		3 & \text{ for }x \geq x_{0}.\end{array}\right. In Example 2 we saw that this function is not continuous at the point x_0. To prove this using the (\varepsilon,\delta)-test, we need to find some \varepsilon > 0 and some x_\delta \in \mathbb{R} such that for all \delta > 0, |x_\delta - x_0| < \delta, but |f(x_\delta) - f(x_0)| > \varepsilon.

Proof. Let \delta > 0 and \varepsilon = 1/2. By choosing x_\delta = x_0 - \delta /2, we have 0 < |x_\delta-x_0| = |x_0 - \frac{\delta}{2} + x_0| = \frac{\delta}{2} < \delta, and |f(x_\delta) - f(x_0)| = |2 - 3| = 1 > \varepsilon. Therefore by Theorem 5 f is not continuous at the point x_{0}.
\square

Interactivity. (\varepsilon, \delta) in example 4.

', fillColor : '#446abd', label : { useMathJax : true, strokeColor : '#446abd' }}); var x1 = board.create('glider', [0, 0, xaxis], { name:'x_{0}', strokeWidth : .3, strokeColor : 'black', fillColor : '#446abd', showinfobox : false, highlight : false, fixed : true, label : { useMathJax : true, offset : [5, 25], strokeColor : '#446abd' }}); board.suspendUpdate(); var x2 = board.create('point', [ function(){return x1.X()-s.Value();}, 0], { visible : false }); var x3 = board.create('point', [function(){return x1.X()+s.Value();},0], { visible : false }); var y1 = board.create('point', [-2, function() { return f(x1.X()); }], { size : 2, name: 'f(x_{0})', strokeWidth : .3, strokeColor : 'black', fillColor : '#446abd', showinfobox : false, highlight : false, label : { useMathJax : true, offset : [5, 25], strokeColor : '#446abd' } }); var yd = board.create('point', [-2, function() { return f(x2.X()); }], { size : 2, name: 'f(x_{\\delta})', strokeWidth : .3, strokeColor : 'black', fillColor : '#bd4444', showinfobox : false, highlight : false, label : { useMathJax : true, offset : [5, 25], strokeColor : '#bd4444' } }); /* doesn't really do anything atm... */ var dist = function() { if (Math.abs(f(x1)-f(x2.X())) != 0 || Math.abs(f(x1)-f(x3.X())) != 0) { return .5; } else { return 0; } }; var y2 = board.create('point', [0, function() { return y1.Y()-dist(); }], { visible : false }); var y3 = board.create('point', [0, function() { return y1.Y()+dist(); }], { visible : false }); var endpoint1 = board.create('point', [0, 2], { name : '', fixed : true, size : 2, fillColor : 'white', strokeWidth : .5, strokeColor : 'black', showinfobox : false }); var endpoint2 = board.create('point', [0, 3], { name : '', fixed : true, size : 2, fillColor : 'black', strokeWidth : .5, strokeColor : 'black', showinfobox : false }); var v1 = board.create('segment', [x1, function() { return [x1.X(), y1.Y()]; }], { strokeColor : '#446abd', strokeWidth : 1, highlight : false }); var v2 = board.create('line', [x2, function() { return [x2.X(), x2.Y()+1]; }], { strokeColor : '#446abd', dash : 2, strokeWidth : 1, highlight : false }); var v3 = board.create('line', [x3, function() { return [x3.X(), x3.Y()+1]; }], { strokeColor : '#446abd', dash : 2, strokeWidth : 1, highlight : false }); var delta = board.create('polygon', [function() { return [x2.X(), -5]; }, function() { return [x2.X(), 6]; }, function() { return [x3.X(), 6]; }, function() { return [x3.X(), -5]; }], { highlight : false, fixed : true, vertices : { visible : false }, borders : { visible : false }, fillColor : '#446abd', fillOpacity : .2 }); var h1 = board.create('segment', [y1, function() { return [x1.X(), y1.Y()]; }], { strokeColor : '#446abd', strokeWidth : 1, highlight : false }); var h2 = board.create('line', [function() { return y2; }, function() { return [y2.X()+1, y2.Y()]; }], { strokeColor : '#bd4444', dash : 2, strokeWidth : 1, highlight : false }); var h3 = board.create('line', [function() { return y3; }, function() { return [y3.X()+1, y3.Y()]; }], { strokeColor : '#bd4444', dash : 2, strokeWidth : 1, highlight : false }); var epsilon = board.create('polygon', [[-5, 3.5],[10, 3.5],[10, 2.5], [-5, 2.5]], { highlight : false, fixed : true, vertices : { visible : false }, borders : { visible : false }, fillColor : '#bd4444', fillOpacity : .2 }); var txt = board.create('text', [4.2, -1.5, function() { return '\\epsilon = 1/2'/*Math.max(Math.abs(y2.Y() - y1.Y()), Math.abs(y1.Y() - y3.Y())).toFixed(2)*/; }], { strokeColor: '#bd4444', useMathJax : true, fixed : true }); board.unsuspendUpdate(); board.fullUpdate(); })();

Properties of continuous functions


This section contains some fundamental properties of continuous functions. We start with the Intermediate Value Theorem for continuous functions, also known as Bolzano's Theorem. This theorem states that a function that is continuous on a given (closed) real interval, attains all values between its values at endpoints of the interval. Intuitively, this follows from the fact that the graph of a function defined on a real interval is a continuous curve.

Theorem 6: Intermediate Value Theorem

If f\colon [a,b]\to \mathbb{R} is continuous and f(a) \lt s \lt f(b), then there is at least one c\in ]a,b[ such that f(c)=s.

Proof.

Interactivity. Theorem 6.

The Intermediate Value Theorem.

Example 1.

Let function f:\mathbb{R} \to \mathbb{R}, where f(x) = x^5 - 3x - 1. Show that there is at least one c \in \mathbb{R} such that f(c) = 0.

Solution. As a polynomial function, f is continuous. And because f(1) = 1^5 - 3 \cdot 1 - 1 = -3 < 0 and f(-1) = (-1)^5 - 3 \cdot (-1) - 1 = 1 > 0, by the Intermediate Value Theorem there is at least one c \in ]-1, 1[ such that f(c) = 0.

Function f(x) = x^5 - 3x - 1.

Example 2.

Let f(x)=x^3-x=x(x^2-1)=x(x-1)(x-1).

By the Intermediate Value Theorem we have f(x) for x or 0 \lt x \lt 1. Similarly, f(x)>0 for -1 \lt x \lt 0 or 1 \lt x, because:

  1. f(x)=0 if and only if x=0 or x=\pm 1, and
  2. f(-2)0, f(1/2) and f(2)>0.

Function f(x) = x^3 - x.

'; }], { anchor : l, strokeColor : 'red', fontSize : 13, fixed : true }); var fleft = board.create('text', [.2, 4.4, function() { return 'f(a)'}], { strokeColor : 'black', fontSize : 13, fixed : true }); var fright = board.create('text', [9.1, 7.6, function() { return 'f(b)'}], { strokeColor : 'black', fontSize : 13, fixed : true }); var xleft = board.create('text', [1.1, -.1, function() { return 'a' }], { strokeColor : 'black', fontSize : 13, fixed : true }); var xright = board.create('text', [9.1, -.1, function() { return 'b' }], { strokeColor : 'black', fontSize : 13, fixed : true }); l.on('drag', function() { if(l.point1.Y() >= 7) { l.point1.moveTo([0, 7]); l.point2.moveTo([1, 7]); } else if(l.point1.Y() <= 4) { l.point1.moveTo([0, 4]); l.point2.moveTo([1, 4]); } }); var intersections = []; intersections[0] = board.create('intersection', [l, g, 0], { name : 'f(c)=s', showinfobox : false, label : { fontSize : 13, offset : [0, -15], strokeColor : 'blue', strokeWidth : .5} }); intersections[1] = board.create('intersection', [l, g, 1], { name : 'f(c)=s', showinfobox : false, label : { fontSize : 13, offset : [8, 22], strokeColor : 'blue', strokeWidth : .5} }); intersections[2] = board.create('intersection', [l, g, 2], { name : 'f(c)=s', showinfobox : false, label : { fontSize : 13, offset : [0, -15], strokeColor : 'blue', strokeWidth : .5} }); board.unsuspendUpdate(); })(); /* Example 1. */ (function() { var board = JXG.JSXGraph.initBoard('jxgbox17', { boundingbox : [-3.5, 2.5, 3.5, -3.5], showcopyright : false, shownavigation : false}); var xaxis = board.create('axis', [[0, 0], [1, 0]], { ticks : { majorHeight : 7, minorTicks : 0, drawZero : true } }); var yaxis = board.create('axis', [[0, 0], [0, 1]], { ticks : { majorHeight : 7, minorTicks : 0, drawZero : true } }); xaxis.defaultTicks.ticksFunction = function() { return 1; }; yaxis.defaultTicks.ticksFunction = function() { return 1; }; var f = function(x) { return (Math.pow(x,5)-3*x-1); } board.create('functiongraph', [f, -3.5, 3.5], { strokeColor : 'black', strokeWidth : 2, highlight : false }); board.fullUpdate(); })(); /* Example 2. */ (function() { var board = JXG.JSXGraph.initBoard('jxgbox18', { boundingbox : [-3.2, 2.9, 3.2, -2.9], showcopyright : false, shownavigation : false}); var xaxis = board.create('axis', [[0, 0], [1, 0]], { ticks : { majorHeight : 7, minorTicks : 0, drawZero : true } }); var yaxis = board.create('axis', [[0, 0], [0, 1]], { ticks : { majorHeight : 7, minorTicks : 0, drawZero : true } }); xaxis.defaultTicks.ticksFunction = function() { return 1; }; yaxis.defaultTicks.ticksFunction = function() { return 1; }; var f = function(x) { return (x*x*x-x); } board.create('functiongraph', [f, -3.2, -1], { strokeColor : 'blue', strokeWidth : 2, highlight : false }); board.create('functiongraph', [f, -1, 0], { strokeColor : 'red', strokeWidth : 2, highlight : false }); board.create('functiongraph', [f, 0, 1], { strokeColor : 'blue', strokeWidth : 2, highlight : false }); board.create('functiongraph', [f, 1, 3.2], { strokeColor : 'red', strokeWidth : 2, highlight : false }); board.fullUpdate(); })();

Next we prove that a continuous function defined on a closed real interval is necessarily bounded. For this result, it is important that the interval is closed. A counter example for an open interval is given after the next theorem.

Theorem 7.

Let f\colon [a,b]\to \mathbb{R} be continuous. Then f is bounded.

Proof.

Note. If f\colon ]a,b[\to \mathbb{R} is continuous, it can be unbounded.

Example 4.

Let f\colon ]0,1]\to \mathbb{R}, where f(x)=1/x. Now \lim_{x\to 0+}f(x)=\infty.

Theorem 8.

Let f\colon [a,b]\to \mathbb{R} be continuous. Then there exist points c,d\in [a,b] such that f(c)\leq f(x)\leq f(d) for all x\in [a,b], i.e. f(c) is minimum and f(d) is maximum of f on the interval [a,b].

Proof.

Function f(x) = 1/x for x > 0.

Example 5.

Let f:[-1,2] \to \mathbb{R}, where f(x) = -x^3 - x + 3. The domain of the function is [-1,2]. To determine the range of the function, we first notice that the function is decreasing. We will now show this.

Let x_1 < x_2. Then x_{1}^3 < x_{2}^3 and -x_{1}^3 > -x_{2}^3.

Because x_1 < x_2, -x_1^3-x_1 > -x_2^3 -x_2 and -x_1^3-x_1 +3 > -x_2^3 -x_2 +3. Thus, if x_1 < x_2 then f(x_1) > f(x_2), which means that the function f is decreasing.

We know that a decreasing function has its minimum value in the right endpoint of the interval. Thus, the minimum value of f:[-1,2] \to \mathbb{R} is f(2) = -2^3 - 2 + 3 = -7. Respectively, a decreasing function has it's maximum value in the left endpoint of the interval and so the maximum value of f:[-1,2] \to \mathbb{R} is f(-1) = -(-1)^3 - (-1) + 3 = 5.

As a polynomial function, f is continuous and it therefore has all the values between it's minimum and maximum values. Hence, the range of f is [-7, 5].

Function -x^3 - x + 3 for [-1, 2].

Example 6.

Suppose that f is a polynomial. Then f is continuous on \mathbb{R} and, by Theorem 7, f is bounded on every closed interval [a,b], a \lt b. Furthermore, by Theorem 3, f must have minimum and maximum values on [a,b].

Note. Theorem 8 is connected to the Intermediate Value Theorem in the following way:

If f\colon [a,b]\to \mathbb{R} be continuous, then there exist points x_1,x_2\in [a,b] such that f([a,b])=[f(x_1),f(x_2)].

4. Derivative

Derivative


The definition of the derivative of a function is given next. We start with an example illustrating the idea behind the formal definition.

Example 0.

The graph below shows how far a cyclist gets from his starting point.


a) Look at the red line. We can see that in three hours, the cyclist moved 20km. The average speed of the whole trip is 6.6 km/h.
b) Now look at the green line. We can see that during the third hour the cyclist moved 10km further. That makes the average speed of that time interval 10 km/h.
Notice that the slope of the red line is 20/3 \approx 6.6 and that the slope of the blue line is 10. These are the same values as the corresponding average speeds.
c) Look at the blue line. It is the tangent of the curve at the point x=2h. Using the same principle as with average speeds, we conclude that after two hours of the departure, the speed of the cyclist was 30/2 km/h = 15 km/h.

Now we will proceed to the general definition:

Definition: Derivative

Let (a,b)\subset \mathbb{R}. The derivative of function f\colon (a,b)\to \mathbb{R} at the point x_0\in (a,b) is f'(x_0):=\lim_{h\to 0} \frac{f(x_0+h)-f(x_0)}{h}. If f'(x_0) exists, then f is said to be differentiable at the point x_0.

Note: Since x = x_0+h, then h=x-x_0, and thus the definition can also be written in the form f'(x_0):=\lim_{x\to x_0} \frac{f(x)-f(x_0)}{x-x_0}.

The derivative can be denoted in different ways:  f'(x_0)=Df(x_0) =\left. \frac{df}{dx}\right|_{x=x_0}, \ \ f'=Df =\frac{df}{dx}.

Interpretation. Consider the curve y = f(x). Now if we draw a line through the points (x_0,f(x_0)) and (x_0+h, f(x_0+h)), we see that the slope of this line is \frac{f(x_0+h)-f(x_0)}{x_0+h-x_0} = \frac{f(x_0+h)-f(x_0)}{h}. When h \to 0, the line intersects with the curve y = f(x) only in the point (x_0, f(x_0)). This line is the tangent of the curve y=f(x) at the point (x_0,f(x_0)) and its slope is \lim_{h\to 0} \frac{f(x_0+h)-f(x_0)}{h}, which is the derivative of the function f at  x_0. Hence, the tangent is given by the equation y=f(x_0)+f'(x_0)(x-x_0).

Interactivity. Move the point of intersection and observe changes on the tangent line of the curve.

Example 1.

Let f\colon \mathbb{R} \to \mathbb{R} be the function f(x) = x^3 + 1. The derivative of f at x_0 = 1 is \begin{aligned}f'(1) &=\lim_{h \to 0} \frac{f(1+h)-f(1)}{h} \\ &=\lim_{h \to 0} \frac{(1+h)^3 + 1 - 1^3 - 1}{h} \\ &=\lim_{h \to 0} \frac{1+3h+3h^2+h^3-1}{h} \\ &=\lim_{h \to 0} \frac{h(3+3h+h^2)}{h} \\ &=\lim_{h \to 0} 3+3h+h^2 \\ &= 3. \end{aligned}

Function  x^3 + 1 and its tangent at the point 1.

Example 2.

Let f\colon \mathbb{R} \to \mathbb{R} be the function f(x)=ax+b. We find the derivative of f(x).

Immediately from the definition we get: \begin{aligned}f'(x) &=\lim_{h\to 0} \frac{f(x+h)-f(x)}{h} \\ &=\lim_{h\to 0} \frac{[a(x+h)+b]-[ax+b]}{h} \\ &=\lim_{h\to 0} a \\ &=a.\end{aligned}

Here a is the slope of the tangent line. Note that the derivative at x does not depend on x because y=ax+b is the equation of a line.

Note. When a=0, we get f(x) = b and f'(x) = 0. The derivative of a constant function is zero.

Example 3.

Let g\colon \mathbb{R} \to \mathbb{R} be the function g(x)=|x|. Does g have a derivative at 0?

Now g'(x_0)= \begin{cases}+1 & \text{when $x_{0}>0$} \\ -1 & \text{when $x_{0}

The graph y=g(x) has no tangent at the point x_0=0: \frac{g(0+h)-g(0)}{h}= \frac{|0+h|-|0|}{h}=\frac{|h|}{h}=\begin{cases}+1 & \text{for $h>0$}, \\ -1 & \text{for $h Thus g'(0) does not exist.

Conclusion. The function g is not differentiable at the point 0.

Remark. Let f\colon (a,b)\to \mathbb{R}. If f'(x) exists for every x\in (a,b) then we get a function f'\colon (a,b)\to \mathbb{R}. We write:

(1) f(x) = f^{(0)}(x),
(2) f'(x) =  f^{(1)}(x) =  \frac{d}{dx}f(x),
(3) f''(x) =  f^{(2)}(x) =  \frac{d^2}{dx^2}f(x),
(4) f'''(x) =  f^{(3)}(x) =  \frac{d^3}{dx^3}f(x),
...

Here f''(x) is called the second derivative of f at x, f^{(3)} is the third derivative, and so on.

We introduce the notation \begin{eqnarray} C^n\bigl( ]a,b[\bigr) =\{ f\colon \, ]a,b[\, \to \mathbb{R} & \mid & f \text{ is } n \text{ times differentiable on the interval } ]a,b[ \nonumber \\ & & \text{ and } f^{(n)} \text{ is continuous}\}. \nonumber \end{eqnarray} These functions are said to be n times continuously differentiable.

Function |x|.

Example 4.

The distance moved by a cyclist (or a car) is given by s(t). Then the speed at the moment t is s'(t) and the acceleration is s''(t).

Linearization and differential
Derivative can also be used to approximate functions. From the definition of the derivative, we get  f'(x_0)\approx \frac{f(x)-f(x_0)}{x-x_0} \Leftrightarrow f(x)\approx f(x_0)+f'(x_0)(x-x_0), where the right-handed side is the linearization or the differential of f at x_0. The differential is denoted by df. The graph of the linearization,  y=f(x_0)+f'(x_0)(x-x_0), is the tangent line drawn on the graph of the function f at the point (x_0,f(x_0)). Later, in multi-variable calculus, the true meaning of the differential becomes clear. For now, it is not necessary to get troubled by the details.
Interactivity.

Properties of derivative


Next we give some useful properties of the derivative. These properties allow us to find derivatives for some familiar classes of functions such as polynomials and rational functions.

Continuity and derivative

If f is differentiable at the point x_0, then f is continuous at the point x_0:  \lim_{h\to 0} f(x_0+h) = f(x_0). Why? Because if f is differentiable, then we get f(x_0)+h\frac{f(x_0+h)-f(x_0)}{h} \rightarrow f(x_0)+0\cdot f'(x_0)=f(x_0), as h \to 0.

Note. If a function is continuous at the point x_0, it doesn't have to be differentiable at that point. For example, the function g(x) = |x| is continuous, but not differentiable at the point 0.

Differentiation Rules

Next we will give some important rules which are often applied in practical problems concerning determination of the derivative of a given function.

Suppose that f and g are differentiable at x.

A Constant Multiplier

(cf)'(x) = cf'(x),\ c \in \mathbb{R}

Proof.

Suppose that f is differentiable at x. We determine: (cf)'(x), where c\in \mathbb{R} is a constant.

\begin{aligned}\frac{(cf)(x+h)-(cf)(x)}{h} \ & \ = \ \frac{cf(x+h)-cf(x)}{h} \\ & \ = \ c \ \frac{f(x+h)-f(x)}{h}\end{aligned}

As h\to 0, we get c \ \frac{f(x+h)-f(x)}{h} \to c f'(x).

\square

The Sum Rule

(f+g)'(x) = f'(x) + g'(x)

Proof.

Suppose that f and g are differentiable at x. We determine (f+g)'(x).

By the definition: \begin{aligned}\frac{(f+g)(x+h)-(f+g)(x)}{h} \ & \ = \ \frac{[f(x+h)+g(x+h)]-[f(x)+g(x)]}{h} \\ & \ = \ \frac{f(x+h)-f(x)}{h}+\frac{g(x+h)-g(x)}{h}\end{aligned}

When h\to 0, we get \frac{f(x+h)-f(x)}{h}+\frac{g(x+h)-g(x)}{h}\to \ f'(x)+g'(x)

\square

The Product Rule

(fg)'(x) = f'(x)g(x) + f(x)g'(x)

Proof.

Suppose that f,g and are differentiable at x. We determine (fg)'(x). \begin{aligned}\frac{(fg)(x+h)-(fg)(x)}{h} & = \frac{f(x+h)g(x+h)-f(x)g(x)}{h} \\ & = \frac{f(x+h)g(x+h)-f(x)g(x+h)+f(x)g(x+h)-f(x)g(x)}{h} \\ & = \frac{f(x+h)-f(x)}{h}\ g(x+h)+f(x)\ \frac{g(x+h)-g(x)}{h}\end{aligned}

When h\to 0, we get \frac{f(x+h)-f(x)}{h}g(x+h)+f(x)\frac{g(x+h)-g(x)}{h}\to f'(x)g(x)+f(x)g'(x).

\square

The Power Rule

\frac{d}{dx} x^n = nx^{n-1} \text{, } n \in \mathbb{Z}

Proof.

For  n\ge 1 we repeteadly apply the product rule, and obtain \begin{aligned}\frac{d}{dx}x^n \ & = \frac{d}{dx}(x\cdot x^{n-1}) \\ & = (\frac{d}{dx}x)x^{n-1}+x\frac{d}{dx}x^{n-1} \\ & \stackrel{dx/dx=1}{=} x^{n-1}+x\frac{d}{dx}x^{n-1} \\ & = x^{n-1}+x\left( x^{n-2}+x\frac{d}{dx}x^{n-2}\right) \\ & = \ldots \\ & = \sum_{k=0}^{n-1} x^{n-1} \\ & = nx^{n-1}.\end{aligned}

The case of negative  n is obtained from this and the product rule applied to the identity  x^n \cdot x^{-n} = 1.

From the power rule we obtain a formula for the derivative of a polynomial. Let P(x)=a_n x^{n}+a_{n-1}x^{n-1}+\ldots+ a_1 x + a_0, where n\in \mathbb{N}. Then \frac{d}{dx}P(x)=na_nx^{n-1}+(n-1)a_{n-1}x^{n-2}+\ldots +2 a_2 x+a_1.

\square

The Reciprocal Rule

\Big(\frac{1}{f}\Big)'(x) = - \frac{f'(x)}{f(x)^2} \text{, } f(x) \neq 0

Proof.

Suppose that f is differentiable at x and f(x)\neq 0. We determine (\frac{1}{f})'(x).

From the definition we obtain: \begin{aligned}\frac{(1/f)(x+h)-(1/f)(x)}{h} & = \frac{1/f(x+h)-1/f(x)}{h} \\ & = \frac{\frac{f(x)}{f(x)f(x+h)}-\frac{f(x+h)}{f(x)f(x+h)}}{h} \\ & = \frac{f(x)-f(x+h)}{h}\frac{1}{f(x)f(x+h)}\end{aligned}

Because f is differentiable at x we get \frac{f(x)-f(x+h)}{h}\frac{1}{f(x)f(x+h)}=-f'(x)/f(x)^2, as h\to 0.

\square

The Quotient Rule

(f/g)'(x) = \frac{f'(x)g(x)-f(x)g'(x)}{g(x)^2},\ g(x) \neq 0

Proof.

Suppose that f,g are differentiable at x and g(x)\neq 0. Then \begin{aligned}(f/g)'(x) & = \Big( f \cdot \frac{1}{g}\Big) '(x) \\ & = f'(x)\frac{1}{g(x)}-f(x)\frac{g'(x)}{g(x)^2} \\ & = \frac{f'(x)g(x)-f(x)g'(x)}{g(x)^2}.\end{aligned}

\square

Interactivity. Vary x and the constant multiplier and see the effect of constant multiplier rule in practice.

Example 1.

\frac{d}{dx}(x^{2006}+5x^3+42)=\frac{d}{dx}x^{2006}+5\frac{d}{dx}x^3+42\frac{d}{dx}1=2006x^{2005}+5\cdot 3x^2.

Example 2.

\begin{aligned}\frac{d}{dx} [(x^4-2)(2x+1)] &= \frac{d}{dx}(x^4-2) \cdot (2x+1) + (x^4-2) \cdot \frac{d}{dx}(2x + 1) \\ &= 4x^3(2x+1) + 2(x^4-2) \\ &= 8x^4+4x^3+2x^4-4 \\ &= 10x^4+4x^3-4.\end{aligned}

Note. We can check the answer by deriving it in another way: \frac{d}{dx} [(x^4-2)(2x+1)] = \frac{d}{dx} (2x^5 +x^4 -4x -2) = 10x^4 +4x^3 -4.

Function  (x^4-2)(2x+1) .

Example 3.

For x \neq 0 we get \frac{d}{dx} \frac{3}{x^3} = 3 \cdot \frac{d}{dx} \frac{1}{x^3} = -3 \cdot \frac{\frac{d}{dx} x^3}{(x^3)^2} = -3 \cdot \frac{3x^2}{x^6}= - \frac{9}{x^4}.

Note. There is another way of solving the problem above by noticing that \frac{1}{x^3} = x^{-3} and differentiating it as a power: \frac{d}{dx} \ \frac{3}{x^3} = 3 \cdot \frac{d}{dx} x^{-3} = 3 \cdot (-3x^{-4})= - \frac{9}{x^4}

Example 4.

\begin{aligned}\frac{d}{dx} \frac{x^3}{1+x^2} & = \frac{(\frac{d}{dx}x^3)(1+x^2)-x^3\frac{d}{dx}(1+x^2)}{(1+x^2)^2} \\ & = \frac{3x^2(1+x^2)-x^3(2x)}{(1+x^2)^2} \\ & = \frac{3x^2+x^4}{(1+x^2)^2}.\end{aligned}

Function x^3 / (1+x^2).

Rolle's Theorem

If f is differentiable at a local extremum x_0\in \, ]a,b[, then f'(x_0)=0.

Proof (idea).

The one-sided limits of the difference quotient have different signs at a local extremum. For example, for a local maximum it holds that \begin{eqnarray} \frac{f(x_0+h)-f(x_0)}{h} = \frac{\text{negative} }{\text{positive}}&\le& 0, \text{ when } h>0, \nonumber \\ \frac{f(x_0+h)-f(x_0)}{h} = \frac{\text{negative}}{\text{negative}}&\ge& 0, \text{ when } h<0 \nonumber \end{eqnarray} and |h| is so small that f(x_0) is a maximum on the interval [x_0-h,x_0+h].

L'Hospital's Rule

There are many different versions of this rule, but we present only the simplest one. Let us assume that f(x_0)=g(x_0)=0 and the functions f,g are differentiable on some interval ]x_0-\delta,x_0+\delta[. If  \lim_{x\to x_0}\frac{f'(x)}{g'(x)} exists, then  \lim_{x\to x_0}\frac{f(x)}{g(x)}=\lim_{x\to x_0}\frac{f'(x)}{g'(x)}.

Proof (idea).

In the special case g'(x_0)\neq 0 the proof is simple:  \frac{f(x)}{g(x)}=\frac{f(x)-f(x_0)}{g(x)-g(x_0)} = \frac{\bigl( f(x)-f(x_0)\bigr) /(x-x_0)}{\bigl( g(x)-g(x_0)\bigr) /(x-x_0)} \to \frac{f'(x_0)}{g'(x_0)}. In the general case we need the so-called generalized mean value theorem, which states that  \frac{f(x)}{g(x)} = \frac{f'(c)}{g'(c)} for some c\in \, ]x_0,x[. Here we have the same point c both in the numerator and the denominator, so we do not even need the continuity of the derivatives!

Derivatives of Trigonometric Functions


In this section, we give differentiation formulas for trigonometric functions \sin, \cos and \tan.

Derivative of Sine

\sin'(t)=\cos(t)

Proof.

Function \sin(x) and its derivative function \cos(x).

Derivative of Cosine

\cos'(t)= - \sin(t)

Proof.

This follows in a similar way as the derivative of Sine, but more easily from the identity \cos(t)=\sin(\pi/2-t) and the Chain rule to be introduced in the following section.

\square

Function \cos(x) and its derivative function -\sin(x).

Derivative of Tangent

\tan'(t) = \frac{1}{\cos^2(t)}=1+\tan^2 t.

Proof.

Because \tan(t)=\frac{\sin(t)}{\cos(t)}, from the quotient rule we obtain \tan'(t)=\frac{\sin'(t)\cos(t)-\sin(t)\cos'(t)}{\cos^2(t)}=\frac{\cos^2(t)+\sin^2(t)}{\cos^2(t)}=\begin{cases}\frac{1}{\cos^2(t)} & \\ 1+\tan^2 t.\end{cases}

\square

Function \tan(x) and its derivative function 1/\cos^2(x).

Example 1.

\frac{d}{dx} (3 \sin(x)) = 3 \sin'(x) = 3 \cos(x).

Example 2.

\frac{d}{dx} \cos^2 (x) = \cos'(x) \cdot \cos(x) + \cos(x) \cdot \cos'(x) = -2\sin(x)\cos(x).

Example 3.

\begin{aligned}
		\frac{d}{dx} \frac{\sin(x) + 1}{\cos(x)} &= \frac{d}{dx} \left( \frac{\sin(x)}{\cos(x)} + \frac{1}{\cos(x)} \right) \\
		&= \tan'(x) - \frac{\cos'(x)}{\cos^2(x)} \\
		&= \frac{1+\sin(x)}{\cos^2 (x)}.\end{aligned}

The Chain Rule


In this section we learn a formula for finding the derivative of a composite function. This important formula is known as the Chain Rule.

The Chain Rule.

Let f\colon \mathbb{R}\to \mathbb{R}, g\colon \mathbb{R}\to \mathbb{R} and f \circ g \colon \mathbb{R}\to \mathbb{R}.

Let g be differentiable at the point x and f at g(x). Then \frac{d}{dx}f(g(x))=f'(g(x))g'(x).

Proof.

Consider

\begin{aligned}\frac{f(g(x+h))-f(g(x))}{h} &= \frac{f(g(x+h))-f(g(x))}{h} \ \frac{g(x+h)-g(x)}{g(x+h)-g(x)} \\
			&= \frac{f(g(x+h))-f(g(x))}{g(x+h)-g(x)} \ \frac{g(x+h)-g(x)}{h}.\end{aligned}

Now let us write k(h):=g(x+h)-g(x). Then g(x+h)=g(x)+k(h) and we get \frac{f(g(x+h))-f(g(x))}{h}=\frac{f(g(x)+k(h))-f(g(x))}{k(h)}\frac{g(x+h)-g(x)}{h}.

Problem. What if k(h)=0? Note that one cannot divide by zero.

Solution. Define E(k):= \begin{cases}0, & \text{for  $k=0$}, \\
			\frac{f(g(x)+k)-f(g(x))}{k}-f'(g(x)), & \text{for  $k\neq 0$},\end{cases} so that \frac{f(g(x+h))-f(g(x))}{h}=[E(k(h))+f'(g(x))]\frac{g(x+h)-g(x)}{h}. Now, because E is continuous, we get [E(k(h))+f'(g(x))]\frac{g(x+h)-g(x)}{h}\to f'(g(x))g'(x). as h\to 0.

\square

Example 1.

The problem is to differentiate the function (2x-1)^3. We take f(x) = x^3 and g(x) = 2x-1 and differentiate the composite function f(g(x)). As f'(x) = 3x^2 \text{ and } g'(x) = 2, we get \frac{d}{dx} (2x-1)^3 = 3(2x-1)^2 \cdot 2 = 6(4x^2-4x+1) = 24x^2-24x+6.

Function (2x-1)^3 and its derivative function.

Example 2.

We need to differentiate the function \sin 3x. Take f(x) = \sin x and g(x) = 3x, then differentiate the composite function f(g(x)). \frac{d}{dx} \sin 3x = \cos 3x \cdot 3 = 3 \cos 3x.

Remark. Let h\colon \mathbb{R}\to \mathbb{R}, g\colon \mathbb{R}\to \mathbb{R} and f\colon \mathbb{R}\to \mathbb{R}. Now \frac{d}{dx}f(g(h(x)))=f'(g(h(x)))\frac{d}{dx}g(h(x))=f'(g(h(x)))g'(h(x))h'(x). Similarly, one may obtain even more complex rules for composites of multiple functions.

Function \sin 3x and its derivative function.

Example 3.

Differentiate the function \cos^3 2x. Take f(x) = x^3, g(x) = \cos x and h(x) = 2x and differentiate the composite function f(g(h(x))). \begin{aligned}\frac{d}{dx} \cos^3 2x &= 3(\cos 2x)^2 \cdot \frac{d}{dx} \cos 2x \\
		&= 3 \cos^2 2x \cdot (-\sin 2x) \cdot 2 \\
		&= -6 \sin 2x \cos^2 2x.\end{aligned}

Function \cos^3 2x and its derivative function.

Extremal Value Problems


We will discuss the Intermediate Value Theorem for differentiable functions, and its connections to extremal value problems.

Definition: Local Maxima and Minima

A function f\colon A\to \mathbb{R} has a a local maximum at the point x_0\in A, if for some h\gt 0 and for all x\in A such that |x-x_0|\lt h, we have f(x)\leq f(x_0).

Similarly, a function f\colon A\to \mathbb{R} has a local minimum at the point x_0\in A , if for some h>0 and for all x\in A such that |x-x_0|\lt h, we have f(x)\geq f(x_0).

A local extreme is a local maximum or a local minimum.

Remark. If x_0 is a local maximum value and f'(x_0) exists, then \begin{cases}f'(x_0) & =\lim_{h\to 0^{+}}\frac{f(x_0+h)-f(x_0)}{h} \leq 0 \\
		f'(x_0) & =\lim_{h\to 0^{-}}\frac{f(x_0+h)-f(x_0)}{h} \geq 0.\end{cases} Hence f'(x_0)=0.

We get:

Theorem 1.

Let x_0\in [a,b] be a local extremal value of a continuous function f\colon [a,b]\to \mathbb{R}. Then either

  1. the derivative f'(x_0) doesn't exist (this includes also cases x_0=a and x_0=b) or

  2. f'(x_0)=0.

Example 1.

Let f: \mathbb{R} \to \mathbb{R} be defined by f(x) = x^3 -3x + 1. Then f'(x) = 3x^2-3 and we can see that at the points x_0 = -1 and x_0 = 1 the local maximum and minimum of f are obtained, f'(-1) = 3 \cdot (-1)^2 - 3 = 0 \text{ and } f'(1) = 3 \cdot 1^2 - 3 = 0.

Function x^3-3x+1 and its derivative function 3x^2-3.

Finding the global extrema

In practice, when we are looking for the local extrema of a given function, we need to check three kinds of points:

  1. the zeros of the derivative

  2. the endpoints of the domain of definition (interval)

  3. points where the function is not differentiable

If we happened to know beforehand that the function has a minimum/maximum, then we start off by finding all the possible local extrema (the points described above), evaluate the function at these points and pick the greatest/smallest of these values.

Example 2.

Let us find the smallest and greatest value of the function f\colon [0,2]\to \mathbf{R}, f(x)=x^3-6x. Since the function is continuous on a closed interval, then it has a maximum and a minimum. Since the function is differentiable, it is sufficient to examine the endpoints of the interval and the zeros of the derivative that are contained in the interval.

The zeros of the derivative: f'(x)=3x^2-6=0 \Leftrightarrow x=\pm \sqrt{2}. Since -\sqrt{2}\not\in [0,2], we only need to evaluate the function at three points, f(0)=0, f(\sqrt{2})=-4\sqrt{2} and f(2)=-4. From these we can see that the smallest value of the function is -4\sqrt{2} and the greatest value is 0, respectively.

Next we will formulate a fundamental result for differentiable functions. The basic idea here is that the change on an interval can only happen, if there is change at some point on the inverval.

Theorem 2.

(The Intermediate Value Theorem for Differentiable Functions). Let f\colon [a,b]\to \mathbb{R} be continuous in the interval [a,b] and differentiable in the interval (a,b). Then f'(x_0)=\frac{f(b)-f(a)}{b-a} for some x_0\in (a,b).

Proof.

Let f be continuous in the interval [a,b] and differentiable in the interval (a,b). Let us define g(x):=f(x)-\frac{f(b)-f(a)}{b-a}(x-a)-f(a).

Now g(a)=g(b)=0 and g is differentiable in the interval (a,b). According to Rolle's Theorem, there exists c\in(a,b) such that g'(c)=0. Hence f'(c)=g'(c)+\frac{f(b)-f(a)}{b-a}=\frac{f(b)-f(a)}{b-a}.

\square

This result has an important application:

Theorem 3.

Let f\colon (a,b)\to \mathbb{R} be a differentiable function. Then

  1. If for all x\in (a,b) \ \ f'(x)\geq 0, then f is increasing,

  2. If for all x\in (a,b) \ \ f'(x)\leq 0, then f is decreasing.

Proof.

Suppose that a \lt x_1 \lt x_2 \lt b.

Then by Theorem 2 there exists x_0\in (x_1,x_2) such that f'(x_0)=\frac{f(x_2)-f(x_1)}{x_2-x_1}.

It follows that f(x_2)-f(x_1)=f'(x_0)(x_2-x_1).

Hence we may conclude that f is increasing for f'(x_0)\geq 0 and decreasing for f'(x_0)\leq 0.

Example 3.

For the polynomial f(x) = \frac{1}{4} x^4-2x^2-7 the derivative is f'(x) = x^3-4x = x(x^2-4) = 0, when x=0, x=2 or x=-2. Now we can draw a table:

x -2 \lt x \lt 0 0 \lt x \lt 2 x>2
x >0 >0
x^2-4 >0 >0
f'(x) >0 >0
f(x) decr. incr. decr. incr.

Function \frac{1}{4} x^4-2x^2-7.

Example 4.

We need to find a rectangle so that its area is 9 and it has the least possible perimeter.

Let x\ (>0) and y\ (>0) be the sides of the rectangle. Then x \cdot y = 9 and we get y=\frac{9}{x}. Now the perimeter is 2x+2y = 2x+2 \frac{9}{x} = \frac{2x^2+18}{x}. In which point does the function f(x) = \frac{2x^2+18}{x} get its minimum value? Function f is continuous and differentiable, when x>0 and using the quotient rule, we get f'(x) = \frac{4x \cdot x-(2x^2+18) \cdot 1}{x^2} = \frac{2x^2-18}{x^2}. Now f'(x) = 0, when \begin{aligned}2x^2-18 &= 0 \\
		2x^2 &= 18 \\
		x^2 &= 9 \\
		x &= \pm 3\end{aligned} but we have defined that x>0 and therefore are only interested in the case x=3. Let's draw a table:

x x>3
f'(x) >0
f(x) decr. incr.

As the function f is continuous, we now know that it attains its minimum at the point x=3. Now we calculate the other side of the rectangle: y=\frac{9}{x}=\frac{9}{3}=3.

Thus, the rectangle, which has the least possible perimeter is actually a square, which sides are of the length 3.

Function \frac{2x^2+18}{x}.

Example 5.

We must make a one litre measure, which is shaped as a right circular cylinder without a lid. The problem is to find the best size of the bottom and the height so that we need the least possible amount of material to make the measure.

Let r > 0 be the radius and h > 0 the height of the cylinder. The volume of the cylinder is 1 dm^3 and we can write \pi r^2 h = 1 from which we get h = \frac{1}{\pi r^2}.

The amount of material needed is the surface area A_{\text{bottom}} + A_{\text{side}} = \pi r^2 + 2 \pi r h = \pi r^2 + \frac{2 \pi r}{\pi r^2} = \pi r^2 + \frac{2}{r}.

Let function f: (0, \infty) \to \mathbb{R} be defined by f(r) = \pi r^2 + \frac{2}{r}. We must find the minimum value for function f, which is continuous and differentiable, when r>0. Using the reciprocal rule, we get f'(r) = 2\pi r -2 \cdot \frac{1}{r^2} = \frac{2\pi r^3 - 2}{r^2}. Now f'(r) = 0, when \begin{aligned}2\pi r^3 - 2 &= 0 \\
		2\pi r^3 &= 2 \\
		r^3 &= \frac{1}{\pi} \\
		r &= \frac{1}{\sqrt[3]{\pi}}.\end{aligned}

Let's draw a table:

r r>\frac{1}{\sqrt[3]{\pi}}
f'(r) >0
f(r) decr. incr.

As the function f is continuous, we now know that it gets its minimum value at the point r= \frac{1}{\sqrt[3]{\pi}} \approx 0.683. Then h = \frac{1}{\pi r^2} = \frac{1}{\pi \left(\frac{1}{\sqrt[3]{\pi}}\right)^2} = \frac{1}{\frac{\pi}{\pi^{2/3}}} = \frac{1}{\sqrt[3]{\pi}} \approx 0.683.

This means that it would take least materials to make a measure, which is approximately 2 \cdot 0.683 dm  = 1.366 dm  \approx 13.7 cm in diameter and 0.683 dm  \approx 6.8 cm high.

Function \pi r^2 + \frac{2}{r}.

', strokeColor : '#b942f5', fillColor : 'black', strokeWidth : 1, fillOpacity : .5, showinfobox : false, label : { useMathJax : true } }); board.create('segment', [x0,function() { return [x0.X(),f(x0.X())]; }], { dash : 1, strokeColor : '#b942f5' }); var g1 = board.create('glider', [function() { return x0.X(); }, function() { return f(x0.X()); }, graph], { visible : false }); var t = board.create('tangent', [g1], { strokeWidth : 1.5, strokeColor : '#b942f5' }); board.create('line', [[function() { return x0.X(); }, function() { return f(x0.X()); }], [function() { return x0.X()+.7; }, function() { return f(x0.X()+.7); }]], { strokeOpacity : .5, strokeWidth : 1.5 }); board.create('line', [[function() { return x0.X(); }, function() { return f(x0.X()); }], [function() { return x0.X()+1.4; }, function() { return f(x0.X()+1.4); }]], { strokeOpacity : .5, strokeWidth : 1.5 }); board.create('line', [[function() { return x0.X(); }, function() { return f(x0.X()); }], [function() { return x0.X()+2.1; }, function() { return f(x0.X()+2.1); }]], { strokeOpacity : .5, strokeWidth : 1.5 }); board.fullUpdate(); })(); (function() { var board = JXG.JSXGraph.initBoard('jxgbox02', { boundingbox : [-5, 7, 5, -4], shownavigation : false, axis : true }); var f = function(x) { return x*x*x+2*x*x; } var graph = board.create('functiongraph', [f, -5, 5], { strokeWidth : 2, strokeColor : 'black' }); var g1 = board.create('glider', [.5, f(.5), graph], { strokeWidth : 1, size : 5.5, strokeColor : 'black', fillColor : 'gray', name : 'df', label : { strokeColor : 'blue' } }); var t = board.create('tangent', [g1], { strokeWidht : 2, strokeOpacity : .8 }); var text1 = board.create('text', [-4, 5, function() { return 'f(x) = x^3+2x^2'}], { useMathJax : true, fontSize : 14, strokeColor : 'black' }); var text2 = board.create('text', [-4, 4, function() { var k = t.getSlope(); var b = g1.Y()-k*g1.X(); var temp = ''; if(k == 0) { temp = 'df=' + b.toFixed(2) +''; } else { if(b < 0) { temp = 'df=' + k.toFixed(2) + 'x' + b.toFixed(2) + ''; } else { temp = 'df=' + k.toFixed(2) + 'x + ' + b.toFixed(2) + ''; } } return temp; } ], { useMathJax : true, fontSize : 14, strokeColor : 'blue' }); })(); (function() { var board = JXG.JSXGraph.initBoard('jxgbox03', { boundingbox : [-3.5, 3, 3.5, -3], axis : false, shownavigation : false, showcopyright : false }); var xaxis = board.create('axis', [[0, 0],[1, 0]], { ticks : { majorHeight : 5 } }); var yaxis = board.create('axis' ,[[0, 0],[0, 1]], { ticks : { majorHeight : 5} }); var f = function(x) { return x*x*x+1; } var graph = board.create('functiongraph', [f, -3.5, 3.5], { strokeWidth : 1.5, strokeColor : 'black' }); var g1 = board.create('glider', [1, f(1), graph], { strokeColor : 'black', size : 2, fillColor : 'red', name : '' }); var t = board.create('tangent', [g1], { strokeColor : 'red', strokeWidth : 1.5 }); var a = board.create('segment', [[function() { return g1.X() }, function() { return g1.Y()-t.getSlope(); }], [function() { return g1.X()-1; }, function() { return g1.Y()-t.getSlope() }]], { strokeWidth : 2, strokeOpacity : .8, dash : 1, highlight : false }); var b = board.create('segment', [[ function() { return g1.X(); }, function() { return g1.Y(); }], [ function() { return a.point1.X(); }, function() { return a.point1.Y(); } ]], { strokeWidth : 2, strokeOpacity : .8, dash : 1, highlight : false }); })(); (function() { var board = JXG.JSXGraph.initBoard('jxgbox04', { boundingbox : [-5, 5, 5, -2], axis : false, shownavigation : false, showcopyright : false }); var xaxis = board.create('axis', [[0, 0], [1, 0]], { highlight : false, drawZero : true, ticks : { minorTicks : 0, majorHeight : 9, label : { highlight : false, offset : [-5, -15] } } }); var yaxis = board.create('axis', [[0, 0], [0, 1]], { highlight : false, ticks : { minorTicks : 0, majorHeight : 9, label : { offset : [-15, 0 ], position : 'lrt', highlight : false } } }); xaxis.defaultTicks.ticksFunction = function() { return 1; }; yaxis.defaultTicks.ticksFunction = function() { return 1; }; var f = function(x) { return Math.abs(x); } board.create('functiongraph', [f, -5, 5], { strokeWidth : 2, strokeColor : '#2183de', highlight : false }); board.create('text', [1.5, 1.2, function() { return 'f(x)=|x|'; }], { useMathJax : true, strokeColor : '#2183de', fontSize : 13, fixed : true, highlight : false }); board.fullUpdate(); })(); (function() { var board = JXG.JSXGraph.initBoard('jxgbox05', { boundingbox : [-1.7, 10.1, 1.7, -10.1], axis : false, shownavigation : false, showcopyright : false }); var xaxis = board.create('axis', [[0, 0], [1, 0]], { highlight : false, drawZero : true, ticks : { minorTicks : 0, majorHeight : 9, label : { highlight : false, offset : [-5, -15] } } }); var yaxis = board.create('axis', [[0, 0], [0, 1]], { highlight : false, ticks : { minorTicks : 0, majorHeight : 9, label : { offset : [-15, 0 ], position : 'lrt', highlight : false } } }); xaxis.defaultTicks.ticksFunction = function() { return 1; }; yaxis.defaultTicks.ticksFunction = function() { return 5; }; var f = function(x) { return (x*x*x*x-2)*(2*x+1); } board.create('functiongraph', [f, -1.7, 1.7], { strokeWidth : 2.5, strokeColor : 'black', highlight : false }); board.fullUpdate(); })(); (function() { var board = JXG.JSXGraph.initBoard('jxgbox06', { boundingbox : [-6.1, 6.5, 6.1, -6.5], axis : false, shownavigation : false, showcopyright : false }); var xaxis = board.create('axis', [[0, 0], [1, 0]], { highlight : false, drawZero : true, ticks : { minorTicks : 0, majorHeight : 9, label : { highlight : false, offset : [-5, -15] } } }); var yaxis = board.create('axis', [[0, 0], [0, 1]], { highlight : false, ticks : { minorTicks : 0, majorHeight : 9, label : { offset : [-15, 0 ], position : 'lrt', highlight : false } } }); xaxis.defaultTicks.ticksFunction = function() { return 2; }; yaxis.defaultTicks.ticksFunction = function() { return 2; }; var f = function(x) { return x*x*x/(1+x*x); } board.create('functiongraph', [f, -6.1, 6.1], { strokeWidth : 2.5, strokeColor : 'black', highlight : false }); board.fullUpdate(); })(); (function() { var board = JXG.JSXGraph.initBoard('jxgbox07', { boundingbox : [-6.77, 6, 6.77, -6], axis : false, shownavigation : false, showcopyright : false }); var xaxis = board.create('axis', [[0, 0], [1, 0]], { highlight : false, drawZero : true, ticks : { minorTicks : 0, majorHeight : 9, label : { highlight : false, offset : [-5, -15] } } }); var yaxis = board.create('axis', [[0, 0], [0, 1]], { highlight : false, ticks : { minorTicks : 0, majorHeight : 9, label : { offset : [-15, 0 ], position : 'lrt', highlight : false } } }); var f = function(x) { return Math.sin(x); } var df = function(x) { return Math.cos(x); } board.create('functiongraph', [f, -6.77, 6.77], { strokeWidth : 1.5, strokeColor : 'red' }); board.create('functiongraph', [df, -6.77, 6.77], { strokeWidth : 1.5, strokeColor : 'black' }); })(); (function() { var board = JXG.JSXGraph.initBoard('jxgbox08', { boundingbox : [-6.77, 6, 6.77, -6], axis : false, shownavigation : false, showcopyright : false }); var xaxis = board.create('axis', [[0, 0], [1, 0]], { highlight : false, drawZero : true, ticks : { minorTicks : 0, majorHeight : 9, label : { highlight : false, offset : [-5, -15] } } }); var yaxis = board.create('axis', [[0, 0], [0, 1]], { highlight : false, ticks : { minorTicks : 0, majorHeight : 9, label : { offset : [-15, 0 ], position : 'lrt', highlight : false } } }); var f = function(x) { return Math.cos(x); } var df = function(x) { return Math.sin(x); } board.create('functiongraph', [f, -6.77, 6.77], { strokeWidth : 1.5, strokeColor : 'blue' }); board.create('functiongraph', [df, -6.77, 6.77], { strokeWidth : 1.5, strokeColor : 'black' }); })(); (function() { var board = JXG.JSXGraph.initBoard('jxgbox09', { boundingbox : [-6.77, 6, 6.77, -6], axis : false, shownavigation : false, showcopyright : false }); var xaxis = board.create('axis', [[0, 0], [1, 0]], { highlight : false, drawZero : true, ticks : { minorTicks : 0, majorHeight : 9, label : { highlight : false, offset : [-5, -15] } } }); var yaxis = board.create('axis', [[0, 0], [0, 1]], { highlight : false, ticks : { minorTicks : 0, majorHeight : 9, label : { offset : [-15, 0 ], position : 'lrt', highlight : false } } }); var f = function(x) { return Math.tan(x); } var df = function(x) { return 1/Math.pow(Math.cos(x),2); } board.create('functiongraph', [f, -6.77, 6.77], { strokeWidth : 1.5, strokeColor : 'green' }); board.create('functiongraph', [df, -6.77, 6.77], { strokeWidth : 1.5, strokeColor : 'black' }); })(); (function() { var board = JXG.JSXGraph.initBoard('jxgbox10', { boundingbox : [-3.2, 4.5, 4.8, -2.5], axis : false, shownavigation : false, showcopyright : false }); var xaxis = board.create('axis', [[0, 0], [1, 0]], { highlight : false, drawZero : true, ticks : { minorTicks : 0, majorHeight : 9, label : { highlight : false, offset : [-5, -15] } } }); var yaxis = board.create('axis', [[0, 0], [0, 1]], { highlight : false, ticks : { minorTicks : 0, majorHeight : 9, label : { offset : [-15, 0 ], position : 'lrt', highlight : false } } }); var f = function(x) { return Math.pow(2*x*x-1, 3); } var df = function(x) { return 6*Math.pow(2*x*x-1, 2); } board.create('functiongraph', [f, 0, 4.8], { strokeWidth : 1.5, strokeColor : 'green' }); board.create('functiongraph', [df, 0, 4.8], { strokeWidth : 1.5, strokeColor : 'black' }); })(); (function() { var board = JXG.JSXGraph.initBoard('jxgbox11', { boundingbox : [-6.77, 6, 6.77, -6], axis : false, shownavigation : false, showcopyright : false }); var xaxis = board.create('axis', [[0, 0], [1, 0]], { highlight : false, drawZero : true, ticks : { minorTicks : 0, majorHeight : 9, label : { highlight : false, offset : [-5, -15] } } }); var yaxis = board.create('axis', [[0, 0], [0, 1]], { highlight : false, ticks : { minorTicks : 0, majorHeight : 9, label : { offset : [-15, 0 ], position : 'lrt', highlight : false } } }); var f = function(x) { return Math.sin(3*x); } var df = function(x) { return 3*Math.cos(x); } board.create('functiongraph', [f, -6.77, 6.77], { strokeWidth : 1.5, strokeColor : 'red' }); board.create('functiongraph', [df, -6.77, 6.77], { strokeWidth : 1.5, strokeColor : 'black' }); })(); (function() { var board = JXG.JSXGraph.initBoard('jxgbox12', { boundingbox : [-6.77, 6, 6.77, -6], axis : false, shownavigation : false, showcopyright : false }); var xaxis = board.create('axis', [[0, 0], [1, 0]], { highlight : false, drawZero : true, ticks : { minorTicks : 0, majorHeight : 9, label : { highlight : false, offset : [-5, -15] } } }); var yaxis = board.create('axis', [[0, 0], [0, 1]], { highlight : false, ticks : { minorTicks : 0, majorHeight : 9, label : { offset : [-15, 0 ], position : 'lrt', highlight : false } } }); var f = function(x) { return Math.pow(Math.cos(2*x), 3); } var df = function(x) { return -6*Math.sin(2*x)*Math.pow(Math.cos(2*x), 2); } board.create('functiongraph', [f, -6.77, 6.77], { strokeWidth : 1.5, strokeColor : 'blue' }); board.create('functiongraph', [df, -6.77, 6.77], { strokeWidth : 1.5, strokeColor : 'black' }); })(); (function() { var board = JXG.JSXGraph.initBoard('jxgbox13', { boundingbox : [-4.73, 4.2, 4.73, -4.2], axis : false, shownavigation : false, showcopyright : false }); var xaxis = board.create('axis', [[0, 0], [1, 0]], { highlight : false, drawZero : true, ticks : { minorTicks : 0, majorHeight : 9, label : { highlight : false, offset : [-5, -15] } } }); var yaxis = board.create('axis', [[0, 0], [0, 1]], { highlight : false, ticks : { minorTicks : 0, majorHeight : 9, label : { offset : [-15, 0 ], position : 'lrt', highlight : false } } }); var f = function(x) { return x*x*x-3*x+1; } var df = function(x) { return 3*x*x-3; } board.create('functiongraph', [f, -3.2, 3.2], { strokeWidth : 1.5, strokeColor : 'blue' }); board.create('functiongraph', [df, -3.2, 3.2], { strokeWidth : 1.5, strokeColor : 'red' }); var p = []; p[0] = board.create('point', [-1, f(-1)]); p[1] = board.create('point', [-1, df(-1)]); p[2] = board.create('point', [1, f(1)]); p[3] = board.create('point', [1, df(1)]); for(var i=0; i < p.length; i++) { p[i].setAttribute({ strokeColor : 'black', size : .5, name : '', fillColor : 'black', fixed : true }); } board.create('segment', [p[0],p[1]], { strokeColor : 'black', dash : 1, highlight : false }); board.create('segment', [p[2],p[3]], { strokeColor : 'black', dash : 1, highlight : false }); board.fullUpdate(); })(); (function() { var board = JXG.JSXGraph.initBoard('jxgbox14', { boundingbox : [-6.76, 1, 6.76, -11], axis : false, shownavigation : false, showcopyright : false }); var xaxis = board.create('axis', [[0, 0], [1, 0]], { highlight : false, drawZero : true, ticks : { minorTicks : 0, majorHeight : 9, label : { highlight : false, offset : [-5, -15] } } }); var yaxis = board.create('axis', [[0, 0], [0, 1]], { highlight : false, ticks : { minorTicks : 0, majorHeight : 9, label : { offset : [-15, 0 ], position : 'lrt', highlight : false } } }); var f = function(x) { return Math.pow(x, 4)/4-2*x*x-7; } board.create('functiongraph', [f, -5.6, 5.6], { strokeWidth : 1.5, strokeColor : 'black' }); var p = []; p[0] = board.create('point', [-2, 0], { }); p[1] = board.create('point', [-2, f(-2)], { }); p[2] = board.create('point', [0, 0], { }); p[3] = board.create('point', [0, f(0)], { }); p[4] = board.create('point', [2, 0], { }); p[5] = board.create('point', [2, f(2)], { }); for(var i=0; i < p.length; i++) { p[i].setAttribute({ visible : false }); } board.create('segment', [p[0],p[1]], { strokeColor : 'red', highlight : false, fixed : true }); board.create('segment', [p[2],p[3]], { strokeColor : 'blue', highlight : false, fixed : true }); board.create('segment', [p[4],p[5]], { strokeColor : 'red', highlight : false, fixed : true }); board.fullUpdate(); })(); (function() { var board = JXG.JSXGraph.initBoard('jxgbox15', { boundingbox : [-1, 15, 10, -1.5], axis : false, shownavigation : false, showcopyright : false }); var xaxis = board.create('axis', [[0, 0], [1, 0]], { highlight : false, drawZero : true, ticks : { minorTicks : 0, majorHeight : 9, label : { highlight : false, offset : [-5, -15] } } }); var yaxis = board.create('axis', [[0, 0], [0, 1]], { highlight : false, ticks : { minorTicks : 0, majorHeight : 9, label : { offset : [-15, 0 ], position : 'lrt', highlight : false } } }); xaxis.defaultTicks.ticksFunction = function() { return 2; }; yaxis.defaultTicks.ticksFunction = function() { return 2; }; var f = function(x) { return (2*x*x+18)/x; } var graph = board.create('functiongraph', [f, -1, 15.9], { strokeColor : 'black', strokeWidth : 2.5, highlight : false }); var p = []; p[0] = board.create('point', [3, 0]); p[1] = board.create('point', [3,f(3)]); for(var i=0; i < p.length; i++) { p[i].setAttribute({ name : '', size : 1, strokeColor : 'black', strokeWidth : .5, fixed : true }); } board.create('segment', [p[0], p[1]], { dash : 1, strokeWidth : 2, strokeColor : 'red', fixed : true, highlight : false }); board.fullUpdate(); })(); (function() { var board = JXG.JSXGraph.initBoard('jxgbox16', { boundingbox : [-1, 13, 5, -1], axis : false, shownavigation : false, showcopyright : false }); var xaxis = board.create('axis', [[0, 0], [1, 0]], { highlight : false, drawZero : true, ticks : { minorTicks : 0, majorHeight : 9, label : { highlight : false, offset : [-5, -15] } } }); var yaxis = board.create('axis', [[0, 0], [0, 1]], { highlight : false, ticks : { minorTicks : 0, majorHeight : 9, label : { offset : [-15, 0 ], position : 'lrt', highlight : false } } }); xaxis.defaultTicks.ticksFunction = function() { return 2; }; yaxis.defaultTicks.ticksFunction = function() { return 2; }; var f = function(x) { return Math.PI*x*x+2/x; } var graph = board.create('functiongraph', [f, 0, 5], { strokeColor : 'black', strokeWidth : 2.5, highlight : false }); var p = []; var a = 1/Math.pow(Math.PI,1/3); p[0] = board.create('point', [a, 0]); p[1] = board.create('point', [a, f(a)]); for(var i=0; i < p.length; i++) { p[i].setAttribute({ name : '', size : 1, strokeColor : 'black', fixed : true }); } board.create('segment', [p[0], p[1]], { dash : 1, strokeWidth : 2, strokeColor : 'red', fixed : true, highlight : false }); board.fullUpdate(); })(); (function() { var board = JXG.JSXGraph.initBoard('jxgbox17', { boundingbox : [-4, 4, 4, -4], axis : false, shownavigation : false, showcopyright : false }); var xaxis = board.create('axis', [[0, 0], [1, 0]], { highlight : false, drawZero : true, ticks : { minorTicks : 0, majorHeight : 9, label : { highlight : false, offset : [-5, -15] } } }); var yaxis = board.create('axis', [[0, 0], [0, 1]], { highlight : false, ticks : { minorTicks : 0, majorHeight : 9, label : { offset : [-15, 0 ], position : 'lrt', highlight : false } } }); xaxis.defaultTicks.ticksFunction = function() { return 1; }; yaxis.defaultTicks.ticksFunction = function() { return 1; }; var s = board.create('slider', [[-3.5, 3.5], [-.5, 3.5], [-10, 0, 10]],{ suffixLabel : ' c=',
			postLabel		:	'', label : { useMathJax : true, fontSize : 13 }, fillColor : '#31d490', withTicks : false, precision : 2, snapWidth : 1 }); var f = function(x) { return x*x*x-x; } var g = function(x) { return s.Value()*(x*x*x-x); } var graph1 = board.create('functiongraph', [f, -4, 4], { strokeColor : '#4260f5', strokeWidth : 2.5, highlight : false, strokeOpacity : 0.6, highlight : false }); var graph2 = board.create('functiongraph', [g, -4, 4], { strokeColor : '#31d490', strokeWidth : 2.5, highlight : false, strokeOpacity : .8, highlight : false }); var g0 = board.create('glider', [0, 0, xaxis], { name : 'x' + '_' + '0', size : 4, strokeColor : 'black', strokeWidth : .7, fillColor : '#42f57e', label : { useMathJax : true } }); var g1 = board.create('glider', [0, 0, graph1], { name : '', size : 2, strokeWidth : .7, strokeColor : 'black', fillColor : '#42f57e', fixed : true, highlight : false }); var l1 = board.create('line', [g0, g1], { straightFirst : false, strokeColor : '#42f57e', dash : 1, highlight : false }); var g2 = board.create('glider', [0, 0, graph2], { visible : false }); var t1 = board.create('tangent', [g1], { strokeColor : '#4260f5', strokeWidth : 2.5, strokeOpacity : 0.6, highlight : false }); var t2 = board.create('tangent', [g2], { strokeColor : '#31d490', strokeWidth : 2.5, strokeOpacity : .8, highlight : false }); board.create('text', [-3.5, 2.8, function() { return 'f(x)=\\color{#4260f5}{x^3-x}'; }], { useMathJax : true, fontSize : 13, strokeColor : 'black', fixed : true, highlight : false }); board.create('text', [-3.5, 2.3, function() { return 'g(x)=c\\color{#31d490}{(x^3-x)}'; }], { useMathJax : true, fontSize : 13, strokeColor : 'black', fixed : true, highlight : false }); board.create('text', [1.5, -2, function() { return 'f\'(x_{0})=\\color{#4260f5}{' + t1.getSlope().toFixed(3) + '}'; }], { useMathJax : true, fontSize : 13, strokeColor : 'black', fixed : true, highlight : false }); board.create('text', [1.5, -2.5, function() { return 'g\'(x_{0})=\\color{#31d490}{' + t2.getSlope().toFixed(3) + '}'; }],{ useMathJax : true, fontSize : 13, strokeColor : 'black', fixed : true, highlight : false }); g0.on('drag', function() { var x = g0.X(); g1.moveTo([x, f(x)]); g2.moveTo([x, g(x)]); board.update(); }); board.fullUpdate(); })(); "; } ], { fontSize : 15 });*/ board.unsuspendUpdate(); } } function addDerivative() { if (JXG.isFunction(f)) { board.create('functiongraph',[JXG.Math.Numerics.D(f), function(){ var c = new JXG.Coords(JXG.COORDS_BY_SCREEN,[0,0],board); return c.usrCoords[1]; }, function(){ var c = new JXG.Coords(JXG.COORDS_BY_SCREEN,[board.canvasWidth,0],board); return c.usrCoords[1]; }], { dash : 2, strokeWidth : 2, strokeColor : '#3ac946' }); } } document.getElementById('plot').onclick = plotter; document.getElementById('clear all').onclick = clearAll; document.getElementById('add tangent').onclick = addTangent; document.getElementById('add Derivative').onclick = addDerivative; })();

5. Taylor polynomial

Taylor polynomial


Example

Compare the graph of \sin x (red) with the graphs of the polynomials 
	x-\frac{x^3}{3!}+\frac{x^5}{5!}-\dots + \frac{(-1)^nx^{2n+1}}{(2n+1)!}
	(blue) for n=1,2,3,\dots,12.

Interaction. The sine function and the polynomial

\displaystyle\sum_{k=0}^{n}\frac{(-1)^{k}x^{2k+1}}{(2k+1)!}

Definition: Taylor polynomial

Let f be k times differentiable at the point x_{0}. Then the Taylor polynomial \begin{align} P_n(x)&=P_n(x;x_0)\\\ &=f(x_0)+f'(x_0)(x-x_0)+\frac{f''(x_0)}{2!}(x-x_0)^2+ \\ & \dots +\frac{f^{(n)}(x_0)}{n!}(x-x_0)^n\\ &=\sum_{k=0}^n\frac{f^{(k)}(x_0)}{k!}(x-x_0)^k\\ \end{align} is the best polynomial approximation of degree n (with respect to the derivative) for a function f, close to the point x_0.

Note. The special case x_0=0 is often called the Maclaurin polynomial.


If f is n times differentiable at x_0, then the Taylor polynomial has the same derivatives at x_0 as the function f, up to the order n (of the derivative).

The reason (case x_0=0): Let 
		P_n(x)=c_0+c_1x+c_2x^2+c_3x^3+\dots +c_nx^n,
	so that \begin{align} P_n'(x)&=c_1+2c_2x+3c_3x^2+\dots +nc_nx^{n-1}, \\ P_n''(x)&=2c_2+3\cdot 2 c_3x\dots +n(n-1)c_nx^{n-2} \\ P_n'''(x)&=3\cdot 2 c_3\dots +n(n-1)(n-2)c_nx^{n-3} \\ \dots && \\ P^{(k)}(x)&=k!c_k + x\text{ terms} \\ \dots & \\ P^{(n)}(x)&=n!c_n \\ P^{(n+1)}(x)&=0. \end{align}

From these way we obtain the coefficients one by one: \begin{align} c_0= P_n(0)=f(0) &\Rightarrow c_0=f(0) \\ c_1=P_n'(0)=f'(0) &\Rightarrow c_1=f'(0) \\ 2c_2=P_n''(0)=f''(0) &\Rightarrow c_2=\frac{1}{2}f''(0) \\ \vdots & \\ k!c_k=P_n^{(k)}(0)=f^{(k)}(0) &\Rightarrow c_k=\frac{1}{k!}f^{(k)}(0). \\ \vdots &\\ n!c_n=P_n^{(n)}(0)=f^{(n)}(0) &\Rightarrow c_k=\frac{1}{n!}f^{(n)}(0). \end{align} Starting from index k=n+1 we cannot pose any new conditions, since P^{(n+1)}(x)=0.

Taylor's Formula

If the derivative f^{(n+1)} exists and is continuous on some interval I=\, ]x_0-\delta,x_0+\delta[, then f(x)=P_n(x;x_0)+E_n(x) and the error term E_n(x) satisfies 
		E_n(x)=\frac{f^{(n+1)}(c)}{(n+1)!}(x-x_0)^{n+1}
	at some point c\in [x_0,x]\subset I. If there is a constant M (independent of n) such that |f^{(n+1)}(x)|\le M for all x\in I, then 
		|E_n(x)|\le \frac{M}{(n+1)!}|x-x_0|^{n+1} \to 0
	as n\to\infty.


\neq omitted here (mathematical induction or integral).


Examples of Maclaurin polynomial approximations: \begin{align} \frac{1}{1-x} &\approx 1+x+x^2+\dots +x^n =\sum_{k=0}^{n}x^k\\ e^x&\approx 1+x+\frac{1}{2!}x^2+\frac{1}{3!}x^3+\dots + \frac{1}{n!}x^n =\sum_{k=0}^{n}\frac{x^k}{k!}\\ \ln (1+x)&\approx x-\frac{1}{2}x^2+\frac{1}{3}x^3-\dots + \frac{(-1)^{n-1}}{n}x^n =\sum_{k=1}^{n}\frac{(-1)^{k-1}}{k}x^k\\ \sin x &\approx x-\frac{1}{3!}x^3+\frac{1}{5!}x^5-\dots +\frac{(-1)^n}{(2n+1)!}x^{2n+1} =\sum_{k=0}^{n}\frac{(-1)^k}{(2k+1)!}x^{2k+1}\\ \cos x &\approx 1-\frac{1}{2!}x^2+\frac{1}{4!}x^4-\dots +\frac{(-1)^n}{(2n)!}x^{2n} =\sum_{k=0}^{n}\frac{(-1)^k}{(2k)!}x^{2k} \end{align}

Example

Which polynomial P_n(x) approximates the function \sin x in the interval [-\pi,\pi] so that the absolute value of the error is less than 10^{-6}?

We use Taylor's Formula for f(x)=\sin x at x_0=0. Then |f^{(n+1)}(c)|\le 1 independently of n and the point c. Also, in the interval in question, we have |x-x_0|=|x|\le \pi. The requirement will be satisfied (at least) if 
		|E_n(x)|\le \frac{1}{(n+1)!}\pi^{n+1} < 10^{-6}.
	This inequality must be solved by trying different values of n; it is true for n\ge 16.

The required approximation is achieved with P_{16}(x), which fo sine is the same as P_{15}(x).

Check from graphs: P_{13}(x) is not enough, so the theoretical bound is sharp!

Taylor polynomial and extreme values


If f'(x_0)=0, then also some higher derivatives may be zero: 
		f'(x_0)=f''(x_0)= \dots = f^{(n)}(x_0) =0,\ f^{(n+1)}(x_0) \neq 0.
	Then the behaviour of f near x=x_0 is determined by the leading term (after the constant term f(x_0)) 
		\frac{f^{(n+1)}(x_0)}{(n+1)!}(x-x_0)^{n+1}.
	of the Taylor polynomial.

This leads to the following result:

Extreme values
  • If n is even, then x_0 is not an extreme point of f.
  • If n is odd and f^{(n+1)}(x_0)>0, then f has a local minimum at x_0.
  • If n is odd and f^{(n+1)}(x_0), then f has a local maximum at x_0.

Newton's method


The first Taylor polynomial P_1(x)=f(x_0)+f'(x_0)(x-x_0) is the same as the linearization of f at the point x_0. This can be used in some simple approximations and numerical methods.

Newton's method

The equation f(x)=0 can be solved approximately by choosing a starting point x_0 (e.g. by looking at the graph) and defining 
	x_{n+1}=x_n-\frac{f(x_n)}{f'(x_n)}
	for n=0,1,2,\dots This leads to a sequence (x_0,x_1,x_2,\dots ), whose terms usually give better and better approximations for a zero of f.


The recursion formula is based on the geometric idea of finding an approximative zero of f by using its linearization (i.e. the tangent line).

Example

Find an approximate value of \sqrt{2} by using Newton's method.

We use Newton's method for the function f(x)=x^2-2 and initial value x_0=2. The recursion formula becomes 
	x_{n+1}= x_n-\frac{x_n^2-2}{2x_n} = \frac{1}{2}\left(x_n+\frac{2}{x_n}\right),
	from which we obtain x_1=1{,}5, x_2\approx 1{,}41667, x_3\approx 1{,}4142157 and so on.

By experimenting with these values, we find that the number of correct decimal places doubles at each step, and x_7 gives already 100 correct decimal places, if intermediate steps are calculated with enough precision.

Taylor series


Taylor series

If the error term E_n(x) in Taylor's Formula goes to zero as n increases, then the limit of the Taylor polynomial is the Taylor series of f (= Maclaurin series for x_0=0).

The Taylor series of f is of the form 
		\sum_{k=0}^{\infty}\frac{f^{(k)}(x_0)}{k!}(x-x_0)^k 
		= \lim_{n\to\infty} \sum_{k=0}^{n}\frac{f^{(k)}(x_0)}{k!}(x-x_0)^k .
	This is an example of a power series.


The Taylor series can be formed as soon as f has derivatives of all orders at x_0 and they are substituted into this formula. There are two problems related to this: Does the Taylor series converge for all values of x?

Answer: Not always; for example, the function 
	f(x)=\frac{1}{1-x}
	has a Maclaurin series (= geometric series) converging only for -1 < x < 1, although the function is differentiable for all x\neq 1: 
	f(x)=\frac{1}{1-x} = 1+x+x^2+x^3+x^4+\dots

Interaction. Newton's method. Set the starting point x_{0} and iterate to find the zeros of the function.
x_{0}=~

If the series converges for some x, then does its sum equal f(x)? Answer: Not always; for example, the function 
	f(x)=\begin{cases}
	e^{-1/x^2}, & x\neq 0,\\
	0, & x=0,\\
	\end{cases}
	satisfies f^{(k)}(0)=0 for all k\in \mathbf{N} (elementary but difficult calculation). Thus its Maclaurin series is identically zero and converges to f(x) only at x=0.

Conclusion: Taylor series should be studied carefully using the error terms. In practice, the series are formed by using some well known basic series.

The graph of e^{-1/x^2}
Examples

\begin{align} \frac{1}{1-x} &= \sum_{k=0}^{\infty} x^k,\ \ |x|< 1 \\ e^x &= \sum_{k=0}^{\infty} \frac{1}{k!}x^k, \ \ x\in \mathbb{R} \\ \sin x &= \sum_{k=0}^{\infty} \frac{(-1)^{k}}{(2k+1)!} x^{2k+1}, \ \ x\in \mathbb{R} \\ \cos x &= \sum_{k=0}^{\infty} \frac{(-1)^{k}}{(2k)!} x^{2k},\ \ x\in \mathbb{R} \\ (1+x)^r &= 1+\sum_{k=1}^{\infty} \frac{r(r-1)(r-2)\dots (r-k+1)}{k!}x^k, |x|<1 \end{align} The last is called the Binomial Series and is valid for all r\in \mathbb{R}. If r=n \in \mathbb{N}, then starting from k=n+1, all the coefficients are zero and in the beginning 
	\binom{n}{k} =\frac{n!}{k!(n-k)!} = \frac{n(n-1)(n-2)\dots (n-k+1)}{k!}.

Compare this to the Binomial Theorem: 
	(a+b)^n=\sum_{k=0}^n\binom{n}{k} a^{n-k}b^k
	=a^n +na^{n-1}b+\dots +b^n
	for n\in\mathbb{N}.

Power series


Definition: Power series

A power series is of the form 
	\sum_{k=0}^{\infty} c_k(x-x_0)^k 
	= \lim_{n\to\infty} \sum_{k=0}^{n}c_k(x-x_0)^k.
	The point x_0 is the centre and the c_k are the coefficients of the series.

The series converges at x if the above limit is defined.

There are only three essentially different cases:

Abel's Theorem.
  • The power series converges only for x=x_0 (and then it consists of the constant c_0 only)
  • The power series converges for all x\in \mathbb{R}
  • The power series converges on some interval ]x_0-R,x_0+R[ (and possibly in one or both of the end points), and diverges for other values of x.

The number R is the radius of convergence of the series. In the first two cases we say that R=0 or R=\infty respectively.

Example

For which values of the variable x does the power series \sum_{k=1}^{\infty} \frac{k}{2^k}x^k converge?

We use the ratio test with a_k=kx^k/2^k. Then 
	\left| \frac{a_{k+1}}{a_k} \right| = 
	\left| \frac{(k+1)x^{k+1}/2^{k+1}}{kx^k/2^k} \right|
	= \frac{k+1}{2k}|x| \to \frac{|x|}{2}
	as k\to\infty. By the ratio test, the series converges for |x|/2, and diverges for |x|/2>1. In the border-line cases |x|/2= 1\Leftrightarrow x=\pm 2 the general term of the series does not tend to zero, so the series diverges.

Result: The series converges for -2< x< 2, and diverges otherwise.

Definition: Sum function

In the interval I where the series converges, we can define a function f\colon I\to \mathbb{R} by setting \begin{equation} \label{summafunktio} f(x) = \sum_{k=0}^{\infty} c_k(x-x_0)^k, \tag{1} \end{equation} which is called the sum function of the power series.

The sum function f is continuous and differentiable on ]x_0-R,x_0+R[. Moreover, the derivative f'(x) can be calculated by differentiating the sum function term by term: 
	f'(x)=\sum_{k=1}^{\infty}kc_k(x-x_0)^{k-1}.
	Note. The constant term c_0 disappears and the series starts with k=1. The differentiated series converges in the same interval x\in \, ]x_0-R,x_0+R[; this may sound a bit surprising because of the extra coefficient k.

Example

Find the sum function of the power series 1+2x+3x^2+4x^3+\dots

This series is obtained by differentiating termwise the geometric series (with q=x). Therefore, \begin{align} 1+2x+3x^2+4x^3+\dots &= D(1+x+x^2+x^3+x^4+\dots ) \\ &= \frac{d}{dx}\left( \frac{1}{1-x}\right) = \frac{1}{(1-x)^2}. \end{align} Multiplying with x we obtain 
	\sum_{k=1}^{\infty}kx^{k} = x+2x^2+3x^3+4x^4+\dots = \frac{x}{(1-x)^2},
	which is valid for |x|.

In the case [a,b]\subset\ ]x_0-R,x_0+R[ we can also integrate the sum function termwise: 
	\int_a^b f(x)\, dx = \sum_{k=0}^{\infty}c_k\int_a^b (x-x_0)^k\, dx.
	Often the definite integral can be extended up to the end points of the interval of convergence, but this is not always the case.

Example

Calculate the sum of the alternating harmonic series.

Let us first substitute q=-x to the geometric series. This yields 
	1-x+x^2-x^3+x^4-\dots =\frac{1}{1-(-x)} = \frac{1}{1+x}.
	By integrating both sides from x=0 to x=1 we obtain 
	1-\frac{1}{2}+\frac{1}{3}-\frac{1}{4}+\dots =\int_0^1\frac{1}{1+x} =\ln 2.
	Note. Extending the limit of integration all the way up to x=1 should be justified more rigorously here. We shall return to integration later on the course.

'; } else { return ''; } } ], { useMathJax : true, fontSize : 14 }); function iterate(){ var x = 0; if(i>10){ return; } if(i==0){ x = document.getElementById("x0").value; if(isNaN(x) || x==""){ alert("Invalid input"); return; } x = Number(x); document.getElementById("x0").disabled = true; } else{ x = points[i-1].X(); } var fx = f(x); slope = derivative(x); var intersection = fx - slope*x; segments[i] = board.create('segment',[[x,0],[x,fx]],{ dash : 1, highlight : false, strokeOpacity : .6, strokeWidth : 1.5, strokeColor : '#32a889' }); tangents[i] = board.create('line',[[x, fx],[x+1, slope+fx]],{ fixed : true, highlight : false, strokeOpacity : .6, strokeWidth : 1.5, strokeColor : '#32a889' }); points[i] = board.create('point',[-intersection/slope,0],{ fixed : true, highlight : false, showInfobox : false, strokeWidth : 1, strokeColor : 'black', fillColor : '#7732a8', name : 'x_{' + (i+1) + '}', label : { useMathJax : true } }); board.update(); i++; } function reset(){ i = 0; document.getElementById("x0").disabled = false; document.getElementById("x0").value = ""; board.removeObject(points), board.removeObject(segments), board.removeObject(tangents); points = [], segments = [], tangents = []; board.fullUpdate(); } document.getElementById("iterate").onclick = iterate; document.getElementById("reset").onclick = reset; board.fullUpdate(); })();

6. Elementary functions

This chapter gives some background to the concept of a function. We also consider some elementary functions from a (possibly) new viewpoint. Many of these should already be familiar from high school mathematics, so in some cases we just list the main properties.

Functions


Definition: Function

A function f\colon A\to B is a rule that determines for each element a\in A exactly one element b\in B. We write b=f(a).


Definition: Domain and codomain

In the above definition of a function A=D_f is the domain (of definition) of the function f and B is called the codomain of f.


Definition: Image of a function

The image of f is the subset f[A]= \{ f(a) \mid a\in A\} of B. An alternative name for image is range.


For example, f\colon \mathbb{R}\to\mathbb{R}, f(x)=x^2, has codomain \mathbb{R}, but its image is f[\mathbb{R} ] =[0,\infty[.

The function in the previous example can also be defined as f\colon \mathbb{R}\to [0,\infty[, f(x)=x^2, and then the codomain is the same as the image. In principle, this modification can always be done, but it is not reasonable in practice.

Example: Try to do the same for f\colon \mathbb{R}\to\mathbb{R}, f(x)=x^6+x^2+x, x\in\mathbb{R}.

  • If the domain A\subset \mathbb{R} then f is a function of one (real) variable: the main object of study in this course.

  • If A\subset \mathbb{R}^n, n\ge 2, then f is a function of several variables (a multivariable function)

Inverse functions


Definition: Injection, surjection and bijection
A function f\colon A \to B is
  • injective (one-to-one) if it has different values at different points; i.e. x_1\neq x_2 \Rightarrow f(x_1)\neq f(x_2), or equivalently f(x_1)= f(x_2) \Rightarrow x_1=x_2.
  • surjective (onto) if its image is the same as codomain, i.e. f[A]=B
  • bijective (one-to-one and onto) if it is both injective and surjective.

Observe: A function becomes surjective if all redundant points of the codomain are left out. A function becomes injective if the domain is reduced so that no value of the function is obtained more than once.

Another way of defining these concepts is based on the number of solutions to an equation:

Definition

For a fixed y\in B, the equation y=f(x) has

  • at most one solution x\in A if f is injective
  • at least one solution x\in A if f is surjective
  • exactly one solution x\in A if f on bijective.

Definition: Inverse function

If f\colon A \to B is bijective, then it has an inverse f^{-1}\colon B \to A, which is uniquely determined by the condition y=f(x) \Leftrightarrow x = f^{-1}(y).


The inverse satisfies f^{-1}(f(a))=a for all a\in A and f(f^{-1}(b))=b for all b\in B.

The graph of the inverse is the mirror image of the graph of f with respect to the line y=x: A point (a,b) lies on the graph of f \Leftrightarrow b=f(a) \Leftrightarrow a=f^{-1}(b) \Leftrightarrow the point (b,a) lies on the graph of f^{-1}. The geometric interpretation of (a,b)\mapsto (b,a) is precisely the reflection with respect to y=x.

If A \subset \mathbb{R} and f\colon A\to \mathbb{R} is strictly monotone, then the function f\colon A \to f[A] has an inverse.

If here A is an interval and f is continuous, then also f^{-1} is is continuous in the set f[A].

Theorem: Derivative of the inverse

Let f\colon \, ]a,b[\, \to\, ]c,d[ be differentiable and bijective, so that it has an inverse f^{-1}\colon \, ]c,d[\, \to\, ]a,b[. As the graphs y=f(x) and y=f^{-1}(x) are mirror images of each other, it seems geometrically obvious that also f^{-1} is differentiable, and we actually have 
	\left(f^{-1}\right)'(x)=\frac{1}{f'(f^{-1}(x))},
	if f'(f^{-1}(x))\neq 0.

Proof.

Differentiate both sides of the equation \begin{align} f(f^{-1}(x)) &= x \\ \Rightarrow f'(f^{-1}(x))\left(f^{-1}\right)'(x) &= Dx = 1, \end{align} and solve for \left(f^{-1}\right)'(x).

\square

Note. f'(f^{-1}(x)) is the derivative of f at the point f^{-1}(x).

one-to-one
1. f\colon A\to B is one-to-one but not onto

onto
2. f\colon A\to B is onto but not one-to-one

one-to-one and onto
3. f\colon A\to B is one-to-one and onto

Transcendental functions


Trigonometric functions


  • Unit of measurement of an angle = rad: the arclength of the arc on the unit circle, that corresponds to the angle.

  • \pi rad = 180 degrees, i.e. 1 rad = 180/\pi \approx 57,\! 3 degrees

  • The functions \sin x, \cos x are defined in terms of the unit circle so that (\cos x,\sin x), x\in [0,2\pi], is the point on the unit circle corresponding to the angle x\in\mathbb{R}, measured counterclockwise from the point (1,0). \tan x = \frac{\sin x}{\cos x}\ (x\neq \pi /2 +n\pi), \cot x = \frac{\cos x}{\sin x}\ (x\neq n\pi)

  • Periodicity: \sin (x+2\pi) = \sin x,\ \cos (x+2\pi)=\cos x, \tan (x+\pi) = \tan x

  • Basic properties (from the unit circle!)
  • \sin 0 = 0, \sin (\pi/2)=1

  • \cos 0=1, \cos (\pi/2)= 0

  • Parity: \sin and \tan are odd functions, \cos is an even function: \sin (-x) = -\sin x, \cos(-x)  = \cos x, \tan (-x) = -\tan x.

  • \sin^2 x + \cos^2 x = 1 for all x\in\mathbb{R}

    Proof: Pythagorean Theorem.

  • Addition formulas:

    \sin (x+y) = \sin x \cos y +\cos x \sin y

    \cos (x+y) = \cos x \cos y -\sin x \sin y

  • Proof: Geometrically, or more easily with vectors and matrices.

    Derivatives: 
	D(\sin x) = \cos x,\ \ D(\cos x) = -\sin x

Interactivity. The connection between the unit circle and the trigonometric functions.
Example

It follows that the functions y(t)=\sin (\omega t) and y(t)=\cos (\omega t) satisfy the differential equation 
	y''(t)+\omega^2y(t)=0,
	that models harmonic oscillation. Here t is the time variable and the constant \omega>0 is the angular frequency of the oscillation. We will see later that all the solutions of this differential equation are of the form 
	y(t)=A\cos (\omega t) +B\sin (\omega t),
	with A,B constants. They will be uniquely determined if we know the initial location y(0) and the initial velocity y'(0). All solutions are periodic and their period is T=2\pi/\omega.

Harmonic oscillator y(t) = y_{0}\cos(\omega t),
where t is the elapsed time in seconds

Arcus functions


The trigonometric functions have inverses if their domain and codomains are chosen in a suitable way.

  • The Sine function 
	\sin \colon [-\pi/2,\pi/2]\to [-1,1]
	is strictly increasing and bijective.

  • The Cosine function 
	\cos \colon [0,\pi] \to [-1,1]
	is strictly decreasing and bijective.

  • The tangent function 
	\tan \colon ]-\pi/2,\pi/2[\, \to \mathbb{R}
	is strictly increasing and bijective.

Arcus functions

Inverses: \arctan \colon \mathbb{R}\to \ ]-\pi/2,\pi/2[, \arcsin \colon [-1,1]\to [-\pi/2,\pi/2], \arccos \colon [-1,1]\to [0,\pi]

This means: x = \tan \alpha \Leftrightarrow \alpha = \arctan x \ \ \text{for } \alpha \in \ ]-\pi/2,\pi/2[ x = \sin \alpha \Leftrightarrow \alpha = \arcsin x \ \ \text{for } \alpha \in \, [-\pi/2,\pi/2] x = \cos \alpha \Leftrightarrow \alpha = \arccos x \ \ \text{for } \alpha \in \, [0,\pi]

Note. Values of the arcus functions should be given in radians, unless we are considering some geometrical applications.

The graphs of \tan and \arctan.

Derivatives of the arcus functions

D \arctan x = \frac{1}{1+x^2},\ x\in \mathbb{R} \tag{1} D\arcsin x = \frac{1}{\sqrt{1-x^2}},\ -1 < x < 1 \tag{2} D\arccos x =  \frac{-1}{\sqrt{1-x^2}},\ -1 < x < 1 \tag{3}

Note. The first result is very useful in integration.

Proof.

Here we will only prove the first result (1). By differentiating both sides of the equation \tan(\arctan x)=x for x\in \mathbb{R}: \bigl( 1+\tan^2(\arctan x)\bigr) \cdot D(\arctan x) = D x = 1 \Rightarrow  D(\arctan x)= \frac{1}{1+\tan^2(\arctan x)} =\frac{1}{1+x^2}.

The last row follows also directly from the formula for the derivative of an inverse.

Example

Show that 
		\arcsin x +\arccos x =\frac{\pi}{2}
	for -1\le x\le 1.

Example

Derive the addition formula for tan, and show that 
	\arctan x+\arctan y = \arctan \frac{x+y}{1+xy}.

Solutions: Voluntary exercises. The first can be deduced by looking at a rectangular triangle with the length of the hypotenuse equal to 1 and one leg of length x.

Introduction: Radioactive decay

Let y(t) model the number of radioactive nuclei at time t. During a short time interval \Delta t the number of decaying nuclei is (approximately) directly proportional to the length of the interval, and also to the number of nuclei at time t: 
	\Delta y = y(t+\Delta t)-y(t) \approx -k\cdot y(t)\cdot \Delta t.
	The constant k depends on the substance and is called the decay constant. From this we obtain 
	\frac{\Delta y}{\Delta t} \approx -ky(t),
	and in the limit as \Delta t\to 0 we end up with the differential equation y'(t)=-ky(t).

Exponential function


Definition: Euler's number

Euler's number (or Napier's constant) is defined as e = \lim_{n\to \infty} \left( 1+\frac{1}{n}\right) ^n = 1+1+\frac{1}{2!}+\frac{1}{3!} +\frac{1}{4!}
	+\dots \approx 2,\! 718281828459\dots


Definition: Exponential function

The Exponential function exp: 
	\exp (x) = \sum_{k=0}^{\infty} \frac{x^k}{k!}=
	\lim_{n\to \infty} \left( 1+\frac{x}{n}\right) ^n = e^x.
	This definition (using the series expansion) is based on the conditions \exp'(x)=\exp(x) and \exp(0)=1, which imply that \exp^{(k)}(0)=\exp(0)= 1 for all k\in\mathbb{N}, so the Maclaurin series is the one above.


The connections between different expressions are surprisingly tedious to prove, and we omit the details here. The main steps include the following:

  • Define \exp\colon\mathbb{R}\to\mathbb{R}, 
	\exp (x) =\sum_{k=0}^{\infty}\frac{x^k}{k!}.
	This series converges for all x\in\mathbb{R} (ratio test).

  • Show: exp is differentiable and satisfies \exp'(x)=\exp(x) for all x\in \mathbb{R}. (This is the most difficult part but intutively rather obvious, because in practice we just differentiate the series term by term like a polynomial.)

  • It has the following properties \exp (0)=1, 
	\exp (-x)=1/\exp (x) \text{ and }
	\exp (x+y)=\exp (x)\,  \exp(y)
	for all x,y\in \mathbb{R}.

    These imply that \exp (p/q)=(\exp (1))^{p/q} for all rational numbers p/q\in \mathbf{Q}.

    By continuity 
	\exp (x) =(\exp (1))^x
	for all x\in \mathbb{R}.

    Since 
	\exp (1) = \sum_{k=0}^{\infty}\frac{1}{k!}
	=\lim_{n\to \infty} \left( 1+\frac{1}{n}\right) ^n=e,
	we obtain the form e^x.

    \square?

Corollary

It follows from above that \exp\colon\mathbb{R}\to\, ]0,\infty[ is strictly increasing, bicective, and 
	\lim_{x\to\infty}\exp(x) = \infty,\ 
	\lim_{x\to-\infty}\exp(x) = 0,\
	\lim_{x\to\infty}\frac{x^n}{\exp (x)} = 0 \text{ for all } n\in \mathbf{N}.


From here on we write e^x=\exp(x). Properties:

  • e^0 = 1
  • e^x >0
  • D(e^x) = e^x
  • e^{-x} = 1/e^x
  • (e^x)^y = e^{xy}
  • e^xe^y =e^{x+y}
for all x,y\in \mathbb{R}.

Differential equation y'=ky

Theorem

Let k\in\mathbb{R} be a constant. All solutions y=y(x) of the ordinary differenial equation (ODE) 
	y'(x)=ky(x),\ x\in \mathbb{R},
	are of the form y(x)=Ce^{kx}, where  C is a constant. If we know the value of y at some point x_0, then the constant C will be uniquely determined.

Proof.

Suppose that y'(x)=ky(x). Then D(y(x)e^{-kx})= y'(x)e^{-kx}+y(x)\cdot (-ke^{-kx}) = ky(x)e^{-kx}-ky(x)e^{-kx}=0 for all x\in\mathbf{R}, so that y(x)e^{-kx}=C= constant. Multiplying both sides with e^{kx} we obtain y(x)=Ce^{kx}.

\square

Euler's formula

Definition: Complex numbers

Imaginary unit i: a strange creature satisfying i^2=-1. The complex numbers are of the form z=x+iy, where x,y\in \mathbb{R}. We will return to these later.


Theorem: Euler's formula

If we substitute ix as a variable in the expontential fuction, and collect real terms separately, we obtain Euler's formula e^{ix}=\cos x+i\sin x.

Proof.

Substitute x=ix in the definition of the exponential function and write the series as the sum of its even (n=2k) and odd (n=2k+1) parts. Note that i^{2k} = (i^2)^k = (-1)^{k} and remember the Taylor series of the trigonometric functions.

\square

As a special case we have Euler's identity e^{i\pi}+1=0. It connects the most important numbers 0, 1, i, e ja \pi and the three basic operations sum, multiplication, and power.

Using e^{\pm ix}=\cos x\pm i\sin x we can also derive the expressions 
	\cos x=\frac{1}{2}\bigl( e^{ix}+e^{-ix}\bigr),\
	\sin x=\frac{1}{2i}\bigl( e^{ix}-e^{-ix}\bigr),
	\ x\in\mathbb{R}.

The graphs of \exp(x) and the partial sums \displaystyle\sum_{k=0}^{n}\frac{x^{k}}{k!}

Logarithms


Definition: Natural logarithm

Natural logarithm is the inverse of the exponential function: 
	\ln\colon \ ]0,\infty[ \ \to \mathbb{R}


Note. The general logarithm with base a is based on the condition 
	a^x = y \Leftrightarrow x=\log_a y
	for a>0 and y>0.

Beside the natural logarithm, in applications also appear the Briggs logarithm with base 10: \lg x = \log_{10} x, and the binary logarithm with base 2: {\rm lb}\, x =\log_{2} x.

Usually (e.g. in mathematical software) \log x is the same as \ln x.

Properties of the logarithm:

  • e^{\ln x} = x for x>0
  • \ln (e^x) =x for x\in\mathbb{R}
  • \ln 1=0, \ln e = 1
  • \ln (a^b) = b\ln a if a>0, b\in\mathbb{R}
  • \ln (ab) = \ln a+\ln b, if a,b>0
  • D\ln |x|=1/x for x\neq 0
  • These follow from the corresponding properties of exp.

    Example

    Substituting x=\ln a and y=\ln b to the formula

    e^xe^y =e^{x+y} we obtain ab =e^{\ln a+\ln b},

    so that \ln (ab) = \ln a +\ln b.

The graph of \ln

Hyperbolic functions


Definition: Hyperbolic functions

Hyperbolic sine sinus hyperbolicus \sinh, hyperbolic cosine cosinus hyperbolicus \cosh and hyperbolic tangent \tanh are defined as \sinh \colon \mathbb{R}\to\mathbb{R}, \ \sinh x=\frac{1}{2}(e^x-e^{-x}) \cosh \colon \mathbb{R}\to [1,\infty[,\ \cosh x=\frac{1}{2}(e^x+e^{-x}) \tanh \colon \mathbb{R}\to \ ]-1,1[, \ \tanh x =\frac{\sinh x}{\cosh x}


Properties: \cosh^2x-\sinh^2x=1; all trigonometric have their hyperbolic counterparts, which follow from the properties \sinh (ix)=i\sin x, \cosh (ix)=\cos x. In these formulas, the sign of \sin^2 will change, but the other signs remain the same.

Derivatives: D\sinh x=\cosh x, D\cosh x=\sinh x.

Hyperbolic inverse functions: the so-called area functions; area and the shortening ar refer to a certain geometrical area related to the hyperbola x^2-y^2=1: \sinh^{-1}x=\text{arsinh}\, x=\ln\bigl( x+\sqrt{1+x^2}\, \bigr) ,\ x\in\mathbb{R} \cosh^{-1}x=\text{arcosh}\, x=\ln\bigl( x+\sqrt{x^2-1}\, \bigr) ,\ x\ge 1

Derivatives of the inverse functions: D \sinh^{-1}x= \frac{1}{\sqrt{1+x^2}} ,\ x\in\mathbb{R} D \cosh^{-1}x= \frac{1}{\sqrt{x^2-1}} ,\ x > 1.

The graph of \cosh
The graph of \sinh
The graph of \tanh
'; }], { useMathJax : true, anchor : g, fixed : true }); var s1 = board.create('segment', [[0, 0], [1, 0]], { fixed : true, highlight : false, dash : 2, strokeWidth : 2, strokeColor : 'black' }); var s2 = board.create('segment', [[0, 0], function() { return [g.X(), g.Y()]; }], { fixed : true, highlight : false, dash : 2, strokeWidth : 2, strokeColor : 'black' }); var x = board.create('angle', [s1, s2, 1, 1], { name : 'x', radius : .4, label : { useMathJax : true, strokeColor : 'black', offset : [0, .5] } }); var info = board.create('text', [-1.2, -1.25, function() { return '(\\cos x,\\sin x)=(' + Math.cos(x.Value()).toFixed(3) + ',' + Math.sin(x.Value()).toFixed(3) + ')'; }], { useMathJax : true, fixed : true }); board.fullUpdate(); })();

7. Area

Area in the plane


We consider areas of plane sets bounded by closed curves. In the more general cases, the concept of area becomes theoretically very difficult.

The area of a planar set is defined by reducing to the areas of simpler sets. The area cannot be "calculated", unless we first have a definition of "area" (although this is common practice in school mathematics).

Starting point

The area of a rectangle
The area of a rectangle is base \times height: A=ab.

A rectangle
Definition: Area of a Parallelogram

The area of a parallelogram is base \times height: 
		 A=ah.


parallelogram
Definition: Area of a triangle

The area of a triangle is (by definition) 
		 A=\frac{1}{2}ah.


triangle

Polygon

A (simple) polygon is a plane set bounded by a closed curve that consists of a finite number of line segments without self-intersections.

polygon
Definition: Area of a polygon

The area of a polygon is defined by dividing it into a finite number of triangles (called a triangulation of the polygon) and adding the areas of these triangles.


triangulation
Theorem.

The sum of the areas of  triangles in a triangulation of a polygon is the same for all triangulations.


General case

For a plane set \color{red} D bounded by a closed curve we can construct inner polygons \color{blue}P_i and outer polygons P_o: \color{blue}P_i\color{black} \subset \color{red}D\color{black}\subset P_o.

A bounded set D has an area if for every \varepsilon >0 there is an inner polygon P_i and an outer polygon P_o, whose areas differ by less than \varepsilon: 
	A(P_o)-A(P_i) This implies that between all areas A(P_i) and A(P_o) there is a unique real number A(D), which is (by definition) the area of D.

Inner and outer polygons

A surprise: The condition that D is bounded by a closed curve (without self-intersections) does not guarantee that it has an area! Reason: The boundary curve can be so "wiggly", that it has positive "area". The first such example was constucted by [W.F. Osgood, 1903]:

Wikipedia: Osgood curve

Example

Derive the formula A=\pi R^2 for a circle with radius R by choosing regular inscrided and circumscribed n-gons as inner and outer polygons, and let n\to\infty.

The solution is a voluntary exercise, where you need the limit \lim_{x\to 0}\frac{\sin x}{x} = 1. Hint: Show that the inscribed and circumscribed areas are 	\pi R^2\frac{\sin (2\pi/n)}{2\pi/n} \ \text{ and }\ \pi R^2\frac{\tan \pi/n}{\pi/n}.

8. Integral

From sum to integral


Definite integral

Geometric interpretation: Let f\colon[a,b]\to\mathbb{R} be such that f(x)\ge 0 for all x\in[a,b]. How can we find the area of the region bounded by the function graph y=f(x), the x-axis and the two lines x=a and x=b?

The answer to this question is given by the definite integral \int_{a}^{b}f(x)\,dx Remark. The general definition of the integral does not necessitate the condition f(x)\ge 0.

Integration of continuous functions

Definition: Partition

Let f\colon[a,b]\to\mathbb{R} be continuous. A finite sequence D=(x_{0},x_{1},x_{2},\dots,x_{n}) of real numbers such that a=x_{0} < x_{1} < x_{2} < \dots < x_{n} = b is called a partition of the interval [a,b].


Geometric interpretation of the definite integral of f from x=a to x=b
Definition: Upper and lower sum

For each partition D we define the related upper sum of the function f as U_{D}(f) = \sum_{k=1}^{n}M_{k}(x_{k}-x_{k-1}),~M_{k} = \max\{f(x)\mid x_{k-1}\le x\le x_{k}\} and the lower sum as L_{D}(f) = \sum_{k=1}^{n}m_{k}(x_{k}-x_{k-1}),~m_{k}=\min\{f(x)\mid x_{k-1}\le x\le x_{k}\}.


If f is a positive function then the upper sum represents the total area of the rectangles circumscribing the function graph and similarly the lower sum is the total area of the inscribed rectangles.

Properties of partitions
  1. Suppose that D_{1} and D_{2} are two partitions of a given interval such that D_{1} is a subsequence of D_{2} (i.e. D_{2} is finer than D_{1}). Then the inequalities

    U_{D_1}(f) \ge U_{D_{2}}(f)~ and ~L_{D_{1}}(f) \le L_{D_{2}}(f)

    always hold.
  2. For any two partitions D_{1} and D_{2} of a given interval the inequality  L_{D_{2}}(f) \le U_{D_{1}}(f) always holds.

Interactive.

f(x)=~

Upper Darboux sumLower Darboux sum
Definition: Integrability

We say that a function f\colon[a,b]\to\mathbb{R} is integrable if for every \epsilon>0 there exists a corresponding partition D of [a,b] such that  U_{D}(f) - L_{D}(f) < \epsilon.


Definition: Integral

Integrability implies that there exists a unique real number I such that L_{D}(f)\le I\le U_{D}(f) for every partition D. This is called the integral of f over the interval [a,b] and denoted by  I = \int_{a}^{b}f(x)\,dx.


Remark. This definition of the integral is sometimes referred to as the Darboux integral.

For non-negative functions f this definition of the integral coincides with the idea of making the difference between the the areas of the circumscribed and the inscribed rectangles arbitrarily small by using ever finer partitions.

Theorem.

A continuous function on a closed interval is integrable.

Proof.

Here we will only provide the proof for continuous functions with bounded derivatives.

Suppose that f\colon[a,b]\to\mathbb{R} is a continuous function and that there exists a constant L>0 such that |f'(x)|\le L for all x\in]a,b[. Let \epsilon>0 and define D to be an equally spaced partition of [a,b] such that \underbrace{|x_{k}-x_{k-1}|}_{=\Delta x} < \frac{\epsilon}{L(b-a)},~\text{for all} k=1,2,\dots,n. Let f(y_{k})=m_{k} and f(z_{k})=M_{k} for some suitable points y_{k},z_{k}\in[x_{k-1},x_{k}]. The mean value theorem then states that M_{k}-m_{k}=f'(c_{k})|z_{k}-y_{k}|\le L\Delta x and thus U_{D}(f)-L_{D}(f) = \sum_{k=1}^{n}(M_{k}-m_{k})\Delta x < \frac{\epsilon}{b-a}\sum_{k=1}^{n}\Delta x = \epsilon.

\square


Definition: Riemann integral

Suppose that f\colon[a,b]\to\mathbb{R} is a continuous function and let (x_{0},x_{1},\dots,x_{n}) be a partition of \left[a,b\right] and (z_{1},z_{2},\dots,z_{n}) be a sequence of real numbers such that z_{k}\in[x_{k-1},x_{k}] for all 1\le k\le n. The partial sums 	S_{n} = \sum_{k=1}^{n}f(z_{k})\Delta x_{k},~\text{where}
	  ~\Delta x_{k}=x_{k}-x_{k-1} are called the Riemann sums of f. Suppose further that the partitions are such that \displaystyle\max_{1\le k\le n}\Delta x_{k}\to 0 as n\to\infty. The integral of f can then be defined as the limit 	\int_{a}^{b}f(x)\,\mathrm{d}x = \lim_{n\to\infty} S_{n}.	This definition of the integral is called the Riemann integral.

Remark. This definition of the integral turns out to be equivalent to that of the Darboux integral i.e. a function is Riemann-integrable if and only if it is Darboux-integrable and the values of the two integrals are always equal.

Example

Find the integral of f(x)=x over the interval [0,1] using Riemann sums.

Let x_{k}=k/n. Then x_{0}=0, x_{n}=1 and x_{k} < x_{k+1} for all 0\le k\le n. Thus the sequence (x_{0},x_{1},x_{2},\dots,x_{n}) is a proper partition of \left[0,1\right]. This partition has the pleasant property hat \Delta x=1/n is a constant. Estimating the Riemann sums we now find that \sum_{k=1}^{n}f(x_{k})\Delta x = \sum_{k=1}^{n}x_{k}\Delta x= 
	\sum_{k=1}^{n}\frac{k}{n}\left(\frac{1}{n}\right) = \frac{1}{n^2}\sum_{k=1}^{n}k = \frac{1}{n^2}\frac{n(n+1)}{2} = \frac{n+1}{2n}\to 
	\frac{1}{2}, as n\to\infty and hence \int_{0}^{1}f(x)\,\mathrm{d}x = \frac{1}{2}.

This is of course the area of the triangular region bounded by the line y=x, the x-axis and the lines x=0 and x=1.

Remark. Any interval [a,b] can be partitioned into equally spaced subintervals by setting \Delta x = (b-a)/n and x_{k} = a + k\Delta x.

Conventions
  1. If the upper and lower limits of integration are the same then the integral is zero:  \int_{a}^{a}f(x)\,dx = 0.
  2. Reversing the limits of integration changes the sign of the integral:  \int_{b}^{a}f(x)\,dx = -\int_{a}^{b}f(x)\,dx.
  3. It also follows that  \int_{a}^{b}f(x)\,dx = \int_{a}^{c}f(x)\,dx + \int_{c}^{b}f(x)\,dx holds for all a,b,c\in\mathbb{R}.

Piecewise-defined functions

Definition: Piecewise continuity

A function f\colon\left[a,b\right]\to\mathbb{R} is called piecewise continuous if it is continuous except at a finite number of points a\le c_{1} < c_{2} < \dots < c_{m} \le b and the one-sided limits of the function are defined and bounded on each of these points. It follows that the restriction of f on each subinterval \left[c_{k-1},c_{k}\right] is continuous if the one-sided limits are taken to be the values of the function at the end points of the subinterval.


Definition: Piecewise integration

Let f\colon\left[a,b\right] be a piecewise continuous function. Then \int_{a}^{b}f(x)\,dx = \sum_{k=1}^{m+1}\int_{c_{k-1}}^{c_{k}}f(x)\,dx, where a=c_{0}< c_{1} < \dots < c_{m+1} = b and f is thought as a continuous function on each subinterval \left[c_{k-1},c_{k}\right]. Usually functions which are continuous yet piecewise defined are also integrated using the same idea.


Example

Consider the function f\colon\left[-1,1\right] defined as 
	f(x) =
	\begin{cases}
	-1 &\text{ for }-1\le x We can now integrate f as follows: 
	\int_{-1}^{1}f(x)\,dx = \int_{-1}^{0}f(x)\,dx + \int_{0}^{1}f(x)\,dx
	
	=\int_{-1}^{0}(-1)\,dx + \int_{0}^{1}1\,dx = -1\cdot(-1-0) + 1\cdot(1-0) = 2.

Integral of the function f(x) =\begin{cases}
	-1 &\text{ for }-1\le x

Important properties


Properties

Suppose that f,g\colon\left[a,b\right]\to\mathbb{R} are piecewise continuous functions. The integral has the following properties

  1. Linearity: If c_{1},c_{2}\in\mathbb{R} then \int_{a}^{b}\big(c_{1}f(x)+c_{2}g(x)\big)\,\mathrm{d}x = c_{1}\int_{a}^{b}f(x)\,\mathrm{d}x+c_{2}\int_{a}^{b}g(x)\,\mathrm{d}x.
  2. If h(x)\ge 0 for all x\in[a,b] then \int_{a}^{b}h(x)\,\mathrm{d}x \ge 0.
  3. If f(x)\le g(x) then \int_{a}^{b}f(x)\,\mathrm{d}x \le \int_{a}^{b}g(x)\,\mathrm{d}x.
  4. As f(x)\le|f(x)| it follows that \int_{a}^{b}f(x)\,\mathrm{d}x \le \int_{a}^{b}|f(x)|\,\mathrm{d}x and taking the absolute value of both sides of the equation gives \left|\int_{a}^{b}f(x)\,\mathrm{d}x\right|\le \int_{a}^{b}|f(x)|\,\mathrm{d}x.
  5. Suppose that p=\inf_{x\in\left[a,b\right]}f(x) and s=\sup_{x\in\left[a,b\right]}f(x). Then p(b-a)\le \int_{a}^{b}f(x)\,dx \le s(b-a).

Fundamental theorem of calculus


Theorem: Mean value theorem

Let f\colon[a,b]\to\mathbb{R} be a continuous function. Then there exists c\in(a,b) such that  f(c)=\frac{1}{b-a}\int_{a}^{b}f(x)\,\mathrm{d}x. This is the mean value of f on the interval [a,b] and we denote it with \overline{f}.

Proof.

Suppose that m and M are the minimum and maximum of f on the interval [a,b], respectively. It follows that  m(b-a)\le \int_{a}^{b}f(x)\,\mathrm{d}x\le M(b-a) or m\le \frac{1}{b-a}\int_{a}^{b}f(x)\,\mathrm{d}x\le M\quad \Leftrightarrow\quad m\le \overline{f}\le M. Thus \overline{f} is between the minimum and maximum of a continuous function f and by the intermediate value theorem it must be that f(c)=\overline{f} for some c\in\,]a,b[.

\square


(First) Fundamental theorem of calculus.

Let f\colon[a,b]\to\mathbb{R} be a continuous function. Then  \frac{\mathrm{d}}{\mathrm{d}x}\int_{a}^{x}f(t)\,\mathrm{d}t = f(x) for all x\in\,]a,b[.

Proof.

Let  F(x) = \int_{a}^{x}f(t)\,\mathrm{d}t. The mean value theorem implies that there exists c\in\,[x,x+h] such that  \frac{F(x+h)-F(x)}{h} = \frac{1}{h}\left(\int_{a}^{x+h}f(t)\,
	\mathrm{d}t-\int_{a}^{x}f(t)\,\mathrm{d}t\right)  =\frac{1}{h}\int_{x}^{x+h}f(t)\,\mathrm{d}t = \frac{1}{h}f(c)(x+h-x) = f(c). As h\to0 we see that c\to x and from the continuity of f it follows that f(c)\to f(x). Thus F'(x)=f(x).

\square

Antiderivative

If F'(x)=f(x) on some open interval then F is the antiderivative (or the primitive function) of f. The fundamental theorem of calculus guarantees that for every continuous function f there exists an antiderivative  F(x) = \int_{a}^{x}f(t)\,dt. The antiderivative is not necessarily expressible as a combination of elementary functions even if f were an elementary function, e.g. f(x) = e^{-x^{2}}. Such primitives are called nonelementary antiderivatives.

Theorem.

Antiderivatives are only unique up to a constant; \int f(x)\,dx = F(x) + C, C\in\mathbb{R} \text{ constant } if F'(x)=f(x).

Proof.

Suppose that F'_{1}(x)=F'_{2}(x)=f(x) for all x. Then the derivative of F_{1}(x)-F_{2}(x) is identically zero and thus the difference is a constant.

\square

(Second) Fundamental theorem of calculus

Let f\colon\left[a,b\right]\to\mathbb{R} be a continuous function and G an antiderivative of f, then \int_{a}^{b}f(x)\,dx = G(x)\Big|_{x=a}^{x=b} = G(b)-G(a).

Proof.

Because F(x)=\int_{a}^{x}f(t)\,dt is an antiderivative of f then due to continuity F(x)-G(x)=C=\text{constant} for all x\in\left[a,b\right]. Substituting x=a we find that C=-G(x). Thus \int_{a}^{x}f(t)\,dt = F(x) = G(x)-G(a) and substituting x=b the result follows.

\square


Suppose that f is a continuous function and that a and b are differentiable functions. Then \frac{d}{dx}\int_{a(x)}^{b(x)}f(t)\,dt = f(b(x))b'(x)-f(a(x))a'(x).

Proof.

Suppose that F is an antiderivative of f. Then from the fundamental theorem of calculus and the chain rule it follows that \frac{d}{dx}\int_{a(x)}^{b(x)}f(t)\,dt  = \frac{d}{dx}\big(F(b(x)) - F(a(x))\big) =\frac{d}{dx}F(b(x)) - \frac{d}{dx}F(a(x)) = F'(b(x))b'(x) - F'(a(x))a'(x)  = f(b(x))b'(x) - f(a(x))a'(x).

\square

Integrals of elementary functions


Constant Functions

Given the constant function f(x) = c,\,c\in\mathbb{R}. The integral \int\limits_a^b f(x)\,\mathrm{d} x = \int\limits_a^b c \, \mathrm{d}x has to be determined now.

Solution by finding a antiderivative

From the previous chapter it is known that g(x) = c\cdot x gives g'(x) = c. This means that c \cdot x is an antiderivative for c. So the following applies \int\limits_a^b c \, \mathrm{d}x = [c \cdot x]_{x=a}^{x=b} = c\cdot b - c \cdot a = c \cdot (b-a).

Remark: Of course, a function h(x) = c \cdot x + d would also be an antiderivative of f, since the constant d is omitted in the derivation. For sake of simplicity c \cdot x can be used, since d can be chosen as d=0 for definite integrals.

Solution by geometry

The area under the constant function forms a rectangle with height c and length b-a. Thus the area is c \cdot (b-a) and this corresponds to the solution of the integral. Illustrate this remark by a sketch.

Linear functions

Given is the linear function  f(x) = mx. We are looking for the integral \int\limits_a^b f(x)\, \mathrm dx=\int\limits_a^b mx\, \mathrm dx.

Solve by finding a antiderivative

The antiderivative of a linear function is in any case a quadratic function, since \frac{\mathrm d x^2}{\mathrm dx} = 2x . The derivative of a quadratic function results in a linear function. Here, it is important to consider the leading factor as in \frac{\mathrm d (m \cdot \frac{1}{2} \cdot x^2)}{\mathrm dx} = mx. Thus the result is \int\limits_a^b mx \mathrm{d}x = \left[\frac{m}{2}x^2 \right]_{x=a}^{x=b}= \frac{m}{2}b^2 - \frac{m}{2}a^2.

Solving by geometry

The integral \int\limits_a^b mx\, \mathrm dx can be seen geometrically, as subtracting the triangle with the edges (0|0), (a|0) and (a| ma) from the triangle with the edges (0|0), (b|0) and (b| mb). Since the area of a triangle ist given by \frac{1}{2} \cdot \mbox{baseline} \cdot \mbox{height}, the area of the first triangle \frac{1}{2}\cdot b \cdot mb = \frac{1}{2}mb^2 and that of the second triangle is analogous \frac{1}{2}ma^2. For the integral the result is \frac{m}{2}b^2 - \frac{m}{2}a^2. This is consistent with the integral calculated using the antiderivative. Illustrate this remark by a sketch.

Power functions

In constant and linear functions we have already seen that the exponent of a function decreases by one when it is derived. So it has to get bigger when integrating. The following applies: \frac{\mathrm d x^n}{\mathrm dx} = n \cdot x^{n-1}. It follows that the antiderivative for x^n must have the exponent n+1, \frac{\mathrm d x^{n+1}}{\mathrm dx} = (n+1) \cdot x^n. By multiplying the last equation with \frac{1}{n+1} we get \frac{\mathrm d}{\mathrm dx}\frac{1}{n+1} x^{n+1} = \frac{n+1}{n+1} \cdot x^n = x^n. Finally the antiderivative is \int x^n\, \mathrm dx = \frac{1}{n+1} x ^{n+1}+c,\,c\in\mathbb{R}.

Examples

  • \int x^2\, \mathrm dx = \frac{1}{3} x^3 +c,\,c\in\mathbb{R}
  • \int x^3\, \mathrm dx = \frac{1}{4} x^4 +c,\,c\in\mathbb{R}
  • \int x^{20}\, \mathrm dx = \frac{1}{21} x^{21} +c,\,c\in\mathbb{R}

The formula \int x^n\, \mathrm dx = [\frac{1}{n+1} x^{n+1}] is also valid, if the exponent of the function is a real number and not equal -1.

Examples

  • \int x^{2,7}\, \mathrm dx = \frac{1}{3,7} x^{3,7}+c,\,c\in\mathbb{R}
  • \int \sqrt{x}\, \mathrm dx = \int x^\frac{1}{2} = \frac{2}{3} x^{\frac{3}{2}}+c,\,c\in\mathbb{R}
  • But: For x\gt0 applies\int x^{-1}\,\mathrm d x=\ln(x)+c,\,c\in\mathbb{R}.

Natural Exponential function

The natural exponential function f(x) = e^x is one of the easiest function to differentiate and integrate. Since the derivation of e^x results in e^x, it follows \int e^x\, \mathrm dx = e^x +c, \, c\in \mathbb{R}.

Example 1

Determine the value of the integral \int_0^1 e^z \,\mathrm{d} z.

\int\limits_0^1e^z\,\mathrm{d}z= e^z\big|_{z=0}^{z=1}=e^1-e^0=e-1.

Example 2

Determine the value of the integral \int_0^b e^{\alpha t} \,\mathrm{d} t. Using the same considerations as above we get \int\limits_0^b e^{\alpha t}\,\mathrm{d}t= \frac{1}{\alpha}e^{\alpha t}\big|_{t=0}^{t=b} =\frac{1}{\alpha}\left(e^{\alpha b}-e^0\right)=\frac{1}{\alpha}\left(e^{\alpha b}-1\right). Important is here, that we have to use the factor \frac1{\alpha}.

Natural Logarithm

The derivative of the natural logarithmic function is \ln'(x) =\frac{1}{x} for x\gt0. It even applies \ln'(x) =\frac{1}{x} to x. These results together result in for the antiderivative of \frac{1}{x}

\int \frac{1}{x}\,\mathrm{d}x = \ln\left(|x|\right) +c , c\in\mathbb{R}.

An antiderivative can be specified for the natural logarithm: \int \ln(x)\,\mathrm{d}x = x\ln(x) - x + c ,\, c\in\mathbb{R}.

Trigonometric function

The antiderivatives of \sin(x) and \cos(x) also result logically if you derive "backwards". We have \int \sin(x)\, \mathrm dx = -\cos(x)+c,\,c\in\mathbb{R}, since  (-\cos(x))' =-(-\sin(x))=\sin(x). Furthermore we know \int \cos(x)\, \mathrm dx = \sin(x)+c,\,c\in\mathbb{R}, since (\sin(x))' = \cos(x) applies.

Example 1

Which area is covered by the sine on the interval [0,\pi] and the x-axis? To determination the area we simply have to evaluate the integral \int_0^{\pi} \sin(\tau) \, \mathrm{\tau}. That means \int_0^{\pi} \sin(\tau) \, \mathrm{d}\tau = \left[-\cos(\tau)\right]_{\tau=0}^{\tau=\pi} = 
	-\cos(\pi) - (-\cos(0)) = -(-1) - (-1) = 2. Again make a sketch for this example.

Example 2

How can the integral \int \cos(\omega t +\phi)\,\mathrm{d}t be expressed analytically?

To determine the integral we use the antiderivative of the cosine: \sin'(x) = \cos(x). However, the inner derivativ has to be considered in the given function and thus we get \int \ \cos(\omega t +\phi)\,\mathrm{d}t =\frac{1}{\omega}\sin(\omega t+\phi)+c,\,c\in\mathbb{R}.

Summary:

The most common antiderivatives follow from the rules of differentiation: \int x^{r}\,dx = \frac{1}{r+1}x^{r+1} + C, ~r\neq-1 \int x^{-1}\,dx = \ln|x| + C \int e^{x}\,dx = e^{x} + C \int \sin x\,dx = -\cos x + C \int \cos x\,dx = \sin x + C \int \frac{dx}{1+x^{2}} = \arctan x + C

Example 1

Evaluate the integrals \displaystyle \int_{-1}^{1}e^{-x}\,dx and \displaystyle\int_{0}^{1}\sin(\pi x)\,dx.

Solution. The antiderivative of e^{-x} is -e^{-x} so we have that \int_{-1}^{1}e^{-x}\,dx = -e^{-1}+e^{1} = 2\sinh1. The antiderivative of \sin(\pi x) is -\frac{1}{\pi}\cos(\pi x) and thus \int_{0}^{1}\sin(\pi x)\,dx = -\frac{1}{\pi}(\cos\pi-\cos0) = \frac{2}{\pi}.

Example 2

Evaluate the integral \displaystyle\int_{0}^{1}\frac{x}{\sqrt{25-9x^{2}}}\,dx.

Solution. The antiderivative might look something like F(x)=a(25-9x^{2})^{1/2}, where we can find the factor a through differentiation: D\big(a(25-9^{2})^{1/2}\big) = a\cdot\frac{1}{2}\cdot(-18x)(25-9x^{2})^{-1/2} = \frac{-9ax}{\sqrt{25-9x^{2}}} hence if a=-1/9 we get the correct antiderivative. Thus \int_{0}^{1}\frac{x}{\sqrt{25-9x^{2}}}\,dx = -\frac{1}{9}\cdot(25-9x^{2})^{1/2}\Big|_{x=0}^{x=1} = -\frac{1}{9}(\sqrt{16}-\sqrt{25}) = \frac{1}{9}. This integral can also be solved using integration by substitution; more on this method later.

Geometric applications


Area of a plane region

Suppose that f and g are piecewise continuous functions. The area of a region bounded by the graphs y=f(x), y=g(x) and the vertical lines x=a and x=b is given by the integral A=\int_{a}^{b}|f(x)-g(x)|\,dx.

Especially if f is a non-negative function on the interval [a,b] and g(x)=0 for all x then the integral A=\int_{a}^{b}f(x)\,dx is the area of the region bounded by the graph y=f(x), the x-axis and the vertical lines x=a and x=b.

Arc length

The arc length \ell of a planar curve y=f(x) between points x=a and x=b is given by the integral \ell = \int_{a}^{b}\sqrt{1+f'(x)^{2}}\,dx.

Heuristic reasoning: On a small interval \left[x,x+\Delta x\right] the arc length of the curve between y=f(x) and y=f(x+\Delta x) is approximately \Delta s \approx \sqrt{\Delta x^{2} + \Delta y^{2}} = \Delta x\sqrt{1+\left(\frac{\Delta y}{\Delta x}\right)^{2}} \approx \Delta x\sqrt{1+f'(x)^{2}}.

Interactive. Arc length approximation using secant vectors. The length of each vector is \Delta s.

Surface of revolution

The area of a surface generated by rotating the graph y=f(x) around the x-axis on the interval \left[a,b\right] is given by A = 2\pi\int_{a}^{b}|f(x)|\sqrt{1+f'(x)^{2}}\,dx. Heuristic reasoning: An area element of the surface is approximately \Delta A \approx \text{perimeter}\cdot\text{length} = 2\pi|f(x)|\cdot\Delta s.

Solid of revolution

Suppose that the cross-sectional area of a solid is given by the function A(x) when x\in\left[a,b\right]. Then the volume of the solid is given by the integral V = \int_{a}^{b}A(x)\,dx. If the graph y=f(x) is rotated around the x-axis between the lines x=a and x=b the volume of the generated figure (the solid of revolution) is V = \pi\int_{a}^{b}f(x)^{2}\,dx. This follows from the fact that the cross-sectional area of the figure at x is a circle with radius f(x) i.e. A(x)=\pi f(x)^{2}.

More generally: Let 0\le g(x)\le f(x) and suppose that the region bounded by y=f(x) and y=g(x) and the lines x=a and x=b is rotated around the x-axis. The volume of this solid of revolution is V = \pi\int_{a}^{b}\big(f(x)^{2}-g(x)^{2}\big)\,dx.

Improper integral


Definition: Improper integral
  • 1st kind: The integral is defined on an unbounded domain, \left[a,\infty\right[,\left]-\infty,b\right] or the entire \mathbb{R}.
  • 2nd kind: The integrand function is unbounded in the domain of integration or a two-sided limit doesn't exist on one or both of the endpoints of the integral

One limitation of the improper integration is that the limit must be taken with respect to one endpoint at a time.

Example

\int_{0}^{\infty}\frac{dx}{\sqrt{x}(1+x)} = \int_{0}^{1}\frac{dx}{\sqrt{x}(1+x)} + \int_{1}^{\infty}\frac{dx}{\sqrt{x}(1+x)} Provided that both of the integrals on the right-hand side converge. If either of the two is divergent then so is the integral.

Definition

Let f\colon\left[a,\infty\right[\to\mathbb{R} be a piecewise continuous function. Then \int_{a}^{\infty}f(x)\,dx = \lim_{R\to\infty}\int_{a}^{R}f(x)\,dx provided that the limit exists and is finite. We say that the improper integral of f converges over \left[a,\infty\right[.

Likewise for f\colon\left]-\infty,b\right]\to\mathbb{R} we define \int_{-\infty}^{b}f(x)\,dx = \lim_{R\to\infty}\int_{-R}^{b}f(x)\,dx provided that the limit exists and is finite.


Example

Find the value of \displaystyle\int_{0}^{\infty}e^{-x}\,dx.

Solution. Notice that \int_{0}^{R}e^{-x}\,dx = \left(-e^{-x}\right)\Bigg|_{x=0}^{R}=1-e^{-R}\to 1 as R\to\infty. Thus the improper integral converges and \int_{0}^{\infty}e^{-x}\,dx = 1.

Definition

Let f\colon\mathbb{R}\to\mathbb{R} be a piecewise continuous function. Then \int_{-\infty}^{\infty}f(x)\,dx = \int_{-\infty}^{0}f(x)\,dx + \int_{0}^{\infty}f(x)\,dx if both of the two integrals on the right-hand side converge.

In the case f(x)\ge0 for all x\in\mathbb{R} the following holds \int_{-\infty}^{\infty}f(x)\,dx = \lim_{R\to\infty}\int_{-R}^{R}f(x)\,dx.


However, this doesn't apply in general. For example, let f(x)=x. Note that even though 
	\int_{-R}^{R}f(x)\,dx = \int_{-R}^{R}x\,dx = \frac{R^2}{2} - \frac{(-R)^2}{2} = 0
	for all R\in\mathbb{R} the improper integral 
	\int_{-\infty}^{\infty}x\,dx = \lim_{R\to\infty}\int_{-R}^{0}x\,dx + \lim_{R\to\infty}\int_{0}^{R}x\,dx = \lim_{R\to\infty}-\frac{(-R)^2}{2} + \lim_{R\to\infty}\frac{R^2}{2} = \infty - \infty
	does not converge.

Improper integrals of the 2nd kind are handled in a similar way using limits. As there are many different (but essentially rather similar) cases, we leave the matter to one example only.

Example

Find the value of the improper integral \displaystyle\int_{0}^{1}\frac{dx}{\sqrt{x}}.

Solution. We get \int_{\epsilon}^{1}\frac{dx}{\sqrt{x}} = \left(2\sqrt{x}\right)\Bigg|_{x=\epsilon}^{x=1} = 2-2\sqrt{\epsilon} \to 2, as \epsilon\to0+. Thus the integral converges and its value is 2.

The improper integral of f(x)=1/\sqrt{x} from x=0 to x=1.

Comparison test


One way of studying the convergence of an improper integral is using the comparison test.
Theorem.
Suppose that f and g are integrable functions such that |f(x)|\le g(x) for a < x < b.
  1. If the improper integral I=\int_{a}^{b}g(x)\,dx converges then so does \displaystyle\int_{a}^{b}f(x)\,dx and its value is less than or equal to I.
  2. If the improper integral \int_{a}^{b}f(x)\,dx diverges then so does \displaystyle\int_{a}^{b}g(x)\,dx.
Example 2

Notice that 0\le\frac{1}{\sqrt{x}(1+x)}\le\frac{1}{\sqrt{x}}, \text{ for }0 < x < 1 and that the integral \int_{0}^{1}\frac{dx}{\sqrt{x}} = 2 converges. Thus by the comparison test the integral \int_{0}^{1}\frac{dx}{\sqrt{x}(1+x)} also converges and its value is less than or equal to 2.

Example 3

Likewise 0\le\frac{1}{\sqrt{x}(1+x)} < \frac{1}{\sqrt{x}(0+x)}=\frac{1}{x^{3/2}}, \text{ for }x\ge1 and because \displaystyle\int_{1}^{\infty}x^{3/2}\,dx=2 converges so does \int_{1}^{\infty}\frac{dx}{\sqrt{x}(1+x)} and its value is less than or equal to 2.

Note. The choice of the dominating function depends on both the original function and the interval of integration.

Example 4

Determine whether the integral \int_{0}^{\infty}\frac{x^2+1}{x^3(\cos^2{x}+1)}\,dx converges or diverges.

Solution. Notice that x^2+1\ge x^2 for all x\in\mathbb{R} and therefore \frac{x^2+1}{x^3(\cos^2{x}+1)} \ge \frac{1}{x\underbrace{(\cos^2{x}+1)}_{\le 2}} \ge \frac{1}{2x}. Now, because the integral \displaystyle\int_{0}^{\infty}\frac{dx}{2x} diverges then by the comparison test so does the original integral.

Integration techniques


Logarithmic integration

Given a quotient of differentiable functions, we know to apply the quotient rule. However, this is not so easy with integration. Here only for a few special cases we will state rules in this chapter.

Logarithmic integration As we already know the derivative of \ln(x), i.e. the natural logarithm to the base e, equal to \frac{1}{x}. According to the chain rule the derivative of differentiable function with positive function values is f\,:\,\frac{\mathrm d}{\mathrm dx} \ln (f(x)) = \frac{f'(x)}{f(x)}. This means that for a quotient of functions where the numerator is the derivative of the denominator yields the rule: \begin{equation} \int \frac{f'(x)}{f(x)}\, \mathrm{d} x= \ln \left(|f(x)|\right) +c,\,c\in\mathbb{R}.\end{equation} Using the absolute value of the function is important, since the logarithm is defined on \mathbb{R}^+.

Examples
  • \int \frac{1}{x}\, \mathrm dx = \int \frac{x'}{x} \mathrm\, dx = \ln(|x|) +c,\,c\in\mathbb{R} .

  • \int \frac{3x^2 + 17}{x^3 +17x - 15}\, \mathrm dx = \ln(|x^3 + 17x - 15|)+c,\,c\in\mathbb{R}.

  • \int \frac{\cos(x)}{\sin(x)}\,\mathrm dx = \ln(|\sin(x)|)+c,\,c\in\mathbb{R}.

Integration of rational functions - partial fraction decomposition

The logarithmic integration works well in special cases of broken rational functions where the counter is a multiple of the derivation of the denominator. However, other cases can sometimes be traced back to this. This method is called partial fractional decomposition, which represents rational functions as the sum of proper rational functions.

Example 1

The function \frac{1}{1-x^2} cannot be integrated at first glance. However, the denominator 1-x^2 can be written as (1-x)(1+x) and the function can finally reads as \dfrac{1}{1-x^2} = \dfrac{\frac{1}{2}}{1+x} + \dfrac{\frac{1}{2}}{1-x} by partial fraction decomposition. This expression can be integrated, as demonstrated now: \begin{eqnarray} \int \dfrac{1}{1-x^2} \,\mathrm dx &= & \int \dfrac{\frac{1}{2}}{1+x} + \dfrac{\frac{1}{2}}{1-x}\, \mathrm dx \\ & =& \frac{1}{2} \int \dfrac{1}{1+x}\, \mathrm dx - \frac{1}{2} \int \dfrac{-1}{1-x}\, \mathrm dx\\ & = &\frac{1}{2} \ln|1+x| +c_1 - \frac{1}{2} \ln|1-x| +c_2\\ &= &\frac{1}{2} \ln \left|\dfrac{1+x}{1-x}\right|+c,\,c\in\mathbb{R}. \end{eqnarray} This procedure is now described in more detail for some special cases.

Case 1: Q(x)=(x-\lambda_1)(x-\lambda_2) with \lambda_1\ne\lambda_2. In this case, R has the representation R(x) = \frac{ax+b}{(x-\lambda_1)(x-\lambda_2)} and can be transformed to \frac{ax+b}{(x-\lambda_1)(x-\lambda_2)} = \frac{A}{(x-\lambda_1)}+\frac{B}{(x-\lambda_2)}. By multiplying with (x-\lambda_1)(x-\lambda_2) it yields ot ax+b = A(x-\lambda_1) + B(x-\lambda_2) = \underbrace{(A+B)}_{\stackrel{!}{=}a}x + \underbrace{(-A\lambda_1-B\lambda_2)}_{\stackrel{!}{=}b}.

A and B are now obtained by the method of equating the coefficients.

Example 2

Determe the partial fraction decomposition of \frac{2x+3}{(x-4)(x+5)}.

Start with the equation \frac{2x+3}{(x-4)(x+5)} = \frac{A}{(x-4)}+\frac{B}{(x+5)} to get the parameters A and B. Multiplication by {(x-4)(x+5)} leads to 2x+3 = A(x+5)+B(x-4) = (A+B)x +5A -4B. Now we get the system of linear equations

\begin{eqnarray}A+B & = & 2 \\ 5A - 4 B &=& 3\end{eqnarray} with the solution A = \frac{11}{9} and B= \frac{7}{9}. The representation with proper rational functions is \frac{2x+3}{(x-4)(x+5)}=\frac{11}{9}\frac{1}{(x-4)}+\frac{7}{9}\frac{1}{(x+5)} The integral of the type \int \frac{ax+b}{(x-\lambda_1)(x-\lambda_2)}\,\mathrm{d} x. is no longer mystic.

With the help of partial fraction decomposition, this integral can now be calculated in the following manner \begin{eqnarray}\int \frac{ax+b}{(x-\lambda_1)(x-\lambda_2)}\mathrm{d} x &=& \int\frac{A}{(x-\lambda_1)}+\frac{B}{(x-\lambda_2)}\mathrm{d} x \\ &=&A\int\frac{1}{(x-\lambda_1)}\mathrm{d} x +B\int\frac{1}{(x-\lambda_2)}\mathrm{d} x \\ & = & A\ln(|x-\lambda_1|) + B\ln(|x-\lambda_2|).\end{eqnarray}

Example 3

Determine the antiderivative for \frac{2x+3}{(x-4)(x+5)}, i.e. \int\frac{2x+3}{(x-4)(x+5)}\,\mathrm{d} x.

From the above example we already know: \int\frac{2x+3}{(x-4)(x+5)}\,\mathrm{d} x = \int\frac{11}{9}\frac{1}{(x-4)}+\frac{7}{9}\frac{1}{(x+5)}\, \mathrm{d} x.

Using the idea explained above immediately follow: \int\frac{11}{9}\frac{1}{(x-4)}+\frac{7}{9}\frac{1}{(x+5)} \,\mathrm{d} x = \frac{11}{9}\int\frac{1}{(x-4)} \mathrm{d} x + \frac{7}{9}\int\frac{1}{(x+5)} \,\mathrm{d} x= \frac{11}{9}\ln(|(x-4)|)+\frac{7}{9}\ln(|(x+5)|). So is the result \int\frac{2x+3}{(x-4)(x+5)}\,\mathrm{d} x=\frac{11}{9}\ln(|(x-4)|)+\frac{7}{9}\ln(|(x+5)|).

Case 2: Q(x)=(x-\lambda)^2.

In this case R has the representation R(x) = \frac{ax+b}{(x-\lambda)^2} and the ansatz \frac{ax+b}{(x-\lambda)^2} = \frac{A}{(x-\lambda)}+\frac{B}{(x-\lambda)^2} is used.

By multiplying the equation with (x-\lambda)^2 we get ax+b = A(x-\lambda) + B. Again equating the coefficients leads us to a system of linear equations in A=a and B=b+A\lambda=b+a\lambda.

So we have \int \frac{ax+b}{(x-\lambda)^2}\,\mathrm{d}x = \int \frac{a}{(x-\lambda)}+\frac{b+a\lambda}{(x-\lambda)^2} \mathrm{d}x =a\ln(|x-\lambda|)-\frac{b+a\lambda}{(x-\lambda)}+c,\,c\in\mathbb{R}.

3. case Q(x)=x^2+mx+n without real zeros.

In this case R has the representation R(x) = \frac{ax+b}{x^2+mx+n} and the representation can not be simplified.

Only the special case R(x) = \frac{2x+m}{x^2+mx+n} is now considered.

In this case we have \int \frac{2x+m}{x^2+mx+n}\,\mathrm{d}x = \ln(|x^2+mx+n|)+c, \quad c\in\mathbb{R}.

Another special case is R(x) = \frac{1}{x^2+1} with \int \frac{1}{x^2+1} \, \mathrm{d} x = \arctan(x) +c,\quad c\in \mathbb{R}.

Integration by Parts

The derivative of a product of two continuously differentiable functions f and g is (f(x)\cdot g(x))' = f'(x)\cdot g(x)+f(x)\cdot g'(x),\quad x\in(a,b).

This leads us to the following theorem:

Theorem: Integration by Parts

Letf and g be continuously differentiable functions on the interval \left[a,b\right]. Then \int_{a}^{b}f'(x)g(x)\,dx = f(b)g(b)-f(a)g(a)-\int_{a}^{b}f(x)g'(x)\,dx Likewise for the indefinite integral it holds that \int f'(x)g(x)\,dx = f(x)g(x)-\int f(x)g'(x)\,dx.

Proof

It follows from the product rule that \frac{d}{dx}(f(x)g(x)) = f'(x)g(x) + f(x)g'(x) or rearranging the terms f'(x)g(x) = \frac{d}{dx}(f(x)g(x)) - f(x)g'(x). Integrating both sides of the equation with respect to x and ignoring the constant of integration now yields \int f'(x)g(x) = f(x)g(x) - \int f(x)g'(x)\,dx.

Example

Solve the integral \displaystyle\int_{0}^{\pi}x\sin x\,dx.

Solution. Set f'(x)=\sin x and g(x) = x. Then f(x)=-\cos x and g'(x) = 1 and the integration by parts gives \int_{0}^{\pi}x\sin x\,dx = -\pi\cos\pi - 0 - \int_{0}^{\pi}(-\cos x)\,dx =\pi+\left(\sin x\right)\Bigg|_{x=0}^{\pi} = \pi.

Notice that had we chosen f and g the other way around this would have led to an even more complicated integral.

Integration by Substitution

Theorem: Integration by substitution

Let f and g be continuously differentiable functions on \left[a,b\right]. Then \int_{a}^{b}f(g(x))g'(x)\,dx = \int_{g(a)}^{g(b)}f(u)\,du.

Proof.

Let F'(x)=f(x). Then \int_{a}^{b}f(g(x))g'(x)\,dx = \int_{a}^{b}(F\circ g)'(x)\,dx = (F\circ g)(b) - (F\circ g)(a) = F(g(b)) - F(g(a)) = \int_{g(a)}^{g(b)}f(u)\,du.

In practise: Substituting u=g(x) we have (heuristically) \frac{du}{dx}=g'(x)\Rightarrow du=g'(x)\,dx and the limits of integration x=a\Rightarrow u=g(a),x=b\Rightarrow u=g(b).


Example 1

Find the value of the integral \displaystyle\int_{0}^{\pi^2}\sin\sqrt{x}\,dx.

Solution. Making the substitution x=t^{2} when t\ge0 we have dx=2t\,dt. Solving the limits from the inverse formula i.e. t=\sqrt{x} we find that t(0)=0 and t(\pi^{2})=\pi. Hence \int_{0}^{\pi^{2}}\sin\sqrt{x}\,dx = \int_{0}^{\pi^{2}}2t\sin t\,dt = 2\int_{0}^{\pi}t\sin t\,dt = 2\pi.

Here the latter integral was solved applying integration by parts in the previous example.

Example 2

Find the antiderivative of \displaystyle\frac{1}{\sqrt{x}(1+x)}.

Solution. Substituting x=t^{2}, t>0 or t=\sqrt{x} gives \int\frac{dx}{\sqrt{x}(1+x)} = \int\frac{2t}{t(1+t^{2})}\,dt = 2\arctan t + C = 2\arctan\sqrt{x} + C.

'; }],{ useMathJax : true, fontSize : 14, fixed : true })); txts.push(b2.create('text',[2,-5, function() { return 'L(f)=' + rsums[1].Value().toFixed(3) + ''; } ],{ useMathJax : true, fontSize : 14, fixed : true })); b1.fullUpdate(); b2.fullUpdate(); }; plotButton.click(); })(); ', label : { useMathJax : true, offset : [0, 30] } }); board.create('integral', [[0.2, 1.4], graph], { fillColor : 'blue', curveRight : { visible : false }, curveLeft : { visible : false }, withLabel : false, baseLeft : { visible : true, showinfobox : false, size : 1, strokeColor : 'black', withLabel : true, name : 'a', label : { useMathJax : true, offset : [0, -1] }}, baseRight : { visible : true, showinfobox : false, size : 1, strokeColor : 'black', withLabel : true, name : 'b', label : { useMathJax : true, offset : [0, -1] }} } ); board.fullUpdate(); })();

9. Differential equations

Introduction


Differential equation is an equation containing an unknown function, e.g.   y = y(x)  , and its derivatives  y'(x), y''(x), \ldots, y^{(n)}(x) . This kind of equation where the unknown function depends on a single variable, is called an ordinary differential equation (ODE) or simply a differential equation. If the unknown function contains several variables, it is called partial differential equation, but they are not covered in this course.

A typical application leading to a differential equation is radioactive decay. If  y=y(t) is the number of radioactive nuclei present at time  t , then during a short time interval  \Delta t the change in this number is approximately  \Delta y \approx -k y(t)\cdot \Delta t, where  k is a positive constant depending on the radioactive substance. The approximation becomes better as  \Delta t \to 0, so that  y'(t) \approx \Delta y/\Delta t \approx -ky(t) . It follows that the differential equation  y'(t)=-ky(t) is a mathematical model for the radioactive decay. In reality, the number of nuclei is an integer, so the function  y(t) is not differentiable (or the derivative is mostly zero!). Therefore, the model describes the properties of some idealized smooth version of  y(t). This is a typical phenomenon in most models.

Order

The order of a differential equation is the highest order of the derivatives appearing in the equation.

For example, the order of the differential equation  y' + 3y = \sin(x) is 1. The order of the differential equation  y'' + 5y' -6y = e^x is 2.

Here the variable of the function y is not visible; the equation is considered to determine y implicitly.

Solutions of a differential equation

A differential equation of order n is of the form 

 \begin{equation} \label{dydef}
F(x, y(x), y'(x),\ldots , y^{(n)}(x)) = 0
\end{equation}

The solution to an ODE is an n times differentiable function y(x) satisfying the above equation for all  x \in I, where I is an open interval in the real axis.

Typically the solution is not unique and there can be an infinite number of solutions. Consider the equation  xy^2 + y' = 0. The equation has the solutions

  •  y_0(x) = 0,\enspace x \in \mathbb{R}
  •  y_1(x) = 2/x^2,\enspace x>0
  •  y_2(x) = 2/x^2,\enspace x
  •  y_3(x) = 2/(x^2 + 3),\enspace x \in \mathbb{R}

Here  y_1,  y_2 and  y_3  are called particular solutions. The general solution is  y(x) = 2/(x^2 + C),\> C \in \mathbb{R}. Particular solutions can be derived from the general solution by assigning the parameter c to some value. Solutions that cannot be derived from the general solution are called special solutions.

Differential equations do not necessarily have any solutions at all. For example, the first order differential equation  \sin(y' + y) = 2 does not have any solutions. If a first order equation can be written in normal form  y' = f(x,y) , where f is continuous, then a solution exists.

Initial condition

Constants in the general solution can be assigned to some values if the solution is required to satisfy additional properties. We may for example demand that the solution equals  y_0 at  x_0 by setting an initial condition  y(x_0) = y_0. With first order equations, only one condition is (usually) needed to make the solution unique. With second order equations, we need two conditions, respectively. In this case, the initial condition is of the form

 \left\{ \begin{array} 
yy(x_0) = y_0 \\ y'(x_0) = y_1 \end{array} \right.

In general, for an equation of order n, we need n extra conditions to make the solution unique. A differential equation and its set initial conditions are jointly referred as an initial value problem.

Example 1.

We saw above that the the general solution to the differential equation  xy^2 + y' = 0 is  y(x) = 2/(x^2 + C). Therefore the solution to the initial value problem

 \left\{\begin{align} xy^2 + y' = 0 \\ y(0) = 1 \end{align} \right.

is  y(x) = 2/(x^2 + 2).

Interactive. In the figure above, some of the solutions to the equation 

 xy^2 + y' = 0 are drawn. Try out and see how the initial condition alters the solution. The initial values can be changed by dragging the points. Is it possible to make the solution curves intersect?

Direction field

The differential equation  y' = f(x,y) can be interpreted geometrically: if the solution curve (i.e. graph of a solution) goes through the point  (x_0, y_0) , then it holds that  y'(x_0) = f(x_0, y_0) , i.e. we can find the slopes of the tangents of the curve even if we do not know the solution itself. Direction field or slope field is a vector field  \vec{i} + f(x_k, y_k)\vec{j} drawn through the points  (x_k, y_k). The direction field provides a fairly accurate image of the behavior of the solution curves.

Interactive. In the picture above we have drawn the direction field corresponding to the differential equation  y' = \sin(xy) . As shown here, an initial value often determines the solution also in the negative direction.

1st Order Ordinary Differential Equations


A manifesting problem in the theory of differential equations is that there is only a relatively small amount of methods for finding solutions that are generally applicable. Even for a fairly simple differential equation a generalized formula does not usually exist, and especially for higher order differential equations it is rare to be able to find an analytic solution. For some equations it is possible, however, and here some of the most common cases are introduced.

Linear 1st order ODE

If a differential equation is of the form

 p_n(x)y^{(n)} + p_{n-1}(x)y^{(n-1)} + \cdots + p_1(x)y' + p_0(x)y = r(x),

then it is called a linear differential equation. The left side of the equation is a linear combination of the derivatives with multipliers  p_k(x) . Thus a first order linear ODE is of the form

 p_1(x)y' + p_0(x)y = r(x). 

If  r(x) = 0 for all x, then the equation is called homogeneous. Otherwise the equation is nonhomogeneous.

Theorem 1.

Consider a normal form initial value problem 

 \left\{\begin{align}y^{(n)} + p_{n-1}(x)y^{(n-1)} + \cdots + p_1(x)y' + p_0(x)y = r(x) \\ y(x_0) = y_0, \: y'(x_0) = y_1, \: \ldots, \: y^{n-1}(x_0) = y_{n-1}. \end{align} \right.  

If the functions  p_k and  r are continuous in the interval  (a,b) containing the initial point x_0, then the initial value problem has a unique solution.

The condition concerning the normality of the equation is crucial. For example, the equation x^2y'' - 4xy' + 6y = 0 may have either zero or an infinite number of solutions depending on the initial condition: substituting  x=0 to the equation automatically forces the initial condition  y(0)=0.

Solving a 1st order linear ODE

A first order linear ODE can be solved by using an integrating factor method. The idea of the method is to multiply both sides of the equation  y' + p(x)y = r(x) by the integrating factor \displaystyle e^{\int p(x) dx} =e^{P(x)}, which allows the equation to be written in the form

 \displaystyle y'(x)e^{P(x)} + p(x)e^{P(x)}y(x) = r(x)e^{P(x)} \Leftrightarrow \frac{d}{dx}\left( y(x)e^{P(x)}\right) = r(x)e^{P(x)}.

Integrating both sides of the equation, we get

\displaystyle y(x)e^{P(x)} = \int r(x)e^{P(x)}\, \mathrm{d}x + C 
\Leftrightarrow y(x)= Ce^{-P(x)} + e^{-P(x)}\int r(x) e^{P(x)}\, \mathrm{d}x. 

It is not advisable to try to remember the formula as it is, but rather keep in mind the idea of how the equation should be modified in order to proceed.

Example 1.

Let us solve the differential equation \displaystyle y'-y = e^x+1. The integrating factor is \displaystyle e^{\int (-1)\, \mathrm{d}x} = e^{-x} so we multiply both sides by this expression:

\displaystyle e^{-x}y'-e^{-x}y = 1+e^{-x}

\displaystyle \frac{d}{dx}(y(x)e^{-x}) = 1+e^{-x}

\displaystyle y(x)e^{-x} = \int 1+e^{-x}\, \mathrm{d}x + C = x - e^{-x} + C

\displaystyle y(x)= e^xx - 1 + Ce^x.

Example 2.

Let us solve the initial value problem

 \left\{\begin{align}xy' = x^2 + 3y \\ y(0) = 1 \end{align} \right.

First, we want to express the problem in normal form:

 \displaystyle y' - \frac{3}{x}y = x.

Now the integrating factor is \displaystyle e^{ \int \frac{3}{x} dx } =\displaystyle e^{ -3 \ln \vert x \vert } =\displaystyle e^{ \ln x^{-3} } =\displaystyle \frac{1}{x^3},\> x>0. Hence, we get

\displaystyle \frac{y'}{x^3} - \frac{3}{x^4}y = \frac{1}{x^2}

\displaystyle \frac{d}{dx}(\frac{y}{x^3}) = \frac{1}{x^2}

\displaystyle \frac{y}{x^3} = \int \frac{1}{x^2}\, 
	\mathrm{d}x + C = - \frac{1}{x} + C

y = Cx^3 - x^2

We have found the general solution. Because  y(0) = C\cdot 0 - 0 = 0, that is, the value of the function does not equal the given initial value, the problem does not have a solution. The main reason for this is that the initial condition is given at  x_0=0, where the normal form of the equation is not defined. Any other choice for  x_0 will lead to a unique solution.

Example 3.

Let us solve the ODE xy'-2y=2 given the initial conditions

  1. y(1)=0
  2. y(0)=0.

From the form y'-(2/x)y=2/x we see that the equation in question is a linear ODE. The integrating factor is


e^{-\int (2/x)\, \mathrm{d}x} = e^{-2\ln |x|} = e^{\ln (1/x^2)} = \frac{1}{x^2}.

Multiplying by the integrating factor, we get


(1/x^2)y'(x)-(2/x^3)y(x) =\frac{2}{x^3} 
\Leftrightarrow 
\frac{d}{dx}\left( \frac{y(x)}{x^2}\right) = \frac{2}{x^3},

so the general solution to the ODE is y(x)=x^2 (-1/x^2+C)=Cx^2-1. From the initial value y(1)=0 it follows that C=1, but the other condition y(0)=0 leads to a contradiction -1=0. Therefore, the solution in the (a) part is y(x)=x^2-1, but a solution satisfying the initial condition of part b) does not exist: by substituting  x=0 , the equation forces  y(0)=-1.

Separable equation

A first order differential equation is separable, if it can be written in the form  y' = f(x)g(y),  where f and g are integrable functions in the domain of interest. Treating formally  y'(x)=dy/dx as a fraction, multiplying by  dx and dividing by  g(y), we obtain  \frac{dy}{g(y)}=f(x)\, dx. Integrating the left hand side with respect to  y and the right hand side with respect to  x, we get

\displaystyle \int \frac{\mathrm{d}y}{g(y)} = \int f(x)\, \mathrm{d}x + C

This method gives the solution to the differential equation in implicit form, which we may further be able to solve explicitly for y =y(x). The justification for this formal treatment can be made by using change of variables in integrals.

Example 4.

Let us solve the differential equation \displaystyle y'+\frac{2}{5}x = 0 by separating the variables. (We could also solve the equation by using the method of integrating factors.)

 \displaystyle y'+\frac{2}{5}y = 0

 \displaystyle \frac{dy}{dx} = -\frac{2}{5}y   

 \displaystyle \int \frac{1}{y}\, \mathrm{d}y = -\frac{2}{5} \int \, \mathrm{d}x   

\displaystyle \ln |y| = -\frac{2}{5}x + C_1

 \displaystyle y =\pm e^{-\frac{2}{5}x+C_1} = \pm e^{-\frac{2}{5}x}e^{C_1} = Ce^{-\frac{2}{5}x}, \: C\neq 0.

In the last step, we wrote C =\pm e^{C_1} for simplicity. The case  C=0 is also allowed, since it leads to the trivial solution y(x)\equiv 0, see below.

Example 5.

Let us solve the initial value problem

 \left\{\begin{align}y' = \frac{x}{y} \\ y(0) = 1 \end{align} \right.

Because the general solution is not required, we may take a little shortcut by applying integrals in the following way:

\displaystyle \frac{dy}{dx} = \frac{x}{y}

\displaystyle \int_1^y y \, \mathrm{d}y =\int_0^x x \, \mathrm{d}x

\displaystyle \frac{1}{2}y^2 - \frac{1}{2} =\frac{1}{2}x^2

The solution is  y=y(x)=\sqrt{x^2+1}.

The trivial solutions of a separable ODE

General solution achieved by applying the method for separable functions typically lacks information about solutions related to the zeros of the function g(y). The reason for this is that in the separation method we need to assume that g(y(x)) \neq 0 in order to be able to divide the expression by g(y(x)). We notice that for each zero \alpha of the function g there exists a corresponding constant solution y(x)\equiv \alpha of the ODE y'=f(x)g(y), since y'(x)\equiv 0=g(\alpha)\equiv g(y(x)). These solutions are called trivial solutions (in contrast to the general solution).

If the conditions of the following theorem hold, then all the solutions to a separable differential equation can be derived from either the general solution or the trivial solutions.

Theorem 2.

Let us consider the initial value problem y'=f(x,y),\ y(x_0)=y_0.

  1. If f is continuous (as a function of two variables), then there exists at least one solution in some interval containing the point x_0.
  2. Also, if f is continuously differentiable with respect to y, then the solution satisfying the initial condition is unique.
  3. The uniqueness also holds, when in addition to (i) the function f is continuously differentiable with respect to x and f(x_0,y_0)\neq 0.

The proof of the theorem is based on a technique known as Picard-Lindelöf iteration, which was invented by Emile Picard and further developed by the Finnish mathematician Ernst Lindelöf (1870-1946), and others.

Applying the previous theorem, we can formulate the following result for separable equations.

Theorem 3.

Let us consider a separable differential equation y'=f(x)g(y), where f is continuous and g is continuously differentiable.

  1. For each zero \alpha of the function g there exists a trivial solution y(x)\equiv \alpha = constant.
  2. All other solutions (= the general solution) can be obtained by applying the previously described method, i.e. separating the variables and calculating the integrals.

The solution curves at each point (x_0,y_0) of the domain of the equation are always unique. In particular, the curves cannot intersect and it is not possible for a single curve to split into two or several parts.

∴ The other solution curves of the ODE cannot intersect the curves y=\alpha corresponding to the trivial solutions. That is, for all the other solutions the condition g(y(x))\neq 0 automatically holds!

Example 6.

Let us solve the linear homogeneous differential equation y'+p(x)y=0 by applying the method of separation.

The equation has the trivial solution y_0(x)\equiv 0. Since the other solutions do not get the value 0 anywhere, it holds that

\begin{aligned}
		\frac{dy}{dx} &= y'= -p(x)y \\
		&\Leftrightarrow \int\frac{\mathrm{d}y}{y}  =  -\int p(x)\, \mathrm{d}x +C_1 \\
		&\Leftrightarrow \ln|y| = -P(x)+C_1 \\
		&\Leftrightarrow |y| =e^{C_1-P(x)} \\
		&\Leftrightarrow y=y(x)=\pm e^{C_1} e^{-P(x)} =Ce^{-P(x)}.\end{aligned}

Here, the expression \pm e^{C_1} has been replaced by a simpler constant C\in\mathbb{R}.

\star Equations expressible as separable

Some differential equations can made separable by using a suitable substitution.

i) ODEs of the form  y'(x)= f\Big(\frac{y(x)}{x}\Big).
Example 7.

Let us solve the differential equation  y'= \frac{x+y}{x-y}. The equation is not separable in this form, but we can make if separable by substituting  u = \frac{y}{x}, resulting to  y' = u + xu'. We get

  u + xu'= \displaystyle \frac{1+u}{1-u}.  

Separating the variables and integrating both sides, we get

  \displaystyle \int \frac{1-u}{1+u^2} \, \mathrm{d}u= \int \frac{1}{x} \, \mathrm{d}x

   \arctan{u} 
	- \displaystyle \frac{1}{2} \ln(u^2 +1)= \ln{x} + C.

Substituting   u = \frac{y}{x} and simplifying yields

   \displaystyle \arctan{\frac{y}{x}} = \ln{C\sqrt{x^2 + y^2}}.

Here, it is not possible to derive an expression for y so we have to make do with just the implicit solution. The solutions can be visualized graphically: 

As we can see, the solutions are spirals expanding in the positive direction that are suitably cut for demonstration purposes. This is clear from the solutions' polar coordinate representation which we obtain by using the substitution

\theta = \displaystyle \arctan{\frac{y}{x}}, r = \sqrt{x^2 + y^2}.

Hence, the solution is

\theta = \ln(Cr) \Leftrightarrow r = Ce^{\theta}.

ii) ODEs of the form \displaystyle y' = f(ax+by+c) 

Another type of differential equation that can be made separable are equations of the form 

\displaystyle y' = f(ax+by+c).

To rewrite the equation as separable, we use the substitution  \displaystyle u = ax+by+c.

Example 8.

Let us find the solution to the differential equation

\displaystyle y' =(x-y)^2 +1.

Here, a natural substitution is \displaystyle u = x-y \Leftrightarrow y = x-u \Rightarrow y' = 1-u'. Substitution yields

 \displaystyle 1-u' =u^2 +1

 \displaystyle \int -\frac{1}{u^2} \, \mathrm{d}u = \int \, \mathrm{d}x

 \displaystyle \frac{1}{u} = x +C

 \displaystyle y = x -  \frac{1}{x +C}.

\star Euler's method

In practice, it is usually not feasible to find analytical solutions to differential equations. In these cases, the only choice for us is to resort to numerical methods. A prominent example of this kind of technique is called Euler's method. The idea behind the method is the observation made earlier with direction fields: even if we do not know the solution itself, we are still able to determine the tangents of the solution curve. In other words, we are seeking solutions for the initial value problem

 \left\{\begin{align}y' = f(x,y) \\ y(x_0) = y_0. \end{align} \right.

In Euler's method, we begin the solving process by choosing the step length  h and using the iteration formula

 \displaystyle y_{k+1} = y_k +  hf(x_k, y_k).

The iteration starts from the index  \displaystyle k=0 by substituting the given initial value to the right side of the iteration formula. Since f(x_k, y_k) = y'(x_k) is the slope of the tangent of the solution at x_k , on each step we move the distance expressed by the step length in the direction of the tangent. Because of this, an error occurs, which grows as the step length is increased.

Example 9.

Use the gadget on the right to examine the solution to the initial value problem

 \left\{\begin{align}y' = \sin(xy) \\ y(x_{0}) = y_{0} \end{align} \right.

obtained by using Euler's method and compare the result to the precise solution.

Interactive. The equation y'=\sin(xy) with the initial condition y(x_{0})=y_{0}. The precise solution is drawn blue while the solution obtained using Euler's method with N number of steps is drawn purple.

2nd and higher order ODEs


For higher order differential equations it is often impossible to find analytical solutions. In this section, we introduce some special cases for which analytical solutions can be found. Most of these cases are linear differential equations. Our focus is on second order differential equations, as they are more common in practical applications and for them it is more likely for an analytical solution to be found, compared to third or higher order differential equations.

Solving a homogeneous ODE

For second order linear equations, there is no easy way to find a general solution. We begin by examining a homogeneous equation

 y’’ + p(x)y’ + q(x)y = 0,

where p and q are continuous functions on their domains. Then, it holds that

1) the equation has linearly independent solutions y_1 and y_2, called fundamental solutions. Roughly speaking, linear independence means that the ratio y_2(x)/y_1(x) is not constant, so that the solutions are essentially different from each other.

2) the general solution can expressed by means of any linearly independent pair of solutions in the form y(x) = C_1y_1(x) + C_2y_2(x) , where  C_1 and  C_2 are constants.

3) if the initial values y(x_0) = a, y'(x_0) = b are fixed, then the solution is unique.

A general method for finding explicitly the fundamental solutions y_1(x) and y_2(x) does not exist. To find the solution, a typical approach is to try to make an educated guess about the solution's form and check the details by substituting this into the equation.

The above results can be generalized to higher order homogeneous equations as well, but then the number of required fundamental solutions and initial conditions increases with respect to the order of the equation.

Example 1.

The equation Â y’’-y= 0 has solutions  y = e^x and  y = e^{-x}. These solutions are linearly independent, so the general solution is of the form  y(x) = C_1e^x + C_2e^{-x}.

Equations with constant coefficients

As a relatively simple special case, let us consider the 2nd order equation

 y’’ + py’ + qy = 0.

In order to solve the equation, we use the guess  y(x) = e^{\lambda x}, where  \lambda is an unknown constant. Substituting the guess into the equation yields

 \lambda^2 e^{\lambda x} + p\lambda e^{\lambda x} + qe^{\lambda x} = 0.

 \lambda^2 + p\lambda + q = 0.

The last equation is called the characteristic equation of the ODE. Solving the characteristic equation allows us to find the solutions for the actual ODE. The roots of the characteristic equation can be divided into three cases:

1) The characteristic equation has two distinct real roots. Then, the ODE has the solutions y_1(x) = e^{\lambda_1x} and y_2(x) = e^{\lambda_2x}.

2) The characteristic equation has a double root. Then, the ODE has the solutions y_1(x) = e^{\lambda x} and y_2(x) = xe^{\lambda x}.

3) The roots of the characteristic equation are of the form \lambda = a \pm bi. Then, the ODE has the solutions y_1(x) = e^{ax}\cos(bx) and y_2(x) = e^{ax}\sin(bx).

The second case can be justified by substitution into the original ODE, and the third case by using the Euler formula  e^{ix}=\cos x+i\sin x. With minor changes, these results can also be generalized to higher order differential equations.

Since the characteristic equation has exactly the same coefficients as the original ODE, it is not necessary to derive it again in concrete examples: just write it down by looking at the ODE!

Example 2.

Let us solve the initial value problem

 \left\{\begin{align}y'' -y' +2y=0 \\ y(0) = 1, y(1)=0 \end{align} \right.

The characteristic equation is \lambda^2 -\lambda -2 = 0, which has the roots  \lambda_1 = 2 and  \lambda_2 = -1. Thus, the general solution is  y(x) = C_1e^{2x} + C_2e^{-x}. The constants can be determined by using the initial conditions:

 \left\{\begin{align}C_1 + C_2=1 \\ e^2C_1 + e^{-1}C_2 = 0 \end{align} \right.

 \left\{\begin{align}C_1 = -\frac{1}{e^3-1} \\ C_2 = \frac{e^3}{e^3-1} \end{align} \right.

Hence, the general solution is  y(x) = \frac{1}{e^3-1} (-e^{2x} + e^{3-x}).

Example 3.

Let us have a look at how the above results hold in higher order equations by solving

 y^{(4)} - 4y''' +14y'' -20y' +25y = 0.

Now, the characteristic equation is  \lambda^4 - 4\lambda^3 +14\lambda^2 -20\lambda +25 = 0, which has the roots  \lambda_1 = \lambda_2 = 1 + 2i and  \lambda_3 = \lambda_4 = 1 - 2i. Thus, the fundamental solutions to the ODE are e^x\sin(2x)e^x\cos(2x)xe^x\sin(2x) and xe^x\cos(2x). The general solution is

 y = C_1e^x\sin(2x) + C_2e^x\cos(2x) + C_3xe^x\sin(2x) + C_4xe^x\cos(2x).

Example 4.

Let  \omega >0 be a constant. The characteristic equation of the ODE 
	y''+\omega^2y=0
	is  \lambda^2+\omega^2=0 with roots  \lambda=\pm i \omega. So  \alpha =0 and  \beta =\omega in Case 3). Since this ODE is a model for harmonic oscillation, we use time  t as variable, and obtain the general solution 
	y(t)=A\cos (\omega t) +B\sin (\omega t),
	with  A,B constants. They will be uniquely determined if we know the initial location y(0) and the initial velocity y'(0). All solutions are periodic and their period is T=2\pi/\omega. In the animation to the right we have  y'(0)=0 and you can choose  \omega and the initial displacement  y(0)=y_0.

Interactive. Harmonic oscillator y(t) = y_{0}\cos(\omega t),
where t is the elapsed time in seconds

Euler's differential equation

Another relatively common type of 2nd order differential equation is Euler's differential equation

 x^2y'' + axy' + by = 0,

where a and b are constants. An equation of this form is solved by using the guess y(x)= x^r. Substituting the guess in the equation yields

 r^2 + (a-1)r + b = 0.

Using the roots of this equation, we obtain the solutions for the ODE in the following way:

1) If the roots are distinct and real, then  y_1(x)= |x|^{r_1} and  y_2(x)= |x|^{r_2}.

2) If the equation has a double root, then  y_1(x)= |x|^{r} and  y_2(x)= |x|^{r}\ln |x|.

3) If the equation has roots of the form r = a \pm bi, then  y_1(x)= |x|^{a}\cos(b\ln |x|) and  y_2(x)= |x|^{a}\sin(b\ln |x|).

Example 5.

Let us solve the equation  x^2y'' - 3xy' + y = 0. Noticing that the equation is Euler's differential equation, we proceed by using the guess y= x^r. Substituting the guess into the equation, we get  r(r-1)x^r - 3rx^r + x^r = 0 \Rightarrow r^2 - 4r + 1 = 0, which yields  r = 2 \pm \sqrt{3}. Therefore, the general solution to the ODE is

y = C_1 x^{2+\sqrt{3}} + C_2x^{2-\sqrt{3}}.

Nonhomogeneous linear differential equations

The general solution to a nonhomogeneous equation

y'' + p(x)y' + q(x)y = r(x)

is the general solution to the corresponding homogeneous equation + particular solution to the nonhomogeneous equation, i.e.

y(x) = C_1y_1(x) + C_2y_2(x) + y_0(x).

The particular solution y_0 is usually found by using a guess that is of the same form as r(x) with general coefficients. Substituting the guess into the ODE, we can solve these coefficients, but only if the guess is of the correct form.

In the table below, we have created a list of possible guesses for second order differential equations with constant coefficients. The form of the guess depends on what kind of elementary functions r(x) consists of. If r(x) is a combination of several different elementary functions, then we need to include corresponding elements for all of these functions in our guess. The characteristic equation of the corresponding homogeneous differential equation is P(\lambda)=\lambda^2+p\lambda+q=0.

r(x) contains
the guess consists of
nth degree polynomial
A_0+A_1x+\dots +A_nx^n ( +A_{n+1}x^{n+1}, if q=P(0)=0)
\sin kx,\ \cos kx
A\cos kx+B\sin kx, if P(ik)\neq 0
\sin kx,\ \cos kx Ax\cos kx+Bx\sin kx, if P(ik)=0
e^{cx}\sin kx,\ e^{cx}\cos kx Ae^{cx}\cos kx+Be^{cx}\sin kx, if P(c+ik)\neq 0
e^{kx} Ae^{kx}, if P(k)\neq 0
e^{kx} Axe^{kx}, if P(k)=0 and P'(k)\neq 0
e^{kx} Ax^2e^{kx}, if P(k)=P'(k)=0

Note. For roots of a second degree polynomial we have to keep in mind that

  • P(k)=0 and P'(k)\neq 0 \Leftrightarrow k\in\mathbb{R} is a simple root of P.

  • P(k)=P'(k)= 0 \Leftrightarrow k\in\mathbb{R} is a double root of P.

  • P(ik)\neq 0 \Leftrightarrow ik\in\mathbb{C} is not a root of P; i.e. \sin kx and \cos kx are not solutions to the homogeneous equation.

Example 6.

Let us find the general solution to the ODE y''+y'-6y=r(x), when

a) r(x)=12e^{-x}

b) r(x)=20e^{2x}.

The solutions are of the form y(x)=C_1e^{-3x}+C_2e^{2x}+y_0(x).

a) Substituting the guess y_0(x)=Ae^{-x}, we get (A -A -6A)e^{-x} =12e^{-x}, which solves for A=-2.

b) In this case a guess of the form Be^{2x} is useless, as it is part of the general solution to the corresponding homogeneous equation and yields just zero when substituted to the left side of the ODE. Here, a right guess is of the form y_0(x)=Bxe^{2x}. Substitution yields


(4B+2B-6B)xe^{2x}+(4B+B)e^{2x} = 20e^{2x},

which solves for B=4.

Using these values for A and B, we can write the general solutions to the given differential equations.

Example 7.

Let us find the solution to the ODE y''+y'-6y=12e^{-x} with the initial conditions y(0)=0, y'(0)=6.

Based on the previous example, the general solution is of the form y(x)=C_1e^{-3x}+C_2e^{2x}-2e^{-x}. Differentiation yields y'(x)=-3C_1e^{-3x}+2C_2e^{2x}+2e^{-x}. From the initial conditions, we get the following pair of equations:


\begin{cases}
0=y(0)=C_1+C_2-2 &\\
6=y'(0)=-3C_1+2C_2+2, &\\
\end{cases}

which solves for C_1=0 and C_2=2. Therefore, the solution to the initial value problem is y(x)=2e^{2x}-2e^{-x}.

Example 8.

A typical application of a second order nonhomogeneous ODE is an RLC circuit containing a resistor (with resistance  R), an inductor (with inductance  L ), a capacitor (with capacitance  C ), and a time-dependent electromotive force  E(t). The electric current  y(t) in the circuit satisfies the ODE  Ly''+Ry'+\frac{1}{C}y=E'(t). Let us solve this ODE with artificially chosen numerical values in the form  y''+10y'+61y=370\sin t.

The homogeneous part has characteristic equation of the form  \lambda^2+10\lambda +61=0 with solutions  \lambda = -5\pm 6i. This gives the solutions  y_1(t)=e^{-5t}\cos(6t) and  y_2(t)=e^{-5t}\sin(6t) for the homogeneous equation. For a particular solution we try y_0(t)=A\cos t +B\sin t. Substituting this into the nonhomogeneous ODE and collecting similar terms yields to  (60A+10B)\cos t +(60B-10A)\sin t = 370\sin t. This equation will be satisfied for all  t (only) if


\begin{cases}
60A+10B=0 &\\
-10A+60B=370. &\\
\end{cases}

which solves for  A=-1 and  B=6. Therefore, the general solution is   y(t)=e^{-5t}(C_1\cos(6t)+C_2\sin(6t)) -\cos t+6\sin t . Note. The exponential terms go to zero very fast and eventually, the current oscillates in the form  y(t)\approx -\cos t+6\sin t.

'); }],{ anchor : p1, anchorY : 'auto', fixed : true, parse : false, fontSize : 16, useMathJax : true }); var x0 = board.getBoundingBox()[0]; var x1 = board.getBoundingBox()[2]; var y0 = board.getBoundingBox()[3]; var y1 = board.getBoundingBox()[1]; var dy = function(x,y) { return Math.sin(x*y); } var f = function(x,yy) { return [dy(x,yy[0])]; } var ode = function() { var data = JXG.Math.Numerics.rungeKutta('euler', [p1.Y()], [p1.X(), x0], 201*(p1.X()-x0)/(x1-x0), f); data = data.reverse(); data = data.slice(0, -1); data = data.concat(JXG.Math.Numerics.rungeKutta('euler', [p1.Y()], [p1.X(), x1], 201*(x1-p1.X())/(x1-x0), f)); return data; } var graph = board.create('curve', [[0],[0]], { strokeColor : '#bd3fe0', strokeWidth : 2 }); graph.updateDataArray = function() { var data = ode(); var h = (x1-x0)/data.length; for(var i=0; i < data.length; i++) { graph.dataX[i] = x0+i*h; graph.dataY[i] = data[i][0]; } }; var slopes = []; var slopefield = function() { for(var x=x0; x<=x1; x=x+0.4){ for(var y=y0; y<=y1; y=y+0.4){ var k = dy(x,y); var length = 3*(2*Math.sqrt(1+k*k)); slopes.push(board.create('arrow', [ [x-1/length,y-k/length], [x+1/length, y+k/length] ],{ fixed : true, highlight : false, strokeOpacity : 0.55, strokeWidth : 1, strokeColor : '#2659f0' })); var yy = slopes[slopes.length-1].point1.Y() if(yy == 0){ slopes[slopes.length-1].setAttribute( { strokeColor : '#18d3af' }); } } } } slopefield(); board.fullUpdate(); })();
'); }],{ anchor : point, anchorY : 'auto', fixed : true, parse : false, fontSize : 16 })); board.create('functiongraph',[function(x) { return g(point,x); },x0,x1],{ strokeWidth : 3, strokeColor : color }); } for(var i=0;i
'); }],{ anchor : p1, anchorY : 'auto', fixed : true, parse : false, fontSize : 16 })); var stepNum = board.create('slider', [[1, 4],[4, 4],[1, 5, 25]],{ name : ' N', snapWidth : 1, precision : 0, fillColor : '#bd59ff', size : 8, withTicks : false, baseline : { strokeWidth : 2 }, highline : { strokeWidth : 4 }, label : { fontSize : 14 } }); /*var numTicks = board.create('ticks', [stepNum.baseline], { ticksDistance : 3/5, minorTicks : 0, majorHeight : 1 });*/ var x0 = board.getBoundingBox()[0]; var x1 = board.getBoundingBox()[2]; var y0 = board.getBoundingBox()[3]; var y1 = board.getBoundingBox()[1]; var dy = function(x,y) { return Math.sin(x*y); } var f = function(x,yy) { return [dy(x,yy[0])]; } var ode = function(interval, steps) { data = JXG.Math.Numerics.rungeKutta('euler', [p1.Y()], [p1.X(), interval], steps, f); return data; } var graph = board.create('curve', [[0],[0]], { strokeColor : '#4287f5', strokeWidth : 3, strokeOpacity : 0.5, dash : 2 }); graph.updateDataArray = function() { var data = ode(x0, 201*(p1.X()-x0)/(x1-x0)); data = data.reverse(); data = data.slice(0, -1); data = data.concat(ode(x1, 201*(x1-p1.X())/(x1-x0))); var h = (x1-x0)/data.length; graph.dataX = []; graph.dataY = []; for(var i=0; i < data.length; i++) { graph.dataX[i] = x0+i*h; graph.dataY[i] = data[i][0]; } }; var approx = board.create('curve', [[0],[0]], { strokeColor : '#bd59ff', strokeWidth : 3 }); approx.updateDataArray = function() { var n = stepNum.Value()+1; var h = a/(n-1); var data = ode(p1.X()+a, n); approx.dataX = []; approx.dataY = []; for(var i=0; i < n; i++) { approx.dataX[i] = p1.X()+i*h; approx.dataY[i] = data[i][0]; } p2.moveTo([approx.dataX[n-1],approx.dataY[n-1]]); }; coords.push(board.create('text',[0,0, function(){ return ('
~(x_{' + stepNum.Value() + '}, y_{' + stepNum.Value() + '})~
'); }],{ anchor : p2, anchorY : 'auto', fixed : true, parse : false, fontSize : 16 })); board.create('line',[[function(){ return p1.X(); },0],[function(){ return p1.X(); },function(){ return p1.Y(); }]],{ dash : 4, StrokeOpacity : 0.5, strokeColor : '#000000' }); board.create('line',[[function(){ return p1.X()+a; },0],[function(){ return p1.X()+a; },function(){ return p1.Y(); }]],{ dash : 4, StrokeOpacity : 0.5, strokeColor : '#000000' }); /*flunky. MathJax needs to be called last */ for(var i=0; i