• 검색 결과가 없습니다.

Successive Substitution: Theory and Example

N/A
N/A
Protected

Academic year: 2022

Share "Successive Substitution: Theory and Example"

Copied!
149
0
0

로드 중.... (전체 텍스트 보기)

전체 글

(1)

Iteration

Suppose we wish to solve f (x) = 0 where:

f (x) = x − 1

2(x + A x)

Here A > 0 is some given constant. Obviously this equation has two solutions, namely x = ±A12

This formula is credited to Heron (a common name) of Alexandria, a Greek engineer and architect who lived somewhere between 100B.C. and A.D. 100. Incidentally the Babylonians also knew of the formula 2000 years earlier.

Let g(x) = 12(x + Ax) so that f(x) = 0 is equivalent to x = g(x)

since

f (x) = 0 = x − 1

2(x + A x) x = 1

2(x + A

x) = g(x)

Let x0 > 0 be a first approximation to A12.

define x1 = g(x0), x2 = g(x1), . . . , xn+1 = g(xn).

This is known as successive substituion or iteration.

(2)

Iteration: Example

f (x) = x − 1

2(x + A

x), A = 5, x0 = 2

We first need to find a g(x) such that f (x) = 0 is that same as x = g(x)

So in our case as we have seen g(x) = 1

2(x + A x)

We also would like to be able to measure the accuracy of our solution. In this case n = |x2n − 5|.

So the sequence {xn} is:

x0 = 2 0 = 1

x1 = 12(2 + 52) = 94 1 = (14)2 x2 = 12(94 + 59

4

) = 16172 2 = (721 )2 x3 = 12(16172 + 1615

72

) = 5184123184 3 = (231841 )2

(3)

Iteration: Example

Say we want to compute square roots and use Newton’s Method. Let R > 0, and let’s set x = p(R).

Then x is a root of the equation x2 − R = 0.

Let’s apply Newton’s method.

f (x) = x2 − R f0(x) = 2x

xn+1 = xnff (x0(xnn))

= xnx2x2n−Rn

= 2x2n−x2x2n−R

n

= x2x2nR

= 12(xnnxRn)

We’ve seen this!

(4)

Iteration: Example

If we want to calculate √

17 with an initial approximation of 4, the sequence {xn} is:

x0 = 4 x1 = 4.12

x2 = 4.123106

x3 = 4.1231056256177

x4 = 4.123105625617660549821409856

So the number of correct digits for the sequence is 1, 3, 6, 14, 28. We converge very rapidly and see the doubling of correct digits.

(5)

Fixed Point Iteration

We compute a sequence of points with a formula of the form

xn+1 = g(xn)

The algorithm defined by such an equation is called functional iteration. In Newton’s method g is:

g(x) = x − f (x) f0(x)

With our iteration formula we can generate sequences that don’t converge, or ones that do, which is what we are interested in. If the sequence {xn} converges then limn→∞ xn exists. Suppose that limn→∞xn = s.

How does s and g relate if g is continous?

g(s) = g( lim

n→∞xn) = lim

n→∞g(xn) = lim

n→∞xn+1 = s So g(s) = s, and we call s a fixed point of g, that is s remains fixed under successive application of the function g(x).

(6)

Contractive Mapping Theorem

A mapping F is said to be contractive if there exists a number λ < 1 such that

|F (x) − F (y)| ≤ λ|x − y|

s

x

s

y F (x) s

F (y) s

Contractive Mapping Theorem Let C be a closed subset of the real line. If F is a contractive mapping of C into C, then F has a unique fixed point and this fixed point is the limit of every sequence obtained by xn+1 = F (xn) with a starting point x0 ∈ C

(7)

Contractive Mapping Theorem: Proof

Lets use the contractive property and the iteration formula to obtain:

|xn−xn−1| = |F (xn−1) −F (xn−2)| ≤ λ|xn−1−xn−2|

Repeat again to get:

|xn−xn−1| ≤ λ|xn−1−xn−2| ≤ λ2|xn−2−xn−3| ≤ . . .

≤ λn−1|x1 − x0| Let’s rewrite xn in the form:

xn = x0 + (x1 − x0) + (x2 − x1) + . . . + (xn − xn−1)

So the sequence {xn} converges if and only if the series

X

n=1

(xn − xn−1)

(8)

converges. To prove it converges it is enough to prove

that

X

n=1

|xn − xn−1|

converges. Using the above comparison

X

n=1

|xn − xn−1| ≤

X

n=1

λn−1|x1 − x0| = 1

1 − λ|x1 − x0| Which converges. Since the sequence converges, let s = limn→∞ xn, so F (s) = s as shown earlier. So the contractive property implies that F is continuous. So is there a unique fixed point? Let x and y be fixed points so.

|x − y| = |F (x) − F (y)| ≤ λ|x − y|

Since λ < 1, |x − y| = 0. s belongs to C since s is the limit of a sequence lying in C.

(9)

Contractive Mapping Theorem: Example

Prove the sequence {Xn} defined as follows converges.

 x0 = −15

xn+1 = 3 − 12|xn| n ≥ 0

|F (x) − F (y)| = |3 − 1

2|x| − 3 + 1

2|y|| = 1

2||y| − |x||

and by the triangle inequality 1

2||y| − |x|| ≤ 1

2|y − x|

So by the Contractive Mapping Theorem, the sequence converges to the unique fixed point of F , which is 2.

(10)

Contractive Mapping Theorem: Example

Use the contractive mapping theorem to compute the fixed point of:

F (x) = 4 + 1

3 sin 2x

|F (x) − F (y)| = 1

3| sin 2x − sin 2y|

Use the Mean-Value Theorem

f0(ξ)(b − a) = f(b) − f(a) F0(x) = 2

3 cos 2x to get

1

3| sin 2x − sin 2y| = 2

3| cos 2ξ||x − y| ≤ 2

3|x − y|

for some ξ between x and y. This shows there is a fixed point of F

(11)

Contractive Mapping Theorem: Example

Using a computer program with an initial value of 4.

x = 4;

m = 20;

for (k = 0; k < M; k++) { x = 4 + 1/3*sin(2*x)

cout << k << ” ” << x << endl;

}

Results in k x

1 4.3297861 2 4.2308951 3 4.2736338

...

15 4.2614840 ...

20 4.2614837

(12)

Fixed Point Theorem 2

Lets build up.

Fixed Point Theorem 1: Let g(x) be continuous on [a, b] with range contained in [a, b]. Suppose that there exists a constant 0 ≤ k < 1 such that |g0(x)| ≤ k for all x ∈ (a, b). Then the equation x = g(x) has exactly one soultion in [a, b]

Proof. Since the range is contained in [a, b], the solution must be in [a, b]. |g0(x)| ≤ k < 1 implies g0(x) exists and is uniformly bounded on (a, b). Suppose there were two different solutions, p and q, in [a, b].

Thus p = g(p), q = g(q), p 6= q. By the Mean-Value Theorem there is a point ξ between p and q such that

g(p) − g(q) = (p − q)g0(ξ) = p − q

which implies g0(ξ) = 1 which is a contradiction.

(13)

s

a

s

b b s

b s

(14)

Fixed Point Theorem 2

Fixed Point Theorem 2:

Let g(x) be continuous on [a, b] with range contained in [a, b]. Suppose that there exists a constant k < 1 such that |g0(x)| ≤ k for all x ∈ (a, b). Then, if x0 ∈ [a, b], the sequence generated by our iterative function (xn+1 = g(xn)) converges to the unique solution to x = g(x) which lies in [a, b].

Moreover an error bound is given by |xn − s| ≤ kn(b − a), where s is the fixed point.

Proof: All points of the sequence are in [a, b] since xn+1 = g(xn) is in [a, b] if xn is in [a, b], and since x0 is in [a, b] it follows that they all are. Let s denote the unique fixed point to g, if for some n, xn = s, then xn+1 = g(xn) = g(s) = s and so xn+1 = s also, so by induction all numbers in the sequence after also = s.

If this isn’t the case, that is xn 6= s for all n. Using the Mean-Value Theorem on the given interval with endpoints s and xn there is some point ξ between s

(15)

and xn such that

g(xn) − g(s) = (xn − s)g0(ξ) = xn+1 − s Since g0 is bounded by k it follows that

|xn+1 − s| ≤ k|xn − s| ≤ k(b − a)for n = 0, 1, . . . in particular,

|x1 − s| ≤ k(b − a)

|x2 − s| ≤ k|x1 − s| ≤ k2(b − a) ...

|xn − s| ≤ k|xn−1 − s| ≤ kn(b − a)

Because |k| < 1 , limn→∞ kn(b − a) = 0, thus limn→∞ |xn − s| = 0, or limn→∞ xn = s.

(16)

Fixed Point Theorem 2: Example

Using Theorem 2 does g(x) = .5(x+5/x) converge in the interval [2, 3]

g0(x) = .5(1 − 5 x2)

Are the hypothesis of Theorem 2 satisfied? Is the range the same as the domain?

Determine max and min values of g(x) on [2, 3]

to verify the range is contained in [2, 3]. g0(x) = 0 when x = √

5. Thus the relative extrema are at x = 2, x = 51/2, x = 3.

g(2) = 9/4 g(51/2) = 51/2 g(3) = 7/3

So if x ∈ [2, 3] then g(x) ∈ [2, 3].

(17)

Fixed Point Theorem 2: Example

Now check to see if there is a k such that |g0(x)| ≤ k < 1. How?

Find max of |g0(x)| on [2, 3] so look at g00(x) = x53. There are no zeros on [2, 3] so we only need to look at endpoints for extrema.

g0(2) = −1/8 g0(3) = 2/9

Thus |g0(x)| ≤ k = 2/9 on [2, 3].

So by Theorem 2, if x0 ∈ [2, 3], the series {xn} converges to 51/2 and a bound on the error is |xn − 51/2| ≤ (2/9)n

(18)

Fixed Point Theorem 3

Let x = g(x) be an equation with a solution p.

Suppose that there exists a δ > 0 and a constant k < 1 such that |g0(x)| ≤ k whenever |x − p| ≤ δ.

If p0 ∈ I = [p − δ, p + δ] then the sequence {pn} generated by iteration (xn+1 = g(xn)) converges to p.

Proof: if x ∈ I and x 6= p then the mean value theorem applies to g(x), on the interval with end points x and p. Thus, there exists a point ξ between x and p such that

g(x) − g(p) = (x − p)g0(ξ) Because |g0(ξ)| ≤ k < 1 it follows that

|g(x) − g(p)| ≤ k|x − p| < |x − p| ≤ δ

Finally, because g(p) = p we have shown that g(x) is closer to p than is x. That is, the range of g(x), for the domain I, is contained in I.

(19)

Fixed Point Theorem 3

Since the hypothesis of Theorem 2 are met so are the conclusions so Theorem 3 is correct.

Corollary III: Let x = g(x) be an equation with a solution p. If g0(x) is continuous at p and g0(p) = 0 then the hypothesis, and therefore the conclusion of theorem 3 are satisified for some interval about p.

This tells us that the closer we get to a solution the faster we will converge.

(20)

Convergence Examples

f (x) = x3 − x − 5

We want a g(x) such that x = g(x) is the same as f (x) = 0

x3 − x − 5 = 0

x3 − 5 = x g1(x) = x3 − 5 x3 − x − 5 = 0

x(x2 − 1) = 5

x = x25−1 g2(x) = x25−1 x3 − x − 5 = 0

x3 = x + 5 g3(x) = (x + 5)13

f (r) = 0 r ≈ 1.9

g10(x) = 3x2 g10(1.9) ≈ 10.83 g20(x) = (x−10x2−1)2 g20(1.9) ≈ −196.8 g30(x) = 13(x + 5)−23 g30(1.9) ≈ 1

3(6.9)23

(21)

Region of Convergence

Let’s expand f (x) about xn

f (x) = f (xn) + f0(xn)(x − xn) + f00(ξ)(x − xn)2 2

Now set x = α where α is a root of f (x);

f (α) = 0 = f (xn) + f0(xn)(α − xn) + f00(ξ)(α − xn)2 2

f (xn) + f0(xn)(α − xn) = −f00(ξ)(α − xn)2 2

f (xn)

f0(xn) + (α − xn) = −f00(ξ)(α − xn)2 2f0(xn) since xn+1 = xnff (x0(xnn)) (Newton’s method)

(α − xn+1) = −f00(ξ)(α − xn)2 2f0(xn)

(22)

Region of Convergence

(α − xn+1) = −f00(ξ)(α − xn)2 2f0(xn) let n = |xn − α|

n+1 = f00(ξ)e2n 2f0(xn)

let m = max

f00(ξ) 2f0(x)

 in I = [α ± x0]

n+1 = 2n or mn+1 = (mn)2

So the sequence {xn} approaches α if m0 < 1 or m|α − x0| < 1

(23)

Region of Convergence: Example

Let f (x) = x2 − C and f(α) = 0 and using the iteration formula: xn+1 = 12 

xn + xC

n

 Case 1: x > α

xn+1 = xn 2



1 + C x2n



= Kxn

Since xC2

n < 1, K < 1, and the minimum of g(x) is α at xn = α. α ≤ xn+1 < xn and the sequence, {x0, x1, . . .} converges to α.

Case 2: 0 < x < α xn+1 = 1

2



xn + C xn



The minimum of g(x) at xn = α and its value is α.

So, xn+1 > α bringing the result of the first iteration to the right side of α. We know from case 1 that any iteration starting at x > α will converge to the root.

(24)

Iteration: Graphical Analysis

We find a g(x) such that f (x) = 0 is that same as x = g(x)

So geometrically what does that tell us about the root of f (x)?

y = g(x) y = x

(25)

Iteration: Graphical Analysis

y = g(x) y = x

Where is the root of f (x)?

(26)

Iteration: Graphical Analysis

y = g(x) y = x

s

x0

(27)

Iteration: Graphical Analysis

y = g(x) y = x

s

x0

(28)

Roots

What does it mean to be a root? Where the function crosses the X-axis

Why do we want to find roots? It is an extremely useful piece of info to know.

How would you find the intersection between two curves? Set them equal and find the root.

A root r, of function f occurs when f (r) = 0.

For example:

f (x) = x2 − 2x − 3

has 2 roots at r = −1 and r = 3. We can verify by substitution. We can also look at f in its factored form.

f (x) = x2 − 2x − 3 = (x + 1)(x − 3)

(29)

Bisection Method

a b

tf (a)

t

f (b)

if f (a)f (b) < 0 we have a root. Why? Yes. The signs are different so it must cross x-axis an odd # of times. Touching the axis isn’t a cross.

Can we have a root if f (a)f (b) > 0. Why? Yes if it crosses an even # of times.

If we pick an x, a < x < b, will either [a, x] or [x, b]

have a root if [a, b] has a root? Yes.

(30)

Bisection Method:Algorithm

Find an interval (The smaller the better) [a, b]

where the the y values at the endpoints are on different sides of the x-axis.

Pick the midpoint, c, of [a, b] and see if that is a root, if it is not see if the root is in [a, c] or [c, b] and repeat on the appropriate new interval.

(31)
(32)

Bisection Method

Our estimate of the root, c, is:

c = a + b 2

if (f (a)f (c) < 0 root is in [a, c]

if (f (b)f (c) < 0 root is in [b, c]

No we have a new interval, if we repeat we should get closer to the root. What happens to the interval at each interation? It is halved.

ci = ai + bi 2 c0 = a0 + b0

2 c1 = a1 + b1

2

(33)

Bisection Method:Convergence

What is the error of c0, c1, c2, . . . , ci?

0 = |r − c0| ≤ a0+b2 0

1 = |r − c1| ≤ a1+b2 1a02+b2 0

2 = |r − c2| ≤ a2+b2 2a12+b2 1a02+b3 0

i = |r − ci| ≤ a20i+1+b0

So does the bisection method converge? Yes!

How fast? O(log2(n))!

So for an error tolerance of  b − a

2i+1 ≤ 

i > log(b − a) − log(2) log(2)

(34)

Bisection Method:Example

Find a root for f (x) = x3 + 2x2 − 11x − 12 = (x + 1)(x − 3)(x + 4)

Find a root in the interval [4,5]?

f (4) = 43+2×42−11×4−12 = 64+32−44−12 = 40

f (5) = 53+2×52−11×5−12 = 125+50−55−12 = 108 Now What?

How about between [2, 5]

f (2) = 23 + 2 × 22 − 11 × 2 − 12 = 8 + 8 − 22 − 12 = −18

f (2 + 5

2 ) = f (3.5) = 3.53 + 2 × 3.52 − 11 × 3.5 − 12 = 16.875 f (2 + 3.5

2 ) = f (2.75) = 2.753 + 2 × 2.752 − 11 × 2.75 − 12 = −6.4 f (2.75 + 3.5

2 ) = f (3.13) = 3.9 f (2.75 + 3.13

2 ) = f (2.94) = −1.6, f (3.04 + 3.13

2 ) = f (2.94) = 1.2

(35)

Bisection Method:Convergence Example

Say we wanted to compute a root to 32-bit binary precision. How many iterations would be needed if a=16 and b=17?

a = (10000.0)2 and b = (10001.0)2, Thus we already know four of the binary digits, so we have 20 left. If the last one is to be correct the error must be less than 220 (you can also use 221 to be conservative) so

b−a

2n+1 < 

17−16

2n+1 < 2−20 2n+1 > 220 n ≥ 20

or

n > log 1−log 2−19 log 2

(36)

Bisection Method:Implementation Issues

What are some issues to be concerned about during implementation? Starting range.

How would you check to see which side of c the root is on? equality of signs.

We can use f (a)f (c) to see if the numbers have different signs. Are there any problems with this?

overflow, underflow.

Are we concerned with the number of calls to the function to evaluate it? Why? Could be expensive.

Will this always converge? Yes if it can be started properly.

(37)

Newton’s Method

x0

v f (x0)

x1

v

x2

v 



Basic idea is to approximate f (x) by a straight line l(x). Lets use a line tangent to the function at a current guess as the approximation.

(38)
(39)

Newton’s Method

The tangent line is the deriviative. So:

l(x) = f0(x0)(x − x0) + f (x0)

So we use the root of l(x) as a new approximation to the root of f (x) and repeat the procedure.

This gives us a sequence:

x1 = x0ff (x0(x00)) x2 = x1ff (x0(x11)) x3 = x2ff (x0(x22))

...

xn+1 = xnff (x0(xnn))

Does this method always converge? Why? No, because the derivative doesn’t really predict the future.

(40)

Newton’s Method:Bad functions

x0

s

x1

s

x2

s LL

LL LL LL LL LL LLL

HH HH HH HH HH HH HH HH HH

x0

s

x1

s

































x0

s

(41)

Newton’s Method:Convergence

Doesn’t always converge. But when it does converge, how fast does it do it?

Let’s use Taylor’s series and the error term Rn. Expand about the point xn

f (x) = f (xn) + f0(xn)(x − xn) + f00(ξ)(x − xn)2 2

Remember ξ is between x and xn. Let’s set x to the root r, so f (r) = 0.

f (r) = 0 = f (xn) + f0(xn)(r − xn) + f00(ξ)(r − xn)2 2

Now divide by f0(xn)

= f (xn)

f0(xn) + f0(xn)(r − xn)

f0(xn) + f00(ξ)(r − xn)2 2f0(xn)

Reduce terms and substitute our iterative formula.

xn − xn+1 + (r − xn) + f00(ξ)(r − xn)2 2f0(xn)

(42)

Newton’s Method:Convergence

Rearrange.

r − xn+1 = − f00(ξ)

2f0(xn)(r − xn)2 What is r − xn+1? It is the n + 1 error or

n+1 = |r − xn+1|

n = |r − xn|

So if we take the absolute value of both sides above we get

n+1 = r − xn+1 = f00(ξ) 2f0(xn)2n

So we are converging quadratically, but only if we are close to the root r and f0(r) 6= 0.

With quadratic convergence we almost double the number of significant digits after each iteration.

if f0(0) = 0 then we get at best linear convergence.

See the book for the analysis.

(43)

Newton’s Method:Example

Find a root for f (x) = x3 + 2x2 − 11x − 12 = (x + 1)(x − 3)(x + 4)

What is f0(x):

f0(x) = 3x2 + 4x − 11

So

xn+1 = xn − f (xn)

f0(xn) = xn − x3n + 2x2n − 11xn − 12 3x2n + 4xn − 11

Start with x0 = 30

x1 = 30 − 28458

2809 = 10.13100748

x2 = 10.13100748 − 3.324046727 = 6.806960753 x3 = 6.806960753 − 2.069106062 = 4.737854691 x4 = 4.737854691 − 1.157209403 = 3.580645288 x5 = 3.580645288 − .4825214818 = 3.098123806 x6 = 3.098123806 − .09455278358 = 3.003571022

(44)

Newton’s Method:Implementation Issues

What problems do we have?

Well we must be able to evaluate f0(x) which isn’t always the case.

We have to evaluate f (x) and f0(x) for each iteration, which could be expensive.

Does not gaurantee convergence!

But it is very fast.

So when did Newton live?

(45)

Secant Method

x0

v f (x0)

x1

v x2

Lets approximate the function with a line, but this time a line between two approximations to the root called a secant line.

(46)
(47)

Secant Method

So what is x2?

Let’s equate the slopes between (x0, f (x0)), (x1, f (x1)) and (x1, f (x1)), (x2, 0):

f (x1) − f(x0)

x1 − x0 = 0 − f(x1) x2 − x1 Rearrange:

x2 = x1 − f(x1) x1 − x0 f (x1) − f(x0)

Do the same thing for x3 using x1 and x2, in general:

xn+1 = xn − f(xn) xn − xn−1 f (xn) − f(xn−1)

This is a two point method since two approximations are needed.

(48)

Secant Method vs Newton’s Method

This is the same as the Newton method except that we approximate f0(x) as

f0(x) = f (xn) − f(xn−1) xn − xn−1

So how does this compare to Newton’s method?

How many function evaluations are needed per iteration? 1

Will it converge? Maybe

How about speed of convergence?

|n+1 ≈ C|n|α α = 1/2(1 + √

5) ≈ 1.62

This isn’t quite as fast as Newton’s method but still superlinear and much faster than bisection. But if you consider number of evaluations it is faster.

(49)

Secant Method:Example

Find a root for f (x) = x3 + 2x2 − 11x − 12 = (x + 1)(x − 3)(x + 4)

What is the secant iteration formula.

xn+1 = xn − f(xn) xn − xn−1 f (xn) − f(xn−1)

Lets use x0 = 5 and x1 = 0 f (5) = 108

f (0) = −12

x2 = x1 − f(x1)f (xx1−x0

1)−f (x0) = 0 − −12−12−108)−12−5 = 12 x3 = x2 − f(x2)f (xx2−x1

2)−f (x1) == −1613 = −1.2307 x4 == −.99176

x4 == −1.000198215

(50)

Secant Method:Implementation Issues

What are some issues to be concerned about during implementation? Starting values. The convergence properties depend on the starting values.

Does not gaurantee convergence!

What if we pick x0 and x1 such that f (x0) = f (x1)?

Who is this method named after?

(51)

Root finding comparison

Bisection Method

• Converges slowly

• Need two starting values on opposite sides of the root

• Always converges.

Newton’s Method

• Converges quickly near the root

• Needs two function evaluations per iteration.

• Need to be able to evaluate f0

• May not converge.

(52)

Root finding comparison

Secant Method

• Converges quickly near the root, but not as fast as Newton’s.

• Needs one function evaluation per iteration.

• Doesn’t need f0

• May not converge.

So what is the best method?

Well hybrids are the most popular, use bisection for so long then switch to Newton’s or Secant.

(53)

Polynomial Roots

P1(z) = anzn + an−1zn−1 + . . . + a0 And in factored form

P1(z) = (z − z1)n1(z − z2)n2 . . . (z − zq)nq

with root zi having multiplicity ni. If an = 1 then the polynomial is monic.

If zi is a root of multiplicity ni for P1 then it is a root of multiplicity ni − 1 of P10.

P1(z) = (z − zi)niS(z)

P10(z) = ni(z − zi)ni−1S(z) + (z − zi)niS0(z)

= (z − zi)ni−1[niS(z) + (z − zi)S0(z)]

= (z − zi)ni−1T (z)

Therefore zi is of multiplicity at least ni − 1 for P10. Is it more?

(54)

Polynomial Roots

If the multiplicity of zi was more than ni − 1 for P10

then zi must be a root of T (z) ⇒ T (zi) = 0 But T (zi) = niS(zi) + (z − zi)S0(zi) = niS(zi) 6= 0 Therefore it must be of multiplicity ni − 1

If the root (multiplicity) structure of P1 is P1(z) = (z − z1)n1(z − z2)n2 . . . (z − zq)nq then P1 and P10 have the common divisor

P2(z) = (z − z1)n1−1(z − z2)n2−1 . . . (z − zq)nq−1 and furthermore, that common divisor is a gcd.

(55)

Polynomial Roots

Remember the Euclidean gcd algorithm? Lets use that algorithm on these polynomials to get

P1(z) = Q1(z)P0(z) + R1(z) P10(z) = Q2(z)R1(z) + R2(z) R1(z) = Q3(z)R2(z) + R3(z) R2(z) = Q4(z)R3(z) + R4(z) ...

Rs−2(z) = Qs(z)Rs−1(z) + Rs(z) Rs−1(z) = Qs+1(z)Rs(z)

Then convert Rs(z) to monic form and call it P2(z) and repeat the Euclidean algorthm for gcd, getting the polynomials P1, P2, P3, . . . , Pq, 1

What happens to the multiplicity of each root as we proceed? They reduce by one.

So what is the multiplicity of root zi for polynomial Pj? ni − (j − 1) = ni − j + 1

(56)

Polynomial Roots

Let’s divide P1 by P2 this gives P1(z)

P2(z) = M1(z) = (z − z1)(z − z2) . . . (z − zq) Let’s define

Mi(z) =

( Pi(z)

Pi+1(z) if i < q Pq(z) if i = q

The polynomials M1, M2, . . . , Mq all have only simple roots.

So for polynomial Mi what can we say about it’s roots? It has all roots that have at least multiplicity i of P and they are simple roots here.

(57)

Polynomial Roots

Let’s also create the polynomials Ni as

Ni(z) =

( Mi(z)

Mi+1(z) if i < q Mq(z) if i = q

The sum of the degrees of the Nis is q and each distinct root of P1(z) is a root of just one Ni and if the root is of multiplicity of m then it is a root of Nm(z).

(58)

Polynomial Roots:Example

P1(z) = z7 − 3z5 + 3z3 − z P10(z) = 7z6 − 15z4 + 9z2 − 1 Lets apply the Euclidean algorithm.

z7−3z5+3z3−z = 1

7z(7z6−15z4+9z2−1)−6

7z5+12

7 z3−6 7z 7z6−15z4+9z2−1 = −49

6 z(−6

7z5+12

7 z3−6

7z)−z4+2z2−1

−6

7z5 + 12

7 z3 − 6

7z = 6

7z(−z4 + 2z2 − 1) + 0

Now make the last polynomial monic we get P2(z) = z4 − 2z2 + 1 and repeat on P2

z4 − 2z2 + 1 = 1

4z(4z3 − 4z) − z2 + 1 4z3 − 4z = −4z(−z2 + 1) + 0

(59)

Polynomial Roots:Example

Set P3(z) = z2 − 1, and repeat z2 − 1 = 1

2z(2z) − 1 2z = −2z(−1) + 0

So P4(z) = 1 Let find the Mi’s and Ni’s.

M1(z) = PP1(z)

2(z) = z3 − z M2(z) = PP2(z)

3(z) = z2 − 1 M3(z) = P3(z) = z2 − 1 N1(z) = MM1(z)

2(z) = z N2(z) = MM2(z)

3(z) = 1

N3(z) = M3(z) = z2 − 1

So the roots of P1(z) are 0, 1 (with multiplicity 3) and -1 (with multiplicity 3) or

P1(z) = (z − 1)3(z + 1)3(z)

(60)

Polynomial Roots:Example 2

What is the factored form of the polynomial

p = 2x9 − 12x8 + 10x7 + 84x6 − 258x5 + 252x4+ 46x3 − 276x2 + 200x − 48

P1(x) = x9 − 6x8 + 5x7 + 42x6 − 129x5 + 126x4 +23x3 − 138x2 + 100x − 24

P2(x) = x5 − 7x4 + 19x3 − 25x2 + 16x − 4 P3(x) = x3 − 4x2 + 5x − 2

P4(x) = x − 1 P5(x) = 1

M1(x) = x4 + x3 − 7x2 − x + 6 = (x − 2)(x + 3)(x − 1)(x + 1) M2(x) = x2 − 3x + 2 = (x − 2)(x − 1)

M3(x) = x2 − 3x + 2 = (x − 2)(x − 1) M4(x) = x − 1

M5(x) = 1

N1(x) = x2 + 4x + 3 = (x + 3)(x + 1) N2(x) = 1

N3(x) = x − 2 N4(x) = x − 1

p(x) = 2(x − 2)3(x + 3)(x − 1)4(x + 1)

(61)

Polynomials

What is a polynomial?

Let φ1(x), φ2(x),. . . φn(x) be a given set of functions. A polynomial p in the base functions {φi(x)}

is any function of the form:

p(x) = a1φ1(x) + a2φ2(x) + . . . + amφm(x)

where the ai are constants. If the base functions are 1, x, x2, . . . , xn then a polynomial

p(x) = a0 + a1x + a2x2 + . . . + anxn =

n

X

i=0

aixi

is called an algebraic polynomial in one variable. If an 6= 0 then the polynomial p is of degree n. We will consider only these types of polynomials.

(62)

Evaluating Polynomials

If we evaluate a polynomial naively how expensive (in terms of computions) is it?

a0

a1x 1 multiply a2xx 2 multiplies a3xxx 3 multiplies ...

anxx . . . x n multiplies

This gives 1 + 2 + 3 + . . . + n = n(n+1)2 multiplies.

Then another n additions to add them together.

Are there better ways? i.e. less calculations.

(63)

Evaluating Polynomials

What if we notice that xn = xn−1x. Since we already calculated xn−1 why calculate it again? So now we have:

a1x 1 multiply a2xx 2 multiplies a3xx2 2 multiplies a4xx3 2 multiplies ...

anxxn−1 2 multiplies

Which is 2n − 1 multiplies and n additions. This is much cheaper than the naive method.

Are there still better ways?

(64)

Evaluating Polynomials

What if we write the polynomial in nested form:

p(x) = a0 + x(a1 + x(a2 + . . . + x(an−1 + x(an)) . . .)

Now we have n multiplications and n additions.

(65)

Synthetic Division (Horner’s Method)

Horner’s only contribution to mathematics was published in 1819, however the same method was proposed by ruffini a few years earlier. But, the method was actually published by chu in 1303, but we know it as Horner’s method because of de Morgan

If a polynomial p and a complex number x0 are given then Horner’s algorithm will give the number p(x0) and the polynomial

q(x) = p(x) − p(x0) x − x0

q is of degree one less then p and by rearranging p(x) = (x − x0)q(x) + p(x0)

If we rewrite the unknown polynomial q and the known polynomial p in the following form

p(x) = a0 + a1x + a2x2 + . . . + anxn

q(x) = b0 + b1x + b2x2 + . . . + bn−1xn−1 and plug in to the above formula we get

(66)

Synthetic Division (Horner’s Method)

bn−1 = an

bn−2 = an−1 + x0bn−1 ...

b0 = a1 + x0b1 p(x0) = a0 + x0b0

We can use this method to evaluate a polynomial quickly with n multiplies and n additions. This method is also good for doing things by hand. To do it by hand set it up the following way

an an−1 an−2 . . . a0 x0 x0bn−1 x0bn−2 . . . x0b0 bn−1 bn−2 bn−3 . . . p(x0)

(67)

Synthetic Division: Example

Evaluate p(x) = x4 − 4x3 + 7x2 − 5x − 2 at x = 3;

First set up the knowns:

1 −4 7 −5 −2 3

Start doing the multiplies and adds:

1 −4 7 −5 −2

3 3 ∗ 0 3 ∗ 1 1 + 0 −4 + 3 and finally

1 −4 7 −5 −2

3 3 −3 12 21

1 −1 4 7 19

so p(3) = 19 and we can write p as

p(x) = (x − 3)(x3 − x2 + 4x + 7) + 19

(68)

Synthetic Division: Deflation

We can also use synthetic division for deflation, which is removing a linear factor from a polynomial. If x0 is a zero of the polynomial p, then x − x0 is a factor of p, and conversely. And the remaining zeros (roots) of p are the n − 1 roots of p(x)/(x − x0).

Example Deflate p(x) = x4 − 4x3 + 7x2 − 5x − 2 by noting that 2 is a root.

First set up the synthetic division and calculate p(2)

1 −4 7 −5 −2

2 2 −4 6 2

1 −2 3 1 0

Since p(2) = 0, we know it is indeed a root. And

x4 − 4x3 + 7x2 − 5x − 2 = (x − 2)(x3 − 2x2 + 3x + 1)

(69)

Synthetic Division: Derivative

Sometime we need not only to calculate p(x) but also p0(x) such as with Netwon’s method. Using synthetic division

p(x) = (x − x0)q(x) + p(x0)

If we use the product rule and differentiate we get p0(x) = q(x) + (x − x0)q0(x)

So p0(x0) is

p0(x0) = q(x0) + (x0 − x0)q0(x0) = q(x0)

Thus p0(x0) can be found by synthetic division on the polynomial q(x)

(70)

Synthetic Division: Derivative Example

For p(x) = 2x5 − 7x3 + 4x − 5 find p(3) and p0(3).

Find p(3) by synthetic division

2 0 −7 0 4 −5

3 6 18 33 99 309

2 6 11 33 103 304 Now find p0(3)

2 6 11 33 103

3 6 36 141 522

2 12 47 174 625 So p0(3) = 625

(71)

Synthetic Division and Taylor expansion

Let p(x) be a polynomial and suppose we desire the coefficients ck in the equation

p(x) = anxn + an−1xn−1 + . . . + a0

= cn(x − x0)n + cn−1(x − x0)n−1 + . . . + c0

Using Taylor’s Theorem we know that ck = p(k)k!(x0). This is expensive to calculate. Is there a more efficient way?

Notice that p(x0) = c0. And this coefficient (c0) can be obtained by applying synthetic division to the polynomial p, using the point x0. We also obtain the polynomial q(x) such that

q(x) = p(x) − p(x0) x − x0

(72)

Synthetic Division and Taylor expansion

q(x) = p(x) − p(x0) x − x0

= cn(x − x0)n + cn−1(x − x0)n−1 + . . . + c0

x − x0 − p(x0)

x − x0

= cn(x−x0)n−1+cn−1(x−x0)n−2+. . .+c1+ c0

x − x0− c0 x − x0

= cn(x − x0)n−1 + cn−1(x − x0)n−2 + . . . + c1 This shows that the coefficient c1 can be obtained by applying synthetic division to the polynomial q with point x0 because c1 = q(x0). We repeat the process until all coefficients ck are found.

(73)

Synthetic Division: Taylor Expansion:Example

Find the Taylor expansion of the polynomial p about the point 3.

p(x) = x4 − 4x3 + 7x2 − 5x − 2

1 −4 7 −5 −2

3 3 −3 12 21

1 −1 4 7 19

3 3 6 30

1 2 10 37

3 3 15

1 5 25

3 3

1 8

So

p(x) = (x−3)4+8(x−3)3+25(x−3)2+37(x−3)+19

(74)

Polynomials: Localizing zeros

Therorem: All zeros of the polynomial p(x) = anxn + an−1xn−1+ . . . + a0 lie in the open disk whose center is at the origin of the complex plane and whose radius is

ρ = 1 + |an|1 max

0≤k≤n |ak|

Example: Find a disk centered at the origin that contains all the zeros of the polynomial

p(x) = x4 − 4x3 + 7x2 − 5x − 2

One such disk would be ρ = 1 + |a4|−1 max

0≤k≤n|ak| = 1 + 7

1 = 8

(75)

Polynomials: Localizing zeros

Take our polynomial p(x) = anxn + an−1xn−1 + . . . + a0 and consider the function s(x) = xnp(1/x).

This gives

s(x) = xn

"

an  1 x

n

+ an−1  1 x

n−1

+ . . . + a0

#

= an + an−1x + . . . + a0xn

s is a polynomial of degree at most n. It has the same coefficients as p except in reverse order. Also for a nonzero complex number z the condition p(z) = 0 is equivalent to s(1/z) = 0. Thus

Therorem:If all the zeros of s are in the disk{z :

|z| ≤ ρ}, then all the nonzero roots (r is a root and r 6= 0) of p are outside of the disk{z : |z| < ρ−1}

(76)

Polynomials: Localizing zeros

Example: Find a disk centered at the origin that contains no zeros of p, where

p(x) = x4 − 4x3 + 7x2 − 5x − 2

Using the above thereom we get

s(x) = −2x4 − 5x3 + 7x2 − 4x + 1

and all zeros of s lie in a disk centered at the origin with a radius of

ρ = 1 + |a4|−1 max

0≤k≤n|ak| = 9 2

Therefore, the zeros of p lie outside the disk of radius 29 and furthermore all zeros of p lie in the ring 29 < |z| < 8 in the complex plane.

참조

관련 문서

If the service gas and calibration gas are the same, the measured gas is ensured to be ±1% of full scale in the specified operating pressure range as to our primary

함수에 사칙 연산과 합성 연산을 적용하는 방법을

[r]

전번 학기 수업

프로그램의 선두에서 시작하고자 하는 경우에는 EDIT 모드에서 RESET 을 누릅니 다... 미러이미지(Mirror Image)에서 오버런 이송시 방 향은

 '(&amp;‹ DI) DDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDD EE I F á &lt;6Fd‹  ‘‹ ”x DI) DDDDDDDDDDDDDDD E.  &lt;6Fd‹ 

미셀구조는 계면활성제의 소수성 부분은 중심부에 모여 핵을 형성하고 친수성 부분은 물과 접... 촉하는 외곽부분을 형성하는 구조로 기름과 같은 소수성물질이

XAFS: X-ray absorption fine structure XES: X-ray emission spectroscopy XRF: X-ray fluorescence.. Use of x-rays; a probe based