Calculus of Variations/CHAPTER VIII

From testwiki
Jump to navigation Jump to search

CHAPTER VIII: THE SECOND VARIATION; ITS SIGN DETERMINED BY THAT OF THE FUNCTION F1.

  • 111 Nature and existence of the substitutions introduced.
  • 112 The total variation.
  • 113,114 The second variation of the function F.
  • 115 The second variation of the integral I. The sign of the second variation in the determination of maximum or minimum values.
  • 116 Discontinuities.
  • 117 The sign of the second variation made to depend upon that of F1.
  • 118 The admissibility of a transformation that has been made. The differential equation J=0.
  • 119 A simple form of the second variation.
  • 120 A general property of a linear differential equation of the second order.
  • 121 The second variation and the function F1. The function F1 cannot change sign and must be different from 0 and in order that there may be a maximum or a minimum.

Article 111.
The substitution x+ϵξ, y+ϵη for x,y causes any point of the original curve to move along a straight line, which makes an angle with the X-axis whose tangent is ηϵ.

This deformation of the curve is insufiEcient, if we require that the point move along a curve other than a straight line.

To avoid this inadequacy we make the more general substitution (by which the regular curve remains regular):

xx+ϵξ1+ϵ22!ξ2+, yy+ϵη1+ϵ22!η2+

where, like ξ, η in our previous development (Art. 75), the quantities ξ1, η1, ξ2, η2 are functions of t, finite, continuous one-valued and capable of being differentiated (as far as necessary) between the limits t0t1. These series are supposed to be convergent for values of ϵ such that |ϵ|<1.

That such substitutions exist may be seen as follows:

Since the curve is regular, the coordinates of consecutive points to P0 and P1 may be expressed by series in the form, say,

(A)x0+ϵa0(1)+ϵ22!a0(2)+, y0+ϵb0(1)+ϵ22!b0(2)+
(B)x1+ϵa1(1)+ϵ22!a1(2)+, y1+ϵb1(1)+ϵ22!b1(2)+

where the coefficients of the powers of ϵ are constants and the series are convergent.

Suppose, now, that we seek to determine the functions of t

(C)x+ϵξ1+ϵ22!ξ2+, y+ϵη1+ϵ22!η2+

such that for t=t0 and t=t1, the expressions (C) will be the same as (A) and (B).

This may be done, for example, by writing

ξ1=t2+α1t+α2,
η1=t2+β1t+β2,

and then determine α1, α2, β1, β2 in such a way that

t02+α1t0+α2=a0(1); t02+β1t0+β2=b0(1),
t12+α1t1+α2=a1(1); t12+β1t1+β2=b1(1).

From this it is seen that

α1=(t1+t0)+a1(1)a0(1)t1t0, etc.

In the same way we may determine quadratic expressions in t for ξ2, η2, etc.

The substitutions thus obtained are of the nature of those which we have assumed to exist, and may evidently be constructed in an infinite number of different ways.

Article 112.
Making the above substitutions in the integral

I=t0t1F(x,y,x,y)dt,

it is seen that

ΔI=t0t1[F(x+ϵξ1+ϵ22!ξ2+,y+ϵη1+ϵ22!η2+,x+ϵξ1+ϵ22!ξ2+,y+ϵη1+ϵ22!η2+)F(x,y,x,y)]dt
=ϵδI+ϵ22!δ2I+ϵ33!δ3I+.

By Taylor's Theorem we have

F(x+ϵξ1+ϵ22!ξ2+,y+ϵη1+ϵ22!η2+,x+ϵξ1+ϵ22!ξ2+,y+ϵη1+ϵ22!η2+)F(x,y,x,y)
=[(ϵξ1+ϵ22!ξ2+)x+(ϵη1+ϵ22!η2+)y+(ϵξ1+ϵ22!ξ2+)x+(ϵη1+ϵ22!η2+)y]F
+12![(ϵξ1+ϵ22!ξ2+)x+(ϵη1+ϵ22!η2+)y+(ϵξ1+ϵ22!ξ2+)x+(ϵη1+ϵ22!η2+)y]2F
+13![]3F+.

The coefficient of ϵ in this expression is the integrand of δI and is zero; while the coefficient of ϵ22! involves terms that are the first partial derivatives of F, and also those that are the second partial derivatives of F.

The first partial derivatives of F that belong to this coefficient, when put under the integral sign, may be written in the form

t0t1[Fxξ2+Fxξ2+Fyη2+Fyη2]dt=t0t1G(yξ2xη2)dt+[]t0t1

(see Art. 79), and this expression is also zero, if we suppose that the end-points remain fixed.

Article 113.
The coefficient of ϵ2 in the preceding development of F by Taylor's Theorem is, neglecting the factor 12!, denoted by δ2F.

We have then

1)δ2F=2Fx2ξ12+22Fxyξ1η1+2Fy2η12+2Fx'2ξ1'2+22Fxyξ1η1+2Fy'2η1'2+2(2Fxxξ1ξ1+2Fyyη1η1+2Fxyξ1η1+2Fyxη1ξ1).

The subscripts may now be omitted and the formula simplified by the introduction of the function F1, which (Art. 73) was defined by the relations:

2)2Fx'2=y'2F1, 2Fxy=xyF1, 2Fy'2=x'2F1;

and by introducing the new notation:

3)L=2FxxyyF1, M=2Fxy+xyF1=2Fxy+yxF1 (owing to the equation G=0), N=2FyyxxF1;

where x, y are used for d2xdt2,d2ydt2.

We have then

δ2F=2Fx2ξ2+22Fxyξη+2Fy2η2+F1(y'2ξ'22xyξη+x'2η'2)+2F1(yyξξ+xxηηxyξηyxηξ)+2(Lξξ+M(ξη+ηξ)+Nηη).

To get an exact differential as a part of the right-hand member of this formula, we write

4)R=Lξ2+2Mξη+Nη2,

an expression which, differentiated with respect to t, becomes

2[Lξξ+M(ξη+ηξ)+Nηη]=dRdtdLdtξ22dMdtξηdNdtη2.

We further write

5)w=yξηx,

where (see Art. 81) w is, neglecting the factor 1x'2+y'2, the amount of the sliding of a point of the curve in the direction of the normal.

Differentiating with respect to t, we have

dwdt=yξxη+yξxη,

from which it follows that

c=(yξxη)2+(yξxη)2+2(yyξξ+xxηηxyξηyxηξ).

Then the expression for the second variation becomes

δ2F=2Fx2ξ2+22Fxyξη+2Fy2η2+F1[(dwdt)2(yξxη)2]+dRdt(dLdtξ2+2dMdtξη+dNdtη2).

If further we write in this expression

6)L1=2Fx2F1y'2dLdt, M1=2Fxy+F1xydMdt, N1=2Fy2F1x'2dNdt,

we have finally

δ2F=F1(dwdt)2+L1ξ2+2M1ξη+N1η2+dRdt.

Article 114.
It follows from 3) that

Lx+My=x2Fxx+y2Fxy.

Owing to the homogeneity of the function F (Chap. IV), it is seen from Euler's Theorem that

F=xFx+yFy,

and consequently,

Fx=x2Fxx+y2Fxy;

and therefore

Fx=Lx+My.

In a similar manner we have

Fy=Mx+Ny.

Differentiating with regard to t, the above expression becomes

ddt(Fy)=2Fx2x+2Fxyy+2Fxxx+2Fxyy=dLdtx+dMdty+Lx+My,

which, owing to 3) is

x(2Fx2F1y'2dLdt)+y(2Fxy+F1yxdMdt)=0;

or from 6)

xL1+yM2=0.

In an analagous manner it may be shown that

xM1+yN1=0.

From these expressions we have at once

L1y'2=M1xy=N1x'2=F2,

where F2 is the factor of proportionality.

It follows that

7)L1=y'2F2,M1=xyF2,N1=x'2F2.

The quantity F2 is defined through these three equations and plays an essential role in the treatment of the second variation.

Owing to the relation 7)

L1ξ2+2M1ξη+N!η2 becomes F2w2,

and consequently,

δ2F=F1(dwdt)+F2w2+dRdt.

Article 115.
The second variation of the integral has therefore the form

8)δ2I=t0t1(F1(dwdt)+F2w2)dt+t0t1dRdtdt.

We suppose that the end-points are fixed so that at these points ξ=0=η, and we further assume that the curve subjected to variation consists of a single regular trace, along which then

R=Lξ2+2Mξη+Nη2

is everywhere continuous, so that

[R]t0t1=0.

Consequently the above integral may be written

8*)t0t1(F1(dwdt)+F2w2)dt.

If the integral I=t0t1F(x,y,x,y)dt is to be a maximum or a minimum for the curve G=0, it is necessary, when the curve is subjected to an indefinitely small variation, that the variation ΔI, which is caused to exist therefrom, have always the same sign, in whatever manner ξ,eta are chosen; and consequently the second variation δ2I must have continuously the same sign as ΔI.

We have repeatedly seen that

ΔI=ϵ22!δ2I+ϵ33!δ3I+,

and for any other value of ϵ1 for example, ϵ1,

Δ1I=ϵ122!δ2I+ϵ133!δ3I+.

If, further, δ2I is negative while ΔI is positive, then we may take ϵ1 so small that the sign of Δ1I depends only upon the first term on the right in the above expansion, and consequently is negative. Therefore the integral I cannot be a maximum or a minimum, since the variation of it is first positive and then negative.

Hence, neglecting for a moment the case when δ2I=0, we have the following theorem:

If the integral I is to be a maximum or a minimum, its second variation must be continuously negative or continuously positive.

When δ2I vanishes for all possible values of ξ, η, it is necessary also that δ2I vanish, since the integral I is to be a maximum or a minimum, and, as in the Theory of Maxima and Minima, we would then have to investigate the fourth variation. In this case the conditions that have to be satisfied are so numerous that a mathematical treatment is very complicated and difficult.

Hence, it is seen that after the condition δI=0 is satisfied, it follows that

for the possibility of a maximum, δ2I must be negative, and
for the possibility of a minimum, δ2I must be positive.

These conditions are necessary, but not sufficient.

Article 116.
In Art. 75 we assumed that ξ,η,ξ,η were continuous functions of t between the limits t0t1. Owing to the assumed existence of ξ,η, we must presuppose the existence of the second derivatives of x and y with respect to t (see Art. 23). From this it also follows that the radius of curvature must vary in a continuous manner. These assumptions have been tacitly made in the derivation of the equation 8) in the preceding article. We shall now free ourselves from the restriction that ξ and η are continuous functions of t, retaining, however, the assumptions regarding the continuity of the quantities x,y,ξ,η,x,y,x,y.

The theorem that Fx and Fy vary in a continuous manner for ox oy the whole curve (Art. 97) in most cases gives a handy means of determining the admissibility of assumptions regarding the continuity of x and y. If, at certain points of the curve G=0, x and y are not continuous, it is always possible to divide the curve into such portions that x and y are continuous throughout each portion. Yet we cannot even then say that x and y are continuous within such a portion, as has been assumed to be true in the above development. If, however, x and y within such a portion of curve are discontinuous, we have only to divide the curve into other portions so that within these new portions x and y no longer suffer any sudden springs. In each of these portions of curve the same conclusions may be made as before in the case of the whole curve, and consequently the assumption regarding the continuous change of x,y throughout the whole curve is not necessary. But if we had limited ourselves to the consideration of a part of the curve in which x,y,x,y,x,y vary in a continuous manner, the continuity of ξ, η in the integration of the integral

dRdtdt

would have been assumed. These assumptions need not necessarily be fulfilled, since the variation of the curve is an arbitrary one, and it is quite possible that such variations may be introduced, where ξ, η become discontinuous, as often as we please. We may, however, drop these assumptions without changing the final results, if only the first named conditions are satisfied. Since the quantities L, M, N depend only upon x,y,x,y,x,y, and since these quantities are continuous, it follows that the introduction of the integral dRdtdt the form given above is always admissible. For if ξ, η were not continuous for the whole trace of the curve, which has been subjected to variation, we could suppose that this curve has been divided into parts, within which the above derivatives varied in a continuous manner, and the integral would then become a sum of integrals of the form

tβtβ+1dRdtdt=[Lξ2+2Mξη+Nη2]tβtβ+1,

where tβ,tβ+1, are the coordinates of the points of division of corresponding values of t. But since ξ, η vary in a continuous manner, we have through the summation of these quantities exactly the same expression

[Lξ2+2Mξη+Nη2]t0t1

as before. The quantities ξ, η are also found under the sign of integration in the right-hand side of 8); but owing to the conception of a definite integral, we may still write it in this form even when these quantities vary in a discontinuous manner ; however, in performing the integration, we must divide the integral corresponding to the positions at which the discontinuities enter into partial integrals. Therefore, we see that the possible discontinuity of ξ, η remains without influence upon the result, if only x,y,x,y,x,y,ξ,η are continuous. Consequently any assumptions regarding the continuity of ξ, η are superfluous; however, in an arbitrarily small portion of the curve which is subjected to variation, the quantities ξ and η must not become discontinuous an infinite number of times since such variation of the curve has been necessarily, once for all excluded.

Article 117.
Following the older mathematicians, Legendre, Jacobi, etc., we may give the second variation a form in which all terms appearing under the sign of integration will have the same sign (plus or minus).

To accomplish this, we add an exact differential ddt(w2v) at under the integral sign in 8), and subtract it from R, the integral thus becoming

δ2I=t0t1(F1(dwdt)2+2vwdwdt+(F2+dvdt)w2)dt+[Rvw2]t0t!.

The expression under the sign of integration is an integral homogeneous quadratic form in w and dwdt. We choose the quantity v so that this expression becomes a perfect square; that is,

9)v2F1(F2+dvdt)=0,

and consequently,

10)δ2I=t0t1F1(dwdt+wvF1)2dt+[Rvw2]t0t1.

We shall see that it is possible to determine a function v, which is finite one-valued and continuous within the interval t0t1, and which satisfies the equation 9). The integral 10) becomes accordingly, if the end-points remain fixed,

10aδ2I=t0t1F1(dwdt+wvF1)2dt

Hence the second variation has the same sign as F1, and it is clear that for the existence of a maximum F1 must be negative, and for a minimum this function must be positive within the interval t0t1 and in case there is a maximum or a minimum, F1 cannot change sign within this interval.

This condition is due to Jacobi. Legendre had previously concluded that we have a maximum when a certain expression corresponding to F1 was negative, and a minimum when it was positive. It is questionable whether the diflferential equation for v is always integrable. Following Jacobi we shall show that such is the case.

Article 118.
Before we go farther, we have yet to prove that the transformation, which we have introduced, is allowable. In spite of the simplicity of the equation 9) we cannot make conclusions regarding the continuity of the function v, which is necessary for the above transformation. ¢ It is therefore essential to show that the equation 9) may be reduced to a system of two linear differential equations, which may be reverted into a linear differential equation of the second order, since for this equation we have definite criteria of determining whether a function which satisfies it remains finite and continuous or not.

Write

v=u1u,

where u1 and u are continuous functions of t, and u0 within the interval t0t1.

Equation 9) becomes then

<u12u2F1(F2+udu1dtu1dudtu2)=0/math>,or:<math>F1u(du1dt+F2u)u1(F1dudt+u1)=0.

Since one of the functions u, u1 may be arbitrarily chosen, we take u so that

11)F1dudt+u1=0;

then, since u0, we have

12)du1dt+F2u=0.

From 11) and 12) it follows that

12a)ddt(F1dudt)F2u=0,

or

13)F1d2udt2+dF1dtdudtF2u=0,

where F1 and F2 are to be considered as given functions of t. We shall denote this difFerential equation by J=0. After u has been determined from this Equation, u1 may be determined from 11), and from u1u=v we have v as a definite function of t.

Article 119.
The expression which has been derived for v seems to contain two arbitrary constants, while the equation 9) has only one. The two constants in the first case, however, may be replaced by one, since the general solution of 13) is

u=c1ϕ1(t)+c2ϕ2(t),


and hence from 11)

v=u1u=F1c1ϕ1(t)+c2ϕ2(t)c1ϕ1(t)+c2ϕ2(t),

an expression which depends only upon the ratio of the two constants.

It follows from the above transformation that

14)δ2I=t0t1F1(dwdtwududt)dt;

but this transformation has a meaning only when it is possible to find a function u within the interval t0t1 which is different from zero, and which satisfies the differential equation J=0.

Article 120.
If we have a linear differential equation of the second order

d2ydx2=P(x)dydx+Q(x)y=0,

and if y1 and y2 are a fundamental system of integrals of this equation, then we have the well known relation due to Abel (see Forsyth's Differential Equations, p. 99)

y1dy2dxy2dy1dx=CeP(x)dx,

or

Δ=|y1y2dy1dxdy2dx|=CeP(x)dx.

If Δ=0, then we would have y1=cy2, and the system is no longer a fundamental system of integrals. This determinant can become zero only at such positions for which P(x) becomes infinitely large; or a change of sign for this determinant can enter only at such positions where P(x) becomes infinite.

In the differential equationy J=0 we have P=ddtln(F1), and if u1, u2 form a fundamental system of integrals of this differential equation, then

Δ=u1du2dtu2du1dt=CF1.

It follows that F1 cannot become infinite or zero within the interval under consideration or upon the boundaries of this interval. Hence, it is again seen that F1 cannot change sign within the interval t0t1.

If F1 and F2 are continuous within the interval t0t1, we have, through differentiating the equation J=0, all higher derivatives of u expressed in terms of u and dudt. Hence, if values of u and dudt are given for a definite value of t, say t, we have a power-series P(tt) for u (see Art. 79), which satisfies the equation J=0.

Article 121.
Suppose that F1 has a definite, positive or negative value for a definite value t of t situated within the interval t0t1, then on account of its continuity it will also be positive or negative for a certain neighborhood of t, say tτ1t+τ2. We may vary the curve in such a manner that within the interval tτ1t+τ2 it takes any form while without this region it remains unchanged.

Consequently the total variation, and therefore also the second variation of I, depends only upon the variation within the region just mentioned, and in accordance with the remarks made above since we may find a function u of the variable t, which is continuous within the given region, which satisfies the differential equation J=0, and which is of such a nature that u and dudt have given at values for t=t, it follows that the transformation which was introduced is admissible, and we have

δ2I=t0t1F1(dwdtdudtwu)2dt.

This quantity is evidently positive when F1 is positive and negative when F1 is negative, so long as

dwdtdudtwu0 (Art. 132).

We have then for the total variation

ΔI=ϵ22!t0t1F1(dwdtdudtwu)2dt+ϵ33!t0t1(ξ,η,ξ,η)dt,

where (ξ,η,ξ,η) denotes an expression of the third dimension in the quantities included within the brackets.

For small values of ϵ it is seen that ΔI has the same sign as the first term on the right-hand side of the above equation. We have, therefore, the following theorem :

The total variation ΔI the integral I is positive when F1 is positive, and negative when F1 is negative throughout the whole interval t0t1.

If F1 could change sign for any position within the interval t0t1, then there would be variations of the curve for which ΔI is positive and others for which ΔI is negative. Hence, for the existence of a maximum or a minimum of I we have the following necessary condition :

In order that there exist a maximum or a minimum of the integral I taken over the curve G=0 within the interval t0t1, it is necessary that F1 have always the same sign within this interval; in the case of a maximum F1 must be continuously negative, and in the case of a minimum this function must be continuously positive.

In this connection it is interesting to note a paper by Prof. W. F. Osgood in the Transactions of the American Mathematical Society, Vol. II, p. 273, entitled:

"On a fundamental property of a minimum in the Calculus of Variations and the proof of a theorem of Weierstrass's."

This paper, which is of great importance, may be much simplified.

Template:BookCat