Discrete Systems and Integrability
Discrete Systems and Integrability
MATH3491/5492:
DISCRETE SYSTEMS AND INTEGRABILITY
Frank W NIJHOFF
University of Leeds, School of Mathematics
0
Aknowledgement:
These lecture notes are an adaptation of a text which is intended as a monograph by J.
Hietarinta, N. Joshi and F.W. Nijhoff on the same subject.
Preface
These Lectures deal with discrete systems, i.e. models of (physical, biological, economic,
etc.) phenomena that are mathematically modelled through equations that involve finite
(as opposed to infinitesimal ) operations.
When modelling continuous phenomena in nature we are often using differential equations
as the mathematical tools to encode the essence of what is going on in the system we are
studying. The differential equations are believed to encapture the main mechanisms behind
the phenomena, based on either microscopic theory or general principles relevant to the
subject. Whatever the source of the model, solving the basic differential equations is then
believed to lead to predictions on the behaviour of the system.
In many circumstances, the phenomena of nature are most adequately described not
in terms of continuous but in terms of discontinuous steps (for instance many models in
biology, economy, neural processes, decision processes, etc.). In that case we are dealing
with discrete-time systems, which can mathematically be descrided through:
• recurrence relations;
• difference equations;
• dynamical mappings.
In addition to these types of systems we will also consider:
• functional relations
which are of a slightly different character, but will also arise in connection to discrete systems.
All of the above will form the main mathematical tools in these Lectures.
To make a comparison with the continuous theory, let us recall the types of differential
equations we know. These fall into two classes:
1. Ordinary differential equations (ODEs), which are of the form:
in which F is some expression (a function of its arguments) and where x is called the
independent variable, whilst y(x) is the dependent variable since it depends on x. The
primes denote differentiation w.r.t. x:
dy d2 y
y ′ (x) = , , y ′′ (x) = ,...
dx dx2
1
2
and the aim is to try and solve y as a function of x, wherever possible, from (0.0.1).
The solution, if we can find it at all, is not unique, and needs further data in order to
render it unique. For instance, this can be done by imposing, in addition to the ODE
itself, a number of initial data, e.g. by fixing values of y and some of its derivatives at
a given value of x, say at x = 0, namely:
y(0) , y ′ (0) , . . .
2. Partial differential equations (PDEs), in which case we have more than one independent
variable. For instance, if we have two independent variables, say x and t, a PDE takes
on the general form:
where F denotes again some expression, and we have in its argument the function
y(x, t), the dependent variable depending on both independenet variables, and its
partial derivatives
phenomena under study. However, in view of eq. (0.0.3) this may be considered as a logical
reversal: since we get the derivative (0.0.3) by the reductive procedure of a limit from the
finite operation (0.0.4) of a difference, shouldn’t we consider the latter as being the more
fudamental?
In modern physics, this question gains even more weight. The underlying structure of
subatomic physics, based on quantum mechanics, leaves no doubt that there is a fundamental
width (given in terms of Planck’s constant ~) beyond which a “continuum” is no longer
perceptible, and even may make no longer sense. Thus, maybe the continuum aspect of
Nature is an illusion, only of validity on the scale of the macroscopic world. Rather than
considering the finite-difference to be the approximation, one actually should consider the
derivative to be the approximation (or simplification) of a more fundamental set of tools. A
famous citation by A. Einstein1 declares:
“To be sure, it has been pointed out that the introduction of a space-time con-
tinuum may be considered as contrary to nature in view of the molecular
structure of everything which happens on a small scale. It is maintained that
perhaps the success of the Heisenberg method points to a purely algebraic de-
scription of nature, that is to the elimination of continuum functions from
physics. Then, however, we must also give up, by principle, the space-time
continuum. It is not unimaginable that human ingenuity will some day find
methods which will make it possible to proceed along such a path. At the
present time, however, such a program looks like an attempt to breathe in
empty space.”
Why then haven’t we then progressed further in the theory of difference equations? It
is fair to say that in contrast to the theory of (ordinary or partial) differential equations,
the theory of difference equations is even today, in the beginning of the 21st century, still
in its infancy. One reason for this discrepancy is that in spite of the simple appearance of
the basic difference operation (0.0.4) the development of comprehensive theory has proved
to be far more difficult. Many more technical difficulties are met in the development of
difference calculus than in differential calculus, mainly due to the nonlocal nature of the
difference operator. Furthermore, it is much more difficult to classify difference equations
than is the case for differential equations: there are just too many possibilities of writing
down difference equations, whereas for differential equations one can easily distinguish them
by their order, degree, well-posedness, or amenability to boundary and initial value problems,
etc.
But there are also historical reasons. Up to the beginning of the 20th century, mathemati-
cians considered the theory of differential equations and the theory of difference equations
very much as two aspects of one and the same endeavor. It is mostly after the two world
wars that a generation of mathematicians working systematically on difference equations
(the famous school of G.D. Birkhoff) disappeared.
It is also important to recall the role of physics as an inspiration in the development of
mathematics, from Newton on. At the beginning of 20th century there were two interesting
developments in physics, which also caught the interest of mathematicians: general relativity
1 A. Einstein, Physics and Reality, 1936. (Reprinted in Essays in Physics, Philosophocal Library, New
York, 1950)
4
and quantum theory. Both relied on “new” mathematics, and attention was drawn on
developing these tools further. It was only in the 1960’s that attention returned on classical
dynamics, on discrete dynamics and the associated chaos on one hand, and on PDE’s and
the associated soliton equations on the other. Major progress was then made on both of
these areas in the 1970’s and 1980’s so that both have now developed into paradigms with
associated mature theory.
What has fundamentally changed in the last two decades is the surge of interest in inte-
grable difference equations (see, e.g, http://www.side-conferences.net). Integrability is
strongly associated with regularity of the behavior, but this regularity is not due to triviality,
but rather due to some underlying mathematical structure that restricts the dynamics in a
nontrivial way. This has already been well established in the theory of soliton equations.
The theory of integrable difference equations (both ordinary and partial) has undergone a
revolution since the early 1980’s, and has also achieved major developments, although many
problems are still open. In fact, the various notions of integrability (of which more in the
following) have proven to be a very strong guiding principle which has driven the subject
forward. The subjects which it has strongly affected include the following
• ordinary difference equations (O∆Es);
• partial difference equations (P∆Es);
• integrable discrete-time systems (mappings);
• discrete and “difference” geometry;
• linear and nonlinear special functions, orthogonal polynomials;
• representation theory of (infinite-dimensional and quantum) algebras & groups;
• combinatorics and graph theory.
Some of these developments go back to results from the 19th century and even earlier,
but it is fair to say that in combining classical results (from Leibniz, Bernoulli, Gauss,
Lagrange, etc.), via the turn of the 19/20th century (Poincaré, Birkhoff, Nörlund, etc.)
with the results of the modern age, the study of integrability in discrete systems forms at
the present time the most promising route towards a general theory of difference equations
and discrete systems.
One of the amazing facts is that in the study of integrable difference equations, we will
encounter all the above type of equations, but moreover these types turn out to be quite
interconnected. In fact, we will see that in many examples we can interpret one and the same
system either as a dynamical map or finite difference equation, or as an analytic difference
equation, or even as functional equation, without damaging the main integrability aspect of
the system.
First, as with differential equations, they divide up between ordinary difference equations
(O∆Es) and partial difference equations (P∆Es), depending on whether there is one or more
than one independent variable. The independent variable is discrete, but now this can have
different interpretations:
a) the independent variable can take on only integer values, so we denote it by n ∈ Z, , and
thus the dependent variable y can be denoted by yn ;
b) the independent variable x can take on all values (real or perhaps even complex), but
in the equation the dependent variable is evaluated at integer shifts of the indepen-
dent variable. Thus, the difference equation will involve evaluations of the dependent
variable y(x) at shifts of x by multiples of a fixed quantity h:
y(x) , y(x + h) , y(x + 2h) , . . . x, h ∈ R .
Remark: These two cases are quite different in nature and the associated problems are
quite distinct. However, case a) can be viewed as specialisation of case b), by setting
yn ≡ y(x0 + nh)
and by fixing x0 as a starting value, we can consider the values of the idependent variable
yn as a stroboscope of the second case.
Remark: Before describing various types of difference equations, let us first make the
obvious observation that prior to the the difference operator (0.0.4) there is an even more
basic operation, namely that of the (lattice) shift. This is simply the operation of evaluating
the dependent variable at a shifted value of the independent variable, and it is useful to
introduce an operator associated with this operation:
Th y(x) = y(x + h) , (0.0.5)
which we will use from time to time. It is easy to see that the difference operator and its
higher orders can be simply expressed in terms of the shifts Th :
1
∆h y(x) =(Th − id)y(x) ,
h
1
∆2h y(x) = T 2 − 2Th + id)y(x) ,
h2 h
... ...
n
n (−1)n X n
∆h y(x) = (−Th )n−j y(x)
hn j=0 j
and thus the nth order difference operator acting on y(x) can be expressed in terms of the
shifted variables y(x), y(x + h), . . . , y(x + nh) . Thus an equation of the form:
F(y(x), ∆h y(x), ∆2h y(x), . . . ; x) = 0
can be rewritten in the form:
F̄(y(x), y(x + h), y(x + 2h), . . . ; x) = 0 ,
6
where the expression F̄ can be straightforwardly be obtained from F by using the above
mentioned substitutions.
←− h −→ → x
Taking into account this remark, we can distinguish the following types of discrete systems:
a) Ordinary finite difference equations: by which we mean recurrence relations of the form
where k and l are fixed integers (see Figure) . We can think of (0.0.6) as an iterative
system where we wish to solve yn at discrete points only. Thus, giving initial values
at a sufficient number of points, generically y0 , y1 , . . . , yk+l , we can hope to iterate
the equation and calculate step-by-step yk+l+1 and subsequent values. The equation
may depend explicitely on n, in which case the equation is nonautonomous, but if
F depends only on n through the dependent variables then the equation (0.0.6) is
autonomous. The order of the equation is generically given by k + l, i.e. the number of
initial data required to achieve a well-defined iteration process. We will sometime refer
to an equation of the form (0.0.6) as an (k + l + 1)-point map (assuming all “points”
at which y is evaluated appear in the expression F).
Finite difference equations of this type can also be viewed as a dynamical mapping
as follows. Assuming that we can solve yn+l uniquely from (0.0.6), leading to an
expression of the form
yn+l = Fn (yn−k , . . . , yn+l−1 ) ,
and introducing the k + l-component vector
y n = (yn−k−l , . . . , yn−1 )t
yn 7→ e n = y n+1 = F n (y n ) ,
y (0.0.7)
where now, even though the equation only involves integer shifts (by an increment h) in
the arument of the dependent variable y(x), the independent variable x is meant to be
a continuous variable, and we would like to solve (0.0.10) for functions y(x) in some
appropriate class of functions. Clearly, there is indeterminacy in this problem: the
7
solutions cannot be determined fully without additional information, and the solution
is determined only up to a shift over h. This implies that the initial data have to
be given over an entire interval on the real line, or alternatively that the role of
“integration constants” is played by periodic functions obeying:
π(x + h) = π(x) .
which is supposed to hold for arbitrary values of the arguments of the functions f , g,
etc. Under general assumptions on these functions (such as continuity, differentiability,
etc.), the imposition that the eq. (0.0.9) should hold for arbitrary values for x, y, . . . , is
often sufficient to almost uniquely fix the functions that solve the functional equation.
The notion of initial values is irrelevant in this context.
Under the assumption that f is differentiable for all real values of its argument, and setting
f ′ (0) = α 6= 0, show that the unique solution is given by f (x) = eαx .
Under the assumption that f , g are both differentiable for all real values of their argu-
ments, and that f is an even function of its argument, determine the general solution of the
functional equation.
c) Partial difference equations: in this case we have more than one independent variable
x, t ∈ R or in the finite difference case n, m ∈ Z. The equations can take the form:
The dependent variable y(x, t) and yn,m respectively depend here on two (or more)
independent variables with discrete shifts. The second case can be reduced to the first
case by setting
yn,m = y(x0 + nδ, t0 + mε) ,
but obviously the two types of problems are quite different in nature, since in the
analytic case we want to solve for y(x, t) as a function of a continuous range of values of
its arguments. The theory of the latter type of equations is still very under-developed.
8
δ −→ n
m
↓
One of the amazing facts is that in the study of integrable difference equations, we
will encounter all these type of equations, but moreover these types turn out to be quite
interconnected. In fact, we will see that in many examnples we can interpret one and
the same system either as a dynamical map or finite difference equation, or as an analytic
difference equation, or even as functional equation, without damaging the main integrability
aspect of the system. In the latter interpretation we are often dealing with solutions in terms
of (nonlinear discrete) special functions.
The subject is also related to a new branch of geometry, called difference geometry, which
is the discrete analogue of the classical differential geometry of curves and surfaces. It turns
out that integrable P∆Es describe discrete quadrilateral surfaces like the one depicted in
the figure below:
The development of this new theory has been one of the more intriguing applications of the
subject.
9
General Literature:
1. G.E. Andrews, R. Askey and R. Roy, Special Functions, (Cambridge Univ. Press,
1999).
2. C.M. Bender and S.A. Orszag, Advanced Mathematical Methods for Scientists and
Engineers, (McGraw-Hill, 1978).
3. P.R. Garabedian, Partial Differential Equations, (Chelsea, 1964).
4. F.B. Hildebrand, Introduction to Numerical Analysis, (McGraw-Hill, 1956; Dover,
1987).
5. E.L. Ince, Ordinary Differential Equations, (Dover, 1956)
6. K. Iwasaki, H. Kimura, S. Shimomura, M. Yoshida, From Gauss to Painlevé, (Vieweg,
1991)
7. L.M. Milne-Thompson, The Calculus of Finite Differences, (AMS Chelsea Publ. 2000),
first edition, 1933.
8. N.E. Nörlund, Vorlesungen über Differenzenrechnung, (Chelsea, 1954), first edition ,
1923.
9. E.T. Whittaker and G.N. Watson, A Course of Modern Analysis, (Cambridge Univ.
Press, 1927 (4th ed.))
10. R. Sauer, Differenzengeometrie, (Springer Verlag, 1970).
11. A.S. Bobenko and Yu. Suris,Discrete Differential Geometry; Integrable structure,
Grad. Studies in Math. vol 98, AMS Publs. 2008, (a preliminary version can be
found on: http:arxiv.org/math/abs/0504358).
12. J. Hietarinta, N. Joshi and F.W. Nijhoff, Discrete Systems and Integrability, (mono-
graph in preparation).
Lecture 1
Elementary Theory of
Difference Equations
In this lecture we will give a brief account some of the basic tools to deal with (ordinary)
difference equations and systems based on discrete operations. Most of the machinery is
rather elementary, as we will concentrate here primarily on linear finite-difference equations,
although some simple nonlinear equations will be dealt with as well. Some of this material
is needed in subsequent lectures dealing with integrable systems. Some analogies with and
distinctions from the theory of differential equations will be mentioned. Probably the best
text for further reading is the classic monograph by L.M.Milne-Thompson, or the relevant
chapters 2 and 5 in the famous book by C.M. Bender & S.A. Orszag. However, there
are many elementary textbooks on the basic theory of finite-difference equations, e.g. the
text by W.G. Kelley & A.C. Peterson. For partial finite-difference equations, material can
be found in the monographs by P.R. Garabedian (chapter 3 and 13) and in the book by
F.B. Hildebrand. These latter texts are motivated by the problem of numerical simulations
through finite-difference equations.
1
∆h y(x) = (Th − id) y(x) , (1.1.1)
h
pointing out that it essentially builds on the shift operator
11
12 LECTURE 1. ELEMENTARY THEORY OF DIFFERENCE EQUATIONS
It is tempting to expect that one could build a theory based on the difference operator
(1.1.1) in a fairly similar way as for the differential operator d/dx , but there are impor-
tant distinctions. Recalling, from elementary calculus, some of the main properties of the
differential operator, they can be summarised as follows:
linearity: for any two functions y(x), w(x) and constants α, β we have:
d dy dw
(αy(x) + βw(x)) = α +β
dx dx dx
product rule: For the product of any two fuctions y(x), w(x) we have the Leibniz rule:
d dy dw
(y(x)w(x)) = w(x) + y(x)
dx dx dx
composition rule: for two functions z(y) and y(x) and forming the composed function
Z(x) = z ◦ y(x) = z(y(x)) by substitution we have the chain rule
dZ
= z ′ (y(x))y ′ (x)
dx
where z ′ (y) = dz/dy , y ′ (x) = dy/dx .
For the difference operator, however, the last rule really stops making sense (at least without
making many further assumptions), whilst the product rule takes a subtly differen form,
namely by computing
1
∆h (y(x)w(x)) = [y(x + h)w(x + h) − y(x)w(x)]
h
1 1
= y(x + h)[w(x + h) − w(x)] − [y(x + h) − y(x)]w(x)
h h
= y(x + h)∆h (x) + (∆h y(x))w(x) ,
which looks similar to the differential product rule, but has one crucial extra element: it
involves necessarily the shift operator Th as well!. Obviously we could write (1.1.4) as well
in the form
∆h (yw) = y∆h w + (∆h y)Th w , (1.1.4)
but in either way the shift operator Th cannot be avoided. Thus, the difference calculus
cannot be built with the difference operator ∆h alone, and we need the shift operator Th as
well, which by itself has the crucial product property:
In fact, a ”difference” calculus can be built on the shift operator alone, but we have to realise
that the latter is essentially a nonlocal operator: it connects two values of the argument,
1.1. ELEMENTARY DIFFERENCE CALCULUS 13
It turns out that the corresponding function can be given in terms of the so-called Pochham-
mer symbol (a)k defined by1
(a)k = a (a − 1) (a − 2) . . . (a − k + 1) , for k > 0, (a)0 ≡ 1 , (1.1.10)
in which k is an integer, but where a could still take on any (real) value. This is also called
the falling factorial2 and it is sometimes denoted by a(k) .
Performing the following computation:
∆(n)k = (n + 1)k − (n)k
= [(n + 1)n(n − 1) · · · (n − k + 2)] − [n(n − 1) · · · (n − k + 2)(n − k + 1)]
= n(n − 1) · · · (n − k + 2)[(n + 1) − (n − k + 1)] = k[n(n − 1) · · · (n − k + 2)]
= k (n)k−1 ,
which shows that f (n; k) = (n)k is a solution of the difference equation (1.1.9) establishing
the analogy with the power function.
Exercise 1.1.4. In the case that k is a negative integer we can introduce the symbol:
(a)k = 1/ ((a + 1)(a + 2) · · · (a + |k|)) k<0, (1.1.11)
(k)
(also denoted as a ). Show that the function f (n; k) = (n)k is a solution of the difference
equation (1.1.9) for the power function also for negative k.
Since ∆ is a linear operator, we can get an analogue to any function defined by a power
series solution, simply by considering linear combinations of power functions of different
degree (the degree being teh variable k). For example, the analogue of the exponential
function exp(ax) , would be
∞
X ak (n)k
e(n; a) = = (1 + a)n . (1.1.12)
k!
k=0
Exercise 1.1.5.
i) Show that the discrete exponential function e(n; a) obeys the difference equation
∆e(n; a) = a e(n; a) ;
ii) show that the second equality in (1.1.12) holds, and that as a consequence we have the
relation
e(n; a)e(m; a) = e(n + m; a) .
1 Note that a does not need to be an integer for this definition to make sense. In fact, one can extend the
Note that the upper index in the summation is n − 1, since ∆ is defined by the forward
differentiation. In both cases the inverse is not unique, but is determined up to a constant:
an integration constant w.r.t. the variable x in the derivative case, and a constant w.r.t. the
variable n in the difference case. In the indefinite integration/summation the constant is
left unspecified, while in the definite integration/summation this constant can be fixed by
specifying the lower limit in the integral/sum, e.g.,
Z x n−1
X
F (x) = f (x) dx and Fn = fj .
0 j=0
This lower limit is purely a matter of choice and the choice depends on the problem at
hand (e.g., initial values when we deal with differential or difference equations). We can also
define the summation so that n is as the lower index:
M
X
Gn := fj , ∆Gn = −fn .
j=n
Exercise 1.1.6. Give the antiderivative of the function f (x) = xk , and the antidifference
of the function f (n; k) = (n)k for both k a positive and negative integer.
The usual rules of linearity apply to the antidifferential:
X X X
(αxn + βyn ) = α xn + β yn . (1.1.13)
The summation symbol can also be used to define a generating function. Let
∞
X ∞
X zk
F (z) := fk z k , or F̄ (z) := fk ,
k!
k=0 k=0
1.1. ELEMENTARY DIFFERENCE CALCULUS 17
This can be useful if the summation can be done to obtain a closed form expression.
f (qx) − f (x)
Tq f (x) = f (qx) , Dq f (x) = (1.1.15)
(q − 1)x
but one should be warned that introducing logarithms may affect the analytic nature of the
solutions in a dramatic way.
The power function coincides here with the continuous power function, since
qk − 1
Dq xk = (k)q xk−1 with (k)q = (1.1.16)
q−1
the (k)q being called q-integers. As a consequence, the so-called q-exponential function can
be defined by
X∞
xk
expq (x) = , (k)q ! = (k)q (k − 1)q . . . (1)q (1.1.17)
(k)q !
k=0
which are the q-analogues of the Pochammer symbols. In terms of these symbols the q-
exponential can also be defined as eq (x) = 1/(x; q)∞ , since we have the q-binomial identity:
X xk ∞
1 x
= = expq . (1.1.19)
(x; q)∞ (q; q)k 1−q
k=0
Exercise 1.1.7. Prove the q-binomial identity P∞(1.1.19) by deriving the difference equation
eq (qx) = (1 − x)eq (x) , and setting eq (x) = k=0 ck xk derive a recurrence relation for the
coefficients ck , whilst c0 = 1.
18 LECTURE 1. ELEMENTARY THEORY OF DIFFERENCE EQUATIONS
The inverse of the q-difference operator Dq is given by the so-called Jackson integral,
defined by
Z x ∞
X
f (x)dq x =: (1 − q)x q k f (q k x) (1.1.20)
0 k=0
Exercise 1.1.8. Show that the analogue of the fundamental theorem of integration holds for
the Jackson integral, i.e., that we have
Z x Z x
Dq f (x)dq x = (Dq f )(x)dq x = f (x)
0 0
This would correspond to the simplest linear inhomogeneous case. The next class of differ-
ence equations are more general linear difference equations. We know that linear differential
equation are far easier to treat, in general, than nonlinear differential equations, because the
linearity allows the superposition of solutions (at least in the homogeneous case). Similarly,
linear difference equation have the same property and, hence, are expected to be simpler
than nonlinear equations. However, both for differential as well as difference equations, one
has to distinguish cases:
• linear homogeneous with constant coefficients;
• linear inhomogeneous with constant coefficients
(but possibly with a nonconstant inhomogeneous term);
• linear with nonconstant coefficients.
In the first case the difference equation is autonomous, i.e., the independent variable enters in
the equation only through the dependent variable. For such equations elementary methods
exist to solve them even in explicit form. For the non-constant coefficient case, the situation
is much more complicated, and often solutions can only be found through power series, and
the latter are subject to subtle considerations of singular points of the equation where the
coefficients may blow up. In this lecture we will mostly restrict ourselves to considering the
constant coefficient case, but in the next lecture we will come back to the case of non-constant
coefficients.
general solution. hence, we need one initial value to fully specify the solution. An equation
of the form y ′′ = F (y, y ′ , x) would be second order, because in general we would need
two initial data to specify the solution fully, e.g. y(x0 ), y ′ (x0 ) at a given value x0 of the
independent variable. Similar considerations apply to finite-difference equations as well: the
order of such an equation would by definition equate the number of initial values needed to
specify fully the solution.
Example: A 2-point equation of the form ∆yn+1 = F (yn ) would be generically of first order,
since we could start from a given value y0 and iterate the solution as a map, thereby fixing the
solution entirely. However, it is not true that for all functions F defining the equation we would have
a first order difference equation, because if we take F (y) := c−y, then the difference equation would
reduce to: yn+1 = c , c being a constant, and the solution of this equation is simply yn ≡ c, ∀n.
The latter would effectively be a 1-point equation, and hence an equation of zeroth order (no initial
values needed to specify the solution, taking into account that c is a constant of the equation, and
hence given a priori). Similar arguments apply to a difference equation of the form yn+1 = G(yn ) ,
which is obviusly connected to the previous form by setting G(y) = F (y) + y.
Example: A 3-point equation of the form yn+1 + yn−1 = F (yn ) would be generically of second
order, since we would for generic F need two initial date, say y−1 , y0 to iterate the equation as a
map:
(yn−1 , yn ) 7→ (yn , yn+1 ) = (yn , −yn−1 + F (yn ))
However, it is not alway right to consider the digits in the shifts of a difference equation as an
indication of the order. Thus, the equation ∆2 yn = F (yn+1 , yn ) does not always need to be of
second order, even though we have a second degree difference operator involved. In fact, if we take
F (e
y, y) = f (y) − 2e
y , then the equation reduces to a 2-point form, and the solution would be of first
order, albeit on a sublattice of integers of either even or odd labeled points.
The general case of first order, constant-coefficient, linear finite-difference equations is not
very exciting. Such equations can be written in the form ayn+1 +byn = 0 , (a, b constant) in
the homogeneous case, leading to the solution yn = (−b/a)n y0 , and in the inhomogeneous
case (with given function fn ) we can follow the procedure of the example below.
The solution of the homogeneous equation (replacing fn by 0) is given above. By a principle which
is called variation of constants, we set yn = ((−b/a)n wn (i.e. replacing the constant y0 by a
function wn of n yet to be determined). Inserting this into the equation yields:
n
b b 1 a n
− a − wn+1 + bwn = fn ⇒ wn+1 − wn = − − fn ,
a a b b
1 X a j
n−1
wn = w0 − − fj , w0 (constant) intial value .
b j=0 b
20 LECTURE 1. ELEMENTARY THEORY OF DIFFERENCE EQUATIONS
Example: We will now consider the general case of second order constant-coefficient, linear
finite-difference equations, the general form of the equation being:
This requires solving the corresponding characteristic equation associated with this difference
equation, which is obtained by exploring a simple trial solution of the form yn ∼ λn , where
λ is to be determined. Inserting this into the euqation we get the characteristic equation:
when λ1 6= λ2 .
3 Note that these solutions are the consequence of the superposition principle: in the case of distinct
roots λn n
1 and λ2 are two independent solutions of the homogeneous equation, and the general solution is an
arbitrary linear combination of these two. In the case of coinciding roots λ1 = λ2 the independent solutions
are λn n
1 and nλ1 and again a linear combination of these two solutions leads to the general solution. The
independent solutions in the latter case can be seen as arising from a limiting procedure of the former case:
λn n
2 −λ1
replacing the independent solutions λn n n
1 , λ2 w.l.o.g. by λ1 and λ2 −λ1 , (note that the latter is itself a linear
combination of the two orginal independent solutions), and taking the limit that the two roots coincide,
i.e. λ2 = λ1 + ǫ, with ǫ → 0, the second solution goes in this limit over into nλn−1 1 , which is the other
independent solution (up to a constant factor) in the coinciding roots case.
1.2. LINEAR (FINITE-)DIFFERENCE EQUATIONS 21
Exercise 1.2.1. Verify that the formula (1.2.22) yields a particular solution of (1.2.21).
Give a similar formula for the case of coinciding roots λ1 = λ2 by setting λ2 = λ1 + ǫ, and
taking the limit ǫ → 0.
3. The general solution of the inhomogeneous equation is now given by the combination yn =
ynhom + ynpart of the general solution of the homogeneous equation (as given in step # 1) and
the particular solution of the inhomogeneous equation (as given in step # 2). The solution
depends on integration constants c1 , c2 , which can subsequently be determined (but only
once we have the full solution of the inhomogeneous problem) by imposing initial conditions,
e.g. by giving y0 and y1 .
F0 (1 − z) + F1 z 1
F (z) = =
1 − z − z2 1 − z − z2
and expanding the r.h.s. w.r.t. powers of z yields the Fibonacci numbers as coefficients.
yn+1 − gn yn = hn , (1.2.23)
where gn (6= 0), hn are some given functions. As before, the complete solution of (1.2.23)
is a sum of the general solution of the associated homogeneous problem and a particular
solution of the full inhomogeneous equation.
It is easy to see, that the homogeneous version of (1.2.23), namely
wn+1 − gn wn = 0 , (1.2.24)
22 LECTURE 1. ELEMENTARY THEORY OF DIFFERENCE EQUATIONS
can be solved by
n−1
Y
wn = w0 gj , (1.2.25)
j=0
Since the left-hand side is now of the form ∆zn , the equation can be simply integrated with
as solution
n−1
X
zn := hk /wk+1 , (1.2.27)
k=k0
where k0 is an arbitrary lower limit of the sum. Thus the general solution of (1.2.23) is
n−1
! n−1
X hk Y
yn := C+ Qk gj . (1.2.28)
k=k0 j=j0 gj j=j0
Here C is the “summation constant”, associated with the homogeneous solution, its value
be determined by an initial condition. This formula (1.2.28) is really only a formal solution,
and in practice we would like to express the sums and products in closed Q form, but
P there
are no general methods for doing that. Sometimes it is useful to note that gj = e log gj .
a(N ) (N −1)
n yn+N + an yn+N −1 + · · · a(1) (0)
n yn+1 + an yn = fn (1.2.29)
(j)
where an , fn are some given functions of n, and where we assume that the product aN 0
n an is
(j)
nonvanishing. In the case of constant coefficients (i.e., when all an = a(j) do not depend on
the independent variable n) the homogeneous problem reduces to solving the corresponding
characteristic equation:
leading to roots λ1 , . . . , λN (some of which may coincide). If all roots are mutually distinct,
the solution would take the form
N
X
yn = cj λnj , cj constant coefficients
j=1
but if some roots coincide we will have to supply different independent solution. In fact, if
a root λj has multiplicity µj (i.e. µj is the power of the factor (λ − λj ) in the factorisation
of the polynomial given by the characteristic equation), then the corresponding linearly
independent solutions are λnj , nλnj , . . . , nµj −1 λnj . The general solution of (1.2.29) in the
inhomogeneous case is given by a linear combination of N linearly independent general
solutions of the homogeneous part plus a particular solution of the inhomogeneous equation.
(j)
In the case of non-constant coefficients (i.e., when the coefficients (1.2.29) an are non-
trivial functions of n) the situation is markedly more complicated, and to find basic solutions
one needs to resort to developing power series in the same way as is done in the Fuchsian
theory of differential equations. Such power series solutions often depend crucially on the
occurrence of singularities of the coefficients. We will refrain from treating this theory at
this juncture. The only issue we want to mention at this point is the nontrivial matter of
linear independency of solutions of the difference equation. In the continuous case of linear
differential equations the linear independence of solutions is expressed through the non-
vanishing of a particular determinant associated with the differential equation, the so-called
Wronski determinant or Wronskian (see Example sheet # 1). In the case of linear differ-
ence equations the role of the Wronskian is replaced by the so-called Casorati determinant
or Casoratian. This determinant is constructed as follows. Suppose that we have N discrete
(j)
functions yn , j = 1, . . . , N , then the Casoratian of these functions is given by
(1) (2) (N )
yn yn ... yn
(1) (2) (N )
yn+1 yn+1 ... yn+1
C(yn(1) , yn(2) , . . . , yn(N ) ) = .. .. .. .. , (1.2.30)
. . . .
(1) (2) (N )
yn+N −1 yn+N −1 ... yn+N −1
and these functions are independent if and only if their Casoratian is nonzero.
Exercise 1.2.2. In the homogeneous case of (1.2.29), i.e., setting fn ≡ 0, show that the
n-dependence of the Casoration is given by:
(0)
(1) (2) N) an
C(yn+1 , yn+1 , . . . , yn+1 ) = (−1)N (N )
C(yn(1) , yn(2) , . . . , ynN ) )
an
(0) (N )
which, provided an /an 6= 0 , guarantees that the indpendence of the solutions is preserved
as n progresses.
24 LECTURE 1. ELEMENTARY THEORY OF DIFFERENCE EQUATIONS
and furthermore
[N ] 1
yn[N ] =: yn+N −1 ⇒ yn+1 = yn+N = (N )
fn − a(N
n
−1) [N ]
yn − · · · − a(1) [2] (0) [1]
n y n − an y n .
an
T
[1] [2] [N ]
Introducing the n-component vector y n = yn , yn , . . . , yn we can now write the
system of equations above, in vector form as follows:
y n+1 = An y n + f n , (1.2.31)
The general solution of the matrix difference equation (1.2.31) is very easily obtained, once
we assume that we know a fundamental matrix solution Y n of the homogeneous matrix
equation, i.e., a N × N matrix solution of the equation
Note that any matrix Ye n obtained from Y n by matrix multiplication from the right by an
invertible constant matrix C, i.e.,
Y n → Ye n = Y n C
is again a fundamental matrix solution of the same homogeneous matrix difference equation.
Whether or not the sum can be computed in closed form depends on the particular nature
of the coefficients in the equation and whether or not we can find a reasonable expression
for the fundamental matrix solution Y n .
The variation of constants approach now amounts to finding a solution of the inhomo-
geneous vector equation (1.2.31) of the form y n = Y n cn where the N -component vector
cn remains to be determined as a function of n. Plugging this form of y n into the inhomo-
geneous equation we obtain:
Y n+1 cn+1 = An Y n cn+1 = An Y n cn + f n
⇒ cn+1 − cn = Y −1
n+1 f n
n−1
X
⇒ cn = c0 + Y −1
j+1 f j
j=0
n−1
X
⇒ y n = Y n cn = Y n c0 + Y n Y −1 −1
j Aj f j .
j=0
Thus, once Y n is known, the last formula provides the solution of the vector difference
equation, and this is also the general solution, because the first term contains an arbitrary
constant N -component vector c0 which corresponds to the N integration constants of the
equation.
In the case of the N th order difference equation (1.2.29) the vector y n can be written as
y n = (yn , yn+1 , . . . , yn+N −1 )T
(1) (N )
and if we know N independent solutions of that equation yn , . . . , yn the matrix Y n is
given by
(1) (2) (N )
yn yn ··· yn
(1) (2) (N )
yn+1 yn+1 ··· yn+1
Yn= .. .. ..
. . .
(1) (2) (N )
yn+N −1 yn+N −1 · · · yn+N −1
from which it is clear that the determinant det(Y n ) can be identified with the Casoratian of
the N independent solutions. The explicit case of N = 2 of this construction will be studied
in the Examples # 1 sheet.
2n2
yn+1 + yn−1 = yn
n2−1
3. Reduction of order: With a known solution of the homogeneous equation the order
of the equation can be reduced.
[1]
where ai , b may depend on n. Suppose yn is a solutions of the homogeneous part. Then substituting
[1]
yn = yn zn yields
the linear difference equation can be often converted into an algebraic equation for Y (z) (in
the case of constant coefficients), or into an ODE (for coefficients depending polynomialy
on the independent discrete variable n). Examples are given on the Example sheet # 1.
Exercise 1.3.1. In the case that the coefficients a,b,c in (1.3.33) are constants, and denoting
by y1 , y2 the roots of the quadratic ay 2 + by + c , give an expression for the general solution
in terms of y1 , y2 . What happens if the roots of this quadratic coincide?
In the general case, when the coefficients are not necessarily constant, we note the fol-
lowing properties:
1. the Riccati equation is form invariant under linear fractional transformations (Möbius
transformations):
αy(x) + β δe
y (x) − β
y(x) 7→ ye(x) = ⇔ y(x) = , (1.3.34)
γy(x) + δ −γey(x) + α
28 LECTURE 1. ELEMENTARY THEORY OF DIFFERENCE EQUATIONS
3. Any four solutions yi (x) , (i = 1, 2, 3, 4) of the same Riccati equation are related
through the relation
(y1 − y2 )(y3 − y4 )
CR[y1 , y2 , y3 , y4 ] =: = constant , (1.3.35)
(y1 − y3 )(y2 − y4 )
where CR stands for cross-ratio. In fact, for any two solutions yi , yj of the Riccati
equation, it is easy to derive:
d
ln(yi − yj ) = a(yi + yj ) + b ,
dx
and, hence, the combination
d
ln((y1 − y2 )(y3 − y4 ))
dx
is invariant under permutations of the indices.
4. Given two solution y1 (x), y2 (x) of the same Riccati equation, a new solution can be
found in the “interpolation form”:
y1 + ρy2 dρ
y(x) = provided ρ(x) solves = a(y1 − y2 )ρ . (1.3.36)
1+ρ dx
Hence, given any two solutions, we can find an interpolating solution of the Riccati
equation through a solution of the linear equation for ρ.
Exercise 1.3.2. Prove the statement in part 4 above, deriving the linear differential equation
for ρ in (1.3.36), given two solutions y1 , y2 of the Riccati equation.
1.3. A SIMPLE NONLINEAR DIFFERENCE EQUATION – DISCRETE RICCATI EQUATION29
is again a solution of the discrete Riccati equation provided that ρn solves the linear
difference equation
(1) (1) (2) (2)
(yn(2) yn+1 + an yn+1 + bn yn(2) + cn )ρn + (yn(1) yn+1 + an yn+1 + bn yn(1) + cn )ρn+1 = 0 .
(1) (2)
By using again the discrete Riccati equation for yn , yn , we can deduce from this
that this linear equation can be written in either of the two equivalent forms:
(2) (1)
y n + an yn+1 + bn
ρn+1 = (1)
ρn = (2)
ρn .
y n + an yn+1 + bn
(i) (j)
3. For any two different solutions yn and yn we have the following relation
(i) (j)
yn+1 − yn+1 cn − an b n
(i) (j)
= (i) (j)
,
yn − yn (an + yn )(an + yn )
and hence we conclude that
(1) (2) (3) (4))
(yn+1 − yn+1 )(yn+1 − yn+1 ) (cn − an bn )2
(1) (2) (3) (4))
= (1) (2) (3) (4)
,
(yn − yn )(yn − yn ) (an + yn )(an + yn )(an + yn )(an + yn )
(1) (2) (3) (4)
for any four solutions yn , yn , yn , yn . Consequently, we infer immediately that for
(1) (2) (3) (4)
any four solutions yn , yn , yn , yn , of the discrete Riccati equation we have that
the cross-ratio
(1) (2) (3) (4)
(yn − yn )(yn − yn )
CR[yn(1) , yn(2) , yn(3) , yn(4) ] = (1) (3) (2) (4)
= constant .
(yn − yn )(yn − yn )
4. By setting yn = fn /gn in (1.3.37) we can show that the equation can be linearised
leading to the linear matricial equation:
fn+1 bn cn fn
=κ , (1.3.39)
gn+1 −1 −an gn
in which κ can be chosen arbitrarily. It is easy to derive from this matricial equation
a second order linear difference equation for fn .
Exercise 1.3.4. Derive the matrix equation (1.3.39) from the discrete Riccati equation
(1.3.37), and by eliminating gn derive a second order difference equation for fn .
Example: An alternative, more direct way to linearize the Riccati equation (1.3.37) is by rewrit-
ing the equation as
(yn + an )(yn+1 + bn ) + cn − an bn = 0 , (1.3.40)
which suggests a substitution of the form
wn
yn + a n = ⇒ wn+1 + (bn − an+1 )wn + (cn − an bn )wn−1 = 0 ,
wn−1
which brings us directly to a linear second order difference equation for wn .
In some isolated cases a judicious grouping of terms can help to find a linearizing trans-
formation.
1.4. PARTIAL DIFFERENCE EQUATIONS 31
Exercise 1.3.5. Linearize the logistic equation (Verhulst equation) dy/dx = ay(1 − y) by
setting y = 1/(w + b). Do the same for a discrete analogue of the logistic equation, namely
yn+1 = ayn (1 − yn+1 ) . Show also that the linearization trick breaks down for another
discretization of the logistic equation, namely yn+1 = ayn (1 − yn ) .
∂y ∂2y
= .
∂t ∂x2
If we discretize the time derivative using the forward difference ∆ and the second order
space derivative with ∆∇ we obtain
1 1
ε (y(x, t + ε) − y(x, t)) = δ 2 (y(x + δ, t) − 2y(x, t) + y(x − δ, t)),
or with discrete indices identifying yn,m = y(x0 + nδ, t0 + mε), relative to some arbitrary
origin (x0 , t0 ), we can then write this equation more simply as:
where α = ε/δ 2 , β = 1 − 2ε/δ 2 . The points that are involved in this equation are indicated
in Figure 1.2
An important problem in solving P∆E’s is the boundary or initial value question. This
often requires a careful consideration of graph of the lattice points involved in the equation.
In the present case we can see from Figure 1.2 that given data on the line t = 0, i.e., m = 0
we can propagate the solution forward in time. If instead we want to propagate into the
positive x-direction, we need initial data on two vertical lines, say n = n0 , n0 + 1. We could
32 LECTURE 1. ELEMENTARY THEORY OF DIFFERENCE EQUATIONS
t m t m
x n x n
a) b)
Figure 1.2: Points involved a) in equation (1.4.41) and b) in (1.4.42).
also use two lines on an angle, such as m − n = s0 , s0 + 1. These initial values correspond
to the equation being second order in the x derivatives.
For practical calculations it is often useful to introduce two shift operators,
T yn,m = yn+1,m , Syn,m = yn,m+1 .
For example the discrete heat-equation (1.4.41) can then be written as
Syn,m = (α + β + αT −1 )yn,m .
If we have now the initial value given as yn,0 = fn then we find the solution in the form
yn,m = (αT + β + αT −1 )m fn .
Next consider the Laplace equation
∂2y ∂ 2y
+ = 0.
∂t2 ∂x2
This could be discretize as the second derivative in the heat equation, but in both x and t.
If we keep the same central point in both discretizations we obtain
yn,m+1 + yn,m−1 + σ(yn+1,m + yn−1,m ) + ρyn,m = 0, (1.4.42)
where σ = ε2 /δ 2 , ρ = −2(1 + ε2 /δ 2 ). The points that are involved in this equation are
indicated in Figure 1.2. From the Figure it is now clear that we need can proceed if we are
given the initial values on two adjacent vertical lines or horizontal lines.
k = n, l = 2n + m
Other classes of mostly nonlinear P∆Es will be considered in subsequent lectures mo-
tivated by the context of integrable systems. Their stencil on the lattice can take different
shapes, e.g. of a four-point hook form:
ε
n
δ
a) b) c)
Figure 1.3: A difference equation may involve a different configuration of points in the
lattice. The cases a), b) and c) correspond to (1.4.43), (1.4.44) and (1.4.45), respectively
34 LECTURE 1. ELEMENTARY THEORY OF DIFFERENCE EQUATIONS
ց
?
ր
From Figure 1.4 we can also infer that the definition of an initial value problem can be
delicate. For example in case b), which is discussed in detail later, we can propagate from
any staircase-like initial values, but any “overhang” could lead to a contradiction in the
initial-value problem (due to overdeterminedness).
The problem of establishing continuum limits is also more complicated for P∆E’s than
for O∆E’s, and in fact not unique: there are more than one way of obtaining contiuous
equations from discrete ones by limits. We return to this problem later, at this point it
suffices to note that sometimes we may have to take the continuum limit along a diagonal
of the lattice in order to obtain nontrivial equations.
Literature:
1. L.M. Milne-Thompson, The Calculus of Finite Differences, (AMS Chelsea Publ. 2000),
first edition, 1933.
2. C.M. Bender and S.A. Orszag: “Advanced Mathematical Methods for Scientists and
Engineers”, (McGraw-Hill, 1978)
3. W.G. Kelley and A.C. Peterson: “Difference Equations: An Introduction with Appli-
cations”, Second Ed. (Academic Press, 2001).
4. P.R. Garabedian, Partial Differential Equations, (Chelsea, 1964).
5. F.B. Hildebrand, Finite-Difference equations and Simulations, (Prentice-Hall, 1968).
6. G. Andrews, R. Askey and R. Roy, Special Functions, (Cambridge Univ. Press, 1999).
Lecture 2
In this part we will discuss how difference equations arise from transformations applied to
differential equations. Thus we find that some functions can be defined both by a differential
equation and by a difference equation. Of course the independent variables of these two types
of equations are not the same, but rather we find interesting duality between parameters
and independent variables.
Often functions defined by a differential equation posses special transformations. One
particularly important class of transformations are those that act on the differential equation
by changing the values of its parameters. When such transformations are iterated, we obtain
a sequence of differential equations (and often a sequence of solutions) that are strung
together by the changing sequence of parameter values.
As particularly simple example consider the differential equation
x w′ = α w. (2.0.1)
Here x is the independent (continuous) variable, the prime stands for x-derivative (w′ = dwdx )
and α is a parameter of the equation. It is easy to find the solution of this equation, it is
wα (x) = k xα , (2.0.2)
where k is required integration constant. From the form of the solution (2.0.2) we can see,
that the simple transformation of multiplying the function wα by x changes α, that is
wα+1 = x wα . (2.0.3)
35
36LECTURE 2. FROM CONTINUOUS TO DISCRETE EQUATIONS VIA TRANSFORMATIONS
which could also be stated as saying that the operation of taking the derivative d/dx and
that of the shift in the parameter α, Tα , commute on solutions, i.e., (Tα w)′ = Tα (w′ ) . This
is a theme we will see coming back many times.
We will first discuss the case of such procedures for ordinary differential equations
(ODEs), using the examples of equations for certain classical special functions, such as
the Bessel, Legendre, and hypergeometric functions. These can be shown not only linear
ODEs, but also some ordinary difference equations (O∆Es) in terms of certain parameters.
There are also nonlinear special functions, such as the elliptic functions and the Painlevé
transcendents, that satisfy nonlinear differential equations, for which similar ideas apply,
but we will not treat them in this Lecture. Next, we will move from ODEs to partial differ-
ential equations (PDEs), and treat briefly some basic theory of soliton equations, and show
how similar ideas there lead to the construction of integrable partial difference equations
(P∆Es).
where the primes denote differentiation in x. This equation arises in many physical applica-
tions, for example in the study of quantum mechanics in harmonic potential. A particular
special solution of this equation is the parabolic cylinder or Weber function w =: Dα (x),
uniquely specified by the asymptotic behaviour
α −x2 /4 α(α − 1) 1
Dα (x) = x e 1− 2
+O , x → +∞. (2.1.5)
2x x4
The simplest special case arises when α is an integer n, this yields the Hermite polynomials
Hn (x), defined by
√
Dn (x) = 2−n/2 exp −x2 /4 Hn (z), z = x/ 2. (2.1.6)
Instead of thinking of Equation (2.1.4) as one equation specified by one fixed value of
α, it is more productive to think of it as an infinite sequence of equations, each of which is
2.1. SPECIAL FUNCTIONS AND LINEAR EQUATIONS 37
or after expanding
e′′ + (α − 1) +
w 1
2 − 41 x2 we = 0. (2.1.11)
which is the same as (2.1.4) except for a new parameter value α̃ = α − 1, and thus w eα̃(x) ∝
wα−1 (x).
Since the equations are linear the proportionality constant is in principle free, and is
determined by some other considerations, like normalization. In this particular case the
solutions of (2.1.4) (the parabolic cylinder functions Dα (x)) are conventionally normalized
to have the asymptotic behaviour as given in (2.1.5). Using (2.1.8) on (2.1.5) we get
2
∂x + x/2 Dα (x) = α xα−1 e−x /4
1 + O 1/x2 = αDα−1 ,
Equation (2.1.9) can be written in two additional ways, using this normalization. First by
solving for the term that appears alone:
Dα−1 (x) = α1 [Dα′ (x) + 21 xDα (x)],
(2.1.12)
Dα (x) = Dα−1 ′
(x) − 12 xDα−1 (x).
The first pair allows us to travel up and down in the chain of Dα (x)’s for different α’s
separated by integers. The second form (2.1.13) allows us to derive a fully discrete equation:
38LECTURE 2. FROM CONTINUOUS TO DISCRETE EQUATIONS VIA TRANSFORMATIONS
substituting α → α + 1 into the second equation of (2.1.13) and then subtracting the two
equations we get
Dα+1 (x) − x Dα (x) + α Dα−1 (x) = 0. (2.1.14)
This is a difference equation, where α is the independent variable and x the parameter.
Let us now reflect on what we have obtained. The conventional theory of ordinary
differential equations regards the parameters α as given and fixed, while x varies in some
domain. By using the Darboux transformation (2.1.8) we can change the parameter α. In
the fully dual point of view, we would keep x is fixed while α changes. At an intermediate
step also have recurrence relations that relate w(x; α), w(x; α̃) and their x-derivatives. In
the theory of integrable equations, such recurrence relations are called (auto-) Bäcklund
transformations.
Exercise 2.1.2. Use the relationship between Weber function and Hermite polynomials
(2.1.6) to derive the difference equation for Hermite polynomials Hn+1 (x) = 2xHn (x) −
2nHn−1 (x), starting from (2.1.14).
with parameter h. Suppose f (x) is a particular solution of this equation for some specific
value h1 of h. Let us define a new function ye by
We say that φe is the Darboux transform of the potential and ye of the wave-function.
What is behind this theorem is a factorisation property. In fact, given that the “poten-
tial” function φ(x) does not depend on the (spectral) parameter h, the solution y(x, h1 ) =
f (x) of (2.1.15) allows one to express this potential as
y ′′ = (h − h1 + v ′ + v 2 )y , where v = ∂x log f .
e
the equation (??) with a new potential φ(x) = φ(x) − 2v ′ given by (2.1.16).
We note that the result on the Weber function of the previous section can be viewed
as an application of Darboux’ theorem: In comparison with (2.1.4) we see that we should
choose φ = x2 /4, f = ex /4 , h1 = 1/2. Then we find that φe = φ − 1 in agreement with
2
(2.1.11).
The pair of equations (2.1.12) are a special case of a con-
struction named after Albert Victor Bäcklund (1845 - 1922) who
worked on transformations of surfaces in differential geometry.
In the the modern era the connection between the latter sub-
ject and the theory of differential equations has become more
prominent. In fact, there is close relation between transforma-
tions between special surfaces in terms of coordinates on these
surfaces and transformations between solutions of (linear and
nonlinear) differential equations. As we shall see in the next
section these transforms will form the basis of the construction
of exact discretizations of those very same differential equations.
We defer the precise definition of a Bäcklund transformation to
later when we deal with PDEs, but only present here a loose Figure 2.2: A.V. Bäcklund
definition of this concept as follows:
40LECTURE 2. FROM CONTINUOUS TO DISCRETE EQUATIONS VIA TRANSFORMATIONS
where ν is a parameter, with specific asymptotic behaviours: The standard Bessel function
of the first kind is defined by1
x ν X
∞ k
−x2 /4
Jν (x) = . (2.1.20)
2 k! Γ(ν + k + 1)
k=0
We will again consider Bessel’s equation as an infinite sequence of equations in the space
of the parameter value ν. To allow us to iterate in this parameter space, we need recurrence
relations for Bessel functions. To find Darboux transformations from which recurrence
relations follow, it is easier to start by transforming equation (2.1.19) into a socalled Sturm-
Liouville form, i.e. without the first derivative term.
Exercise 2.1.3. Transforming the dependent variable w by setting w(x) = p(x)y(x), and
substituting this into eq. (2.1.19), show that we can derive a second order ODE for
√ y without
a term containing y ′ , by choosing 2x p′ (x) + p(x) = 0, which implies p(x) = 1/ x.
√
Transforming variables by taking w(x; ν) = y(x; ν)/ x, we get
ν 2 − 1/4
y ′′ + 1 − y=0 (2.1.21)
x2
Eq. (2.1.21) corresponds to Darboux’ equation (2.1.15) by setting φ(x) = (ν 2 −1/4)/x2 and
h = −1. Following Darboux’ theorem 2.1.1 we should first construct a suitablefunction f ,
1 Recall the definition of the Γ-function, which is used in this formula, as the function obeying the difference
which we can take in the form f := xσ which corresponds to the value h1 = 0 in (2.1.15),
and where we need to take σ = (1/2) ± ν. Taking, for convenience, the lower sign we obtain
the following transformation from (2.1.17)
Thus, we see here again an example here the Darboux transformation implies an integer
step in the parameter, namely the parameter ν, and therefore we can identify
up to an arbitrary constant factor, which (as follows from an explicit computation) we can
set equal to 1, to be consistent with the series representation (2.1.20) of the solution of the
Bessel equation.
Using (2.1.22)) we find easily the factorization
Now solving for the derivative terms with (2.1.24) and shifting ν → ν + 1 in the second
equation we obtain
xwν′ = −ν wν + x wν−1 ,
(2.1.28)
xwν′ = ν wν − x wν+1 ,
The equations (2.1.28) and (2.1.29) hold for the Bessel functions of the first kind Jν (x).
42LECTURE 2. FROM CONTINUOUS TO DISCRETE EQUATIONS VIA TRANSFORMATIONS
ψxx + uψ = λψ , (2.2.31a)
ψt = 4ψxxx + 6uψx + 3ux ψ, (2.2.31b)
whose consistency condition leads to (2.2.30). The first equation (2.2.31a) has the form of
a linear spectral problem for the differential operator
L = ∂x2 + u , (2.2.32)
where the coefficient u = u(x, t) plays the role of a potential. The parameter λ is an
eigenvalue of the operator L and can, in principle, depend on t if u depends on it, but
by the definition of an eigenvalue of a differential operator it should independent of x. The
2.2. BÄCKLUND TRANSFORMATIONS FOR NON-LINEAR PDES 43
second equation (2.2.31b) describes the (linear) time-evolution of the function ψ(x, t), where
again the same u enters in the coefficients. The system (2.2.31) is overdetermined: the two
linear equations can only be compatible with each other if additional conditions hold for the
coefficients, which are all expressed in terms of u. Furthermore, we shall assume that the
time-evolution descibes an isospectral deformation 2 of the linear spectral problem, i.e., we
assume that λ is, in fact, independent of t.
Theorem 2.2.1. Under the assumption of isospectrality, i.e., λt = 0 , the linear system
(2.2.31) is self-consistent, i.e., (ψxx )t = (ψt )xx , iff either ψ ≡ 0 or the potential u = u(x, t)
obeys the Korteweg-de Vries (KdV) equation (2.2.30).
Proof. First, using (2.2.31a) rewrite (2.2.31b) as follows:
ψt = (4λ + 2u)ψx − ux ψ .
and now consider the cross-derivatives:
(ψxx )t = ((λ − u)ψ)t = (λt − ut )ψ + (λ − u) [(4λ + 2u)ψx − ux ψ]
(ψt )xx = (4λ + 2u)ψxxx + 4ux ψxx + 2uxxψx − ux ψxx − 2uxx ψx − uxxx ψ
= (4λ + 2u) ((λ − u)ψ)x + 3ux (λ − u)ψ − uxxxψ
⇒ (ψt )xx − (ψxx )t = (ut − uxxx − 6uux) ψ − λt ψ .
Hence, under the condition of isospectrality, λt = 0 , we see that the system is compatible,
i.e. (ψt )xx = (ψxx )t provided u(x, t) obeys the KdV equation.
The linear system of equations associated with a non-linear PDE is symptomatic of its
integrability through the inverse scattering method. Such a linear system, consisting of a
spectral problem and an equation for the time evolution, is called a Lax pair, after P.D.
Lax who gave a systematic framework for describing such linear problems in his celebrated
paper [Lax,1968]. Although for virtually all soliton equations Lax pairs have been found,
there is no fully algorithmic method known to produce a Lax pair for a given equation.
potential u(x, t) as t varies, which in principle may affect the spectrum of the operator, i.e., the collection of
eigenvalues. However, if the deformation is ”isospectral” the spectrum is preserved while varying t, meaning
that the eigenvalues λ do not change as t varies.
44LECTURE 2. FROM CONTINUOUS TO DISCRETE EQUATIONS VIA TRANSFORMATIONS
u = λ − vx − v 2 . (2.2.35)
Furthermore, after taking derivatives w.r.t. x on both sides of (4.2.29) and expressing all
terms using v we get a PDE governing v:
The latter equation (for λ = 0) is known as the modified KdV equation (MKdV), and
it differs from the KdV equation (2.2.30) notably in the nonlinear term. The differential
substitution (2.2.35) allows one to find a solution of the KdV equation given a solution of
the MKdV equation: if v solves the MKdV (2.2.36) and u is defined by (2.2.35), then u
solves KdV (2.2.30).
e = λ + vx − v 2 ,
u (2.2.37a)
u = λ − vx − v 2 . (2.2.37b)
e + u = 2(λ − v 2 )
u (2.2.38a)
e − u = 2vx .
u (2.2.38b)
The latter can be integrated if we introduce the variable w by taking u = wx . For the KdV
equation this change of variables leads to
after one integration in x. (Note that we have omitted an irrelevant integration constant.)
This equation for w is called the potential KdV equation (PKdV).
In terms of this new dependent variable the equation (2.2.38b) can be integrated to
e − w = 2v , and inserting it into the first relation (2.2.38a) we obtain
w
1 2
e + w)x = 2λ −
(w e − w) .
(w (2.2.40a)
2
2.2. BÄCKLUND TRANSFORMATIONS FOR NON-LINEAR PDES 45
Example: As a specific example, consider the simplest case where the seed solution w of the
PKdV equation (2.2.39) is the trivial solution w ≡ 0. Setting w ≡ 0 in (2.2.41) yields
1 2
w
ex = 2λ − w
e ,
2
which can be integrated by separation of variables and yields
w(x,
e t) = 2k tanh (kx + c(t)) , λ = k2 .
Substituting this expression into the (2.2.42) (with w = 0) reveals that we must take ct = 4k3 t + c0 ,
were c0 is a constant. Thus we have obtained the solution:
e t) = 2k tanh kx + 4k3 t + c0 .
w(x, (2.2.43)
This implies that for the solution of the KdV equation we obtain the solution
u ex = 2k2 sech2 kx + 4k3 t + c0 ,
e(x, t) = w (2.2.44)
which is the famous formula for the 1-soliton solution of the KdV equation.
λ 1 2
e
BTλ : w 7→ w e + w)x = 2λ −
(w e − w) ,
(w (2.2.45a)
2
µ 1 2
b
BTµ : w 7→ w b + w)x = 2µ − (w
(w b − w) , (2.2.45b)
2
where we have used the notation w, e w
b to denote the solution obtained by applying the BT
with parameter λ, µ, respectively.
There are now two ways to compose these BTs: either start with BTλ and and subse-
quently apply BTµ , or the other way around. In this way we get iterated solutions which
we can denote by wbe and w eb respectively,
be = BTµ ◦ BTλ w,
w eb = BTλ ◦ BTµ w.
w
The highly nontrivial result is that, under certain conditions, both ways of composing BTs
lead to the same result: w be = w,
eb and hence the two BTs commute. This is the famous
permutability property of the BTs.
2.2. BÄCKLUND TRANSFORMATIONS FOR NON-LINEAR PDES 47
Theorem 2.2.3. The BTs given by (2.2.45a),(2.2.45b) for different parameters λ and µ
generate solutions (with a suitable choice of integration constants) for which we have the
following commutation diagram of BTs:
BTλ e
w BTµ
w be = w
w eb
BTµ BTλ
b
w
The proof of the permutability property is quite deep and relies on the spectral properties
that play at the background of the equations.
Proof. A direct proof is computational. Using the additional relations:
µ b
1 b 2
e 7→ w
BTµ : w e be + w
w e e−w
w e ,
= 2µ − (2.2.45c)
x 2
1 e 2
λ e e
b 7→ w
BTλ : w b b+w
w b = 2λ − b−w
w b . (2.2.45d)
x 2
be as follows:
e in terms of w and w
Iterating the first Bäcklund chain we can solve w
1 b be − w)x + 2(λ − µ)
(w
e=
w e + w) +
(w ,
2 be − w
w
and reinserting this into the BT we obtain:
2
λ+µ be + w)x + ∂ 2 log(w
= (w be − w) + 1 ∂x log(w
be − w)
x
2
1 b 2 (λ − µ)2
+ (w e − w) + 2 ,
8 be − w)2
(w
and the latter relation is symmetric under the interchange of λ and µ. A similar symmetry
can be derived for the t-part of the BT. Hence, starting from an arbitrary seed w we can
find solutions, by appropriate choice of integration constants, which are symmetric under
interchange of λ and µ.
The consequences of the permutability property are far reaching, and we will give an
explicit realization as follows. In fact, using all four Bäcklund relations (2.2.45a)-(2.2.45d),
and setting w be = w,
eb we can now eliminate all the derivatives from the four eqs. (2.2.45) we
obtain a purely algebraic equation of the form:
be − w)(w
(w b − w)
e = 4(µ − λ). (2.2.46)
be without having to
This allows us to obtain directly the iterated BT transformed variable w
derive the solution through the Riccati equations of the BT.
48LECTURE 2. FROM CONTINUOUS TO DISCRETE EQUATIONS VIA TRANSFORMATIONS
Exercise 2.2.1. Construct a 2-soliton solution by starting from the seed solution w ≡ 0,
and two 1-soliton solutions of the type (2.2.43) (with different parameters k and l, where
be from (2.2.46). Note that the phases c0 may
λ = k 2 , µ = l2 , respectively) and solving for w
be taken to be different for w e and w.
b Verify explicitly that the wbe so constructed actually
solves (2.2.39).
i
If, for simplification, we denote e 2 θ = w, we can write (2.2.50) as
bew
λ(w b − ww) bew
e = µ(w e − ww).
b (2.2.51)
2.2. BÄCKLUND TRANSFORMATIONS FOR NON-LINEAR PDES 49
e
ee
w
ee
w
b
ee
e
w w
w be
w
b
w b
be
w
bb
w
b
bb
w
Exercise 2.2.2. Starting from the trivial solution of the sine-Gordon equation θ ≡ 0, use
e t) of the same equation containing the parameter
the BT (2.2.48) to obtain a solution θ(x,
λ, namely:
e t) = 4 tan−1 exp λx + t + ϕ0
θ(x, , (2.2.52)
λ
where ϕ0 is a (constant) phase. (Hint: Use the integral
Z
dϕ ϕ
= ln tan + c.
sin ϕ 2
after which we can write (2.2.46) and (2.2.51) as difference equations of the form:
and
respectively, the shifts along the lattice wn,m 7→ wn+1,m and wn,m 7→ wn,m+1 corresponding
to the application of the Bäcklund transformations BTλ and BTµ .
We will now change our point of view: at each elementary plaquette of the above lattice
of solutions we have a relation of the form (2.2.54) or (2.2.55) (or something else for other
equations), and we will now elevate these equations as being the main equations of interest.
In a sense having reached this point we can “forget” about the original PDE’s from whence
the construction originated, and place the permutability equations at the centre of our focus.
Later we will indeed consider lattice equations, i.e., P∆E’s like (2.2.54) or (2.2.55) on
their own merit and study their remarkable properties. In particular we will show that these
equations are integrable in some precise sense.
Literature:
1. E.T. Whittaker and G.N. Watson, A course of Modern Analysis, fourth edition, 1927,
(Cambridge University Press, 2002).
2. J.G. Darboux Lecons sur la theorie generale des surfaces et les applications geometriques
du calcul infinitesimal (Gauthier-Villars, Paris, 1914)
3. V.B. Matveev and M.A. Salle, Darboux transformations and solitons (Springer, 1991)
4. C. Rogers and W.K. Schief, Bäcklund and Darboux Transformations (Cambridge Uni-
versity Press, 2002).
5. D.J. Korteweg and G. de Vries, On the change of form of long waves advancing in a
rectangular channel, and on a new type of long stationary waves, Philos. Mag. (5) 39
(1895) 422–443.
6. C.S. Gardner, J.M. Greene, M.D. Kruskal and R.M. Miura, Method for solving the
Korteweg-de Vries equation, Phys. Rev. Lett. 19 (1967) 1095–1097.
7. P.D. Lax, Integrals of nonlinear equations of evolution and solitary waves, Commun.
Math. Phys. 21 (1968) 467–490.
8. R.M. Miura, The Korteweg-de Vries Equation: A Survey of Results, SIAM Review 18
# 3 (1976) 412–459.
9. G.L. Lamb, Elements of Soliton Theory, (Wiley Interscience, 1980).
10. M.J. Ablowitz and H. Segur, Solitons and the Inverse Scattering Transform, (SIAM,
1982).
11. F. Calogero and A. Degasperis, Spectral Transform and Solitons, vol. 1, (North-
Holland Publ., Amsterdam, 1982).
12. A.C. Newell, Solitons in Mathematics and Physics, (SIAM, Philadelphia, 1985).
13. P.G. Drazin and R.S. Johnson, Solitons: An Introduction, (Cambridge University
Press, 1989).
14. M.J. Ablowitz and P.A. Clarkson, Solitons, Nonlinear Evolution Equations and Inverse
Scattering, (Cambridge Univ. Press, 1991).
Lecture 3
Integrability of P∆Es
We will investigate here partial difference equations (P∆E’s) of the following canonical form
(which we will call quadrilateral P∆Es)
e, u
Q(u, u b
b, u
e) = 0 , (3.1.1)
51
52 LECTURE 3. INTEGRABILITY OF P∆ES
u e
u
- -
? ?
? - -?
b
e
u
b
u
The notation is inspired by the one for Bäcklund transformations, which as we have
seen in the previous Lecture, give rise to purely algebraic equation as a consequence of their
permutability property. However, here we will forget, at first instance, about this connection,
and consider lattice equations of the form (3.1.1) in their own right as a partial difference
equation (P∆E) on a two-dimensional lattice. Thus, n and m, play the role of independent
(discrete) variables, very much like x and t being the continuous variables for an equation
like the KdV equation
Even though the form (3.1.1) seems very restrictive, it will turn out that it is in a sense the
most elementary form as a model type of equations, and on the other hand remarkably rich.
In fact, as we shall see later, P∆Es of the form (3.1.1) are “rich enough” to be candidates
for discretisations of PDEs of arbitrary order in the spatial and temporal variables.
A classification of P∆Es in the same way as of PDE s, does not yet exist. Nonetheless, we
may consider a P∆E of the form (3.1.1) to have features reminiscent of hyperbolic PDEs.
In fact, if the equation Q = 0 can be solved uniquely for each elementary quadrilateral,
then one may pose initial value problems (IVPs) in ways very similar to hyperbolic type of
equations such as the KdV equation itself (i.e. as a discrete nonlinear evolution equation).
This can be done as follows.
The naive approach would be to consider IVP where we would assign values of the
dependenet variable u along horizontal array of vertices in the lattice, i.e. values for un,0 ,
for all n, considering the variable m to be the temporal discrete variable. It is easy to see,
however, that such an IVP would lead to a nonlocal problem if we want to use (3.1.1) as an
iteration scheme to find all values un,m for m > 0. In fact, to calculate any value un,1 say
for given n, we would need to involve all initial values un′ ,0 with n′ < n, and furthermore
have to assume limiting behaviour as n → −∞. This would be a complicated procedure.
However, there is nothing that tells us that we should identify the n- and m axes as the
spatial and temporal axes respectively. The lattice picture allows us to play other and more
3.1. QUADRILATERAL P∆ES 53
natural games, and it is the equation itself that gives the lead in provding is with the natural
IVP that we could impose. Thus, changing the perspective slightly, we may tilt the lattice
and rather consider initial value data to be imposed on configurations like a “sawtooth” (or
a “ladder”), such as:
R R R
The analogy with hyperbolic PDEs can be further seen from the consideration of the
“memory” each point in the lattice has of the initial values that are involved in its determi-
nation upon iteration. In fact, any given point has a backward shadow (the analogue of the
so-called “lightcone”) of points the values of u on which determine the value at that point,
as indicated by the picture:
the point at the bottom of the cone being fully determined by the initial values at the top
of the cone and only and exclusively by those values! This is very much reminiscent again
of what happens in the case of hyperbolic PDEs.
u
b
u u e
u
e u
b
u
It is easy to see that the lattice equation (2.2.54) derived in Lecture 2 from the Bäcklund
transform of KdV is precisely of the form given above. However, not all equations of the
form (3.1.2) will be of interest to us. We will be interested in those equations (for specific
choices of the coefficients k1 , . . . , k6 ) which have the additional property of being integrable.
What we mean by this will be explored in the next section.
b, u
Q(u, u eub
e; p, q) = 0 , (3.2.1)
and if we demand that the values for u at each vertex can be solved uniquely, implying
multilinearity in each of these variables around the quadrilateral, then we are led again to
eq. (3.1.2) with the coefficients k0 , . . . , k6 depending on the lattice parameters p and q in a
specific way. The question is what criteria to use in order to chose that dependence!
It is here that we will restrict ourselves to quadrilateral P∆Es which we regard to be
integrable. The question of what is the proper definition integrability, and to answer that
question in general is very difficult: a one-fits-all definition (for all the possible type of
systems that we would like to regard as being integrable) is possibly not possible to give
in a precise mathematical sense. However, if we restrict ourselves here to P∆Es, and in
particular quadrilateral P∆Es, then we can aspire to be a bit more precise. We will explore
the definition of an integrable quadrilateral P∆E by means of the following example, namely
the example of the P∆E arising from the BTs for the KdV equation, eq. (2.2.54).
First, we remark that the presence of the lattice parameters p, q is crucial: whereas
normally we would like to consider the parameters to be chosen once and for all and then
remain fixed (thus specifying a specifi equation) when solving the equation on the lattice,
here we will argue that we should look at (2.2.54) as defining a whole parameter-family of
equations, and that it makes sense to look at them altogether with p and q variable. In
3.2. INTEGRABILITY AND CONSISTENCY-AROUND-THE-CUBE 55
doing this, however, we must attach each parameter to a specific diecrete variable such as
p being associated with the variable n, and q with m. This is also natural from the way in
which we derived (2.2.54) from the BT construction: each BT is attached to a parameter (λ,
µ,. . . ), and with each parameter we can build a direction in an infinite-dimensional lattice
of BTs.
The main point we want to make now is the following:
Let us illustrate this by means of the example of (2.2.54), in which we will identify (for
later convenience) the parameters 4λ := p2 and 4µ := q 2 . What the statement above
suggests is that we can “embed” the equation in a multi-dimensional lattice by conbsidering
the dependent variable w, not only to depend on n and m (with associated lattice paameters
p and q espectively), but that w may in fact depend on an infinity of lattice variables each
associated with its parameter, as follows:
w = wn,m,h,... = w(n, m, h, . . . ; p, q, r, . . . )
and with each of these variables we have a corresponding elementary shift on the lattice:
e := wn+1,m,h
w , b := wn,m+1,h
w , w := wn,m,h+1 ...
Rewriting the equation (2.2.54) in the form (2.2.46) with the substitutions for λ and µ
investigate what happens if we impose a copy of the same equation in all three lattice
directions. This would lead to the system of equations:
b − w)(w
(w e be = p2 − q 2 ,
− w) (3.2.2a)
e
(w − w)(w −we ) = p2 − r 2 , (3.2.2b)
(w b = r2 − q 2 .
b − w)(w − w) (3.2.2c)
These equations are consistent if the evaluations along the cube are independent of the way
of calculating the final point. Imposing initial values:
w := a , e =: b
w , b =: c
w , w =: d
56 LECTURE 3. INTEGRABILITY OF P∆ES
w e
w
e
w
w
b
w be
w
b
w b
e
w
In fact:
2 2
e = (p − q )w
b ew b + (r2 − p2 )w w
b + (q 2 − r2 )ww e
w . (3.2.3)
(r2 − q 2 )we + (q 2 − p2 )w + (p2 − r2 )w
b
independent of the way in which he value at this vertex is calculated! This property, implying
that under relevant initial value problems on the 3-dimensional lattice the iteration of the
solution can be performed in an unambiguous way (namely independent of the way by which
we perform the calculation of the iterates) will be referred to as the consistency-around-the-
cube (CAC) property. It is this property that we shall consider to be the main hallmarfk of
the integrability of the equation.
Remark: We note also, for later reference, that in the above case the formula for w b
e is
independent of the value w at the opposite end of the main diagonal across the cube. This
property we will refer to as the tetrahedron property.
b−u
(p − q + u b
e) = p2 − q 2 ,
e)(p + q + u − u (3.2.4)
and this is, in fact, equivalent to the equation (2.2.54) by the change of (dependent)
variable w = u − np − mq − c (c a constant with respect to n and m). We will show
in Lecture 4 that this equation reduces to the potential KdV equation after a double
continuum limit, so that we rightly regard it as a discretisation of the latter equation.
Lattice potential MKdV equation:
v − veb
p(vb v − vbb
ve) = q(ve ve) (3.2.5)
Solutions of this equation are related to the solutions of the previous one (3.2.4) via
the relations
pe
v − qb
v b pv + qb
ve
b−u
p−q+u e= , e=
p+q+u−u , (3.2.6)
v e
v
3.3. LAX PAIR FOR LATTICE POTENTIAL KDV 57
which is the “Schwarzian KdV equation”. This equation is invariant under Möbius
transformations:
αz + β
z 7→ Z = .
γz + δ
Solutions of eq. (3.2.7) are related to the ones of (3.2.5) via the relations:
p(z − ze) = ve
v , q(z − zb) = vb
v. (3.2.8)
Exercise 3.2.1. Show by explicit computation that the P∆Es given by (3.2.4)-(3.2.7) possess
the consistency-around-the-cube property.
Exercise 3.2.2. Show that by using the relations (3.2.6) one can derive (3.2.4) from (3.2.5)
and vice versa.
Noting that eqs. (3.3.1) are both fractional linear in W , we can linearise these equations by
the substitution:
F
W = ,
G
58 LECTURE 3. INTEGRABILITY OF P∆ES
leading to
Fe wF + (k 2 − p2 − ww)G
e
= ,
Ge F − wGe
Fb wF + (k 2 − q 2 − ww)G
b
= ,
b
G F − wGb
and since at least one of the two functions F or G can be chosen freely, this allows us to
split in each of these equations the numerator and denominator to give:
( (
Fe = γ wF + [k 2 − p2 − ww]Ge , Fb = γ ′ wF + [k 2 − q 2 − ww]G
b ,
e = γ (F − wG) respectively b = γ ′ (F − wG)
G e , G b ,
(3.3.2)
in which γ and γ ′ are to be specified later. What happens next is obvious: we introduce the
two-component vector
F
φ= ,
G
and write (3.3.2) as a system of two 2×2 matrix equations:
e = Lφ ,
φ b = Mφ ,
φ (3.3.3a)
How does this linear system work? The consistency relation of the linear problem (3.3.3a)
b
e =φ e
b , which is the condition that simply expresses
is obtained from the condition that φ
that φ must be a proper function of the lattice variables n and m. Calculating the left
hand-side of this condition we get on the one hand
b [ =L
e = (Lφ) b = LM
bφ b φ,
φ
e ^
b = (M fφe =M
f Lφ .
φ φ) = M
Equating both sides we see that a sufficient condition for the consistency is the matricial
equation:
b
LM =M fL , (3.3.4)
which we will loosely refer to as the Lax equation, but it is sometimes also referred to as
a discrete zero-curvature condition (the reason for this terminology will be explained in
(3.3.1)). Pictorially the Lax equation is illustrated by the following diagram
3.3. LAX PAIR FOR LATTICE POTENTIAL KDV 59
φ L e
φ
f
M
M
b
e
b
φ φ
b
L
where the vectors φ are located at the vertices of the quadrilateral and in which the matrices
L and M are attached to the edges linking the vertices.
Using the explicit form (3.3.3b) of the matrices L and M , and working out the condition
(3.3.4) we find
!
′ b k 2 − p2 − w
w bwbe w k 2 − q 2 − ww b
bγ
γ be =
1 −w 1 −w b
!
′ we k2 − q2 − w be
ew w k 2 − p2 − w w e
=γeγ be
1 −w 1 −w e
and we will chose the γ and γ ′ (which were unspecified so far) such that the relation for the
determinants of this equation:
γ b det(M ) = γ
bγ ′ det(L) f ) det(L)
e′ γ det(M (3.3.5)
is trivially satisified. Since in this example the determinants of L and M are given by
det(L) = p2 − k 2 , det(M ) = q 2 − k 2 ,
L = 1 + δL1 + · · · , M = 1 + ǫM 1 + · · · ,
60 LECTURE 3. INTEGRABILITY OF P∆ES
e, u
Q(u, u b
b, u
e; p, q) = 0 ,
using the notation indicated in the following diagram (renaming the lattice parameters in
order to avoid confusion with previously used notation),
u p e
u
- -
q ? ?q
? - -?
b
u p b
e
u
subject to the following restrictions:
a) Linearity: Q is multilinear in its arguments, i.e., it is linear in each vertex-variable u,e
u,
u b
b,u
e;
b) Symmetry: Q is invariant under the group D4 of symmetries of the square, generated by
the interchangements:
e, u
Q(u, u b
b, u
e; p, q) = ±Q(u, u
b, u b
e, u u, u, b
e; q, p) = ±Q(e e, u
u b; p, q) ; (3.4.1)
c) Tetrahedral Condition: in the consistency check, the evaluation of the point on the cube
b
e is independent of u.
given by u
3.4. ∗ CLASSIFICATION OF QUADRILATERAL P∆ES 61
Q-list:
Q1 : b)(e
p(u − u b
e) − q(u − u
u−u e)(b b
e) = δ 2 pq(q − p)
u−u (3.4.2a)
Q2 : b)(e
p(u − u b
e) − q(u − u
u−u e)(b b
e) + pq(p − q)(u + u
u−u e+u b
b+u
e) =
= pq(p − q)(p2 − pq + q 2 ) (3.4.2b)
Q3 : p(1 − q 2 )(ub
u+u b
e) − q(1 − p2 )(ue
eu u+u bu b
e) =
2 2
2 2 b 2 (1 − p )(1 − q )
= (p − q ) (b uue + uu
e) + δ (3.4.2c)
4pq
Q4 : p(ue
u+u bu b
e) − q(ub
u+u eub
e) =
pQ − qP b b
= 2 2
(b
uue + uu
e) − pq(1 + ue uubu
e) (3.4.2d)
1−p q
where P 2 = p4 − γp2 + 1 , Q2 = q 4 − γq 2 + 1 .
H-list:
H1 : b
e)(b
(u − u e) = p2 − q 2
u−u (3.4.3a)
H2 : b
e)(e
(u − u b) = (p − q)(u + u
u−u e+u b
e) + p2 − q 2
b+u (3.4.3b)
H3 : p(ue bb
u+u e) − q(ub
u eu
u+u b
e) = δ 2 (p2 − q 2 ) (3.4.3c)
A-list:
A1 : p (u + ub)(e b
e) − q(u + u
u+u e)(bu+be) = δ 2 pq(p2 − q 2 )
u (3.4.4a)
A2 : 2
p(1 − q )(ub eu
u+u b 2
e) − q(1 − p )(ue
u+u b
bu 2 2
e) + (p − q ) (1 + ue
uu b
bu
e) = 0
(3.4.4b)
The Q-equations are related through the following coalescence diagram, i.e. displaying
how the equations relate to eachother through certain limits on the parameters:
Q3 (Q3 )δ=0
Q4
Q2 Q1 (Q1 )δ=0
Soliton type solutions have been constructed by now for all equations in the ABS list [Nijhoff,
Atkinson & Hietarinta, 2009; Atkinson & Nijhoff, 2010].
62 LECTURE 3. INTEGRABILITY OF P∆ES
Remarks: It is in itself remarkable that the list of equations found is so short. In fact,
all equations in the list Q are in fact special subcases of the last equation, Q4, which was
discovered by V. Adler in 1997. The latter equation can also be expressed as:
h i
A [(u − b)(bu − b) − (a − b)(c − b)] (e b
e − b) − (a − b)(c − b) +
u − b)(u
h i
+B [(u − a)(e u − a) − (b − a)(c − a)] (b
u − a)(ub
e − a) − (b − a)(c − a) =
= ABC(a − b) (3.4.5)
cf. [9]. Here the extra parameters c, C are related through C 2 = 4c3 − g2 c − g3 and to
(a, A) and (b, B) through the relations:
[Hietarinta, 2005] which is somewhat simpler. In (3.4.6) the sn denote the Jacobi elliptic
functions, cf . Appendix A, of modulus k, i.e. sn(α) = sn(α; k), etc.
It should be noted that all previous examples presented in section 3.2 can be recognised
as special subcases of the equations in the lists Q,H and A, and which can be obtained by
“degeneration” of Adler’s equation in either of its forms. All these equations possess Lax
pairs, but it was only recently that explicit solutions were found for most of the new cases
including of Adler’s equation, cf. e.g. [12].
Adler’s equation is the discrete analogue of a famous soliton equation discovered in 1980
by I. Krichever and S. Novikov, which reads:
3 u2xx − (4u3 − g2 u − g3 )
ut = uxxx − , (3.4.7)
2 ux
and which generalises the Schwarzian KdV equation.
We recall that the difference between the continuous KdV equation (2.2.30) and its potential
version (2.2.39), as discussed in Lecture 2, was simply that its solutions were connected
3.6. SINGULARITY CONFINEMENT 63
through taking a derivative, u = wx . However, on the lattice there are many ways in
which can do the analogue of “taking the derivative”, for instance we can replace it by
taking a difference involving two neighbouring vertices, like ∆n wn,m ≡ wn+1,m − wn,m or
∆m wn,m ≡ wn,m+1 − wn,m , but these are not the only choices. Alternatively we can take
a difference between vertices farther away or across diagonals, such as:
It is then straightforward from (3.2.4) (using the notation in terms of e- and bshifts) to
derive the following equation for Q:
e= a − a
b−Q
Q ⇔
be
R−R =
a
−
a
, (3.5.3)
Q Q be b
R R e
where a ≡ p2 − q 2 . The equations for Q and R are simply related by the fact that QR = a.
A Lax pair for either of the equations (3.5.3) can be easily derived, by starting from
the Lax pair for the potential KdV equation. In fact, rather than writing it as a first order
matricial system, one may derive from the Lax pair a three-point linear equation in terms
of one of the components of the vector φ on which the Lax matrices act. In this way we
obtain the following scalar Lax pair:
be =
ϕ b+Λϕ ,
Qϕ (3.5.4a)
e =
ϕ b+ Rϕ ,
ϕ (3.5.4b)
remembering that QR = a.
Exercise 3.5.1. Derive the Lax pair (3.5.4) from the set of equations (3.3.2) (with γ =
γ ′ = 1) by setting ϕ = G and identifying the spectral variable by Λ = k 2 − q 2 .
Exercise 3.5.2. Show that the consistency condition of the Lax pair (3.5.4) viewed as an
overdetermined discrete linear system, leads to the lattice KdV equation (3.5.3).
As an example let us consider the lattice KdV equation (3.5.3), (setting for convenience
the parameter a = 1 w.l.o.g.)
1 1
Rn+1,m+1 − Rn,m = − .
Rn+1,m Rn,m+1
We will now study how the the initial data progresses, when one hits a singularity. The
initial data is given at the solid line on Figure 3.1, a, b, 0, c, d. When one proceeds from
these initial values one obtains infinity at two places, then one 0 at the next level and finally
two ambiguities of the type ∞ − ∞.
A more detailed analysis with the initial value 01 = ε (small) yields the following values
at the subsequent iterations
1 1 1 1
∞1 = b + − , ∞2 = c + − ,
ε a d ε
at the first stage, and on the next
1 1 1 1
s=a+ − , t=d+ − ,
∞1 f g ∞2
1 1 1 1 2
02 = ε + − = −ε + b − c + − ε + ...
∞2 ∞1 a d
b c
a 01 d
f ∞1 ∞2 g
s 02 t
?1 ?2
Figure 3.1: Propagation of singularities in a 2D map. Here n grows in the SE direction and
m in the SW direction.
3.6. SINGULARITY CONFINEMENT 65
1 1 1 1
?1 = ∞1 + − =c+ − + O(ε)
02 s d a − 1/f
1 1 1 1
?2 = ∞2 + − =b− + + O(ε)
t 02 a d + 1/g
1 λ
Rn+1,m+1 − Rn,m = −
Rn+1,m Rn,m+1
with λ 6= 1, the above fine cancellation would no longer happen, and singularities would
again occur at ?1 and ?2 , and would persist throughout! In that case the singularities are no
longer confined to a finite number of iteration steps, and we conclude that the corresponding
map is not integrable.
1
wn+1,m+1 = wn,m − ,
wn,m+1 − wn+1,m
w00 w20
w10
Literature
1. M.J. Ablowitz and F.J. Ladik, A nonlinear difference scheme and inverse scattering, Stud.
Appl. Math. 55 (1976) 213—229; On the solution of a class of nonlinear partial difference
equations, ibid. 57 (1977) 1–12.
2. R. Hirota, Nonlinear partial difference equations I-III, J. Phys. Soc. Japan 43 (1977), 1424–
1433, 2074–2089.
3. E. Date, M. Jimbo and T. Miwa, Method for generating discrete soliton equations I-V, J.
Phys. Soc. Japan 51 (1982) 4116–4131, 52 (1983) 388–393, 761–771.
4. B. Grammaticos, A. Ramani and V. Papageorgiou, Do integrable mappings have the Painlevé
property?, Phys. Rev. Lett. bf 67 (1991) 1825–1828.
5. F.W. Nijhoff, G.R.W. Quispel and H.W. Capel, Direct linearization of nonlinear difference-
difference equations, Phys. Lett. 97A (1983) 125–128; G.R.W. Quispel, F.W. Nijhoff, H.W.
Capel and J. van der Linden, Linear integral equations and nonliner difference-difference
equations, Physica 125A (1984) 344–380.
6. F.W. Nijhoff and H.W. Capel, The discrete Korteweg-de Vries equation, Acta Applicandae
Mathematicae 39 (1995) 133–158.
7. F.W. Nijhoff and A.J. Walker, The discrete and continuous Painlevé VI hierarchy and the
Garnier systems, Glasgow Math. J. 43A (2001) 109–123.
8. V.E. Adler, A.I. Bobenko and Yu.B. Suris, Classification of integrable equations on quad-
graphs, Commun. Math. Phys. 233 (2003) 513–543.
9. F.W. Nijhoff, Lax pair for the Adler (lattice Krichever-Novikov) system, Phys. Lett. 297A
(2002), 49–58.
10. J. Hietarinta, Searching for CAC-maps, J. Nonlin. Math. Phys. 12 Suppl. 2 (2005) 223–230.
11. F.W. Nijhoff, J. Atkinson and J. Hietarinta, Soliton Solutions for ABS Lattice Equations: I
Cauchy Matrix Approach, J. Phys. A: Math. Theor. 42 40 (2009) 404005 (34pp).
12. J. Atkinson and F.W. Nijhoff, A constructive approach to the soliton solutions of integrable
lattice equations, Commun. Math. Phys. 299 (2010) 283–304.
Lecture 4
Integrable nonlinear equations stand out among other differential or difference equations
by the fact that one can construct infinite classes of solutions in explicit form, i.e. exact
solutions which can be explicitely written down. This distinguishes them from most other
model equations that occur in the applied mathematical sciences.
The partial difference equations (P∆Es) that we have considered in Lecture 3 admit
several types of special solutions. Such solutions include:
• rational solutions;
• soliton solutions;
• periodic solutions;
• similarity solutions.
These different solutions require different techniques to obtain them, and to write them
down in terms of explicit formulae. For example, to obtain soliton solutions we can use
the inverse scattering transform method, or alternatively the technique of Bäcklund trans-
formations, whilst period solutions rely on a method called finite-gap integration, which
relies on techniques from algebraic geometry. In most cases the problem of obtaining such
special solutions amounts to the investigation of reductions of the P∆Es: the choice of the
class of solutions restricts the parameter-space from an infinite-dimensional space to a finite-
dimensional one. Effectively this means that the original P∆E is reduced to an ordinary
difference equation (O∆E) or a system of O∆Es in the process of obtaining the solution,
which could often be expressed through the imposition of additional compatible constraints.
The reduced equations, can then be studied in their own right, and they may take different
shapes, such as:
• dynamical mappings;
67
68 LECTURE 4. SPECIAL SOLUTIONS & CONTINUUM LIMITS
which then can subsequently solved from that perspective. Thus, we get a solution scheme
that looks like:
P∆E → O∆E system → explicit solution .
In this lecture we will concentrate on two types of special solutions: i) single soliton solutions
obtainable through BTs; ii) simple periodic solutions as dynamical mappings.
• auto-Bäcklund tranforms, i.e. nonlinear transformations that bring you from one so-
lution, of the P∆E under consideration, to another solution of the same equation;
In either case, one of the first problems we encounter is to find an initial solution, which
could be a trivial solution (like the zero solution, if the equation admits it), but sometimes
it might be problematic to find such a simple solution. Whenever we have such a solution,
we can use it as a ”seed solution” on which we can implement the BT. Here we will give
some examples, both in the case of a non-auto-BT, as well as in the case of an auto-BT,
how this works in the discrete case (the continuous case was already treated in Lecture 2).
e
−2ww = p2 + v + ve , (4.1.1a)
b
−2ww = q 2 + v + vb , (4.1.1b)
4.1. SOLUTIONS FROM DISCRETE BÄCKLUND TRANSFORMS 69
which was first given in [Atkinson, 2008]. By eliminating the variable w from both equations
(4.1.1a) and (4.1.1b), using shifts in the variables n and m on the lattice, it is easy to show
that v obeys the lattice equation
v − vb)(v − b
(e v ) = (p2 − q 2 ) v + ve + vb + b
e v + p2 + q 2 .
e (4.1.2)
Similarly, by eliminating v from the eqs. (4.1.1) it follows that w obeys the lattice potential
KdV equation
b − w)(w
(w e be = (p2 − q 2 ) .
− w) (4.1.3)
We can now take a known solution of (4.1.3), for instance the solution
and insert this into the relations (4.1.1). Thus, making explicit the dependence on the
variables n and m, we obtain from (4.1.1a)
where the latter equality follows from direct computation. The latter equation being an
O∆E of the form yn + yn+1 = 0 , where yn = vn,m + (w0 + np + mq)2 , we can directly
integrate to obtain a solution of the form yn = c(−1)n , with c a constant w.r.t. the variable
n. Thus, we obtain
vn,m = cm (−1)n − (w0 + np + mq)2 ,
where cm is constant w.r.t. n (but not necessarily w.r.t. m!) as a solution of the first relation
of the non-auto-BT (4.1.1). Subsequently, we have to satisfy also the second relation of the
non-auto-BT, but in this case, due to the symmetry of the relations as well as of the initial
solution, it is easy to see that the above expression for vn,m also yields a solution to the
second relation, i.e. (4.1.1b), provided we take cm = c(−1)m with now c to be constant also
w.r.t. the variable m Hence, we arrive at the solution of H2, (4.1.2), namely
Note that the solutions (4.1.4) and (4.1.5) can be extended to solutions on a multidimensional
lattice by assuming that the initial value w0 depends on additional variables associated with
additional lattice directions. Hence, we can etend the solutions easily to a 3-dimensional
lattice by setting
and it is automatic that this solution will obey copies of the equations (4.1.3) and (4.1.2)
on each 2-dimensional sublattice of the 3-dimensional lattice of variables n, m and h. Thus,
we could call the solution (4.1.6) to be a covariantly extended solution of the quadrilateral
lattice equation.
70 LECTURE 4. SPECIAL SOLUTIONS & CONTINUUM LIMITS
be − w)(w
(w b = p2 − q 2 ,
e − w) (4.1.7)
for which we have seen in Lecture 3 that it possesses the C.A.C. property. This means that
we can impose the same equation in any number of lattice directions of a multidmensional
lattice, each direction of which carries its own lattice parameter. Thus, we can impose on
the same function w of independent discrete variables the additional equations;
e − w)(w
(w e − w) = p2 − k 2 , (4.1.8a)
b b − w) =
(w − w)(w 2
q −k . 2
(4.1.8b)
In the same way that, once the inital solution w is given, the spatial part of the BT
(2.2.41) is a Riccati equation (and hence linearisable), in the discrete case both parts of
the discrete BT (4.1.8) are discrete Riccati equations. In a similar way as we have done
in subsection 2.2.4 we shall now show how to obtain the 1-soliton solution of (4.1.7) by
implementation of the discrete BT.
In connection with the C.A.C. property which establishes the multidimensional consistency
of the equation, the main thing to recognise now is that we need also multidimensional
consistency of the solution. This means that additional variables associated with other
lattice directions, may be hidden in the initial value ξ0 , in such a way that, for instance,
ξ 0 = ξ0 + k . Taking this into account, we can now search for a form of the solution wn,m of
the BT (4.1.8), which we assume to be in the form
fn,m
w n,m = w (0)
n,m + , (4.1.9)
gn,m
where the factorisation into the functions f and g wll be taken such that we get a linear set
of equations for these two functions (based on the fact that eqs. (4.1.8) is a set of discrete
4.1. SOLUTIONS FROM DISCRETE BÄCKLUND TRANSFORMS 71
and this can be split into the following (matricial) linear system
!
fe e (0) − w(0) ,
w 0 f p+k 0 f
= = ,
ge −1 , w (0)
e −w (0) g −1 p−k g
and a simlar computation holds for the equation for the vector (f, g)T in the other lattice
direction following from (4.1.8b). Thus, we obtain in both lattice directions the linear
matricial equations
φn+1,m = P φn,m , φn,m+1 = Qφn,m , (4.1.10)
with
p+k 0 q+k 0
P = , Q= .
−1 p−k −1 q−k
Solving this system would lead to the solution of the BT, and this can be done on the basis
of the following observations:
1. the two equations in (4.1.10) can be simultaneously solved, because in fact the two
b e
e = φ.
b
matrices P and Q commute, P Q = QP (check this!), which implies that φ
n m
Thus, the solution of the vector φ is simply given by φn,m = P Q φ0,0 in terms of
an initial vector φ0,0 .
2. crucially the matrices P and Q are lower triangular constant matrices, which implies
that it is easy to compute arbitrary powers of such matrices, following the general rule:
!
a 0 N aN , 0
A= ⇒ A = N
−bN .
c b c a a−b , bN
f0.0 /g0,0 and thus reconstituting (4.1.9). The end result, after some simple algebra, is the
(1)
solution wn,m ≡ wn,m given by
n m
ρn,m (w 0,0 − ξ0 − k) p+k q+k
w n,m = ξ0 +k+np+mq+ 1 , ρn,m ≡ ,
1 − 2k (w 0,0 − ξ0 − k)(ρn,m − 1) p−k q−k
(4.1.12)
and this is the 1-soliton solution of the lattice equation (4.1.7). Note that we see the return
of the plane-wave factors ρn,m which is the ingredient in this exact solution through which
the main dependence on the independent discrete variables n, m enters in the solution.
Exercise 4.1.2. Compare the 1-soliton solution (4.1.12) of the lattice potential KdV equa-
tion with the 1-soliton solution (2.2.43) of the continuous equation, by recalling that the
discrete plane wave factors ρ become continuous exponential functions after a continuum
limit, (see section 4.4).
Finally, the solution of the actual lattice kdV equation, which according to section 3.5 is
the equation (3.5.3) in terms of either Qn,m = wn,m −wn+1,m+1 or Rn,m = wn,m+1 −wn+1,m ,
cf. (3.5.2), can obviously be obtained directly from (4.1.12) in explicit terms.
Remark: Higher soliton solutions for the lattice equation (4.1.7) can be obtained by it-
erating this procedure, or more simply, by applying the lattice equation once again, but
now viewed as a permutability condition from three known solutions, namely the seed solu-
tion w(0) , the 1-soliton solution w(1) with parameter k = k1 and another 1-soliton solution
w(2) with prameter k = k2 , to yield a new 2-soliton solution w(12) (depending on the two
parameters k1 and k2 ) according to the diagram:
k1 w(1) k2
w(0) w(12)
k2 k1
w(2)
This procedure is, however, not the most effective one to obtain higher soliton solutions,
and alternative direct methods can yield these solutions in closed form, using special types
of determinants. (See, e.g., the results in [12]).
equation we are looking at? Thus, studying systematically continuum limits of given P∆Es
is of interest anyway. In doing this there a two important features to take on board: i)
Continuum limits are not unique – one can have more than one differential equation arising
from one and the same difference equation; ii) in performing a continuum limits one should
respect the solution structure – making sure that the continuum limit of the equations
reflects a limit on the solutions as well. To make sure this is the case, we will motivate the
limits by investigating first what happens to the corresponding linear equations associated
witht he nonlinear equations of interest.
h dy h2 d2 y
y(x + h) = y(x) + + + ··· (4.2.1)
1! dx 2! dx2
into the difference equations, and then by expanding power-by-power in the lattice param-
eter h that we obtain the transition from difference to differential equation. Typically the
equation that emerges as “the continuum limit” of the discrete equation is the coefficient of
the dominant term in this expansion as h → 0.
Performing this sequence of steps on a given lattice equation there are two questions to
answer, namely
• among the various parameters present in the lattice equation, how do we identify the
parameter (or combination of parameters) to take as the one tending to zero in order
to shrink the lattice?
• how do we determine the behaviour of the independent and dependent variables under
the limit on this chosen parameter?
In general a brute force or naive continuum limit may easily lead to a total collapse of the
equation, where the Taylor expansions applied to the equation lead to mismatch of orders,
and hence to a situation where the limit results in no equation at all (due to conflicting
constraints emerging from the expansion) or to a trivial equation in leading order in the
lattice parameter. To avoid this problem and to answer the questions above, it turns out it
is useful to first study the continuum limits of the linearised equation before we attack the
74 LECTURE 4. SPECIAL SOLUTIONS & CONTINUUM LIMITS
full nonlinear equation. We will, thus, first derive the linearised form of the lattice equations
under consideration and derive a special class of solutions of these linear equations to use
these as a guidance on how to take nontrivial and consistent limits.
By a linearisation of a nonlinear lattice equation such as (2.2.46) we mean the linear
equation obtained by expanding the dependent variable around a specific known solution
of the nonlinear equation and taking the dominant term in the expansion. The simplest
linearisations are obtained by taking a trivial solution, such as the zero solution (if it exists).
In the case of the example (2.2.46), taking w ≡ 0 is not allowed since it does not lead to a
solution of that equation, but we can modify the equation slightly by changing it into
It is easily verified that the linear lattice equation (4.2.3) obeys the consistency-around-the-
cube property of section 3.2 in the same way as the full nonlinear equation, and hence we can
consistently embed this equation in a higher dimensional lattice by writing the compatible
system:
ρ − ρe) = (p − q)(b
(p + q)(b ρe − ρ) , (4.2.4a)
(p + k)(ρ − ρe) = (p − k)(eρ − ρ) , (4.2.4b)
(q + k)(ρ − ρb) = (q − k)(bρ − ρ) , (4.2.4c)
where ¯ denotes the shift in the third direction associated with lattice parameter k. The
linear system (4.2.4) has many solutions, but we will fix a specific class of solution by
demanding that the specific variable k is associated with a shift ρ 7→ ρ such that ρ = 0.
(One way of thinking of what this means is that the solution ρ with this constraint is
k
obtained from an (inverse) Bäcklund transformation ρ 7→ ρ with seed solution ρ = 0.)
Solving the two relations (4.2.4b) and (4.2.4c) with ρ = 0:
p−k q−k
ρe = ρ , ρb = ρ,
p+k q+k
we obtain the solution: n m
p−k q−k
ρn,m = ρ0,0 , (4.2.5)
p+k q+k
with ρ0,0 some arbitrary initial value. It is straightforward to show by direct computation
that (4.2.5) is a solution of (4.2.3).
The solutions (4.2.5) we will refer to as lattice plane-wave factors and have already seen
how they appeared in section 4.1. Thus, not only do they play a role as approximated
4.2. CONTINUUM LIMITS 75
solutions (namely as solutions of the linearised version of the nonlinear lattice equation
(4.2.2)) but they also are the key ingredient in the exact solution of the full nonlinear
equation. It is these solutions that we will exploit in the next section to formulate the
precise limits on parameters and discrete variables to get nontrivial limiting equations.
we obtain:
m m m
q−k −2k ξ (−2k)
lim = lim 1 + = lim 1 + = e−2kξ ,
m→∞
q→∞ q + k m→∞
q→∞ q + k m→∞ m (1 − kξ/m)
m=ξq m=ξq
(4.2.7)
using the fact the extra term in the denominator within the brackets becomes negligeable
as m → ∞.
Let us now investigate the effect of this limit on the equations, e.g. the linear equation
(4.2.4a). The idea is to re-interpret the dependent variable as
m
ρ = ρn,m =: ρn (ξ) , ξ = ξ0 + , (4.2.8)
q
where ξ0 is some initial value. This means that any shift in the discrete variable m increments
in the argument of the function ρ by 1/q. The next thing is for 1/q small, to consider Taylor
series expansions of the form:
1 1 1 1 2
ρn,m+1 = ρn (ξ + ) = ρn (ξ) + ∂ξ ρn (ξ) + ∂ ρn (ξ) + · · · ,
q q 2 q2 ξ
This linear equation is a mixed form of differential equation (w.r.t variable ξ) and discrete
(w.r.t. variable n, and is hence a semi-discrete equation. By construction, it can be directly
verified that the form: n
p−k
ρn (ξ) = e−2kξ ρ0 (0) ,
p+k
is a solution of this equation.
Inspired by this result, let us now turn to the nonlinear equation for u, (4.2.2), and
perform exactly the same limit there. Thus, introducing in a similar was as for the linear
equation the reinterpretation of the discrete variables
m
un,m =: un (ξ) , ξ = ξ0 + , (4.2.10)
q
Expanding in powers of 1/q, noting that the dominant terms of order O(q 2 ) and order O(q)
cancel identically, we obtain as coefficient of the leading term of order O(1) the following
equation
∂ξ (un + un+1 ) = 2p(un+1 − un ) − (un+1 − un )2 , (4.2.11)
which is consequently the continuum limit of the lattice equation (4.2.2) under this limit as
q → ∞. Eq. (4.2.11) is a nonlinear D∆E, which is of first order in the derivative w.r.t. the
continuous variable ξ and of first order in the discrete variable n as well.
Remark: By comparing eq. (4.2.11) with eq. (2.2.41) we recognise that by the change of
variables x = 2ξ , p2 = 4λ2 , w
b − w = un+1 − un − p , ∂ξ un = 2wx , we recover the spatial
part of the BT of the KdV equation.
Other lattice equations: We can perform simlar limits on the other members of the
KdV family of lattice equations, namely on (3.2.5) amd (3.2.7), leading to
2
∂ξ (vn+1 vn ) = p(vn+1 − vn2 ) (4.2.12)
and
(∂ξ zn )(∂ξ zn+1 ) = p2 (zn − zn+1 )2 , (4.2.13)
respectively. We shall show in section 4.3 that all these D∆Es are integrable by virtue of
the existence of semi-discrete analogues of the Lax pairs in all cases.
4.2. CONTINUUM LIMITS 77
R=ρ e = ρe
R
b be b
ee b
R R = ρb R = ρe
Let us investigate what happens on the level of the plane-wave factors ρ given by (4.2.5)
with this change of variables (n, m) 7→ (N = n + m, m) . Rearranging factors we get:
n+m m N m
p−k q−k p+k p−k 2(q − p)k
ρn,m = ρ0,0 = 1+ ρ0,0 := RN,m .
p+k q+k p−k p+k (q + k)(p − k)
and focusing on what happens with the second factor in this limit we observe:
m m
2δk 2δk
lim 1 + = lim 1 +
m→∞
δ→0 (p + δ + k)(p − k) m→∞
δ→0 (p2 − k 2 ) + (p − k)δ
δm=τ δm=τ
m
τ 2k 2kτ
= lim 1 + = exp , (4.2.16)
m→∞ m (p2 − k 2 ) + (p − k)τ /m) p2 − k 2
78 LECTURE 4. SPECIAL SOLUTIONS & CONTINUUM LIMITS
using the fact the extra term in the denominator within the brackets becomes negligeable as
m → ∞. Thus, the limit (4.2.16) makes good sense on the level of the plane-wave factors.
To investigate what happens with the linear lattice equation (4.2.15) under the limit
(4.2.16), we set
R = RN,m =: RN (τ ) , τ = τ0 + mδ , (4.2.17)
(allowing for some constan background value τ0 of the continuous variable), and applying
the Taylor expansion:
1 2 2
RN,m+1 = RN (τ + δ) = RN (τ ) + δ ∂τ RN (τ ) + δ ∂τ RN (τ ) + · · · .
2
Inserting this into the eq. (4.2.15) we get:
1 2
(2p + δ) RN +1 + δ ṘN +1 + δ R̈N +1 + · · · − RN +1
2
1 2
= −δ RN +2 + δ ṘN +2 + δ R̈N +2 + · · · − RN ,
2
which in leading order O(δ) yields as dominant term in the expansion the equation:
2pṘN = RN −1 − RN +1 . (4.2.18)
U = UN,m =: UN (τ ) , τ = τ0 + mδ , (4.2.22)
we can perform a similar Taylor expansion on U as we did for R. Inserting this into (4.2.21)
we get
1
−δ + UN +1 + δ U̇N +1 + δ 2 ÜN +1 + · · · − UN +1 ×
2
1
× 2p + δ + UN − UN +2 + δ U̇N +2 + δ 2 ÜN +2 + · · · = −(2p + δ)δ ,
2
4.2. CONTINUUM LIMITS 79
or equivalently
2p
U̇N = 1 − . (4.2.23)
2p + UN −1 − UN +1
In spite of the fact that eq. (4.2.23) was derived starting from the same lattice equation,
it is quite a different nonlinear D∆E from (4.2.11), as is evident by inspecting the orders:
this equation is first order in the derivative w.r.t. the continuous variable τ , and second
order w.r.t. the discrete variable N , whereas (4.2.11) is first order in both the discrete and
continuous variables but the derivative w.r.t. ξ acts on both un and un+1 . We will assess in
section 4.3 that both equations (4.2.11) and (4.2.23) are integrable in the sense that there
exists an associated Lax pair in both cases.
Other examples: The process described above to obtain a mixed continuum limit we
refer to as the skew continuum limit of the lattice equation. It can be applied also to the
other members of the lattice KdV family, namely to eqs. (3.2.5) and (3.2.7). In those cases
we obtain
VN −1 − VN +1
p∂τ log VN = , (4.2.24)
VN −1 + VN +1
and
2 (ZN −1 − ZN )(ZN − ZN +1 )
ŻN = (4.2.25)
p ZN −1 − ZN +1
respectively.
have
N
p−k 2kτ 2kτ k k
exp = exp + N ln 1 − − ln 1 +
p+k p2 − k 2 p2 − k 2 p p
2kτ k2 k4 k 1 k3 1 k5
= exp 1 + 2 + 4 + · · · − 2N + + + ···
p2 p p p 3 p3 5 p5
τ N 3 τ 1 N 5 τ 1 N
= exp 2k − + 2k − + 2k − + ···
p2 p p4 3 p3 p6 5 p5
3
t+k5 t′ +···
−→ ekx+k .
This leads to the identification of the variables of the full continuum limit as the coefficients
of the various powers of k in the expansion of the exponents of the plane-wave factors,
namely
τ τ 1 ξ
x=2 −ξ , t=2 − , (4.2.26)
p2 p4 3 p2
in which as before ξ = N/p . Thus, this calculation suggests, first, that the solution (4.2.19)
can be identified as follows:
1 1 1 1
RN (τ ) := R(ξ, τ ) ⇒ RN ±1 (τ ) = R(ξ ± , τ ) = R ± Rξ + 2 Rξξ ± 3 Rξξξ + · · · ,
p p 2p 6p
R(ξ, τ ) := R(x, t) ,
∂R ∂R ∂x ∂R ∂t 2
= + = −2Rx − Rt , (4.2.27a)
∂ξ ∂x ∂ξ ∂t ∂ξ 3p2
∂R ∂R ∂x ∂R ∂t 2 2
= + = 2 Rx + Rt . (4.2.27b)
∂τ ∂x ∂τ ∂t ∂τ p p4
Since the two seps (i.e. Taylor expansion of the shift in N and change of variables (ξ, τ ) 7→
(x, t)) involve the parameter p, the expansion in powers of 1/p and the selection of the
dominant term can only be done after having done both of these steps. Thus, starting from
the semi-discrete equation (4.2.18) we get
2 1
2p∂τ R = − ∂ξ R − 3 ∂ξ3 R − · · · ⇒
p 3p
3
2 2 2 2 1 2
⇒ 2p ∂x + ∂t R = − −2∂x − ∂t R − −2∂ x − ∂t R − ··· ,
p2 p4 p 3p2 3p3 3p2
and we observe that while the leading term of order O(1/p) cancels identically, the next
term of order O(1/p3 ) yields precisely the equation:
Rt = Rxxx ,
4.2. CONTINUUM LIMITS 81
as expected.
Turning now to the nonlinear equation (4.2.23) and performing the same continuum limit
there, setting first
1 1 1 1
UN (τ ) := U (ξ, τ ) ⇒ UN ±1 (τ ) = U (ξ ± , τ ) = U ± U ξ + 2 U ξξ ± 3 U ξξξ + · · · ,
p p 2p 6p
and, second, to use similar expressions as (4.2.27) to implement the change of variables
(ξ, τ ) 7→ (x, t) to obtain:
" 3 #−1
2 2 1 2 1 2
∂x + 4 ∂t U = 1 − 1 − 2 −2∂x − 2 ∂t U − 4 −2∂x − 2 ∂t U − ···
p2 p p 3p 6p 3p
" 3 # 2
1 2 1 2 1 2
= − 2 −2∂x − 2 ∂t U + 4 −2∂x − 2 ∂t U + · · · − 2 −2∂x − 2 ∂t U + · · · − · · ·
p 3p 6p 3p p 3p
expanding the demoninator on the right-hand side in a power series. Again we observe
that the terms of order O(1/p) cancel identically, and that the next dominant order
O(1/p3 ) yields the equation:
Ut = Uxxx − 3Ux2 , (4.2.28)
which is the potential KdV equation (coinciding with (2.2.39) in terms of the variable w up
to a change of sign in the independent variables, i.e. after x 7→ −x, t 7→ −t we recover the
equation for w).
We see that we have come now full circle: we started out with the continuous KdV
equation, derived its Bäcklund transformations, which using the permutability theorem led
to the construction of a lattice of solutions, the relations between these solutions being
reinterpreted as a partial difference equation on the twodimensional lattice. As a dynamical
equation the latter was seen as a discretisation of some continuum equations, both semi-
discrete as well as fully discrete, and the full continuum limit now turns out to be the
equation we started out from, namely the KdV itself. Perhaps this is not so surprising,
but what is remarkable is that the equations at all levels are compatible with eachother:
the continuous and discrete equations can be imposed simultaneously on one and the same
variable u, and, in fact, the variable U of (2.2.39) is nothing else than the variable w of
Lecture # 2.
Remark: Performing similar continuum limits on the semi-discrete equations (4.2.24) and
(4.2.25) we recover the following fully continuous equations, namely
Vx Vxx
Vt = Vxxx − 3 , (4.2.29)
V
82 LECTURE 4. SPECIAL SOLUTIONS & CONTINUUM LIMITS
Literature
1. G.L. Lamb, Elements of Soliton Theory, (Wiley Interscience, 1980).
2. M.J. Ablowitz and H. Segur, Solitons and the Inverse Scattering Transform, (SIAM, 1982).
3. R.K. Dodd, J.C. Eilbeck, J.D. Gibbon and H.C. Morris, Solitons and Nonlinear Wave Equa-
tions, (Academic Press, 1982).
5. P.G. Drazin and R.S. Johnson, Solitons. An Introduction, (Cambridge Univ. Press, 1989).
6. T. Miwa, M. Jimbo and E. Date, Solitons, Differential Equations, Symmetries and Infinite
Dimensional Lie Algebras, (Cambridge University Press, 2000).
7. M.J. Ablowitz and P.A. Clarkson, Solitons, Nonlinear Evolution Equations and Inverse Scat-
tering, LMS Series (Cambridge Univ. Press, 1991).
8. R. Hirota, Exact solution of the Korteweg-de Vries equation for multiple collisions of solitons,
Phys. Rev. Lett. 27 (1971), 1192.
9. R. Hirota, Direct Method for finding exact solutins of nonlinear evolution equations, Lect.
Notes in Math. 515 (Springer Verlag, 1976), p. 40.
10. R. Hirota, Direct Methods in Soliton Theory, in: R.K. Bullough and P.J. Caudrey, eds.,
Solitons, (Springer Verlag, 1980).
11. J. Atkinson, Bäcklund transformations for integrable lattice equations, J. Phys. A: Math.
Theor. 41 (2008) 135202 (8 pp.).
12. J. Atkinson, J. Hietarinta and F.W. Nijhoff, Seed and soliton solutions for Adler’s lattice
equation, J. Phys. A: Math. Theor. 40 1 (2007) F1–F8.
13. J. Atkinson, J. Hietarinta and F.W. Nijhoff, Soliton solutions for Q3, J. Phys. A: Math
Theor. 41 14 (2008) 142001 (11pp)
14. V.G. Papageorgiou, F.W. Nijhoff and H.W. Capel, Integrable Mappings and Nonlinear Inte-
grable Lattice Equations, Physics Letters 147A (1990), 106–114.
15. H.W. Capel, F.W. Nijhoff, H.W. Capel, Complete Integrability of Lagrangian Mappings and
Lattices of KDV Type, Physics Letters 155A (1991), 377–387.
16. G.R.W. Quispel, H.W. Capel, V.G. Papageorgiou and F.W. Nijhoff, Integrable Mappings
derived from Soliton Equations, Physica 173A (1991), 243–266.
Contents
83
84 CONTENTS
3 Integrability of P∆Es 51
3.1 Quadrilateral P∆Es . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 51
3.2 Integrability and Consistency-around-the-Cube . . . . . . . . . . . . . . . . . 54
3.3 Lax pair for lattice potential KdV . . . . . . . . . . . . . . . . . . . . . . . . 57
3.4 ∗ Classification of Quadrilateral P∆Es . . . . . . . . . . . . . . . . . . . . . . 60
3.5 Lattice KdV Equation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 62
3.6 Singularity confinement . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 63