2.6.1 The linear case
We have seen several times now that the wave equation is in general re-written as a system of two first-order equations. For this reason it is fundamental to study systems of linear first-order ODEs. To this aim, we let u = (u1, u2, . . . , ud)T = u(x, t)∈ Rdbe a column vector satisfying:
A(x, t)∂u
∂t + B(x, t)∂u
∂x = C(x, t)u + d(x, t), (2.32)
where A, B, and C, are real n× n matrices and d ∈ Rd. We prescribe Cauchy data along the curve t = φ(x). In this first order case we prescribe values of u:
u(x, φ(x)) = f (x).
Differentiating the previous equation, we have:
ux+ φ0ut= f0, and substituting into (2.32):
(A− φ0B)ut = Cf + d− Bf0.
The curve φ is a characteristic curve if the derivatives of u cannot be found from the Cauchy data alone. In this case this amounts to saying that the matrix multiplying utis singular:
det (A− φ0B) = 0.
Noting that φ0 = dt/dx, we can write the following differential equation of order n:
det (Adx− Bdt) = 0,
called the characteristic differential equation.
(X, T ) t
x Pn . . .
P3 P2
P1
FIGURE 2.12: Characteristic curves obtained by solving (backwards) the differential equation (2.34) for each eigenvaluei starting from the point (X, T ).
For example, assuming matrix A invertible at t = 0, we can prescribe initial data on the x-axis:
u(x, 0) = f (x).
For small t, we can multiply equation (2.32) by A−1 to obtain:
ut+ ˜Bux = ˜Cf + ˜d,
where ˜B = A−1B, ˜C = A−1C, and ˜e = A−1d. The characteristic differential equation becomes then:
det
B˜−dx dtI
= 0, i = 1, . . . , n,
which states that the speed of propagation of the i-th wave is the i-th eigenvalue of matrix ˜B:
dxi/dt = λi(x, t), i = 1, . . . , n, (2.33)
which is a set of n differential equations. Note that, dx/dt = λi, i = 1, . . . , n are the eigenvalues of matrix ˜B(x, t). The system (2.32) is said to be hyperbolic if ˜B has a set of linearly independent (right) eigenvectors and corresponding real eigenvalues. In other words, ˜B is such that
Br˜ i = λiri, i = 1, . . . , n,
where the eigenvectors ri are all linearly independent (form a basis for Rdand depend continuously on x, t). In the hyperbolic case, matrix ˜B can be diagonalized as:
B = U ΛU˜ −1 = U ΛVT,
where matrices U and V contain the ordered (with respect to the eigenvalues λi) right and left eigen-vectors ri and li as columns and the diagonal matrix Λ contains the corresponding eigenvalues λi (see Appendix D.3). Since U−1 = VT we can transform system (2.32) in canonical form by setting u = U v, and thus v = VTu, called the “characteristic variables”:
vt+ Λvx = ˜Cv + ˜d,
with ˜C = VTCU − VTUt− VTBUx and ˜d = VTd. A few basic facts about linear algebra and left and right eigenvectors are reviewed in Appendix D. Also the initial conditions are transformed:
VTf (x) = g(x) for t = 0. This way we have achieved a complete decomposition of the system obtaining a set of first order differential equations for each component vi of the vector function v that are coupled only on the right hand side. The i-th equation takes the form:
dvi
Thus, we can use the method of characteristic to solve each equation obtaining a family of charac-teristics for each i = 1, . . . , n. The specific characteristic curve of the i-th family is obtained by
“back-tracking”, i.e., solving the differential equation starting from a generic point (X, T ) to obtain points Pi on the x-axis (Figure 2.12). We would like to remark that the (qualitative) curves on the left of point (X, T ) are relative to positive eigenvalues λi > 0, while the ones on the right correspond to negative λi. Indicating with x(t) = γi(t, X, T ) the backtracked characteristic curve of the i-th family, the general solution of (2.34) is obtained by integration along the curve γi:
vi(X, T ) = gi(γi(0, X, T )) +
We remark here that the main difference that we note with respect to the scalar case is that now the i-th wave (in the new variables vi) is obtained as a superposition of all other waves vj, j = 1, . . . , n traveling along different curves and with different characteristic speeds. In the integral above the domain of integration changes with the wave number i, and for this reason in the coefficients ˜cij and ˜di the substitution x(τ ) = γi(τ, X, T ) has been made where the curve γi is indexed with the wave number. We remark that, as seen in Section 2.2 for Burgers equation, the time interval where this solution is actually well defined cannot be too large, and we say that this solution is valid for sufficiently small T , usually without being more specific as done with Burgers equation.
In the previous paragraphs we have looked at a pure initial value problem, where initial data were prescribe at t = 0 on the entire x-axis. Now we want to consider the presence of a boundary. For this we assume that the equation is defined for x > 0, so that we need to impose a boundary condition at x = 0 for all times t > 0. However, we have to specify these boundary conditions only for those wave families corresponding to positive eigenvalues. For the other characteristic curves the domain of dependence does not include the t-axis (Figure 2.12). If we rank the eigenvalues from the
(X, T ) t
Pn x . . .
Pr
Pj
P1
FIGURE 2.13: Characteristic curves obtained by solving (backwards) the differential equation (2.34) for each eigenvaluei starting from the points (X, T ) and from the boundary point Pj.
smallest (λd) to the largest (λ1), i.e., λd < λd−1 < . . . < 0 < λr < λr−1 < . . . < λ1, we have the situation shown in Figure 2.13, where from point Pj a new set of characteristics has been identified by backtracking. We recall again that all these qualitative considerations work only in the case of smooth solutions, when u is continuous together with its first derivative. In more complex situations, as we have seen for the Burgers equation where shocks develop, we can still proceed with a similar analysis, but to do so we need to properly define how discontinuous solutions have to be interpreted.
Example 2.13. To exemplify the above development we address here the case of a system of first order PDEs with constant coefficients. Let u(x, t), f (x) ∈ Rd. Then our constant coefficient system takes the form:
ut+ Aux = 0, u(x, 0) = f (x), (2.35)
where matrix A is a hyperbolic matrix, i.e., it has real eigenvalues and eigenvectors and is diagonal-izable. We denote by ri, li, λi the (constant) right and left eigenvectors and the eigenvalues of A, respectively, and as before we collect the right eigenvectors as columns of the matrix U and the left eigenvectors as columns of the matrix V . Recalling (D.11), we can express the component of u using the basis U as ui =hli, ui. Then scalar multiplication of (2.35) by l1, . . . , ldyields equations for the i-th wave:
hli, uti + hli, Auxi = ui,t+ λiui,x = 0 ui(x, 0) =hli, f (x)i = fi(x).
We obtain a decoupled system of Cauchy problems that can be solved independently using the for-mula (2.3):
ui(x, t) = fi(x− λit).
t
x t
x FIGURE 2.14: Superposition of two linear waves (left) and complex interactions of two nonlinear waves (right).
Then the solution to our problem is given by:
u(x, t) =
n
X
i=1
fi(x− λit)ri. (2.36)
This can be seen easily by taking time and space derivatives to obtain:
ut(x, t) =
n
X
i=1
−λihli, fx(x− λit)i ri =−Aux(x, t).
Again, we note that the solution is the superposition of n waves each traveling with characteristic speed λi.
2.6.2 The quasi-linear case
We have seen that in a linear system of conservation laws the different waves do not interact and the solution u is a linear superposition of the different waves. On the other hand, if the system is nonlinear the different waves may interact in a complex manner. The situation is qualitatively shown in Figure 2.14, where the linear waves move undeformed along the characteristic lines, while nonlinear waves move along curves and their interaction may generate new waves.
That this is true can be seen from the following discussion [2]. Consider the quasi-linear system of conservation laws:
ut+ A(u)ux = 0.
The matrix A(u) is a function of u and so are its eigenvalues λi(u) and its left and right eigenvectors li(u) and ri(u). As usual we denote with U (u) = {r1(u), . . . , rd(u)} the matrix of the basis vectors formed with the right eigenvectors, with V (u) = {l1(u), . . . , ld(u)} the matrix of the dual basis, i.e.
the left eigenvectors, and with Λ(u) the diagonal matrix of the eigenvalues of A. Using (D.11), we can express the spatial gradient∇ u = uxas a linear combination of the right eigenvectors riof A as:
ut= (VTut)TU =
n
X
i=1
hli(u), uti ri =−Aux = A(VTux)TU = −
n
X
i=1
λi(u)hli(u), uti ri(u).
Differentiating ux with respect to t and ut with respect to x, equating the mixed derivatives, after some calculations (for details see [2]), we obtain:
(ux,i)t+ (λiux,i)x =
n
X
j=1,k=1 j>k
Gijk, (2.37)
with Gijk given by:
Gijk = (λj − λk)hli, [rj, rk]i ux,jux,k,
where [rj, rk] = Jrkrj− Jrjrk = [directional derivative of rkalong rj]- [directional derivative or rj
along rk] is called the Lie bracket. The quantityJrk denotes the Jacobian matrix of vector rj. There-fore,Jrkrjindicates the directional derivative of rkin the direction of rj. We note that equation (2.37) is in conservation form and decoupled on the left-hand-side. The source term, however, couples all the waves. In fact, Gijk can be interpreted as the amount of i-waves originating from the interaction of j-waves with k-waves [2]. The quantity λj − λk describes the wave speed difference, the term ux,jux,k the product of the densities of the j-waves and the k-waves, and hli, [rj, rk]i gives the i-th component of the Lie bracket in the U basis .