MIDTERM REVIEW AND SAMPLE EXAM Abstract. These notes outline the material for the upcoming exam. Note that the review is divided into the two main topics we have covered thus far, namely, ordinary differential equations and linear algebra with applications to solving linear systems Contents 1. Ordinary Differential Equations. 1.1. Reduction to a first-order system. 1.2. First-order systems of ODEs: existence and uniqueness theory. 1.3. Autonomous first-order systems: critical points, linearization, and stability analysis. 1.4. Solving ODEs: analytic methods (and numerical methods to be covered later) 2. Linear Algebra 2.1. Vectors and Matrices 2.2. Vector Spaces 2.3. Linear systems of equations. 3. Midterm Sample Exam 1 2 4 8 12 15 15 17 20 23 1. Ordinary Differential Equations. An nth-order ordinary differential equation in implicit form reads (1) F (t, y, y 0 , ..., y (n) ) = 0, F : Ωt × Ω → Rm , where y ∈ Rm , Ωt ⊆ R, Ω ⊆ Rm×(n+1) . Componentwise, the ODE system can be written as F1 (t, y, y 0 , ... , y (n) ) 0 F2 (t, y, y 0 , ... , y (n) ) 0 = .. . .. . . Fm (t, y, y 0 , ... , y (n) ) Date: Today is October 18, 2012. 1 0 2 MIDTERM REVIEW AND SAMPLE EXAM Equivalently, in explicit form the system y (n) = f (t, y, y 0 , ..., y (n−1) ) can be written in its component form as (n) y1 f1 (t, y, y 0 , ... , y (n−1) ) (n) y2 f2 (t, y, y 0 , ... , y (n−1) ) . = (2) .. . . . fm (t, y, y 0 , ... , y (n−1) ) (n) ym where now f : Ωt × Ω → Rm with y ∈ Rm , Ωt ⊆ R, example, take m = n = 2, then 00 y1 f1 (t, y1 , y2 , y10 , y20 ) = . 00 f2 (t, y1 , y2 , y10 , y20 ) y2 Ω ⊆ Rm×n . As an Definition 1.1 (Solutions of ODEs). The solution of a system of ODEs (2) is a function u ∈ Rm such that u(n) = f (t, u, u0 , ..., u(n−1) ). Remark 1.2 (Initial conditions). Generally, the solution of the system (2) admits infinitely many solutions, depending on the initial conditions. To obtain a unique solution, the initial conditions must be specified. Definition 1.3 (Initial Value Problem). An initial value problem (IVP) is a system of ODEs (2) together with appropriate initial conditions, y (i) (t0 ) = ci ∈ Rm , i = 1, ..., n − 1. 1.1. Reduction to a first-order system. Here, we show that general nth-order systems of ODEs are equivalent to first-oder systems of larger dimension. This, in turn, implies that we can work with first order systems almost exclusively in analyzing and solving general nonlinear problems. An nth-order system of ODEs (2) can be written as a first-order nonlinear system as follows. Let z1 = y and zi := z (i−1) , i = 2, ..., n. Then, the equivalent first-order system is z10 = z2 z20 = z3 .. . 0 zn−1 = zn zn0 = f (t, z1 , ..., zn ). Note that zi ∈ Rm for i = 1, ..., n. If we write the components of zi as zij , j = 1, , , ., m, then we can write the first-order system for z in terms of its m × n components as follows: 0 zij = zi+1,j , 0 znj = fj (t, z11 , ..., z1m , ..., zn1 , .., znm ), i = 1, ..., n, j = 1, ..., m, j = 1, .., m. MIDTERM REVIEW AND SAMPLE EXAM 3 Example 1.4. Take n = 2 and m = 1, then we have the scalar second-order equation y 00 = f (t, y, y 0 ), f : Ωt × Ω → R, Ω ⊂ R. Note that, if we set f (t, y, y 0 ) = − mk1 y, then we have the equation of a mass-spring system m1 y 00 + ky = 0. To write the equation as a first order system, we set z1 = y and z2 = z10 = y 0 such that z20 = y 00 and the second order ODE becomes z10 = z2 z20 = f (t, z1 , z2 ) = − k z1 . m1 Thus, we have the equivalent linear first-order system 0 1 z y 0 z = Az, A = , z= 1 = . k − m1 0 z2 y0 Example 1.5. Take n = m = 2. Then, as mentioned above, we have a system of two second-order ODEs 00 y1 f1 (t, y1 , y2 , y10 , y20 ) = . 00 f2 (t, y1 , y2 , y10 , y20 ) y2 We thus let z11 = y1 , z12 = y2 , z21 = y10 , and z22 = y20 and arrive at the first-order system 0 z11 0 z12 0 z21 0 z22 = = = = z21 , z22 , f1 (t, z11 , z12 , z21 , z22 ), f2 (t, z11 , z12 , z21 , z22 ). An example of such a system is given by the mass spring system of ODEs from the homework: m1 y100 = −k1 y1 + k2 (y2 − y1 ), m2 y200 = −k2 (y2 − y1 ), or k2 k1 y1 + (y2 − y1 ), m1 m1 k2 = − (y2 − y1 ). m2 y100 = − y200 4 MIDTERM REVIEW AND SAMPLE EXAM If we let m1 = m2 = 1, k1 = 3, and k2 = 2 we have y100 = −5y1 + 2y2 , y200 = 2y1 − 2y2 , giving f1 (t, y1 , y2 , y10 , y20 ) = −5y1 + 2y2 , f2 (t, y1 , y2 , y10 , y20 ) = 2y1 − 2y2 . Thus, the first-order system of ODEs for z becomes 0 z11 0 z12 0 z21 0 z22 = = = = z21 , z22 , −5z11 + 2z12 , 2z11 − 2z12 . Remark 1.6 (Reduction to first-order system). To summarize, we have shown that in general an nth-order system of ODEs is equivalent to a firstoder system of larger dimension. This, in turn, implies that we can work with first order systems almost exclusively in analyzing and solving ODEs.1 While in general we are only able to solve a small class of first-order systems analytically, numerical methods have been developed for a variety of first-order systems and provide tools for treating more general nonlinear problems. 1.2. First-order systems of ODEs: existence and uniqueness theory. Before reviewing various solution techniques, we give some definitions and state a general existence and uniqueness result for first-order systems of ODEs. We begin with the local existence and uniqueness result. Theorem 1.7 (Local Existence and Uniqueness). Consider a first-order IVP 0 y1 f1 (t, y1 , ..., yn ) .. .. .= , y(t0 ) = y0 . . 0 fn (t, y1 , ..., yn ) yn Assume that fi (), i = 1, ..., n, are continuous functions with continuous ∂fi (), i, j = 1, ..., n, in some domain Ωt × Ω such that partial derivatives ∂y j (t0 , y0 ) ∈ Ωt × Ω, where Ωt ⊆ R and Ω ⊆ Rn . Then, their exists a unique solution to the IVP for t0 − α < t < t0 + α, where α > 0. 1We note that in some special cases, (e.g., second-order ODEs) it is preferable to work directly with the given equation. As we have seen in the homework (see also the sample exam) the solution to second-order linear systems with constant coefficients can be found by substituting y = xeωt into the system of ODEs, which then reduces solving the system to solving an eigenvalue problem. MIDTERM REVIEW AND SAMPLE EXAM 5 Note that the result of the theorem is referred to as a local result since it only holds in some neighborhood about t0 . Example 1.8. Consider the first-order IVP y 0 = f (t, y), y(t0 ) = y0 . Then, we have the following three possibilities: ∂f • The IVP has a unique solution near (t0 , y0 ) if f and ∂f ∂t , and ∂y are continuous near (t0 , y0 ). As an example, consider the equation y0 = y2, y0 = 1, 1 which has the unique solution y = − t−1 for t < 1, but blows up as t → 1. • The IVP has a solution but it is not unique if only f is continu∂f ous near (t0 , y0 ) and not ∂f ∂t , and ∂y . As an example, consider the equation √ y 0 = y, y0 = 0, which has the solutions y = 0 and y = 14 t2 . • The IVP does not have a solution. As an example, consider the equation |y 0 | + |y| = 0, y(0) = c. 1.2.1. Linear first-order systems of ODEs. Linear first-order systems are an important class of ODEs since they are well understood theoretically. Moreover, the established techniques that have been developed for first-order linear problems can be used to analyze and solve more general nonlinear and higher-order problems. Definition 1.9 (Linear ODE systems). A first-order system of ODEs is linear if it can be written as y 0 = A(t)y + g(t), where aij = aij (t). Equivalently, we can write this system as y10 = a11 (t)y1 + ... + a1n (t)yn + g1 (t) .. . 0 yn = an1 (t)y1 + ... + ann (t)yn + gn (t) Recall that, the system is homogenous if g(t) = 0. Definition 1.10 (General Solution of nonhomogenous problem). The general solution to a nonhomogenous system of linear ODEs is y = y (h) + y (p) , where y (h) is a solution to the homogenous problem and y (p) is any solution to the nonhomogenous problem. 6 MIDTERM REVIEW AND SAMPLE EXAM Remark 1.11 (Principle of Superposition). The set of all solutions to the homogenous system, y 0 − Ay = 0, forms a vector space, which yields the so-called principle of superposition. That is, (1) y = 0 is a solution to the homogenous system and (2) if y (1) , y (2) are solutions to the homogenous system, then (αy (1) + βy (2) )0 − A(αy (1) + βy (2) ) = α((y (1) )0 − Ay (1) ) + β((y (2) )0 − Ay (2) ) = α0 + β0 = 0, where α, β ∈ R are arbitrary constants. Thus, applying this result n − 1 times gives the principle of superposition: if y (1) , ..., y (n) are solutions to the linear homogenous system, then so is y = c1 y (1) + c2 y (2) + ... + cn y (n) , where c1 , ..., cn are arbitrary constants. Definition 1.12 (General Solution of homogenous problem). The general solution to a homogenous system of linear ODEs is y = c1 y (1) + c2 y (2) + ... + cn y (n) , where y (1) , ..., y (n) form a basis (fundamental solution set) of the system. Definition 1.13 (Linear independence). The solutions (functions of t) y (1) , ..., y (n) are linearly independent if and only if (1) (2) (n) y1 (t) y1 (t) . . . y1 (t) (1) y2 (t) y2(2) (t) . . . y2(n) (t) 6= 0, W (t) = det . .. (1) (2) (n) yn (t) yn (t) . . . yn (t) for some t ∈ I. Example 1.14. For a 2 × 2 system (1) (2) y (t) y1 (t) W (t) = det 1(1) (2) y2 (t) y2 (t) ! which is the same as the Wronskian for a second-order linear ODE since, as seen in Example 1.4, this equation can be reduced to a 2 × 2 system. This follows since the solutions to the 2 × 2 system are y (1) = y and y (2) = y 0 . Remark 1.15. Given the superposition principle for homogenous problems, it follows that y = y (h) + y (p) satisfies the nonhomogenous problem: (y (h) + y (p) )0 − A(y (h) + y (p) ) = y (h) − Ay (h) + y (p) − Ay (p) = 0 + g. MIDTERM REVIEW AND SAMPLE EXAM 7 Remark 1.16 (Existence and Uniqueness: Linear ODE Systems). The existence and uniqueness theorem simplifies in the case of a first-order linear system of ODEs. Here we note that ∂fi = aij (t), ∂yj i, j = 1, ..., n. The theorem thus reduces to the following result. Theorem 1.17. Let aij (t), i, j = 1, ..., n, be continuous functions of t on some interval I = (α, β), with t0 ∈ I. Then there exists a unique solution to the IVP on the interval I. Example 1.18. Consider the scalar fist-order linear IVP: y 0 + p(t)y = q(t), y(t0 ) = c0 . By the theorem, if p(t) and q(t) are continuous on an interval I containing t0 , then, there exists a unique solution to the IVP on I. Example 1.19. Consider the scalar second-order linear IVP: y 00 + p(t)y 0 + q(t)y = g(t), y(t0 ) = c1 , y( t0 )0 = c2 . This problem can be reduced to a first-order system and thus by the theorem, if p(t), q(t), and g(t) are continuous on an interval I containing t0 , then, there exists a unique solution to the IVP on I. Remark 1.20 (Linear ODE). There are various interpretations of the equation y 0 + p(t)y = q(t), y(t0 ) = y0 that lead to some additional useful insights: • Physically, we can interpret the equation as prescribing the velocity of a point particle moving on a line in time, t. • Geometrically, we can interpret the equation as specifying the slope of the graph of a function y(t). It the slope is plotted pointwise as a vector field (direction field), then the solution curves must be tangent to the direction field. Note that the slope is constant along curves f (t, y) = c, called the isoclines. • We can write the solution to the equation explicitly as: y = y (h) + y (p) , where y (h) solves the homogenous problem y 0 + p(t)y = 0 and y (p) is any solution to the nonhomogenous problem y 0 + p(t)y = q(t). 8 MIDTERM REVIEW AND SAMPLE EXAM This follows because the equation is linear: y 0 + p(t)y = (y (h) + y (p) )0 + p(t)(y (h) + y (p) ) = (y (h) )0 + p(t)y (h) + (y (p) )0 + p(t)y (p) = 0 + q(t). We solve the homogenous problem using separation of variables, which has the general solution y (h) (t) = ce R −p(t)dt . To find a particular solution R of the nonhomogenous equation we use an integrating factor µ = e −p(t) and obtain Z R (p) y = e p(t) q(t)dt. Thus, the general solution of the nonhomogenous problem is Z R R (h) (p) −p(t)dt y = y + y = ce + e p(t) q(t)dt. 1.3. Autonomous first-order systems: critical points, linearization, and stability analysis. An autonomous first-order system of nonlinear ODEs can be linearized using Taylor expansion and under appropriate assumptions the type and stability of the critical points of the nonlinear system can be analyzed using the resulting linear system. The following discussion summarizes this result. Definition 1.21 (Autonomous first-order system). An autonomous nonlinear first-order system is given by y 0 = f (y), where the right hand side f does not depend explicitly on t. 1.3.1. Critical points via the Phase Plane method and linearization. Definition 1.22 (Critical points). The critical points, yc , of the autonomous nonlinear system y 0 = f (y) are points for which f is undefined or that satisfy f (yc ) = 0. As shown in the sample exam, we can assume that yc = 0. Applying Taylor expansion near the critical point we have ∂f1 ∂f1 ··· ∂y1 ∂yn . . . .. .. . f (y) = f (0) + Jf (0)(y − 0) + H.O.T., Jf (0) = .. ∂fn ∂fn ··· ∂y1 ∂yn 0 Now, since f (0) = 0, if we let A = Jf (0) and let h(y) = H.O.T., then we can write the autonomous system as y 0 = f (y) = Ay + h(y). MIDTERM REVIEW AND SAMPLE EXAM 9 If we drop the function h(y) we obtain the linearized system such that near the origin y 0 ≈ Ay. We have the following result concerning the use of this approximation. Theorem 1.23 (Linearization). Consider the autonomous first-order system y 0 = f (y). If fi , i = 1, ..., n, are continuous and have continuous partial derivatives in a neighborhood of the critical point, yc , and det A 6= 0, then the kind and stability of the critical points of the nonlinear system are the same as the system y 0 = Ay obtained by linearization. We note that exceptions occur when the eigenvalues of A are purely imaginary. This result requires analysis of the critical points of the linearized system y 0 = Ay, which we review next for n = 2. We note that similar analysis can be conducted for general n × n systems. Remark 1.24. We note that in our analysis we use the Phase Plane method, in which we consider the components of the solution y1 (t) and y2 (t) as defining parametric curves in the y1 y2 -plane (the phase plane). If we plot all such trajectories for a given ODE system, then we obtain the phase portrait. Note that y1 = y2 = 0 is a critical point of the system since the slope of the trajectory at the critical point is undefined: dy2 y0 a21 y1 + a22 y2 0 = 20 = = . dy1 y1 a11 y1 + a12 y2 0 1.3.2. Classification of critical points. To determine the type of each critical point we compute the eigenpairs (λ, x) of A to find the general solution to the homogenous system y 0 = Ay and then study the behavior as t → ±∞. There are a total of 4 cases to consider. Examples of finding the solution to such systems for the case of a center and saddle point are provided in the sample exam. (1) Node: λ1 , λ2 ∈ R and λ1 λ2 > 0. We call the node proper if all the trajectoreis have a distinct tangent at the origin. In this case we have λ1 = λ2 . The node is improper if all trajectories have same tangent at the origin, except for two of them. In this case, λ1 6= λ2 . The node is degenerate if A has only a single eigenvector. In this case, we solve for the first eigenvector x and then solve for the generalized eigenvector u, by solving the system (A − λI)u = x. Note that, the eigenvectors are linearly independent provided A is symmetric or skew symmetric. (2) Saddle point: λ1 , λ2 ∈ R and λ1 λ2 < 0. In this case, we have two incoming and two outgoing trajectories, all others miss the origin. (3) Center: λ1 , λ2 ∈ C and λ1 = iµ, λ2 = −iµ. The trajectories are closed around the origin. (4) Spiral: λ1 , λ2 ∈ C and real(λi ) 6= 0, i = 1, 2. Here the trajectories spiral to or from the origin. 10 MIDTERM REVIEW AND SAMPLE EXAM Remark 1.25. The eigenvalues are the roots of the characteristic polynomial of a11 a12 A= . a21 a22 That is, the eigenvalues λ, satisfy a11 − λ a12 det(A − λI) = det a21 a22 − λ = (a11 − λ)(a22 − λ) − a21 a12 = λ2 − trace(A)λ + det(A) = 0, where trace(A) = a11 + a22 and det(A) = a11 a22 − a21 a12 . Now, the roots λ1 and λ2 satisfy (λ − λ1 )(λ − λ2 ) = λ2 − (λ1 + λ2 )λ + λ1 λ2 = 0. Hence, we have that trace(A) = λ1 + λ2 and det(A) = λ1 λ2 , where √ √ 1 1 λ1 = (p + ∆) and λ2 = (p − ∆), 2 2 with p = trace(A), q = det(A), and ∆. The type of the critical point can be categorized according to the quantities p = trace(A), q = det(A) and ∆ = p2 − 4q. The following table summarizes this classification. q = λ1 λ2 ∆ = (λ1 − λ2 )2 Node q>0 ∆≥0 Saddle point q<0 Real opposite sign q>0 pure imaginary Type p = λ1 + λ2 Center p=0 Spiral p 6= 0 ∆<0 Eigenvlaues Real same sign Complex not pure imaginary Table 1. Eigenvalue criteria for critical points. 1.3.3. Stability analysis for 2 × 2 autonomous systems. The stability of critical (fixed points of a system of constant coefficient linear autonomous differential equations of first order can be analyzed using the eigenvalues of the corresponding matrix A. Definition 1.26. The autonomous system has a constant solution, an equilibrium point of the corresponding dynamical system. This solution is (1) asymptotically stable as t → ∞, (”in the future”) if and only if for all eigenvalues λ of A, real(λ) < 0. (2) asymptotically stable as t → −∞ (”in the past”) if and only if for all eigenvalues λ of A, real(λ) > 0. MIDTERM REVIEW AND SAMPLE EXAM 11 Figure 1. Classification of equilibrium points of a linear autonomous system. These profiles also arise for non-linear autonomous systems in linearized approximations. (3) unstable if there exists an eigenvalue λ of A with real(λ) > 0 for t → ∞. The stability of a critical point can also be categorized according to the values of p = trace(A), q = det(A), and ∆ = p2 − 4q. The following table summarizes the classification. Type of Stability Asymptotically stable Stable Unstable p = λ1 + λ2 q = λ1 λ2 p<0 p≤0 p>0 q>0 q>0 or q < 0 Table 2. Criteria for stability. 12 MIDTERM REVIEW AND SAMPLE EXAM 1.4. Solving ODEs: analytic methods (and numerical methods to be covered later). Here, we describe various techniques for solving ODEs. We begin by reviewing methods for solving a single ODE, or scalar differential equation. We then proceed to solving systems of ODEs. A list of the techniques we covered in this course is as follows: • Linear first-order ODE: separation of variables. • Non-linear first-order ODE: exact equations and integrating factors, linearization, and reduction to linear form. • Linear first-order constant coefficient (systems) of ODE(s): the general solution of the nonhomogenous problems, the homogenous solution and the eigenproblem, and the particular solution and the methods of undetermined coefficients and variation of parameters. • Numerical methods for solving ODEs: Euler’s method as a simple example. 1.4.1. Non-linear first-order ODE. Some nonlinear ODEs can be reduced to a linear ODE, for example, the first-order Bernoulli equation y 0 + p(t)y = g(t)y α , α ∈ R. If we take u(t) = [y(t)]1−α , then u0 (t) = (1 − α)y(t)−α y 0 . Substituting into the ODE gives u0 (t) = (1 − α)y(t)−α y 0 = (1 − α)y(t)−α (gy α − py) = (1 − α)(g − pu) or u0 + (1 − α)pu = (1 − α)g, which is a first-order linear system for u. An important example of the Bernoulli equation results when we set α = 2, p(t) = −A, and g(t) = −B in which case we have y 0 = Ay − By 2 . The equation for u is then u0 (t) + Au = B, which has solution u(t) = ce−At + B/A, implying y= 1 1 = −At . u ce + B/A MIDTERM REVIEW AND SAMPLE EXAM 13 1.4.2. Linear second-order ODE: the general solution. Here, we consider the linear second order IVP y 00 + p(t)y 0 + q(t)y = g, y(t0 ) = c1 , y( t0 )0 = c2 . We assume that p, q, g are continuous functions on some interval I containing t0 such that there exists a unique solution y. Remark 1.27. The solution set of the homogenous problem forms a vector space, that is, (1) y = 0 is a solution and, (2) if α, β ∈ R and x and y are solutions to the homogenous problem, then αx + βy is a solution: (αx + βy)00 + p(αx + βy)0 + q(αx + βy) = α(x00 + px0 + qx) + β(y 00 + py 0 + qy) = α · 0 + β · 0 = 0. Remark 1.28 (General solution). All solutions of the homogenous problem can be written as y = c1 y (1) + c2 y (2) , where c1 , c2 ∈ R are arbitrary constants that are uniquely determined by the initial conditions, provided that the solutions y (1) and y (2) form a basis or fundamental system. This, in turn, holds true if and only if y (1) and y (2) are linearly independent, or the Wronskian W (t) = (y (2) (t))0 y (1) (t) − y (2) (t)(y (1) (t))0 6= 0. We note that it is sufficient to check that this condition holds for any value of t. Example 1.29 (Constant coefficients). Consider the case of a homogenous second-order linear constant coefficient ODE: y 00 + ay 0 + by = 0. Then, we derive the solution by setting y = eλt into the ODE, which gives (after canceling the common exponential term) the characteristic (quadratic) polynomial: λ2 + aλ + b = 0, whose two roots, λ1 , λ2 give us a solution of the form: y = c1 eλ1 t + c2 eλ2 t . There are three possible cases for the roots √ √ a a2 − 4b2 a a2 − 4b λ1 = − + λ2 = − − . 2 , 2 2 (1) Two distinct real roots: λ1 6= λ2 ∈ R. In this case, the solution is y = c1 eλ1 t + c2 eλ2 t . (2) Double real roots: λ1 = λ2 = − a2 ∈ R. In this case, the solution is a a y = c1 e− 2 t + c2 te− 2 t . 14 MIDTERM REVIEW AND SAMPLE EXAM (3) Two complex conjugate roots: λ1 = − a2 + iµ, λ2 = − a2 − iµ. In this √ case, µ = 4b − a2 > 0 and the solution is a y = e− 2 t (c1 cos(µt) + c2 sin(µt)). Here, Euler’s formula was used: ea+ib = ea eib = ea (cos b + i sin b). Example 1.30 (Euler-Cauchy equation). Another important second-order linear ODE is the Euler-Cauchy equation: t2 y 00 + aty 0 + by = 0. This system can be reduces to a constant coefficient problem in x by substituting y = tr , since then t0 = rt(r−1) and t00 = r(r − 1)t(r−2) , implying that r(r − 1) + ar + b = r2 + (a − 1)r + b = 0. The roots r1 and r2 of this quadratic polynomial give the solutions to the system: y = c1 tr1 + c2 tr2 . Given the solution to the homogenous problem, one can find the general solution to the corresponding nonhomogenous ODE using various techniques, for example the methods of undetermined coefficients and variation of parameters. We review these techniques for ODE systems in the next section, noting that they can also be applied in the case of a scalar equation in a similar way. 1.4.3. Systems of linear constant coefficient ODEs. Here, we consider solving constant coefficient linear ODE systems y 0 = Ay + g(t), A ∈ Rn×n , y ∈ Rn . As discussed in Section 1.2, we solve for the general solution of the nonhomogenous system y = y (h) + y (p) , by first computing the solution for y (h) , the solution of the homogenous problem, and then using the methods of undetermined coefficients or variation of parameters to find y (p) , the particular solution. Examples of how to use the latter methods to find y (p) are found in the sample exam. The general solution of the homogenous system is given by (see, Definition 1.12) y = c1 x(1) eλ1 t + c2 x(2) eλ2 t + ... + cn x(n) eλn t , where λ1 , λ2 , ..., λn are the eigenvalues of A, i.e, the roots of the characteristic polynomial det(A − λI) (a polynomial of degree n in λ), and x(1) , ..., x(n) the corresponding eigenvectors. We note that if the λi , i = 1, .., n, are distinct, then one can show that the eigenvectors are linearly independent. MIDTERM REVIEW AND SAMPLE EXAM 15 2. Linear Algebra We first state basic definitions and axioms for vectors and matrices. Then, we review some related concepts from Linear Algebra and apply these ideas to the solution of linear systems. 2.1. Vectors and Matrices. We consider rectangular matrices a11 . . . a1n .. ∈ Rm×n . .. A = ... . . am1 . . . amn Note that when n = 1 we obtain a column vector x1 .. x = . ∈ Rm xm and when m = 1 we obtain a row vector y = y1 . . . ym ∈ Rm . The basic operations of addition and multiplication among constants α ∈ R, vectors x ∈ Rn , and matrices A ∈ Rm×n are as follows: • Addition of two matrices C = A + B ∈ Rm×n is defined elementwise such that cij = ai,j + bi,j and results in the matrix a11 . . . a1n b11 . . . b1n a11 + b11 . . . a1n + b1n .. + .. .. = .. .. .. .. .. A+B = ... . . . . . . . . . am1 . . . amn bm1 . . . bmn am1 + bm1 . . . amn + bmn • Multiplication of a matrix, A, by a constant, α, is defined elementwise and results in the matrix: αA = [αaij ], i = 1, ..., m, j = 1, , , ., , , n. • Multiplication of a vector, x, by a matrix, A, results in the vector: Pn a1j xj Pj=1 n a2j xj j=1 Ax = ∈ Rm . .. . Pn j=1 amj xj • Multiplication of a matrix in the matrix Pn l=1 a1l bl1 .. AB = Pn . l=1 aml bl1 A ∈ Rm×n by a matrix B ∈ Rn×k results Pn ... l=1 a1l blk .. m×k .. . ∈R . Pn . ... l=1 aml blk Note that in general AB 6= BA. 16 MIDTERM REVIEW AND SAMPLE EXAM Given the matrix a11 a21 A = .. . a12 a22 .. . a1n a2n .. , . ... ... .. . am1 am2 . . . amn its traspose is defined as a11 a12 AT = .. . a21 a22 .. . a1n a2n . . . am1 . . . am2 .. . .. . . . . . amn Note that (AB)T = B T AT . There are several classes of matrices that arise often in parctice: (1) A square matrix, D, is diagonal if dij = 0 for i 6= j: D= d11 0 0 .. . d22 .. . ... 0 ... .. . .. . 0 0 .. . . 0 dnn (2) The identity matrix is a diagonal matrix where all the diagonal elements are equal to one: ... 0 . . .. . . 0 1 I = . . . . .. . . . . 0 0 ... 0 1 1 0 (3) A square matrix is symmetric if AT = A or aij = aji , i, j = 1, ..., n. (4) A square matrix is skew symmetric if AT = −A or aij = −aji , i, j = 1, ..., n. Note that skew symmetric matrices have a zero diagonal aii = −aii = 0, i = 1, ..., n. (5) An upper triangular matrix, U , is defined as uij = 0, i > j: u11 u12 . . . u1n 0 u22 . . . u2n U = .. .. . .. .. . . . . 0 ... 0 unn MIDTERM REVIEW AND SAMPLE EXAM (6) A lower triangular matrix, l11 l L = 21 ... ln1 17 L, is defined as lij = 0, j > i: 0 ... 0 . . l22 . . .. . .. . . . 0 . ln2 . . . lnn Definition 2.1. The inverse of a square matrix, A, is denoted by A−1 and satisfies: A−1 A = AA−1 = I. Example 2.2. Consider the case where n = 2 such that: a11 a12 A= . a21 a22 The inverse of A is A −1 1 = det (A) a22 −a12 , −a21 a11 where det(A) = a11 a22 − a21 a12 is the determinant of the matrix. This implies that A−1 exists if and only if det(A) 6= 0. To check that this is indeed the inverse of the matrix of A we compute 1 a22 −a12 a11 a12 −1 A A = a21 a22 det (A) −a21 a11 1 a11 a12 a22 −a12 = −a21 a11 det (A) a21 a22 1 0 = . 0 1 Theorem 2.3. The result in the example for n = 2 also holds for general matrices A ∈ Rn×n , that is, A−1 exists if and only if det (A) 6= 0. 2.2. Vector Spaces. Definition 2.4. A vector space, V, is a mathematical structure formed by a collection of elements called vectors, which may be added together and multiplied (”scaled”) by numbers, called scalars. Note that the elements of a vector space need not be vectors v ∈ Rm , they can also be functions, matrices, etc. The operations of vector addition and scalar multiplication must satisfy certain requirements, called axioms, listed below. Let u, v, w ∈ V be arbitrary vectors and let α, β ∈ K be arbitrary scalars. (1) Associativity of addition u + (v + w) = (u + v) + w. (2) Commutativity of addition: u + v = v + u. (3) Identity element of addition: there exists an element 0 ∈ V, called the zero vector, such that v + 0 = v for all v ∈ V. 18 MIDTERM REVIEW AND SAMPLE EXAM (4) Inverse elements of addition: for every v ∈ V, there exists an element −v ∈ V, called the additive inverse of v, such that v + (−v) = 0. (5) Distributivity of scalar multiplication with respect to vector addition: α(u + v) = αu + αv. (6) Distributivity of scalar multiplication with respect to field addition: (α + β)v = αv + βv. (7) Compatibility of scalar multiplication: α(βv) = (αβ)v. (8) Identity element of scalar multiplication: 1v = v, where 1 denotes the multiplicative identity in the field K. The requirement that vector addition and scalar multiplication be (external) binary operations includes (by definition of binary operations) a property called closure: that u + v and αv are in V for all α ∈ K, and u, v ∈ V. This follows since a binary operation on a set is a calculation involving two elements of the set (called operands) and producing another element of the set, in this case V. For the field K, the notion of an external binary operation is needed, which is defined as a binary function from K ×S to S. P Examples of vector spaces are V = Rm and V = {p(x)|p(x) = i αi xi }. The latter polynomial space is infinite dimensional, whereas the former Euclidian space is finite dimensional. Definition 2.5. A nonempty subset W of a vector space V that is closed under addition and scalar multiplication (and therefore contains the 0-vector of V) is called a subspace of V. Remark 2.6. To prove that W is a subspace, it is sufficient to prove (1) 0 ∈ W and (2) for any u, v ∈ W and any α, βK, αu + βv ∈ W. Definition 2.7. If S = {v1 , ...vn } is a finite subset of elements of a vector space V, then the span is n X span(S) = u|u = αi vi , αi ∈ K . i=1 The span of S may also be defined as the set of all linear combinations of the elements of S, which follows from the above definition. Definition 2.8. A basis B of a vector space V over a field K is a linearly independent subset of V that spans V. In more detail, suppose that B = {v1 , ..., vn } is a finite subset of a vector space V over a field K. Then, B is a basis if it satisfies the following conditions: (1) The linear independence property: if α1 v1 + ... + αn vn = 0, then α1 = ... = αn = 0; and (2) The spanning property: for every v ∈ V it is possible to choose α1 , ..., αn ∈ K such that v = α1 v1 + ... + αn vn . MIDTERM REVIEW AND SAMPLE EXAM 19 Definition 2.9. The dimension of a vector space V is dim V = |B|, where B is a basis for V. Remark 2.10. We proved the following useful results in class: (1) The coefficients αi are called the coordinates of the vector v with respect to the basis B, and by the first property they are uniquely determined. (2) Given a vector space V with dim V = n, any linearly independent set of n vectors forms a basis of V and any collection of n + 1 vectors is linearly dependent. Thus, the dimension is the maximum number of linearly independent vectors in V. The above results all follow from the following basic result. Lemma 2.11. Let V be a vector space. Assume that the set of vectors V = {v1 , ..., vn } spans V and that the set of vectors W = {w1 , ..., wm } is linear independent. Then, m ≤ n and a set of the form {w1 , ..., wm , vi1 , ..., vin−m } spans V . Proof. Assume vi 6= 0 for some i so that V = 6 {0}, else wi ∈ V can not be linearly independent for any m ≥ 1. Since V spans V it follows that P w1 = i αi vi and since w1 6= 0, we have that αj 6= 0 for some j. Thus X 1 vj = w1 − αi vi , αj i6=j implying that the set {w1 , v1 , ..., vj−1 , vj+1 , ..., vn } spans V. Repeating this argument, since this updated set spans V X w2 = βw1 + αi vi 6= 0, i6=j and thus {w1 , w2 , v1 , ..., vj−1 , vj+1 , ..., vn } must be linearly dependent. Now, since w1 and w2 are linearly independent, it must be that αk 6= 0 for some k 6= j (else w2 = βw1 ). Thus, the set {w1 , w2 , v1 , ..., vj−1 , vj+1 , ..., vk−1 , vk+1 , ..., vn } spans V. Repeating the same argument another m-2 times gives that {w1 , ..., wm , vi1 , ..., vin−m } spans V. Next, assume m > n. Then after n steps {w1 , ..., wn } 20 MIDTERM REVIEW AND SAMPLE EXAM spans V and after n + 1 steps {w1 , ..., wn , wn+1 } is linearly dependent, a contradiction. Thus, m ≤ n. 2.3. Linear systems of equations. Here, we consider the matrix equation Ax = b, A ∈ Rm×n , b ∈ Rm , and x ∈ Rn , which represents a system of m linear equations with n unknowns: a1,1 x1 + a1,2 x2 + ... + a1,n xn = b1 , a2,1 x1 + a2,2 x2 + ... + a2,n xn = b2 , .. . am,1 x1 + am,2 x2 + ... + am,n xn = bn , where, x1 , ..., xn are the unknowns, ai,j , 1 ≤ i, j ≤ n are given coefficients and b1 , ..., bn are constants, imposing constraints that these equations must satisfy. Note that the unknowns appear in each of the equations and as such we must find their values such that all equations are satisfied simultaneously. There are n columns of A (vectors) denoted as a1j a2j (j) a = .. , j = 1, ..., n, . amj and m row vectors a(i) = ai1 ai2 . . . ain , i = 1, ..., m. Definition 2.12. The column space, a vector space, is defined as colsp(A) = span(a(1) , ..., a(n) and the row space, also a vector space, is defined as rowsp(A) = span(a(1) , ..., a(m) . Definition 2.13. The maximum number of linearly independent rows in a matrix A is called the row rank of A, which is equal to the maximum number of linearly independent columns in A, referred to ass the column rank. Since the two are equal, we will not distinguish between them and write rank(A) to denote both. Remark 2.14. Note that from our discussion of vector spaces we have that rank(A) = dim colsp(A) = dim rowsp(A). Finally, we proved the following existence and uniqueness theorem for this linear system. Theorem 2.15. Consider the linear system Ax = b, A ∈ Rm×n , b ∈ Rm , . and x ∈ Rn and let A˜ = [A..b]. Then, the system ˜ = rank(A); (1) is consistent (the solution exists) if rank(A) ˜ (2) has a unique solution if rank(A) = rank(A) = n; MIDTERM REVIEW AND SAMPLE EXAM 21 ˜ = rank(A) < n. (3) has infinitely many solutions if rank(A) Proof. The first two parts are proved in the sample exam. Thus, we prove ˜ = rank(A) = r < n. Then only the last statement here. Assume rank(A) note that n X b= xi a(i) , i=1 where there are r linearly independent columns in A and then n − r which are linearly combinations of these r columns. We reorder the columns of the ˆ which has as its first r columns those columns from matrix A to obtain A, A that are linearly independent: r n X X b= x ˆi a ˆ(i) + x ˆi a ˆ(i) . i=1 i=r+1 Note that these first r columns form a basis, B = {ˆ a(1) , ..., a ˆ(r) } for colsp(A) = ˆ Now, since the last n − r columns of Aˆ can be written as linear colsp(A). combinations of elements of B, we can write them as follows: r X a(i) = αij a ˆ(j) , i = r + 1, ..., n, j=1 which gives the system b= r X x ˆi a ˆ(i) + i=1 n X x ˆi i=r+1 r X αij a ˆ(j) . j=1 Collecting terms, we arrive at a reduced system: r X b= yi a ˆ(i) , i=1 where yi = x ˆi + βi , n X βi = x ˆl αli . l=r+1 Now, by the second result of the theorem, i.e., (2), the yi , i = 1, ..., r are uniquely determined. Thus, once we fix the values of xi , i = r + 1, ..., n, then we can solve for the values of xi , i = 1, ..., r. Example 2.16. Consider the matrix 1 2 0 A = 1 2 1 0 0 0 and the right hand side b1 b = b2 . b3 22 MIDTERM REVIEW AND SAMPLE EXAM Note that rank(A) = 2 and in order for ˜ = 2. Now, that b3 = 0, such that rank(A) many solutions to Ax = b. Our approach 1 0 Aˆ = 1 1 0 0 a solution to exist we must have we show that there exist infinitely follows the proof above. Let 2 2 . 0 Then, b = = = = 1 0 2 x ˆ1 1 + x ˆ2 1 + x ˆ3 2 0 0 0 1 0 1 x ˆ1 1 + x ˆ2 1 + 2ˆ x3 1 0 0 0 1 0 (ˆ x1 + 2ˆ x3 ) 1 + x ˆ2 1 0 0 1 0 y1 1 + y2 1 , 0 0 with y1 = x ˆ1 + 2ˆ x3 and y2 = x ˆ2 . Now, the solution to the system in y is y1 = b1 and y2 = b2 − b1 . Thus, b1 = x ˆ1 + 2ˆ x3 and x ˆ2 = b2 − b1 . Thus, x ˆ1 = b1 − 2ˆ x3 x ˆ2 = b2 − b1 x ˆ3 ∈ R. Note that x ˆ3 is a free variable and can be chosen arbitrarily. MIDTERM REVIEW AND SAMPLE EXAM 23 3. Midterm Sample Exam (1) (4.1 #14.) Undamped motions of an elastic spring are governed by the equation my 00 + ky = 0 or my 00 = −ky, where m is the mass, k the spring constant, and y(t) the displacement of the mass from its equilibrium position. Modeling the masses on the two springs we obtain the following system of ODEs: m1 y100 = −k1 y1 + k2 (y2 − y1 ) m2 y200 = −k2 (y2 − y1 ) for unknown displacements y1 (t) of the first mass m1 and y2 (t) of the second mass m2 . The forces acting on the first mass give the first equation and the forces acting on the second mass give the second ODE. Let m1 = m2 = 1, k1 = 3, and k2 = 2 which gives the system 00 −5 2 y1 y1 00 = . y = 00 y2 2 −2 y2 Solve the equation by substituting the function y = xeωt into the ODE. Solution. y 00 = ω 2 xeωt = Axeωt . Setting ω 2 = λ and dividing by eωt gives the eigenproblem Ax = λx, which we solve for x and λ to find the solution. Now, det(A − λI) = (−5 − λ)(−2 − λ) − 4 = λ2 + 7λ + 6 = (λ + 1)(λ + 6) = 0. Thus, the eigenvalues are λ1 = −1 and λ2 = −6. The eigenvector for λ1 = −1 is obtained by solving −4 2 x1 0 = . 2 −1 x2 0 1 Which gives the eigenvector x(1) = . Similarly, the eigenvector 2 2 for λ2 = −6 is x(2) = . −1 √ √ √ √ Notice that ω = ± λ, −1 = i and −6 = i 6. Thus, √ √ y = x(1) c1 eit c2 e−it + x(2) c3 ei 6t + c4 e−i 6t . Now, using Euler’s formula eit = cos(t) + i sin(t) it follows that √ √ y = a1 x(1) cos(t) + b1 x(1) sin(t) + a2 x(2) cos( 6t) + b2 x(2) sin( 6t), 24 MIDTERM REVIEW AND SAMPLE EXAM where a1 = c1 + c2 , b1 = i(c1 − c2 ), a2 = c3 + c4 , b2 = i(c3 − c4 ). These four arbitrary constants are specified by the initial conditions. Remarks: in components, the solution reads: √ √ y1 = a1 cos(t) + b1 sin(t) + 2a2 cos( 6t) + 2b2 x(2) sin( 6t), √ √ y2 = 2a1 cos(t) + 2b1 sin(t) − a2 cos( 6t) − b2 x(2) sin( 6t) The first two terms in y1 and y2 give a slow harmonic motion, and the last two a fast motion. The slow motion occurs if both masses are moving in the same direction, for example, if a1 = 1 and the other three constants are zero. The fast motion occurs if at each instant the two masses are moving in opposite directions, so that one spring is extended and the other compressed. For example if a2 = 1 and the other constants are zero. Depending on the initial conditions, one or the other of these motions, or superposition of both of them will result. (2) (Review # 28) Find the location of the critical points of the system: y10 = cos(y2 ) y20 = 3y1 Can the type of the critical points be determined from the linearized system? Be sure to justify this claim. If your answer is yes, then find the type of the critical points. Solution. The critical points are given by (y1 , y2 ) such that f1 (y2 ) = cos(y2 ) = f2 (y1 ) = 3y1 = 0. Thus, their are infinitely many critical points given by π , 0, 2n + 1 2 where n is any integer. The transformation y → y˜ that maps the critical points to the origin is given by π y˜1 = y1 and y˜2 = y2 − (2n + 1) . 2 Thus, we obtain a system in y˜ by substituting π y1 = y˜1 and y2 = y˜2 + (2n + 1) 2 into the system for y. Letting n = 0 we have the critical point (0, π2 ), for which the system in y˜ reads π y˜10 = cos(˜ y2 + ) = − sin(˜ y2 ) 2 y˜20 = 3˜ y1 MIDTERM REVIEW AND SAMPLE EXAM 25 To determine the type of this critical point we linearize the system: y˜0 ≈ Jf (0)˜ y , where Jf (0) is the Jacobian matrix 0 − cos(0) 0 −1 Jf (0) = = . 3 0 3 0 Note that f1 (˜ y2 ) = − sin(˜ y2 ) ∈ C 1 , f2 (˜ y1 ) = 3˜ y1 ∈ C 1 and det Jf (0) 6= 0 such that the type and stability of the critical points of the nonlinear system coincide with the linearized system. Now, √ 2 the eigenvalues satisfy λ + 3 = 0, so that λ± = ±i 3. Thus, p = trace(Jf (0)) = λ+ + λ− = 0 and det(Jf (0)) = λ+ λ− = 3 and, thus, the critical point is a center. Note that by periodicity of cos(), all the critical points (0, (4n + 1) π2 ) are centers as well, since in such case π π π cos y˜2 + 4n + 1 = cos y˜2 + 2nπ + = cos y˜2 + . 2 2 2 Next, consider n = −1 such that the critical point is (0, − π2 ). Then, π y˜10 = cos(˜ y2 − ) = sin(˜ y2 ) 2 y˜20 = 3˜ y1 implying 0 cos(0) 0 1 Jf (0) = = . 3 0 3 0 √ Here, the eigenvalues satisfy λ2 − 3 = 0 and, thus, λ± = ± 3. This gives p = 0 and q = −3 which implies the critical point is a saddle point. Note that due to periodicity, all of the critical points of the form (0, (4n − 1) π2 ) are saddle points. (3) Consider the nonhomogenous system of ODEs: 0 y (t) = Ay(t) + g(t), −3 1 A= , 1 −3 −6 −2t g(t) = e . 2 Show that a unique solution exists for initial conditions y(0) = s. Then compute the solution. Solution. To show that a unique solution exists we consider the right hand side of the equaiton: −3y1 + y2 − 6e−2t f (t, y1 , y2 ) = . y1 − 3y2 + 2e−2t Since the equation is a constant coefficient linear system, it is sufficient to note that the exponential terms corresponding to the rhs, e−2t , are continuously differentiable at t = 0 (actually for any t). 26 MIDTERM REVIEW AND SAMPLE EXAM Next, we compute the solution to the homogenous system g(t) = 0. The eigenvalues satisfy (λ + 2)(λ + 4) = 0, and so λ1 = −2 and λ2 = −4. The corresponding eigenvectors are 1 1 (1) (2) x = and x = , 1 −1 giving the homogenous solution 1 1 −2t (h) e−4t . e + c2 y = c1 −1 1 Now, to find a particular solution, we can use (1) the method of undetermined coefficients or (2) variation of parameters. For the undetermined coefficients approach, we consider the particular solution y (p) = ute−2t + ve−2t , since the e−2t in g(t) is a solution to the homogenous problem. Our task is then to determine the coefficients u and v. This can be done by plugging y (p) into the system of ODEs: −6 −2t (p) 0 −2t −2t −2t −2t −2t (y ) = ue − 2ute − 2ve = Aute + Ave + e . 2 Now, equating coefficients • of the te−2t terms gives Au = −2u. Thus, u is a solution to the homogenous system −3 + 2 1 u1 0 = . 1 −3 + 2 u2 0 Thus, u1 = u2 = c, for any c ∈ R. • of the e−2t terms gives c u − 2v = Av + w, u = , c −6 w= 2 or c v1 −3 1 v1 −6 −2 = + . c v2 1 −3 v2 2 This gives the system −1 1 v1 6+c = , 1 −1 v2 −2 + c which is consistent if c = −2, since then, the right hand side is an element of the column space of the given matrix: 4 −1 ∈ span . −4 1 MIDTERM REVIEW AND SAMPLE EXAM 27 Note that since rank(A) < n = 2, there exist infinitely many solutions to this linear system: 1 −1 −1 4 = (v1 − v2 ) , + v2 = v1 −1 1 1 −4 implying v2 −v1 = 4 or v2 = v1 +4, where v1 is the free variable. Finally, plugging u and v into the definition of y (p) and adding this to y (h) gives the general solution to the nonhomogenous problem: 1 −2 v1 1 −2t −4t −2t (h) (p) e + te + e−2t . e + c2 y = y + y = c1 −1 −2 v1 + 4 1 Note that the solution is valid for any v1 ∈ R. Once v1 has been selected, the constants c1 and c2 are uniquely determined by the initial conditions, y1 (t0 ) = s1 and y2 (t0 ) = s2 . Next, given a solution to the homogenous system, we find a particular solution using the method of variation of parameters. This approach is easy to understand if we write the solution of the homogenous system, y (h) , solving y 0 = Ay + g, g ≡ 0, in terms of the fundamental matrix. Let, 1 −2t 1 (1) (2) y = e , y = e−4t . 1 −1 Then, y (h) (t) = Y (t)c, where Y = y (1) y (2) = e−2t e−4t e−2t −e−4t and c = c1 . c2 The idea is then to write the particular solution in the form y (p) = Y (t)u(t) and plug this into the nonhomogenous system to find u(t). Substituting into the equation gives Y 0 u + Y u0 = AY u + g. Noting that Y 0 = AY , it follows that Y 0 u = AY u, which gives Y 0 u = g ⇒ u0 = Y −1 g, where Y −1 −1 = −2e−6t −4t 1 e2t e2t −e −e−4t = . −e−2t e−2t 2 e4t −e4t Thus, u0 = 1 2 e2t e2t e4t −e4t −2 −6e−2t = . −4e2t 2e−2t 28 MIDTERM REVIEW AND SAMPLE EXAM Integrating we get u(t) = −2t −2e2t and y (p) −2t −2t −2 v1 e e−4t −2t − 2 −2t −2t = Y (t)u(t) = −2t = e = te + e−2t −2t + 2 −2 v1 + 4 −2e2t e −e−4t for v1 = −2. Linear Algebra (4) Show that the set of all 3 × 3 skew symmetric matrices, i.e., V = {A ∈ R3×3 such that AT = −A}, is a vector space and find the dimension and a basis for V . Solution. Note that if A ∈ R3×3 such that AT = −A, then we can write A as 0 a1 a2 0 a3 , A = −a1 −a2 −a3 0 so that at most three entries of the nine total entries in the matrix A can be chosen independently. Now, taking 0 b1 b2 0 b3 , B = −b1 −b2 −b3 0 we have 0 αa1 + βb1 αa2 + βb2 0 αa3 + βb3 , αA + βB = −(αa1 + βb1 ) −(αa2 + βb2 ) −(αa3 + βb3 ) 0 which is again a skew symmetric matrix. Moreover the 3 × 3 zero matrix is skew symmetric, since then aij = −aji , as required. Thus, V is a vector space. A basis of V is 0 1 0 −1 0 0 , 0 0 0 0 0 1 0 0 0 , −1 0 0 and 0 0 0 0 0 1 , 0 −1 0 so that dim V = 3. (5) Consider the linear system of equations x ∈ Rn , and b ∈ Rm . Define the augmented matrix A˜ = A b ∈ Rm×(n+1) . (a) Show that ˜ = rank(A). (b) Show that the linear system is consistent iff rank(A) the system has a unique solution iff this common rank is equal to n. Ax = b, A ∈ Rm×n , MIDTERM REVIEW AND SAMPLE EXAM 29 P Solution. To prove (a), note that b = ni=1 xi a(i) and, also that ˜ = dim colsp(A) = rank(A), implying b ∈ colsp(A) iff rank(A) rank(A). Next, to prove (b) we assume that there exist two solutions, say x, y ∈ Rn such that Ax = Ay = b. Then, n X Ax − Ay = A(x − y) = (xi − yi )a(i) = 0. i=1 Now, since rank(A) = n, it follows that the columns of A are linearly independent, implying xi − yi = 0, for i = 1, ..., n, or x = y. In the other direction, we assume that Ax = b has a unique solution and then pick a particular right hand side b = 0. Then, x = 0 is a solution and by assumption it is unique. This, then implies that n X Ax = xi a(i) = 0 i=1 has only the trivial solution x = 0. Thus, the columns of a(i) , i = 1, ..., n, are linearly independent and rank(A) = n. By our as˜ = sumption that a solution exists and part (a) it follows that rank(A) rank(A).
© Copyright 2024