Linear Algebra
Contents
6. Linear Algebra#
Linear Algebra is a discipline where vector spaces and linear mapping between them are studied. In physics and astronomy, several phenomena can be readily written in terms of linear variables, what makes Computational Linear Algebra a very important topic to be covered throughout this course. We shall cover linear systems of equations, techniques for calculating inverses and determinants and factorization methods.
An interesting fact of Computational Linear Algebra is that it does not comprises numerical approaches as most of the methods are exact. The usage of a computer is then necessary because of the large number of calculations rather than the non-soluble nature of the problems. Numerical errors come then from round-off approximations.
See: http://pages.cs.wisc.edu/~amos/412/lecture-notes/lecture14.pdf
Contents
from IPython.display import HTML, Markdown, YouTubeVideo,Latex
%pylab inline
import numpy as np
import matplotlib.pyplot as plt
## JSAnimation import available at https://github.com/jakevdp/JSAnimation
#from JSAnimation import IPython_display
from matplotlib import animation
#Interpolation add-on
import scipy.interpolate as interp
xrange=range
%pylab is deprecated, use %matplotlib inline and import the required libraries.
Populating the interactive namespace from numpy and matplotlib
https://stackoverflow.com/a/50748163
try:
from google.colab.output._publish import javascript
from IPython.display import Math as math
from IPython.display import Latex as latex
url = "https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.3/latest.js?config=default"
def Math(*args,**kwargs):
javascript(url=url)
return math(*args,**kwargs)
def Latex(*args,**kwargs):
#print(args[0].replace('$',''))
javascript(url=url)
return math(args[0],**kwargs)
except:
from IPython.display import Math, Latex
6.1. Matrices in Python#
One of the most useful advantages of high level languages like Python, is the manipulation of complex objects like matrices and vectors. For this part we are going to use advanced capabilities for handling matrices, provided by the library NumPy.
NumPy, besides the extreme useful NumPy array objects, which are like tensors of any rank, also provides the Matrix objects that are overloaded with proper matrix operations.
6.1.1. Numpy arrays#
A matrix can be represented by an array of two indices
#NumPy Arrays
M1 = np.array( [[ 5 ,-4, 0],
[-4 , 7,-3],
[ 0 ,-3, 5]] )
M2 = np.array( a[[3,-2,1],[-1,5,4],[1,-2,3]] )
print( f'M1=\n{M1}, with shape={M1.shape},\n\nM2=\n{M2}' )
M1=
[[ 5 -4 0]
[-4 7 -3]
[ 0 -3 5]], with shape=(3, 3),
M2=
[[ 3 -2 1]
[-1 5 4]
[ 1 -2 3]]
Note tha the previous definitions correspond to integer arrays.
WARNING: Be careful with integer arrays. In the next example, a single float entry convert the full matrix from an integer array to a float array:
M1[0,0]
5
float
np.array([[ 5, -4, 0],
[-4, 7, -3.],
[ 0, -3, 5]])#[0,0]
array([[ 5., -4., 0.],
[-4., 7., -3.],
[ 0., -3., 5.]])
M1/3
array([[ 1.66666667, -1.33333333, 0. ],
[-1.33333333, 2.33333333, -1. ],
[ 0. , -1. , 1.66666667]])
(M1/3).astype(int)
array([[ 1, -1, 0],
[-1, 2, -1],
[ 0, -1, 1]])
6.1.1.1. Special arrays:#
Let
zero matrix
n=3
np.zeros( (n,n) )
array([[0., 0., 0.],
[0., 0., 0.],
[0., 0., 0.]])
Ones matrix
np.ones( (n,n) )
array([[1., 1., 1.],
[1., 1., 1.],
[1., 1., 1.]])
np.ones( (n,n) )*1j
array([[0.+1.j, 0.+1.j, 0.+1.j],
[0.+1.j, 0.+1.j, 0.+1.j],
[0.+1.j, 0.+1.j, 0.+1.j]])
Identity matrix
np.identity(n)
array([[1., 0., 0.],
[0., 1., 0.],
[0., 0., 1.]])
np.diag([1,2,4])
array([[1, 0, 0],
[0, 2, 0],
[0, 0, 4]])
Random matrix with entries between 0 and 1
np.random.random()
0.9460452544468382
#np.random.seed(986554575)
np.random.random((n,n))
array([[0.05016015, 0.81396332, 0.17984221],
[0.46070592, 0.46019579, 0.9506249 ],
[0.28245888, 0.83429461, 0.47399034]])
np.random.uniform(0,10,(n,n))
array([[5.16683546, 5.73486742, 4.15495224],
[0.75064986, 0.18470119, 3.9795813 ],
[0.59178719, 6.74641986, 7.40144976]])
Integer Random matrix
np.random.randint(0,10,(n,n)) #0 to 9
array([[9, 1, 9],
[1, 8, 4],
[0, 4, 2]])
np.random.randint(-10,10,(n,n))
array([[ 4, -3, -5],
[ 7, 2, -9],
[-10, 5, -8]])
6.1.1.2. Analytical matrices#
Some times it is necessary to work with analytical Matrices. In such a case we can use the symbolic module Sympy
import sympy
x =sympy.Symbol('x') ## declare analytical varibles
sympy.init_printing() ## Use LaTeX to print sympy obejects
import numpy as np
#NumPy Arrays
M1 = np.array( [[ 5 ,-4, 0],
[-4 , 7,-3],
[ 0 ,-3, 5]] )
M2 = np.array( [[3,-2,1],[-1,5,4],[1,-2,3]] )
sympy.Matrix(M2)
In the following we will focus only in numerical matrices as numpy
arrays, but also use sympy
to print the matrices in a more readable way
6.1.1.3. Matrix operations (numpy)#
Both as functions or array’s methods
Transpose
from IPython import display
#Latex(r'M_2^{\rm T}=')
sympy.Matrix( M2.transpose())
Matrix addition
sympy.Matrix(M1)
sympy.Matrix(M2)
sympy.Matrix( M1+M2 )
Complex arrays are allowed
Mc=M1+M2*1j
sympy.Matrix( M1+M2*1j )
with the corresponding complex matrix operations
display ( sympy.Matrix( Mc.conjugate() ) )
display ( sympy.Matrix( Mc.conjugate().transpose() ) ) ## hermitian-conjugate
Matrix multiplication
dot(a, b, out=None)
Dot product of two arrays. Specifically,
If both
a
andb
are 1-D arrays, it is inner product of vectors (without complex conjugation).If both
a
andb
are 2-D arrays, it is matrix multiplication, but using :func:matmul
ora @ b
is preferred.If either
a
orb
is 0-D (scalar), it is equivalent to :func:multiply
and usingnumpy.multiply(a, b)
ora * b
is preferred.…
M1 = np.array( [[ 5 ,-4, 0],
[-4 , 7,-3],
[ 0 ,-3, 5]] )
M2 = np.array( [[3,-2,1],
[-1,5,4],
[1,-2,3]] )
#Multiplication
sympy.Matrix( np.dot( M1, M2 ) )
sympy.Matrix( M1.dot(M2) )
Recommended way
sympy.Matrix(M1@M2)
np.array( [1,2])@np.array( [[1],[2]])
array([5])
#Multiplication is not commutative
sympy.Matrix ( M2@M1 )
Complex matrices
σ_2=np.array(
[[0,-1j],
[1j,0]])
sympy.Matrix(σ_2)
Trace
np.trace(σ_2)
0j
Determinant
np.linalg.det(σ_2)
(-1+0j)
Raise a square matrix to the (integer) power n
.
sympy.Matrix ( np.linalg.matrix_power(σ_2,2) )
(σ_2@σ_2).astype(float)
/tmp/ipykernel_12325/501469815.py:1: ComplexWarning: Casting complex values to real discards the imaginary part
(σ_2@σ_2).astype(float)
array([[1., 0.],
[0., 1.]])
odd power
sympy.Matrix ( np.linalg.matrix_power(σ_2,7) )
even power
sympy.Matrix ( np.linalg.matrix_power(σ_2,14) )
Inverse of a matrix
sympy.Matrix( np.linalg.inv(M2) )
sympy.Matrix ( ( M2@np.linalg.inv(M2) ).round(15) )
( M2@np.linalg.inv(M2) ).round(15).astype(int)
array([[1, 0, 0],
[0, 1, 0],
[0, 0, 1]])
sympy.Matrix ( ( M2@np.linalg.inv(M2) ).round(15).astype(int) )

6.1.1.4. Element by element operations#
print(M1)
M2
[[ 5 -4 0]
[-4 7 -3]
[ 0 -3 5]]
array([[ 3, -2, 1],
[-1, 5, 4],
[ 1, -2, 3]])
M1*M2
array([[ 15, 8, 0],
[ 4, 35, -12],
[ 0, 6, 15]])
#Division
print( M1/M2 )
[[ 1.66666667 2. 0. ]
[ 4. 1.4 -0.75 ]
[ 0. 1.5 1.66666667]]
sympy.Matrix( M1@np.linalg.inv(M2).round(3) )
6.1.1.5. Submatrix from matrix:#
M1 = np.array( [[ 5 ,-4, 0],
[-4 , 7,-3],
[ 0 ,-3, 5]] )
Elements of the matrix
M1[1,2]
-3
6.1.1.6. Extract rows#
M1[1,:]
array([-4, 7, -3])
or
M1[1]
array([-4, 7, -3])
In general:
M1[list of rows to extract]
M1[[1]]
array([[-4, 7, -3]])
M1 = np.array( [[ 5 ,-4, 0],
[-4 , 7,-3],
[ 0 ,-3, 5]] )
M1[[0,2]]
array([[ 5, -4, 0],
[ 0, -3, 5]])
6.1.1.7. Extract columns#
M1[:,1]
array([-4, 7, -3])
and convert back into a column
np.c_[M1[:,1]]
array([[-4],
[ 7],
[-3]])
6.1.1.8. Reordering#
M1 = np.array( [[ 5 ,-4, 0],
[-4 , 7,-3],
[ 0 ,-3, 5]] )
Reverse row order to 2,1,0
M1[[2,1,0]]
array([[ 0, -3, 5],
[-4, 7, -3],
[ 5, -4, 0]])
or
np.r_[[M1[2]],[M1[1]],[M1[0]]]
array([[ 0, -3, 5],
[-4, 7, -3],
[ 5, -4, 0]])
Reverse column order to 2,1,0
np.c_[ M1[:,2],M1[:,1],M1[:,0] ]
array([[ 0, -4, 5],
[-3, 7, -4],
[ 5, -3, 0]])
6.1.1.9. It is possible to reasign a full set of rows to a matrix#
M1 = np.array( [[ 5,-4, 0],
[-4, 7,-3],
[ 0,-3, 5]] )
M1[[0,2]]=[[0 ,0,0.3],
[0.3,0,0 ]]
M1
array([[ 0, 0, 0],
[-4, 7, -3],
[ 0, 0, 0]])
Note that because M1
has only integer numbers, it is assumed to be an integer matrix. In order to allow float assignments, it is first necessary to convert to a float matrix:
M1=M1.astype(float)
M1[[0,2]]=[[0 ,0,0.3],
[0.3,0,0 ]]
M1
array([[ 0. , 0. , 0.3],
[-4. , 7. , -3. ],
[ 0.3, 0. , 0. ]])
WARNING
In some cases a reasignment of a list or array points to the same space in memory. To really copy an array to other variable it is always recommended to use the .copy()
method.
A=np.array([1,2])
A
array([1, 2])
Reassign memory space
C=A
C
array([1, 2])
A change is reflected in both variables
C[0]=5
print(f'A={A} → C={C}')
A=[5 2] → C=[5 2]
A[1]=8
print(f'A={A} → C={C}')
A=[5 8] → C=[5 8]
To keep the first memory space
B=A.copy() ## two different memory spaces
B[0]=4
print('A =',A)
print('B =',B)
A = [5 2]
B = [4 2]
6.1.1.10. Matrix object#
numpy also has a Matrix object with simplified operations. However it is recommended to use the general array object even for specific matrix operations. This helps to avoid incompatibilities with usual array operations. For example, as shown below the multiplication symbol, *
, behaves different for arrays that for matrices, and could induce to errors.
#NumPy Matrix
M1 = np.matrix( [[5,-4,0],[-4,7,-3],[0,-3,5]] )
M2 = np.matrix( [[3,-2,1],[-1,5,4],[1,-2,3]] )
print (M1, "\n")
print (M2)
[[ 5 -4 0]
[-4 7 -3]
[ 0 -3 5]]
[[ 3 -2 1]
[-1 5 4]
[ 1 -2 3]]
#Addition
sympy.Matrix( M1+M2 )

#Multiplication
sympy.Matrix( M1*M2 )

6.1.2. Basic operations with matrices#
In order to simplify the following methods, we introduce here 3 basic operations over linear systems:
The
-th row can be multiplied by a non-zero constant , and then a new row used in place of , i.e. . We denote this operation as .The
-th row can be multiplied by a non-zero constant and added to some row . The resulting value take the place of , i.e. . We denote this operation as .Finally, we define a swapping of two rows as
, denoted as
6.1.2.1. Extract rows from an array#
6.1.3. Activity #
Write three routines that perform, given a matrix
row_lamb(i, λ, A)
:i
is the row to be changed,λ
the multiplicative factor andA
the matrix. This function should return the new matrix with the performed operation .row_comb(i, j, λ, A)
:i
is the row to be changed,j
the row to be added,λ
the multiplicative factor andA
the matrix. This function should return the new matrix with the performed operation .row_swap(i, j, A)
:i
andj
are the rows to be swapped. This function should return the new matrix with the performed operation .
import numbers
isinstance(np.array([2])[0],numbers.Integral)
True
M1=M1.astype(float)
M1
array([[ 2., -4., 0.],
[-4., 7., -3.],
[ 0., -3., 5.]])
M1.dtype==np.dtype('int64')
False
def row_sawp(i,j,A):
'''
i and j are the rows to be swapped
'''
B=A.copy()
B[[i,j]]=B[[j,i]]
return B
def row_comb(i,j,λ,A):
'''
`i` is the row to be changed,
`j` the row to be added
`λ` the multiplicative factor
`A` the matrix.
operation
---------
E_i→ E_i+λ*E_j
'''
B=A.copy()
if B.dtype==np.dtype('int64'):
B=B.astype(float)
B[i]=B[i]+λ*B[j]
return B
def row_lamb(i,λ,A):
'''
`i` is the row to be changed,
`λ` the multiplicative factor
`A` the matrix.
operation
---------
λ → λ*E_i
'''
B=A.copy()
if B.dtype==np.dtype('int64'):
B=B.astype(float)
B[i]=λ*B[i]
return B
Check with
import numpy as np
import sympy
M1=np.array( [[5,-4,0],[-4,7,-3],[0,-3,5]] )
M1
array([[ 5, -4, 0],
[-4, 7, -3],
[ 0, -3, 5]])
row_lamb(1,3.4,M1)
array([[ 5. , -4. , 0. ],
[-13.6, 23.8, -10.2],
[ 0. , -3. , 5. ]])
row_comb(0,1,5/4,M1)
array([[ 0. , 4.75, -3.75],
[-4. , 7. , -3. ],
[ 0. , -3. , 5. ]])
row_sawp(0,1,M1)
array([[-4, 7, -3],
[ 5, -4, 0],
[ 0, -3, 5]])
6.2. Gaussian Elimination#
6.2.1. Example 1#
Using Kirchhoff’s circuit laws, it is possible to solve the next system:
obtaining the next equations:
or
Defining the variables
A first method to solve linear systems of equations is the Gaussian elimination. This procedure consists of a set of recursive steps performed in order to diagonalise the matrix of the problem. A suitable way to introduce this method is applying it to some basic problem. To do so, let’s take the result of the Example 1:
Constructing the associated augmented matrix, we obtain
6.2.1.1. Activity #
Use numpy functions to build the augmented matrix M1
M1 = np.array( [[5,-4,0],[-4,7,-3],[0,-3,5]] )
sympy.Matrix(M1)
np.array( [[1],[0],[-2]] )
array([[ 1],
[ 0],
[-2]])
M1_aug=np.c_[ M1, [[1],[0],[-2]] ]
sympy.Matrix(M1_aug)
or
M1_aug=np.hstack( (M1,[[1],[0],[-2]]))
sympy.Matrix(M1_aug)
At this point, we can eliminate the coefficients of the variable
[-4,7,-3⋮0]+[4,-16/5,0⋮4/5]=[0,19/5,-3⋮4/5]
The coefficient in the third equation is already null. We then obtain:
M1_LU=row_comb(1,0,4/5,M1_aug)
M1_LU
array([[ 5. , -4. , 0. , 1. ],
[ 0. , 3.8, -3. , 0.8],
[ 0. , -3. , 5. , -2. ]])
Now, we proceed to eliminate the coefficient of
[0,-3,5⋮-2]+[0,(3*5/19)*(19/5),-(3*5/19)*3⋮(3*5/19)*4/5]→[0,0,50/19⋮-26/19]
M1_LU=row_comb(2,1,3*5/19,M1_LU)
M1_LU
array([[ 5. , -4. , 0. , 1. ],
[ 0. , 3.8 , -3. , 0.8 ],
[ 0. , 0. , 2.63157895, -1.36842105]])
M1_LU
array([[ 5. , -4. , 0. , 1. ],
[ 0. , 3.8 , -3. , 0.8 ],
[ 0. , 0. , 2.63157895, -1.36842105]])
The final step is to solve for
From this, it is direct to conclude that
row_lamb(2,19/50,M1_LU)
array([[ 5. , -4. , 0. , 1. ],
[ 0. , 3.8 , -3. , 0.8 ],
[ 0. , 0. , 1. , -0.52]])
The full implementation in numpy is with np.linalg.solve
np.linalg.solve?
np.linalg.solve(M1,[1,0,-2])
array([ 0.04, -0.2 , -0.52])
np.linalg.solve(M1,[[1],[0],[-2]])
array([[ 0.04],
[-0.2 ],
[-0.52]])
6.3. Linear Systems of Equations#
A linear system is a set of equations of
where
A linear system has solution if and only if
Although there is an intuitive way to solve this type of systems, by just adding and subtracting equations until reaching the desire result, the large number of variables of some systems found in physics and astronomy makes necessary to develop iterative and general approaches. Next, we shall introduce matrix and vector notation and some basic operations that will be the basis of the methods to be developed in this section.
Quadratic eqution…
6.3.1. Matrices and vectors#
A
In the same way, it is possible to define a
and a column vector of constant coefficients
The system of equations
can be then written in a more convenient way as
We can also introducing the
6.3.1.1. Implementation in Scipy#
import scipy.linalg as linalg
linalg.lu?
M1_aug
array([[ 5, -4, 0, 1],
[-4, 7, -3, 0],
[ 0, -3, 5, -2]])
P,L,U=linalg.lu(M1_aug)
U
array([[ 5. , -4. , 0. , 1. ],
[ 0. , 3.8 , -3. , 0.8 ],
[ 0. , 0. , 2.63157895, -1.36842105]])
x3=U[2,3]/U[2,2]
'x3=-26/50={:.2f}'.format(x3)
'x3=-26/50=-0.52'
x2=(U[1,3]-U[1,2]*x3)/U[1,1]
round(x2,2)

x1=( U[0,3]-U[0,1]*x2-U[0,2]*x3 )/U[0,0]
round(x1,2)

6.3.2. General Gaussian elimination#
Now, we shall describe the general procedure for Gaussian elimination:
Give an augmented matrix
.Find the first non-zero coefficient
associated to . This element is called pivot.Apply the operation
. This guarantee the first row has a non-zero coefficient .Apply the operation
. This eliminates the coefficients associated to in all the rows but in the first one.Repeat steps 2. to 4. but for the coefficients of
and then and so up to . When iterating the coefficients of , do not take into account the first -th rows as they are already sorted.Once you obtain a diagonal form of the matrix, apply the operation
. This will make the coefficient of in the last equation equal to 1.The final result should be an augmented matrix of the form: $
$The solution to the problem is then obtained through backward substitutions, i.e. $
$
See custom implementation of general Gaussian elimination
#Gaussian Elimination
import scipy
def Gaussian_Elimination( A0 ):
#Making local copy of matrix
P,L,A=scipy.linalg.lu(A0)
n = len(A)
#Finding solution
x = np.zeros( n )
x[n-1] = A[n-1,n]
for i in range( n-1, -1, -1 ):
x[i] = ( A[i,n] - sum(A[i,i+1:n]*x[i+1:n]) )/A[i,i]
#Upper diagonal matrix and solutions x
return A, x
np.random.seed(3)
M = np.random.random( (4,5) )
M
array([[0.5507979 , 0.70814782, 0.29090474, 0.51082761, 0.89294695],
[0.89629309, 0.12558531, 0.20724288, 0.0514672 , 0.44080984],
[0.02987621, 0.45683322, 0.64914405, 0.27848728, 0.6762549 ],
[0.59086282, 0.02398188, 0.55885409, 0.25925245, 0.4151012 ]])
U,x=Gaussian_Elimination(M)
Upper triangular
U
array([[ 0.89629309, 0.12558531, 0.20724288, 0.0514672 , 0.44080984],
[ 0. , 0.63097203, 0.16354802, 0.47919953, 0.62205662],
[ 0. , 0. , 0.52490983, -0.06699671, 0.21531002],
[ 0. , 0. , 0. , 0.32582312, 0.00303691]])
solutions
x
array([0.27395594, 0.8721633 , 0.41137442, 0.00932074])
M1_aug
array([[ 5, -4, 0, 1],
[-4, 7, -3, 0],
[ 0, -3, 5, -2]])
Gaussian_Elimination(M1_aug)
(array([[ 5. , -4. , 0. , 1. ],
[ 0. , 3.8 , -3. , 0.8 ],
[ 0. , 0. , 2.63157895, -1.36842105]]),
array([ 0.04, -0.2 , -0.52]))
6.3.3. Pivoting Strategies#
The previous method of Gaussian Elimination for finding solutions of linear systems is mathematically exact, however round-off errors that appear in computational arithmetics can represent a real problem when high accuracy is required.
In order to illustrate this, consider the next situation:
Using four-digit arithmetic we obtain:
1. Constructing the augmented matrix:
2. Applying the reduction with the first pivot, we obtain:
where:
In this step, we have taken the first four digits. This leads us to:
The exact system is instead
Using the solution
The exact solution is however:
The source of such a large error is that the factor
6.3.4. Computing time#
As mentioned before, Algebra Linear methods do not invole numerical approximations excepting round-off errors. This implies the computing time depends on the number of involved operations (multiplications, divisions, additions and subtractions). It is possible to demonstrate that the numbers of required divisions/multiplications is given by:
and the required additions/subtractions:
These numbers are calculated separately as the computing time per operation for divisions and multiplications are similar and much larger than computing times for additions and subtractions. According to this, the overall computing time, for large
6.3.5. Example 2#
Using the library datetime
of python, evaluate the computing time required for Gaussian_Elimination
to find the solution of a system of
#Importing datatime
from datetime import datetime
#Number of repeats
Nrep = 500
#Size of matrix
n = 20
def Gaussian_Time( n, Nrep ):
#Arrays of times
Times = []
#Cicle for number of repeats
for i in xrange(Nrep):
#Generating random matrix
M = np.matrix( np.random.random( (n,n+1) ) )
#Starting time counter
tstart = datetime.now()
#Invoking Gaussian Elimination routine
Gaussian_Elimination(M)
#Killing time counter
tend = datetime.now()
#Saving computing time
Times.append( (tend-tstart).microseconds )
#Numpy Array
Times = np.array(Times)
print "The mean computing time for a %dx%d matrix is: %lf microseconds"%(n,n,Times.mean())
#Histrogram
plt.figure( figsize=(8,5) )
histo = plt.hist( Times, bins = 30 )
plt.xlabel( "Computing Time [microseconds]" )
plt.ylabel( "Ocurrences" )
plt.grid()
return Times.mean()
Gaussian_Time( n, Nrep )
The mean computing time for a 20x20 matrix is: 2762.628000 microseconds
2762.6280000000002

6.3.6. Activity #
Using the previous code, estimate the computing time for random matrices of size
6.3.7. Partial pivoting#
A suitable method to reduce round-off errors is to choose a pivot more conveniently. As we saw before, a small pivot generally implies larger propagated errors as they appear usually as dividends. The partial pivoting method consists then of a choosing of the largest absolute coefficient associated to
This way, propagated multiplication errors would be minimized.
6.3.8. Activity #
Create a new routine Gaussian_Elimination_Pivoting
from Gaussian_Elimination
in order to include the partial pivoting method. Compare both routines with some random matrix and with the exact solution.
6.4. Matrix Inversion#
Asumming a nonsingular matrix
A corollary of this definition is that
Once defined the Gaussian Elimination method, it is possible to extend it in order to find the inverse of any nonsingular matrix. Let’s consider the next equation:
This can be rewritten as a set of
These systems can be solved individually by using Gaussian Elimination, however we can mix all the problems, obtaining the augmented matrix:
Now, applying Gaussian Elimination we can obtain a upper diagonal form for the first matrix. Completing the steps using forwards elimination we can convert the first matrix into the identity matrix, obtaining
Where the second matrix is then the inverse
6.4.1. Activity #
Using the previous routine Gaussian_Elimination_Pivoting
, create a new routine Inverse
that calculates the inverse of any given squared matrix.
6.5. Determinant of a Matrix#
The determinant of a matrix is a scalar quantity calculated for square matrix. This provides important information about the matrix of coefficients of a system of linear of equations. For example, any system of
6.5.1. Calculating determinants#
Next, we shall define some properties of determinants that will allow us to calculate determinants by using a recursive code:
1. If
2. If
3. The cofactor
4. The determinant of a
or
This is, it is possible to use both, a row or a column for calculating the determinant.
6.5.2. Computing time of determinants#
Using the previous recurrence, we can calculate the computing time of the previous algorithm. First, let’s consider the number of required operations for a
The determinant is then given by:
the number of required multiplications was
Now, using the previous formula for the determinant
For a
For a
If
In computers, this is a prohibitive computing time so other schemes have to be proposed.
6.5.3. Activity #
Evaluate the computing time of the Determinant
routine for matrix sizes of
6.5.4. Properties of determinants#
Determinants have a set of properties that can reduce considerably computing times. Suppose
1. If any row or column of
2. If two rows or columns of
3. If
4. If
5. If
6. If
7.
8.
9. Finally and most importantly: if
As we analysed before, Gaussian Elimination takes a computing time scaling like
6.5.5. Activity #
Using the Gaussian_Elimination
routine and tracking back the performed operations, construct a new routine called Gaussian_Determinant
. Make the same analysis of the computing time as the previous activity. Compare both results.
6.5.6. Existence of inverse#
A matrix
If the matrix
Example From the previous example
np.random.seed(3)
M = np.random.random( (4,5) )
M
array([[0.5507979 , 0.70814782, 0.29090474, 0.51082761, 0.89294695],
[0.89629309, 0.12558531, 0.20724288, 0.0514672 , 0.44080984],
[0.02987621, 0.45683322, 0.64914405, 0.27848728, 0.6762549 ],
[0.59086282, 0.02398188, 0.55885409, 0.25925245, 0.4151012 ]])
M[:,0]
array([0.5507979 , 0.89629309, 0.02987621, 0.59086282])
np.c_[M[:,0]]
array([[0.5507979 ],
[0.89629309],
[0.02987621],
[0.59086282]])
A=np.c_[ tuple( [ np.c_[M[:,i]] for i in range(4) ] ) ]
A
array([[0.5507979 , 0.70814782, 0.29090474, 0.51082761],
[0.89629309, 0.12558531, 0.20724288, 0.0514672 ],
[0.02987621, 0.45683322, 0.64914405, 0.27848728],
[0.59086282, 0.02398188, 0.55885409, 0.25925245]])
b=np.c_[ M[:,4] ]
b
array([[0.89294695],
[0.44080984],
[0.6762549 ],
[0.4151012 ]])
such that
Check that
np.linalg.det(A)

sympy.Matrix( np.dot( np.linalg.inv(A) , b).round(4) )
6.5.6.1. Numpy implementation#
np.linalg.solve(A,b)
array([[0.27395594],
[0.8721633 ],
[0.41137442],
[0.00932074]])
np.linalg.solve(A,M[:,4])
array([0.27395594, 0.8721633 , 0.41137442, 0.00932074])
6.6. Matrix diagonalization#
6.7. LU Factorization#
As we saw before, the Gaussian Elimination algorithm takes a computing time scaling as
The Gauss-Jordan algorithm can reduce even more this problem in order to solve it directly, yielding:
From the upper diagonal form to the completely reduced one, it is necessary to perform
Now, let
For solving this system we can then:
1. Solve the equivalent system
2. Once we know
The global computing time is then
6.7.1. Activity #
In order to compare the computing time that Gaussian Elimination takes and the previous time for the LU factorization, make a plot of both computing times. What can you conclude when
6.7.2. Derivation of LU factorization#
Although the LU factorization seems to be a far better method for solving linear systems as compared with say Gaussian Elimination, we was assuming we already knew the matrices
You may wonder then, what advantage implies the use of this factorization? Well, matrices
First, let’s assume a matrix
henceforth,
The previous operation over the matrix
This is called the first Gaussian transformation matrix. From this, we have
where
Repeating the same procedure for the next pivots, we obtain then
where the
and
Note
so we can define
Now, taking the equation
and defining the inverse of
we obtain
where the lower diagonal matrix
$
6.7.3. Algorithm for LU factorization#
The algorithm is then given by:
1. Give a square matrix
2. Apply the operation
3. Construct the matrix
with
4. Repeat the steps 2 and 3 for the next column until reaching the last one.
5. Return the matrices
6.7.4. Activity #
Create a routine called LU_Factorization
that, given a matrix
M1
array([[ 5, -4, 0],
[-4, 7, -3],
[ 0, -3, 5]])
import scipy
P,L,U=scipy.linalg.lu(M1)
U
array([[ 5. , -4. , 0. ],
[ 0. , 3.8 , -3. ],
[ 0. , 0. , 2.63157895]])
The same obtained before
scipy.linalg.lu?
6.7.5. Eigenvalues and Eigenvectors activity #
6.7.5.1. Electron interacting with a magnetic field#
An electron is placed to interact with an uniform magnetic field. To give account of the possible allowed levels of the electron in the presence of the magnetic field it is necessary to solve the next equation
where the hamiltonian is equal to
Then, by solving the problem
And solving the problem
The function scipy.optimize.root can be used to get roots of a given equation. The experimental value of
Find the allowed energy levels.
Find the autofunctions and normalize them.
Hint: An imaginary number in python can be written as 1j.