Previously we saw that an #(n\times n)#-matrix #A# of rank #n# has some nice properties: according to theorem Invertibilty and rank the inverse of #A# exists and according to Unique solution and maximal rank each system of linear equations with coefficient matrix #A# has exactly one solution.
We also discussed that determining the rank of a matrix is possible by reducing rows and columns to the reduced echelon form and counting the number of independent rows (or columns) of the result. We will however also be able to see whether the rank is #n# or not by calculating the so-called determinant of the #(n\times n)#-matrix. The calculation of a this number is fairly simple, but the underlying theory is more complicated.
The determinant of a #(2\times 2)#-matrix
\[
A=\left(\begin{array}{cc} a & b \\ c & d \end{array}\right)
\] is the number #a\cdot d-b\cdot c#.
The usual notation for the determinant of #A# is #\det(A)# or \(
\left|\begin{array}{cc} a & b \\ c & d \end{array}\right|\).
If #A = \matrix{2&3\\ 4 &6}#, then \[\det(A) = 2\cdot 6- 3\cdot 4 = 0\]
The linear map #\mathbb{R}^2\to\mathbb{R}^2# with matrix #A# is invertible if and only if #\det(A)\ne0#.
In this case, the inverse of #A# is equal to
\[A^{-1} =\dfrac{1}{\det(A)}\cdot \matrix{d&-b\\ -c&a}\]
This theorem including the proof we saw earlier in The inverse of a matrix. Here we give an alternative proof by verifying that #\det(A)\neq 0 \Leftrightarrow \text{rank}(A)=2#.
Suppose #\text{rank}(A)\lt 2#. Then, at least one of the two rows of #A# is dependent on the other. Suppose the second row is a multiple of the first, say #\rv{c,d}=\lambda \rv{a,b}#. Then \[\det(A)=a\cdot (\lambda b)-b\cdot (\lambda a)=0\] The other case (in which the first row is dependent on the second) can be proven similarly.
Conversely, suppose #\det(A)=0#. If one of the two rows is equal to the zero vector, then this row is dependent on the other, so #\text{rank}(A)\lt 2#. Now we assume that #\rv{a,b}\ne\vec{0}#. If #a\neq 0#, then it follows from #a\,d=b\,c# that #d=\frac{b\,c}{a}# and thus #\rv{c,d}=\frac{c}{a}\rv{a,b}#, so the second row is dependent on the first one. If #a = 0#, then #b\ne 0# and it follows from #a\,d=b\,c# that #c = 0#, so #\rv{c,d} =\frac{d}{b}\cdot\rv{a,b}#, and we see again that the second row is dependent on the first one.
Finally, it can easily be verified that \[A\, \matrix{d&-b\\ -c&a}=\matrix{a&b\\ c&d}\,\matrix{d&-b\\ -c&a} = \matrix{a\,d-b\,c&0\\ 0&a\,d-b\,c} = \det(A)\cdot I_2 \] from which it follows that, if #\det(A)\ne0#, the matrix \(\frac{1}{\det(A)}\matrix{d&-b\\ -c&a}\) is the inverse of #A#.
The number #\det(A)# also plays a role in solving a system #A\vec{x}=\vec{p}# in two unknowns: if #A_1# is the matrix obtained from #A# by replacing the first column by #\vec{p}#, and #A_2# is the matrix obtained from #A# by replacing the second column by #\vec{p}#, then the unique solution of the system
\[
\lineqs{ax_1 + bx_2 &=&p_1 \\ cx_1 + dx_2 &=&p_2}
\] is equal to
\[
x_1 = \frac{\det(A_1)}{\det(A)}=\frac{p_1 d-bp_2}{ad-bc}, \quad
x_2 = \frac{\det(A_2)}{\det(A)}=\frac{ap_2-p_1 c}{ad-bc}
\] provided #\det(A)\neq 0#. You can verify this by calculation, but a nice proof valid for all dimensions #n# will be given later. If #n=3# and #A# is a #(3\times 3)#-matrix, then #A\vec{x}=\vec{p}# has a similar solution for #x_1#, #x_2#, and #x_3# in terms of the determinants of #A#, #A_1#, #A_2#, and #A_3#.
A third topic where #\det(A)# plays a role is the computation of surface area. We have not defined area nor the subtleties that arise when coordinates are being used. What we observe here will only serve as an illustration. The oriented surface area of a parallelogram defined by #\rv{a,b}# and #\rv{c,d}# is equal to #a\cdot d-b\cdot c=\det(A)#. It is called oriented, because the outcome can also be negative, depending on the order in which we write down the vectors. The 'real' surface is obtained by taking the absolute value.
The expression #a\cdot d-b\cdot c# is not linear, and thus appears to fall outside the scope of linear algebra. Fortunately, the expression is multilinear. Certain properties of the #(2\times 2)#-determinant model generalize to #(n\times n)#-determinants.
Regard the determinant as a function of the two rows of the matrix: #\det(\vec{a}_1, \vec{a}_2)#. Then #\det# satisfies the following three properties (for all choices of vectors and scalars):
- bilinearity (linearity in both arguments):
\[
\begin{array}{rcl}
\det(\lambda_1 \vec{b}_1 +\lambda_2 \vec{b}_2, \vec{a}_2)&=&
\lambda_1 \det(\vec{b}_1, \vec{a}_2) + \lambda_2 \det(\vec{b}_2 ,\vec{a}_2)\\
\det(\vec{a}_1, \lambda_1 \vec{b}_1 +\lambda_2 \vec{b}_2)&=&
\lambda_1 \det(\vec{a}_1, \vec{b}_1) + \lambda_2\det(\vec{a}_1 ,\vec{b}_2)
\end{array}
\]
- antisymmetry: #\det(\vec{a}_1,\vec{a}_2)=-\det(\vec{a}_2,\vec{a}_1)# (swapping two vectors produces a minus sign); as a consequence, if the two vectors are identical to each other, then #\det(\vec{a},\vec{a})=0#;
- normalization: #\det(\vec{e}_1,\vec{e}_2)=1#.
The determinant is unique in the sense that, if a function of pairs of vectors has the above three properties of #\det#, it is the determinant function #\det#.
These properties are easy to verify. For example, the second property follows from
\[
\left|\begin{array}{cc} a_{11} & a_{12} \\ a_{21} & a_{22}\end{array}\right|
=a_{11}a_{22}-a_{12}a_{21} =-(a_{21}a_{12}-a_{22}a_{11})=-
\left|\begin{array}{cc} a_{21} & a_{22}\\ a_{11} & a_{12} \end{array}\right|
\]
Uniqueness: Suppose that #D# is a function of pairs of vectors, which has the aforementioned three properties of #\det#. Then we have #D(\vec{a}, \vec{a})=0#. By using the bilinearity, we find:
\[
\begin{array}{rcl}
D(\rv{a,b},\rv{c,d}) & =&D(a\vec{e}_1 + b\vec{e}_2,c\vec{e}_1 + d\vec{e}_2)\\ &&\phantom{xxx}\color{blue}{\text{row vectors written in terms of standard basis } \vec{e}_1,\,\vec{e}_2}\\
& =&aD(\vec{e}_1,c\vec{e}_1 + d\vec{e}_2)+ bD(\vec{e}_2,c\vec{e}_1 + d\vec{e}_2)\\ &&\phantom{xxx}\color{blue}{\text{linearity in the first argument}}\\
& =&ac D(\vec{e}_1, \vec{e}_1)+ad D(\vec{e}_1,\vec{e}_2)
+ bc D(\vec{e}_2,\vec{e}_1)+bd D(\vec{e}_2, \vec{e}_2)\\ &&\phantom{xxx}\color{blue}{\text{linearity in the second argument}}\\ & =& ad\,D(\vec{e}_1,\vec{e}_2) + bc\, D(\vec{e}_2,\vec{e}_1)\\ &&\phantom{xxx}\color{blue}{D(\vec{a}, \vec{a})=0}\\
& =& (ad-bc) D(\vec{e}_1,\vec{e}_2) \\ &&\phantom{xxx}\color{blue}{D(\vec{e}_2, \vec{e}_1)=-D(\vec{e}_1, \vec{e}_2)}\\ &=&ad-bc\\ &&\phantom{xxx}\color{blue}{D(\vec{e}_1, \vec{e}_2)=1}\\ &=&\det(\rv{a,b},\rv{c,d}) \\ &&\phantom{xxx}\color{blue}{\text{definition of }\det}
\end{array}
\]
It may seem overkill to describe a simple expression like #a\cdot d-b\cdot c# with these properties, but in higher dimensions such a definition will be essential because explicit formulas are unmanageable for calculations.
The reason for working out the calculations in this simple situation is that the higher dimensional case shows the same kind of calculation, where we will refrain from elaborating all details.
Note that the minus sign is caused by a change in the order of the vectors in the second to last transition.