ebooksgratis.com

See also ebooksgratis.com: no banners, no cookies, totally FREE.

CLASSICISTRANIERI HOME PAGE - YOUTUBE CHANNEL
Privacy Policy Cookie Policy Terms and Conditions
Kernel (matrix) - Wikipedia, the free encyclopedia

Kernel (matrix)

From Wikipedia, the free encyclopedia

Main article: kernel (mathematics)

In linear algebra, the kernel or null space (also nullspace) of a matrix A is the set of all vectors x for which Ax = 0. The null space of a matrix with n columns is a linear subspace of n-dimensional Euclidean space.[1]

The nullspace (or kernel) of the matrix A is exactly the same thing as the nullspace (or kernel) of the linear mapping defined by the matrix-vector multiplication \mathbf{x} \mapsto \mathbf{Ax}, that is, the set of vectors that map to the zero vector.

Contents

[edit] Finite dimensional linear operator

In linear algebra, the null space (also nullspace) of a matrix A is the set of all vectors x for which Ax = 0. The null space of a matrix with n columns is a linear subspace of n-dimensional Euclidean space.[1]

If we regard the matrix as a linear transformation, then the null space is precisely the kernel of the mapping (i.e. the set of vectors that map to zero). For this reason, the kernel of a linear transformation between abstract vector spaces is sometimes referred to as the null space of the transformation.

[edit] Definition

The null space of an m × n matrix A is the set

\text{Null}(\mathbf{A}) = \left\{ \textbf{x}\in\textbf{R}^n : \mathbf{A}\textbf{x} = \textbf{0} \right\}\text{,}[2]

where 0 denotes the zero vector with m components. The matrix equation Ax = 0 is equivalent to a homogeneous system of linear equations:

\mathbf{A}\textbf{x}=\textbf{0} \;\;\Leftrightarrow\;\;  \begin{alignat}{6}
a_{11} x_1 &&\; + \;&& a_{12} x_2 &&\; + \cdots + \;&& a_{1n} x_n &&\; = 0&    \\
a_{21} x_1 &&\; + \;&& a_{22} x_2 &&\; + \cdots + \;&& a_{2n} x_n &&\; = 0&    \\
\vdots\;\;\; &&     && \vdots\;\;\; &&              && \vdots\;\;\; && \vdots\,& \\
a_{m1} x_1 &&\; + \;&& a_{m2} x_2 &&\; + \cdots + \;&& a_{mn} x_n &&\; = 0. &
\end{alignat}

From this viewpoint, the null space of A is the same as the solution set to the homogeneous system.

[edit] Example

Consider the matrix

\mathbf{A} = \begin{bmatrix}\,\,\,2 & 3 & 5 \\ -4 & 2 & 3\end{bmatrix}

The null space of this matrix consists of all vectors (xyz) ∈ R3 for which

\begin{bmatrix}\,\,\,2 & 3 & 5 \\ -4 & 2 & 3\end{bmatrix}\begin{bmatrix} x \\ y \\ z\end{bmatrix} = \begin{bmatrix} 0 \\ 0 \end{bmatrix}\text{.}

This can be written as a homogeneous system of linear equations involving x, y, and z:

\begin{alignat}{7}
 2x &&\; + \;&& 3y &&\; + \;&& 5z &&\; = \;&& 0 \\
-4x &&\; + \;&& 2y &&\; + \;&& 3z &&\; = \;&& 0.
\end{alignat}

The null space of A is precisely the set of solutions to these equations (in this case, a line through the origin in R3).

[edit] Subspace properties

The null space of an m × n matrix is a subspace of Rn. That is, the set Null(A) has the following three properties:

  1. Null(A) always contains the zero vector.
  2. If x ∈ Null(A) and y ∈ Null(A), then x + y ∈ Null(A).
  3. If x ∈ Null(A) and c is a scalar, then c x ∈ Null(A).

Here are the proofs:

  1. A0 = 0.
  2. If Ax = 0 and Ay = 0, then A(x + y) = Ax + Ay0 + 0 = 0.
  3. If Ax = 0 and c is a scalar, then A(cx) = cAxc0 = 0.

[edit] Basis

The null space of a matrix is not affected by elementary row operations. This makes it possible to use row reduction to find a basis for the null space:

Input An m × n matrix A.
Output A basis for the null space of A
  1. Use elementary row operations to put A in reduced row echelon form.
  2. Interpreting the reduced row echelon form as a homogeneous linear system, determine which of the variables x1x2, ..., xn are free. Write equations for the dependent variables in terms of the free variables.
  3. For each free variable xi, choose the vector in the null space for which xi = 1 and the remaining free variables are zero. The resulting collection of vectors is a basis for the null space of A.

For example, suppose that the reduced row echelon form of A is

\left[ \begin{alignat}{6}
1 && 0 && -3 && 0 &&  2 && -8 \\
0 && 1 &&  5 && 0 && -1 && 4 \\
0 && 0 &&  0 && 1 &&  7 && -9 \\
0 && \;\;\;\;\;0 &&  \;\;\;\;\;0 && \;\;\;\;\;0 &&  \;\;\;\;\;0 && \;\;\;\;\;0 \end{alignat} \,\right]\text{.}

Then x3, x5, and x6 are free, with

\begin{alignat}{7}
x_1 &&\; = \;&&  3x_3 &&\; - \;&& 2x_5 &&\; + \;&& 8x_6  & \\
x_2 &&\; = \;&& -5x_3 &&\; + \;&&  x_5 &&\; - \;&& 4x_6  & \\
x_4 &&\; = \;&&       &&\; - \;&& 7x_5 &&\; + \;&& 9x_6  &.
\end{alignat}

Therefore, the three vectors

\begin{array}{r}  \\  \\ x_3 \rightarrow \\  \\ x_5 \rightarrow \\ x_6 \rightarrow \end{array}

\left[\!\! \begin{array}{r} 3 \\ -5 \\ \mathbf{1} \\ 0 \\ \mathbf{0} \\ \mathbf{0} \end{array} \right],\;
\left[\!\! \begin{array}{r} -2 \\ 1 \\ \mathbf{0} \\ -7 \\ \mathbf{1} \\ \mathbf{0} \end{array} \right],\;
\left[\!\! \begin{array}{r} 8 \\ -4 \\ \mathbf{0} \\ 9 \\ \mathbf{0} \\ \mathbf{1} \end{array} \right]

are a basis for the null space of A.

[edit] Relation to the row space

Main article: Rank-nullity theorem

The product of a matrix A and a vector x can be written in terms of the dot product of vectors:

\mathbf{A}\textbf{x} = \begin{bmatrix} \textbf{a}_1 \cdot \textbf{x} \\ \textbf{a}_2 \cdot \textbf{x} \\ \vdots \\ \textbf{a}_m \cdot \textbf{x} \end{bmatrix}\text{.}

Here a1, ..., am denote the row vectors of the matrix A. It follows that x is in the null space of A if and only if x is orthogonal (or perpendicular) to each of the row vectors of A.

The row space of a matrix A is the span of the row vectors of A. By the above reasoning, the null space of A is the orthogonal complement to the row space. That is, a vector x lies in the null space of A if and only if it is perpendicular to every vector in the row space of A.

The dimension of the row space of A is called the rank of A, and the dimension of null space of A is called the nullity of A. These quantities are related by the equation

\text{rank}(\mathbf{A}) + \text{nullity}(\mathbf{A}) = n\text{.}\,

Here n denotes the number of columns of the matrix A. The equation above is known as the rank-nullity theorem.

[edit] Nonhomogeneous equations

The null space also plays a role in the solution to a nonhomogeneous system of linear equations:

\mathbf{A}\textbf{x}=\textbf{b}\;\;\;\;\;\;\text{or}\;\;\;\;\;\;\begin{alignat}{7}
a_{11} x_1 &&\; + \;&& a_{12} x_2 &&\; + \cdots + \;&& a_{1n} x_n &&\; = \;&&& b_1      \\
a_{21} x_1 &&\; + \;&& a_{22} x_2 &&\; + \cdots + \;&& a_{2n} x_n &&\; = \;&&& b_2      \\
\vdots\;\;\; &&     && \vdots\;\;\; &&              && \vdots\;\;\; &&     &&& \;\vdots \\
a_{m1} x_1 &&\; + \;&& a_{m2} x_2 &&\; + \cdots + \;&& a_{mn} x_n &&\; = \;&&& b_m      \\
\end{alignat}

If u and v are two possible solutions to the above equation, then

\mathbf{A}(\textbf{u}-\textbf{v}) = \mathbf{A}\textbf{u} - \mathbf{A}\textbf{v} = \textbf{b} - \textbf{b} = \textbf{0}\,

Thus, the difference of any two solutions to the equation Ax = b lies in the null space of A.

It follows that any solution to the equation Ax = b can be expressed as the sum of a fixed solution v and an arbitrary element of the null space. That is, the solution set to the equation Ax = b is

\left\{ \textbf{v}+\textbf{x} \,:\, \textbf{x}\in\text{Null}(\mathbf{A})\, \right\}\text{,}

where v is any fixed vector satisfying Av = b. Geometrically, this says that the solution set to Ax = b is the translation of the null space of A by the vector v. See also Fredholm alternative.

[edit] Left null space

The left null space of a matrix A consists of all vectors x such that xTA = 0T, where T denotes the transpose of a column vector. The left null space of A is the same as the null space of AT. The left null space of A is the orthogonal complement to the column space of A, and is the cokernel of the associated linear transformation. The null space, the row space, the column space, and the left null space of A are the four fundamental subspaces associated to the matrix A.

[edit] Null space of a transformation

If V and W are vector spaces, the null space (or kernel) of a linear transformation TV → W is the set of all vectors in V that map to zero:

\ker(T) = \left\{\textbf{v}\in V : T(\textbf{v}) = \textbf{0} \right\}\text{.}

If we represent the linear transformation by a matrix, then the kernel of the transformation is precisely the null space of the matrix.

[edit] Numerical computation of null space

Algorithms based on row or column reduction, that is, Gaussian elimination, presented in introductory linear algebra textbooks and in the preceding sections of this article, are not suitable for a practical computation of the null space because of numerical accuracy problems in the presence of rounding errors. Namely, the computation may greatly amplify the rounding errors, which are inevitable in all but textbook examples on integers, and so give completely wrong results. For this reason, methods based on introductory linear algebra texts are generally not suitable for implementation in software; rather, one should consult contemporary numerical analysis sources for an algorithm like the one below, which does not amplify rounding errors unnecessarily.

A state-of-the-art approach is based on singular value decomposition (SVD). This approach can be also easily programmed using standard libraries, such as LAPACK. SVD of matrix A computes unitary matrices U and V and a rectangular diagonal matrix S of the same size as A with nonnegative diagonal entries, such that

 \mathbf{USV} = \mathbf{A}.

Denote the columns of V by

v_1,\dots,v_n,

the diagonal entries of S by

s_1,\dots,s_{\min\{m,n\}},

and put

s_{\min\{m,n\}+1}=\cdots=s_n=0.

(The numbers si are called the singular values of A.) Then the columns vi of V such that the corresponding  s_i\,=\,0 form an orthonormal basis of the nullspace of A. In a numerical computation, the singular values si are taken to be zero when they are less than some small tolerance. In the Matlab function null, the tolerance is taken to be

 \max\{m,n\} \max \{s_i\} \varepsilon,

where \varepsilon is the machine epsilon of the computer, that is, the smallest number such that in the floating point arithmetics of the computer, 1\,+\,\varepsilon>1. For the IEEE 64 bit floating point format,  \varepsilon\,\approx\, 2.2\cdot 10^{-16}.

Computation of the SVD of a matrix generally costs about the same as several matrix-matrix multiplications with matrices of the same size, when state-of-the art implementation (accurate up to rounding precision) is used, such as in LAPACK. This is true even if, in theory, the SVD cannot be computed by a finite number of operations, so an iterative method with stopping tolerances based on rounding precision must be employed. The cost of the SVD approach is several times higher than computing the null space by reduction, but it should be acceptable whenever reliability is important. It is also possible to compute the null space by the QR decomposition, with the numerical stability and the cost both being between those of the SVD and the reduction approaches.

[edit] Notes

  1. ^ a b Linear algebra, as discussed in this article, is a very well-established mathematical discipline for which there are many sources. Almost all of the material in this article can be found in Lay 2005, Meyer 2001, and Strang 2005.
  2. ^ This equation uses set-builder notation.

[edit] See also

[edit] References

See also: List of linear algebra references

[edit] Textbooks

[edit] Free Online books

[edit] Numerical analysis textbooks

Wikibooks
Wikibooks has a book on the topic of


aa - ab - af - ak - als - am - an - ang - ar - arc - as - ast - av - ay - az - ba - bar - bat_smg - bcl - be - be_x_old - bg - bh - bi - bm - bn - bo - bpy - br - bs - bug - bxr - ca - cbk_zam - cdo - ce - ceb - ch - cho - chr - chy - co - cr - crh - cs - csb - cu - cv - cy - da - de - diq - dsb - dv - dz - ee - el - eml - en - eo - es - et - eu - ext - fa - ff - fi - fiu_vro - fj - fo - fr - frp - fur - fy - ga - gan - gd - gl - glk - gn - got - gu - gv - ha - hak - haw - he - hi - hif - ho - hr - hsb - ht - hu - hy - hz - ia - id - ie - ig - ii - ik - ilo - io - is - it - iu - ja - jbo - jv - ka - kaa - kab - kg - ki - kj - kk - kl - km - kn - ko - kr - ks - ksh - ku - kv - kw - ky - la - lad - lb - lbe - lg - li - lij - lmo - ln - lo - lt - lv - map_bms - mdf - mg - mh - mi - mk - ml - mn - mo - mr - mt - mus - my - myv - mzn - na - nah - nap - nds - nds_nl - ne - new - ng - nl - nn - no - nov - nrm - nv - ny - oc - om - or - os - pa - pag - pam - pap - pdc - pi - pih - pl - pms - ps - pt - qu - quality - rm - rmy - rn - ro - roa_rup - roa_tara - ru - rw - sa - sah - sc - scn - sco - sd - se - sg - sh - si - simple - sk - sl - sm - sn - so - sr - srn - ss - st - stq - su - sv - sw - szl - ta - te - tet - tg - th - ti - tk - tl - tlh - tn - to - tpi - tr - ts - tt - tum - tw - ty - udm - ug - uk - ur - uz - ve - vec - vi - vls - vo - wa - war - wo - wuu - xal - xh - yi - yo - za - zea - zh - zh_classical - zh_min_nan - zh_yue - zu -