|
|
(One intermediate revision by one other user not shown) |
Line 1: |
Line 1: |
| {{too technical|date=April 2013}}
| | Even out youthful dogs crapper digest from puppy arthritis. Although pawl arthritis is Sir Thomas More vulgar among sr. dogs, your puppy tin can [http://Www.google.Co.uk/search?hl=en&gl=us&tbm=nws&q=draw+arthritis&gs_l=news draw arthritis] when it is caused by any form of combat injury to joints. Also, youth dogs that are fairly on the weighed down incline tin can are likewise in greater risk of arthritis. About hot dog arthritis rear end lead off early, without you knowing, and backside exclusively pop viewing effects when your wienerwurst ages. Since pup arthritis is non unwashed and merely occurs in select cases when hurt or harm plays a persona in negative whatever of your Young dogs joints, this typewrite of pawl arthritis is usually exclusively characterized by gimpiness and unvarying junction aches and pains. <br><br>Whirligig Track Arthritis Symptoms <br><br>Claudication is the world-class of wholly andiron arthritis sym. Dogs with arthritis commonly show up no early tattletale signs apart from a enfeebling in its movements, so rent this as your low clew. Your weenie whitethorn of a sudden recover it laborious to drive up or movement about flying and easily. A deal out of detent owners error these symptoms for convention signs of senescence in quondam dogs. However, in Young dogs, when vague signs of gimpiness commence to show, your detent power be in peril of arthritis. Also, if you are sore enough, you volition about potential be able-bodied to observe abnormalities in your dogs deportment. If your dogs look to sound off or see chronic pain sensation about its joints, chances are luxuriously that that equates to chase arthritis. Just about click arthritis cases are besides acompanied with inflammation, since the damaged articulatio bequeath beau up. Away from these easily identified symptoms, your trail tail end likewise abide from fever, and a commute in its appetency. These symptoms are harder to name since it derriere be caused by something else. However, if these symptoms are attended with the early symptoms mentioned earlier, it agency that in that location is stronger cogent evidence that your track is experiencing an attack of arthritis. <br><br>How to Treat Dog Arthritis <br><br>If your weenie starts display whatever of the symptoms above, dont disregard the condition and control for the early symptoms. Tail arthritis in old dogs buns be prevented when cad arthritis in Young dogs are right hardened. Also, if your dogs arthritis is now paying sufficient attending to, the faster it throne be toughened. Veterinarians commonly dictate botheration relievers, and anti-seditious medicinal drug to facilitate with the intumescence and the pain sensation. At that place are besides various types of therapy to process a frump with arthritis. Nonpareil of the almost uncouth therapies is the utilize of bee venom products. Flavoring therapies are also identical good in relieving hurt experient by dogs with arthritis. However, about handling methods and medications are geared towards relieving the nuisance and soreness that your firedog feels. Hound arthritis is a degenerative disease. In one case it sets on, its jolly operose to overturn the effects. <br><br>How to Void Domestic dog Arthritis <br><br>To ward off trail arthritis, pull in certain that tied as a Young puppy, your andiron gets the mighty amount of work out. Dont Lashkar-e-Toiba your frank loiter around, organic process older. Dont o'er exercise it with excessive exercises either. Also, since fill out dogs carry higher risks of acquiring track arthritis, nominate surely to see to it and influence your dogs dieting and nutrition. If your dog-iron starts look rounder and seems heavier, act immediately by keen dorsum on its meals and taking it KO'd to work.<br><br>If you cherished this short article and you would like to acquire additional information relating to [http://www.shop-pet.com/ ปัตตาเลี่ยนสุนัข] kindly pay a visit to our own website. |
| {| align=right width=248px style="margin-left:1em"
| |
| |[[Image:Projectivisation F5P^1.svg|120px]][[Image:Projectivisation F5P^1.svg|120px]]<br/>[[Image:Projectivisation F5P^1.svg|120px]][[Image:Projectivisation F5P^1.svg|120px]]
| |
| |-
| |
| | style="font-size:87%" |One-dimensional subspaces in the two-dimensional vector space over the [[finite field]] '''F'''<sub>5</sub>. The [[origin (mathematics)|origin]] (0, 0), marked with green circles, belongs to any of six 1-subspaces, while each of 24 remaining points belongs to exactly one; a property which holds for 1-subspaces over any field and in all dimensions. All '''F'''<sub>5</sub><sup>2</sup> (i.e. a 5 × 5 square) is pictured four times for a better visualization
| |
| |}
| |
| The concept of a '''linear subspace''' (or '''vector subspace''') is important in [[linear algebra]] and related fields of [[mathematics]].
| |
| A linear subspace is usually called simply a ''subspace'' when the context serves to distinguish it from other kinds of [[subspace (disambiguation)|subspace]]s. | |
| | |
| == Definition and useful characterization of subspace ==
| |
| | |
| Let ''K'' be a [[field (mathematics)|field]] (such as the field of [[real number]]s), and let ''V'' be a [[vector space]] over ''K''.
| |
| As usual, we call elements of ''V'' ''vectors'' and call elements of ''K'' ''[[scalar (mathematics)|scalars]]''. Ignoring the full extent of mathematical generalization, scalars can be understood simply as [[number]]s.
| |
| Suppose that ''W'' is a [[subset]] of ''V''.
| |
| If ''W'' is a vector space itself (which means that it is [[closure (mathematics)|closed]] under operations of [[addition]] and [[scalar multiplication]]), with the same vector space operations as ''V'' has, then ''W'' is a '''subspace''' of ''V''.
| |
| | |
| To use this definition, we don't have to prove that all the properties of a vector space hold for ''W''.
| |
| Instead, we can prove a theorem that gives us an easier way to show that a subset of a vector space is a subspace.
| |
| | |
| '''Theorem:'''
| |
| Let ''V'' be a vector space over the field ''K'', and let ''W'' be a subset of ''V''.
| |
| Then ''W'' is a subspace [[if and only if]] ''W'' satisfies the following three conditions:
| |
| #The zero vector, '''0''', is in ''W''.
| |
| #If '''u''' and '''v''' are elements of ''W'', then the sum '''u''' + '''v''' is an element of ''W'';
| |
| #If '''u''' is an element of ''W'' and ''c'' is a scalar from ''K'', then the product ''c'''''u''' is an element of ''W'';
| |
| | |
| '''Proof:'''
| |
| Firstly, property 1 ensures ''W'' is nonempty. Looking at the definition of a [[vector space]], we see that properties 2 and 3 above assure closure of ''W'' under addition and scalar multiplication, so the vector space operations are well defined. Since elements of ''W'' are necessarily elements of ''V'', axioms 1, 2 and 5–8 of a vector space are satisfied. By the closure of ''W'' under scalar multiplication (specifically by 0 and -1), vector space's definitional axiom [[identity element]] of [[vector_addition#Addition_and_subtraction|addition]] and axiom [[inverse element]] of addition are satisfied.
| |
| | |
| Conversely, if ''W'' is subspace of ''V'', then ''W'' is itself a vector space under the operations induced by
| |
| ''V'', so properties 2 and 3 are satisfied. By property 3, −'''w''' is in ''W'' whenever '''w''' is, and it follows that
| |
| ''W'' is closed under subtraction as well. Since
| |
| ''W'' is nonempty, there is an element '''x''' in ''W'', and <math> \bold x - \bold x = {\bold 0} </math>
| |
| is in ''W'', so property 1 is satisfied. One can also argue that since ''W'' is nonempty, there is an element '''x''' in ''W'', and 0 is in the field ''K'' so <math> 0 \bold x = {\bold 0} </math> and therefore property 1 is satisfied.
| |
| | |
| == Examples ==
| |
| '''Example I:'''
| |
| Let the field ''K'' be the set '''R''' of [[real number]]s, and let the vector space ''V'' be the [[real coordinate space]] '''R'''<sup>3</sup>.
| |
| Take ''W'' to be the set of all vectors in ''V'' whose last component is 0.
| |
| Then ''W'' is a subspace of ''V''.
| |
| | |
| ''Proof:''
| |
| #Given '''u''' and '''v''' in ''W'', then they can be expressed as '''u''' = (''u''<sub>1</sub>, ''u''<sub>2</sub>, 0) and '''v''' = (''v''<sub>1</sub>, ''v''<sub>2</sub>, 0). Then '''u''' + '''v''' = (''u''<sub>1</sub>+''v''<sub>1</sub>, ''u''<sub>2</sub>+''v''<sub>2</sub>, 0+0) = (''u''<sub>1</sub>+''v''<sub>1</sub>, ''u''<sub>2</sub>+''v''<sub>2</sub>, 0). Thus, '''u''' + '''v''' is an element of ''W'', too.
| |
| #Given '''u''' in ''W'' and a scalar ''c'' in '''R''', if '''u''' = (''u''<sub>1</sub>, ''u''<sub>2</sub>, 0) again, then ''c'''''u''' = (''cu''<sub>1</sub>, ''cu''<sub>2</sub>, ''c''0) = (''cu''<sub>1</sub>, ''cu''<sub>2</sub>,0). Thus, ''c'''''u''' is an element of ''W'' too.
| |
| | |
| '''Example II:'''
| |
| Let the field be '''R''' again, but now let the vector space be the [[Cartesian plane]] '''R'''<sup>2</sup>.
| |
| Take ''W'' to be the set of points (''x'', ''y'') of '''R'''<sup>2</sup> such that ''x'' = ''y''.
| |
| Then ''W'' is a subspace of '''R'''<sup>2</sup>.
| |
| | |
| ''Proof:''
| |
| #Let '''p''' = (''p''<sub>1</sub>, ''p''<sub>2</sub>) and '''q''' = (''q''<sub>1</sub>, ''q''<sub>2</sub>) be elements of ''W'', that is, points in the plane such that ''p''<sub>1</sub> = ''p''<sub>2</sub> and ''q''<sub>1</sub> = ''q''<sub>2</sub>. Then '''p''' + '''q''' = (''p''<sub>1</sub>+''q''<sub>1</sub>, ''p''<sub>2</sub>+''q''<sub>2</sub>); since ''p''<sub>1</sub> = ''p''<sub>2</sub> and ''q''<sub>1</sub> = ''q''<sub>2</sub>, then ''p''<sub>1</sub> + ''q''<sub>1</sub> = ''p''<sub>2</sub> + ''q''<sub>2</sub>, so '''p''' + '''q''' is an element of ''W''.
| |
| #Let '''p''' = (''p''<sub>1</sub>, ''p''<sub>2</sub>) be an element of ''W'', that is, a point in the plane such that ''p''<sub>1</sub> = ''p''<sub>2</sub>, and let ''c'' be a scalar in '''R'''. Then ''c'''''p''' = (''cp''<sub>1</sub>, ''cp''<sub>2</sub>); since ''p''<sub>1</sub> = ''p''<sub>2</sub>, then ''cp''<sub>1</sub> = ''cp''<sub>2</sub>, so ''c'''''p''' is an element of ''W''.
| |
| | |
| In general, any subset of the [[real coordinate space]] '''R'''<sup>''n''</sup> that is defined by a system of homogeneous [[linear equation]]s will yield a subspace.
| |
| (The equation in example I was ''z'' = 0, and the equation in example II was ''x'' = ''y''.)
| |
| Geometrically, these subspaces are points, lines, planes, and so on, that pass through the point '''0'''.
| |
| | |
| === Examples related to calculus ===
| |
| | |
| '''Example III:'''
| |
| Again take the field to be '''R''', but now let the vector space ''V'' be the set '''R'''<sup>'''R'''</sup> of all [[function (mathematics)|function]]s from '''R''' to '''R'''.
| |
| Let C('''R''') be the subset consisting of [[continuous function|continuous]] functions.
| |
| Then C('''R''') is a subspace of '''R'''<sup>'''R'''</sup>.
| |
| | |
| ''Proof:''
| |
| #We know from calculus that {{nowrap|0 ∈ C('''R''') ⊂ '''R'''<sup>'''R'''</sup>}}.
| |
| #We know from calculus the sum of continuous functions is continuous.
| |
| #Again, we know from calculus that the product of a continuous function and a number is continuous.
| |
| | |
| '''Example IV:'''
| |
| Keep the same field and vector space as before, but now consider the set Diff('''R''') of all [[differentiable]] functions.
| |
| The same sort of argument as before shows that this is a subspace too.
| |
| | |
| Examples that extend these themes are common in [[functional analysis]].
| |
| | |
| == Properties of subspaces ==
| |
| | |
| A way to characterize subspaces is that they are [[Closure (mathematics)|closed]] under [[linear combination]]s.
| |
| That is, a nonempty set ''W'' is a subspace [[if and only if]] every linear combination of ([[finite set|finite]]ly many) elements of ''W'' also belongs to ''W''.
| |
| Conditions 2 and 3 for a subspace are simply the most basic kinds of linear combinations.
| |
| | |
| In a [[topological vector space]] ''X'', a subspace ''W'' need not be [[closed set|closed]] in general, but a [[finite-dimensional]] subspace is always closed.<ref>See {{cite url|title=Basic Facts About Hilbert Space|url=http://www.math.colostate.edu/~pauld/M645/BasicHS.pdf|format=pdf|accessdate=September 17, 2012|author=Paul DuChateau}} for [[Hilbert space]]s</ref> The same is true for subspaces of finite [[codimension]], i.e. determined by a finite number of continuous [[linear functional]]s.
| |
| | |
| ==Descriptions==
| |
| Descriptions of subspaces include the solution set to a homogeneous [[system of linear equations]], the subset of Euclidean space described by a system of homogeneous linear [[parametric equations]], the [[linear span|span]] of a collection of vectors, and the [[null space]], [[column space]], and [[row space]] of a [[matrix (mathematics)|matrix]]. Geometrically (especially, over the field of [[real number]]s and its subfields), a subspace is a [[flat (geometry)|flat]] in an ''n''-space that passes through the origin.
| |
| | |
| A natural description of an 1-subspace is the [[scalar multiplication]] of one non-[[additive identity|zero]] vector '''v''' to all possible scalar values. 1-subspaces specified by two vectors are equal if and only if one vector can be obtained from another with scalar multiplication:
| |
| :<math>\exist c\in K: \mathbf{v}' = c\mathbf{v}\text{ (or }\mathbf{v} = \frac{1}{c}\mathbf{v}'\text{)}</math>
| |
| This idea is generalized for higher dimensions with [[linear span]], but criteria for [[equality (mathematics)|equality]] of ''k''-spaces specified by sets of ''k'' vectors are not so simple.
| |
| | |
| A [[duality (mathematics)|dual]] description is provided with [[linear functional]]s (usually implemented as [[linear equation]]s). One non-[[additive identity|zero]] linear functional '''F''' specifies its [[kernel (linear algebra)|kernel]] subspace '''F''' = 0 of [[codimension]] 1. Subspaces of codimension 1 specified by two linear functionals are equal if and only if one functional can be obtained from another with scalar multiplication (in the [[dual space]]):
| |
| :<math>\exist c\in K: \mathbf{F}' = c\mathbf{F}\text{ (or }\mathbf{F} = \frac{1}{c}\mathbf{F}'\text{)}</math>
| |
| It is generalized for higher codimensions with a [[system of equations]]. The following two subsections will present this latter description in details, and [[#Span of vectors|the remaining]] four subsections further describe the idea of liner span.
| |
| | |
| ===Systems of linear equations===
| |
| The solution set to any homogeneous [[system of linear equations]] with ''n'' variables is a subspace in the [[coordinate space]] ''K''<sup>''n''</sup>:
| |
| | |
| :<math>\left\{ \left[\!\! \begin{array}{c} x_1 \\ x_2 \\ \vdots \\ x_n \end{array} \!\!\right] \in K^n : \begin{alignat}{6}
| |
| a_{11} x_1 &&\; + \;&& a_{12} x_2 &&\; + \cdots + \;&& a_{1n} x_n &&\; = 0& \\
| |
| a_{21} x_1 &&\; + \;&& a_{22} x_2 &&\; + \cdots + \;&& a_{2n} x_n &&\; = 0& \\
| |
| \vdots\;\;\; && && \vdots\;\;\; && && \vdots\;\;\; && \vdots\,& \\
| |
| a_{m1} x_1 &&\; + \;&& a_{m2} x_2 &&\; + \cdots + \;&& a_{mn} x_n &&\; = 0&
| |
| \end{alignat} \right\}. </math>
| |
| | |
| For example (over real or [[rational number]]s), the set of all vectors (''x'', ''y'', ''z'') satisfying the equations
| |
| | |
| :<math>x + 3y + 2z = 0 \;\;\;\;\text{and}\;\;\;\; 2x - 4y + 5z = 0</math>
| |
| | |
| is a one-dimensional subspace. More generally, that is to say that given a set of ''n'' independent functions, the dimension of the subspace in ''K''<sup>''k''</sup> will be the dimension of the [[null set]] of ''A'', the composite matrix of the ''n'' functions.
| |
| | |
| ===Null space of a matrix===
| |
| {{main|Null space}}
| |
| In a finite-dimensional space, a homogeneous system of linear equations can be written as a single [[matrix (mathematics)|matrix]] equation:
| |
| | |
| :<math>A\mathbf{x} = \mathbf{0}.</math>
| |
| | |
| The set of solutions to this equation is known as the [[null space]] of the matrix. For example, the subspace described above is the null space of the matrix
| |
| | |
| :<math>A = \left[ \begin{alignat}{3} 1 && 3 && 2 &\\ 2 && \;\;-4 && \;\;\;\;5 &\end{alignat} \,\right]\text{.}</math>
| |
| | |
| Every subspace of ''K''<sup>''n''</sup> can be described as the null space of some matrix (see [[#Algorithms|algorithms]], below).
| |
| | |
| ===Linear parametric equations===
| |
| The subset of ''K''<sup>''n''</sup> described by a system of homogeneous linear [[parametric equations]] is a subspace:
| |
| | |
| :<math>\left\{ \left[\!\! \begin{array}{c} x_1 \\ x_2 \\ \vdots \\ x_n \end{array} \!\!\right] \in K^n : \begin{alignat}{7}
| |
| x_1 &&\; = \;&& a_{11} t_1 &&\; + \;&& a_{12} t_2 &&\; + \cdots + \;&& a_{1m} t_m & \\
| |
| x_2 &&\; = \;&& a_{21} t_1 &&\; + \;&& a_{22} t_2 &&\; + \cdots + \;&& a_{2m} t_m & \\
| |
| \vdots \,&& && \vdots\;\;\; && && \vdots\;\;\; && && \vdots\;\;\; & \\
| |
| x_n &&\; = \;&& a_{n1} t_1 &&\; + \;&& a_{n2} t_2 &&\; + \cdots + \;&& a_{nm} t_m & \\
| |
| \end{alignat} \text{ for some } t_1,\ldots,t_m\in K \right\}. </math>
| |
| | |
| For example, the set of all vectors (''x'', ''y'', ''z'') parameterized by the equations
| |
| | |
| :<math>x = 2t_1 + 3t_2,\;\;\;\;y = 5t_1 - 4t_2,\;\;\;\;\text{and}\;\;\;\;z = -t_1 + 2t_2</math>
| |
| | |
| is a two-dimensional subspace of ''K''<sup>3</sup>, if ''K'' is a [[number field]] (such as real or rational numbers).<ref name="fields">Generally, ''K'' can be any field of such [[characteristic (algebra)|characteristic]] that the given integer matrix has the appropriate [[rank (matrix theory)|rank]] in it. All fields include [[integer]]s, but some integers may equal to zero in some fields.</ref>
| |
| | |
| ===Span of vectors===
| |
| {{main|Linear span}}
| |
| In linear algebra, the system of parametric equations can be written as a single vector equation:
| |
| | |
| :<math>\begin{bmatrix} x& \\ y& \\ z& \end{bmatrix} \;=\; t_1 \!\begin{bmatrix} 2& \\ 5& \\ -1& \end{bmatrix} + t_2 \!\begin{bmatrix} 3& \\ -4& \\ 2& \end{bmatrix}.</math>
| |
| | |
| The expression on the right is called a [[linear combination]] of the vectors
| |
| (2, 5, −1) and (3, −4, 2). These two vectors are said to '''span''' the resulting subspace.
| |
| | |
| In general, a '''linear combination''' of vectors '''v'''<sub>1</sub>, '''v'''<sub>2</sub>, ... , '''v'''<sub>''k''</sub> is any vector of the form
| |
| | |
| :<math>t_1 \mathbf{v}_1 + \cdots + t_k \mathbf{v}_k.</math>
| |
| | |
| The set of all possible linear combinations is called the '''span''':
| |
| | |
| :<math>\text{Span} \{ \mathbf{v}_1, \ldots, \mathbf{v}_k \}
| |
| = \left\{ t_1 \mathbf{v}_1 + \cdots + t_k \mathbf{v}_k : t_1,\ldots,t_k\in K \right\} .</math>
| |
| | |
| If the vectors '''v'''<sub>1</sub>, ... , '''v'''<sub>''k''</sub> have ''n'' components, then their span is a subspace of ''K''<sup>''n''</sup>. Geometrically, the span is the flat through the origin in ''n''-dimensional space determined by the points '''v'''<sub>1</sub>, ... , '''v'''<sub>''k''</sub>.
| |
| | |
| ; Example
| |
| : The ''xz''-plane in '''R'''<sup>3</sup> can be parameterized by the equations
| |
| ::<math>x = t_1, \;\;\; y = 0, \;\;\; z = t_2.</math>
| |
| | |
| :As a subspace, the ''xz''-plane is spanned by the vectors (1, 0, 0) and (0, 0, 1). Every vector in the ''xz''-plane can be written as a linear combination of these two:
| |
| | |
| ::<math>(t_1, 0, t_2) = t_1(1,0,0) + t_2(0,0,1)\text{.}\,</math>
| |
| | |
| :Geometrically, this corresponds to the fact that every point on the ''xz''-plane can be reached from the origin by first moving some distance in the direction of (1, 0, 0) and then moving some distance in the direction of (0, 0, 1).
| |
| | |
| ===Column space and row space===
| |
| {{main|Row and column spaces}}
| |
| A system of linear parametric equations in a finite-dimensional space can also be written as a single matrix equation:
| |
| | |
| :<math>\mathbf{x} = A\mathbf{t}\;\;\;\;\text{where}\;\;\;\;A = \left[ \begin{alignat}{2} 2 && 3 & \\ 5 && \;\;-4 & \\ -1 && 2 & \end{alignat} \,\right]\text{.}</math>
| |
| | |
| In this case, the subspace consists of all possible values of the vector '''x'''. In linear algebra, this subspace is known as the [[column space]] (or [[image (mathematics)|image]]) of the matrix ''A''. It is precisely the subspace of ''K''<sup>''n''</sup> spanned by the column vectors of ''A''.
| |
| | |
| The [[row space]] of a matrix is the subspace spanned by its row vectors. The row space is interesting because it is the [[orthogonal complement]] of the null space (see below).
| |
| | |
| ===Independence, basis, and dimension===
| |
| {{main|Linear independence|Basis (linear algebra)|Dimension (vector space)}}
| |
| [[File:Basis for a Plane.PNG|thumb|280px|right|The vectors '''u''' and '''v''' are a basis for this two-dimensional subspace of '''R'''<sup>3</sup>.]]
| |
| In general, a subspace of ''K''<sup>''n''</sup> determined by ''k'' parameters (or spanned by ''k'' vectors) has dimension ''k''. However, there are exceptions to this rule. For example, the subspace of ''K''<sup>3</sup> spanned by the three vectors (1, 0, 0), (0, 0, 1), and
| |
| (2, 0, 3) is just the ''xz''-plane, with each point on the plane described by infinitely many different values of {{nowrap| ''t''<sub>1</sub>, ''t''<sub>2</sub>, ''t''<sub>3</sub>}}.
| |
| | |
| In general, vectors '''v'''<sub>1</sub>, ... , '''v'''<sub>''k''</sub> are called '''linearly independent''' if
| |
| | |
| :<math>t_1 \mathbf{v}_1 + \cdots + t_k \mathbf{v}_k \;\ne\; u_1 \mathbf{v}_1 + \cdots + u_k \mathbf{v}_k</math>
| |
| | |
| for
| |
| (''t''<sub>1</sub>, ''t''<sub>2</sub>, ... , ''t<sub>k</sub>'') ≠ (''u''<sub>1</sub>, ''u''<sub>2</sub>, ... , ''u<sub>k</sub>'').<ref>This definition is often stated differently: vectors '''v'''<sub>1</sub>, ..., '''v'''<sub>''k''</sub> are linearly independent if
| |
| {{nowrap| ''t''<sub>1</sub>'''v'''<sub>1</sub> + ··· + ''t<sub>k</sub>'''''v'''<sub>''k''</sub> ≠ '''0'''}} for {{nowrap| (''t''<sub>1</sub>, ''t''<sub>2</sub>, ..., ''t<sub>k</sub>'') ≠ (0, 0, ..., 0)}}. The two definitions are equivalent.</ref>
| |
| If {{nowrap| '''v'''<sub>1</sub>, ..., '''v'''<sub>''k''</sub> }} are linearly independent, then the '''coordinates''' {{nowrap| ''t''<sub>1</sub>, ..., ''t<sub>k</sub>''}} for a vector in the span are uniquely determined.
| |
| | |
| A '''basis''' for a subspace ''S'' is a set of linearly independent vectors whose span is ''S''. The number of elements in a basis is always equal to the geometric dimension of the subspace. Any spanning set for a subspace can be changed into a basis by removing redundant vectors (see [[#Algorithms|algorithms]], below).
| |
| | |
| ; Example
| |
| : Let ''S'' be the subspace of '''R'''<sup>4</sup> defined by the equations
| |
| ::<math>x_1 = 2 x_2\;\;\;\;\text{and}\;\;\;\;x_3 = 5x_4.</math>
| |
| :Then the vectors (2, 1, 0, 0) and (0, 0, 5, 1) are a basis for ''S''. In particular, every vector that satisfies the above equations can be written uniquely as a linear combination of the two basis vectors:
| |
| | |
| ::<math>(2t_1, t_1, 5t_2, t_2) = t_1(2, 1, 0, 0) + t_2(0, 0, 5, 1).\,</math>
| |
| | |
| :The subspace ''S'' is two-dimensional. Geometrically, it is the plane in '''R'''<sup>4</sup> passing through the points (0, 0, 0, 0), (2, 1, 0, 0), and (0, 0, 5, 1).
| |
| | |
| ==Operations and relations on subspaces==
| |
| === Inclusion ===
| |
| <!-- some illustration, please -->
| |
| The [[inclusion relation|set-theoretical inclusion]] binary relation specified a [[partial order]] on the set of all subspaces (of any dimension).
| |
| | |
| A subspace cannot lie in any subspace of lesser dimension. If dim ''U'' = ''k'', a finite number, and ''U'' ⊂ ''W'', then dim ''W'' = ''k'' if and only if ''U'' = ''W''.
| |
| | |
| ===Intersection===
| |
| [[File:IntersectingPlanes.png|thumb|right|In '''R'''<sup>3</sup>, the intersection of two-dimensional subspaces is one-dimensional]]
| |
| Given subspaces ''U'' and ''W'' of a vector space ''V'', then their [[intersection (set theory)|intersection]] ''U'' ∩ ''W'' := {'''v''' ∈ ''V'' : '''v''' is an element of both ''U'' and ''W''} is also a subspace of ''V''.
| |
| | |
| ''Proof:''
| |
| # Let '''v''' and '''w''' be elements of ''U'' ∩ ''W''. Then '''v''' and '''w''' belong to both ''U'' and ''W''. Because ''U'' is a subspace, then '''v''' + '''w''' belongs to ''U''. Similarly, since ''W'' is a subspace, then '''v''' + '''w''' belongs to ''W''. Thus, '''v''' + '''w''' belongs to ''U'' ∩ ''W''.
| |
| # Let '''v''' belong to ''U'' ∩ ''W'', and let ''c'' be a scalar. Then '''v''' belongs to both ''U'' and ''W''. Since ''U'' and ''W'' are subspaces, ''c'''''v''' belongs to both ''U'' and ''W''.
| |
| # Since ''U'' and ''W'' are vector spaces, then '''0''' belongs to both sets. Thus, '''0''' belongs to ''U'' ∩ ''W''.
| |
| | |
| For every vector space ''V'', the [[zero vector space|set {'''0'''}]] and ''V'' itself are subspaces of ''V''.
| |
| | |
| ===Sum===
| |
| If ''U'' and ''W'' are subspaces, their '''sum''' is the subspace
| |
| | |
| :<math>U + W = \left\{ \mathbf{u} + \mathbf{w} \colon \mathbf{u}\in U, \mathbf{w}\in W \right\}.</math>
| |
| | |
| For example, the sum of two lines is the plane that contains them both. The dimension of the sum satisfies the inequality
| |
| | |
| :<math>\max(\dim U,\dim W) \leq \dim(U + W) \leq \dim(U) + \dim(W).</math>
| |
| | |
| Here the minimum only occurs if one subspace is contained in the other, while the maximum is the most general case. The dimension of the intersection and the sum are related:
| |
| | |
| :<math>\dim(U+W) = \dim(U) + \dim(W) - \dim(U \cap W).</math>
| |
| | |
| === Lattice of subspaces ===
| |
| Aforementioned two operations make the set of all subspaces a bounded [[distributive lattice]], where the [[zero vector space|{0} subspace]], the [[least element]], is an [[identity element]] of the sum operation, and the identical subspace ''V'', the greatest element, is an [[identity element]] of the intersection operation.
| |
| | |
| === Other ===
| |
| {{unreferenced section|date=April 2013}}
| |
| If ''V'' is an [[inner product space]], then the [[orthogonal complement]] ⊥ of any subspace of ''V'' is again a subspace. This operation, understood as [[negation]] (¬), makes the lattice of subspaces a (possibly [[infinite set|infinite]]) [[Boolean algebra (structure)|Boolean algebra]].
| |
| | |
| In a [[pseudo-Euclidean space]] there are orthogonal complements too, but such operation does not form a Boolean algebra (nor a [[Heyting algebra]]) because of [[null vector|null]] subspaces, for which ''N'' ∩ ''N''<sup>⊥</sup> = ''N'' ≠ {0}. The same case presents the <sup>⊥</sup> operation in [[symplectic vector space]]s.
| |
| | |
| ==Algorithms==
| |
| Most algorithms for dealing with subspaces involve [[row reduction]]. This is the process of applying [[elementary row operation]]s to a matrix until it reaches either [[row echelon form]] or [[reduced row echelon form]]. Row reduction has the following important properties:
| |
| # The reduced matrix has the same null space as the original.
| |
| # Row reduction does not change the span of the row vectors, i.e. the reduced matrix has the same row space as the original.
| |
| # Row reduction does not affect the linear dependence of the column vectors.
| |
| | |
| ===Basis for a row space===
| |
| :'''Input''' An ''m'' × ''n'' matrix ''A''.
| |
| :'''Output''' A basis for the [[row space]] of ''A''.
| |
| :# Use elementary row operations to put ''A'' into row echelon form.
| |
| :# The nonzero rows of the echelon form are a basis for the row space of ''A''.
| |
| See the article on [[row space]] for an [[Row space#Basis|example]].
| |
| | |
| If we instead put the matrix ''A'' into reduced row echelon form, then the resulting basis for the row space is uniquely determined. This provides an algorithm for checking whether two row spaces are equal and, by extension, whether two subspaces of ''K''<sup>''n''</sup> are equal.
| |
| | |
| ===Subspace membership===
| |
| :'''Input''' A basis {'''b'''<sub>1</sub>, '''b'''<sub>2</sub>, ..., '''b'''<sub>''k''</sub>} for a subspace ''S'' of ''K''<sup>''n''</sup>, and a vector '''v''' with ''n'' components.
| |
| :'''Output''' Determines whether '''v''' is an element of ''S''
| |
| :# Create a (''k'' + 1) × ''n'' matrix ''A'' whose rows are the vectors '''b'''<sub>1</sub>, ... , '''b'''<sub>''k''</sub> and '''v'''.
| |
| :# Use elementary row operations to put ''A'' into row echelon form.
| |
| :# If the echelon form has a row of zeroes, then the vectors {{nowrap| {'''b'''<sub>1</sub>, ..., '''b'''<sub>''k''</sub>, '''v'''} }} are linearly dependent, and therefore {{nowrap| '''v''' ∈ ''S'' }}.
| |
| | |
| ===Basis for a column space===
| |
| :'''Input''' An ''m'' × ''n'' matrix ''A''
| |
| :'''Output''' A basis for the [[column space]] of ''A''
| |
| :# Use elementary row operations to put ''A'' into row echelon form.
| |
| :# Determine which columns of the echelon form have [[Row echelon form|pivots]]. The corresponding columns of the original matrix are a basis for the column space.
| |
| See the article on [[column space]] for an [[Column space#Basis|example]].
| |
| | |
| This produces a basis for the column space that is a subset of the original column vectors. It works because the columns with pivots are a basis for the column space of the echelon form, and row reduction does not change the linear dependence relationships between the columns.
| |
| | |
| ===Coordinates for a vector===
| |
| :'''Input''' A basis {'''b'''<sub>1</sub>, '''b'''<sub>2</sub>, ..., '''b'''<sub>''k''</sub>} for a subspace ''S'' of ''K''<sup>''n''</sup>, and a vector {{nowrap| '''v''' ∈ ''S''}}
| |
| :'''Output''' Numbers ''t''<sub>1</sub>, ''t''<sub>2</sub>, ..., ''t''<sub>''k''</sub> such that {{nowrap|1= '''v''' = ''t''<sub>1</sub>'''b'''<sub>1</sub> + ··· + ''t''<sub>''k''</sub>'''b'''<sub>''k''</sub>}}
| |
| :# Create an [[augmented matrix]] ''A'' whose columns are '''b'''<sub>1</sub>,...,'''b'''<sub>''k''</sub> , with the last column being '''v'''.
| |
| :# Use elementary row operations to put ''A'' into reduced row echelon form.
| |
| :# Express the final column of the reduced echelon form as a linear combination of the first ''k'' columns. The coefficients used are the desired numbers {{nowrap| ''t''<sub>1</sub>, ''t''<sub>2</sub>, ..., ''t''<sub>''k''</sub>}}. (These should be precisely the first ''k'' entries in the final column of the reduced echelon form.)
| |
| If the final column of the reduced row echelon form contains a pivot, then the input vector '''v''' does not lie in ''S''.
| |
| | |
| ===Basis for a null space===
| |
| :'''Input''' An ''m'' × ''n'' matrix ''A''.
| |
| :'''Output''' A basis for the null space of ''A''
| |
| :# Use elementary row operations to put ''A'' in reduced row echelon form.
| |
| :# Using the reduced row echelon form, determine which of the variables {{nowrap| ''x''<sub>1</sub>, ''x''<sub>2</sub>, ..., ''x<sub>n</sub>''}} are free. Write equations for the dependent variables in terms of the free variables.
| |
| :# For each free variable ''x<sub>i</sub>'', choose a vector in the null space for which {{nowrap|1= ''x<sub>i</sub>'' = 1}} and the remaining free variables are zero. The resulting collection of vectors is a basis for the null space of ''A''.
| |
| See the article on [[null space]] for an [[Kernel_(matrix)#Basis|example]].
| |
| | |
| ===Equations for a subspace===
| |
| :'''Input''' A basis {'''b'''<sub>1</sub>, '''b'''<sub>2</sub>, ..., '''b'''<sub>''k''</sub>} for a subspace ''S'' of ''K''<sup>''n''</sup>
| |
| :'''Output''' An (''n'' − ''k'') × ''n'' matrix whose null space is ''S''.
| |
| :# Create a matrix ''A'' whose rows are {{nowrap| '''b'''<sub>1</sub>, '''b'''<sub>2</sub>, ..., '''b'''<sub>''k''</sub>}}.
| |
| :# Use elementary row operations to put ''A'' into reduced row echelon form.
| |
| :# Let {{nowrap| '''c'''<sub>1</sub>, '''c'''<sub>2</sub>, ..., '''c'''<sub>''n''</sub> }} be the columns of the reduced row echelon form. For each column without a pivot, write an equation expressing the column as a linear combination of the columns with pivots.
| |
| :# This results in a homogeneous system of ''n'' − ''k'' linear equations involving the variables '''c'''<sub>1</sub>,...,'''c'''<sub>''n''</sub>. The {{nowrap| (''n'' − ''k'') × ''n''}} matrix corresponding to this system is the desired matrix with nullspace ''S''.
| |
| ; Example
| |
| :If the reduced row echelon form of ''A'' is
| |
| | |
| ::<math>\left[ \begin{alignat}{6}
| |
| 1 && 0 && -3 && 0 && 2 && 0 \\
| |
| 0 && 1 && 5 && 0 && -1 && 4 \\
| |
| 0 && 0 && 0 && 1 && 7 && -9 \\
| |
| 0 && \;\;\;\;\;0 && \;\;\;\;\;0 && \;\;\;\;\;0 && \;\;\;\;\;0 && \;\;\;\;\;0 \end{alignat} \,\right] </math>
| |
| | |
| :then the column vectors {{nowrap| '''c'''<sub>1</sub>, ..., '''c'''<sub>6</sub>}} satisfy the equations
| |
| | |
| ::<math> \begin{alignat}{1}
| |
| \mathbf{c}_3 &= -3\mathbf{c}_1 + 5\mathbf{c}_2 \\
| |
| \mathbf{c}_5 &= 2\mathbf{c}_1 - \mathbf{c}_2 + 7\mathbf{c}_3 \\
| |
| \mathbf{c}_6 &= 4\mathbf{c}_2 - 9\mathbf{c}_3.
| |
| \end{alignat}</math>
| |
| | |
| :It follows that the row vectors of ''A'' satisfy the equations
| |
| | |
| ::<math> \begin{alignat}{1}
| |
| x_3 &= -3x_1 + 5x_2 \\
| |
| x_5 &= 2x_1 - x_2 + 7x_3 \\
| |
| x_6 &= 4x_2 - 9x_3.
| |
| \end{alignat}</math>
| |
| | |
| :In particular, the row vectors of ''A'' are a basis for the null space of the corresponding matrix.
| |
| | |
| ==See also==
| |
| * [[Signal subspace]]
| |
| * [[Multilinear subspace learning]]
| |
| * [[Cyclic subspace]]
| |
| | |
| ==Textbooks==
| |
| * {{Citation
| |
| | last = Axler
| |
| | first = Sheldon Jay
| |
| | year = 1997
| |
| | title = Linear Algebra Done Right
| |
| | publisher = Springer-Verlag
| |
| | edition = 2nd
| |
| | isbn = 0-387-98259-0
| |
| }}
| |
| * {{Citation
| |
| | last = Lay
| |
| | first = David C.
| |
| | date = August 22, 2005
| |
| | title = Linear Algebra and Its Applications
| |
| | publisher = Addison Wesley
| |
| | edition = 3rd
| |
| | isbn = 978-0-321-28713-7
| |
| }}
| |
| * {{Citation
| |
| | last = Meyer
| |
| | first = Carl D.
| |
| | date = February 15, 2001
| |
| | title = Matrix Analysis and Applied Linear Algebra
| |
| | publisher = Society for Industrial and Applied Mathematics (SIAM)
| |
| | isbn = 978-0-89871-454-8
| |
| | url = http://www.matrixanalysis.com/DownloadChapters.html
| |
| }}
| |
| * {{Citation
| |
| | last = Poole
| |
| | first = David
| |
| | year = 2006
| |
| | title = Linear Algebra: A Modern Introduction
| |
| | publisher = Brooks/Cole
| |
| | edition = 2nd
| |
| | isbn = 0-534-99845-3
| |
| }}
| |
| * {{Citation
| |
| | last = Anton
| |
| | first = Howard
| |
| | year = 2005
| |
| | title = Elementary Linear Algebra (Applications Version)
| |
| | publisher = Wiley International
| |
| | edition = 9th
| |
| }}
| |
| * {{Citation
| |
| | last = Leon
| |
| | first = Steven J.
| |
| | year = 2006
| |
| | title = Linear Algebra With Applications
| |
| | publisher = Pearson Prentice Hall
| |
| | edition = 7th
| |
| }}
| |
| | |
| ==External links==
| |
| * {{planetmath reference|id=624|title=Vector subspace}}.
| |
| * {{aut|[[Gilbert Strang]]}}, [http://ocw.mit.edu/OcwWeb/Mathematics/18-06Spring-2005/VideoLectures/detail/lecture10.htm MIT Linear Algebra Lecture on the Four Fundamental Subspaces] at Google Video, from [[MIT OpenCourseWare]]
| |
| | |
| ==References==
| |
| {{Reflist}}
| |
| | |
| {{DEFAULTSORT:Linear Subspace}}
| |
| [[Category:Linear algebra]]
| |
| [[Category:Articles containing proofs]]
| |
| [[Category:Operator theory]]
| |
| [[Category:Functional analysis]]
| |
| | |
| [[ru:Векторное подпространство]]
| |