Jump to content

Row and column vectors: Difference between revisions

From Wikipedia, the free encyclopedia
Content deleted Content added
Sh3l6orrr (talk | contribs)
No edit summary
Madyno (talk | contribs)
restore, a colum vector is not a vector in the normal sense, but a matrix
Line 1: Line 1:
In [[linear algebra]], a '''column vector''' or '''column matrix''' is an ''m'' × 1 [[matrix (mathematics)|matrix]], that is, a matrix consisting of a single column of ''m'' elements,
{{One source|date=May 2021}}
In [[linear algebra]], a '''column vector''' is a column of entries, for example,


:<math>\boldsymbol{x} = \begin{bmatrix} x_1 \\ x_2 \\ \vdots \\ x_m \end{bmatrix} \,. </math>
:<math>\boldsymbol{x} = \begin{bmatrix} x_1 \\ x_2 \\ \vdots \\ x_m \end{bmatrix} \,. </math>


Similarly, a '''row vector''' is a row of entries<ref>{{harvtxt|Meyer|2000}}, p. 8</ref>
Similarly, a '''row vector''' or '''row matrix''' is a 1 &times; ''m'' [[matrix (mathematics)|matrix]], that is, a matrix consisting of a single row of ''m'' elements<ref>{{harvtxt|Meyer|2000}}, p. 8</ref>


:<math>\boldsymbol a = \begin{bmatrix} a_1 & a_2 & \dots & a_n \end{bmatrix} \,. </math>
:<math>\boldsymbol x = \begin{bmatrix} x_1 & x_2 & \dots & x_m \end{bmatrix} \,. </math>


Throughout, boldface is used for both row and column vectors. The [[transpose]] (indicated by T) of a row vector is the column vector
Throughout, boldface is used for the row and column vectors. The [[transpose]] (indicated by T) of a row vector is a column vector


:<math>\begin{bmatrix} x_1 \; x_2 \; \dots \; x_m \end{bmatrix}^{\rm T} = \begin{bmatrix} x_1 \\ x_2 \\ \vdots \\ x_m \end{bmatrix} \,,</math>
:<math>\begin{bmatrix} x_1 \; x_2 \; \dots \; x_m \end{bmatrix}^{\rm T} = \begin{bmatrix} x_1 \\ x_2 \\ \vdots \\ x_m \end{bmatrix} \,,</math>


and the transpose of a column vector is the row vector
and the transpose of a column vector is a row vector


:<math>\begin{bmatrix} x_1 \\ x_2 \\ \vdots \\ x_m \end{bmatrix}^{\rm T} = \begin{bmatrix} x_1 \; x_2 \; \dots \; x_m \end{bmatrix} \,.</math>
:<math>\begin{bmatrix} x_1 \\ x_2 \\ \vdots \\ x_m \end{bmatrix}^{\rm T} = \begin{bmatrix} x_1 \; x_2 \; \dots \; x_m \end{bmatrix} \,.</math>


The set of all row vectors with ''n'' entries forms an ''n''-dimensional [[vector space]]; similarly, the set of all column vectors with ''m'' entries forms an ''m''-dimensional vector space.
The set of all row vectors forms a [[vector space]] called '''row space'''; similarly, the set of all column vectors forms a vector space called '''column space'''. The dimensions of the [[row and column spaces]] equals the number of entries in the row or column vector.


The space of row vectors with ''n'' entries can be regarded as the [[dual space]] of the space of column vectors with ''n'' entries, since any linear functional on the space of column vectors can be represented as the left-multiplication of a unique row vector.
The column space can be viewed as the [[dual space]] to the row space, since any linear functional on the space of column vectors can be represented uniquely as an [[inner product]] with a specific row vector.


== Notation ==
== Notation ==
Line 30: Line 29:
:<math>\boldsymbol{x} = \begin{bmatrix} x_1, x_2, \dots, x_m \end{bmatrix}^{\rm T}</math>
:<math>\boldsymbol{x} = \begin{bmatrix} x_1, x_2, \dots, x_m \end{bmatrix}^{\rm T}</math>


Some authors also use the convention of writing both column vectors and row vectors as rows, but separating row vector elements with [[comma]]s and column vector elements with [[semicolon]]s (see alternative notation 2 in the table below).{{fact|date=March 2021}}
Some authors also use the convention of writing both column vectors and row vectors as rows, but separating row vector elements with [[comma]]s and column vector elements with [[semicolon]]s (see alternative notation 2 in the table below).
{| class="wikitable"
{| class="wikitable"
|-
|-
Line 52: Line 51:
[[Matrix multiplication]] involves the action of multiplying each row vector of one matrix by each column vector of another matrix.
[[Matrix multiplication]] involves the action of multiplying each row vector of one matrix by each column vector of another matrix.


The [[dot product]] of two column vectors '''a''' and '''b''' is equivalent to the matrix product of the transpose of '''a''' with '''b''',
The [[dot product]] of two vectors '''a''' and '''b''' is equivalent to the matrix product of the row vector representation of '''a''' and the column vector representation of '''b''',


:<math>\mathbf{a} \cdot \mathbf{b} = \mathbf{a}^\intercal \mathbf{b} = \begin{bmatrix}
:<math>\mathbf{a} \cdot \mathbf{b} = \mathbf{a} \mathbf{b}^\mathrm{T} = \begin{bmatrix}
a_1 & \cdots & a_n
a_1 & a_2 & a_3
\end{bmatrix}\begin{bmatrix}
\end{bmatrix}\begin{bmatrix}
b_1 \\ \vdots \\ b_n
b_1 \\ b_2 \\ b_3
\end{bmatrix} = a_1 b_1 + \cdots + a_n b_n \,, </math>
\end{bmatrix} = a_1 b_1 + a_2 b_2 + a_3 b_3 \,, </math>


By the symmetry of the dot product, the [[dot product]] of two column vectors '''a''' and '''b''' is also equivalent to the matrix product of the transpose of '''b''' with '''a''',
which is also equivalent to the matrix product of the row vector representation of '''b''' and the column vector representation of '''a''',


:<math>\mathbf{b} \cdot \mathbf{a} = \mathbf{b}^\intercal \mathbf{a} = \begin{bmatrix}
:<math>\mathbf{b} \cdot \mathbf{a} = \mathbf{b} \mathbf{a}^\mathrm{T} = \begin{bmatrix}
b_1 & \cdots & b_n
b_1 & b_2 & b_3
\end{bmatrix}\begin{bmatrix}
\end{bmatrix}\begin{bmatrix}
a_1 \\ \vdots \\ a_n
a_1 \\ a_2 \\ a_3
\end{bmatrix} = a_1 b_1 + \cdots + a_n b_n\,. </math>
\end{bmatrix}\,. </math>


The matrix product of a column and a row vector gives the [[outer product]] of two vectors '''a''' and '''b''', an example of the more general [[tensor product]]. The matrix product of the column vector representation of '''a''' and the row vector representation of '''b''' gives the components of their dyadic product,
The matrix product of a column and a row vector gives the [[outer product]] of two vectors '''a''' and '''b''', an example of the more general [[tensor product]]. The matrix product of the column vector representation of '''a''' and the row vector representation of '''b''' gives the components of their dyadic product,


:<math>\mathbf{a} \otimes \mathbf{b} = \mathbf{a} \mathbf{b}^\intercal = \begin{bmatrix}
:<math>\mathbf{a} \otimes \mathbf{b} = \mathbf{a}^\mathrm{T} \mathbf{b} = \begin{bmatrix}
a_1 \\ a_2 \\ a_3
a_1 \\ a_2 \\ a_3
\end{bmatrix}\begin{bmatrix}
\end{bmatrix}\begin{bmatrix}
Line 82: Line 81:
which is the [[transpose]] of the matrix product of the column vector representation of '''b''' and the row vector representation of '''a''',
which is the [[transpose]] of the matrix product of the column vector representation of '''b''' and the row vector representation of '''a''',


:<math>\mathbf{b} \otimes \mathbf{a} = \mathbf{b} \mathbf{a}^\intercal = \begin{bmatrix}
:<math>\mathbf{b} \otimes \mathbf{a} = \mathbf{b}^\mathrm{T} \mathbf{a} = \begin{bmatrix}
b_1 \\ b_2 \\ b_3
b_1 \\ b_2 \\ b_3
\end{bmatrix}\begin{bmatrix}
\end{bmatrix}\begin{bmatrix}
Line 92: Line 91:
\end{bmatrix} \,. </math>
\end{bmatrix} \,. </math>


==Matrix transformations==
==Preferred input vectors for matrix transformations==
{{POV|date=February 2021}}
{{main|Transformation matrix}}
Frequently a row vector presents itself for an operation within ''n''-space expressed by an ''n'' × ''n'' matrix ''M'',
An ''n'' × ''n'' matrix ''M'' can represent a [[linear map]] and act on row and column vectors as the linear map's [[transformation matrix]]. For a row vector ''v'', the product ''vM'' is another row vector ''p'':


:<math> v M = p \,.</math>
:<math> v M = p \,.</math>


Another ''n'' × ''n'' matrix ''Q'' can act on ''p'',
Then ''p'' is also a row vector and may present to another ''n'' × ''n'' matrix ''Q'',


:<math> p Q = t \,. </math>
:<math> p Q = t \,. </math>


Then one can write ''t'' = ''p Q'' = ''v MQ'', so the [[matrix product]] transformation ''MQ'' maps ''v'' directly to ''t''. Continuing with row vectors, matrix transformations further reconfiguring ''n''-space can be applied to the right of previous outputs.
Conveniently, one can write ''t'' = ''p Q'' = ''v MQ'' telling us that the [[matrix product]] transformation ''MQ'' can take ''v'' directly to ''t''. Continuing with row vectors, matrix transformations further reconfiguring ''n''-space can be applied to the right of previous outputs.


When a column vector is transformed to another column vector under an ''n'' × ''n'' matrix action, the operation occurs to the left,
In contrast, when a column vector is transformed to become another column under an ''n'' × ''n'' matrix action, the operation occurs to the left,


:<math> p^\mathrm{T} = M v^\mathrm{T} \,,\quad t^\mathrm{T} = Q p^\mathrm{T} </math>,
:<math> p^\mathrm{T} = M v^\mathrm{T} \,,\quad t^\mathrm{T} = Q p^\mathrm{T} </math>,


leading to the algebraic expression ''QM v''<sup>T</sup> for the composed output from ''v''<sup>T</sup> input. The matrix transformations mount up to the left in this use of a column vector for input to matrix transformation.
leading to the algebraic expression ''QM v''<sup>T</sup> for the composed output from ''v''<sup>T</sup> input. The matrix transformations mount up to the left in this use of a column vector for input to matrix transformation.

The column vector approach to matrix transformation leads to a [[right-to-left]] orientation for successive transformations. In [[geometric transformation]]s described by matrices, the two approaches are related by the [[transpose]] operator. Though equivalent, the fact that directionality of English text is [[left-to-right]] has led some English authors to have a preference for the row vector input to matrix transformation:

For instance, this row vector input convention has been used to good effect by Raiz Usmani,<ref>Raiz A. Usmani (1987) ''Applied Linear Algebra'' [[Marcel Dekker]] {{isbn|0824776224}}. See Chapter 4: "Linear Transformations"</ref> where on page 106 the convention allows the statement "The product mapping ''ST'' of ''U'' into ''W'' [is given] by:
:<math>\alpha (ST) = (\alpha S) T = \beta T = \gamma</math>."
(The Greek letters represent row vectors).

[[Ludwik Silberstein]] used row vectors for spacetime events; he applied Lorentz transformation matrices on the right in his [[List of important publications in physics#Special|Theory of Relativity]] in 1914 (see page 143).
In 1963 when [[McGraw-Hill]] published ''Differential Geometry'' by [[Heinrich Guggenheimer]] of the [[University of Minnesota]], he used the row vector convention in chapter 5, "Introduction to transformation groups" (eqs. 7a,9b and 12 to 15). When [[H. S. M. Coxeter]] reviewed<ref>Coxeter [http://www.ams.org/mathscinet/pdf/188842.pdf Review of ''Linear Geometry''] from [[Mathematical Reviews]]</ref> ''Linear Geometry'' by [[Rafael Artzy]], he wrote, "[Artzy] is to be congratulated on his choice of the 'left-to-right' convention, which enables him to regard a point as a row matrix instead of the clumsy column that many authors prefer." [[J. W. P. Hirschfeld]] used right multiplication of row vectors by matrices in his description of projectivities on the [[Galois geometry]] PG(1,q).<ref>[[J. W. P. Hirschfeld]] (1979) ''Projective Geometry over Finite Fields'', page 119, [[Clarendon Press]] {{isbn|0-19-853526-0}}</ref>

In the study of stochastic processes with a [[stochastic matrix]], it is conventional to use a row vector as the [[stochastic vector]].<ref>[[John G. Kemeny]] & [[J. Laurie Snell]] (1960) ''Finite Markov Chains'', page 33, D. Van Nostrand Company</ref>


== See also ==
== See also ==

Revision as of 14:10, 29 November 2022

In linear algebra, a column vector or column matrix is an m × 1 matrix, that is, a matrix consisting of a single column of m elements,

Similarly, a row vector or row matrix is a 1 × m matrix, that is, a matrix consisting of a single row of m elements[1]

Throughout, boldface is used for the row and column vectors. The transpose (indicated by T) of a row vector is a column vector

and the transpose of a column vector is a row vector

The set of all row vectors forms a vector space called row space; similarly, the set of all column vectors forms a vector space called column space. The dimensions of the row and column spaces equals the number of entries in the row or column vector.

The column space can be viewed as the dual space to the row space, since any linear functional on the space of column vectors can be represented uniquely as an inner product with a specific row vector.

Notation

To simplify writing column vectors in-line with other text, sometimes they are written as row vectors with the transpose operation applied to them.

or

Some authors also use the convention of writing both column vectors and row vectors as rows, but separating row vector elements with commas and column vector elements with semicolons (see alternative notation 2 in the table below).

Row vector Column vector
Standard matrix notation
(array spaces, no commas, transpose signs)
Alternative notation 1
(commas, transpose signs)
Alternative notation 2
(commas and semicolons, no transpose signs)

Operations

Matrix multiplication involves the action of multiplying each row vector of one matrix by each column vector of another matrix.

The dot product of two vectors a and b is equivalent to the matrix product of the row vector representation of a and the column vector representation of b,

which is also equivalent to the matrix product of the row vector representation of b and the column vector representation of a,

The matrix product of a column and a row vector gives the outer product of two vectors a and b, an example of the more general tensor product. The matrix product of the column vector representation of a and the row vector representation of b gives the components of their dyadic product,

which is the transpose of the matrix product of the column vector representation of b and the row vector representation of a,

Preferred input vectors for matrix transformations

Frequently a row vector presents itself for an operation within n-space expressed by an n × n matrix M,

Then p is also a row vector and may present to another n × n matrix Q,

Conveniently, one can write t = p Q = v MQ telling us that the matrix product transformation MQ can take v directly to t. Continuing with row vectors, matrix transformations further reconfiguring n-space can be applied to the right of previous outputs.

In contrast, when a column vector is transformed to become another column under an n × n matrix action, the operation occurs to the left,

,

leading to the algebraic expression QM vT for the composed output from vT input. The matrix transformations mount up to the left in this use of a column vector for input to matrix transformation.

The column vector approach to matrix transformation leads to a right-to-left orientation for successive transformations. In geometric transformations described by matrices, the two approaches are related by the transpose operator. Though equivalent, the fact that directionality of English text is left-to-right has led some English authors to have a preference for the row vector input to matrix transformation:

For instance, this row vector input convention has been used to good effect by Raiz Usmani,[2] where on page 106 the convention allows the statement "The product mapping ST of U into W [is given] by:

."

(The Greek letters represent row vectors).

Ludwik Silberstein used row vectors for spacetime events; he applied Lorentz transformation matrices on the right in his Theory of Relativity in 1914 (see page 143). In 1963 when McGraw-Hill published Differential Geometry by Heinrich Guggenheimer of the University of Minnesota, he used the row vector convention in chapter 5, "Introduction to transformation groups" (eqs. 7a,9b and 12 to 15). When H. S. M. Coxeter reviewed[3] Linear Geometry by Rafael Artzy, he wrote, "[Artzy] is to be congratulated on his choice of the 'left-to-right' convention, which enables him to regard a point as a row matrix instead of the clumsy column that many authors prefer." J. W. P. Hirschfeld used right multiplication of row vectors by matrices in his description of projectivities on the Galois geometry PG(1,q).[4]

In the study of stochastic processes with a stochastic matrix, it is conventional to use a row vector as the stochastic vector.[5]

See also

Notes

  1. ^ Meyer (2000), p. 8
  2. ^ Raiz A. Usmani (1987) Applied Linear Algebra Marcel Dekker ISBN 0824776224. See Chapter 4: "Linear Transformations"
  3. ^ Coxeter Review of Linear Geometry from Mathematical Reviews
  4. ^ J. W. P. Hirschfeld (1979) Projective Geometry over Finite Fields, page 119, Clarendon Press ISBN 0-19-853526-0
  5. ^ John G. Kemeny & J. Laurie Snell (1960) Finite Markov Chains, page 33, D. Van Nostrand Company

References

  • Axler, Sheldon Jay (1997), Linear Algebra Done Right (2nd ed.), Springer-Verlag, ISBN 0-387-98259-0
  • Lay, David C. (August 22, 2005), Linear Algebra and Its Applications (3rd ed.), Addison Wesley, ISBN 978-0-321-28713-7
  • Meyer, Carl D. (February 15, 2001), Matrix Analysis and Applied Linear Algebra, Society for Industrial and Applied Mathematics (SIAM), ISBN 978-0-89871-454-8, archived from the original on March 1, 2001
  • Poole, David (2006), Linear Algebra: A Modern Introduction (2nd ed.), Brooks/Cole, ISBN 0-534-99845-3
  • Anton, Howard (2005), Elementary Linear Algebra (Applications Version) (9th ed.), Wiley International
  • Leon, Steven J. (2006), Linear Algebra With Applications (7th ed.), Pearson Prentice Hall