language-icon Old Web
English
Sign In

Orthogonalization

In linear algebra, orthogonalization is the process of finding a set of orthogonal vectors that span a particular subspace. Formally, starting with a linearly independent set of vectors {v1, ... , vk} in an inner product space (most commonly the Euclidean space Rn), orthogonalization results in a set of orthogonal vectors {u1, ... , uk} that generate the same subspace as the vectors v1, ... , vk. Every vector in the new set is orthogonal to every other vector in the new set; and the new set and the old set have the same linear span. In linear algebra, orthogonalization is the process of finding a set of orthogonal vectors that span a particular subspace. Formally, starting with a linearly independent set of vectors {v1, ... , vk} in an inner product space (most commonly the Euclidean space Rn), orthogonalization results in a set of orthogonal vectors {u1, ... , uk} that generate the same subspace as the vectors v1, ... , vk. Every vector in the new set is orthogonal to every other vector in the new set; and the new set and the old set have the same linear span. In addition, if we want the resulting vectors to all be unit vectors, then the procedure is called orthonormalization. Orthogonalization is also possible with respect to any symmetric bilinear form (not necessarily an inner product, not necessarily over real numbers), but standard algorithms may encounter division by zero in this more general setting.

[ "Algorithm", "Applied mathematics", "Mathematical optimization", "Matrix (mathematics)", "Geometry", "Gram–Schmidt process" ]
Parent Topic
Child Topic
    No Parent Topic