The conjugate gradient algorithm can be interpreted as the following optimization problem: minimize defined by
where and are symmetric and positive definite.
At each iteration is minimized along an A-conjugate direction, constructing orthogonal residuals:
where is a Krylov subspace:
Minimum residual algorithms work by minimizing the Euclidean norm over . At each iteration, is the vector in that gives the smallest residual.
The biconjugate gradient algorithm belongs to a more general class of Petrov-Galerkin methods, where orthogonality is enforced in a different i-dimensional subspace ( remains in ):