language-icon Old Web
English
Sign In

Sherman–Morrison formula

In mathematics, in particular linear algebra, the Sherman–Morrison formula, named after Jack Sherman and Winifred J. Morrison, computes the inverse of the sum of an invertible matrix A {displaystyle A} and the outer product, u v T {displaystyle uv^{T}} , of vectors u {displaystyle u} and v {displaystyle v} . The Sherman–Morrison formula is a special case of the Woodbury formula. Though named after Sherman and Morrison, it appeared already in earlier publications. Suppose A ∈ R n × n {displaystyle Ain mathbb {R} ^{n imes n}} is an invertible square matrix and u {displaystyle u} , v ∈ R n {displaystyle vin mathbb {R} ^{n}} are column vectors. Then A + u v T {displaystyle A+uv^{T}} is invertible iff 1 + v T A − 1 u ≠ 0 {displaystyle 1+v^{T}A^{-1}u eq 0} . In this case, Here, u v T {displaystyle uv^{T}} is the outer product of two vectors u {displaystyle u} and v {displaystyle v} . The general form shown here is the one published by Bartlett. ( ⇐ {displaystyle Leftarrow } ) To prove that the backward direction ( 1 + v T A − 1 u ≠ 0 ⇒ A + u v T {displaystyle 1+v^{T}A^{-1}u eq 0Rightarrow A+uv^{T}} is invertible with inverse given as above) is true, we verify the properties of the inverse. A matrix Y {displaystyle Y} (in this case the right-hand side of the Sherman–Morrison formula) is the inverse of a matrix X {displaystyle X} (in this case A + u v T {displaystyle A+uv^{T}} ) if and only if X Y = I {displaystyle XY=I} . We first verify that the right hand side ( Y {displaystyle Y} ) satisfies X Y = I {displaystyle XY=I} . ( ⇒ {displaystyle Rightarrow } ) Reciprocally, if 1 + v T A − 1 u = 0 {displaystyle 1+v^{T}A^{-1}u=0} , then letting x = A − 1 u {displaystyle x=A^{-1}u} , one has ( A + u v T ) x = 0 {displaystyle (A+uv^{T})x=0} thus ( A + u v T ) {displaystyle (A+uv^{T})} has a non-trivial kernel and is therefore not invertible. If the inverse of A {displaystyle A} is already known, the formula provides anumerically cheap wayto compute the inverse of A {displaystyle A} corrected by the matrix u v T {displaystyle uv^{T}} (depending on the point of view, the correction may be seen as a perturbation or as a rank-1 update).The computation is relatively cheap because the inverse of A + u v T {displaystyle A+uv^{T}} does not have to be computed from scratch (which in general is expensive),but can be computed by correcting (or perturbing) A − 1 {displaystyle A^{-1}} . Using unit columns (columns from the identity matrix) for u {displaystyle u} or v {displaystyle v} , individual columns or rows of A {displaystyle A} may be manipulatedand a correspondingly updated inverse computed relatively cheaply in this way. In the general case, where A − 1 {displaystyle A^{-1}} is a n {displaystyle n} -by- n {displaystyle n} matrixand u {displaystyle u} and v {displaystyle v} are arbitrary vectors of dimension n {displaystyle n} , the whole matrix is updated and the computation takes 3 n 2 {displaystyle 3n^{2}} scalar multiplications. If u {displaystyle u} is a unit column, the computation takes only 2 n 2 {displaystyle 2n^{2}} scalar multiplications. The same goes if v {displaystyle v} is a unit column. If both u {displaystyle u} and v {displaystyle v} are unit columns, the computation takes only n 2 {displaystyle n^{2}} scalar multiplications.

[ "Linear system", "Numerical analysis", "Matrix (mathematics)", "Inverse" ]
Parent Topic
Child Topic
    No Parent Topic