Cholesky Decomposition Calculator
Welcome to the Cholesky decomposition calculator. In this accompanying text to the tool, we'll learn all there is to know about the Cholesky factorization, which decomposes a matrix into a product of matrices. We'll specifically cover how to calculate the Cholesky decomposition and an example of Cholesky decomposition for a matrix.
What is a matrix decomposition?
Before we can learn about the Cholesky decomposition, we must define what a matrix decomposition is. In the world of linear algebra, a matrix decomposition (or a matrix factorization) is the factorization of a matrix into a product of matrices. So, just as you can factorize into groups of products (like or ) you can factorize a matrix like below… but how?
Factorizing a matrix is much harder than a number. Lucky for us, mathematicians have discovered many different methods of performing matrix decompositions. The most famous of these methods are the LU decomposition, the QR decomposition, the singular value decomposition (SVD), and the Cholesky decomposition. Which method you'd want to use depends on the problem you're trying to solve.
What is the Cholesky decomposition?
Knowing what matrix decomposition is, we can go on to define the Cholesky decomposition. The Cholesky decomposition (or the Cholesky factorization) is the factorization of a matrix into the product of a lower triangular matrix and its transpose. We can rewrite this decomposition in mathematical notation as:
To be Cholesky-decomposed, matrix needs to adhere to some criteria:
-
must be symmetric, i.e. .
-
By extension, this means must be square.
-
must be positive definite (meaning its eigenvalues must all be positive).
If doesn't tick all the items on our list, no suitable can exist.
How to calculate the Cholesky decomposition?
The goal of any matrix decomposition method is to find the factorization's terms, and so we want to find the lower triangular matrix . The Cholesky decomposition has no single mathematical formula, but it is easily obtained by hand for a small matrix. For larger matrices, there's an algorithmic process to follow.
Let's begin by looking at the simple matrix case symbolically. We first define:
…and:
Note that the elements of above its diagonal are zero, as is a lower triangular matrix.
We know from the definition of the Cholesky factorization that , so let's take a look at the right-hand side of this equation.
Knowing the above and that , we can just equate corresponding elements of and and solve the equations:
Notice that we need earlier elements of to solve for the later elements: needs , which needs .
We can use the equations above to solve for the matrix that we defined at the top:
The principles of the above example apply to the Cholesky factorization of any sized matrix. For larger matrices, we can generalize the process with the following two equations.
For elements on 's diagonal:
For elements off 's diagonal:
How to use the Cholesky decomposition calculator?
With the help of our calculator, you can easily calculate if you know what is. Remember, must be symmetric and positive definite. If it's not, the calculator can't give you as no that complies with exists.
-
Select 's shape. Our Cholesky decomposition solver supports , , and matrices.
-
Give the calculator your matrix . The matrix's elements are separated by row — see the graphical representation at the top of the calculator if you're unsure.
-
Find the result below. Our Cholesky decomposition solver will calculate and display it below your matrix. If is not symmetric and positive definite, the calculator will notify you accordingly.
Example Cholesky decomposition of a 3×3 matrix
Now that we know how to compute the Cholesky factorization for a matrix of any size, let's do an example. We'll use the general-case algorithmic equations we've just discussed to solve given the following matrix :
We can jump straight into the solving process, starting at the top left and moving left-to-right, top-to-bottom.
And so, after all our effort, we've fully obtained ,
Finally, to test our answer, we can see if really is equal to :
Perfect! We've just performed a complete decomposition of . You can also try out this example in our Cholesky decomposition calculator.
What are the applications of the Cholesky decomposition?
Matrix decomposition methods exploit the structure of the factorization's terms to make solving system of equations easier. Remember how we said that is lower triangular? Lower triangular matrices are especially easy to work with, and therefore the Cholesky decomposition is frequently the method of choice in solving systems of equations.
FAQ
How do you determine whether a matrix has a Cholesky decomposition?
For the matrix A to have a Cholesky decomposition, it must be symmetric, and it must be positive definite (meaning it must have only positive eigenvalues). If A does not adhere to these requirements, it cannot have a Cholesky decomposition, meaning no matrix L that satisfies L · LT = A can exist.
What does the Cholesky decomposition do?
Like any matrix decomposition method, the Cholesky decomposition takes a matrix A and factorizes it. It produces a lower triangular matrix L, which when multiplied with its transpose LT produces the original matrix A. This is valuable in many matrix operations, as the structure of a lower triangular matrix can be exploited to make the operations compute much faster.
What is a symmetric matrix?
A symmetric matrix is one that is equal to its transpose. Redefined mathematically, symmetric matrices satisfy the condition A = AT. This has the implication that the matrix must also be square. Graphically, a symmetric matrix is symmetric along its main diagonal.
What is a positive definite matrix?
A positive definite matrix is a matrix with only positive eigenvalues. The formal mathematical definition is that matrix A is positive definite if z⃗T·A·z⃗ > 0 for every nonzero column vector z⃗. Alternatively, A is only positive definite if it has a Cholesky decomposition.
What is the Cholesky decomposition of the identity matrix?
The identity matrix's Cholesky decomposition is also the identity matrix. This is because the identity matrix's transpose is itself (ɪT = ɪ) and so ɪ · ɪT = ɪ·ɪ = ɪ. In general, the Cholesky decomposition L of a diagonal matrix D is also diagonal, and its diagonal entries are the square roots of D's. When the diagonal entries are all 1 (like they are in ɪ), we get the identity matrix back.
A | = |
|