## Iterative Methods: Part 2

Last time, we looked at 2D Poisson’s equation and discussed how to arrive at a matrix equation using the finite difference method. The matrix, which represents the discrete Laplace operator, is sparse, so we can use an iterative method to solve the equation efficiently.

Today, we will look at Jacobi, Gauss-Seidel, Successive Over-Relaxation (SOR), and Symmetric SOR (SSOR), and a couple of related techniques—red-black ordering and Chebyshev acceleration. In addition, we will analyze the convergence of each method for the Poisson’s equation, both analytically and numerically.

## Iterative Methods: Part 1

Once again, we consider solving the equation $A\vec{x} = \vec{b}$, where $A \in \mathbb{R}^{n \times n}$ is nonsingular. When we use a factorization method for dense matrices, such as LU, Cholesky, SVD, and QR, we make $O(n^{3})$ operations and store $O(n^{2})$ matrix entries. As a result, it’s hard to accurately model a real-life problem by adding more variables.

Luckily, when the matrix is sparse (e.g. because we used finite difference method or finite element method), we can use an iterative method to approximate the solution efficiently. Typically, an iterative method incurs $O(n^{2})$ operations, if not fewer, and requires $O(n)$ storage.

Over the next few posts, we will look at 5 iterative methods:

• Jacobi
• Gauss-Seidel
• Successive Over-Relaxation (SOR)
• Symmetric SOR (SSOR)
• Conjugate Gradient (CG).

In addition, we will solve 2D Poisson’s equation using these methods. We will compare the approximate solutions to the exact to illustrate the accuracy of each method.

## Isogeometric Analysis Library

Today, I make my Isogeometric Analysis library available to you.

Written in Matlab, it reflects the time and effort that I spent in graduate school. It also shows unfortunate age, as B-splines and NURBS are fully supported, but not T-splines and other basis that are state-of-the-art. One day, I hope to return.

You can find my Isogeometric Analysis library here:

If you like regular, vanilla Finite Element Analysis:

## Hearing Perturbation Theory

In numerical linear algebra, we create ways for a computer to solve a linear system of equations $A\vec{x} = \vec{b}$. In doing so, we analyze how efficiently and accurately we can find the solution $\vec{x}$.

Perturbation theory concerns how much error we incur in the solution $\vec{x}$ when we perturb (spoil) the data $A$ and $\vec{b}$. A classic statement tells us that the amount of error depends on the condition number of the matrix $A$.

I will define and prove the statement, and help you understand it by “hearing” it.

## 15 Keywords

As an officer of Central Austin Toastmasters, I use Meetup to advertise my club and welcome new guests. Meetup allows me to set 15 keywords that best describe my club. It also lets you (a Meetup member) set keywords that describe your interests. Meetup uses your keywords to recommend groups that would interest you.

My goal is to attract more Meetup members so that more will visit my Toastmasters club. Hence, I should select keywords that are not only relevant to my club, but will also likely coincide with your interests.

Which 15 keywords should I use to attract more members?

I wrote a simple program to help me decide on these keywords.