Band matrix

From Wikipedia, the free encyclopedia

In mathematics, particularly matrix theory, a band matrix is a sparse matrix, whose non-zero entries are confined to a diagonal band, comprising the main diagonal and zero or more diagonals on either side.

Formally, an n×n matrix A=(ai,j ) is a band matrix if all matrix elements are zero outside a diagonally bordered band whose range it determined by constants k1 and k2:

a_{i,j}=0 \quad\mbox{if}\quad j<i-k_1 \quad\mbox{ or }\quad j>i+k_2; \quad k_1, k_2 \ge 0.\,

The quantities k1 and k2 are the left and right half-bandwidth, respectively. The bandwidth of the matrix is k1 + k2 + 1 (in other words, the smallest number of adjacent diagonals to which the non-zero elements are confined).

A band matrix with k1 = k2 = 0 is a diagonal matrix; a band matrix with k1 = k2 = 1 is a tridiagonal matrix; when k1 = k2 = 2 one has a pentadiagonal matrix and so on. If one puts k1 = 0, k2 = n−1, one obtains the definition of a lower triangular matrix; similarly, for k1 = n−1, k2 = 0 one obtains an upper triangular matrix.

Contents

[edit] Applications

In numerical analysis, matrices from finite element or finite difference problems are often banded. Such matrices can be viewed as descriptions of the coupling between the problem variables; the bandedness corresponds to the fact that variables are not coupled over arbitrarily large distances. Such matrices can be further divided - for instance, banded matrices exist where every element in the band is nonzero. These often arise when discretising one-dimensional problems.[citation needed]

Problems in higher dimensions also lead to banded matrices, in which case the band itself also tends to be sparse. For instance, a partial differential equation on a square domain (using central differences) will yield a matrix with a half-bandwidth equal to the square root of the matrix dimension, but inside the band only 5 diagonals are nonzero. Unfortunately, applying Gaussian elimination (or equivalently an LU decomposition) to such a matrix results in the band being filled in by many non-zero elements. As sparse matrices lend themselves to more efficient computation than dense matrices, there has been much research focused on finding ways to minimise the bandwidth (or directly minimise the fill in) by applying permutations to the matrix, or other such equivalence or similarity transformations.[citation needed]

From a computational point of view, working with band matrices is always preferential to working with similarly dimensioned dense square matrices. A band matrix can be likened in complexity to a rectangular matrix whose row dimension is equal to the bandwidth of the band matrix. Thus the work involved in performing operations such as multiplication falls significantly, often leading to huge savings in terms of calculation time and complexity.

[edit] Examples and special cases

The following are special cases of band matrices:

The inverses of Lehmer matrices are constant tridiagonal matrices, and are thus band matrices.

[edit] Notes

The representation of band matrices in the LAPACK Fortran package is not consistent with that found in the EISPACK package.

[edit] External links