To install click the Add extension button. That's it.

The source code for the WIKI 2 extension is being checked by specialists of the Mozilla Foundation, Google, and Apple. You could also do it yourself at any point in time.

4,5
Kelly Slayton
Congratulations on this excellent venture… what a great idea!
Alexander Grigorievskiy
I use WIKI 2 every day and almost forgot how the original Wikipedia looks like.
Live Statistics
English Articles
Improved in 24 Hours
Added in 24 Hours
What we do. Every page goes through several hundred of perfecting techniques; in live mode. Quite the same Wikipedia. Just better.
.
Leo
Newton
Brights
Milds

From Wikipedia, the free encyclopedia

In mathematics, particularly matrix theory, a band matrix or banded matrix is a sparse matrix whose non-zero entries are confined to a diagonal band, comprising the main diagonal and zero or more diagonals on either side.

YouTube Encyclopedic

  • 1/3
    Views:
    10 019
    1 741
    51 197
  • Zero, identity, diagonal, triangular, banded matrices | Lecture 3 | Matrix Algebra for Engineers
  • PLUS TWO MATHEMATICS MALAYALAM " MATRICES "PART 2
  • Sparse Matrices - Intro to Parallel Programming

Transcription

Band matrix

Bandwidth

Formally, consider an n×n matrix A=(ai,j ). If all matrix elements are zero outside a diagonally bordered band whose range is determined by constants k1 and k2:

then the quantities k1 and k2 are called the lower bandwidth and upper bandwidth, respectively.[1] The bandwidth of the matrix is the maximum of k1 and k2; in other words, it is the number k such that if .[2]

Examples

Applications

In numerical analysis, matrices from finite element or finite difference problems are often banded. Such matrices can be viewed as descriptions of the coupling between the problem variables; the banded property corresponds to the fact that variables are not coupled over arbitrarily large distances. Such matrices can be further divided – for instance, banded matrices exist where every element in the band is nonzero. These often arise when discretising one-dimensional problems.[citation needed]

Problems in higher dimensions also lead to banded matrices, in which case the band itself also tends to be sparse. For instance, a partial differential equation on a square domain (using central differences) will yield a matrix with a bandwidth equal to the square root of the matrix dimension, but inside the band only 5 diagonals are nonzero. Unfortunately, applying Gaussian elimination (or equivalently an LU decomposition) to such a matrix results in the band being filled in by many non-zero elements.

Band storage

Band matrices are usually stored by storing the diagonals in the band; the rest is implicitly zero.

For example, a tridiagonal matrix has bandwidth 1. The 6-by-6 matrix

is stored as the 6-by-3 matrix

A further saving is possible when the matrix is symmetric. For example, consider a symmetric 6-by-6 matrix with an upper bandwidth of 2:

This matrix is stored as the 6-by-3 matrix:

Band form of sparse matrices

From a computational point of view, working with band matrices is always preferential to working with similarly dimensioned square matrices. A band matrix can be likened in complexity to a rectangular matrix whose row dimension is equal to the bandwidth of the band matrix. Thus the work involved in performing operations such as multiplication falls significantly, often leading to huge savings in terms of calculation time and complexity.

As sparse matrices lend themselves to more efficient computation than dense matrices, as well as in more efficient utilization of computer storage, there has been much research focused on finding ways to minimise the bandwidth (or directly minimise the fill-in) by applying permutations to the matrix, or other such equivalence or similarity transformations.[3]

The Cuthill–McKee algorithm can be used to reduce the bandwidth of a sparse symmetric matrix. There are, however, matrices for which the reverse Cuthill–McKee algorithm performs better. There are many other methods in use.

The problem of finding a representation of a matrix with minimal bandwidth by means of permutations of rows and columns is NP-hard.[4]

See also

Notes

References

  • Atkinson, Kendall E. (1989), An Introduction to Numerical Analysis, John Wiley & Sons, ISBN 0-471-62489-6.
  • Davis, Timothy A. (2006), Direct Methods for Sparse Linear Systems, Society for Industrial and Applied Mathematics, ISBN 978-0-898716-13-9.
  • Feige, Uriel (2000), "Coping with the NP-Hardness of the Graph Bandwidth Problem", Algorithm Theory - SWAT 2000, Lecture Notes in Computer Science, vol. 1851, pp. 129–145, doi:10.1007/3-540-44985-X_2.
  • Golub, Gene H.; Van Loan, Charles F. (1996), Matrix Computations (3rd ed.), Baltimore: Johns Hopkins, ISBN 978-0-8018-5414-9.
  • Press, WH; Teukolsky, SA; Vetterling, WT; Flannery, BP (2007), "Section 2.4", Numerical Recipes: The Art of Scientific Computing (3rd ed.), New York: Cambridge University Press, ISBN 978-0-521-88068-8.

External links

This page was last edited on 27 January 2024, at 08:32
Basis of this page is in Wikipedia. Text is available under the CC BY-SA 3.0 Unported License. Non-text media are available under their specified licenses. Wikipedia® is a registered trademark of the Wikimedia Foundation, Inc. WIKI 2 is an independent company and has no affiliation with Wikimedia Foundation.