LU decomposition

In numerical analysis and linear algebra, lower–upper (LU) decomposition or factorization factors a matrix as the product of a lower triangular matrix and an upper triangular matrix (see matrix decomposition). The product sometimes includes a permutation matrix as well. LU decomposition can be viewed as the matrix form of Gaussian elimination. Computers usually solve square systems of linear equations using LU decomposition, and it is also a key step when inverting a matrix or computing the determinant of a matrix. The LU decomposition was introduced by the Polish astronomer Tadeusz Banachiewicz in 1938.[1] To quote: "It appears that Gauss and Doolittle applied the method [of elimination] only to symmetric equations. More recent authors, for example, Aitken, Banachiewicz, Dwyer, and Crout … have emphasized the use of the method, or variations of it, in connection with non-symmetric problems … Banachiewicz … saw the point … that the basic problem is really one of matrix factorization, or “decomposition” as he called it."[2] It is also sometimes referred to as LR decomposition (factors into left and right triangular matrices).

Definitions

LDU decomposition of a Walsh matrix

Let A be a square matrix. An LU factorization refers to the factorization of A, with proper row and/or column orderings or permutations, into two factors – a lower triangular matrix L and an upper triangular matrix U:

In the lower triangular matrix all elements above the diagonal are zero, in the upper triangular matrix, all the elements below the diagonal are zero. For example, for a 3 × 3 matrix A, its LU decomposition looks like this:

Without a proper ordering or permutations in the matrix, the factorization may fail to materialize. For example, it is easy to verify (by expanding the matrix multiplication) that . If , then at least one of and has to be zero, which implies that either L or U is singular. This is impossible if A is nonsingular (invertible). This is a procedural problem. It can be removed by simply reordering the rows of A so that the first element of the permuted matrix is nonzero. The same problem in subsequent factorization steps can be removed the same way; see the basic procedure below.

LU factorization with partial pivoting

It turns out that a proper permutation in rows (or columns) is sufficient for LU factorization. LU factorization with partial pivoting (LUP) refers often to LU factorization with row permutations only:

where L and U are again lower and upper triangular matrices, and P is a permutation matrix, which, when left-multiplied to A, reorders the rows of A. It turns out that all square matrices can be factorized in this form,[3] and the factorization is numerically stable in practice.[4] This makes LUP decomposition a useful technique in practice.

LU factorization with full pivoting

An LU factorization with full pivoting involves both row and column permutations:

where L, U and P are defined as before, and Q is a permutation matrix that reorders the columns of A.[5]

Lower-diagonal-upper (LDU) decomposition

A Lower-diagonal-upper (LDU) decomposition is a decomposition of the form

where D is a diagonal matrix, and L and U are unitriangular matrices, meaning that all the entries on the diagonals of L and U are one.

Rectangular matrices

Above we required that A be a square matrix, but these decompositions can all be generalized to rectangular matrices as well.[6] In that case, L and D are square matrices both of which have the same number of rows as A, and U has exactly the same dimensions as A. Upper triangular should be interpreted as having only zero entries below the main diagonal, which starts at the upper left corner. Similarly, the more precise term for U is that it is the row echelon form of the matrix A.

Example

We factor the following 2-by-2 matrix:

One way to find the LU decomposition of this simple matrix would be to simply solve the linear equations by inspection. Expanding the matrix multiplication gives

This system of equations is underdetermined. In this case any two non-zero elements of L and U matrices are parameters of the solution and can be set arbitrarily to any non-zero value. Therefore, to find the unique LU decomposition, it is necessary to put some restriction on L and U matrices. For example, we can conveniently require the lower triangular matrix L to be a unit triangular matrix, so that all the entries of its main diagonal are set to one. Then the system of equations has the following solution:

Substituting these values into the LU decomposition above yields

Existence and uniqueness

Square matrices

Any square matrix admits LUP and PLU factorizations.[3] If is invertible, then it admits an LU (or LDU) factorization if and only if all its leading principal minors[7] are nonzero[8] (for example does not admit an LU or LDU factorization). If is a singular matrix of rank , then it admits an LU factorization if the first leading principal minors are nonzero, although the converse is not true.[9]

If a square, invertible matrix has an LDU (factorization with all diagonal entries of L and U equal to 1), then the factorization is unique.[8] In that case, the LU factorization is also unique if we require that the diagonal of (or ) consists of ones.

In general, any square matrix could have one of the following:

  1. a unique LU factorization (as mentioned above);
  2. infinitely many LU factorizations if two or more of any first (n−1) columns are linearly dependent or any of the first (n−1) columns are 0;
  3. no LU factorization if the first (n−1) columns are non-zero and linearly independent and at least one leading principal minor is zero.

In Case 3, one can approximate an LU factorization by changing a diagonal entry to to avoid a zero leading principal minor.[10]

Symmetric positive-definite matrices

If A is a symmetric (or Hermitian, if A is complex) positive-definite matrix, we can arrange matters so that U is the conjugate transpose of L. That is, we can write A as

This decomposition is called the Cholesky decomposition. If is positive definite, then the Cholesky decomposition exists and is unique. Furthermore, computing the Cholesky decomposition is more efficient and numerically more stable than computing some other LU decompositions.

General matrices

For a (not necessarily invertible) matrix over any field, the exact necessary and sufficient conditions under which it has an LU factorization are known. The conditions are expressed in terms of the ranks of certain submatrices. The Gaussian elimination algorithm for obtaining LU decomposition has also been extended to this most general case.[11]

Algorithms

Closed formula

When an LDU factorization exists and is unique, there is a closed (explicit) formula for the elements of L, D, and U in terms of ratios of determinants of certain submatrices of the original matrix A.[12] In particular, , and for , is the ratio of the -th principal submatrix to the -th principal submatrix. Computation of the determinants is computationally expensive, so this explicit formula is not used in practice.

Using Gaussian elimination

The following algorithm is essentially a modified form of Gaussian elimination. Computing an LU decomposition using this algorithm requires floating-point operations, ignoring lower-order terms. Partial pivoting adds only a quadratic term; this is not the case for full pivoting.[13]

Generalized explanation

Notation

Given an N × N matrix , define as the original, unmodified version of the matrix . The parenthetical superscript (e.g., ) of the matrix is the version of the matrix. The matrix is the matrix in which the elements below the main diagonal have already been eliminated to 0 through Gaussian elimination for the first columns.

Below is a matrix to observe to help us remember the notation (where each represents any real number in the matrix):

Procedure

During this process, we gradually modify the matrix using row operations until it becomes the matrix in which all the elements below the main diagonal are equal to zero. During this, we will simultaneously create two separate matrices and , such that .

We define the final permutation matrix as the identity matrix which has all the same rows swapped in the same order as the matrix while it transforms into the matrix . For our matrix , we may start by swapping rows to provide the desired conditions for the n-th column. For example, we might swap rows to perform partial pivoting, or we might do it to set the pivot element on the main diagonal to a non-zero number so that we can complete the Gaussian elimination.

For our matrix , we want to set every element below to zero (where is the element in the n-th column of the main diagonal). We will denote each element below as (where ). To set to zero, we set for each row . For this operation, . Once we have performed the row operations for the first columns, we have obtained an upper triangular matrix which is denoted by .

We can also create the lower triangular matrix denoted as , by directly inputting the previously calculated values of via the formula below.

Example

If we are given the matrixwe will choose to implement partial pivoting and thus swap the first and second row so that our matrix and the first iteration of our matrix respectively becomeOnce we have swapped the rows, we can eliminate the elements below the main diagonal on the first column by performing such that,Once these rows have been subtracted, we have derived from the matrix Because we are implementing partial pivoting, we swap the second and third rows of our derived matrix and the current version of our matrix respectively to obtainNow, we eliminate the elements below the main diagonal on the second column by performing such that . Because no non-zero elements exist below the main diagonal in our current iteration of after this row subtraction, this row subtraction derives our final matrix (denoted as ) and final matrix:After also switching the corresponding rows, we obtain our final matrix:Now these matrices have a relation such that .

Relations when no rows are swapped

If we did not swap rows at all during this process, we can perform the row operations simultaneously for each column by setting where is the N × N identity matrix with its n-th column replaced by the transposed vector In other words, the lower triangular matrix

Performing all the row operations for the first columns using the formula is equivalent to finding the decomposition Denote so that .

Now let's compute the sequence of . We know that has the following formula.

If there are two lower triangular matrices with 1s in the main diagonal, and neither have a non-zero item below the main diagonal in the same column as the other, then we can include all non-zero items at their same location in the product of the two matrices. For example:

Finally, multiply together and generate the fused matrix denoted as (as previously mentioned). Using the matrix , we obtain

It is clear that in order for this algorithm to work, one needs to have at each step (see the definition of ). If this assumption fails at some point, one needs to interchange n-th row with another row below it before continuing. This is why an LU decomposition in general looks like .

LU Crout decomposition

Note that the decomposition obtained through this procedure is a Doolittle decomposition: the main diagonal of L is composed solely of 1s. If one would proceed by removing elements above the main diagonal by adding multiples of the columns (instead of removing elements below the diagonal by adding multiples of the rows), we would obtain a Crout decomposition, where the main diagonal of U is of 1s.

Another (equivalent) way of producing a Crout decomposition of a given matrix A is to obtain a Doolittle decomposition of the transpose of A. Indeed, if is the LU-decomposition obtained through the algorithm presented in this section, then by taking and , we have that is a Crout decomposition.

Through recursion

Cormen et al.[14] describe a recursive algorithm for LUP decomposition.

Given a matrix A, let P1 be a permutation matrix such that

,

where , if there is a nonzero entry in the first column of A; or take P1 as the identity matrix otherwise. Now let , if ; or otherwise. We have

Now we can recursively find an LUP decomposition . Let . Therefore

which is an LUP decomposition of A.

Randomized algorithm

It is possible to find a low rank approximation to an LU decomposition using a randomized algorithm. Given an input matrix and a desired low rank , the randomized LU returns permutation matrices and lower/upper trapezoidal matrices of size and respectively, such that with high probability , where is a constant that depends on the parameters of the algorithm and is the -th singular value of the input matrix .[15]

Theoretical complexity

If two matrices of order n can be multiplied in time M(n), where M(n) ≥ na for some a > 2, then an LU decomposition can be computed in time O(M(n)).[16] This means, for example, that an O(n2.376) algorithm exists based on the Coppersmith–Winograd algorithm.

Sparse-matrix decomposition

Special algorithms have been developed for factorizing large sparse matrices. These algorithms attempt to find sparse factors L and U. Ideally, the cost of computation is determined by the number of nonzero entries, rather than by the size of the matrix.

These algorithms use the freedom to exchange rows and columns to minimize fill-in (entries that change from an initial zero to a non-zero value during the execution of an algorithm).

General treatment of orderings that minimize fill-in can be addressed using graph theory.

Applications

Solving linear equations

Given a system of linear equations in matrix form

we want to solve the equation for x, given A and b. Suppose we have already obtained the LUP decomposition of A such that , so .

In this case the solution is done in two logical steps:

  1. First, we solve the equation for y.
  2. Second, we solve the equation for x.

In both cases we are dealing with triangular matrices (L and U), which can be solved directly by forward and backward substitution without using the Gaussian elimination process (however we do need this process or equivalent to compute the LU decomposition itself).

The above procedure can be repeatedly applied to solve the equation multiple times for different b. In this case it is faster (and more convenient) to do an LU decomposition of the matrix A once and then solve the triangular matrices for the different b, rather than using Gaussian elimination each time. The matrices L and U could be thought to have "encoded" the Gaussian elimination process.

The cost of solving a system of linear equations is approximately floating-point operations if the matrix has size . This makes it twice as fast as algorithms based on QR decomposition, which costs about floating-point operations when Householder reflections are used. For this reason, LU decomposition is usually preferred.[17]

Inverting a matrix

When solving systems of equations, b is usually treated as a vector with a length equal to the height of matrix A. In matrix inversion however, instead of vector b, we have matrix B, where B is an n-by-p matrix, so that we are trying to find a matrix X (also a n-by-p matrix):

We can use the same algorithm presented earlier to solve for each column of matrix X. Now suppose that B is the identity matrix of size n. It would follow that the result X must be the inverse of A.[18]

Computing the determinant

Given the LUP decomposition of a square matrix A, the determinant of A can be computed straightforwardly as

The second equation follows from the fact that the determinant of a triangular matrix is simply the product of its diagonal entries, and that the determinant of a permutation matrix is equal to (−1)S where S is the number of row exchanges in the decomposition.

In the case of LU decomposition with full pivoting, also equals the right-hand side of the above equation, if we let S be the total number of row and column exchanges.

The same method readily applies to LU decomposition by setting P equal to the identity matrix.

Code examples

C code example

/* INPUT: A - array of pointers to rows of a square matrix having dimension N
 *        Tol - small tolerance number to detect failure when the matrix is near degenerate
 * OUTPUT: Matrix A is changed, it contains a copy of both matrices L-E and U as A=(L-E)+U such that P*A=L*U.
 *        The permutation matrix is not stored as a matrix, but in an integer vector P of size N+1 
 *        containing column indexes where the permutation matrix has "1". The last element P[N]=S+N, 
 *        where S is the number of row exchanges needed for determinant computation, det(P)=(-1)^S    
 */
int LUPDecompose(double **A, int N, double Tol, int *P) {

    int i, j, k, imax; 
    double maxA, *ptr, absA;

    for (i = 0; i <= N; i++)
        P[i] = i; //Unit permutation matrix, P[N] initialized with N

    for (i = 0; i < N; i++) {
        maxA = 0.0;
        imax = i;

        for (k = i; k < N; k++)
            if ((absA = fabs(A[k][i])) > maxA) { 
                maxA = absA;
                imax = k;
            }

        if (maxA < Tol) return 0; //failure, matrix is degenerate

        if (imax != i) {
            //pivoting P
            j = P[i];
            P[i] = P[imax];
            P[imax] = j;

            //pivoting rows of A
            ptr = A[i];
            A[i] = A[imax];
            A[imax] = ptr;

            //counting pivots starting from N (for determinant)
            P[N]++;
        }

        for (j = i + 1; j < N; j++) {
            A[j][i] /= A[i][i];

            for (k = i + 1; k < N; k++)
                A[j][k] -= A[j][i] * A[i][k];
        }
    }

    return 1;  //decomposition done 
}

/* INPUT: A,P filled in LUPDecompose; b - rhs vector; N - dimension
 * OUTPUT: x - solution vector of A*x=b
 */
void LUPSolve(double **A, int *P, double *b, int N, double *x) {

    for (int i = 0; i < N; i++) {
        x[i] = b[P[i]];

        for (int k = 0; k < i; k++)
            x[i] -= A[i][k] * x[k];
    }

    for (int i = N - 1; i >= 0; i--) {
        for (int k = i + 1; k < N; k++)
            x[i] -= A[i][k] * x[k];

        x[i] /= A[i][i];
    }
}

/* INPUT: A,P filled in LUPDecompose; N - dimension
 * OUTPUT: IA is the inverse of the initial matrix
 */
void LUPInvert(double **A, int *P, int N, double **IA) {
  
    for (int j = 0; j < N; j++) {
        for (int i = 0; i < N; i++) {
            IA[i][j] = P[i] == j ? 1.0 : 0.0;

            for (int k = 0; k < i; k++)
                IA[i][j] -= A[i][k] * IA[k][j];
        }

        for (int i = N - 1; i >= 0; i--) {
            for (int k = i + 1; k < N; k++)
                IA[i][j] -= A[i][k] * IA[k][j];

            IA[i][j] /= A[i][i];
        }
    }
}

/* INPUT: A,P filled in LUPDecompose; N - dimension. 
 * OUTPUT: Function returns the determinant of the initial matrix
 */
double LUPDeterminant(double **A, int *P, int N) {

    double det = A[0][0];

    for (int i = 1; i < N; i++)
        det *= A[i][i];

    return (P[N] - N) % 2 == 0 ? det : -det;
}

C# code example

public class SystemOfLinearEquations
{
    public double[] SolveUsingLU(double[,] matrix, double[] rightPart, int n)
    {
        // decomposition of matrix
        double[,] lu = new double[n, n];
        double sum = 0;
        for (int i = 0; i < n; i++)
        {
            for (int j = i; j < n; j++)
            {
                sum = 0;
                for (int k = 0; k < i; k++)
                    sum += lu[i, k] * lu[k, j];
                lu[i, j] = matrix[i, j] - sum;
            }
            for (int j = i + 1; j < n; j++)
            {
                sum = 0;
                for (int k = 0; k < i; k++)
                    sum += lu[j, k] * lu[k, i];
                lu[j, i] = (1 / lu[i, i]) * (matrix[j, i] - sum);
            }
        }

        // lu = L+U-I
        // find solution of Ly = b
        double[] y = new double[n];
        for (int i = 0; i < n; i++)
        {
            sum = 0;
            for (int k = 0; k < i; k++)
                sum += lu[i, k] * y[k];
            y[i] = rightPart[i] - sum;
        }
        // find solution of Ux = y
        double[] x = new double[n];
        for (int i = n - 1; i >= 0; i--)
        {
            sum = 0;
            for (int k = i + 1; k < n; k++)
                sum += lu[i, k] * x[k];
            x[i] = (1 / lu[i, i]) * (y[i] - sum);
        }
        return x;
    }
}

MATLAB code example

function LU = LUDecompDoolittle(A)
    n = length(A);
    LU = A;
    % decomposition of matrix, Doolittle's Method
    for i = 1:1:n
        for j = 1:(i - 1)
            LU(i,j) = (LU(i,j) - LU(i,1:(j - 1))*LU(1:(j - 1),j)) / LU(j,j);
        end
        j = i:n;
        LU(i,j) = LU(i,j) - LU(i,1:(i - 1))*LU(1:(i - 1),j);
    end
    %LU = L+U-I
end

function x = SolveLinearSystem(LU, B)
    n = length(LU);
    y = zeros(size(B));
    % find solution of Ly = B
    for i = 1:n
        y(i,:) = B(i,:) - LU(i,1:i)*y(1:i,:);
    end
    % find solution of Ux = y
    x = zeros(size(B));
    for i = n:(-1):1
        x(i,:) = (y(i,:) - LU(i,(i + 1):n)*x((i + 1):n,:))/LU(i, i);
    end    
end

A = [ 4 3 3; 6 3 3; 3 4 3 ]
LU = LUDecompDoolittle(A)
B = [ 1 2 3; 4 5 6; 7 8 9; 10 11 12 ]'
x = SolveLinearSystem(LU, B)
A * x

See also

Notes

  1. ^ Schwarzenberg-Czerny, A. (1995). "On matrix factorization and efficient least squares solution". Astronomy and Astrophysics Supplement Series. 110: 405. Bibcode:1995A&AS..110..405S.
  2. ^ Dwyer (1951).
  3. ^ a b Okunev & Johnson (1997), Corollary 3.
  4. ^ Trefethen & Bau (1997), p. 166.
  5. ^ Trefethen & Bau (1997), p. 161.
  6. ^ Lay, Lay & McDonald (2021), p. 133, 2.5: Matrix Factorizations.
  7. ^ Rigotti (2001), Leading Principal Minor.
  8. ^ a b Horn & Johnson (1985), Corollary 3.5.5
  9. ^ Horn & Johnson (1985), Theorem 3.5.2.
  10. ^ Nhiayi, Ly; Phan-Yamada, Tuyetdong (2021). "Examining Possible LU Decomposition". North American GeoGebra Journal. 9 (1).
  11. ^ Okunev & Johnson (1997).
  12. ^ Householder (1975).
  13. ^ Golub & Van Loan (1996), pp. 112, 119.
  14. ^ Cormen et al. (2009), p. 819, 28.1: Solving systems of linear equations.
  15. ^ Shabat, Gil; Shmueli, Yaniv; Aizenbud, Yariv; Averbuch, Amir (2016). "Randomized LU Decomposition". Applied and Computational Harmonic Analysis. 44 (2): 246–272. arXiv:1310.7202. doi:10.1016/j.acha.2016.04.006. S2CID 1900701.
  16. ^ Bunch & Hopcroft (1974).
  17. ^ Trefethen & Bau (1997), p. 152.
  18. ^ Golub & Van Loan (1996), p. 121.

References

References

Computer code

Online resources

Read other articles:

Indian academic (born 1956) This article has multiple issues. Please help improve it or discuss these issues on the talk page. (Learn how and when to remove these template messages) This biography of a living person needs additional citations for verification. Please help by adding reliable sources. Contentious material about living persons that is unsourced or poorly sourced must be removed immediately from the article and its talk page, especially if potentially libelous.Find sources: Gee…

Love Is All There IsSampul DVDSutradaraJoseph BolognaRenée TaylorProduserElliott KastnerDitulis olehGabriel BolognaJoseph BolognaWilliam Shakespeare (drama)Erik ShapiroRenée TaylorPemeran Lainie Kazan Joseph Bologna Barbara Carrera Renée Taylor William Hickey Dick Van Patten Abe Vigoda Connie Stevens Paul Sorvino Angelina Jolie Nathaniel Marston Penata musikJeff BealSinematograferAlan JonesPenyuntingNicholas EliopoulosDennis M. O'ConnorDistributorThe Samuel Goldwyn CompanyTanggal rilis …

Battaglia della seconda harraparte della Seconda Fitna islamicaLa battaglia fu combattuta appena a nord di Medina su un terreno roccioso e basaltico simile a quello dell'immagine, scattata nei pressi dell'insediamentoData26 agosto 683 LuogoMedina CausaMancato riconoscimento medinese del califfato di Yazīd ibn Muʿāwiya EsitoVittoria tattica omayyade SchieramentiOmayyadiMedinesi (Ansar, Muhajirun, inclusi i Quraysh) e Meccani ComandantiMuslim ibn ʿUqbaMarwan ibn al-HakamʿAbd Allāh b. Ḥan…

This article includes a list of general references, but it lacks sufficient corresponding inline citations. Please help to improve this article by introducing more precise citations. (October 2019) (Learn how and when to remove this message) Princess of Chiang Mai Dara RasamiPrincess of Chiang MaiPrincess consort of SiamTenure3 February 1887 – 23 October 1910BornChao Dara Rasmi(1873-08-26)26 August 1873Kingdom of Chiang MaiDied9 December 1933(1933-12-09) (aged 60)Chiang Mai, SiamSpouseChu…

Indian politician (1920–2021) Raghunandan Lal BhatiaBhatia in 200523rd Governor of BiharIn office10 July 2008 – 28 June 2009Chief MinisterNitish KumarPreceded byR. S. GavaiSucceeded byDevanand Konwar16th Governor of KeralaIn office23 June 2004 – 10 July 2008Chief MinisterA. K. AntonyOommen ChandyV. S. AchuthanandanPreceded byT. N. Chaturvedi(Additional Charge)Succeeded byR. S. Gavai Personal detailsBorn(1920-07-03)3 July 1920Amritsar, Punjab, British IndiaDied14 M…

Argentine footballer Adolfo Zumelzú Zumelzú in 1929Personal informationFull name Adolfo Bernabé ZumelzúDate of birth (1902-01-05)5 January 1902Place of birth Buenos Aires, ArgentinaDate of death 29 March 1973(1973-03-29) (aged 71)Position(s) Half-backYouth career San IsidroSenior career*Years Team Apps (Gls) San Isidro 1925 Racing Club 1930 Estudiantil Porteño 1931 Tigre 7 (0) Sportivo Palermo International career1927–1930 Argentina 13 (4) Medal record Men's Football Representing …

Family of birds Not to be confused with Hylidae, a family of amphibians. Hyliidae Green hylia (Hylia prasina) Scientific classification Domain: Eukaryota Kingdom: Animalia Phylum: Chordata Class: Aves Order: Passeriformes Superfamily: Sylvioidea Family: HyliidaeBannerman, 1923 Genera Hylia Pholidornis Hyliidae is a family of passerine birds which contains just two species, the green hylia (Hylia prasina) and the tit hylia (Pholidornis rushiae). Physiological similarities and molecular phylogenet…

Due Carrarecomune Due Carrare – VedutaPiazza del Municipio e Via Roma a Carrara San Giorgio LocalizzazioneStato Italia Regione Veneto Provincia Padova AmministrazioneSindacoDavide Moro (lista civica) dal 21-9-2020 Data di istituzione21-3-1995 TerritorioCoordinate45°17′N 11°50′E / 45.283333°N 11.833333°E45.283333; 11.833333 (Due Carrare)Coordinate: 45°17′N 11°50′E / 45.283333°N 11.833333°E45.283333; 11.833333 (Due Ca…

Hospital in Virginia, United StatesChippenham HospitalHospital Corporation of AmericaGeographyLocationRichmond, Virginia, United StatesCoordinates37°30′54″N 77°31′30″W / 37.51496°N 77.52509°W / 37.51496; -77.52509OrganizationFundingFor-profit hospitalServicesEmergency departmentLevel I trauma centerBeds466HelipadFAA LID: 6VA3HistoryOpened1972LinksWebsitewww.hcavirginia.com/locations/chippenham-hospitalListsHospitals in Virginia Chippenham Hospital is a for-pro…

 烏克蘭總理Прем'єр-міністр України烏克蘭國徽現任杰尼斯·什米加尔自2020年3月4日任命者烏克蘭總統任期總統任命首任維托爾德·福金设立1991年11月后继职位無网站www.kmu.gov.ua/control/en/(英文) 乌克兰 乌克兰政府与政治系列条目 宪法 政府 总统 弗拉基米尔·泽连斯基 總統辦公室 国家安全与国防事务委员会 总统代表(英语:Representatives of the President of Ukraine) 总理…

Sri Lankan King Kashyapa I (Kasyapa I)King of AnuradhapuraReign473–495PredecessorDhatusenaSuccessorMoggallanaBornc.448Died495 (aged 46-47)IssuePrincess Bodhi Princess UthpalawannaHouseHouse of MauryaDynastyMoriyaFatherDhatusenaMotherA non-royal wife of King Dhatusena Kashyapa I, also known as Kasyapa I or Kassapa I, was a king of Sri Lanka, who ruled the country from 473 to 495 CE. He was the second king of the royal Moriya dynasty of Sri Lanka. Kashyapa is credited with the construction of th…

عنتالمنطقة الجنوبية في إسرائيلمدن عراد أشدود عسقلان بئر السبع ديمونا إيلات كريات جات كريات ملاخي نتيفوت أوفاكيم رهط سديروت مجالس محلية عرعرة النقب حورة كسيفة اللقية لهافيم ميتار متسبي ريمون عومر شقيب السلام تل السبع يروحام المجالس الإقليمية القيصوم بير توفيا بني شمعون وا…

Botanical component of pest control for plants This article needs additional citations for verification. Please help improve this article by adding citations to reliable sources. Unsourced material may be challenged and removed.Find sources: Weed control – news · newspapers · books · scholar · JSTOR (February 2014) (Learn how and when to remove this message) Weeding redirects here. For the library process, see Weeding (library). A field of beets being wee…

Kiziba refugee camp in the west of Rwanda, 2014 Refugee camp in Beirut, c. 1920–25 Temporary settlement for refugees Refugee camp (located in present-day eastern Congo-Kinshasa) for Rwandans following the Rwandan genocide of 1994 A camp in Guinea for refugees from Sierra Leone Nahr el-Bared, Palestinian refugee camp in North Lebanon in 2005Mitzpe Ramon, development camp for Jewish refugees, southern Israel, 1957 A refugee camp is a temporary settlement built to receive refugees and peopl…

Gouvernement Henri Brisson (1) Troisième République « Le nouveau Ministère » à la une du Journal illustré, 19 avril 1885. Données clés Président de la République Jules Grévy Président du Conseil Henri Brisson Formation 6 avril 1885 Fin 29 décembre 1885 Durée 8 mois et 23 jours Composition initiale Coalition Gauche radicale - Gauche républicaine - Union démocratique - Union républicaine - Centre gauche Représentation IIIe législature 418  /  557 G…

هذه المقالة يتيمة إذ تصل إليها مقالات أخرى قليلة جدًا. فضلًا، ساعد بإضافة وصلة إليها في مقالات متعلقة بها. (أبريل 2019) جينجر كلارك معلومات شخصية الميلاد 7 مارس 1879   ووستير  الوفاة 10 مايو 1943 (64 سنة)   ليك تشارلز  مواطنة الولايات المتحدة  الحياة العملية المهنة لاعب كرة …

هذه المقالة بحاجة لصندوق معلومات. فضلًا ساعد في تحسين هذه المقالة بإضافة صندوق معلومات مخصص إليها. يفتقر محتوى هذه المقالة إلى الاستشهاد بمصادر. فضلاً، ساهم في تطوير هذه المقالة من خلال إضافة مصادر موثوق بها. أي معلومات غير موثقة يمكن التشكيك بها وإزالتها. (مارس 2016) هذه المقا…

Place in Lubusz Voivodeship, PolandŻagańOld town FlagCoat of armsŻagańCoordinates: 51°37′N 15°19′E / 51.617°N 15.317°E / 51.617; 15.317Country PolandVoivodeship LubuszCountyŻagańGminaŻagań (urban gmina)Established12th centuryFirst mentioned1202Town rightsbetween 1248 and 1260Government • MayorSławomir KowalArea • Total39.92 km2 (15.41 sq mi)Population (2019-06-30[1]) • Total25,731&#…

Cet article est une ébauche concernant les armes. Vous pouvez partager vos connaissances en l’améliorant (comment ?) selon les recommandations des projets correspondants. International MaxxPro Caractéristiques de service Type véhicule de transport de troupes, génie militaire Production Constructeur Navistar International Production 2007, en cours Unités produites 9 000 Caractéristiques générales Équipage 2 plus 8 Longueur 6,5 à 7,2 m Largeur 2,5 m Hauteur 3 m Garde au sol 35 cm…

Mathematical concept Algebraic structure → Group theoryGroup theory Basic notions Subgroup Normal subgroup Quotient group (Semi-)direct product Group homomorphisms kernel image direct sum wreath product simple finite infinite continuous multiplicative additive cyclic abelian dihedral nilpotent solvable action Glossary of group theory List of group theory topics Finite groups Cyclic group Zn Symmetric group Sn Alternating group An Dihedral group Dn Quaternion group Q Cauchy's theorem Lagrange's…