EFFICIENT COMPUTATION FOR EIGENVALUE DECOMPOSITION AND SINGULAR VALUE DECOMPOSITION OF MATRICES
For eigenvalue decomposition, a first set of at least one variable is derived based on a first matrix being decomposed and using Coordinate Rotational Digital Computer (CORDIC) computation. A second set of at least one variable is derived based on the first matrix and using a look-up table. A second matrix of eigenvectors of the first matrix is then derived based on the first and second variable sets. To derive the first variable set, CORDIC computation is performed on an element of the first matrix to determine the magnitude and phase of this element, and CORDIC computation is performed on the phase to determine the sine and cosine of this element. To derive the second variable set, intermediate quantities are derived based on the first matrix and used to access the look-up table.
Latest QUALCOMM INCORPORATED Patents:
- Techniques for listen-before-talk failure reporting for multiple transmission time intervals
- Techniques for channel repetition counting
- Random access PUSCH enhancements
- Random access response enhancement for user equipments with reduced capabilities
- Framework for indication of an overlap resolution process
This application is a continuation of and claims the benefit of priority from U.S. patent application Ser. No. 11/096,839 (now allowed), entitled “Efficient Computation for Eigenvalue Decomposition and Singular Value Decomposition of Matrices” and filed Mar. 31, 2005, which claims the benefit of priority from U.S. Provisional Patent Application Ser. No. 60/628,324, entitled “Eigenvalue Decomposition and Singular Value Decomposition of Matrices Using Jacobi Rotation” and filed Nov. 15, 2004, both of which are assigned to the assignee of this application and are fully incorporated herein by reference for all purposes.
BACKGROUND1. Field
The present invention relates generally to communication, and more specifically to techniques for decomposing matrices.
2. Background
A multiple-input multiple-output (MIMO) communication system employs multiple (T) transmit antennas at a transmitting entity and multiple (R) receive antennas at a receiving entity for data transmission. A MIMO channel formed by the T transmit antennas and the R receive antennas may be decomposed into S spatial channels, where S≦min {T, R}. The S spatial channels may be used to transmit data in a manner to achieve higher overall throughput and/or greater reliability.
The MIMO channel response may be characterized by an R×T channel response matrix H, which contains complex channel gains for all of the different pairs of transmit and receive antennas. The channel response matrix H may be diagonalized to obtain S eigenmodes, which may be viewed as orthogonal spatial channels of the MIMO channel. Improved performance may be achieved by transmitting data on the eigenmodes of the MIMO channel.
The channel response matrix H may be diagonalized by performing either singular value decomposition of H or eigenvalue decomposition of a correlation matrix of H. The singular value decomposition provides left and right singular vectors, and the eigenvalue decomposition provides eigenvectors. The transmitting entity uses the right singular vectors or the eigenvectors to transmit data on the S eigenmodes. The receiving entity uses the left singular vectors or the eigenvectors to receive data on the S eigenmodes.
Eigenvalue decomposition and singular value decomposition are computationally intensive. There is therefore a need in the art for techniques to efficiently decompose matrices.
SUMMARYTechniques for efficiently decomposing matrices are described herein. According to an embodiment of the invention, a method is provided in which a first set of at least one variable (e.g., cosine c1, sine s1, and magnitude r) is derived based on a first matrix to be decomposed and using Coordinate Rotational Digital Computer (CORDIC) computation. A second set of at least one variable (e.g., variables c and s) is derived based on the first matrix and using a look-up table. A second matrix of eigenvectors is then derived based on the first and second sets of at least one variable.
According to another embodiment, an apparatus is described which includes a CORDIC processor, a look-up processor, and a post-processor. The CORDIC processor derives a first set of at least one variable based on a first matrix to be decomposed. The look-up processor derives a second set of at least one variable based on the first matrix and using a look-up table. The post-processor derives a second matrix of eigenvectors based on the first and second sets of at least one variable.
According to yet another embodiment, an apparatus is described which includes means for deriving a first set of at least one variable based on a first matrix to be decomposed and using CORDIC computation, means for deriving a second set of at least one variable based on the first matrix and using a look-up table, and means for deriving a second matrix of eigenvectors based on the first and second sets of at least one variable.
According to yet another embodiment, a method is provided in which CORDIC computation is performed on an element of a first matrix to determine the magnitude and phase of the element. CORDIC computation is also performed on the phase of the element to determine the sine and cosine of the element. A second matrix of eigenvectors is then derived based on the magnitude, sine, and cosine of the element.
According to yet another embodiment, an apparatus is described which includes means for performing CORDIC computation on an element of a first matrix to determine the magnitude and phase of the element, means for performing CORDIC computation on the phase of the element to determine the sine and cosine of the element, and means for deriving a second matrix of eigenvectors based on the magnitude, sine, and cosine of the element.
According to yet another embodiment, a method is provided in which intermediate quantities are derived based on a first matrix to be decomposed. At least one variable is then derived based on the intermediate quantities and using a look-up table. A second matrix of eigenvectors is derived based on the at least one variable.
According to yet another embodiment, an apparatus is described which includes a pre-processor, a look-up table, and a post-processor. The pre-processor derives intermediate quantities based on a first matrix to be decomposed. The look-up table provides at least one variable based on the intermediate quantities. The post-processor derives a second matrix of eigenvectors based on the at least one variable.
According to yet another embodiment, an apparatus is described which includes means for deriving intermediate quantities based on a first matrix to be decomposed, means for deriving at least one variable based on the intermediate quantities and using a look-up table, and means for deriving a second matrix of eigenvectors based on the at least one variable.
According to yet another embodiment, a method is provided in which multiple iterations of Jacobi rotation are performed on a first matrix of complex values with multiple Jacobi rotation matrices. Each Jacobi rotation matrix is derived by performing eigenvalue decomposition using CORDIC computation, a look-up table, or both. A unitary matrix with orthogonal vectors is then derived based on the multiple Jacobi rotation matrices.
According to yet another embodiment, an apparatus is described which includes means for performing multiple iterations of Jacobi rotation on a first matrix of complex values with multiple Jacobi rotation matrices and means for deriving a unitary matrix with orthogonal vectors based on the multiple Jacobi rotation matrices. Each Jacobi rotation matrix is derived by performing eigenvalue decomposition using CORDIC computation, a look-up table, or both.
According to yet another embodiment, a method is provided in which multiple matrices of complex values are obtained for multiple transmission spans. Multiple iterations of Jacobi rotation are performed on a first matrix of complex values for a first transmission span to obtain a first unitary matrix with orthogonal vectors. Each iteration of the Jacobi rotation utilizes eigenvalue decomposition using CORDIC computation, a look-up table, or both. Multiple iterations of the Jacobi rotation are performed on a second matrix of complex values for a second transmission span to obtain a second unitary matrix with orthogonal vectors. The first unitary matrix is used as an initial solution for the second unitary matrix.
According to yet another embodiment, an apparatus is described which includes means for obtaining multiple matrices of complex values for multiple transmission spans, means for performing multiple iterations of Jacobi rotation on a first matrix of complex values for a first transmission span to obtain a first unitary matrix with orthogonal vectors, and means for performing multiple iterations of the Jacobi rotation on a second matrix of complex values for a second transmission span to obtain a second unitary matrix with orthogonal vectors. Each iteration of the Jacobi rotation utilizes eigenvalue decomposition using CORDIC computation, a look-up table, or both. The first unitary matrix is used as an initial solution for the second unitary matrix.
Various aspects and embodiments of the invention are described in further detail below.
The word “exemplary” is used herein to mean “serving as an example, instance, or illustration.” Any embodiment described herein as “exemplary” is not necessarily to be construed as preferred or advantageous over other embodiments.
The decomposition techniques described herein may be used for single-carrier and multi-carrier communication systems. Multiple carriers may be obtained with orthogonal frequency division multiplexing (OFDM), some other multi-carrier modulation techniques, or some other construct. OFDM effectively partitions the overall system bandwidth into multiple (K) orthogonal frequency subbands, which are also called tones, subcarriers, bins, and frequency channels. With OFDM, each subband is associated with a respective subcarrier that may be modulated with data. For clarity, much of the following description is for a single-carrier MIMO system.
A MIMO channel formed by multiple (T) transmit antennas and multiple (R) receive antennas may be characterized by an R×T channel response matrix H, which may be given as:
where entry for hi,j, for i=1, . . . , R and j=1, . . . , T, denotes the coupling or complex channel gain between transmit antenna j and receive antenna i.
The channel response matrix H may be diagonalized to obtain multiple (S) eigenmodes of H, where S≦min {T, R}. The diagonalization may be achieved by performing either singular value decomposition of H or eigenvalue decomposition of a correlation matrix of H.
The eigenvalue decomposition may be expressed as:
R=HH·H=V·Λ·VH, Eq. (2)
where R is a T×T correlation matrix of H;
V is a T×T unitary matrix whose columns are eigenvectors of R;
Λis a T×T diagonal matrix of eigenvalues of R; and
“H” denotes a conjugate transpose.
The unitary matrix V is characterized by the property VH·V=I, where I is the identity matrix. The columns of the unitary matrix are orthogonal to one another, and each column has unit power. The diagonal matrix Λ contains possible non-zero values along the diagonal and zeros elsewhere. The diagonal elements of Λ are eigenvalues of R. These eigenvalues are denoted as {λ1, λ2, . . . , λS} and represent the power gains for the S eigenmodes. R is a Hermitian matrix whose off-diagonal elements have the following property: ri,j=rj,i*, where “*” denotes the complex conjugate.
The singular value decomposition may be expressed as:
H=U·Σ·VH, Eq. (3)
where U is an R×R unitary matrix of left singular vectors of H;
Σ is an R×T diagonal matrix of singular values of H; and
V is a T×T unitary matrix of right singular vectors of H.
U and V each contain orthogonal vectors. Equations (2) and (3) indicate that the right singular vectors of H are also the eigenvectors of R. The diagonal elements of Σ are the singular values of H. These singular values are denoted as {σ1, σ2, . . . , σS} and represent the channel gains for the S eigenmodes. The singular values of H are also the square roots of the eigenvalues of R, so that σi=√{square root over (λi)} for i=1, . . . , S.
A transmitting entity may use the right singular vectors in V to transmit data on the eigenmodes of H, which typically provides better performance than simply transmitting data from the T transmit antennas without any spatial processing. A receiving entity may use the left singular vectors in U or the eigenvectors in V to receive the data transmission on the eigenmodes of H. Table 1 shows the spatial processing performed by the transmitting entity, the received symbols at the receiving entity, and the spatial processing performed by the receiving entity. In Table 1, s is a T×1 vector with up to S data symbols to be transmitted, x is a T×1 vector with T transmit symbols to be sent from the T transmit antennas, r is an R×1 vector with R received symbols obtained from the R receive antennas, n is an R×1 noise vector, and ŝ is a T×1 vector with up to S detected data symbols, which are estimates of the data symbols in s.
Eigenvalue decomposition and singular value decomposition of a complex matrix may be performed with an iterative process that uses Jacobi rotation, which is also commonly referred to as Jacobi method or Jacobi transformation. The Jacobi rotation zeros out a pair of off-diagonal elements of the complex Hermitian matrix by performing a plane rotation on the matrix. For a 2×2 complex Hermitian matrix, only one iteration of the Jacobi rotation is needed to obtain the two eigenvectors and two eigenvalues for this matrix. For a larger complex matrix with dimension greater than 2×2, the iterative process performs multiple iterations of the Jacobi rotation to obtain the desired eigenvectors and eigenvalues, or singular vectors and singular values, for the larger complex matrix. Each iteration of the Jacobi rotation on the larger complex matrix uses the eigenvectors of a 2×2 submatrix, as described below.
Eigenvalue decomposition of a simple 2×2 Hermitian matrix R2×2 may be performed as follows. The Hermitian matrix R2×2 may be expressed as:
where A, B, and D are arbitrary real values, and θb is an arbitrary phase.
The first step of the eigenvalue decomposition of R2×2 is to apply a two-sided unitary transformation, as follows:
where Rre is a symmetric real matrix containing real values and having symmetric off-diagonal elements at locations (1, 2) and (2,1).
The symmetric real matrix Rre is then diagonalized using a two-sided Jacobi rotation, as follows:
where angle θ may be expressed as:
A 2×2 unitary matrix V2×2 of eigenvectors of R2×2 may be derived as:
The two eigenvalues λ1 and λ2 may be computed directly from the elements of Rre as follows:
Equation (9) is the solution to a characteristic equation of R2×2. In equation (9), λ1 is obtained with the plus sign for the second quantity on the right hand side, and λ2 is obtained with the minus sign for the second quantity, where λ1≧λ2.
The elements of V2×2 may be computed directly from the elements of R2×2, as follows:
where r1,1, r1,2 and r2,1 are elements of R2×2, r is the magnitude of r1,2, and <r1,2 is the phase of r1,2, which is also denoted as θ=<r1,2.
Equation set (10) performs a complex Jacobi rotation on the 2×2 Hermitian matrix R2×2 to obtain the matrix V2×2 of eigenvectors of R2×2. The computations in equation set (10) are derived to eliminate the arc-tangent operation in equation (7) and the cosine and sine operations in equation (8). However, the square root and division operations in equation set (10) present their own implementation difficulties, so a simpler implementation is desirable. Equation set (10) is computed once for eigenvalue decomposition of the 2×2 Hermitian matrix R2×2. A complex matrix larger than 2×2 may be decomposed by performing eigenvalue decomposition of many 2×2 submatrices, as described below. Thus, it is highly desirable to compute equation set (10) as efficiently as possible in order to reduce the amount of time needed to decompose the larger complex matrix.
Variables r, c1, s1 and hence g1 in equation set (10) may be efficiently computed using CORDIC processor 110. A CORDIC processor implements an iterative algorithm that allows for fast hardware calculation of trigonometric functions such as sine, cosine, magnitude, and phase using simple shift and add/subtract hardware. Variables r, c1 and s1 may be computed in parallel to reduce the amount of time needed to perform eigenvalue decomposition. The CORDIC processor computes each variable iteratively, with more iterations producing higher accuracy for the variable.
A complex multiply of two complex numbers, R=R+jRim and C=Cre+jCim, may be expressed as:
The magnitude of Y is equal to the product of the magnitudes of R and C. The phase of Y is equal to the sum of the phases of R and C.
The complex number R may be rotated by up to 90 degrees by multiplying R with a complex number Ci having the following form: Ci=1±jKi, where Ci,re=1 and Ci,im=±Ki. Ki is decreasing powers of two and has the following form:
Ki=2−i, Eq. (13)
where i is an index that is defined as i=0, 1, 2, . . . .
The complex number R may be rotated counter-clockwise if the complex number Ci has the form Ci=1+jKi. The phase of Ci is then <Ci=arctan (Ki). Equation set (12) may then be expressed as:
Yre=Rre−Ki·Rim=Rre−2−i·Rim, and Eq. (14a)
Yim=Rim+Ki·Rre=Rim+2−i·Rre. Eq. (14b)
The complex number R may be rotated clockwise if the complex number C, has the form Ci=1−jKi. The phase of Ci is then <Ci=−arctan (Ki). Equation set (12) may then be expressed as:
Yre=Rre+Ki·Rre+2−i·Rim, and Eq. (15a)
Yim=Rim−K·R=Rim−2−i·Rre. Eq. (15b)
The counter-clockwise rotation in equation set (14) and the clockwise rotation in equation set (15) by the complex number Ci may be achieved by shifting both Rim and Rre by i bit positions, adding/subtracting the shifted Rim to/from Rre to obtain Yre, and adding/subtracting the shifted Rre to/from Rim to obtain Yim. No multiplies are needed to perform the rotation.
Table 2 shows the value of Ki, the complex number Ci, the phase of Ci, the magnitude of Ci, and the CORDIC gain gi for each value of i from 0 through 7. As shown in Table 2, for each value of i, the phase of Ci, is slightly more than half the phase of Ci−1. A given target phase may be obtained by performing a binary search and either adding or subtracting each successively smaller phase value θi. Index i denotes the number of iterations for the binary search, and more iterations give a more accurate final result.
Since the magnitude of Ci is greater than 1.0 for each value of i, multiplication of R with Ci results in the magnitude of R being scaled by the magnitude of Ci. The CORDIC gain for a given value of i is the cumulative magnitude of Ci for the current and prior values of i. The CORDIC gain for i is obtained by multiplying the CORDIC gain for i−1 with the magnitude of Ci, or gi=gi−1·|Ci|. The CORDIC gain is dependent on the value of i but converges to a value of approximately 1.647 as i approaches infinity.
In equation set (10), r is the magnitude of element r1,2 and B is the phase of element r1,2. The magnitude and phase of r1,2 may be determined by CORDIC processor 110 as follows. A variable {tilde over (r)}1,2 is formed with the absolute values of the real and imaginary parts of r1,2, or {tilde over (r)}1,2=abs (Re {r1,2})+j abs (Im {r1,2}). Thus, {tilde over (r)}1,2 sits on the first quadrant of an x-y plane. The phase θ is initialized to zero. {tilde over (r)}1,2 is then iteratively rotated such that its phase approaches zero.
For each iteration starting with i=0, {tilde over (r)}1,2 is deemed to have (1) a positive phase if the imaginary part of {tilde over (r)}1,2 is positive or (2) a negative phase if the imaginary part of {tilde over (r)}1,2 is negative. If the phase of {tilde over (r)}1,2 is negative, then {tilde over (r)}1,2 is rotated counter-clockwise by θi (or equivalently, θi is added to the phase of {tilde over (r)}1,2) by multiplying {tilde over (r)}1,2 with Ci=1+jKi, as shown in equation set (14). Conversely, if the phase of {tilde over (r)}1,2 is positive, then {tilde over (r)}1,2 is rotated clockwise by θi (or equivalently, θi is subtracted from the phase of {tilde over (r)}1,2) by multiplying {tilde over (r)}1,2 with Ci=1−jKi, as shown in equation set (15). {tilde over (r)}1,2 is thus updated in each iteration by either a counter-clockwise or clockwise rotation. The phase θ is updated by (1) adding θi to the current value of θ if θi was added to the phase of {tilde over (r)}1,2 or (2) subtracting θi from the current value of θ if θi was subtracted from the phase of {tilde over (r)}1,2. θ thus represents the cumulative phase that has been added to or subtracted from the phase of {tilde over (r)}1,2 to zero out the phase.
The final result becomes more accurate as more iterations are performed. Ten iterations are typically sufficient for many applications. After all of the iterations are completed, the phase of {tilde over (r)}1,2 should be close to zero, the imaginary part of {tilde over (r)}1,2 should be approximately zero, and the real part of {tilde over (r)}1,2 is equal to the magnitude of r1,2 scaled by the CORDIC gain, or r=Re {{tilde over (r)}1,2}/gi. The final value of θ is the total phase rotation to zero out the phase of {tilde over (r)}1,2. The phase of {tilde over (r)}1,2 is thus equal to −θ. The phase θ may be represented by a sequence of sign bits, z1 z2 z3 . . . , where zi=1 if was subtracted from θ and zi=−1 if θi was added to θ.
The computation of the magnitude and phase of r1,2 may performed as follows. First the variables are initialized as:
i=0, Eq. (16a)
x0=abs(Re{r1,2}), Eq. (16b)
y0=abs(Im{r1,2}), and Eq. (16c)
θtot(i)=0. Eq. (16d)
A single iteration of the CORDIC computation may be expressed as:
In equations (17b) and (17c), a counter-clockwise rotation is performed if the phase of xi+jyi is positive and zi=1, and a clockwise rotation is performed if the phase of xi+jyi is negative and zi=−1. After all of the iterations are completed, the magnitude is set as r=xi+1 and the phase is set as θ=θtot(i+1). The scaling by the CORDIC gain may be accounted for by other processing blocks.
In equation set (10), c1 is the cosine of r1,2 and s1 is the sine of r1,2. The cosine and sine of r1,2 may be determined by CORDIC processor 110 as follows. A unit magnitude complex number R′ is initialized as R′=x0′+jy0′=1+j0 and is then rotated by −θ. For each iteration starting with i=0, the complex number R′ is rotated (1) counter-clockwise by θi by multiplying R′ with Ci=1+jKi if sign bit zi indicates that 9 was subtracted from 9 or (2) clockwise by θi by multiplying R′ with Ci=1jKi if sign bit zi indicates that θi was added to θ. After all of the iterations are completed, c1 is equal to the real part of the final R′ scaled by the CORDIC gain, or c1=xi′/gi, and s1 is equal to the imaginary part of the final R′ scaled by the CORDIC gain, or s1=yi′/gi.
The computation of the cosine and sine of r1,2 may be performed as follows. First the variables are initialized as:
i=0, Eq. (18a)
x0′=1, and Eq. (18b)
y0′=0. Eq. (18c)
A single iteration of the CORDIC computation may be expressed as:
xi+1′=xi′−zi·2−i·yi′, Eq. (19a)
yi+1′=yi′+zi·2−i·xi′, and Eq. (19b)
i=i+1. Eq. (19c)
In equations (19a) and (19b), for each iteration i, R′ is rotated in the direction indicated by sign bit zi. After all of the iterations are completed, the cosine is set as c1=xi+1′ and the sine is set as s1=yi+1′. The scaling by the CORDIC gain may be accounted for by other processing blocks.
The cosine c1 and the sine s1 may also be computed in parallel with the magnitude r by initializing the variable R′ to a CORDIC gain scaled version of 1+j0, or R′=(1+j0)/g where g is the CORDIC gain for the number of iterations to be performed. At each iteration, the CORDIC rotation performed for the magnitude r is determined, and an opposite CORDIC rotation is performed on the variable R′. For this scheme, it is not necessary to determine the phase θ.
Within CORDIC unit 210, a demultiplexer (Demux) 208 receives element r1,2, provides abs (Re {r1,2}) as x0, and provides abs (Im {r1,2}) as y0. A multiplexer (Mux) 212a receives x0 on a first input and xi from a delay element 219a on a second input, provides x0 on its output when i=0, and provides xi on its output when i>0. The output of multiplexer 212a is xi for the current iteration. A shifter 214a receives and shifts xi to the left by i bits and provides a shifted xi. A multiplexer 212b receives yo on a first input and yi from a delay element 219b on a second input, provides y0 on its output when i=0, and provides yi on its output when i>0. The output of multiplexer 212b is yi for the current iteration. A shifter 214b receives and shifts yi to the left by i bits and provides a shifted yi. A sequencer 222 steps through index i and provides appropriate controls for the units within CORDIC processor 110. A sign detector 224 detects the sign of yi and provides sign bit zi, as shown in equation (17a).
A multiplier 216a multiplies the shifted xi with sign bit zi. A multiplier 216b multiplies the shifted yi with sign bit zi. Multipliers 216a and 216b may be implemented with bit inverters. A summer 218a sums the output of multiplier 216b with xi and provides xi+1 for the current iteration (which is also xi for the next iteration) to delay element 219a and a switch 220. A summer 218b subtracts the output of multiplier 216a from yi and provides yi+1 for the current iteration (which is also yi for the next iteration) to delay element 219b. Switch 220 provides xi+1 as the magnitude r after all of the iterations are completed.
Within CORDIC unit 230, a multiplexer 232a receives x0′=1 on a first input and xi′ from a delay element 239a on a second input, provides x0′ on its output when i=0, and provides xi′ on its output when i>0. The output of multiplexer 232a is xi′ for the current iteration. A shifter 234a receives and shifts xi′ to the left by i bits and provides a shifted xi′. A multiplexer 232b receives y0′=0 on a first input and yi′ from a delay element 239b on a second input, provides y0′ on its output when i=0, and provides yi′ on its output when i>0. The output of multiplexer 232b is yi′ for the current iteration. A shifter 234b receives and shifts yi′ to the left by i bits and provides a shifted yi′.
A multiplier 236a multiplies the shifted xi′ with sign bit zi from detector 224. A multiplier 236b multiplies the shifted yi′ with sign bit zi. Multipliers 236a and 236b may also be implemented with bit inverters. A summer 238a subtracts the output of multiplier 236b from xi′ and provides xi+1′ for the current iteration (which is also xi′ for the next iteration) to delay element 239a and a switch 240a. A summer 238b adds the output of multiplier 236a with yi′ and provides yi+1′ for the current iteration (which is also yi′ for the next iteration) to delay element 239b and a switch 240b. After all of the iterations are completed, switch 240a provides xi+1′ as cosine c1, and switch 240b provides yi+1′ as sine s1.
Look-Up TableIn equation set (10), variables c and s are functions of only τ, and intermediate variables x and t are used to simplify the notation. abs(τ) ranges from 0 to ∞, variable c ranges from 0.707 to 1.0, and variable s ranges from 0.707 to 0.0. A large range of values for τ is thus mapped to a small range of values for c and also a small range of values for s. Hence, an approximate value of r should give good approximations of both c and s.
A look-up table (LUT) may be used to efficiently compute variables c and s based on the dividend/numerator and the divisor/denominator for r. The use of the look-up table avoids the need to perform a division to compute r in equation (10e), a square root to compute x in equation (10f), a division to compute t in equation (10g), a division and a square root to compute c in equation (10h), and a multiply to compute s in equation (10i). Since division and square root operations are computationally intensive, the use of the look-up table can greatly reduce the amount of time needed to perform eigenvalue decomposition. The computation of variables c and s using the look-up table may be performed as follows.
The dividend is equal to r2,2−r1,1, and the absolute value of the dividend is represented as a binary floating-point number of the form 1.bn
The divisor is equal to 2·r and is a positive value because r is the magnitude of r1,2. The divisor is also represented as a binary floating-point number of the form 1.bd
The fractional bits of the mantissa for the dividend (which are bn
In general, the look-up table may be of any size. A larger size look-up table can provide greater accuracy in the computation of variables c and s. In a specific embodiment, the look-up table has a size of 2K×16, an 11-bit input address, a 8-bit output for variable c, and a 8-bit output for variable s. The 11-bit input address is composed of three fractional bits bn
Look-up table 320 receives the fractional bits {bn} for the dividend, the fractional bits {bd} for the divisor, and the exponent difference Δm as an input address. Look-up table 320 provides the stored values for variables c and s based on the input address. An output unit 322 appends a ‘1’ for the leftmost fractional bit for variable c, further appends a plus sign bit (‘+’), and provides the final value of c. An output unit 324 appends a plus sign bit (‘+’) for variable s and provides the final value of s.
Look-up table 320 may be designed to account for the CORDIC gain in the computation of r, c1, and s1 so that the elements of V2×2 have the proper magnitude. For example, since the magnitude r is used to form the address for look-up table 320, the CORDIC gain for the magnitude r may be accounted for in the addressing of the look-up table. In another embodiment, look-up table 320 stores a rotation sequence for a CORDIC processor, which then computes variables c and s with the rotation sequence. The rotation sequence is the sequence of sign bits zi and may be stored using fewer bits than the actual values for variables c and s. However, the CORDIC processor would require some amount of time to compute variables c and s based on the rotation sequence.
In the description above, the variables r, c1 and s1 are derived with a CORDIC processor and the variables c and s are derived with a look-up table. The variables r, c1 and s1 may also be derived with one or more look-up tables of sufficient size to obtain the desired accuracy for r, ci and s1. The variables r, c1 and s1 and the variables c and s may also be computed in other manners and/or using other algorithms (e.g., power series). The choice of which method and algorithm to compute each set of variables may be dependent on various factors such as the available hardware, the amount of time available for computation, and so on.
Eigenvalue DecompositionEigenvalue decomposition of an N×N Hermitian matrix that is larger than 2×2, as shown in equation (2), may be performed with an iterative process. This iterative process uses the Jacobi rotation repeatedly to zero out off-diagonal elements in the N×N Hermitian matrix. For the iterative process, N×N unitary transformation matrices are formed based on 2×2 Hermitian sub-matrices of the N×N Hermitian matrix and are repeatedly applied to diagonalize the N×N Hermitian matrix. Each N×N unitary transformation matrix contains four non-trivial elements (elements other than 0 or 1) that are derived from elements of a corresponding 2×2 Hermitian sub-matrix. The resulting diagonal matrix contains the real eigenvalues of the N×N Hermitian matrix, and the product of all of the unitary transformation matrices is an N×N matrix of eigenvectors for the N×N Hermitian matrix.
In the following description, index i denotes the iteration number and is initialized as i=0. R is an N×N Hermitian matrix to be decomposed, where N>2. An N×N matrix Di is an approximation of diagonal matrix Λ of eigenvalues of R and is initialized as D0=R. An N×N matrix Vi is an approximation of matrix V of eigenvectors of R and is initialized as V0=I.
A single iteration of the Jacobi rotation to update matrices Di and Vi may be performed as follows. First, a 2×2 Hermitian matrix Dpq is formed based on the current Di as follows:
where dp,q is the element at location (p,q) in Di; and
pε{1, . . . , N}, qε{1, . . . , N}, and p≠q.
Dpq is a 2×2 submatrix of Di, and the four elements of Dpq are four elements at locations (p,p), (p,q), (q,p) and (q,q) in Di. The values for indices p and q may be selected in a predetermined or deterministic manner, as described below.
Eigenvalue decomposition of Dpq is then performed as shown in equation set (10) to obtain a 2×2 unitary matrix Vpq of eigenvectors of Dpq. For the eigenvalue decomposition of ppq, R2×2 in equation (4) is replaced with ppq, and V2×2 from equation (10j) or (10k) is provided as Vpq.
An N×N complex Jacobi rotation matrix Tpq is then formed with matrix Vpq Tpq is an identity matrix with the four elements at locations (p,p), (p,q), (q,p) and (q,q) replaced with the (1,1), (1,2), (2,1) and (2,2) elements, respectively, of Vpq. Tpq has the following form:
where v1,1, v1,2, v2,1 and v2,2 are the four elements of Vpq. All of the other off-diagonal elements of Tpq are zeros. Equations (10j) and (10k) indicate that Tpq is a complex matrix containing complex values for v2,1 and v2,2. Tpq is also called a transformation matrix that performs the Jacobi rotation.
Matrix Di is then updated as follows:
Di+1=TpqH·Di·Tpq. Eq. (22)
Equation (22) zeros out two off-diagonal elements dp,q and dq,p at locations (p,q) and (q,p), respectively, in Di. The computation may alter the values of other off-diagonal elements in Di.
Matrix Vi is also updated as follows:
Vi+1=Vi·Tpq Eq. (23)
Vi may be viewed as a cumulative transformation matrix that contains all of the Jacobi rotation matrices Tpq used on Di.
The Jacobi rotation matrix Tpq may also be expressed as a product of (1) a diagonal matrix with N−1 ones elements and one complex-valued element and (2) a real-valued matrix with N−2 ones along the diagonal, two real-valued diagonal elements, two real-valued off-diagonal elements, and zeros elsewhere. As an example, for p=1 and q=2, Tpq has the following form:
where g1 is a complex value and c and s are real values. The update of Di in equation (22) may then be performed with 12(N−2)+8 real multiplies, and the update of Vi in equation (23) may be performed with 12N real multiplies. A total of 24N−16 real multiples are then performed for one iteration. Besides the Tpq structure, the number of multiplies to update Di is reduced by the fact that Di remains Hermitian after the update and that there is a 2×2 diagonal sub-matrix after the update with real-valued eigenvalues as the diagonal elements.
Each iteration of the Jacobi rotation zeros out two off-diagonal elements of Di. Multiple iterations of the Jacobi rotation may be performed for different values of indices p and q to zero out all of the off-diagonal elements of Di. The indices p and q may be selected in a predetermined manner by sweeping through all possible values.
A single sweep across all possible values for indices p and q may be performed as follows. The index p is stepped from 1 through N−1 in increments of one. For each value of p, the index q is stepped from p+1 through N in increments of one. An iteration of the Jacobi rotation to update Di and Vi may be performed for each different combination of values for p and q. For each iteration, Dpq is formed based on the values of p and q and the current Di for that iteration, Vpq is computed for Dpq as shown in equation set (10), Tpq is formed with Vpq as shown in equation (21) or (24), Di is updated as shown in equation (22), and Vi is updated as shown in equation (23). For a given combination of values for p and q, the Jacobi rotation to update D and V, may also be skipped if the magnitude of the off-diagonal elements at locations (p,q) and (q,p) in Di are below a predetermined threshold.
A sweep consists of N·(N−1)/2 iterations of the Jacobi rotation to update Di and Vi for all possible values of p and q. Each iteration of the Jacobi rotation zeros out two off-diagonal elements of Di but may alter other elements that might have been zeroed out earlier. The effect of sweeping through indices p and q is to reduce the magnitude of all off-diagonal elements of Di, so that Di approaches the diagonal matrix Λ. Vi contains an accumulation of all Jacobi rotation matrices that collectively give Di. Vi thus approaches V as Di approaches Λ.
Any number of sweeps may be performed to obtain more and more accurate approximations of V and Λ. Computer simulations have shown that four sweeps should be sufficient to reduce the off-diagonal elements of Di to a negligible level, and three sweeps should be sufficient for most applications. A predetermined number of sweeps (e.g., three or four sweeps) may be performed. Alternatively, the off-diagonal elements of Di may be checked after each sweep to determine whether Di is sufficiently accurate. For example, the total error (e.g., the power in all off-diagonal elements of Di) may be computed after each sweep and compared against an error threshold, and the iterative process may be terminated if the total error is below the error threshold. Other conditions or criteria may also be used to terminate the iterative process.
The values for indices p and q may also be selected in a deterministic manner. As an example, for each iteration i, the largest off-diagonal element of Di is identified and denoted as dp,q. The iteration is then performed with Dpq containing this largest off-diagonal element dp,q and three other elements at locations (p,p), (q,p), and (q,q) in Di. The iterative process may be performed until a termination condition is encountered. The termination condition may be, for example, completion of a predetermined number of iterations, satisfaction of the error criterion described above, or some other condition or criterion.
Upon termination of the iterative process, the final Vi is a good approximation of V, and the final Di is a good approximation of Λ. The columns of Vi may be used as the eigenvectors of R, and the diagonal elements of Di may be used as the eigenvalues of R. The eigenvalues in the final Di are ordered from largest to smallest because the eigenvectors in Vpq for each iteration are ordered. The eigenvectors in the final Vi are also ordered based on their associated eigenvalues in Di.
Except for the first iteration, the computation of Tpq and the updates of Di and Vi do not have to proceed in a sequential order, assuming that the computations do not share the same hardware units. Since the updates of Di and Vi involve matrix multiplies, it is likely that these updates will proceed in a sequential order using the same hardware. The computation of Tpq for the next iteration can start as soon as the off-diagonal elements of Di have been updated for the current iteration. The computation of Tpq may be performed with dedicated hardware while Vi is updated. If the Jacobi rotation computation is finished by the time the Vi update is done, then the Di update for the next iteration can start as soon as the Vi update for the current iteration is completed.
For iteration i, the values for indices p and q are selected in a predetermined manner (e.g., by stepping through all possible values for these indices) or a deterministic manner (e.g., by selecting the index values for the largest off-diagonal element) (block 512). A 2×2 matrix Dpq is then formed with four elements of matrix Di at the locations determined by indices p and q (block 514). Eigenvalue decomposition of Dpq is then performed as shown in equation set (10) to obtain a 2×2 matrix Vpq of eigenvectors of Dpq (block 516). This eigenvalue decomposition may be efficiently performed as described above for
An N×N complex Jacobi rotation matrix Tpq is then formed based on matrix Vpq, as shown in equation (21) or (24) (block 518). Matrix Di is then updated based on Tpq as shown in equation (22) (block 520). Matrix Vi is also updated based on Tpq, as shown in equation (23) (block 522).
A determination is then made whether to terminate the eigenvalue decomposition of R (block 524). The termination criterion may be based on the number of iterations or sweeps already performed, an error criterion, and so on. If the answer is ‘No’ for block 524, then index i is incremented (block 526), and the process returns to block 512 for the next iteration. Otherwise, if termination is reached, then matrix Di is provided as an approximation of diagonal matrix Λ, and matrix Vi is provided as an approximation of matrix V of eigenvectors of R (block 528).
For a multi-carrier MIMO system (e.g., a MIMO system that utilizes OFDM), multiple channel response matrices H(k) may be obtained for different subbands. The iterative process may be performed for each channel response matrix H(k) to obtain matrices Di (k) and Vi(k), which are approximations of diagonal matrix Λ(k) and matrix V(k) of eigenvectors, respectively, of R(k)=HH(k)·H(k).
A high degree of correlation typically exists between adjacent subbands in a MIMO channel. This correlation may be exploited by the iterative process to reduce the amount of computation to derive Di (k) and Vi(k) for all subbands of interest. For example, the iterative process may be performed for one subband at a time, starting from one end of the system bandwidth and traversing toward the other end of the system bandwidth. For each subband k except for the first subband, the final solution Vi(k−1) obtained for the prior subband k±1 may be used as an initial solution for the current subband k. The initialization for each subband k may be given as: V0(k)=Vi(k−1) and D0(k)=V0H(k)·R(k)·V0(k). The iterative process then operates on the initial solutions of D0(k) and V0(k) for subband k until a termination condition is encountered.
The concept described above may also be used across time. For each time interval t, the final solution Vi(t−1) obtained for a prior time interval t−1 may be used as an initial solution for the current time interval t. The initialization for each time interval t may be given as: V0(t)=Vi(t−1) and D0(t)=V0H(t)·R(t)·V0(t), where R(t)=HH/(t)·H(t) and H(t) is the channel response matrix for time interval t. The iterative process then operates on the initial solutions of D0(t) and V0(t) for time interval t until a termination condition is encountered.
Singular Value DecompositionThe iterative process may also be used for singular value decomposition (SVD) of an arbitrary complex matrix H larger than 2×2. H has a dimension of R×T, where R is the number of rows and T is the number of columns. The iterative process for singular value decomposition of H may be performed in several manners.
In a first SVD embodiment, the iterative process derives approximations of the right singular vectors in V and the scaled left singular vectors in U·Σ. For this embodiment, a T×T matrix Vi is an approximation of V and is initialized as V0=I. An R×T matrix Wi is an approximation of U·Σ and is initialized as W0=H.
For the first SVD embodiment, a single iteration of the Jacobi rotation to update matrices Vi and Wi may be performed as follows. First, a 2×2 Hermitian matrix Mpq is formed based on the current Wi. Mpq is a 2×2 submatrix of WiH·W and contains four elements at locations (p,p), (p,q), (q,p) and (q,q) in WiH·Wi. The elements of Mpq may be computed as follows:
where wp is column p of Wi, wq is column q of Wi, and wl,p is the element at location (l,p) in Wi. Indices p and q are such that pε{1, . . . , T}, qε{1, . . . , T}, and p#q. The values for indices p and q may be selected in a predetermined or deterministic manner, as described above.
Eigenvalue decomposition of Mpq is then performed as shown in equation set (10) to obtain a 2×2 unitary matrix Vpq of eigenvectors of Mpq. For this eigenvalue decomposition, R2×2 is replaced with Mpq, and V2×2 is provided as Vpq.
A T×T complex Jacobi rotation matrix Tpq is then formed with matrix Vpq. Tpq is an identity matrix with the four elements at locations (p,p), (p,q), (q,p) and (q,q) replaced with the (1,1), (1,2), (2,1) and (2,2) elements, respectively, of Vpq Tpq has the form shown in equations (21) and (24).
Matrix Vi is then updated as follows:
Vi+1=Vi·Tpq. Eq. (26)
Matrix Wi is also updated as follows:
Wi+1=Wi·Tpq. Eq. (27)
The iterative process is performed until a termination condition is encountered, which may be a predetermined number of sweeps or iterations, satisfaction of an error criterion, and so on. Upon termination of the iterative process, the final Vi is a good approximation of V, and the final Wi is a good approximation of U·Σ. When converged, WiH·Wi=ΣT·Σ and U=Wi·Σ−1, where “T” denotes a transpose. For a square diagonal matrix, the final solution of Σ may be given as: {circumflex over (Σ)}=(WiH·Wi)1/2. For a non-square diagonal matrix, the non-zero diagonal values of {circumflex over (Σ)} are given by the square roots of the diagonal elements of WiH·Wi. The final solution of U may be given as: Û=Wi·{circumflex over (Σ)}−1.
The left singular vectors of H may be obtained by performing the first SVD embodiment and solving for scaled left singular vectors H·V=U·Σ and then normalizing. The left singular vectors of H may also be obtained by performing the iterative process for eigenvalue decomposition of H·HH.
In a second SVD embodiment, the iterative process directly derives approximations of the right singular vectors in V and the left singular vectors in U. This SVD embodiment applies the Jacobi rotation on a two-sided basis to simultaneously solve for the left and right singular vectors. For the second SVD embodiment, a T×T matrix Vi is an approximation of V and is initialized as V0=I. An R×R matrix Ui is an approximation of U and is initialized as U0=I. An R×T matrix Di is an approximation of Σ and is initialized as D0=H.
For the second SVD embodiment, a single iteration of the Jacobi rotation to update matrices Vi, Ui, and Di may be performed as follows. First, a 2×2 Hermitian matrix Xp
where dp
Eigenvalue decomposition of Xp
Another 2×2 Hermitian matrix Yp
where {tilde over (d)}p
Eigenvalue decomposition of Yp
Matrix Vi is then updated as follows:
Vi+1=Vi·Tp
Matrix Ui is updated as follows:
Ui+1=Ui·Sp
Matrix Di is updated as follows:
Di+1=Sp
The iterative process is performed until a termination condition is encountered. Upon termination of the iterative process, the final Vi is a good approximation of {tilde over (V)}, the final Ui is a good approximation of U, and the final Di is a good approximation of {tilde over (Σ)}, where {tilde over (V)} and {tilde over (Σ)} may be rotated versions of V and Σ, respectively. Vi and Di may be unrotated as follows:
{circumflex over (Σ)}=Di·P, and Eq. (33a)
{circumflex over (V)}=Vi·P, Eq. (33b)
where P is a T×T diagonal matrix with diagonal elements having unit magnitude and phases that are the negative of the phases of the corresponding diagonal elements of Di. {circumflex over (Σ)} and {circumflex over (V)} are the final approximations of Σand V, respectively.
For a multi-carrier MIMO system, the iterative process may be performed for each channel response matrix H(k) to obtain matrices Vi(k), Ui(k), and Di(k), which are approximations matrices V(k), U(k), and Σ(k), respectively, for that H(k). For the first SVD embodiment, for each subband k except for the first subband, the final solution Vi(k−1) obtained for the prior subband k−1 may be used as an initial solution for the current subband k, so that V0(k)=Vi(k−1) and W0(k)=H(k)·V0(k). For the second SVD embodiment, for each subband k except for the first subband, the final solutions Vi(k−1) and Ui(k−1) obtained for the prior subband k−1 may be used as initial solutions for the current subband k, so that V0(k)=Vi(k−1), U0(k)=Ui(k−1), and D0(k)=U0H(k)·H(k)·V0(k). The concept may also be used across time or both frequency and time, as described above.
SystemOn the downlink, at access point 610, a transmit (TX) data processor 614 receives traffic data from a data source 612 and other data from a controller 630. TX data processor 614 formats, encodes, interleaves, and modulates the data and generates data symbols, which are modulation symbols for data. A TX spatial processor 620 multiplexes the data symbols with pilot symbols, performs spatial processing with eigenvectors or right singular vectors if applicable, and provides Nap streams of transmit symbols. Each transmitter unit (TMTR) 622 processes a respective transmit symbol stream and generates a downlink modulated signal. Nap downlink modulated signals from transmitter units 622a through 622ap are transmitted from antennas 624a through 624ap, respectively.
At user terminal 650, Nut antennas 652a through 652ut receive the transmitted downlink modulated signals, and each antenna provides a received signal to a respective receiver unit (RCVR) 654. Each receiver unit 654 performs processing complementary to the processing performed by transmitter units 622 and provides received symbols. A receive (RX) spatial processor 660 performs spatial matched filtering on the received symbols from all receiver units 654a through 654ut and provides detected data symbols, which are estimates of the data symbols transmitted by access point 610. An RX data processor 670 processes (e.g., symbol demaps, deinterleaves, and decodes) the detected data symbols and provides decoded data to a data sink 672 and/or a controller 680.
A channel estimator 678 processes received pilot symbols and provides an estimate of the downlink channel response, Ĥ(k), for each subband of interest. Controller 680 may decompose each matrix Ĥ(k) to obtain {circumflex over (V)}(k) and {circumflex over (Σ)}(k), which are estimates of V(k) and Σ(k) for H(k). Controller 680 may derive a downlink spatial filter matrix Mdn(k) for each subband of interest based on {circumflex over (V)}(k), as shown in Table 1. Controller 680 may provide Mdn(k) to RX spatial processor 660 for downlink spatial matched filtering and {circumflex over (V)}(k) to a TX spatial processor 690 for uplink spatial processing.
The processing for the uplink may be the same or different from the processing for the downlink. Data from a data source 686 and signaling from controller 680 are processed (e.g., encoded, interleaved, and modulated) by a TX data processor 688, multiplexed with pilot symbols, and further spatially processed by TX spatial processor 690 with {circumflex over (V)}(k) for each subband of interest. The transmit symbols from TX spatial processor 690 are further processed by transmitter units 654a through 654ut to generate Nut uplink modulated signals, which are transmitted via antennas 652a through 652ut.
At access point 610, the uplink modulated signals are received by antennas 624a through 624ap and processed by receiver units 622a through 622ap to generate received symbols for the uplink transmission. An RX spatial processor 640 performs spatial matched filtering on the received data symbols and provides detected data symbols. An RX data processor 642 further processes the detected data symbols and provides decoded data to a data sink 644 and/or controller 630.
A channel estimator 628 processes received pilot symbols and provides an estimate of either HT(k) or U(k) for each subband of interest, depending on the manner in which the uplink pilot is transmitted. Controller 630 may receive ĤT(k) for multiple subbands and decompose each matrix ĤT(k) to obtain Û(k). Controller 680 may also derive an uplink spatial filter matrix Mup(k) for each subband of interest based on Û(k). Controller 630 provides Mup(k) to RX spatial processor 640 for uplink spatial matched filtering and Û(k) to TX spatial processor 620 for downlink spatial processing.
Controllers 630 and 680 control the operation at access point 610 and user terminal 650, respectively. Memory units 632 and 682 store data and program codes used by controllers 630 and 680, respectively. Controllers 630 and/or 680 may perform eigenvalue decomposition and/or singular value decomposition of the channel response matrices obtained for its link.
The decomposition techniques described herein may be implemented by various means. For example, these techniques may be implemented in hardware, software, or a combination thereof. For a hardware implementation, the processing units used to perform decomposition may be implemented within one or more application specific integrated circuits (ASICs), digital signal processors (DSPs), digital signal processing devices (DSPDs), programmable logic devices (PLDs), field programmable gate arrays (FPGAs), processors, controllers, micro-controllers, microprocessors, other electronic units designed to perform the functions described herein, or a combination thereof.
For a software implementation, the decomposition techniques may be implemented with modules (e.g., procedures, functions, and so on) that perform the functions described herein. The software codes may be stored in a memory unit (e.g., memory unit 632 or 682 in
Headings are included herein for reference and to aid in locating certain sections. These headings are not intended to limit the scope of the concepts described therein under, and these concepts may have applicability in other sections throughout the entire specification.
The previous description of the disclosed embodiments is provided to enable any person skilled in the art to make or use the present invention. Various modifications to these embodiments will be readily apparent to those skilled in the art, and the generic principles defined herein may be applied to other embodiments without departing from the spirit or scope of the invention. Thus, the present invention is not intended to be limited to the embodiments shown herein but is to be accorded the widest scope consistent with the principles and novel features disclosed herein.
Claims
1. A computer-program storage apparatus for decomposing a matrix comprising a memory unit having software codes stored thereon, the software codes being executable by one or more processors and the software codes comprising:
- code for deriving a first set of at least one variable based on a first matrix to be decomposed and using Coordinate Rotational Digital Computer (CORDIC) computation;
- code for deriving a second set of at least one variable based on the first matrix and using a look-up table; and
- code for deriving a second matrix of eigenvectors based on the first and second sets of at least one variable.
2. The computer-program storage apparatus of claim 1, wherein the code for deriving the first set of at least one variable comprises:
- code for performing CORDIC computation on an element of the first matrix to determine magnitude and phase of the element; and
- code for performing CORDIC computation on the phase of the element to determine sine and cosine of the element, and
- wherein the first set of at least one variable comprises the sine and cosine of the element.
3. The computer-program storage apparatus of claim 1, wherein the code for deriving the second set of at least one variable comprises:
- code for deriving intermediate quantities based on the first matrix; and
- code for deriving the second set of at least one variable based on the intermediate quantities and using the look-up table.
4. The computer-program storage apparatus of claim 3, wherein the code for deriving the intermediate quantities comprises:
- code for deriving a divisor for an intermediate variable based on a first element of the first matrix;
- code for converting the divisor into a first binary floating-point number;
- code for deriving a dividend for the intermediate variable based on second and third elements of the first matrix;
- code for converting the dividend into a second binary floating-point number; and
- code for forming the intermediate quantities based on the first and second floating-point numbers.
5. A computer-program storage apparatus for decomposing a matrix comprising a memory unit having software codes stored thereon, the software codes being executable by one or more processors and the software codes comprising:
- code for deriving intermediate quantities based on a first matrix to be decomposed;
- code for deriving at least one variable based on the intermediate quantities and using a look-up table; and
- code for deriving a second matrix of eigenvectors based on the at least one variable.
6. The computer-program storage apparatus of claim 5, wherein the code for deriving the intermediate quantities based on the first matrix comprises:
- code for deriving a divisor for an intermediate variable based on a first element of the first matrix;
- code for converting the divisor into a first binary floating-point number;
- code for deriving a dividend for the intermediate variable based on second and third elements of the first matrix;
- code for converting the dividend into a second binary floating-point number; and
- code for forming the intermediate quantities based on the first and second floating-point numbers.
7. The computer-program storage apparatus of claim 6, wherein the code for forming the intermediate quantities based on the first and second floating-point numbers comprises:
- code for deriving a first intermediate quantity based on a mantissa of the first floating-point number;
- code for deriving a second intermediate quantity based on a mantissa of the second floating-point number; and
- code for deriving a third intermediate quantity based on exponents of the first and second floating-point numbers.
8. The computer-program storage apparatus of claim 7, wherein the code for deriving the at least one variable comprises:
- code for forming an input address for the look-up table based on the first, second, and third intermediate quantities; and
- code for accessing the look-up table with the input address.
9. A computer-program storage apparatus for decomposing a matrix comprising a memory unit having software codes stored thereon, the software codes being executable by one or more processors and the software codes comprising:
- code for performing Coordinate Rotational Digital Computer (CORDIC) computation on an element of a first matrix to determine magnitude and phase of the element;
- code for performing CORDIC computation on the phase of the element to determine sine and cosine of the element; and
- code for deriving a second matrix of eigenvectors based on the magnitude, sine, and cosine of the element.
10. The computer-program storage apparatus of claim 9, wherein the CORDIC computation on the element to determine the magnitude and phase of the element and the CORDIC computation on the phase of the element to determine sine and cosine of the element are performed in parallel.
11. The computer-program storage apparatus of claim 9, wherein the CORDIC computation on the element to determine the magnitude and phase of the element and the CORDIC computation on the phase of the element to determine sine and cosine of the element are performed for a predetermined number of iterations.
12. A computer-program storage apparatus for decomposing a matrix comprising a memory unit having software codes stored thereon, the software codes being executable by one or more processors and the software codes comprising:
- code for performing a plurality of iterations of Jacobi rotation on a first matrix of complex values with a plurality of Jacobi rotation matrices, each Jacobi rotation matrix being derived by performing eigenvalue decomposition of a correlation submatrix using Coordinate Rotational Digital Computer (CORDIC) computation, a look-up table, or both the CORDIC computation and the look-up table; and
- code for deriving a first unitary matrix with orthogonal vectors based on the plurality of Jacobi rotation matrices.
13. The computer-program storage apparatus of claim 12, wherein the code for performing the plurality of iterations of the Jacobi rotation comprises, for each iteration:
- code for forming the submatrix based on the first matrix;
- code for decomposing the submatrix using the CORDIC computation, the look-up table, or both the CORDIC computation and the look-up table to obtain eigenvectors of the submatrix;
- code for forming a Jacobi rotation matrix with the eigenvectors; and
- code for updating the first matrix with the Jacobi rotation matrix.
14. The computer-program storage apparatus of claim 13, wherein the code for decomposing the submatrix comprises:
- code for deriving a first set of at least one variable based on the submatrix and using the CORDIC computation;
- code for deriving a second set of at least one variable based on the submatrix and using the look-up table; and
- code for deriving the eigenvectors of the submatrix based on the first and second sets of at least one variable.
15. The computer-program storage apparatus of claim 12, further comprising:
- code for deriving a diagonal matrix of eigenvalues based on the plurality of Jacobi rotation matrices.
16. The computer-program storage apparatus of claim 12, further comprising:
- code for deriving a second matrix of complex values based on the plurality of Jacobi rotation matrices; and
- code for deriving a second unitary matrix with orthogonal vectors based on the second matrix.
17. The computer-program storage apparatus of claim 16, further comprising:
- code for deriving a diagonal matrix of singular values based on the second matrix.
18. The computer-program storage apparatus of claim 12, further comprising:
- code for deriving a second unitary matrix with orthogonal vectors based on the plurality of Jacobi rotation matrices.
19. The computer-program storage apparatus of claim 18, further comprising:
- code for deriving a diagonal matrix of singular values based on the plurality of Jacobi rotation matrices.
20. A computer-program storage apparatus for decomposing a matrix comprising a memory unit having software codes stored thereon, the software codes being executable by one or more processors and the software codes comprising:
- code for obtaining a plurality of matrices of complex values for a plurality of transmission spans;
- code for performing a plurality of iterations of Jacobi rotation on a first matrix of complex values for a first transmission span to obtain a first unitary matrix with orthogonal vectors, wherein each iteration of the Jacobi rotation utilizes eigenvalue decomposition of a correlation submatrix using Coordinate Rotational Digital Computer (CORDIC) computation, a look-up table, or both the CORDIC computation and the look-up table; and
- code for performing a plurality of iterations of the Jacobi rotation on a second matrix of complex values for a second transmission span to obtain a second unitary matrix with orthogonal vectors, wherein the first unitary matrix is used as an initial solution for the second unitary matrix, wherein the first and second matrices are among the plurality of matrices, and wherein the first and second transmission spans are among the plurality of transmission spans.
21. The computer-program storage apparatus of claim 20, further comprising:
- for each remaining one of the plurality of matrices of complex values, code for performing a plurality of iterations of the Jacobi rotation on the matrix of complex values to obtain a unitary matrix with orthogonal vectors, wherein another unitary matrix obtained for another one of the plurality of matrices is used as an initial solution for the unitary matrix.
22. The computer-program storage apparatus of claim 20, further comprising:
- code for selecting the plurality of matrices in sequential order for decomposition.
23. The computer-program storage apparatus of claim 20, wherein the plurality of transmission spans correspond to a plurality of frequency subbands in a multi-carrier communication system.
24. The computer-program storage apparatus of claim 20, wherein the plurality of transmission spans correspond to a plurality of time intervals.
25. The computer-program storage apparatus of claim 20, wherein the plurality of matrices of complex values are a plurality of channel response matrices for a plurality of frequency subbands.
Type: Application
Filed: Mar 9, 2010
Publication Date: Jul 1, 2010
Applicant: QUALCOMM INCORPORATED (San Diego, CA)
Inventors: Steven J. Howard (San Diego, CA), John W. Ketchum (San Diego, CA), Mark S. Wallace (San Diego, CA), Jay Rodney Walton (San Diego, CA)
Application Number: 12/720,017
International Classification: G06F 17/16 (20060101); G06F 7/487 (20060101);