1
Convergence of Linear Interference Cancellation Multiuser Receivers Alex Grant and Christian Schlegel Abstract— We consider the convergence in norm of several iterative implementations of linear multiuser receivers, under the assumption of long random spreading sequences. We find that asymptotically, linear parallel interference cancellation diverges for systems loads of greater than about 17%. Using known results from the theory of iterative solutions for linear systems we derive optimal or near optimal relaxation parameters for parallel (first and second order stationary, Chebyshev) and serial cancellation (successive relaxation) methods. An analytic comparison of the asymptotic convergence factor for the various methods is given. Simulations are used to verify results for finite size systems. Keywords— Interference cancellation, iterative methods, multi-stage receivers, multi-user receivers, CDMA, random sequences
I. Introduction Interference cancellation is a simple multiuser detection technique in which estimates (or statistics) of other users’ interference are subtracted from the received signal, in order to improve the estimate for a given user. This process can proceed iteratively (sometimes called a multi-stage receiver), by repeating the subtraction process. Such iterations may be linear or non-linear. Linear methods use linear estimates of the interference, whereas non-linear methods have no such restriction. We mainly restrict our attention to linear schemes, previously considered in [1], [2], [3], [4], [5], [6], [7], [8], [9], [10], [11], [12], [13], [14], [15], [16], [17], [18], [19]. For an introduction to multiuser detection, see [20]. It is known that iterated linear interference cancellation for the detection of linear multiple access channels is equivalent to well known iterative methods for the solution of systems of linear equations [10], [16], [17]. In particular, parallel interference cancellation is equivalent to the method of Jacobi, while serial cancellation corresponds to the Gauss-Seidel method. Our research was motivated by the realization that parallel interference cancellation converges to the decorrelator [4], [5], [8], [9]. Related to these receivers are the approximate decorrelators [1], [15], [20], based on the viewpoint of series expansions for the linear filters. Given the mathematical literature devoted to the iterative solution of linear systems (see [21], [22], [23], [24], for textbook treatments), it makes sense to apply those results to the design of multiuser detectors, rather than try to derive known results from engineering applications. In the recent engineering literature there has been a numA. Grant is with the Institute for Telecommunications Research, University of South Australia C. Schlegel is with the Deptartment of Electrical Engineering University of Utah
ber of papers on iterative detection and decoding. While most of these detectors use some form of non-linear cancellation, usually in the form of a hard-decision device placed in the iteration loop (or between decoder stages), their signal flow structure is identical to the well-known iterative linear algorithms described in this paper. The original multistage detector [25], [26] corresponds to the Jacobi iteration while the fine tuned successive interference canceler discussed by Divsalar et. al. [27], [28], [29] is in fact a minor modification of the successive relaxation method discussed in Section III-E. Iterative methods have also been successfully applied to coded multiuser systems [30], [31], [32]. We restrict our scope to detection of uncoded data. We analyze the convergence properties of several linear iterative implementations of both the decorrelator and the minimum mean squared error filter. Convergence is measured with respect to a given norm (e.g. Euclidean). In particular, we consider the method of Jacobi and improvements thereon, corresponding to weighted parallel cancellation (such as the Chebyshev method). We also consider the Gauss-Seidel, and more generally, the successive relaxation methods corresponding to (weighted) serial cancellation. Finally we also present results concerning the conjugate gradient method, which is a parameter free iteration. For the case of random spreading sequences use the theory of large random matrices to give asymptotically (for large systems) optimal weighting parameters that are easily calculated knowing only the number of users and the length of the spreading sequences. We also consider the convergence speeds of the various methods (with respect to an arbitrary given norm). Related work may be found in [33], where the authors use random matrix results to find the spectral efficiency of various iterative detectors. In Section II we introduce our discrete-time, synchronous, linear model and describe the optimal receiver followed by the decorrelator and minimum mean squared error filters. Section III is the main part of the paper, where we consider the various iterative implementations of the linear receivers. In Section IV we use computer simulations to suggest that our results are also valid for small systems. We conclude in Section V with a discussion and comparison of the various implementations. In the Appendix we present those results from random matrix theory required in our proofs. We shall use the following notations. The vector x ∈ Cn is an column vector with complex entries xi , i = 1, 2, . . . , n. Likewise A ∈ Cm×n is a matrix with complex entries Aij , i = 1, . . . , m, j = 1, . . . , n. The superscripts T , ∗ and −1 denote respectively transposition, Hermitian adjoint and matrix inverse. In is the n × n identity matrix. Denote by
2
S (R) the spectrum [34, p. 35] of R. The spectral radius of a matrix A is ρ (A) = max{|λ| , λ ∈ S (A)}. We shall also refer to λmin = min S (A) and λmax = max S (A). The operator E denotes expectation. II. Linear Multiple Access and Linear Detectors We assume a linear multiple access system (e.g. CodeDivision Multiple Access) where K users access the channel each using a signature signal of duration N chips. Transmission is symbol-synchronous. Such a system can be described by the linear relation [20] r = AWd + n, where A ∈ CN ×K has unit energy columns which are the discrete signature signals of the K users, W = diag(w1 , . . . , wK ) is a diagonal matrix of the users’ amplitudes, d ∈ {−1, +1}K has elements di , the transmitted binary symbol for user i, and n ∈ CN is a sampled circularly symmetric complex noise vector with covariance matrix E [nn∗ ] = σ 2 IN . We shall be particularly interested in the case where the signature sequences A are chosen randomly (in a fashion known to the receiver) for each symbol interval. ˆ which The optimal detector outputs the hypothesis d maximizes the conditional probability distribution p(r|d). In additive Gaussian noise this is equivalent to minimizing the squared Euclidean distance ˆ = arg d
min
d∈{−1,+1}K
kr − AWdk2 ,
min
d∈{−1,+1}K
dT WT RWd − 2dT W Re(y),
M∈CK×N
= WRW + σ 2 I ¡
¢−1
WA∗
The calculation of either the LMMSE filter or the decor¡ ¢ relator requires a matrix inversion of complexity O K 3 , which is a large computational burden if the signature sequences are changing from symbol interval to symbol interval. For the remainder of the paper we shall assume that the entries of A are chosen randomly according to the following rule, modeling the use of spreading sequences with period much longer than the data symbol duration. Definition 1: For each symbol interval, the real or complex spreading sequences A shall be selected randomly as follows. Each element Aij , i = 1, . . . , N , j = 1, . . . , K shall be chosen independently from a given distribution, with mean E[Aij ] = 0, variance E[A2ij ] = 1/N and finite higher moments. Such sequence selection is quite general, and includes the ubiquitous uniform selection of chips from {+1, −1}. Our convergence analyses will draw upon results from random matrix theory, summarized in Appendix V. III. Iterative Solution Methods
in turn equivalent to minimizing the quadratic ˆ = arg d
Minimizing the expected Euclidean distance from the filter output to the transmitted data, the LMMSE filter is given by [20] £ ¤ H = arg min E kHr − dk2
(1)
where R = A∗ A is the K × K normalized correlation matrix between the signature signals of the users and y = A∗ r is the output of a bank of matched filters. Brute force evaluation of (1) requires the computation of 2K terms. In certain special cases however optimal detection is possible with polynomial complexity [35], [36], [37], [38]. Of theoretical and practical interest are the class of linear detectors which form decisions based on the output of ˆ = sgn Hr. We are particularly invarious linear filters d terested in the decorrelator [39] and the linear minimum mean squared error (LMMSE) [40], [41], [42], [43] filters. We shall consider the synchronous case. The asynchronous extension is not difficult [20]. The decorrelator may be derived by relaxing the constraints on the minimization (1). Instead of a combinatorial minimization, we find ˆ = arg min kr − AWdk2 , d d∈RK
ˆ = A† r, where A† is which has the closed form solution d the Moore-Penrose generalized inverse of A [34, pp. 421422]. If A is non-singular1 , A† = A−1 . If A has linearly −1 independent columns, A† = (A∗ A) A∗ . 1 Note that A† exists for any matrix A and can always be expressed in terms of the singular value decomposition of A.
For a given n×n real or complex matrix M and n-vector b, solving a system of linear equations Mx = b,
(2) ¡
¢ 3
for the unknown vector x requires in general O n operations (unless M has some special property that makes it easily invertible). It is possible however to design an iteration that converges to the solution. Letting M = S − T (a splitting of M), we see that Sx = Tx + b has the same solution as (2). Furthermore, for any initial guess x0 , is is easy to show that the iteration Sxk+1 = Txk + b, k = 0, 1, 2, . . . converges in any given norm to the true solution if and only if the spectral radius of the iteration matrix B = S−1 T satisfies ρ (B) < 1. See [21], [23] for details. For a useful iteration we require that S be easily invertible ¡ ¢ (resulting in low complexity iteration steps) and ρ S−1 T be as small as possible (increasing the rate of convergence). For linear multiuser detection, we are interested in the iterative solution of Md = b where for the decorrelator, M = R, b = y and for the LMMSE M = WRW + σ 2 I, b = Wy. Let D, L and U be the diagonal, strictly lower triangular and strictly upper triangular parts of M respectively. We shall consider linear parallel cancellation, based on S = D (Jacobi, stationary methods and Chebyshev) and linear serial cancellation based on S = D + ωL for a real parameter ω ∈ [0, 2] (Gauss-Seidel, successive relaxation). We shall consider convergence criteria for large systems which henceforth means β = K/N is held constant and K → ∞.
3
A. General Convergence Issues We consider convergence in normed spaces [34, Chapter 5]. While this may not be directly related to bit error rate, it is still a useful criteria for the following reasons. Firstly it is tractable and gives insight into the iterative process. Secondly, the Euclidean norm is of interest for Gaussian channels and systems using linear filters as pre-processors for minimum distance decoding. We have already stated that for fast convergence we should minimize ρ (B). In practice it depends on the norm of interest and the number of steps that will be used. Let ˆ − xk , where x ˆ is a solution to (2). It is easy to show ek = x that the error vector at step k is ek = Bk e0 . Thus for any relevant vector norm k · k, and its induced matrix norm [34, Definition 5.6.1] we have kek k = kBk e0 k ≤ kBk k ke0 k, according to the properties of induced matrix norms [34, Theorem 5.6.2]. Thus the error at step k at most kBk k times the error of the initial guess. We therefore make the following definition [23]. Definition 2: Let B = S−1 T be the iteration matrix. For a given matrix norm, k · k, the convergence factor for k steps is given by kBk k. If it is known a-priori how many steps will be used (or can be afforded) for a particular iteration, it is kBk k that should be minimized. However it is known that for any matrix norm, limk→∞ kBk k1/k = ρ (B), i.e. the average convergence factor, per step for k steps approaches the spectral radius for large k [34, Corollary 5.6.14]. Note also 1/k that for B symmetric, kBk k2 = ρ (B), which means that the error of the initial guess in Euclidean space is decreased geometrically with rate ρ (B). In light of these results (especially the latter, pertaining to the Euclidean norm) and for tractability we hereafter concentrate on the spectral radius of the iteration matrix, sometimes called the asymptotic convergence factor. The techniques we describe could be applied to the finite k case, or for the use of any other norm. See [44], [45] and [20, Problem 5.12] for optimization of convergence for some finite k receivers. Not only is the spectral radius of the iteration matrix of interest, but clearly reduction of the norm of the initial error is also important. This amounts to choosing an inital guess that is not too far (at least on average) from the solution. Of course this guess must be obtained with low computational burden (otherwise we could simply choose the exact solution as our first guess). This issue was considered in [14], where they found via simulation that initialisation using the all-zero vector performed worse than an initial guess consisting of the matched filter output y. The authors also considered an even better “sliding window” method for initialisation, which was only possible due to the asynchronous nature of their system. In this paper, we shall use x0 = y as the initial guess throughout, as we wish to concentrate on the properties of the iteration matrices. We consider in greater detail the issue of optimising x0 in a forthcoming paper [46].
B. Jacobi The Jacobi iteration has a particularly simple implementation, since the diagonal matrix S = D is easily invertible. The iteration is as follows: xk+1 = D−1 (y − (M − I)xk ) .
(3)
For the decorrelator, the Jacobi iteration corresponds to parallel interference cancellation (see Figure 1, with τi = 1 and initial guess d0 = y). dk+1 = y − (R − I)dk .
(4)
This is equivalent to implementing the matrix inversion via the series expansion R−1 =
∞ X i=0
(−1)i (R − I)i ,
convergent if ρ (R) < 2. This convergence criterion has been mentioned by several authors [20]. In the following lemma2 , we give a more meaningful convergence criterion for systems using long pseudo-random sequences. Theorem 1: For large systems, the Jacobi iteration (4) √ converges to the decorrelator if and only if K < N ( 2−1)2 , that is for system loads of less than about 17%. Furthermore, the asymptotic convergence factor, ρJ is given by p ρJ = β + 2 β. (5) Proof: The iteration is convergent if and only if ρ (I − R) < 1. Now λ ∈ S (R) implies 1 − λ ∈ S (I − R) [34, Theorem 1.1.6]. Hence the iteration converges if and only if maxλ∈S(R) |1 − λ| < 1. If we fix β = K/N √ , and send K → ∞, ρ (R) converges in probability to ( β − 1)2 , according to Lemma 2. This implies √ convergence to the decorrelator if and only if K < N ( 2−1)2 . (Since |1 − λmin | ≤ |1 − λmax | for β ∈ [0, 1]). [Figure 1 about here.] For the LMMSE filter, S = W2 + σ 2 I, T = W(I − R)W and the iteration is given by ¡ ¢−1 xk+1 = W2 + σ 2 I W [y − (R − I)Wxk ] , (6)
which is just a signal-to-noise ratio weighted version of the parallel cancellation (4). Using similar reasoning as before, we can derive convergence criteria for the Jacobi implementation of the LMMSE filter. Theorem 2: For large systems, the Jacobi iteration (6) converges to the LMMSE detector if µq ¶2 −1 K cos−1 √ . kmax + 1 π 4 β
(17)
For large kmax , an approximate condition is k
λmax > √ . kmax 8 β Proof: Let B = I − τk R be the iteration matrix for stage k. First, we note that, according to [34, Theorem 1.1.6], ρ (B) = max {|1 − τk λmax | , |1 − τk λmin |} . It is easily verified for 0 < β < 1, that τk is positive and monotonically increasing with k. Now if both τk λmax and τk λmin are less than 1, the matrix is stable. For any other case, a necessary condition for instability is τk λmax > 2. Hence Bk is unstable if µ ¶ k − 1/2 λmin cos (18) π ( 2 + 1)2 , the required property
6 For the equal power LMMSE, simply add σ 2 to the right hand side of (16).
7 A sufficient condition only. There are other types of matrices that can also result in convergence.
E. Successive Relaxation
6
|x∗ Lx| < 1/2 is violated. We therefore seek another approach. Note that for 0 < ω < 2, the second term of (20) is positive, since we know that in this range the spectral radius is less than 1. The upper bound results from γ, which is a maximization over all unit length complex vectors x, rather than setting x equal to the eigenvector corresponding to the largest eigenvalue of S−1 T (see [23, p. 242] for details). Instead of choosing γ according to (21), we propose to use an (appropriately maximized) lower bound to this value. This means that we no longer get an upper bound for the spectral radius, however we shall see that the resulting ω compares well with simulation, as shown in Figure 4. ˜ and L ˜ be defined as in the Theorem. Then (since Let M ˜ M is positive definite), ¯ ¯ ˜ = ¯¯x∗ D− 21 (D − L − L∗ ) D− 12 x¯¯ x∗ Mx ¯ ¯ ¯ ˜ ¯ ≤ 1 + 2 ¯x∗ Lx ¯.
¯ ¯ ¯ ˜ ¯2 ˜ − 1)2 . Thus for all unit x ∈ Cn , Hence ¯x∗ Lx ¯ ≥ 14 (x∗ Mx we have ·¯ ¸ ¯ 1 ¯ ∗ ˜ ¯2 1 Lx¯ − γ≥ ¯x ˜ 4 x∗ Mx · ¸ 1 1 1 ∗˜ ≥ (x Mx − 1)2 − ˜ 4 4 x∗ Mx 1 ∗˜ 1 = x Mx − . 4 2
˜ Hence γ ≥ λmax (M)/4 − 1/2 and we choose as our relaxation parameter ω 0 as follows ω 0 = arg min 1 − ¡ 1 ω∈[0,2]
2 ω
−
¢ 1 2
−1
˜ λ−1 min (M) +
ω 2 q ˜ ˜ λmin (M)λmax (M) − 1 =2 . ˜ max (M) ˜ −1 λmin (M)λ
˜ λmax (M) 4
1 2
+
1 ω
˜ = R and we can substitute the For the decorrelator, M limiting values for the extremal eigenvalues from Lemma 2 to find 2 0 . (23) = ωDEC 2−β Substitution of ω 0 into the right hand side of (20) results in an asymptotic convergence factor, ρSR of (bearing in mind the approximations used) 1
The conjugate gradient method [22] was considered for multiple-access detection in [14], [17]. This method is parameter free, provides fast convergence and has a finite termination property. The latter property means that in the absence of numerical errors, the method finds the exact solution of a K × K system after K iterations. For a system matrix M with condition number κ = λmax /λmin , the error at iteration k with respect to the weighted norm kekM = kM1/2 ek2 obeys k
(κ − 1) kek kM ≤ 2 √ ke0 kM √ 2k 2k ( κ + 1) + ( κ − 1)
(25)
Taking √ k → ∞,√we obtain the asymptotic convergence factor ( κ − 1)/( κ + 1). Considering the decorrelator and large systems this √ results in an asymptotic convergence factor ρCG = β, which is the same as the second order stationary method. This result however is misleading. In practice more accurate approximations than predicted by (25) are obtained. IV. Simulation Results The analysis presented in this paper has been for very large systems, K → ∞. We now present simulation results that indicate the validity of the results for more realistically sized systems, (bearing in mind that if the system is small enough, we can always implement direct inversion). Figure 3 shows simulated bit error rates for the decorrelator and for 5 steps of the Jacobi method. The number of users was kept fixed, K = 8, equal powers were assigned to the users, corresponding to a signal to noise ratio of Eb /N0 = 7dB. [Figure 3 about here.]
−
(22)
ρSR = β 4 .
F. Conjugate Gradient
(24)
Hence the asymptotic convergence factor for successive relaxation is always worse than that for the second order stationary method. In the case of the LMMSE filter with unit power users, ˜ = (R + σ 2 I)/(1 + σ 2 ), and ω 0 for large systems may be M ˜ = (λmin + found easily from Equation (22), with λmin (M) ˜ = (λmax + σ 2 )/(1 + σ 2 ). σ 2 )/(1 + σ 2 ) and λmax (M)
From Figure 3, we see that the performance of the Jacobi method starts to diverge from the decorrelator for β > .15, agreeing well with the value of .17 predicted by theory. The curve labeled ‘Hard’ is the result of performing the Jacobi iteration, taking however a hard decision at each stage (i.e. a non-linear iteration). It is interesting to note that although the hard canceler works better than Jacobi, it still has similar convergence problems (not shared by the other methods we have discussed). Note that the hard-decision cancellation method offers only slight improvement over the decorrelator, and only for very lightly loaded systems. This result should be compared to the non-linear cancellation literature, such as [25], [28]. In fact, the iteration given in [25] is identical to Figure 2 with a hard-decision device inserted before the weighing coefficients Rij . Figure 4 compares the “optimal” parameters for successive relaxation, obtained by simulation and by analysis, Equation (23). The simulation was performed by generating 200 sets of unit energy, length N = 32 binary random sequences for K = 1 through K = 32 users, and finding the ω which minimized the spectral radius of the corresponding iteration matrix. The resulting ω were averaged for each K, N pair.
7
[Figure 4 about here.] From this figure, we see that the analysis agrees fairly well with the simulation, (especially for small β), bearing in mind that the analysis was not for the expected value of ω 0 , and that it is only an approximation. In any case, as the next figure shows, the choice of ω for successive relaxation may perhaps be better approached from an entirely different perspective. We remark that for low β (the region in which linear receivers work well), we predict ω 0 close to 1. It may therefore be desirable to use the Gauss-Seidel method, which would avoid some multiplications. Figure 5 shows the simulated bit error rate versus iteration for conjugate gradient (CG), successive relaxation (SOR), Gauss-Seidel (GS), Chebyshev, first (S1) and second order (S2) stationary methods. The simulation parameters are K = 8, N = 16, Eb /N0 = 10dB (all users transmit with equal powers). The SOR method uses ω 0 as the relaxation parameter. Jacobi is not shown, as β = 1/2 is outside its convergence region. The exact decorrelator and LMMSE performances are included for comparison. A maximum of eight iterations is shown. This is because with eight users, direct inversion would be more efficient than performing nine or more iterations. [Figure 5 about here.] In terms of BER, the conjugate gradient method achieves the decorrelator performance first, requiring only three iterations. Note that the convergence of this method is not monotone, and for iterations 3-6 conjugate gradient obtains better performance than the decorrelator. We also see that although the optimized SOR converges slightly faster than Gauss-Seidel, that the latter method out-performs SOR. Once again this is because convergence is from below (this was also observed in [14]). Of the parallel methods, the Chebyshev method performs best, both in terms of BER, and in terms of providing a smooth convergence (once again from below). It is interesting to note that the Chebyshev method actually achieves the best BER of any method (approaching the LMMSE performance), due to its overshoot of the decorrelator performance. Similar phenomena have been observed by other researchers, and have been exploited in [44]. The stationary methods display a “ping-pong” effect, which has been previously reported by others. It is clear that further research is required in order to optimize convergence, or to minimize the BER of such methods. Figure 6 shows the BER performance of the iterative implementations of the LMMSE filter. The iterative methods are labeled as for Figure 5. Once again, the simulation parameters are K = 8, N = 16, Eb /N0 = 10dB. From the figure, we see that the conjugate gradient method achieves the LMMSE performance after only 5 iterations. Successive relaxation takes an extra iteration and Gauss-Seidel requires 7 iterations. Note however that for iteration 1 and 2, Gauss-Seidel outperforms SOR. Of the parallel methods, Chebyshev again outperforms the stationary methods as expected. We now see that whereas for the decorrelator, the slow convergence of the
Chebyshev method was to its advantage in obtaining good performance, it now converges from above, making it less attractive. [Figure 6 about here.] In order to investigate the convergence properties for a more lightly loaded system, Figure 7 shows the simulated bit error rate performance versus iteration for a system with K = 8 and N = 32. The SNR is once again 10 dB. Only the results for implementation of the LMMSE are shown, since at this load, the performance of the decorrelator and LMMSE are not that different. [Figure 7 about here.] The stationary methods are still slow to converge. We now see that the conjugate gradient, successive relaxation and Gauss-Seidel all offer similar performance. This is in contrast to remarks made in [14]. We further investigated this phenomenon with larger dimension simulations (32 users), and we observed similar characteristics. In practice, there are other issues in selection of an implementation method. Although conjugate gradient is in general fastest, it is slightly more complex that either parallel or serial cancellation (it requires the computation of an additional three inner products). The parallel and serial cancellation methods also have the advantage that they can be pipelined, to reduce latency. V. Discussion Using the theory of large random matrices, we have presented convergence analyses of various iterative implementations for the decorrelator and LMMSE. Let us now compare these methods, based upon the asymptotic convergence factor, which may be of more interest for coded systems which use these receivers as a first stage (we have already seen in Section IV that uncoded BER and convergence factor are only loosely related). Figure 8, shows the asymptotic convergence factor for the Jacobi (5), first (13) and second (15) order stationary and successive relaxation (24) methods. For the latter three methods, optimal or near optimal choice of parameters has been used. [Figure 8 about here.] Starting with small spreading factors, we see that Jacobi, the first order stationary method and successive relaxation have similar performance, whereas the second order stationary method (and asymptotically, the first order Chebyshev method) outperforms all, even for highly loaded systems. Notable is the cut-off at β ≈ 0.17 for the Jacobi method, a problem that does not affect the other methods. Note also how these methods approach 1 for highly loaded β → 1 systems. Because the first order stationary method is only a simple modification of the Jacobi method, but avoids the convergence problems, it may be desirable for low spreading factors. Even so, its convergence factor increases quite quickly with β, and if better performance is desired, the second
8
order method would be used, resulting in a slight, but perhaps justified increase in complexity. Lying in between the first and second order methods is successive relaxation. Table I summarizes the main results of this paper. Further research is needed into the convergence properties of non-linear cancellation schemes. Results in this direction can be found in [47]. Extensions of our results to spacetime processing may be found in [49]. [Table 1 about here.] Appendix We are interested in the asymptotic behavior of the eigenvalues of R = A∗ A, where A is chosen according to Definition 1. The following lemma shows that the extremal eigenvalues of R converge to specific values. Lemma 2 (Extremal eigenvalues of R) Let λmin = min S (R) and λmax = max S (R) be the smallest and largest eigenvalues respectively of the complex positive semi-definite matrix R = A∗ A, where A is chosen randomly according to Definition 1. Then, holding β = K/N < 1 constant, and sending K → ∞, λmin → λmax →
³p
³p
β−1 β+1
´2 ´2
of WRW using the method described in [60], [59]. Suppose that for large systems, the spectral distribution of W converges to a fixed distribution H(t), and let Z 1 tdH(t) z(m) = − + β . (26) m 1 + tm (according to Equation (1.2) in [59]). The support of the limiting spectral distribution of WRW is given by the range of values of (26) where z(m) is increasing. Note however that we do not imply that the extreme eigenvalues converge in probability to the minimum and maximum of this support (although we could use it as a not altogether unreasonable assumption). Furthermore, it is in general difficult to find the support. For example, a power distribution with p point masses results in the need to solve a p + 1-th order polynomial. In the more natural case where the power distribution is unknown a-priori, it is probably an unreasonable computational burden to exactly calculate the limiting values of the eigenvalues. Simple bounds may therefore be useful. For example, both R and W are Hermitian, and hence Weyl’s inequality [34, p. 181] may be applied. References
,
[1]
[2]
almost surely. The proof for this lemma can be found in [50] (smallest eigenvalue) and [51] for the largest eigenvalue. In fact the distribution of the eigenvalues tends to a deterministic limit [52], [53], [50], [54], [55], a phenomenon exploited by several authors to prove asymptotic results concerning the capacity and spectral efficiency of the randomsequence CDMA channel, both for optimal joint decoding [56] and using the decorrelator and LMMSE [57], [58]. A stronger result than Lemma 2 is in fact true, namely that in the limit, with probability 1, there are no eigenvalues outside the support of the limiting spectral distribution [59]. In order to derive results for the LMMSE filter, we shall have occasion to find the extremal eigenvalues of WRW + σ 2 I. Using [34, Theorem 1.1.6] we need only find the extreme eigenvalues of WRW (or alternatively W 2 R, by the obvious similarity transform). In the equal power case (where without loss of generality we can assume unit power), the limiting values of the eigenvalues are found easily, and under the same limiting conditions as Lemma 2 are given by ³p ´2 ¡ ¢ β − 1 + σ2 min S R + σ 2 I → ´2 ³p ¡ ¢ β + 1 + σ2 . max S R + σ 2 I → In the non-equal power case, there is a little more work to do. For large R, chosen according to Definition 1, we can find the support of the limiting spectral distribution
[3] [4]
[5] [6] [7]
[8] [9]
[10]
[11] [12]
[13]
R. Kohno, M. Hatori, and H. Imai, “Cancellation techniques of co-channel interference in asynchronous spread spectrum multiple access systems,” Electronics and Communications, vol. 66-A, pp. 20–29, 1983. P. Patel and J. Holtzman, “Analysis of a simple successive interference cancellation scheme in a DS/CDMA system,” IEEE J. Select. Areas Commun., vol. 12, no. 5, pp. 796–807, June 1994. A. Kaul and B. Woerner, “Analytic limits on the performance of adaptive multistage interference cancellation,” IEE Electron. Lett., vol. 30, pp. 2093–2094, Dec. 1994. V. Ghazi-Moghadam, L. B. Nelson, and M. Kaveh, “Parallel interference cancellation for CDMA systems,” in 33rd Annual Allterton Conference on Communications, Control and Computing, Monticello, IL, 1995, pp. 216–224. K. Jamal and E. Dahlman, “Multi-stage interference cancellation for DS-CDMA,” in IEEE Veh. Technol. Conf., Atlanta, Georgia, Apr. 1996, pp. 671–675. S. Moshavi, E. G. Kanterakis, and D. L. Schilling, “Multistage linear receivers for DS-CDMA systems,” Int. J. Wireless Inform. Networks, vol. 3, pp. 1–17, 1996. H. Elders-Boll, M. Herper, and A. Busboom, “Adaptive receivers for mobile DS-CDMA communication systems,” in IEEE Vehicular Technology Conference, Phoenix, AZ, May 1997, vol. 3, pp. 2128–2132. R. M. Buehrer and B. D. Woerner, “The asymptotic multiuser efficiency of m-stage interference cancellation receivers,” in 8th IEEE PIMRC, Helsinki, Finland, Sept. 1997, pp. 570–574. L. K. Rasmussen, A.-L. Johansson, and T. J. Lim, “One-shot filtering equivalence for linear successive interference cancellation in CDMA,” in IEEE Veh. Technol. Conf., Phoenix, Arizona, May 1997, pp. 2163–2167. H. Elders-Boll, A. Busboom, and H. D. Schotten, “Spreading sequences for zero-forcing DS-CDMA multiuser detectors,” in 8th IEEE Int. Symp. PIMRC ’97, Helsinki, Finland, Sept. 1997, pp. 53–57. D. Guo, “Linear parallel interference cancellation in CDMA,” M.S. thesis, National University of Singapore, Singapore, 1998. A.-L. Johansson, Successive interference cancellation in DSCDMA systems, Ph.D. thesis, Chalmers University of Technology, Department of Signals and Systems, G¨ oteborg, Sweden, Oct. 1998. H. Elders-Boll, H. D. Schotten, and A. Busboom, “The influence of spreading sequences on iterative multiuser detectors,” in IEEE International Symposium on Spread Spectrum Techniques and Applications, Sun City, South Africa, Sept. 1998, vol. 1, pp. 174–178.
9
[14] M. J. Juntti, B. Aazhang, and J. O. Lilleberg, “Iterative iplementations of linear multiuser detection for asynchronous CDMA systems,” IEEE Trans. Commun., vol. 46, no. 4, pp. 503–508, April 1998. [15] N. B. Mandayam and S. Verd´ u, “Analysis of an approximate decorrelating detector,” Wireless Personal Communications, vol. 6, pp. 97–111, 1998. [16] H. Elders-Boll, A. Busboom, and H.-D. Schotten, “Implementation of linear multiuser detectors for asynchronous CDMA systems by linear interference cancellation algorithms,” in 48th IEEE Vehicular Technology Conference, Ottawa, Canada, May 1998, pp. 1849–1853. [17] H. Elders-Boll, H.-D. Schotten, and Axel Busboom, “Efficient implementation of linear multiuser detectors for asynchronous CDMA systems by linear interference cancellation,” European Transactions on Telecommunications, vol. 9, no. 5, pp. 427–438, Sept-Oct 1998, Special Issue on Code Division Multiple Access Techniques for Wireless Communication Systems. [18] L. K. Rasmussen, T. J. Lim, and A.-L. Johansson, “A matrixalgebraic approach to successive interference cancellation in CDMA,” IEEE Trans. Commun., vol. 48, no. 1, pp. 145–151, Jan. 2000. [19] D. Guo, L. K. Rasmussen, S. Sun, , and T. J. Lim, “A matrixalgebraic approach to linear parallel interference cancellation in CDMA,” IEEE Trans. Commun., vol. 41, no. 1, pp. 152–161, Jan. 2000. [20] S. Verd´ u, Multiuser Detection, Cambridge University Press, Cambridge, 1998. [21] G. Strang, Linear Algebra and its Applications, Harcourt Brace and Company, 3 edition, 1988. [22] G. H. Golub and C. F. van Loan, Matrix Computations, Johns Hopkins University Press, Baltimore, Maryland, 2 edition, 1989. [23] O. Axelsson, Iterative Solution Methods, Cambridge University Press, 1994. [24] R. Barrett, M. Berry, T. F. Chan, J. Demmel, J. Donato, J. Dongarra, V. Eijkhout, R. Pozo, C. Romine, and H. Van der Vorst, Templates for the Solution of Linear Systems: Building Blocks for Iterative Methods, 2nd Edition, SIAM, Philadelphia, PA, 1994. [25] M.K. Varanasi and B. Aazhang, “Multistage detection in asynchronous code-division multiple-access communication,” IEEE Trans. Commun., vol. 38, no. 4, pp. 509–519, April. 1990. [26] M.K. Varanasi and B. Aazhang, “Near optimum detection in synchronous code-division multiple-access systems,” IEEE Trans. Commun., vol. 39, no. 5, pp. 725–736, May. 1991. [27] D. Divsalar and M. K. Simon, “Improved parallel interference cancellation for CDMA,” JPL publication, , no. 95-21, Oct. 1995. [28] D. Divsalar, M. K. Simon, and D. Raphaeli, “Improved parallel interference cancellation for CDMA,” IEEE Trans. Commun., vol. 46, no. 2, pp. 258–268, Feb. 1998. [29] R. M. Buehrer and S. P. Nicoloso, “Comments on partial parallel interference cancellation for CDMA,” IEEE Trans. Commun., vol. 47, no. 5, pp. 658–661, May 1999. [30] T. R. Giallorenzi and S. G. Wilson, “Suboptimum multiuser receivers for convolutionally coded asynchronous DS-CDMA systems,” IEEE Trans. Commun., vol. 44, no. 9, pp. 1183–1196, Sept. 1996. [31] M. Moher, “An iterative multiuser decoder for near-capacity communications,” IEEE Trans. Commun., vol. 46, no. 7, pp. 870–880, July 1998. [32] P. D. Alexander, A. J. Grant, and M. C. Reed, “Iterative detection on code-division multiple-access with error control coding,” European Transactions on Telecommunications, vol. 9, no. 5, pp. 419–426, Sept.-Oct. 1998. [33] R. R. M¨ uller and S. Verd´ u, “Spectral efficiency of low-complexity multiuser detectors,” in IEEE Int. Symp. Inform. Theory, Sorrento, Italy, June 2000, p. 439. [34] R. A. Horn and C. R. Johnson, Matrix Analysis, Cambridge University Press, 1990. [35] R. Learned, A. Willsky, and D. Boroson, “Low complexity optimal joint detection for oversaturated multiple access communications,” IEEE Trans. Sig. Proc., vol. 45, no. 1, pp. 113–123, Jan. 1997. [36] S. Ulukus and R. Yates, “Optimum mutliuser detection is tractable for synchronous cdma using m-sequences,” IEEE Commun. Lett., vol. 2, pp. 89–91, 1998. [37] C. Sankaran and A. Ephremides, “Solving a class of opti-
[38] [39] [40] [41] [42] [43]
[44]
[45] [46] [47]
[48] [49] [50] [51] [52] [53] [54] [55] [56] [57] [58] [59]
[60]
mum multiuser detection problems with polynomial complexity,” IEEE Trans. Inform. Theory, vol. 44, pp. 1958–1961, 1998. C. Schlegel and A. Grant, “Polynomial complexity optimal detection of certain multiple access systems,” To appear, Sep. 2000. R. Lupas and S. Verd´ u, “Linear multiuser detectors for synchronous code–division multiple–access channels,” IEEE Trans Inform. Theory, vol. 35, no. 1, pp. 123–136, Jan. 1989. Z. Xie, R. T. Short, and C. K. Rushforth, “A family of suboptimum detectors for coherent multiuser communications,” IEEE. J. Select. Areas Commun., vol. 8, no. 4, pp. 683–690, May 1990. P. B. Rapajic and B. S. Vucetic, “Adaptive receiver structures for asynchronous CDMA systems,” IEEE. J. Select. Areas Commun., vol. 12, no. 4, pp. 685–697, May 1994. U. Madhow and M. L. Honig, “MMSE interference suppression for direct–sequence spread–spectrum CDMA,” IEEE Trans. Commun., vol. 42, no. 12, pp. 3178–3188, Dec. 1994. S. L. Miller, “An adaptive direct-sequence code-division multiple-access receiver for multiuser interference rejection,” IEEE Trans. Commun, vol. 43, no. 2/3/4, pp. 1746–1755, February/March/April 1994. D. Guo, L. K. Rasmussen, S. Sumei, T. J. Lim, and C. Cheah, “MMSE-based linear parallel interference cancellation in CDMA,” in Proc. IEEE Int. Symp. Spread Spectrum Techniques and Appl., Sun City, South Africa, Sept. 1998, pp. 917–921. L. K. Rasmussen, Y. Ma D. Guo, , and T. J. Lim, “Aspects on linear parallel interference cancellation in CDMA,” in IEEE Int. Symp. Inform. Theory, Cambridge, USA, Aug 1998, p. 37. M. Navarro, A. Grant, and S. Perreau, “Iterative detection for Rayleigh space-time channels,” Submitted to IEEE. Trans. Commun. D. R. Brown, M. Motani, V. V. Veeravalli, H. V. Poor, and Jr C. R. Johnson, “On the performance of linear parallel interference cancellation,” IEEE. Trans. Inform. Theory, 1999, Submitted. D. M. Young, Iterative Solution of Large Systems, Academic Press, Orlando, Florida, 1971. M. Navarro and A. Grant, “Iterative detection for Rayleigh space-time channels,” in IEEE Veh. Technol. Conf., Boston, 24-28 Sept. 2000. Z. D. Bai and Y. Q. Yin, “Limit of the smallest eigenvalue of a large dimensional sample covariance matrix,” Ann. Probab., vol. 21, pp. 1275–1294, 1993. Z. D. Bai, J. W. Silverstein, and Y. Q. Yin, “A note on the largest eigenvalue of a large dimensional sample covariance matrix,” J. Multivariate Anal., vol. 26, pp. 166–168, 1988. J. W. Silverstein, “Eigenvalues and eigenvectors of large dimensional sample covariance matrices,” Contemporary Mathematics, vol. 50, pp. 153–159, 1986. Y. Q. Yin and Z. D. Bai, “Spectra for large-dimensional random matrices,” Contemporary Mathematics, vol. 50, pp. 161–167, 1986. J. W. Silverstein and Z. D. Bai, “On the emperical distribution of eigenvalues of a class of large dimensional random matrices,” J. Multivariate Anal., vol. 54, no. 2, pp. 175–192, Aug. 1995. J. W. Silverstein, “Strong convergence of the empirical distribution of eigenvalues of large dimensional random matrices,” J. Multivariate Anal., vol. 55, no. 2, pp. 331–339, Aug. 1995. A. J. Grant and P. D. Alexander, “Random sequence multisets for synchronous code-division multiple-access channels,” IEEE Trans. Inform. Theory, vol. 44, no. 7, pp. 2832–2836, Nov. 1998. D. Tse and S. Hanly, “Linear multiuser receivers: Effective interference, effective bandwidth and user capacity,” IEEE Trans. Inform. Theory, vol. 45, no. 2, pp. 641–657, Mar. 1999. S. Verd´ u and S. Shamai, “Spectral efficiency of CDMA with random spreading,” IEEE Trans. Inform. Theory, vol. 45, no. 2, pp. 622–640, Mar. 1999. Z. D. Bai and J. W. Silverstein, “No eigenvalues outside the support of the limiting spectral distribution of large dimensional sample covariance matrices,” Ann. Probab., vol. 26, no. 1, pp. 316–339, 1998. J. W. Silverstein and S.-I. Choi, “Analysis of the limiting spectral distribution of large dimensional random matrices,” J. Multivariate Anal., vol. 54, no. 2, pp. 295–309, 1995.
10
Alex Grant Alex Grant received the B.E. and Ph.D. degrees from the University of South Australia in 1993 and 1996 respectively. In 1997, he was a post-doctoral research fellow at the Laboratory for Signal and Information Processing, Swiss Federal Institute of Technology (ETH), Zurich. Since March 1998 he has been with the Institute for Telecommunication Research, where he is now a Senior Lecturer. Alex’s research interests include information theory and its applications to multiple user and other communications problems. Alex is Leader of the Coding and Information Theory Research Group at ITR and serves as Chariman for the Australian Chapter of the IEEE Information Theory Society.
Christian Schlegel received the Dipl. El. Ing. ETH degree from the Federal Institute of Technology, Zurich, in 1984, and the M.S. and Ph.D. degrees in electrical engineering from the University of Notre Dame, Notre Dame, IN, in 1986 and 1988, respectively. In 1988 he joined the Communications Group at the research center of Asea Brown Boveri, Ltd., in Baden, Switzerland, where he was involved in mobile communications research. He spent the 1991/92 academic year as Visiting Assistant Professor at the University of Hawaii at Manoa, Hawaii, before joining the Digital Communications Group at the University of South Australia in Adelaide, Australia, where he headed the mobile research centre (MCRC). From 1994–1996 he was with the University of Texas at San Antonio, and since 1996 he has been with the University of Utah, Salt Lake City. His interests are in the area of digital communications, coded modulation, mobile radio systems, multiple access communications, space-time coding for high-capacity wireless systems, as well as in the implementation of complex communications systems and the design of analog VLSI decoders for digital systems. He is the author of the research monograph “Trellis Coding” published by IEEE Press in 1997, and is currently working on his new book entitled “Coordinated Multiple User Communications”, coauthored with Dr. Alex Grant. Dr. Schlegel received an NSF 1997 Career Award in support of his research in multiuser communications. Dr. Schlegel is a senior member of the IEEE Information Theory and Communication Societies. In the past he has organized conference sessions and served as consultant on digital communications and error control coding projects on numerous occasions. He is an associate editor for coding theory and techniques for the IEEE Transactions on Communications, and the technical co-chair of the IEEE Information Theory Workshop 2001 to be held in Cairns, Australia. He has also served on several technical progam committees, most recently for the IEEE VTC’99 conference in Houston, TX, the IEEE VTC’2000 conference held in Tokyo, Japan, and Globecom 2001 to be held in San Antonio, TX.
11
List of Figures 1 2 3 4 5 6 7 8
First order iterative implementations of the decorrelator, including: Jacobi, τ k = 1, First order stationary, τk = 1/(1 + β) and Chebyshev, Equation (16). . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Successive relaxation implementation of the decorrelator. (Gauss-Seidel, ω = 1). . . . . . . . . . . . . . . . Bit error rate versus spreading factor, β using the Jacobi implementation of the decorrelator. 8 users, 5 iterations, Equal powers, Eb /N0 = 7dB. Also shown is parallel cancellation of hard decisions (labeled Hard). Comparison of optimal parameters for the successive relaxation method, obtained by simulation and analysis. Iterative implementations of the decorrelator. Simulated BER convergence properties. K = 8, N = 16, Eb /N0 = 10dB. Simulated decorrelator and LMMSE performance also shown. . . . . . . . . . . . . . . . . Iterative implementations of the LMMSE. Simulated BER convergence properties. K = 8, N = 16, E b /N0 = 10dB. Simulated LMMSE performance also shown. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Iterative implementations of the LMMSE. Simulated BER convergence properties. K = 8, N = 32, E b /N0 = 10dB. Simulated LMMSE performance also shown. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Asymptotic convergence factors for various implementations of the decorrelator. . . . . . . . . . . . . . . .
12 13 14 15 16 17 18 19
Figures
12
(0)
-
Matched Filter 1
-
Matched Filter i
d1
(0)
di
HH - τ0 R1jH H © ©© ©
H H - τ0 RijH H © ©© ©
r
-
-
Matched Filter j
(0)
Matched Filter K
H
H H r - τ0 H © © © ©
yj
dK
H H H - τ0 RKjH © © © © Stage 0
(1)
d1
(1)
di
C
C C
C
HH - τ1 R1jH H © ©© ©
H H - τ1 RijH H © ©© ©
C
C − C − WC ? ²¯ - + ±° 6−
yj
(1) dj
H
H H - τ0 H © © © ©
r (1)
dK
C C
C
C
C C
C − WC ? ²¯ - + ±° 6− −
(2)
dj
yj
H H H - τ1 RKjH © © © © Stage 1
Fig. 1. First order iterative implementations of the decorrelator, including: Jacobi, τ k = 1, First order stationary, τk = 1/(1 + β) and Chebyshev, Equation (16).
Figures
13
©© © H R1j ¾ HH
(0)
Matched Filter 1
d1
-
(0)
Matched Filter i
di
-
r
£
Matched Filter j
(0)
Matched Filter K
dK
£
£ H £ r - 1 − ωH - +n £ ©© © £ − ? 6 °£ − HH yj - +n - ω H © ©© − 6
(0)
dj
£
£
©© © H Rij ¾ HH £
(1) d1
©© © H R1j ¾ HH
d1
(1)
©© © H Rij ¾ HH £
di
di
£
H
H HH - RKj ©© ©
£
yj
(1)
(2)
£ H £ - 1 − ωH - +n - d(2) j £ ©© © £− H − ? 6 °£ - +n - ωHH © ©© − 6 H
dK
Stage 1
£
£
(2)
H HH - RKj ©© © Stage 2
Fig. 2. Successive relaxation implementation of the decorrelator. (Gauss-Seidel, ω = 1).
Figures
14
0
10
√ β = ( 2 − 1)2 Jacobi −1
10
BER
Hard
−2
10
Decorrelator
−3
10
−4
10
0
0.2
0.4
0.6
0.8
1
β = K/N Fig. 3. Bit error rate versus spreading factor, β using the Jacobi implementation of the decorrelator. 8 users, 5 iterations, Equal powers, Eb /N0 = 7dB. Also shown is parallel cancellation of hard decisions (labeled Hard).
Figures
15
2
1.9
1.8
1.7
1.6
ω
ω0
1.5
Simulation 1.4
1.3
1.2
1.1
1
0
0.1
0.2
0.3
0.4
0.5
0.6
0.7
0.8
0.9
1
β = K/N Fig. 4. Comparison of optimal parameters for the successive relaxation method, obtained by simulation and analysis.
Figures
16
−1
10
BER
CG SOR GS Chebyshev S1 S2
−2
10
Decorrelator LMMSE −3
10
0
1
2
3
4
5
6
7
8
Number of iterations Fig. 5. Iterative implementations of the decorrelator. Simulated BER convergence properties. K = 8, N = 16, E b /N0 = 10dB. Simulated decorrelator and LMMSE performance also shown.
Figures
17
−1
10
BER
CG SOR GS Chebyshev S1 S2
−2
10
LMMSE −3
10
0
1
2
3
4
5
6
7
8
Number of iterations Fig. 6. Iterative implementations of the LMMSE. Simulated BER convergence properties. K = 8, N = 16, E b /N0 = 10dB. Simulated LMMSE performance also shown.
Figures
18
−1
10
CG SOR GS Chebyshev S1 S2
−2
BER
10
−3
10
−4
10
LMMSE −5
10
0
1
2
3
4
5
6
7
8
Number of iterations Fig. 7. Iterative implementations of the LMMSE. Simulated BER convergence properties. K = 8, N = 32, E b /N0 = 10dB. Simulated LMMSE performance also shown.
Figures
19
1 ρS1 ρJ
0.8
Convergence factor
ρSR ρS2
0.6
0.4
0.2
0 0
0.2
0.4
0.6
0.8
1
β = K/N Fig. 8. Asymptotic convergence factors for various implementations of the decorrelator.
Figures
20
List of Tables I
Summary: Iterative implementations of the decorrelator. . . . . . . . . . . . . . . . . . . . . . . . . . . . .
21
Tables
21
Method
Type
Convergence
Optimal parameters
Jacobi Stationary 1 Stationary 2
Parallel Parallel Parallel
β < 0.17 β