Skip to main content

Low complexity interference alignment algorithms for desired signal power maximization problem of MIMO channels

Abstract

In this article, we investigate the interference alignment (IA) solution for a K-user MIMO interference channel. Proper users’ precoders and decoders are designed through a desired signal power maximization model with IA conditions as constraints, which forms a complex matrix optimization problem. We propose two low complexity algorithms, both of which apply the Courant penalty function technique to combine the leakage interference and the desired signal power together as the new objective function. The first proposed algorithm is the modified alternating minimization algorithm (MAMA), where each subproblem has closed-form solution with an eigenvalue decomposition. To further reduce algorithm complexity, we propose a hybrid algorithm which consists of two parts. As the first part, the algorithm iterates with Householder transformation to preserve the orthogonality of precoders and decoders. In each iteration, the matrix optimization problem is considered in a sequence of 2D subspaces, which leads to one dimensional optimization subproblems. From any initial point, this algorithm obtains precoders and decoders with low leakage interference in short time. In the second part, to exploit the advantage of MAMA, it continues to iterate to perfectly align the interference from the output point of the first part. Analysis shows that in one iteration generally both proposed two algorithms have lower computational complexity than the existed maximum signal power (MSP) algorithm, and the hybrid algorithm enjoys lower complexity than MAMA. Simulations reveal that both proposed algorithms achieve similar performances as the MSP algorithm with less executing time, and show better performances than the existed alternating minimization algorithm in terms of sum rate. Besides, from the view of convergence rate, simulation results show that the MAMA enjoys fastest speed with respect to a certain sum rate value, while hybrid algorithm converges fastest to eliminate interference.

Introduction

Interference alignment (IA) technique is recently brought to show that each user can achieve half degree of freedom (DoF) in the K-user interference channel. It jointly optimizes precoding matrices for all transmitters, so that all interferences at one receiver fall into a reduced dimensional subspace. Then by multiplying decoding matrix orthogonal to this subspace, the certain receiver can extract the desired signals without interference. By utilizing the IA techniques, Cadambe and Jafar[1] showed that the achieved sum capacity of the K-user interference channel scales linearly with the number of users, in the high signal-to-noise-ratio (SNR) regime. Generally, the IA solutions are required to satisfy the following conditions simultaneously:

  1. (1)

    all the interferences are eliminated;

  2. (2)

    all the subspaces for desired signals are full rank;

  3. (3)

    precoders and decoders are required to be orthogonal.

The Yetis et al.[2] related the feasibility of the IA conditions in fully connected interference channel to the problem of determining the solvability of a multivariate polynomial system, with arbitrary antenna configurations. The achievable DoFs are also discussed based on this polynomial system, relating to users’ number of antennas and the number of users[3]. The analysis is further extended to that in partially connected channels[46].

From the signal processing point of view, the procedure of IA is to solve precoders and decoders according to the three conditions with a feasible IA system. However, the solution to this feasible problem is still not known in general. There are available closed form solutions only for certain cases, such as 3-user MIMO channel with N antennas each user equips with and N / 2 DoFs each user requires, and K-user channel where each user equips with K − 1 antennas and wishes to achieve 1 DoF. For general cases, the system is turned into an optimization problem minimizing the total leakage interference and preserving the orthogonality of precoders and decoders as constraints, which is denoted as leakage interference minimization (LIM) problem. With the solution, the IA condition 2 can be almost surely satisfied if channels have no special structures[7]. LIM problem is proved to be NP-hard when the number of antennas each user equips with are greater than 3[8]. Thus, iterative algorithms rather than the analytical solutions should be considered. Gomadam et al.[7] exploited the channel reciprocity and proposed an alternating minimization algorithm (AMA) to design precoders and decoders in a distributed way. Then each subproblem is equivalent to an eigenvalue problem with eigenvalue decomposition required. In the AMA, although the leakage interference can be perfectly canceled after convergence, its performance in terms of sum rate is not optimal. In fact, it is pointed out that for general interference channels the constructed LIM problem has a large number of different IA solutions obtained from different initial points, which lead to different achieved sum rate values[9]. The main reason is that the AMA only eliminates the interference in the desired signal space without considering the system sum rate, which results in a suboptimal sum rate achieved with finite signal power.

Gomadam et al.[7] also noticed the disadvantage of AMA, and then proposed the Max signal-to-interference-and-noise-ratio (SINR) algorithm as well. In each iteration, the basic idea of the Max-SINR algorithm is to choose the precoders and decoders stream by stream, with aim to maximize the SINR of each stream instead of minimizing the leakage interference. Due to the relaxation of the IA condition 3, the IA condition 2 does not hold anymore in this algorithm, which means the required DoFs might not be satisfied. This analysis accords with the performances shown in[7], that it achieves higher sum rate than the AMA sin intermediate SNR regime, however suffers from the loss of required DoFs in the high SNR regime. Many other algorithms can also be designed to perform like Max-SINR, however none of them can achieve the optimal DoFs without IA conditions[10]. Further, IA scheme provides the receivers interference-free subspaces, with which receivers completely get rid of complicated cancelation of interference. Thus, it is important to improve IA algorithms in general SNR scenarios.

Prior study in[11] proposes an iterative algorithm using the gradient descent method to solve the new IA model, where the utility function of either sum rate or the desired signal power is maximized with the IA conditions 1 and 3 as constraints. The corresponding MSP algorithm is shown to obtain higher sum rate than the AMA generally regardless of the initial point and higher than Max-SINR in high SNR regime. However, it requires a series of eigenvalue decompositions and compact singular value decompositions (SVDs), which lead to high computational complexity. Besides, the MSP algorithm has much slower convergence rate than the AMA.

Nevertheless, in practical systems, receivers have limited computational complexity, which might be a bottleneck for the complexity of the algorithms. Besides, channel reciprocity requires TDD operations which restrict the executing time of the algorithms. Therefore, the principal question here is how to design proper algorithms to balance the achieved sum rate with computational complexity and computing time. In this article, we aim to propose algorithms to maintain the advantage of the MSP algorithm with faster convergence, lower complexity and less executing time. Two efficient algorithms are proposed to solve the desired signal power maximization IA problem. First we propose a modified alternating minimization algorithm (MAMA) with Courant penalty function technique. Then, to further reduce the algorithm complexity, a new algorithm with Householder transformation (AHT) is proposed, where a two-dimensional subspace method is applied to solve subproblems. It acts to obtain precoders and decoders with low leakage interference in short time, and then the MAMA continues iterating to get perfect IA, which forms the hybrid algorithm.

The remaining article is organized as follows: In Section “Desired signal power maximization interference alignment model”, the desired signal power maximization model is presented. The two algorithms, MAMA and hybrid algorithm, are proposed in Sections “Modified alternating minimization algorithm” and “A hybrid algorithm”, respectively. The computational complexity of the proposed algorithms and MSP algorithm are analyzed in Section “Analysis of computational complexity”. Numerical results and further remarks are shown in Section “Simulations”.

Notation: Lowercase and uppercase boldface represent vectors and matrices, respectively. represents the complex domain. Re(a) means the real part of scalar a. tr(A) and A F are the trace and the Frobenius norm of matrix A, respectively. I d represents the d × d identity matrix.K represents the set of the user indices {1, 2, …, K}.CN(μ,σ2) means the complex Gaussian distribution with mean μ and variance σ2. And we use E X (·) to denote the statistical expectation with the variable X. O(n) means the same order amount of n.

Desired signal power maximization IA model

Consider a K-user interference MIMO channel (M × N, d)K as in Figure1, where each transmitter equips with M antennas, each receiver with N antennas and each user pair wishes to achieve d DoFs. Suppose s k d×1 denotes the transmit signal vector of the k th user with power covariance asE( s k s k H )=(P/d) I d , where P is the total transmit power of each user. For convenience we unify the transmit power of each stream, i.e. P / d = 1. After receiving the signal y k , the k th receiver multiplies decoding matrix to it on the left, which is expressed as:

U k H y k = U k H H kk V k s k desired signal + l k , l K U k H H kl V l s l interference + U k H n k noise ,
(1)

where H kl N×M denotes the channel matrix between the k th transmitter and the l th receiver, and V k M × d , U k N×d represent the precoder and decoder of the k th user, respectively. The three terms of (1) on the right side represent the desired signal, the interference from other users and the noise with distribution ofCN(0, σ k 2 I N ) at the k th receiver, respectively. In this article we assume all the noises have the same covariance, that is σ k 2 = σ 2 ,kK.

Figure 1
figure1

K -user MIMO interference channel.

Following the feasible condition for IA system, IA scheme is defined as[7]:

U k H H kl V l = 0 , k l , k , l K ;
(2)
rank ( U k H H kk V k ) = d , k K ;
(3)
V k H V k = I d , U k H U k = I d , k K.
(4)

The original idea of IA only consists of (2) and (3). Noticing that for any precoding and decoding matrices {V k , U k , kK} that satisfy these two conditions, {V k P k , U k Q k , kK} also satisfy (2) and (3) as long as {P k ,Q k d×d,kK} are all non-singular matrices. This indicates that the solutions of the IA system are not unique and the solution matrices form d-dimensional subspaces. Therefore, we require the columns of V k ,U k ,kK to be the orthogonal basis of the corresponding subspaces, which is the condition (4).

Besides the requirement of (2) and (4), we wish to maximize the desired signal power, in order to achieve sufficiently high sum rate. Suppose all the transmit signals are statistically independent of each other. We can obtain the expected total desired signal power

P S = k K E s ( U k H H kk V k s k 2 2 ) = k K U k H H kk V k F 2 .

Based on the above analysis, we present the desired signal power maximization (PM) model as follows:

max U k , V k k K P S ( U k , V k ) = k K U k H H kk V k F 2
(5a)
s. t. U k H H kl V l = 0 , k l , k , l K ,
(5b)
V k H V k = I d , U k H U k = I d , k K .
(5c)

This model was first brought in[11], in which its performance is compared with the sum rate maximization (SRM) model. The PM model can achieve similar performance as the SRM model, while its related optimization problem is much simpler. Therefore it is a good way to approximate the SRM model by the PM model[11]. Thus in this article we only focus on solving (5) to design proper precoders and decoders.

The sum of squares of the residuals of constraints (5b) is the total leakage interference, which is given by:

P I U k , V k = k K E s l k , l K U k H H kl V l s l 2 = k K l k , l K U k H H kl V l F 2 .
(6)

The essential idea of the AMA given by[7] is to minimize PI alternately for U k and V k . From the formulation of (5), we know that the MSP algorithm in[11] solves it by increasing PS while reducing PI towards zero at the same time. Here we briefly introduce the framework of MSP. In each iteration for each precoder and decoder, an AMA step is first taken to reduce PI. And then the iterative point go along the gradient direction of PS to increase it, whose stepsize gradually shrinks to 0 in order to perfectly align the interference by simply taking AMA steps in the last few iterations. MSP algorithm requires a series of eigenvalue decompositions and SVDs, which bring in high computational complexity and complicated matrix computations. Therefore we wish to propose algorithms for (5) with lower complexity and less complicated computations.

Modified alternating minimization algorithm (MAMA)

In this section, we present our first low complexity algorithm, the MAMA. The main difficulty for solving the highly nonlinear nonconvex optimization problem (5) is to deal with the nonlinear constraint (5b). Noticing that (5a) and the penalty term (6) for (5b) have quite similar expressions, we can combine these two together by the Courant penalty function technique[12] as in the next section.

Courant penalty function technique

Courant penalty function technique is a classic penalty function technique, which avoids dealing with constraints by moving them to the objective function. The basic idea is to replace the constraints with a penalty term scaled by a parameter C. The penalty term is the sum of squares of all the constraint violations. In (5), we apply Courant penalty technique to (5b) and keep the constraints (5c) in order to obtain the following simple constrained problem:

min U k , V k k K P ( U k , V k , C ) = C P I ( U k , V k ) P S ( U k , V k )
(7a)
s. t. V k H V k = I d , U k H U k = I d , k K ,
(7b)

where PI is defined by (6).

It is well known that if (5) is feasible, the solution of (7) converges to that of (5) as C approaches infinity[13]. Thus the solution of (5) can be approximated by that of (7). As pointed out in[13], the approximated error between the solutions of (7) and (5) is of the levelO( 1 C ), which concerns about the choice of the penalty parameter C. Therefore, theoretically, (7) with larger C would get better approximation of (5). However, in real computations, large C may lead to ill-conditioned objective function and consequently bring in numerical calculation difficulties. To avoid such difficulties caused by unnecessarily large C in the first few iterations, we initially set C as a small positive number such as C = 1. If sufficient reduction in PI is not achieved, we increase C to σ0C, where σ0 > 1 is a constant. The solution of (7), (U k (C), V k (C), kK) provides a good initial point for solving (7) again when C is replaced by σ0C. In this way we force PI to reduce towards 0 eventually and the optimal solution of (5) can be obtained.

The overall algorithm

Via Courant penalty function technique our main task is now how to solve (7), which is quite similar to the LIM problem in which PI is to be minimized. We can rewrite (7) as:

min U k , V k k K P ( U k , V k , C )
(8a)
= k K tr U k H C l k , l K H kl V l V l H H kl H H kk V k V k H H kk H U k = k K tr V k H C l k , l K H lk H U l U l H H lk H kk H U k U k H H kk V k s. t. V k H V k = I d , U k H U k = I d , k K.
(8b)

The only difference between the LIM problem and (8) is the objective function, as LIM minimizes the total leakage interference PI subject to (8b). Thus, we can borrow the idea of the AMA in[7], to iterate V k , kK and U k , kK alternatively in each iteration. Fixing all V k , kKin (8), U k , kK become independent of each other. In this case, (8) turns into K independent subproblems with formulas as:

min U k tr U k H C l k , l K H kl V l V l H H kl H H kk V k V k H H kk H U k s. t. U k H U k = I d .
(9)

We can obtain the closed-form solution of (9) as U k = ν min d ( J k ),kK, which means the columns of U k are eigenvectors corresponding to the d smallest eigenvalues of

J k = C l k H kl V l V l H H kl H H kk V k V k H H kk H .

Similar solutions of V k can be achieved when fixing all U k , kK. Of course, we need a technique to update the penalty parameter C according to PI during iterations.

Based on the above debold-scriptions, our MAMA for the PM model (5) is stated as follows (Algorithm 1).

Algorithm 1: modified alternating minimization algorithm

  1. 1.

    Set initial precoders V k and decoders U k , kK, the initial penalty parameter C and σ 0 > 1. P ̄ I = P I ( U k , V k ).

  2. 2.

    For decoders: U k = ν min d ( J k ),kK from above.

  3. 3.

    For precoders: V k = ν min d ( J ~ k ),kK, where J ~ k =C l k H lk H U l U l H H lk H kk H U k U k H H kk .

  4. 4.

    If the algorithm converges, then stop and output V k , U k , kK. If P I ( U k , V k )> P ̄ I 1/C, then increase the penalty parameter C := σ 0 C. If P I ( U k , V k )< P ̄ I , let P ̄ I = P I ( U k , V k ). Go to step 2.

In the original AMA of[7], eigenvectors corresponding to the d smallest eigenvalues of positive semi-definite matrices { l k H kl V l V l H H kl H , k K } and those of { l k H kl H U l U l H H kl , k K } are required. In our MAMA, the matrices J k and J ~ k are not necessarily positive semi-definite, since the objective function value of (8) might be negative with feasible solution of quite low PI and high PS.

Although our MAMA avoids calculating SVD which is required by the MSP algorithm given by[11], eigenvalue decompositions are still required for solving subproblems. This lends an impetus to further improvement.

A hybrid algorithm

In order to further reduce the complexity of our MAMA, we propose a hybrid technique in this section. First, we give an AHT, which is a very low complexity algorithm for (5). This algorithm is free from complicated matrix computation, such as eigenvalue decomposition, SVD and QR factorization, and enjoys lower computational complexity than MAMA.

Algorithm with Householder transformation

In this algorithm, we still use the Courant penalty function technique. Thus we focus on problem (7). The only constraints of (7) are orthogonal constraints. We require all the iterative points U k and V k generated by our algorithms are feasible, namely (7b) are always satisfied. By requiring feasibility, basically we are solving an unconstrained optimization problem on the Grassmann manifold. Otherwise, if we allow iterations go outside the feasible region, we will have to adopt some technique, such as projection, to draw the iterations back to the feasible region, which can be complicated. For example, the projection to the set of orthogonal matrices can be a non-differentiable operator and time-consuming to compute.

Preserving orthogonality

Let V k i denotes the precoder of the k th user in the i th iteration. Suppose both V k i and V k i + 1 are orthogonal, there should exist a unitary matrix P k i M × M such that

P k i V k i = V k i + 1 .

The straightforward approach is to search for all unitary matrices P k i M × M , but this can be computationally very expensive. We consider a special class of unitary matrices, the Householder matrices, which can be represented as P = I − 2ppH with pHp = 1. The use of Householder transformation can not only preserve orthogonality, but also turn the problem from a matrix optimization into a vector optimization problem. In the i th iteration with fixed V k i , U k i ,kK as certain feasible solution of (7), we try to obtain orthogonal vectors V k , U k , kK and set

V k i + 1 =( I M 2 v k v k H ) V k i , U k i + 1 =( I N 2 u k u k H ) U k i ,kK
(10)

as the precoders and decoders in the (i + 1)th iteration. Here, V k , U k , kK are regarded as optimal (or suboptimal) solutions to minimize the objective function value of (7),P( U k i + 1 , V k i + 1 ,C)=P(( I N 2 u k u k H ) U k i ,( I M 2 v k v k H ) V k i ,C).

As precoding and decoding matrices are updated according to Householder transformation in each iteration, the new algorithm is denoted as AHT.

Alternating directions

Substituting (10) into (7), we get (11) with variables V k , U k , kK:

min u k , v k k K k K C l k , l K ( U l i + 1 ) H H lk V k i + 1 F 2 ( U k i + 1 ) H H kk V k i + 1 F 2 = k K C l k ( U l i ) H ( I N 2 u l u l H ) H lk ( I M 2 v k v k H ) V k i F 2 ( U k i ) H ( I N 2 u k u k H ) H kk ( I M 2 v k v k H ) V k i F 2 s. t. v k H v k = 1 , u k H u k = 1 , k K .
(11)

With the fact that

( U l i + 1 ) H H lk ( I M 2 v k v k H ) V k i F 2 = tr ( V k i ) H ( I M 2 v k v k H ) H lk H U l i + 1 ( U l i + 1 ) H H lk ( I M 2 v k v k H ) V k i = 4 tr ( V k i ) H v k v k H H lk H U l i + 1 ( U l i + 1 ) H H lk v k v k H V k i 2 tr ( V k i ) H v k v k H H lk H U l i + 1 ( U l i + 1 ) H H lk V k i + ( V k i ) H H lk H U l i + 1 ( U l i + 1 ) H H lk v k v k H V k i + tr ( V k i ) H H lk H U l i + 1 ( U l i + 1 ) H H lk V k i = 4 ( v k H H lk H U l i + 1 ( U l i + 1 ) H H lk v k ) ( v k H V k i ( V k i ) H v k ) 2 ( v k H H lk H U l i + 1 ( U l i + 1 ) H H lk V k i ( V k i ) H v k + v k H V k i ( V k i ) H H lk H U l i + 1 ( U l i + 1 ) H H lk v k ) + tr ( V k i ) H H lk H U l i + 1 ( U l i + 1 ) H H lk V k i ,
(12)

we can rewrite the objective function of (11) as:

k K 4 ( v k H A k v k ) v k H V k i ( V k i ) H v k 2 v k H A k V k i ( V k i ) H v k + v k H V k i ( V k i ) H A k v k + tr ( V k i ) H A k V k i ,
(13)

where A k =C l k , l K H lk H U l i + 1 ( U l i + 1 ) H H lk H kk H U k i + 1 ( U k i + 1 ) H H kk ,kK.

Fixing all u l , lK in (11), U l i + 1 ,lK are determined according to (10), and thus V k , kK become independent of each other. Therefore, (11) is decomposed into K subproblems with the same form as follows:

min x M f ( x ) = ( x H Ax ) ( x H Bx ) x H Cx s. t. x H x = 1 .
(14)

Here x represents V k .A=2 C l k , l K H lk H U l i + 1 ( U l i + 1 ) H H lk H kk H U k i + 1 ( U k i + 1 ) H H kk ,B=2 V k i ( V k i ) H and C = (AB + BA)/2. Similarly, fixing all v k ,kK, it yields K subproblems like (14) from (11). Based on the analysis above, the alternating direction method can be applied to optimize V k , kK and U k , kK alternatively.

Subproblem: 2D subspace method

Subproblem (14) is a nonlinear optimization on the Stiefel manifold in complex field, which is not easy to handle. With the aim of low complexity, we abandon the second order method where Hessian matrix information is required. Instead, we apply a 2D subspace method to solve it iteratively[14]. We define the subspace spanned by x0 and g ~ as

S ( x 0 , g ~ ) = x : x = b x 0 + a g ~ .a [ 1 , 1 ] , b = 1 a 2 ,

where x0 is the current iterative feasible point and g ~ represents its normalized gradient of the Lagrangian function of (14). In each step, the feasible domain of x shrinks from the Stiefel manifold xHx = 1 toS( x 0 , g ~ ).

Here g ~ is calculated as follows. First, the Lagrangian function of (14) is expressed as:

L(x,λ)=f(x)λ( x H x1),
(15)

where λ is the Lagrange multiplier. Then the gradient of (15) is

g= g 0 λ x 0 ,
(16)

where g0 is the gradient of the objective function f at the point x0:

g 0 = 2 [ ( x 0 H Ax 0 ) Bx 0 + ( x 0 H Bx 0 ) Ax 0 Cx 0 ] .

As a necessary condition of the first order optimality condition of (14), we require g H x 0 =0. Thusλ=2( x 0 H Ax 0 )( x 0 H Bx 0 ) x 0 H Cx 0 is deduced due to x 0 H x 0 =1. Once λ is computed, g can be chosen by (16) and we can let g ~ =g/g 2 .

To maintainS( x 0 , g ~ ) on the Stiefel manifold, any pointxS( x 0 , g ~ ) should satisfy x H x= b 2 x 0 2 + a 2 g ~ 2 =1. Thus a2 + b2 = 1, a [−1, 1], namelyb=± 1 a 2 . For any a [−1, 1], the objective function of (14) share the same function value at 1 a 2 x 0 +a g ~ and 1 a 2 x 0 a g ~ . Therefore, we only consider the caseb= 1 a 2 to avoid redundant solutions.

Taking the expression ofxS( x 0 , g ~ ) into (14), it turns into a one dimensional constraint optimization problem with variable a:

min a [ 1 , 1 ] p 1 a 1 a 2 + p 2 a 2 + p 3 a 3 1 a 2 + p 4 a 4 ,
(17)

where the current iterative point x0 corresponds to a = 0, and

p 1 = 2 [ Re ( x 0 H A g ~ ) ( x 0 H Bx 0 ) + Re ( x 0 H B g ~ ) ( x 0 H Ax 0 ) Re ( x 0 H C g ~ ) ] , p 2 = ( x 0 H Ax 0 ) ( g ~ H B g ~ ) + ( x 0 H Bx 0 ) ( g ~ H A g ~ ) 2 ( x 0 H Ax 0 ) ( x 0 H Bx 0 ) + 4 Re ( x 0 H A g ~ ) Re ( x 0 H B g ~ ) + x 0 H Cx 0 g ~ H C g ~ , p 3 = 2 [ Re ( x 0 H B g ~ ) ( g ~ H A g ~ x 0 H Ax 0 ) + Re ( x 0 H A g ~ ) ( g ~ H B g ~ x 0 H Bx 0 ) ] , p 4 = ( g ~ H A g ~ x 0 H Ax 0 ) ( g ~ H B g ~ x 0 H Bx 0 ) 4 Re ( x 0 H A g ~ ) Re ( x 0 H B g ~ ) .

Newton’s method with line search is applied to solve (17) and a local minimizer is guaranteed to be found[13].

The method to solve subproblem (14) is summarized as follows:

Subalgorithm: two dimensional subspace method

  1. 1.

    Given any initial feasible point x 0 and the maximum iteration number T inner, t 1 = 1.

  2. 2.

    Calculate g ~ and construct subproblem (17).

  3. 3.

    Solve (17) by Newton’s method with line search and x 0 := 1 a 2 x 0 +a g ~ . If the objective function of (14) does not reduce or t 1 = T inner, stop and output x 0; else t 1 := t 1 + 1, go to Step 2.

The maximum iteration number Tinner relates to the efficiency of the algorithm. It should not be too large or too small, in order to avoid spending too much time for subproblem or insufficient reduction of the objective function value. We will discuss its specific choice in Section “Simulations”.

As the projection to the Stiefel manifold xHx = 1 is very easy to compute, projected gradient method is also a common approach to solve (14). In the following theorem and remark, we compare our 2D subspace method with the projected gradient method and show the advantage of our method.

Theorem 1

Denote the projected gradient step started from x 0 with stepsize α as:

x(α)= x 0 α g 0 x 0 α g 0 .
(18)

On one hand, for any vectorxS( x 0 , g ~ ), there exists a stepsize α , such that x is expressed as (18); on the other hand, for any α , the projected gradient step x ( α ) can be expressed as the linear combination of x 0 and g ~ , which meansx(α)S( x 0 , g ~ ).

Proof. See the Appendix.

Remark 1

Theorem 1 shows that step in certain 2D subspace and the projected gradient step can be mutually expressed. Thus the minimization of the objective function f on the 2D subspace S( x 0 , g ~ ) and that along the projected gradient step are essentially the same. The 2D subspace problem is easy to handle, as shown above; but the search for (18) may be quite difficult, because the projection operator would make the objective function nondifferentiable and lose the common Taylor expansion, consequently it is nontrivial to search for proper stepsize α . Due to these considerations, we use the 2D subspace method rather than the projected gradient method to solve (14).

Framework of AHT algorithm

According to the above discussions, the framework of the AHT algorithm can be concluded as follows (Algorithm 2).

Algorithm 2: algorithm with Householder transformation

  1. 1.

    Set initial precoders and decoders as V k 0 , U k 0 ,kK, parameter T inner in 2D subspace, the maximum inner iteration number T outer, the initial penalty parameter C, parameter σ 0 > 1 for C and the stopping parameter ε. i = 0.

  2. 2.

    Solve (11) iteratively: fix U k i , V k i ,kK, and set the initial normalized vectors U k , V k , kK. t = 1.

    1. (2.1)

      For precoders: fix U k , kK. For kK, construct A, B, C in (14) and solve it with 2D subspace method to obtain V k .

    2. (2.2)

      For decoders: fix V k , kK. Similarly to Step 2.1, obtain new u k ,kK by solving (14).

    3. (2.3)

      Calculate U k i + 1 , V k i + 1 ,kK from (10). If t = Touter or P I ( U k i + 1 , V k i + 1 ) P I ( U k i , V k i )1/C, go to Step 3; else t := t + 1, go to step 2.1.

  3. 3.

    If P I ( U k i + 1 , V k i + 1 )> P I ( U k i , V k i )1/C, increase the penalty parameter C := σ 0 C. If P I ( U k i + 1 , V k i + 1 )<ε, stop and output U k := U k i + 1 , V k := V k i + 1 ,kK; else i := i + 1, go to Step 2.

In our simulations, t is usually no more than 5 when P I ( U k i + 1 , V k i + 1 ) P I ( U k i , V k i )1/C is achieved in Step 2. Touter, the maximum number of inner iterations inside Step 2, acts to avoid extreme situation with too much time to solve subproblems. Thus we suggest Touter = 5.

Hybrid algorithm

Although the AHT algorithm enjoys lower complexity than MAMA as analyzed in the next section, it solves subproblem (14) inexactly by successive 2D subspace minimizations. This reduces the objective function value rapidly at first, however the speed becomes much more slowly when the iterative point approaches a local optimal solution. Simulation also verifies that it is quite difficult to align the interference perfectly with the AHT algorithm. In contrast, MAMA solves subproblem (9) exactly with closed form solution. The sequence generated by MAMA converges fast near the solution, which is complementary to that of AHT.

In order to make full use of the low complexity of AHT and the fast local convergence property of MAMA, it seems reasonable for us to combine these two methods together into a hybrid algorithm. We believe that such a hybrid algorithm will enjoy the properties of low complexity and fast local convergence, which are verified by our simulation tests reported in Section “Simulations”.

This hybrid algorithm consists of two stages. In the first stage, it uses the AHT algorithm to iterate from any starting point to a point near the solution. Then, the algorithm switches to MAMA, namely it applies MAMA to converge to the solution, by starting from the point obtained by AHT (Algorithm 3).

Algorithm 3: hybrid algorithm

  1. 1.

    Set initial precoders and decoders as V k 0 , U k 0 ,kK, the penalty parameter C = 1.

  2. 2.

    Input the initial iterative point and penalty parameter into the AHT algorithm. Set its stopping criterion as P I ( U k , V k )<0.01 P I ( U k 0 , V k 0 ). Output the corresponding V k , U k , kK and the current penalty parameter C AHT = C.

  3. 3.

    Input V k , U k , kK and C = C AHT as the initial iterative point and penalty parameter of MAMA, respectively. Iterate until convergence by the MAMA. Output the corresponding precoders and decoders V k , U k , kK.

In some sense, it can be viewed that the AHT algorithm acts as a method to rapidly find a good initial point with low interference for MAMA. Starting from the good initial point, MAMA can converge very fast to a local optimal solution with perfect IA.

Analysis of computational complexity

In this section, we analyze the computational complexity of IA algorithms for PM model, including our two proposed algorithms MAMA and AHT, as well as the MSP algorithm from[11].

Advantages of low complexity algorithms

Before presenting the detailed analysis on the computational complexities of different algorithms, we provide the reasons for constructing low computational complexity IA algorithms for the PM model.

  1. 1.

    No matter our algorithms or the existed IA algorithms, they all exploit the channel reciprocity to design the transmit precoders and receiver decoders in a distributed way. Such reciprocity is based on the time division duplex (TDD) operation with synchronized time-slot in the practical system. However, due to the time varying property of wireless channel, perfect reciprocity may be hard to achieve and result in residual interference at the receivers. This requires algorithms with short computing time to reduce the possibility of performance loss due to the imperfect reciprocity in practical systems [15].

  2. 2.

    Furthermore, that the receivers generally have limited computational complexity, might be a bottleneck for the complexity of the algorithms. Thus we need to design algorithms with low complexity and simple computation, to satisfy the computation restriction of communication equipments.

  3. 3.

    Also, algorithms with lower complexity are easier to be extended to large scale problems. For small scale problems, algorithms with low or high complexity do not differ much from each other. However as the problem dimension grows, high complexity algorithms can be very very slow for large scale problems because of too much computing time, while algorithms with lower complexity are more preferable.

Detailed analysis

Here we consider the number of complex multiplications as the complexity criterion. The AHT, MAMA and MSP algorithm are analyzed and compared. Their main computations in one iteration are listed as follows:

  1. 1.

    The computation of H kl V l V l H H kl H and H lk H U l U l H H lk ,k,lK are required for all the three algorithms. Let H ~ kl = H kl V l . Then we can calculate H kl V l V l H H kl H as H ~ kl H ~ kl H . Similarly we can compute H lk H U l U l H H lk by introducing H ̄ lk = H lk H U l . As k and l traverse all the elements in K, the entire complexity is K 2[(dMN + d N 2) + (dMN + d M 2)] = K 2 d(2MN + N 2 + M 2).

  2. 2.

    Besides term 1, the AHT algorithm requires to compute U k U k H and V k V k H , with complexity of Kd(M 2 + N 2).

  3. 3.

    P I is computed in both the AHT and MAMA. Noticing that

    P I = k K l k , l K U k H H kl V l F 2 = k K l k , l K V k H H lk H U l F 2 ,

we can compute U k H H ~ kl F 2 if N < M, or V k H H ̄ lk F 2 otherwise, where H ~ kl and H ̄ lk are the pre-computed parameters in term 1. With the above analysis, the corresponding complexity isK(K1) d 2 (min(M,N)+1).

  1. 4.

    AHT algorithm requires to update precoders and decoders by computing U k i + 1 , V k i + 1 ,kK in (10). With U k i + 1 = U k i 2 u k ( u k H U k i ), the complexity is K[(d 2 + dM) + (d 2 + dN)] = Kd(2d + M + N).

  2. 5.

    In each inner iteration of our 2D subspace method for (14) in the AHT algorithm, it mainly requires computation of Ax, Bx, Cx = [A(Bx) + B(Ax)] /2, Ag and Bg, the complexity of which is 6N 2, where n is the dimension of x. Suppose for each 2D subspace method, there are mostly T inner inner iterations, we can see that each iteration of AHT requires the total complexity of 6K T inner (M 2 + N 2).

  3. 6.

    One eigenvalue decomposition of matrix with dimension M × M requires 9M 3 complexity [16]. Both the MAMA and MSP algorithm need K eigenvalue decompositions of M × M matrices and K with dimension N × N. The entire complexity is 9K(M 3 + N 3).

  4. 7.

    The complexity of one compact SVD of matrix with dimension M × d(d < M) is 14M d 2 + 8d 3 [16]. The MSP algorithm requires K compact SVDs for matrices with dimension M × d and K with dimension N × d, whose complexity is K d 2(14M + 14N + 16d).

Table1 shows the specific computational complexity of the three algorithms. Comparing the AHT and MAMA, besides the common terms, AHT owns term 2,4,5, while MAMA has term 6. With the fact that 1 ≤ d ≤ min(M, N), we can deduce that:

Kd ( 2 d + M + N ) + Kd ( M 2 + N 2 ) 2 Fd ( M + N ) + Fd ( M 2 + N 2 ) 3 K ( M 2 + N 2 ) < 3 K ( M 3 + N 3 ) .
(19)

For term 5, as long as Tinner ≤ min(M, N), we have

6K T inner ( M 2 + N 2 )6K( M 3 + N 3 ).
(20)

Adding both sides of (19) and (20), we conclude that AHT has lower complexity than MAMA.

Table 1 Comparison of computational complexity

Similarly, we compare term 3 and term 7 to see the difference between the complexity of MAMA and MSP. As long as K ≤ 23 (which is usually the case in practical considered IA problems), the following inequality holds:

K ( K 1 ) d 2 ( min ( M , N ) + 1 ) 22 K d 2 ( min ( M , N ) + 1 ) K d 2 ( 22 min ( M , N ) + 22 d ) K d 2 ( 14 M + 14 N + 16 d ) .

So the complexity of MAMA is lower than that of the MSP algorithm. Thus the algorithms ranked from low to high complexity in one iteration are AHT, MAMA and MSP. This also implies that the hybrid algorithm has lower complexity than MAMA and MSP.

Our above analysis compares the complexity of each iteration of different algorithms. However, in order to compare the computational complexity of different algorithms, we need to estimate the total number of iterations of all the algorithms under considerations, which is not easy. Therefore, we try to explain it by convergence curves with consuming time in simulations.

Simulations

In this section, we analyze our proposed algorithms the MAMA, AHT and hybrid algorithm by simulations, and compare them with the existed AMA[7] and MSP algorithm proposed in[11]. (5 × 5, 2)4 interference channels are considered, that is, K = 4, M = N = 5, d = 2, which satisfy the general feasibility condition[2, 3]. Each component of H kl , k, lK is i.i.d complex Gaussian distributionCN(0, 1).

We use sum rate as the measure of quality of service. The sum rate of the K-user MIMO interference channel is expressed as follows:

R = k K log I N + F k 1 L kk = k K log I N + σ 2 I N + l k , l K L kl 1 L kk ,
(21)

where L kl = H kl V l V l H H kl H ,k,lK, F k = σ 2 I N + l k , l K L kl and σ2 is the covariance of the additive white Gaussian noise at the receivers. Here we define the SNR as SNR = P / σ2 = d / σ2.

In the AHT algorithm, we set Touter = 5 and initially C = 1. For each figure, 250 random realization of different channel coefficients are generated to evaluate the average performance. For each realization the initial values of precoders and decoders V k 0 , U k 0 ,kK are randomly generated and remain the same in the compared algorithms.

Parameter analysis

As mentioned before, the choice of Tinner relates to the efficiency of the AHT algorithm and also the hybrid algorithm. In Figure2 we plot the convergence curves of the hybrid algorithms with parameter Tinner = 3, 5, 10, respectively. SNR equals 30 dB and σ0 = 20 here.

Figure 2
figure2

Convergence of the average sum rate of the hybrid algorithms with Tinner = 3, 5, 10.

Figure2 reveals that, both the converged average sum rate and the convergence rate of the hybrid algorithms with different Tinner are quite similar. The curve with Tinner = 10 seems to be the worst, which accords with our complexity analysis as Tinner > min(M, N) = 5. In the upcoming simulations we set Tinner = 5 with which the hybrid algorithm obtains the highest sum rate, to improve the efficiency of our proposed algorithms.

Besides Tinner, we also compare the performances of the hybrid algorithms with different σ0. σ0 determines the growth speed of the penalty parameter C, which represents the weight of PI in the objective function of (7). If σ0 is too small, it would take many iterations for C to grow large, and thus affects the efficiency of the hybrid algorithm. On the other side, too large σ0 forces C to grow too fast, which makes (7) degenerate into the LIM problem in short time and that PS cannot be increased efficiently.

Based on the aforementioned analysis, we plot the convergence curves of the hybrid algorithms with σ0 = 10, 20, 50 in Figure3. SNR is also set as 30 dB here. The three curves are very close to each other, which indicates that the algorithm is not sensitive to σ0. In the following simulations we set σ0 = 20.

Figure 3
figure3

Convergence of the average sum rate of the hybrid algorithms with σ0 = 10, 20, 50.

AHT algorithm

In this section, the relative iteration number to achieve certain leakage interference of the AHT and MSP algorithm are displayed in Figure4, to show the property of AHT algorithm. The stopping criterion is PI ≤ 0.01. The average number of iterations of the MSP and AHT algorithm to solve one problem are 1233 and 583, respectively. To analyze the relative convergence performance, we use the normalized iterations, the ratio of the iterations at the certain point and the total iterations, as the x-axis. And the y-axis represents the total leakage interference PI. Figure4 implies that to achieve the interference of 0.5 from the initial point, the AHT algorithm requires 7.5% of its total iterations while MSP requires 45%. Compared to the MSP algorithm, the AHT algorithm iterates rapidly to a point with low leakage interference PI. Actually the main consuming time of AHT algorithm is used to reduce PI from less than 1 towards 0.01, and the convergence rate becomes much slower when PI is smaller. The performances here accords with the previous analysis.

Figure 4
figure4

Comparison of AHT and MSP: total leakage interference versus normalized iteration number.

Comparison of different IA algorithms

In this section, the performances of different IA algorithms are compared and analyzed, including our proposed MAMA, AHT and hybrid algorithm, as well as the existed MSP and AMA. The number of iterations of each algorithm to solve each problem is set as 3000, in order to perfectly align the interference. To take full advantage of AHT, we set its stopping parameterε=0.01 P I ( U k 0 , V k 0 ), where V k 0 , U k 0 ,kK are the initial precoders and decoders.

The average sum rate and relative computing time of the four algorithms (MAMA, Hybrid, MSP and AMA) with respect to different SNRs are depicted in Figures5 and6, respectively. Both our proposed algorithms show good performances, especially the hybrid algorithm. As reflected in Figure5, the curves representing the average sum rate of MSP, MAMA and the hybrid algorithms are quite close to each other. The solutions of the three algorithms are quite different, whereas they get similar sum rate under the same SNR. And all the three algorithms gain about 5 bps/Hz higher sum rate than the AMA under medium and high SNR. In Figure6, both of our proposed two algorithms require much less computing time than the MSP algorithm. Particularly, the hybrid algorithm gain almost as high sum rate as MSP with as little time as the AMA.

Figure 5
figure5

Comparison of MAMA, Hybrid, MSP and AMA: average sum rate versus SNR.

Figure 6
figure6

Comparison of MAMA, Hybrid, MSP and AMA: average relative computing time versus SNR. Time of AMA is set as standard.

To further compare the convergence performances of these four algorithms as well as AHT algorithm, we plot their convergence curves of sum rate with respect to the running time in Figure7. As AHT algorithm is difficult to converge and eliminate interference, it is shown as the first stage of the hybrid algorithm. Its output point is used to continue iteration in the hybrid algorithm. SNR are set as 5, 15 and 30 dB to represent the scenarios of low, medium and high SNR, respectively. In all three scenarios, both MAMA and the hybrid algorithm achieve similar converged average sum rate as the MSP algorithm, which gain more sum rate than the AMA. To achieve a certain sum rate, MAMA consumes the least time. We also observe an interesting phenomenon, that in the low and medium SNR scenario each of the MAMA, hybrid and MSP algorithm achieves high sum rate during iteration before convergence, after that it reduces and converges to a lower rate. This may be relevant to the value of σ2. As shown at the beginning of this section, SNR value is inversely proportional to σ2. For the low SNR scenario, σ2 is considerably high and σ2I N takes the main part of F k in (21). Then, the sum rate mainly increases with large PS. All the three concerned algorithms are designed to increase PS in the first few iterations and gradually reduce the weight of this requirement. This may lead to the phenomenon that PS increases to a peak value and then decreases gradually, which explains the similar phenomenon of sum rate in Figure7.

Figure 7
figure7

Convergence of the average sum rate of MAMA, AHT, Hybrid, MSP and AMA with respect to SNR = 5, 15, 30 dB.

Similar as Figure7, for the five compared algorithms (MSP, AHT, Hybrid, MAMA and AMA), we also plot the convergence curves of the leakage interference with respect to the running time in Figure8. As the convergence curves of the same algorithm are quite close for different SNR scenarios, here we only display the scenario that SNR is 30 dB for representation. Among the algorithms, the hybrid algorithm converges fastest to a point with interference less than 10−4. The MAMA also converges faster than the MSP and AMA.

Figure 8
figure8

Convergence of the leakage interference of MAMA, AHT, Hybrid, MSP and AMA with respect to SNR = 30 dB.

Comparing the proposed MAMA and hybrid algorithm from both aspects of computing time and convergence rate in Figures6,7 and8, we should admit that the improvement of the hybrid algorithm over MAMA is limited in the examples that we tested. The main reason of this phenomenon is due to the quite small scale of the test problems, for which lower complexity algorithm does not have much gain. Nevertheless, as the dimension of the problem increases, the hybrid algorithm will save more consuming time and benefit more. Moreover, in similar applications in other fields, there might be large scale similar matrix optimization problems, we believe that the hybrid algorithm will improve MAMA greatly.

Remark 2

We have pointed out in Section “A hybrid algorithm” that the sequence generated by the hybrid algorithm converges to a local optimal solution. Although it is not guaranteed to be global optimal, the obtained local optimal solution performs with high sum rate (almost the same as MSP algorithm and higher than the AMA, as depicted in Figure 5 ) and perfect IA. Such solution satisfies our requirement for an IA solution.

The performances of MAMA and the hybrid algorithm in this section indicates that the Courant penalty function technique is an effective way to combine the leakage interference and the desired signal power together. Through this technique, both algorithms achieve high average sum rate in short time. In the last few iterations when C becomes very large, the algorithm essentially tries to minimize P I with the part of − P S being nearly ignored. Therefore, MAMA (also the hybrid algorithm) eventually behaves very similar to the original AMA, which can perfectly align the interference.

Conclusion

This article proposed two low complexity algorithms for the desired signal power maximization IA problems of MIMO channels. The IA constraints are added to the objective function and combined with the desired signal power via the Courant penalty function technique. First, a MAMA is proposed following the similar approach of the AMA. Then, a hybrid algorithm is proposed to further reduce complexity. In the hybrid algorithm, the AHT is proposed to iterate rapidly from any initial point to precoders and decoders with low leakage interference. This step provides a good point around the local optimal solution. From this point, MAMA is applied to converge fast to the local optimal solution satisfying our requirement. Analysis shows that, among the compared algorithms the hybrid algorithm has the lowest computational complexity, followed by MAMA, with MSP being the highest. Simulations indicate that both the hybrid algorithm and MAMA achieve similar sum rate as the MSP algorithm with less computing time, and higher sum rate than the AMA.

Appendix

Proof of Theorem 1

As stated in (16), g0 is the linear combination of x0 and g ~ . Suppose

g 0 = c 1 x 0 + c 2 g ~ ,

where c1 and c2 are scalars.

First prove that the projected gradient step can be expressed by 2D subspace step. For the projected gradient step x(α) with any α, we would like to find corresponding scalars a and b, such that

x(α)= x 0 α g 0 x 0 α g 0 = ( 1 c 1 α ) x 0 c 2 α g ~ ( 1 c 1 α ) x 0 c 2 α g ~ =b x 0 +a g ~ .
(22)

Let the coefficients of x0 and g ~ remain unchanged, then

a = c 2 α ( 1 c 1 α ) x 0 c 2 α g ~ , b = 1 c 1 α ( 1 c 1 α ) x 0 c 2 α g ~ .

Next show that 2D subspace step can also be expressed by the projected gradient step. For anyxS( x 0 , g ~ ), we havex=b x 0 +a g ~ . As we wish to find α such that x = x(α), then (22) holds. Similar to the proof above, we can induce that

b a = c 1 α 1 c 2 α , α = a c 1 a c 2 b .

Thus we have proved Theorem 1.

References

  1. 1.

    Cadambe VR, Jafar SA: Interference alignment and degrees of freedom of the K-user interference channel. IEEE Trans. Inf. Theory 2008, 54(8):3425-3441.

    MathSciNet  Article  Google Scholar 

  2. 2.

    Yetis CM, Gou T, Jafar SA, Kayran AH: On feasibility of interference alignment in MIMO interference networks. IEEE Trans. Signal Process 2010, 58(9):4771-4782.

    MathSciNet  Article  Google Scholar 

  3. 3.

    Razaviyayn M, Lyubeznik G, Luo ZQ: On the degrees of freedom achievable through interference alignment in a MIMO interference channel. IEEE Trans. Signal Process 2012, 60(2):812-821.

    MathSciNet  Article  Google Scholar 

  4. 4.

    Choi SW, Chung S-Y: On the multiplexing gain of K-user partially connected interference channel,. preprint,. Jun 2008 http://arxiv.org/pdf/0806.4737

    Google Scholar 

  5. 5.

    Guillaud M, Gesbert D: Interference alignment in the partially connected K-user MIMO interference channel. European Signal Processing Conference (EUSIPCO) (Barcelona, Spain, 2011), pp. 1095–1099

    Google Scholar 

  6. 6.

    Lee N, Park D, Kim Y-D: Degrees of freedom on the K-user MIMO interference channel with constant channel coefficients for downlink communications. Proc. IEEE Global Telecommun. Conf., (GLOBECOM) (Hawaii, USA, 2009), pp. 1–6

    Google Scholar 

  7. 7.

    Gomadam K, Cadambe VR, Jafar SA: A distributed numerical approach to interference alignment and applications to wireless interference networks. IEEE Trans. Inf. Theory 2011, 57(6):3309-3322.

    MathSciNet  Article  Google Scholar 

  8. 8.

    Liu YF, Dai YH, Luo ZQ: On the complexity of leakage interference minimization for interference alignment. IEEE Int. Workshop Signal Process. Advances Wireless Comm (San francisco, USA, 2011), pp. 471–475

    Google Scholar 

  9. 9.

    Schmidt DA, Utschick W, Honig ML: Large system performance of interference alignment in single-beam MIMO networks. Proc. IEEE Global Telecommun. Conf., (GLOBECOM) (Miami, USA, 2010), pp. 1–6

    Google Scholar 

  10. 10.

    Shen H, Li B, Tao MX, Wang XD: MSE-based transceiver designs for the MIMO interference channel. IEEE Trans. Wirel. Commun 2010, 9(11):3480-3489.

    Article  Google Scholar 

  11. 11.

    Santamaria I, Gonzalez O, Heath Jr RW, Peters SW: Maximum sum-rate interference alignment algorithms for MIMO channels. Proc. IEEE Global Telecommun. Conf., (GLOBECOM) (Miami, USA, 2010), pp. 1–6

    Google Scholar 

  12. 12.

    Nocedal J, Wright SJ: Numerical Optimization. (Springer, Berlin, 1999), pp. 488–522

    Book  Google Scholar 

  13. 13.

    Yuan YX: Computational Mehthod for Nonlinear Optimization (in Chinese). (Science Press, Beijing, 2008), pp. 16–20; 158–162

    Google Scholar 

  14. 14.

    Yuan YX(Higher Education Press, Beijing, 2007), pp. 206–218

  15. 15.

    Ghauch HG, Papadias CB: Interference alignment: a one-sided approach. Proc. IEEE Global Telecommun. Conf., (GLOBECOM) (Houston, USA, 2011), pp. 1–6

    Google Scholar 

  16. 16.

    Golub GH, Van Loan CF: Matrix Computations,. (The Johns Hopkins University Press, London, 1996), pp. 248–255; 414–425

    Google Scholar 

Download references

Acknowledgements

The authors would like to thank the editor and two anonymous referees for their comments and suggestions which improved the article greatly. This work was partly supported by the National Natural Science Foundation of China (NSFC) grants 10831006, 11021101, by CAS grant kjcx-yw-s7 and by the Fundamental Research Funds for the Central Universities.

Author information

Affiliations

Authors

Corresponding author

Correspondence to Cong Sun.

Additional information

Competing interests

The authors declare that they have no competing interests.

Authors’ original submitted files for images

Rights and permissions

Open Access This article is distributed under the terms of the Creative Commons Attribution 2.0 International License (https://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Reprints and Permissions

About this article

Cite this article

Sun, C., Yang, Y. & Yuan, Y. Low complexity interference alignment algorithms for desired signal power maximization problem of MIMO channels. EURASIP J. Adv. Signal Process. 2012, 137 (2012). https://doi.org/10.1186/1687-6180-2012-137

Download citation

Keywords

  • Interference alignment
  • Power maximization
  • Courant penalty function
  • Alternating minimization algorithm
  • Householder transformation