ieee transactions on signal processing, vol. 61, no. 23 ... · tively. the beamformer output at the...

15
IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 61, NO. 23, DECEMBER 1, 2013 6103 Robust Adaptive Beamforming for General-Rank Signal Model With Positive Semi-Denite Constraint via POTDC Arash Khabbazibasmenj, Student Member, IEEE, and Sergiy A. Vorobyov, Senior Member, IEEE Abstract—The robust adaptive beamforming (RAB) problem for general-rank signal model with an additional positive semi-def- inite constraint is considered. Using the principle of the worst-case performance optimization, such RAB problem leads to a differ- ence-of-convex functions (DC) optimization problem. The existing approaches for solving the resulted non-convex DC problem are based on approximations and nd only suboptimal solutions. Here, we aim at nding the globally optimal solution for the non-convex DC problem and clarify the conditions under which the solution is guaranteed to be globally optimal. Particularly, we rewrite the problem as the minimization of a one-dimensional optimal value function (OVF). Then, the OVF is replaced with another equivalent one, for which the corresponding optimization problem is convex. The new one-dimensional OVF is minimized iteratively via polynomial time DC (POTDC) algorithm. We show that the POTDC converges to a point that satises Karush-Kuhn-Tucker (KKT) optimality conditions, and such point is the global optimum under certain conditions. Towards this conclusion, we prove that the proposed algorithm nds the globally optimal solution if the presumed norm of the mismatch matrix that corresponds to the desired signal covariance matrix is sufciently small. The new RAB method shows superior performance compared to the other state-of-the-art general-rank RAB methods. Index Terms—Difference-of-convex functions (DC) program- ming, non-convex programming, semi-denite programming relaxation, robust adaptive beamforming, general-rank signal model, polynomial time DC (POTDC). I. INTRODUCTION I T is well known that when the desired signal is present in the training data, the performance of adaptive beamforming methods degrades dramatically in the presence of even a very slight mismatch in the knowledge of the desired signal covari- ance matrix. The mismatch between the presumed and actual Manuscript received December 11, 2012; revised May 18, 2013 and August 25, 2013; accepted August 29, 2013. Date of publication September 10, 2013; date of current version November 05, 2013. The associate editor coordinating the review of this manuscript and approving it for publication was Dr. Antonio De Maio. This work was supported in part by the Natural Science and Engi- neering Research Council (NSERC) of Canada. Parts of this paper were pre- sented at CAMSAP San Juan, Puerto Rico, 2011. A. Khabbazibasmenj is with the Department of Electrical and Computer En- gineering, University of Alberta, Edmonton, AB T6G 2V4 Canada (e-mail: [email protected]). S. A. Vorobyov is with the Department of Signal Processing and Acoustics, Aalto University, FI-00076 AALTO, Finland, on leave from the Department of Electrical and Computer Engineering, University of Alberta, Edmonton, AB T6G 2V4, Canada (e-mail: [email protected]). Color versions of one or more of the gures in this paper are available online at http://ieeexplore.ieee.org. Digital Object Identier 10.1109/TSP.2013.2281301 source covariance matrices occurs because of, for example, dis- placement of antenna elements, time varying environment, im- perfections of propagation medium, etc. The main goal of any robust adaptive beamforming (RAB) technique is to provide ro- bustness against any such mismatches. Most of the RAB methods have been developed for the case of point source signals when the rank of the desired signal co- variance matrix is equal to one [1]–[11]. Among the principles used for such RAB methods design are i) the worst-case perfor- mance optimization [2]–[6]; ii) probabilistic based performance optimization [8]; and iii) estimation of the actual steering vector of the desired signal [9]–[11]. In many practical applications such as, for example, the incoherently scattered signal source or source with uctuating (randomly distorted) wavefront, the rank of the source covariance matrix is higher than one. Although the RAB methods of [1]–[11] provide excellent robustness against any mismatch of the underlying point source assumption, they are not perfectly suited to the case when the rank of the desired signal covariance matrix is higher than one. The RAB for the general-rank signal model based on the ex- plicit modeling of the error mismatches has been developed in [12] based on the worst-case performance optimization prin- ciple. Although the RAB of [12] has a simple closed form so- lution, it is overly conservative because the worst-case correla- tion matrix of the desired signal may be indenite or even nega- tive denite [13]–[15]. Thus, less conservative approaches have been developed in [13]–[15] by considering an additional posi- tive semi-denite (PSD) constraint to the worst-case signal co- variance matrix. The major shortcoming of the RAB methods of [13]–[15] is that they nd only a suboptimal solution and there may be a signicant gap to the global optimal solution. For ex- ample, the RAB of [13] nds a suboptimal solution in an iter- ative way, but there is no guarantee that such iterative method converges [15]. A closed-form approximate suboptimal solu- tion is proposed in [14], however, this solution may be quite far from the globally optimal one as well. All these shortcomings motivate us to look for new efcient ways to solve the afore- mentioned non-convex problem globally optimally. 1 We propose a new method that is based on recasting the orig- inal non-convex difference-of-convex functions (DC) program- ming problem as the minimization of a one dimensional optimal value function (OVF). Although the corresponding optimiza- tion problem of the newly introduced OVF is non-convex, it can be replaced with another equivalent problem. Such optimiza- tion problem is convex and can be solved efciently. The new 1 Some preliminary results have been presented in [16]. 1053-587X © 2013 IEEE

Upload: others

Post on 24-Sep-2020

2 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 61, NO. 23 ... · tively. The beamformer output at the time instant is given as (2) where is the complex beamforming vector of the antenna

IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 61, NO. 23, DECEMBER 1, 2013 6103

Robust Adaptive Beamforming for General-RankSignal Model With Positive Semi-Definite

Constraint via POTDCArash Khabbazibasmenj, Student Member, IEEE, and Sergiy A. Vorobyov, Senior Member, IEEE

Abstract—The robust adaptive beamforming (RAB) problemfor general-rank signal model with an additional positive semi-def-inite constraint is considered. Using the principle of the worst-caseperformance optimization, such RAB problem leads to a differ-ence-of-convex functions (DC) optimization problem. The existingapproaches for solving the resulted non-convex DC problem arebased on approximations and find only suboptimal solutions. Here,we aim at finding the globally optimal solution for the non-convexDC problem and clarify the conditions under which the solutionis guaranteed to be globally optimal. Particularly, we rewritethe problem as the minimization of a one-dimensional optimalvalue function (OVF). Then, the OVF is replaced with anotherequivalent one, for which the corresponding optimization problemis convex. The new one-dimensional OVF is minimized iterativelyvia polynomial time DC (POTDC) algorithm. We show that thePOTDC converges to a point that satisfies Karush-Kuhn-Tucker(KKT) optimality conditions, and such point is the global optimumunder certain conditions. Towards this conclusion, we prove thatthe proposed algorithm finds the globally optimal solution if thepresumed norm of the mismatch matrix that corresponds to thedesired signal covariance matrix is sufficiently small. The newRAB method shows superior performance compared to the otherstate-of-the-art general-rank RAB methods.

Index Terms—Difference-of-convex functions (DC) program-ming, non-convex programming, semi-definite programmingrelaxation, robust adaptive beamforming, general-rank signalmodel, polynomial time DC (POTDC).

I. INTRODUCTION

I T is well known that when the desired signal is present inthe training data, the performance of adaptive beamforming

methods degrades dramatically in the presence of even a veryslight mismatch in the knowledge of the desired signal covari-ance matrix. The mismatch between the presumed and actual

Manuscript received December 11, 2012; revised May 18, 2013 and August25, 2013; accepted August 29, 2013. Date of publication September 10, 2013;date of current version November 05, 2013. The associate editor coordinatingthe review of this manuscript and approving it for publication was Dr. AntonioDe Maio. This work was supported in part by the Natural Science and Engi-neering Research Council (NSERC) of Canada. Parts of this paper were pre-sented at CAMSAP San Juan, Puerto Rico, 2011.A. Khabbazibasmenj is with the Department of Electrical and Computer En-

gineering, University of Alberta, Edmonton, AB T6G 2V4 Canada (e-mail:[email protected]).S. A. Vorobyov is with the Department of Signal Processing and Acoustics,

Aalto University, FI-00076 AALTO, Finland, on leave from the Departmentof Electrical and Computer Engineering, University of Alberta, Edmonton, ABT6G 2V4, Canada (e-mail: [email protected]).Color versions of one or more of the figures in this paper are available online

at http://ieeexplore.ieee.org.Digital Object Identifier 10.1109/TSP.2013.2281301

source covariance matrices occurs because of, for example, dis-placement of antenna elements, time varying environment, im-perfections of propagation medium, etc. The main goal of anyrobust adaptive beamforming (RAB) technique is to provide ro-bustness against any such mismatches.Most of the RAB methods have been developed for the case

of point source signals when the rank of the desired signal co-variance matrix is equal to one [1]–[11]. Among the principlesused for such RAB methods design are i) the worst-case perfor-mance optimization [2]–[6]; ii) probabilistic based performanceoptimization [8]; and iii) estimation of the actual steering vectorof the desired signal [9]–[11]. In many practical applicationssuch as, for example, the incoherently scattered signal source orsource with fluctuating (randomly distorted) wavefront, the rankof the source covariance matrix is higher than one. Although theRAB methods of [1]–[11] provide excellent robustness againstany mismatch of the underlying point source assumption, theyare not perfectly suited to the case when the rank of the desiredsignal covariance matrix is higher than one.The RAB for the general-rank signal model based on the ex-

plicit modeling of the error mismatches has been developed in[12] based on the worst-case performance optimization prin-ciple. Although the RAB of [12] has a simple closed form so-lution, it is overly conservative because the worst-case correla-tion matrix of the desired signal may be indefinite or even nega-tive definite [13]–[15]. Thus, less conservative approaches havebeen developed in [13]–[15] by considering an additional posi-tive semi-definite (PSD) constraint to the worst-case signal co-variance matrix. The major shortcoming of the RABmethods of[13]–[15] is that they find only a suboptimal solution and theremay be a significant gap to the global optimal solution. For ex-ample, the RAB of [13] finds a suboptimal solution in an iter-ative way, but there is no guarantee that such iterative methodconverges [15]. A closed-form approximate suboptimal solu-tion is proposed in [14], however, this solution may be quite farfrom the globally optimal one as well. All these shortcomingsmotivate us to look for new efficient ways to solve the afore-mentioned non-convex problem globally optimally.1

We propose a new method that is based on recasting the orig-inal non-convex difference-of-convex functions (DC) program-ming problem as the minimization of a one dimensional optimalvalue function (OVF). Although the corresponding optimiza-tion problem of the newly introduced OVF is non-convex, it canbe replaced with another equivalent problem. Such optimiza-tion problem is convex and can be solved efficiently. The new

1Some preliminary results have been presented in [16].

1053-587X © 2013 IEEE

Page 2: IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 61, NO. 23 ... · tively. The beamformer output at the time instant is given as (2) where is the complex beamforming vector of the antenna

6104 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 61, NO. 23, DECEMBER 1, 2013

one-dimensional OVF is then minimized by the means of thepolynomial time DC (POTDC) algorithm (see also [17], [18]).We prove that the point found by the POTDC algorithm forthe RAB for general-rank signal model with positive semi-def-inite constraint is a Karush-Kuhn-Tucker (KKT) optimal point.Moreover, we prove a number of results that lead us to the equiv-alence between the claim of global optimality for the problemconsidered and the convexity or strict quasi-convexity of thenewly obtained one-dimensional OVF. The global optimality ofthe proposed POTDC method is then proved under some con-ditions. As an additional check, we also develop a tight lower-bound for such OVF that is used in the simulations to furtherconfirming global optimality.The rest of the paper is organized as follows. System model

and preliminaries are given in Section II, while the problemis formulated in Section III. The new proposed method isdeveloped in Section IV followed by our simulation results inSection V. Finally, Section VI presents our conclusions. Thispaper is reproducible research, and the software needed togenerate the simulation results can be obtained from the IEEEXplore together with the paper.

II. SYSTEM MODEL AND PRELIMINARIES

The narrowband signal received by a linear antenna arraywith omni-directional antenna elements at the time instantcan be expressed as

(1)

where , , and are the statistically independentvectors of the desired signal, interferences, and noise, respec-tively. The beamformer output at the time instant is given as

(2)

where is the complex beamforming vector of theantenna array and stands for the Hermitian transpose.The beamforming problem is formulated as finding the beam-forming vector which maximizes the beamformer outputsignal-to-interference-plus-noise ratio (SINR) given as

(3)

where andare the desired signal and interfer-

ence-plus-noise covariance matrices, respectively, andstands for the statistical expectation.Depending on the nature of the desired signal source, its cor-

responding covariance matrix can be of an arbitrary rank, i.e.,, where rank denotes the rank oper-

ator. Indeed, in many practical applications, for example, in thescenarios with incoherently scattered signal sources or signalswith randomly fluctuating wavefronts, the rank of the desiredsignal covariance matrix is greater than one [12]. The onlyparticular case in which, the rank of is equal to one is thecase of the point source.

The interference-plus-noise covariance matrix is typ-ically unavailable in practice and it is substituted by the datasample covariance matrix

(4)

where is number of the training data samples. The problemof maximizing the SINR (3) (here we always use sample ma-trix estimate instead of ) is known as minimum vari-ance distortionless response (MVDR) beamforming and can bemathematically formulated as

(5)

The solution to the MVDR beamforming problem (5) can befound as [1]

(6)

which is known as the sample matrix inversion (SMI) MVDRbeamformer for general-rank signal model. Here standsfor the principal eigenvector operator.In practice, the actual desired signal covariance matrix

is usually unknown and only its presumed value is avail-able. The actual source correlation matrix can be modeledas , where and denote an unknownmismatch and the presumed correlation matrices, respectively.It is well known that the MVDR beamformer is very sensitiveto such mismatches [12]. RABs also address the situationwhen the sample estimate of the data covariance matrix (4)is inaccurate (for example, because of small sample size) and

, where is an unknown mismatch matrix tothe data sample covariance matrix. In order to provide robust-ness against the norm-bounded mismatches and

(here denotes the Frobenius norm of a matrix),the RAB of [12] uses the worst-case performance optimizationprinciple of [2] and finds the solution as

(7)

Although the RAB of (7) has a simple closed-form expression,it is overly conservative because the constraint that the matrix

has to be PSD is not considered [13]. For example,the worst-case desired signal covariance matrix in (7)can be indefinite or negative definite if is rank deficient.Indeed, in the case of incoherently scattered source, has thefollowing form , wheredenotes the normalized angular power density, is the desiredsignal power, and is the steering vector towards direction .For a uniform angular power density on the angular bandwidth, the approximate numerical rank of is equal to[19]. This leads to a rank deficient matrix if the angular

power density does not cover all the directions. Therefore, theworst-case covariance matrix is indefinite or negativedefinite. Note that the worst-case data sample covariance matrix

is always positive definite.

Page 3: IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 61, NO. 23 ... · tively. The beamformer output at the time instant is given as (2) where is the complex beamforming vector of the antenna

KHABBAZIBASMENJ AND VOROBYOV: ROBUST ADAPTIVE BEAMFORMING 6105

III. PROBLEM FORMULATION

Decomposing as , the RAB problem for anorm-bounded mismatch to the matrix is given as[13]

(8)

For every in the optimization problem (8) whose norm is lessthan or equal to , the expressionrepresents a non-convex quadratic constraint with respect to .Because there exists infinite number of mismatches , therealso exists infinite number of such non-convex quadratic con-straints. By finding the minimum possible value of the quadraticterm with respect to for a fixed ,the infinite number of such constraints can be replaced with asingle constraint. Hence, we consider the following optimiza-tion problem

(9)

This problem is convex and its optimal value can be expressedas a function of as given by the following lemma.Lemma 1: The optimal value of the optimization problem (9)

as a function of is equal to

otherwise.(10)

Proof: See Appendix I-A.It follows from (10) that the desired signal can be totally re-

moved from the beamformer output if . Basedon Lemma 1, the constraint in (8) can be equivalently replacedby the constraint

(11)

Moreover, the maximum of the quadratic termin the objective function of the problem in (8) with respect to, can be easily derived as . There-

fore, the RAB problem (8) can be equivalently written in a sim-pler form as

(12)

Due to the non-convex DC constraint, the problem (12) is anon-convex DC programming problem [17], [18]. DC optimiza-tion problems are believed to be NP-hard in general [20], [21].There is a number of methods that can be applied to addressDC programming problems of type (12). Among these methodsare the generalized poly block algorithm, the extended gen-eral power iterative (GPI) algorithm [22], DC iteration-basedmethod [23], etc. However, the existing methods do not guar-antee to find the globally optimal solution of a DC programmingproblem in polynomial time.

Recently, the problem (12) has also been suboptimally solvedusing an iterative semi-definite relaxation (SDR)-based algo-rithm in [13] which also does not result in the globally optimalsolution and for which the convergence even to a KKT optimalpoint is not guaranteed. A closed-form suboptimal solution forthe aforementioned non-convex DC problem has been also de-rived in [14]. Despite its computational simplicity, the perfor-mance of the method of [14] may be far from the global op-timum and even the KKT optimal point. Another iterative algo-rithm has been proposed in [15], but it modifies the problem (12)and solves the modified problem instead, which again gives noguarantees of finding the globally optimal solution of the orig-inal problem (12).

IV. NEW PROPOSED METHOD

A. Main Idea and OVF

Here, we aim at solving the problem (12) globally optimallyin polynomial time. For this goal, we design a POTDC-type al-gorithm (see also [17], [18]) that can be used for solving a classof DC programming problems in polynomial time. By intro-ducing the auxiliary optimization variable and setting

, the problem (12) can be equivalently rewrittenas

(13)

Note that is restricted to be greater than or equal to one be-cause is greater than or equal to one due to the constraintof the problem (12). For future needs, we find the set of all ’sfor which the problem (13) is feasible. Let us define the fol-lowing set for a fixed value of

(14)

It is trivial that for every , the quadratic termis non-negative as is a positive semi-definite

matrix. Using the minimax theorem [24], it can be easily veri-fied that the maximum value of the quadratic termover is equal to andthis value is achieved by

(15)

Here stands for the largest eigenvalue operator. Due tothe fact that for any , the scaled vector liesinside the set , the quadratic term can takevalues only in the intervalover .Considering the later fact and also the optimization problem

(13), it can be concluded that is feasible if and only ifwhich implies that

(16)

Page 4: IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 61, NO. 23 ... · tively. The beamformer output at the time instant is given as (2) where is the complex beamforming vector of the antenna

6106 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 61, NO. 23, DECEMBER 1, 2013

or, equivalently, that

(17)

The function is strictly increasing and it isalso less than or equal to one for . Therefore, it can beimmediately found that the problem (13) is infeasible for any

if . Thus, hereafter, it is assumed that. Moreover, using (17) and the fact that the

function is strictly increasing, it can be found thatthe feasible set of the problem (13) corresponds to

(18)

As we will see in the following sections, for developing thePOTDC algorithm for the problem (13), an upper-bound for theoptimal value of in (13) is needed. Such upper-bound is ob-tained in terms of the following lemma.Lemma 2: The optimal value of the optimization

variable in the problem (13) is upper-bounded by

, where isany arbitrary feasible point of the problem (13).

Proof: See Appendix I-B.Using Lemma 2, the problem (13) can be equivalently stated

as

(19)

where

(20)

and

(21)

For a fixed value of , the inner optimization problem in (19) isnon-convex with respect to . Based on the inner optimizationproblem in (19) when is fixed, we define the following OVF

(22)

Using the OVF (22), the problem (19) can be equivalentlyexpressed as

(23)

The corresponding optimization problem of for a fixedvalue of is non-convex. In what follows, we aim at replacing

with an equivalent OVF whose corresponding optimiza-tion problem is convex.Introducing the matrix and using the fact that

for any arbitrary matrix , (herestands for the trace of a matrix), the OVF (22) can be equiva-lently recast as

(24)

By dropping the rank-one constraint in the corresponding opti-mization problem of for a fixed value of , ,a new OVF denoted as can be defined as

(25)

For brevity, we will refer to the optimization problems thatcorrespond to the OVFs and when is fixed, asthe optimization problems of and , respectively. Notealso that compared to the optimization problem of , the op-timization problem of is convex. Moreover, it is easy tocheck that the optimization problem of is a hidden convexproblem, i.e., the duality gap between this problem and its dualis zero [25]–[28]. Since both of the optimization problems of

and have the same dual problem, it can be immedi-ately concluded that the OVFs and are equivalent,i.e., for any . Furthermore, based onthe optimal solution of the optimization problem of whenis fixed, the optimal solution of the optimization problem ofcan be constructed [25]–[28]. Based on the later fact, the

original problem (23) can be expressed as

(26)

It is noteworthy to mention that based on the optimal solu-tion of (26) denoted as , we can easily obtain the optimalsolution of the original problem (23) or, equivalently, the op-timal solution of the problem (19). Specifically, since the OVFs

and are equivalent, is also the optimal solutionof the problem (23) and, thus, also the problem (19). Moreover,the optimization problem of is convex and can be easilysolved. In addition, using the results in [25]–[28] and based onthe optimal solution of the optimization problem of , theoptimal solution of the optimization problem of can beconstructed. Thus, we concentrate on the problem (26).Since for every fixed value of , the corresponding optimiza-

tion problem of is a convex semi-definite programming(SDP) problem, one possible approach for solving (26) is basedon exhaustive search over . In other words, can be found

Page 5: IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 61, NO. 23 ... · tively. The beamformer output at the time instant is given as (2) where is the complex beamforming vector of the antenna

KHABBAZIBASMENJ AND VOROBYOV: ROBUST ADAPTIVE BEAMFORMING 6107

by using an exhaustive search over a fine grid on the intervalof . Although this search method is inefficient, it can beused as a benchmark.Using the definition of the OVF , the problem (26) can

be equivalently expressed as

(27)

Note that replacing by results in a much simplerproblem. Indeed, compared to the original problem (19), inwhich the first constraint is non-convex, the correspondingfirst constraint of (27) is convex. All the constraints and theobjective function of the problem (27) are convex except for theconstraint which is non-convex onlyin a single variable . It makes the problem (27) non-convexoverall. This single non-convex constraint can be rewritten as

where all the terms are linearwith respect to and except for the concave term of .The latter constraint can be handled iteratively by buildinga POTDC-type algorithm (see also [17], [18]) based on theiterative linear approximation of the non-convex termaround suitably selected points. It is interesting to mention thatthis iterative linear approximation can be also interpreted interms of the DC-iteration approach over the single non-convexterm . The fact that iterations are needed only over a singlevariable helps to reduce dramatically the number of iterations ascompared to the traditional DC-iteration approach and allowsfor simple algorithm shown below.

B. Iterative POTDC Algorithm

Let us consider the optimization problem (27) and replacethe term by its linear approximation around , i.e.,

. It leads to the following SDP problem

(28)

To demonstrate the POTDC algorithm graphically and also tosee how the linearization points are selected in different itera-tions, let us define the following OVF based on the optimizationproblem (28)

(29)

where in denotes the linearization point. The OVFcan be also obtained through in (25) by replacing

the term in with its linearapproximation around . Since and its linear approxima-tion have the same values at , and take the samevalues at this point. The following lemma establishes the rela-tionship between the OVFs and .Lemma 3: The OVF is a convex upper-bound offor any arbitrary , i.e., ,and is convex with respect to . Furthermore,

the OVFs and are directionally differentiable atthe point and the values of these OVFs as well as theirright and left derivatives are equal at . In other words,under the condition that is differentiable at , istangent to at the point .

Proof: See Appendix I-C.In what follows, we explain intuitively how the proposed

POTDC method works. For the sake of clarity, it is assumedin this explanation only that the OVF is differentiableover the interval , however, similar interpretation canbe made generally even for non-differentiable . Moreover,as we will see later, the differentiability of the OVF is notneeded for establishing the optimality results for the POTDCmethod.Let us consider an arbitrary point, denoted as ,

as an initial linearization point, i.e., . Based on Lemma3, is a convex function with respect to which is thetangent to at the linearization point , and it is alsoan upper-bound to . Let denote the global minimizerof that can be easily obtained due to the convexity of

with polynomial time complexity.Since is the tangent to at and it is

also an upper-bound for , it can be concluded that is adescent point for , i.e., as it is shown inFig. 1. Specifically, the fact that is the tangent toat and is the global minimizer of impliesthat

(30)

Furthermore, since is an upper-bound for ,. Due to the later fact and also the (30), it is

concluded that .Choosing as the linearization point in the second itera-

tion, and finding the global minimizer of over the in-terval denoted as , another descent point can be ob-tained, i.e., . This process is continued untilconvergence.The iterative descent method can be described as shown in

Algorithm 1. The following lemma about the convergence ofAlgorithm 1 and the optimality of the solution obtained by thisalgorithm is in order. Note that this lemma makes no assump-tions about the differentiability of the OVF .Lemma 4: The following statements regarding Algorithm 1

are true:i) The optimal value of the optimization problem in Algo-rithm 1 is non-increasing over iterations, i.e.,

Page 6: IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 61, NO. 23 ... · tively. The beamformer output at the time instant is given as (2) where is the complex beamforming vector of the antenna

6108 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 61, NO. 23, DECEMBER 1, 2013

Fig. 1. Iterative method for minimizing the OVF . The convex OVFis the upper-bound to which is tangent to it at , and its

minimum is denoted as . The point is used to establish another convexupper-bound function denoted as and this process continues.

ii) The sequence of the optimal values in Algorithm 1 con-verges. Note that the termination condition is not consid-ered for this statement.

iii) If Algorithm 1 converges (without considering termina-tion condition), such a limiting point is regular and it sat-isfies the KKT optimality conditions.Proof: See Appendix I-D.

Algorithm 1: The iterative POTDC algorithm

Require: An arbitrary , the terminationthreshold , set equal to 1.

repeat

Solve the following optimization problem usingto obtain and

and set

,

,

until

for .

Note that the termination condition in Algorithm 1, i.e.,

is used for stopping the algorithm when the value achieved isdeemed to be close enough to the optimal solution. The factthat the sequence of optimal values generated by Algorithm 1 isnon-increasing and convergent has been used for choosing thetermination condition. Despite its simplicity, this terminationcondition may stop the iterative algorithm prematurely. In orderto avoid this situation, one can define the termination conditionbased on the approximate satisfaction of the KKT optimalityconditions.The point obtained by Algorithm 1 is guaranteed to be the

global optimum of the problem considered if the OVFis a convex function of . It is also worth noting that even amore relaxed property of the OVF is sufficient to guar-antee global optimality. Specifically, if defined in (25) isa strictly quasi-convex function of , then it is stillguaranteed that we find the global optimum of the optimizationproblem (12) [29].The worst-case computational complexity of a general stan-

dard SDP problem can be expressed as ,where and denote, respectively, the number of constraintsand the number of variables of the standard SDP problem [30]and stands for the big-O (the highest order of complexity).The total number of variables in the SDP problem in Algorithm1, which includes the real and imaginary parts of and the realvariable , is equal to . The computational complexityof Algorithm 1 is equal to that of the SDP optimization problemin Algorithm 1, that is, thus , times the number of itera-tions (see also Simulation Example 1 in the next section).The RAB algorithm of [13] is iterative as well and its com-

putational complexity is equal to times the numberof iterations. The complexity of the RABs of [12] and [14]is . The comparison of the overall complexity of theproposed POTDC algorithm with that of the DC iteration-basedmethod is also performed in Simulation Example 4 in the nextsection. Although the computational complexity of the newproposed method may be slightly higher than that of some otherRABs, it finds the globally optimal solution as it is shown inSection IV-C. Moreover, it results in a superior performance asit is shown in Section V. Thus, next we show that under certainconditions the proposed POTDC method is guaranteed to findthe globally optimal solution of a reformulated optimizationproblem that corresponds to the general-rank RAB problem.

C. Global Optimality

For studying the conditions under which the proposedPOTDC method is guaranteed to find the global optimum of(12), we consider a reformulation of (12). Specifically, sincethis problem is feasible, it can be equivalently expressed as

(31)

Note that the constraint can be dropped asmaximizing the objective function in (31) implies that this con-straint is satisfied at the optimal point. By dropping this con-

Page 7: IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 61, NO. 23 ... · tively. The beamformer output at the time instant is given as (2) where is the complex beamforming vector of the antenna

KHABBAZIBASMENJ AND VOROBYOV: ROBUST ADAPTIVE BEAMFORMING 6109

straint, the problem (31) can be further expressed as the fol-lowing homogenous problem

(32)

Since (32) is homogenous, without loss of generality, the termcan be fixed to be equal to one. By doing so

and introducing the auxiliary variable , the problem (32) canbe equivalently rewritten as

(33)

where takes values in a closed interval as it is shown next.Specifically, the problem (33) is feasible if and only if

where ,, and stands for the smallest eigenvalue oper-

ator. In similar steps as in Section IV-A, i.e., by introducingand relaxing rank-one constraint, (33) can be

equivalently recast as

(34)

where the optimal solution of (33) can be extracted preciselyfrom the optimal solution of the problem (34). Thus, hereafterwe focus on the problem (34). This problem is a DC optimiza-tion problem which can be addressed using the POTDC algo-rithm. Specifically, the proposed POTDCmethod can be appliedto (34) by successively linearizing the term around suitablyselected points. Moreover, all the related results hold true in thiscase. In order to find the conditions which guarantee the globaloptimality of the POTDCmethod, let us introduce the followingOVF

(35)

Similar to the convexity proof for the OVF (see Lemma4), it can be easily verified that the OVF is concave. Basedon the definition of the OVF , the problem (34) can befurther simplified as

(36)

Note that since is a concave function, is also aconcave function and as a result, the objective function of theproblem (36) is the difference of two concave functions. The fol-lowing theorem shows when the problem (34), or equivalently,(36) is guaranteed to be solvable globally optimally by our pro-posed method.Theorem 1: For any arbitrary and whose

corresponding OVF is strictly concave and continuouslydifferentiable, and provided that is sufficiently small, the pro-posed POTDCmethod finds the globally optimal solution of theproblem (34), or equivalently, (36).

Proof: See Appendix I-E. The explicit condition for to besufficiently small is specified in the proof and it is not repeatedin the theorem formulation because of the space limitations.Note that the result of Theorem 1 also holds when OVF

is non-differentiable. In this case, the proof followssimilar steps, but it is slightly more technical and thereforeomitted because of the space limitations.Additionally, note that Theorem 1 does not imply that if

is not sufficiently small or the OVF is not continuouslydifferentiable, the POTDC method does not find the globallyoptimal solution. In other words, the condition of the theorem issufficient but not necessary. Indeed, according to our numericalresults, the globally optimal solution is always achieved by theproposed method.

D. Lower-Bounds on the Optimal Value of Problem (27)

We also aim at developing a tight lower-bound for the optimalvalue of the optimization problem (27). Such lower-bound isalso used for assessing the performance of the proposed iterativealgorithm.As it was mentioned earlier, although the objective func-

tion of the optimization problem (27) is convex, its feasibleset is non-convex due to the second constraint of (27). Alower-bound for the optimal value of (27) can be achieved byreplacing the second constraint of (27) by its correspondingconvex-hull. However, such lower-bound may not be tight. Inorder to obtain a tight lower-bound, we can divide the sector

into subsectors and solve the optimization problem(27) over each subsector in which the second constraint of (27)has been replaced with the corresponding convex hull. Theminimum of the optimal values of such optimization problemover the subsectors is the lower-bound for the problem (27).It is obvious that by increasing , the lower-bound becomestighter.

V. SIMULATION RESULTS

Let us consider a uniform linear array (ULA) of 10 omni-di-rectional antenna elements with the inter-element spacing ofhalf wavelength. Additive noise in antenna elements is mod-eled as spatially and temporally independent complex Gaussiannoise with zero mean and unit variance. Throughout all sim-ulation examples, it is assumed that in addition to the desiredsource, an interference source with the interference-to-noiseratio (INR) of 30 dB impinges on the antenna array. For ob-taining each point in the simulation examples, 100 independentruns are used unless otherwise is specified and the sample datacovariance matrix is estimated using snapshots.The new proposed method is compared in terms of the output

SINR to the general-rank RAB methods of [12]–[14] and tothe rank-one worst-case RAB of [2]. Moreover, the proposedmethod and the aforementioned general-rank RAB methods arealso compared in terms of the achieved values for the objectivefunction of the problem (12). The diagonal loading parametersof and are chosen for the proposedRAB and the RAB methods of [13] and [14], and the parame-ters of and are chosen for the RAB of [12].The initial point (see Algorithm 1) in the first iteration ofthe proposed method equals to unless otherwise is

Page 8: IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 61, NO. 23 ... · tively. The beamformer output at the time instant is given as (2) where is the complex beamforming vector of the antenna

6110 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 61, NO. 23, DECEMBER 1, 2013

Fig. 2. Example 1: Output SINR versus SNR;INR=30 dB and .

specified. The termination threshold for the proposed methodis chosen to be equal to . For obtaining a lower-bound onthe optimal value of the optimization problem (27), the interval

is divided in to 50 subsectors.

A. Simulation Example 1

In this example, the desired and interference sources are lo-cally incoherently scattered with Gaussian and uniform angularpower densities with central angles of 30 and 10 , respec-tively. The angular spreads of the desired and the interferingsources are assumed to be 4 and 10 , respectively. The pre-sumed knowledge of the desired source is different from the ac-tual one and is characterized by an incoherently scattered sourcewith Gaussian angular power density whose central angle andangular spread are 34 and 6 , respectively. Note that, the pre-sumed knowledge about the shape of the angular power densityof the desired source is correct while the presumed central angleand angular spread deviate from the actual one.In Figs. 2 and 3, the output SINR and the objective func-

tion values of the problem (12), respectively, are plotted versusSNR. It can be observed from the figures that the proposed newmethod based on the POTDC algorithm has superior perfor-mance over the other RABs. Moreover, Fig. 3 confirms that thenew proposed method achieves the global minimum of the opti-mization problem (12) since the corresponding objective valuecoincides with the lower-bound on the objective function ofthe problem (12). Fig. 4 shows the convergence of the iterativePOTDC method in terms of the average of the optimal valuefound by the algorithm over iterations for SNR=15 dB. It canbe observed that the proposed algorithm converges to the globaloptimum in about 4 iterations.

B. Simulation Example 2

In the second example, we study how the rank of the actualcorrelation matrix of the desired source affects the perfor-mance of the proposed general-rank RAB and other methods

Fig. 3. Example 1: Objective function value of the problem (12) versus SNR;INR=30 dB, and .

Fig. 4. Example 1: Objective function value of the problem (12) versus thenumber of iterations;SNR=15 dB, INR=30 dB, and .

tested. The same simulation set up as in the previous example isconsidered. The only difference is that the actual angular spreadof the desired source varies and so does the actual rank of thedesired source covariance matrix. The angular spread of the de-sired user is chosen to be 1 , 2 , 5 , 9 , and 14 . Figs. 5 and6 show, respectively, the output SINR and the objective func-tion values of the problem (12) versus the rank of the actualcorrelation matrix of the desired source for different methodswhen SNR=10 dB. It can be seen from the figures that the pro-posedmethod outperforms the other methods in all rank in termsof the objective value of the optimization problem (12) and it

Page 9: IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 61, NO. 23 ... · tively. The beamformer output at the time instant is given as (2) where is the complex beamforming vector of the antenna

KHABBAZIBASMENJ AND VOROBYOV: ROBUST ADAPTIVE BEAMFORMING 6111

Fig. 5. Example 2: Output SINR versus the actual rank of ; SNR=10 dB,INR=30 dB, and .

Fig. 6. Example 2: Objective function value of the problem (12) versus theactual rank of ; SNR=10 dB, INR=30 dB, and .

achieves the globally optimal solution as it coincides with thelower-bound.

C. Simulation Example 3

In this example, we also consider the locally incoherentlyscattered desired and interference sources. However, comparedto the previous examples, there is a substantial error in theknowledge of the desired source angular power density.The interference source is modeled as in the previous exam-

ples, while the angular power density of the desired source is

Fig. 7. Example 3: Actual and presumed angular power densities of general-rank source.

Fig. 8. Example 3: Output SINR versus SNR; INR=30 dB and .

assumed to be a truncated Laplacian function distorted by se-vere fluctuations. The central angle and the scale parameter ofthe Laplacian distribution is assumed to be 30 and 0.1, respec-tively, and it is assumed to be equal to zero outside of the interval[15 , 45 ] as it has been shown in Fig. 7. The presumed knowl-edge of the desired source is different from the actual one and ischaracterized by an incoherently scattered source with Gaussianangular power density whose central angle and angular spreadare 34 and 6 , respectively.Fig. 8 depicts the corresponding output SINR of the problem

(12) obtained by the beamforming methods tested versus SNR.It can be concluded from the figure that the proposed methodhas superior performance over the other methods.

Page 10: IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 61, NO. 23 ... · tively. The beamformer output at the time instant is given as (2) where is the complex beamforming vector of the antenna

6112 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 61, NO. 23, DECEMBER 1, 2013

TABLE IAVERAGE NUMBER OF THE ITERATIONS

TABLE IIAVERAGE CPU TIME

D. Simulation Example 4

Finally, we compare the efficiency of the proposed POTDCmethod to that of the DC iteration-based method that can bewritten for the problem (12) as

(37)

where denotes the inner product, stands for the gra-dient operator, and the function is replaced withthe first two terms of the Taylor expansion of around

. First, is initialized, and in the next iterations,is selected as the optimal obtained from solving (37) in theprevious iteration. Thus, the iterations are performed over thewhole vector of variables of the problem.The simulation set up is the same as in Simulation Example 1

except that different number of antennas is used. For a fair com-parison, the initial point in the proposed method and in(37) are chosen randomly. Particularly, the initialization pointfor the proposed POTDC method is chosen uniformly over theinterval while the imaginary and real parts of the initialvector in (37) are chosen independently as zero mean, unitvariance, Gaussian random variables. If the so-generatedis not feasible, another initialization point is generated and thisprocess continues until a feasible point is resulted. Note that thetime which is consumed during the generation of a feasible pointis negligible and it has not been considered in the average CPUtime comparison. Table I shows the average number of iterationstill convergence for the aforementioned methods versus the sizeof the antenna array. The termination threshold is set to ,

, and each number in the table is obtained by av-eraging the results over 200 runs. It can be seen from the tablethat the number of iterations for the proposed method is essen-tially fixed while it increases for the DC-iteration method as thesize of the array, and thus the size of the problem (12), increases.The latter phenomenon can be justified by considering the DC it-eration-type interpretation of the proposed method over the onedimensional OVF of . The dimension of is indepen-dent of the size of the array (thus, the size of the optimizationproblem), while the size of the search space over iterations forthe DC iteration-based method (37), that is, , increases asincreases. The average (over 200 runs) CPU time for the afore-mentioned methods is also shown in Table II. Both methodshave been implemented in Matlab using CVX software and runon the same PC with Intel(R) Core(TM)2 CPU 2.66 GHz.

Table II confirms that the proposed method is more efficientthan the DC iteration-based one in terms of the time required forconvergence. It is worth noting also that although the number ofvariables in the matrix of the optimization problem (28) isin general (since has to be a Hermitian matrix) after therank-one constraint is relaxed, the probability that the optimalhas rank one is very high as shown in [11], [31]–[33]. Thus,

in almost all cases, for different data sets, the actual dimensionof the problem (28) is . As a result, the average com-plexity of solving (28) is significantly smaller than the worst-case complexity, which is also guaranteed to be polynomial.

VI. CONCLUSION

We have considered the RAB problem for general-rank signalmodel with additional positive semi-definite constraint. SuchRAB problem corresponds to a non-convex DC optimizationproblem. We have studied this non-convex DC problem and de-signed the POTDC-type algorithm for solving it. It has beenproved that the point found by the POTDC algorithm for theRAB for general-rank signal model with positive semi-definiteconstraint is a KKT optimal point. Moreover, the problem con-sidered can be solved globally optimally under certain condi-tions. Specifically, we have proved that if the presumed norm ofthe mismatch that corresponds to the covariance matrix of thedesired source is sufficiently small, then the proposed POTDCmethod finds the globally optimal solution of the correspondingoptimization problem. The resulted RAB method shows supe-rior performance compared to the other existing methods interms of the output SINR and the resulted objective value. Italso has complexity that is guaranteed to be polynomial. Noneof the existing methods used for DC programming problemsguarantee that the global optimum can be found in polynomialtime, even under some conditions. Thus, the fundamental de-velopment of this work is the claim of global optimality and thefact that this claim boils down to convexity of the OVF (25).It implies that certain relatively simple DC programming prob-lems, which have been believed to be NP-hard, are actually notNP-hard under certain conditions.

APPENDIX

A. Proof of Lemma 1

The optimization problem (9) can be equivalently expressedas

(38)

First, note that based on the Cauchy-Schwarz inequality it can befound that . The latter implies thatunder the condition that , the norm of the vectoris always less than or equal to . Depending on whether thenorm of is greater than or smaller than , two differentcases are possible. First, let us consider the case that

. Then, by choosing as it is

Page 11: IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 61, NO. 23 ... · tively. The beamformer output at the time instant is given as (2) where is the complex beamforming vector of the antenna

KHABBAZIBASMENJ AND VOROBYOV: ROBUST ADAPTIVE BEAMFORMING 6113

guaranteed that and the matrix product be-comes equal to . The former can be verified simply asfollows

(39)

where the last inequality is due to the assumption that. By such , the objective value of the problem (9) be-

comes equal to its smallest non-negative value, i.e., zero.Next we consider the case when . Then,

choosing results in the vectorto be parallel to the vector . Since

(it can be verified by following similar steps as in (39)) and, as itwas discussed earlier, the norm of the vector is always lessthan or equal to , it can be concluded that is parallelto and it has the largest possible magnitude. In what fol-lows, we show that the optimal solution in this case is equal to. The following train of inequalities is in order

(40)

(41)

where the first inequality is due to the triangular inequality andthe second one is due to the Cauchy-Schwarz inequality. Since

is parallel to and it has the largest possible magnitude,the inequalities (40) and (41) are both active when , i.e.,the equality holds, and therefore, is the optimal solution.This completes the proof.

B. Proof of Lemma 2

First, we verify whether the optimal solution of the optimiza-tion problem (13), or equivalently, the following problem

(42)

is achievable or not.Let denote any arbitrary feasible point of the

problem (42). It is easy to see that if, then is greater than

or equal to . The latter implies that if theoptimal solution is achievable, it lies inside the sphere of

. Based on this fact,the optimization problem (42) can be recast as

(43)

The feasible set of the new constraint in (43) is bounded andclosed. Moreover, it can be easily shown that the feasible set ofthe constraint is also closed. Specifically,due to the fact that first constraint of the problem (43) is a sub-level set of the following continuous function

, its feasible set is closed [34]. Since both of the feasiblesets of the constraints are closed and one of them is bounded,the feasible set of the problem (43), which is the intersection ofthese two sets, is also closed and bounded. The latter implies thatthe feasible set of the problem (43) is compact. Therefore, alsobased on the fact that the objective function of (43) is continues,the optimal solution of (43), or equivalently (13), is achievable.Let denote the optimal solution of the problem

(13) and let us define the following auxiliary optimizationproblem

(44)

It can be seen that if is a feasible point of (44), then the pairis also a feasible point of (13), which implies that the

optimal value of (44) is greater than or equal to that of (13).However, since is a feasible point of (44) and the valueof the objective function at this feasible point is equal to theoptimal value of (13), i.e., it is equivalent to ,it can be concluded that both of the optimization problems (13)and (44) have the same optimal value.Let us define another auxiliary optimization problem based

on (44) as

(45)

which is obtained from (44) by dropping the last constraintof (44). The feasible set of (44) is a subset of the feasibleset of (45). Thus, the optimal value of (45) is smaller thanor equal to the optimal value of (44), and thus also, the op-timal value of (13). Using the maximin theorem [24], it iseasy to verify that .Since is smaller than or equal to the optimal value of(13), it is upper-bounded by , whereis an arbitrary feasible point of (13). The latter implies that

. Thiscompletes the proof.

C. Proof of Lemma 3

First, we prove that is a convex function with re-spect to . For this goal, let and denote the op-timal solutions of the optimization problems of and

, respectively, i.e.,

and , where and are any

two arbitrary points in the interval . It is trivial to verifythat is a feasible point of the corresponding

Page 12: IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 61, NO. 23 ... · tively. The beamformer output at the time instant is given as (2) where is the complex beamforming vector of the antenna

6114 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 61, NO. 23, DECEMBER 1, 2013

optimization problem of (see the defi-nition (29)). Therefore,

(46)

which proves that is a convex function with respect to.In order to show that is greater than or equal to, it suffices to show that the feasible set of the optimiza-

tion problem of is a subset of the feasible set of theoptimization problem of . Let denote a feasible pointof the optimization problem of , it is easy to verify that

is also a feasible point of the optimization problem ofif the inequality holds. Thisinequality can be rearranged as

(47)

and it is valid for any arbitrary . Therefore, is also a fea-sible point of the optimization problem of which impliesthat .In order to show that the right and left derivatives are equal,

we use the result of [35] which gives expressions for the direc-tional derivatives of a parametric SDP. Specifically, the direc-tional derivatives for the following OVF

(48)

are derived in [35], where and are a scalar andan matrix, respectively, is the real valued vectorof optimization variables, and is the real valued vectorof optimization parameters. Let be an arbitrary fixed point.If the optimization problem of poses certain properties,then according to [35, Theorem 10] it is directionally differ-entiable at . These properties are (i) the functionsand are continuously differentiable, (ii) the optimiza-tion problem of is convex, (iii) the set of optimal solu-tions of the optimization problem of denoted as isnonempty and bounded, (iv) the Slater condition for the op-timization problem of holds true, and (v) the inf-com-pactness condition is satisfied. Here inf-compactness conditionrefers to the condition of the existence of and acompact set such that

for all in a neighborhood of where denotesthe -dimensional Euclidean space. If for all the optimiza-tion problem of is convex and the set of optimal solutionsof is non-empty and bounded, then the inf-compactnessconditions holds automatically.The directional derivative of at in a directionis given by

(49)

where is the set of optimal solutions that corresponds tothe dual problem of the optimization problem of , and

denotes the Lagrangian defined as

(50)

where denotes the Lagrange multiplier matrix.Let us look again to the definitions of the OVFs and

(25) and (29), respectively, and define the followingblock diagonal matrix

(51)

as well as another block diagonal matrix denoted aswhich has exactly same structure as the ma-

trix with only difference that the elementin is replaced by

in .Then the OVFs and can be equivalently recast as

(52)

and

(53)

It is straightforward to see that the functions, , and are

continuously differentiable. Furthermore, it is easy to verifythat both optimization problems of and canbe expressed as

(54)

The problem (54) is convex and its solution set is non-empty andbounded. Indeed, let and denote two optimal solutionsof the problem above. The Euclidean distance between and

can be expressed as

(55)

where the last line is due to the fact that the matrix productis positive semi-definite and, therefore,

, and also the fact that for any arbitrary positive semi-definitematrix . From (55), it can be seen that thedistance between any two arbitrary optimal solutions of (54) isfinite and, therefore, the solution set is bounded. Moreover, it is

Page 13: IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 61, NO. 23 ... · tively. The beamformer output at the time instant is given as (2) where is the complex beamforming vector of the antenna

KHABBAZIBASMENJ AND VOROBYOV: ROBUST ADAPTIVE BEAMFORMING 6115

easy to verify that the dual problem for the optimization problem(54) can be expressed as

(56)

where and are the Lagrange multipliers. The optimiza-tion problem (54) is a convex SDP problem which satisfies theSlater’s conditions as the point is a strictlyfeasible point for its dual problem (56). Thus, the optimizationproblem (54) satisfies the strong duality. It can also be shownthat the inf-compactness condition is satisfied by verifying thatthe optimization problems of and are convex andtheir corresponding solution sets are bounded for any . There-fore, both of the OVFs and are directionally dif-ferentiable at .Using the result of [35, Theorem 10], the directional deriva-

tives of and can be respectively computed as

(57)and

(58)where and denote the optimal solution sets of the opti-mization problem (54) and its dual problem, respectively. Usingthe definitions of and , it can be seenthat the terms and are equalat and, therefore, the directional derivatives are equiva-lent. The latter implies that the left and right derivatives ofand are equal at .

D. Proof of Lemma 4

i) The optimization problem in Algorithm 1 at itera-tion is obtained by linearizing around

. Since and are feasiblefor the optimization problem at iteration , it can bestraightforwardly concluded that the optimal valueof the objective at iteration is less than or equalto the optimal value at the previous iteration, i.e.,

.ii) Since the sequence of the optimal values, i.e.,

is non-increasing andbounded from below (every optimal value is non-nega-tive), the sequence of the optimal values converges.

iii) The proof follows straightforwardly from [36, Proposi-tion 3.2]. Moreover, every feasible point of the problem(27) is a regular point. Specifically, if and denotea feasible point of the problem (27), the gradients of theequality and inequality constraints of this problem atand can be expressed, respectively, as

(59)

(60)

(61)

(62)

where denotes the vectorization operator,and denote, respectively, the real and imaginaryparts of a complex number. Note that only one of the con-straints or can be active. The gradients ,, and (or ) are linearly independent unless is

proportional to the identity matrix , i.e., whereis some coefficient. Therefore, assuming first that isnot proportional to the identity matrix, the linear inde-pendence constraint qualification (LICQ) holds at everyfeasible point. In the case when is proportional to theidentity matrix, the problem (12) can be expressed as

(63)

Then, the optimal solution of the problem (63) can be triv-ially obtained as pro-vided that . For the case when , the problem(63) is not feasible.

Since every point obtained in any iteration of Algorithm 1is a feasible point of the problem (27), the sequence generatedby Algorithm 1 is a sequence of all regular points. Thus, thissequence of regular points converges to a regular point and suchpoint satisfies the KKT optimality conditions.

E. Proof of Theorem 1

The OVF is continuously differentiable under the con-dition that the optimal solution of its corresponding optimiza-tion problem is unique [37]. Although the following proof is es-tablished based on the continuous differentiability assumptionof the OVF , it can be generalized to the case when theOVF is not continuously differentiable.Let , denote the optimal maximizer of the

concave OVF . Note that since the OVF is assumedto be strictly concave, its global maximizer, i.e., , is unique.Moreover, the strict concavity of the OVF together thefact that for any imply thatfor . If , then it is obvious that the function

is strictly decreasing over the intervalfor any arbitrary and, therefore, the POTDC methodfinds the globally optimal solution, i.e., . Next, we assumethat which implies thatdue to the strict concavity of . The optimal solution of theproblem (36) denoted as is less than or equal to . In orderto show it, let us assume that . Using the fact thatis the optimal solution of (36), it can be trivially concluded that

(64)

Page 14: IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 61, NO. 23 ... · tively. The beamformer output at the time instant is given as (2) where is the complex beamforming vector of the antenna

6116 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 61, NO. 23, DECEMBER 1, 2013

Since it was assumed that , the inequality (64) canbe rewritten as

(65)

From (65), it can be concluded that as-suming . The latter contradicts the fact that isthe global maximizer of the OVF and, thus, .Using the fact that , the problem (36) can be further

rewritten as

(66)

Let us now assume that while the following argumentscan be straightforwardly resulted also when . It is ob-vious that the function is strictly decreasing overthe interval and, therefore, the POTDCmethod will notbe stuck over this interval for any arbitrary .Since is the global maximizer of the OVF, we equivalently say that . In what follows,

we show that there exists such that for any the func-tion is strictly quasi-concave over the interval

. For this goal, we define the following new function

(67)

Note that since is continuously differentiable, iscontinuous and, therefore, is a continuous function. Basedon the definition of and due to the strict concavity of ,it can be concluded that the multiplicative termin the definition of is a strictly decreasing function on theinterval which approaches zero as approaches .Moreover, the term is equal to zero at the point

. The latter implies that the function is non-zeroover the interval , while at the point .Using the properties that the term is strictly

decreasing over the interval and the OVFis strictly concave together with the fact that the the term

approaches zero when , it is straight-forward to show that there exists , suchthat the function is strictly decreasing over the interval

. Moreover, using the strictly decreasing property ofthe term over the interval , it canbe easily concluded that for any

. Since the lower-bound of over the interval, i.e., , is less than and also

is a continuous function, there exists such that. The latter implies that

(68)

Based on the fact that is strictly decreasing over the intervaland, therefore, over the interval and also based

on (68), it can be concluded that for any ,

if , while if . Inother words,

(69)

and

(70)

which implies that by selecting equal to where, or equivalently, if , then the func-

tion is strictly quasi-concave, and therefore, thePOTDC method finds the globally optimal solution. This com-pletes the proof.

REFERENCES[1] A. B. Gershman, “Robust adaptive beamforming in sensor arrays,” J.

Electron. Commun., vol. 53, no. 6, pp. 305–314, Dec. 1999.[2] S. A. Vorobyov, A. B. Gershman, and Z.-Q. Luo, “Robust adaptive

beamforming using worst-case performance optimization: A solutionto the signal mismatch problem,” IEEE Trans. Signal Process., vol. 51,pp. 313–324, Feb. 2003.

[3] J. Li, P. Stoica, and Z. Wang, “On robust capon beamforming and di-agonal loading,” IEEE Trans. Signal Process., vol. 51, pp. 1702–1715,Jul. 2003.

[4] S. J. Kim, A. Magnani, A. Mutapcic, S. Boyd, and Z. Q. Luo, “Robustbeamforming via worst-case SINRmaximization,” IEEE Trans. SignalProcess., vol. 56, no. 6, pp. 1539–1547, Apr. 2008.

[5] S. A. Vorobyov, A. B. Gershman, Z.-Q. Luo, and N. Ma, “Adaptivebeamforming with joint robustness against mismatched signal steeringvector and interference nonstationarity,” IEEE Signal Process. Lett.,vol. 11, no. 2, pp. 108–111, Feb. 2004.

[6] J. Li, P. Stoica, and Z. Wang, “Doubly constrained robust capon beam-former,” IEEE Trans. Signal Process., vol. 52, pp. 2407–2423, Sep.2004.

[7] R. G. Lorenz and S. P. Boyd, “Robust minimum variance beam-forming,” IEEE Trans. Signal Process., vol. 53, pp. 1684–1696, May2005.

[8] S. A. Vorobyov, H. Chen, and A. B. Gershman, “On the relationshipbetween robust minimum variance beamformers with probabilistic andworst-case distortionless response constraints,” IEEE Trans. SignalProcess., vol. 56, pp. 5719–5724, Nov. 2008.

[9] A. Hassanien, S. A. Vorobyov, and K. M. Wong, “Robust adaptivebeamforming using sequential programming: An iterative solutionto the mismatch problem,” IEEE Signal Process. Lett., vol. 15, pp.733–736, 2008.

[10] A. Khabbazibasmenj, S. A. Vorobyov, and A. Hassanien, “Ro-bust adaptive beamforming via estimating steering vector based onsemidefinite relaxation,” in Proc. 44th Ann. Asilomar Conf. Signals,Syst., Comput., Pacific Grove, CA, Nov. 2010, pp. 1102–1106.

[11] A. Khabbazibasmenj, S. A. Vorobyov, and A. Hassanien, “Robustadaptive beamforming based on steering vector estimation with aslittle as possible prior information,” IEEE Trans. Signal Process., vol.60, pp. 2974–2987, Jun. 2012.

[12] S. Shahbazpanahi, A. B. Gershman, Z.-Q. Luo, and K. M. Wong,“Robust adaptive beamforming for general rank signal models,” IEEETrans. Signal Process., vol. 51, pp. 2257–2269, Sep. 2003.

[13] H. H. Chen and A. B. Gershman, “Robust adaptive beamforming forgeneral-rank signal models with positive semi-definite constraints,” inProc. IEEE ICASSP, Las Vegas, NV, Apr. 2008, pp. 2341–2344.

[14] C.W. Xing, S. D.Ma, and Y. C.Wu, “On low complexity robust beam-forming with positive semidefinite constraints,” IEEE Trans. SignalProcess., vol. 57, pp. 4942–4945, Dec. 2009.

[15] H. H. Chen and A. B. Gershman, “Worst-case based robust adaptivebeamforming for general-rank signal models using positive semi-defi-nite covariance constraint,” in Proc. IEEE ICASSP, Prague, Czech Re-public, May 2011, pp. 2628–2631.

[16] A. Khabbazibasmenj and S. A. Vorobyov, “A computationally efficientrobust adaptive beamforming for general-rank signal models with pos-itive semi-definiteness constraint,” in Proc. IEEE CAMSAP, San Juan,Puerto Rico, Dec. 2011, pp. 185–188.

Page 15: IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 61, NO. 23 ... · tively. The beamformer output at the time instant is given as (2) where is the complex beamforming vector of the antenna

KHABBAZIBASMENJ AND VOROBYOV: ROBUST ADAPTIVE BEAMFORMING 6117

[17] A. Khabbazibasmenj, S. A. Vorobyov, F. Roemer, and M. Haardt,“Polynomial-time DC (POTDC) for sum-rate maximization intwo-way AF MIMO relaying,” in Proc. IEEE ICASSP, Kyoto, Japan,Mar. 2012, pp. 2889–2892.

[18] A. Khabbazibasmenj, F. Roemer, S. A. Vorobyov, and M. Haardt,“Sum-rate maximization in two-way AF MIMO relaying: Polynomialtime solutions to a class of DC programming problems,” IEEE Trans.Signal Process., vol. 60, no. 10, pp. 5478–5493, Oct. 2012.

[19] A. Pezeshki, B. D. Van Veen, L. L. Scharf, H. Cox, and M. Lundberg,“Eigenvalue beamforming using a multi-rank MVDR beamformerand subspace selection,” IEEE Trans. Signal Process., vol. 56, pp.1954–1967, May 2008.

[20] R. Horst, P. M. Pardalos, and N. V. Thoai, Introduction to Global Op-timization. Dordrecht, The Netherlands: Kluwer Academic, 1995.

[21] R. Horst and H. Tuy, Global Optimization: Deterministic Ap-proaches. New York, NY, USA: Springer, 1996.

[22] J. Zhang, F. Roemer, M. Haardt, A. Khabbazibasmenj, and S. A.Vorobyov, “Sum rate maximization for multi-pair two-way relayingwith single-antenna amplify and forward relays,” in Proc. ICASSP,Kyoto, Japan, Mar. 2012, pp. 2477–2480.

[23] A. L. Yuille and A. Rangarajan, “The concave-convex procedure,”Neural Comput., vol. 15, pp. 915–936, 2003.

[24] S. Haykin, Adaptive Filter Theory, 3rd ed. Englewood Cliffs, NJ,USA: Prentice-Hall, 1995.

[25] A. Beck and Y. C. Eldar, “Strong duality in nonconvex quadratic op-timization with two quadratic constraints,” SIAM J. Optimiz., vol. 17,no. 3, p. 844860, 2006.

[26] Y. Huang andD. P. Palomar, “Rank-constrained separable semidefiniteprogramming with applications to optimal beamforming,” IEEE Trans.Signal Process., vol. 58, pp. 664–678, Feb. 2010.

[27] Y. Ye and S. Zhang, “New results on quadratic minimization,” SIAMJ. Optimiz., vol. 14, pp. 245–267, 2003.

[28] W. Ai, Y. Huang, and S. Zhang, “New results on Hermitian matrixrank-one decomposition,” Math. Program., pp. 253–383, Aug. 2009.

[29] M. S. Bazaraa, C. M. Shetty, and H. D. Sherali, Nonlinear Program-ming: Theory and Algorithms, 2nd ed. New York, NY, USA: Wiley,1993.

[30] Y. Nesterov and A. Nemirovsky, Interior Point Polynomial Algorithmsin Convex Programming. Philadelphia, PA, USA: SIAM, 1994.

[31] Z.-Q. Luo, W.-K. Ma, A. M.-C. So, Y. Ye, and S. Zhang, “Semidefiniterelaxation of quadratic optimization problems,” IEEE Signal Process.Mag., vol. 27, no. 3, pp. 20–34, May 2010.

[32] K. T. Phan, S. A. Vorobyov, N. D. Sidiropoulos, and C. Tellambura,“Spectrum sharing in wireless networks via QoS-aware secondarymulticast beamforming,” IEEE Trans. Signal Process., vol. 57, pp.2323–2335, Jun. 2009.

[33] Z.-Q. Luo, N. D. Sidiropoulos, P. Tseng, and S. Zhang, “Approxima-tion bounds for quadratic optimization with homogeneous quadraticconstraints,” SIAM J. Optimiz., vol. 18, no. 1, pp. 1–28, Feb. 2007.

[34] C. Udriste, “Convex functions and optimization methods on Rie-mannian manifolds,” in Series: Mathematics and Its Applica-tions. New York, NY, USA: Springer, 1994, vol. 297.

[35] A. Shapiro, “First and second order analysis of nonlinear semidefiniteprograms,” Math. Programm. Ser.B, vol. 77, pp. 301–320, 1997.

[36] A. Beck, A. Ben-Tal, and L. Tetruashvili, “A sequential parametricconvex approximation method with applications to nonconvex Trusstopology design problems,” J. Global Optimiz., vol. 47, no. 1, pp.29–51, 2010.

[37] D. Goldfarb and K. Scheinberg, “On parametric semidefinite program-ming,” Appl. Numer. Math., vol. 29, no. 3, pp. 361–377, 1999.

Arash Khabbazibasmenj (S’08) received the B.Sc.and M.Sc. degrees in electrical engineering (commu-nications) from Amirkabir University of Technology,Tehran, Iran, and the University of Tehran, Tehran,Iran, in 2006 and 2009, respectively, and the Ph.D.degree in electrical engineering (signal processing)from the University of Alberta, Edmonton, Alberta,Canada, in 2013.He is currently working as a Postdoctoral Fellow

in the Department of Electrical and Computer En-gineering, University of Alberta. During spring and

summer 2011, he was also a visiting student at Ilmenau University of Tech-nology, Germany. His research interests include signal processing and optimiza-tion methods in radar, communications, and related fields.Dr. Khabbazibasmenj is a recipient of the Alberta Innovates Graduate Award

in ICT.

Sergiy A. Vorobyov (M’02–SM’05) received theM.Sc. and Ph.D. degrees in systems and control fromKharkiv National University of Radio Electronics,Ukraine, in 1994 and 1997, respectively.He is a Professor with the Department of Signal

Processing and Acoustics, Aalto University, Finland,and is currently on leave from the Department ofElectrical and Computer Engineering, Universityof Alberta, Edmonton, Canada. He has been withthe University of Alberta as an Assistant Professorfrom 2006 to 2010, Associate Professor from 2010

to 2012, and Full Professor since 2012. Since his graduation, he also heldvarious research and faculty positions at Kharkiv National University ofRadio Electronics, Ukraine; the Institute of Physical and Chemical Research(RIKEN), Japan; McMaster University, Canada; Duisburg-Essen Universityand Darmstadt University of Technology, Germany; and the Joint ResearchInstitute between Heriot-Watt University and Edinburgh University, U.K.He has also held short-term visiting positions at Technion, Haifa, Israel andIlmenau University of Technology, Ilmenau, Germany. His research interestsinclude statistical and array signal processing, applications of linear algebra,optimization, and game theory methods in signal processing and communica-tions, estimation, detection and sampling theories, and cognitive systems.Dr. Vorobyov is a recipient of the 2004 IEEE Signal Processing Society Best

Paper Award, the 2007 Alberta Ingenuity New Faculty Award, the 2011 CarlZeiss Award (Germany), the 2012 NSERC Discovery Accelerator Award, andother awards. He served as an Associate Editor for the IEEE TRANSACTIONS ONSIGNAL PROCESSING from 2006 to 2010 and for the IEEE SIGNAL PROCESSINGLETTERS from 2007 to 2009. He was a member of the Sensor Array and Multi-Channel Signal Processing Committee of the IEEE Signal Processing Societyfrom 2007 to 2012. He is a member of the Signal Processing for Communi-cations and Networking Committee since 2010. He has served as the TrackChair for Asilomar 2011, Pacific Grove, CA, the Technical Co-Chair for IEEECAMSAP 2011, Puerto Rico, and the Tutorial Chair for ISWCS 2013, Ilmenau,Germany.