Fast Gradient-Free Optimization in Distributed Multi-User MIMO Systems

06/09/2020
by   Olivier Bilenne, et al.
0

In this paper, we develop a gradient-free optimization methodology for efficient resource allocation in Gaussian MIMO multiple access channels. Our approach combines two main ingredients: (i) the entropic semidefinite optimization method of matrix exponential learning (MXL); and (ii) a one-shot gradient estimator which achieves low variance through the reuse of past information. Owing to this reuse mechanism, the proposed gradient-free MXL algorithm with callbacks (MXL0^+) retains the convergence speed of gradient-based methods while requiring minimal feedback per iteration-a single scalar. In more detail, in a MIMO multiple access channel with K users and M transmit antennas per user, the MXL0^+ algorithm achieves ϵ-optimality within poly(K,M)/ϵ^2 iterations (on average and with high probability), even when implemented in a fully distributed, asynchronous manner. For cross-validation, we also perform a series of numerical experiments in medium- to large-scale MIMO networks under realistic channel conditions. Throughout our experiments, the performance of MXL0^+ matches-and sometimes exceeds-that of gradient-based MXL methods, all the while operating with a vastly reduced communication overhead. In view of these findings, the MXL0^+ algorithm appears to be uniquely suited for distributed massive MIMO systems where gradient calculations can become prohibitively expensive.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset