Channel Estimation for an Antenna Array
20230353426 · 2023-11-02
Inventors
- Yejian CHEN (Stuttgart, DE)
- Stefan Wesemann (Kornwestheim, DE)
- Jafar MOHAMMADI (Stuttgart, DE)
- Thorsten WILD (Stuttgart, DE)
Cpc classification
International classification
Abstract
A method of channel estimation for an antenna array is disclosed. The method includes, receiving a signal transmitted by the antenna array, obtaining a neural network model trained for channel estimation using the received signal, inputting a representation of the received signal into the neural network model and generating a channel estimate for the received signal, and deciding whether to employ a further neural network model for the channel estimation.
Claims
1. A method of channel estimation for an antenna array, the method comprising: receiving a signal transmitted by a first spatial dimension of the antenna array and by a second spatial dimension of the antenna array, obtaining a first neural network model trained for channel estimation for the first spatial dimension of the antenna array, and a second neural network model trained for channel estimation for the second spatial dimension of the array using the received signal, inputting a representation of the received signal for the first spatial dimension of the antenna array into the first neural network model and generating a channel estimate for the first spatial dimension of the array, and inputting a representation of the received signal for the second spatial dimension of the antenna array into a second neural network model and generating a channel estimates for the second spatial dimension of the antenna array, combining the channel estimates for the first spatial dimension of the antenna array and for the second spatial dimension of the antenna array to generate a channel estimate for the antenna array, and deciding whether to employ a further neural network model for the channel estimation.
2. The method of claim 1, comprising: deciding to employ a further neural network for the channel estimation, obtaining a further neural network model trained for channel estimation using the channel estimate generated by the neural network model, and inputting the channel estimate into the further neural network model to generate a further channel estimate for the received signal.
3. The method of claim 1, wherein deciding whether to employ a further neural network model for the channel estimation comprises determining a computational capacity for performing operations of a further neural network model.
4. The method of claim 1, wherein obtaining the neural network model comprises training a neural network model for channel estimation using labelled input data, storing parameters of the trained neural network model in memory, and retrieving the stored parameters from the memory.
5. The method of claim 2, wherein obtaining the further neural network model comprises training a further neural network model using a channel estimation of the neural network model and a label of the input data, storing parameters of the trained further neural network in memory, and retrieving the stored parameters from the memory.
6. The method of claim 1, wherein inputting a representation of the received signal into the neural network model comprises inputting a spatial or sample covariance matrix representation of the received signal.
7. The method of claim 1, wherein obtaining the neural network model comprises training a neural network model for channel estimation using labelled input data corresponding to a plurality of spatially different spatial positions, storing parameters of the trained neural network model in memory, and retrieving the stored parameters from the memory.
8. The method of claim 1, wherein combining the channel estimates for the first spatial dimension of the antenna array and for the second spatial dimension of the antenna array comprises computing an arithmetic mean or a geometric mean of the channel estimates.
9. The method of claim 1, wherein the first and second spatial dimensions are horizontal and vertical dimensions respectively of the two-dimensional antenna array.
10. The method of claim 1, wherein obtaining a first neural network model trained for channel estimation for the first spatial dimension of the antenna array, and a second neural network model trained for channel estimation for the second spatial dimension of the array, comprises training first and second neural network models respectively for channel estimation using labelled training data.
11. The method of claim 1, wherein obtaining a first neural network model trained for channel estimation for the first spatial dimension of the antenna array, and a second neural network model trained for channel estimation for the second spatial dimension of the array, comprises training first and second neural network models respectively for channel estimation using labelled training data corresponding to a plurality of spatially different spatial positions.
12. The method of claim 1, further comprising receiving a signal transmitted by a third spatial dimension of the antenna array and by a fourth spatial dimension of the antenna array, obtaining a third neural network model trained for channel estimation for the third spatial dimension of the antenna array, and a fourth neural network model trained for channel estimation for the fourth spatial dimension of the array, inputting a representation of the received signal for the third spatial dimension of the antenna array into the third neural network model and generating a channel estimate for the third spatial dimension of the array, and inputting a representation of the received signal for the fourth spatial dimension of the antenna array into the fourth neural network model and generating a channel estimate for the fourth spatial dimension of the antenna array, and combining the channel estimates for each of the first to the fourth spatial dimensions of the antenna array to generate a channel estimate for the antenna array.
13. The method of claim 12, wherein the third and fourth spatial dimensions are mutually different diagonal dimensions of the two-dimensional antenna array.
14. A non-transitory computer readable medium comprising a computer program comprising instructions, which, when executed with an apparatus, cause the apparatus to perform the method of claim 1.
15-16. (canceled)
17. The apparatus as in claim 20 where the apparatus comprises the antenna array.
18. The apparatus as in claim 17 where the apparatus comprises a base station for a wireless communication network, the base station comprising the antenna array, where the base station and the antenna array are configured for wirelessly communicating with remote user equipment.
19. The apparatus as in claim 18 where the apparatus comprises a wireless communication network, the network comprising the base station and the remote user equipment in wireless communication with the base station via the antenna array.
20. An apparatus comprising: at least one processor; and at least one non-transitory memory storing instructions that, when executed with the at least one processor, cause the apparatus to perform: receiving a signal transmitted by a first spatial dimension of an antenna array and by a second spatial dimension of the antenna array, obtaining a first neural network model trained for channel estimation for the first spatial dimension of the antenna array, and a second neural network model trained for channel estimation for the second spatial dimension of the array using the received signal, inputting a representation of the received signal for the first spatial dimension of the antenna array into the first neural network model and generating a channel estimate for the first spatial dimension of the array, and inputting a representation of the received signal for the second spatial dimension of the antenna array into a second neural network model and generating a channel estimates for the second spatial dimension of the antenna array, combining the channel estimates for the first spatial dimension of the antenna array and for the second spatial dimension of the antenna array to generate a channel estimate for the antenna array, and deciding whether to employ a further neural network model for the channel estimation.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
[0036] In order that the present invention may be more readily understood, embodiments of the invention will now be described, by way of example, with reference to the accompanying drawings, in which:
[0037]
[0038]
[0039]
[0040]
[0041]
[0042]
[0043]
[0044]
[0045]
[0046]
[0047]
[0048]
[0049]
DETAILED DESCRIPTION OF THE DISCLOSURE
[0050] A wireless communication network 101 embodying an aspect of an invention of the present disclosure is illustrated schematically in the Figures.
[0051] Referring firstly to
[0052] Referring next to
[0053] Computer 201 embodying an aspect of the invention comprises central processing unit 202, flash memory 203, random-access memory 204, input/output interface 205, and system bus 206. The computer 201 is configured to run neural network models for estimation of channel in communications between the base station 102 and the handset 103.
[0054] Central processing unit 202 is configured for execution of instructions of a computer program. Flash memory 203 is configured for non-volatile storage of computer programs for execution by the central processing unit 202. Random-access memory 204 is configured as read/write memory for storage of operational data associated with computer programs executed by the central processing unit 202. Input/output interface 205 is provided for connection of external computing devices and/or other peripheral hardware to computer 201, to facilitate control of the computer 201 and inputting of input data. The components 202 to 205 of the computer 201 are in communication via system bus 206.
[0055] In the embodiment, the flash memory 203 has a computer program for channel estimation using neural network models stored thereon. The computer 201 is thus configured, in accordance with the instructions of the computer program, to train neural network models for estimating channel in signals transmitted between the antenna array 205 of the base station 202 and the plural user equipment in communication with the base station 202, such as handset 103, sample signals received by the user equipment, such as handset 103, from the base station 202, and process the sampled signal on the central processing unit 202 using the trained neural network models to thereby generate one or more channel estimates predictions for the received signal. The computer 201 is then configured to transmit the channel estimations to a signal decoder of the handset 103 in order to allow correct recovery of the transmitted signal ‘X’ from the received signal ‘Y’ by compensating for the channel ‘H’.
[0056] Referring in particular to
[0057] At stage 301, the computer program causes the central processing unit 202 to train neural network models for channel estimation, and store the trained neural network models in the flash memory 203.
[0058] At stage 302, the computer program causes the central processing unit 202 to generate channel estimations for a signal transmitted by the base station 102 and received by the handset 103 using the neural network models trained at stage 301.
[0059] Referring in particular to
[0060] At stage 401, labelled training data is obtained for channel estimation of signals transmitted by the base station 102 to the handset 103.
[0061] At stage 401, the labelled training data is reshaped vertically and horizontally for vertical and horizontal spatial dimensions respectively of the antenna array 105 of the base station 102.
[0062] At stage 402, the labelled training data is input into parallelised first and second neural network models.
[0063] A stage 403, the first and second neural network models are executed on the input to thereby generate channel estimates H.sub.v, H.sub.h for vertical and horizontal spatial dimensions respectively of the antenna array 105.
[0064] At stage 404, parameters, such as weights, of the first and second neural network models respectively are stored in flash memory 203 of the computer 201.
[0065] At stage 405, the channel estimates H.sub.v, H.sub.h are combined to obtain at stage 406 a channel estimate H for the signal received by the handset 103.
[0066] At stage 407, the channel estimates H are compared to the corresponding label of the input labelled training data, to determine a magnitude of an error in the channel estimates H compared to the label. A determination is then made as to whether the accuracy of the channel estimates H satisfy a threshold accuracy value. For example, the threshold accuracy value could be a value manually input by a user of the system denoting a desired accuracy of channel estimates obtained during the inference stage 302. If the determination a stage 407 is answered in the negative, indicating that the channel estimates H are sufficiently accurate, the method proceeds to termination at stage 408.
[0067] In the alternative, if the determination at stage 407 is answered in the affirmative, indicating that channel estimates of the first and second neural network models are insufficiently accurate, the weights of the first and second neural network models are updated, and the channel estimates H obtained by stage 406 are output and substituted in place of the corresponding observations of the labelled training data input at stage 402. The method of stages 402 to 407 is then repeated, using the updated weights for the first and second neural network models, by inputting the channel estimates H into the updated first and second neural network models.
[0068] Stages 402 to 407 may then be performed repeatedly, i.e. iteratively, until the determination at stage 407 is finally answered in the negative, at which time the training phase is ended at stage 408.
[0069] Referring in particular to
[0070] In the embodiment, the Kronecker covariance model is employed to perform training of the neural networks for the vertical and horizontal spatial dimensions respectively. With the Kronecker model, the spatial covariance matrix of a two-dimensional antenna array may be approximated as the Kronecker product of a vertical covariance matrix and a horizontal covariance matrix. The training for the full spatial covariance matrix of a two-dimensional array, denoted as highly complex brute-force training, can thus be replaced by low complexity subspace training with two neural networks in horizontal and vertical spatial domains by separating the three-dimensional channel into azimuth and elevation dimensions and treating the dimensions as independent two-dimensional channels. This thus achieves a complexity cost saving factor. The two subspace channel estimates may then be combined e.g. simply-averaged, to obtain channel estimates. Furthermore, as a consequence of treating array dimensions independently, we obtain for each antenna element two estimates, one from the horizontal estimator and one from the vertical estimator. This additional horizontal/vertical combining gain can be expected to improve the accuracy of the channel estimation.
[0071] Thus, referring in particular to
[0072] Thus, referring next in particular to
H.sub.h=[h.sub.1.sup.(h) . . . h.sub.M.sup.(h)] Equation 1:
H.sub.v=[h.sub.1.sup.(v) . . . h.sub.N.sup.(v)], Equation 2:
[0073] Considering MMSE channel estimation for both spatial dimensions, i.e. subspaces individually, matrices H.sub.h and H.sub.v may thus be given by equations 3 and 4 respectively, where W.sub.h and W.sub.v denote N×N and M×M MMSE weighting matrices, and the channel observation is denoted as M×N matrix Y.
H.sub.h=W.sub.hY.sup.T=R.sub.h(R.sub.h+σ.sup.2I.sub.N).sup.−1Y.sup.T Equation 3:
H.sub.v=W.sub.vY=R.sub.v(R.sub.v+σ.sup.2I.sub.M).sup.−1Y Equation 4:
[0074] Thus, the arithmetic mean of the independent channel estimates of the vertical and horizontal spatial dimensions may be given by equation 5.
Ĥ.sub.a=0.5(H.sub.h.sup.T+H.sub.v)=0.5(W.sub.vY+YW.sub.h.sup.T). Equation 5:
[0075] Equation 5 may be reformulated as the vectorized expression of equation 6, where the MN×MN matrix of equation 7 is the effective weighting with respect to an arithmetic mean of the subspace channel estimates.
vec(Ĥ.sub.a)=0.5(I.sub.N.Math.W.sub.v+W.sub.h.Math.I.sub.M)vec(Y),Equation 6:
W.sub.a=0.5(I.sub.N.Math.W.sub.v+W.sub.h.Math.I.sub.M Equation 7:
[0076] It is alternatively possible to consider the geometric mean of the subspace channel estimates, which may be given by equation 8, where the MN×MN matrix W.sub.g given by equation 9 is the effective weighting with the geometric mean.
vec(Ĥ.sub.g)=(I.sub.N.Math.W.sub.v).sup.0.5(W.sub.h.Math.I.sub.M).sup.0.5vec(Y)=W.sub.h.sup.0.5.Math.W.sub.v.sup.0.5)vec(Y), Equation 8:
W.sub.g=W.sub.h.sup.0.5.Math.W.sub.v.sup.0.5 Equation 9:
[0077] Thus, the geometric mean of the independent channel estimates of the vertical and horizontal spatial dimensions may be given by equation 10.
Ĥ.sub.g=W.sub.v.sup.0.5Y(W.sub.h.sup.0.5).sup.T. Equation 10:
[0078] The equations indicate that the geometric mean can provide the channel estimates to a greater accuracy. However, in computing the geometric mean
, the matrix square root operation (⋅)0.5 has to be invoked for matrices W.sub.v and W.sub.h, which introduces additional complexity due to the non-linear operation.
[0079] Referring next in particular to
[0080] In the example of
[0081] Referring next to
[0082] The subspace training may then be carried out for training dataset A. The neural network models of horizontal and vertical subspaces may be passed to training dataset B for de-noising. Subsequently a new subspace training may be started with the de-noised training dataset B. Similarly, the neural network models of subspaces will be passed to training dataset A. This constitutes one iteration. The training weights are delivered from one iteration to the next iteration. This introduces independent subspace diversity to increase the effective SNR for the training in the next iteration.
[0083] Referring next to
[0084] At stage 1001, observations are made of a signal transmitted by the antenna array 105 of the base station 102.
[0085] At stage 1002, a determination of a number of channel estimation iterations to be performed is made. In an example, the determination at stage 1102 takes account of an available computational capacity of CPU 202 of computer 201 for performing computational operations. If the network is under high load, e.g. when a large number of user equipment is served, and thus many instances of channel estimation are required, equation 11 holds.
n.sub.UE×n.sub.iteration=C.sub.SoC Equation 11:
[0086] Where, n.sub.UE is the number of active UEs, n.sub.iteration is the number of iterations to be performed, and C.sub.SoC is the computational capacity of the CPU 202 available for channel estimation tasks. From equation 11, the algorithm adaptively adjusts the number of channel estimation iterations to the load of the network. This requirement could be stated as an input that lets the algorithm limit the number of iterations according to Equation 11:
n.sub.iteration=C.sub.SoC/n.sub.UE. Equation 11:
[0087] After being combined in horizontal and vertical subspaces, the channel estimates still exhibit residual additive noise, which is Gaussian but with reduced variance, compared to the noise in the channel observation input. The key idea of the inference method depicted schematically in
[0088] The determination at stage 1002 may thus allow for ready adaptation of the complexity/accuracy of the method in dependence on an available computational capacity for performing computational operations of the iterations.
[0089] At stage 1003, the vertical neural network and horizontal network models for the i-th iteration, as generated during the training phase 301, are loaded.
[0090] At stage 1004, the noisy observation of the received signal is input into the first and second neural networks, and channel estimates for the vertical and horizontal spatial dimensions respectively are obtained.
[0091] At stage 1005, the channel estimates for the vertical and horizontal spatial dimensions obtained at stage 1004 are combined, for example, by computing an arithmetic mean value using Equation 5, to thereby generate a channel estimate at stage 1006.
[0092] At stage 1007, a determination is made as to whether further iterations of stages 1003 to 1006 are required, by reference to the determination of the required number of iteration obtained at stage 1002. If the determination is answered in the negative indicating that further iterations are not required, the method proceeds to termination at stage 1008.
[0093] In the alternative, if the determination at stage 1007 is determined in the affirmative, indicating that further iterations are required, the weights of the neural networks models are updated to weights corresponding to a second iteration as learned at stage 301, and channel estimate is input into the updated neural networks, and stages 1003 to 1007 are repeated until the determination at stage 1007 is answered in the negative, indicating that all iterations prescribed by the determination at stage 1002 have been performed, and the channel estimate may then be output to a signal decoder of the handset 103.
[0094] Referring finally to
[0095] Thus, in the modification to the method, in addition to training of horizontal and vertical neural networks, two additional diagonal spatial dimension neural networks may be similarly trained at stage 301, and deployed at inference stage 302 in parallel with the first and second neural networks. As indicated in the Figure, in the example the (8-2)-th antenna element will thus be processed in vertical subspace by the first neural network, horizontal subspace by the second neural network, and mutually different diagonal subspaces by the third and fourth neural networks. The use of the two additional neural networks for the diagonal dimensions is motivated by the recognition that the decomposition of the full-dimensional (spatial) channel covariance matrix into the Kronecker-product of horizontal and vertical spatial covariance matrices is only an approximation. The spatial correlation information that is lost by this approximation may be included in the channel estimation procedure using the diagonal array dimensions.
[0096] Referring next to
[0097] Referring firstly in particular to
[0098] This semi-universality of the neural network models has two key advantages compared to the neural network models described with reference to
[0099] Referring next to
[0100] Referring finally to
[0101] In this modification to the treatment of a received signal, the high-dimensional data (observations and labels) structures may be reconstructed to lower-dimensional, two-dimensional, subspaces used as the inputs of a neural network model, which may be represented as sample covariance matrices of the received signal. After training in these lower-dimensional subspaces successively (with or without combining the channel estimates for the subspaces), the noise variance of the original high-dimensional data will be iteratively improved.
[0102] Although the present invention and its advantages have been described in detail, it should be understood that various changes, substitutions and alterations can be made herein without departing from the spirit and scope of the invention as defined by the appended claims. In the claims, the word “comprising” does not exclude other elements or steps, and the indefinite article “a” or “an” does not exclude a plurality.