Integrating volterra series model and deep neural networks to equalize nonlinear power amplifiers

11451419 · 2022-09-20

Assignee

Inventors

Cpc classification

International classification

Abstract

The nonlinearity of power amplifiers (PAs) has been a severe constraint in performance of modern wireless transceivers. This problem is even more challenging for the fifth generation (5G) cellular system since 5G signals have extremely high peak to average power ratio. Non-linear equalizers that exploit both deep neural networks (DNNs) and Volterra series models are provided to mitigate PA nonlinear distortions. The DNN equalizer architecture consists of multiple convolutional layers. The input features are designed according to the Volterra series model of nonlinear PAs. This enables the DNN equalizer to effectively mitigate nonlinear PA distortions while avoiding over-fitting under limited training data. The non-linear equalizers demonstrate superior performance over conventional nonlinear equalization approaches.

Claims

1. A distortion-compensating processor, comprising: an input port configured to receive a distorted signal y(n) representing information x(n) distorted by a channel h; at least one automated processor, configured to: decompose the distorted signal y(n) as a Volterra series expansion y ( n ) = .Math. d = 0 D .Math. k = 0 P b k d x ( n - d ) .Math. "\[LeftBracketingBar]" x ( n - d ) .Math. "\[RightBracketingBar]" k - 1 ;  and implement an equalizer operating on separate terms of the Volterra series expansion to produce an output z(n) having reduced distortion with respect to the distorted signal y(n), the equalizer comprising a deep neural network comprising a plurality of neural network hidden layers trained with respect to distortion of the channel h to provide an input-output response and z ( n ) = .Math. k = 1 P .Math. d 1 = 0 D .Math. .Math. d k = 0 D f d 1 , .Math. , d k .Math. i = 1 k r ( n - d i ) ; and an output port configured to present the output, wherein: r(n) is a response of the channel h to the nonlinearly distorted radio frequency signal y(n); p is a respective nonlinearity order; d is a memory depth parameter; D is a total memory length; P is a total nonlinearity order; k is a nonlinear order; and b.sub.kd are nonlinear response parameters.

2. The distortion-compensating processor according to claim 1, wherein the distorted signal is received from a radio receiver.

3. The distortion-compensating processor according to claim 1, wherein the distorted signal is distorted by amplification by a radio frequency power amplifier and transmission through a radio frequency communication channel.

4. The distortion-compensating processor according to claim 1, wherein: the undistorted radio frequency signal x(n) is distorted by an analog process to produce the nonlinearly distorted radio frequency signal y(n), which passes through the channel h; r ( n ) = .Math. = 0 L h y ( n - ) + v ( n ) is a response signal sequence, wherein r(n) is stacked together into M+1 dimensional vectors r(n)=[r(n), . . . , r(n−M)].sup.T, where (⋅).sup.T denotes transpose, such that r(n)=HG(n)×(n)+v(n); custom character is a set of finite-impulse response channel coefficients; custom character is an equalization delay; v(n) is an additive white Gaussian noise component signal sequence; H is an (M+1)×(M+L+1) dimensional channel matrix H = [ h o .Math. h L h o .Math. h L ] ; G(n)=diag{V.sub.y(n)e.sup.jψ.sup.y(n), . . . , V.sub.n(y-M-L)e.sup.jψ.sup.y(n-M-L)} is an (M+L+1)×(M+L+1) dimensional diagonal matrix of the nonlinear responses,
x(n)=[x(n), . . . ,x(n−M−L)].sup.T, and
v(n)=[v(n), . . . ,v(n−M)].sup.T; f.sup.T=G′(n)[ƒ.sub.0, . . . , ƒ.sub.M], where [ƒ.sub.0, . . . , ƒ.sub.M]H≈[0, . . . , 1, . . . , 0] is computed by the equalizer to equalize the channel h; G ( n ) = 1 V y ( n - d ) e - j ψ y ( n - d ) is computed by the equalizer to equalize the analog process; {circumflex over (r)}(n) is a resulting sequence after linear channel equalization; z(n)=ƒ.sup.T r(n)≈x(n−d) represents the output with equalization delay d; and z ( n ) = .Math. d = 0 D .Math. p = 0 P g k d r ^ ( n - d ) .Math. "\[LeftBracketingBar]" r ^ ( n - d ) .Math. "\[RightBracketingBar]" k - 1 is a Volterra series model of the equalizer with coefficients g.sub.kd such that z(n)≈x(n−custom character) for some equalization delay custom character, wherein: coefficients g.sub.kd are determined according to min { g kd } .Math. n = L N .Math. "\[LeftBracketingBar]" x ( n - L ) - .Math. d = 0 D .Math. k = 1 P g k d r ^ ( n - d ) .Math. "\[LeftBracketingBar]" r ^ ( n - d ) .Math. "\[RightBracketingBar]" k - 1 .Math. "\[RightBracketingBar]" 2 with training symbols x(n) and received samples {circumflex over (r)}(n), z(n)=arg min.sub.∀x(n)|{circumflex over (r)}(n)−V.sub.ye.sup.jψ.sup.yx(n)|.sup.2 is a maximum likelihood estimation operator for a nonlinear equalization output signal sequence having reduced distortion, V.sub.y is an amplitude of a signal sequence y(n), ψ.sub.y is a phase change of the signal sequence y(n); and the deep neural network equalizer is trained to determine the channel coefficients custom character, analog process responses V.sub.y, ψ.sub.y, and the channel equalizer ƒ.sup.T.

5. The distortion-compensating processor according to claim 4, wherein: vector a = [ g 0 0 , g 0 1 , .Math. , g P D ] T , vector x = [ x ( 0 ) ; .Math. , x ( N - L ) ] T , vector B = [ r ^ ( L ) r ^ ( L ) .Math. "\[LeftBracketingBar]" r ^ ( L ) .Math. "\[RightBracketingBar]" .Math. r ^ ( L - D ) .Math. "\[LeftBracketingBar]" r ^ ( L - D ) .Math. "\[RightBracketingBar]" P - 1 .Math. .Math. r ^ ( N ) r ^ ( N ) .Math. "\[LeftBracketingBar]" r ^ ( N ) .Math. "\[RightBracketingBar]" .Math. r ^ ( N - D ) .Math. "\[LeftBracketingBar]" r ^ ( N - D ) .Math. "\[RightBracketingBar]" P - 1 ] , and min { g kd } .Math. n = L N .Math. "\[LeftBracketingBar]" x ( n - L ) - .Math. d = 0 D .Math. k = 1 P g k d r ^ ( n - d ) .Math. "\[LeftBracketingBar]" r ^ ( n - d ) .Math. "\[RightBracketingBar]" k - 1 .Math. "\[RightBracketingBar]" 2 min a .Math. x - Ba .Math. 2 , having solution a=B.sup.+x, where B.sup.+=(B.sup.HB).sup.−1B is the pseudo-inverse of the matrix B.

6. The distortion-compensating processor according to claim 1, wherein the Volterra series expansion comprises at least fifth order terms, and the deep neural network comprises at least two convolutional network layers.

7. The distortion-compensating processor according to claim 1, wherein the deep neural network comprises at least three one-dimensional layers, each layer having at least 10 feature maps, and a fully connected layer subsequent to the at least three layers.

8. The distortion-compensating processor according to claim 1, wherein the distorted signal comprises a transmitted orthogonal frequency multiplexed radio frequency signal.

9. The distortion-compensating processor according to claim 1, further comprising a demodulator, configured to demodulate the output as the set of symbols representing the information.

10. A method of compensating for a distortion, comprising: receiving a distorted signal y(n) representing information x(n) distorted and communicated through a channel h; decomposing the distorted signal y(n) with a Volterra processor to produce a Volterra series expansion of form y ( n ) = .Math. d = 0 D .Math. k = 0 P b k d x ( n - d ) .Math. "\[LeftBracketingBar]" x ( n - d ) .Math. "\[RightBracketingBar]" k - 1 ; and equalizing the distorted signal y(n) with an automated nonlinear equalizer, comprising a deep neural network having a plurality of neural network hidden layers trained with respect to distortion of the channel h to have an input-output response z ( n ) = .Math. k = 1 P .Math. d 1 = 0 D .Math. .Math. d k = 0 D f d 1 , .Math. , d k .Math. i = 1 k r ( n - d i ) , which receives the Volterra series expansion, and produces an output z(n) having reduced distortion with respect to the distorted signal y(n), wherein: r(n) is a response of channel h to the nonlinearly distorted radio frequency signal y(n); p is a respective nonlinearity order; d is a memory depth parameter; D is a total memory length; P is a total nonlinearity order; k is a nonlinear order; and b.sub.kd are nonlinear response parameters.

11. The method according to claim 10, wherein the input comprises a radio frequency orthogonal frequency multiplexed signal amplified and distorted by a radio frequency power amplifier, received though a radio receiver.

12. The method according to claim 10, further comprising: computing a maximum likelihood estimation for a nonlinear equalization z(n)=arg min.sub.∀x(n)|({circumflex over (r)}(n)−V.sub.ye.sup.jψ.sup.yx(n)|.sup.2, to produce the output z(n)=ƒ.sup.T r(n)≈x(n−d), with equalization delay d; training the deep neural network to determine the channel coefficients custom character of the channel h, having a channel amplitude response V.sub.y, a channel phase response ψ.sub.y, and a channel equalization response ƒ.sup.T; approximating G′(n) with a Volterra series model z ( n ) = .Math. d = 0 D .Math. k = 0 P g k d r ^ ( n - d ) .Math. "\[LeftBracketingBar]" r ^ ( n - d ) .Math. "\[RightBracketingBar]" k - 1 , with coefficients g.sub.kd designed such that z(n)≈x(n−custom character) for equalization delay custom character, and estimated according to min { g kd } .Math. n = L N .Math. "\[LeftBracketingBar]" x ( n - L ) - .Math. d = 0 D .Math. k = 1 P g k d r ^ ( n - d ) .Math. "\[LeftBracketingBar]" r ^ ( n - d ) .Math. "\[RightBracketingBar]" k - 1 .Math. "\[RightBracketingBar]" 2 , with training symbols x(n) and received samples {circumflex over (r)}(n); wherein: channel h, produces response r ( n ) = .Math. = 0 L h y ( n - ) + v ( n ) from y(n); custom character is a set of finite-impulse response channel coefficients; custom character is an equalization delay; v(n) is an additive white Gaussian noise component; r(n) is stacked together into M+1 dimensional vectors r(n)=[r(n), . . . , r(n−M)].sup.T, where (⋅).sup.T denotes transpose, such that r(n)=HG(n)×(n)+v(n); H is an (M+1)×(M+L+1) dimensional channel matrix H = [ h o .Math. h L h o .Math. h L ] ; G(n)=diag{V.sub.y(n)e.sup.jψ.sup.y(n), . . . , V.sub.n(y-M-L)e.sup.jψ.sup.y(n-M-L)} is an (M+L+1)×(M+L+1) dimensional diagonal matrix which consists of the nonlinear responses, x(n)=[x(n), . . . , x(n−M−L)].sup.T, and v(n)=[v(n), . . . , v(n−M)].sup.T; f.sup.T=G′(n)[ƒ.sub.0, . . . , ƒ.sub.M], where [ƒ.sub.0, . . . , ƒ.sub.M]H≈[0, . . . , 1, . . . , 0] represents a channel equalization; G ( n ) = 1 V y ( n - d ) e - j ψ y ( n - d ) represents an analog processor equalization; and {circumflex over (r)}(n) is a linearized representation of r(n).

13. The method according to claim 12, wherein: vector a = [ g 0 0 , g 0 1 , .Math. , g P D ] T , vector x = [ x ( 0 ) ; .Math. , x ( N - L ) ] T , vector B = [ r ^ ( L ) r ^ ( L ) .Math. "\[LeftBracketingBar]" r ^ ( L ) .Math. "\[RightBracketingBar]" .Math. r ^ ( L - D ) .Math. "\[LeftBracketingBar]" r ^ ( L - D ) .Math. "\[RightBracketingBar]" P - 1 .Math. .Math. r ^ ( N ) r ^ ( N ) .Math. "\[LeftBracketingBar]" r ^ ( N ) .Math. "\[RightBracketingBar]" .Math. r ^ ( N - D ) .Math. "\[LeftBracketingBar]" r ^ ( N - D ) .Math. "\[RightBracketingBar]" P - 1 ] , and min { g kd } .Math. n = L N .Math. "\[LeftBracketingBar]" x ( n - L ) - .Math. d = 0 D .Math. k = 1 P g k d r ^ ( n - d ) .Math. "\[LeftBracketingBar]" r ^ ( n - d ) .Math. "\[RightBracketingBar]" k - 1 .Math. "\[RightBracketingBar]" 2 min a .Math. x - Ba .Math. 2 , having solution a=B.sup.+x, where B.sup.+=(B.sup.HB).sup.−1B is the pseudo-inverse of the matrix B.

14. The method according to claim 10, wherein: the Volterra series expansion comprises at least fifth order terms; and the deep neural network comprises at least three layers, each layer having at least 10 feature maps, and a fully connected layer subsequent to the at least three layers.

15. The method according to claim 10, further comprising demodulating the output z(n) as the set of information symbols.

16. A non-linear distortion-compensating processor, comprising: an input configured to receive a non-linearly distorted radio frequency signal r(n) representing information x(n) distorted by a radio frequency communication system to produce a distorted signal y(n) communicated through a channel h; at least one automated processor, configured to: decompose the non-linearly distorted radio frequency signal r(n) received from the channel h as a Volterra series expansion y ( n ) = .Math. d = 0 D .Math. k = 0 P b k d x ( n - d ) .Math. "\[LeftBracketingBar]" x ( n - d ) .Math. "\[RightBracketingBar]" k - 1 ; process the Volterra series expansion with a deep neural network comprising a plurality of neural network hidden layers trained with respect to the distortion of the radio frequency communication system comprising the non-linear distortion and channel distortion, to equalize the distortion, the process having an input-output response z ( n ) = .Math. k = 1 P .Math. d 1 = 0 D .Math. .Math. d k = 0 D f d 1 , .Math. , d k .Math. i = 1 k r ( n - d i ) ;  and demodulate z(n) to extract the information x(n) modulated in the non-linearly distorted radio frequency signal r(n), wherein: p is a respective nonlinearity order; d is a memory depth parameter; D is a total memory length; P is a total nonlinearity order; k is a nonlinear order; and b.sub.kd are nonlinear response parameters.

17. The non-linear distortion-compensating processor according to claim 16, wherein the non-linearly distorted radio frequency signal is an orthogonal frequency multiplexed signal which is distorted by amplification by a power amplifier and transmission over the communication channel h.

18. The non-linear distortion-compensating processor according to claim 16, wherein the deep neural network comprises at least three convolutional network layers, each layer having at least 10 feature maps, and a fully connected layer subsequent to the at least three layers.

19. The non-linear distortion-compensating processor according to claim 16, wherein the deep neural network is trained with data comprising symbol-specific pairs of the information x(n) and the corresponding non-linearly distorted radio frequency signal r(n) representing the non-linearly distorted information y(n) communicated through the channel h.

20. The non-linear distortion-compensating processor according to claim 16, wherein the response r(n) of the channel h is r ( n ) = .Math. = 0 L h y ( n - ) + v ( n ) ; and the at least one automated processor is further configured to: compute a maximum likelihood estimation for a nonlinear equalization z(n)=arg min.sub.∀x(n)|{circumflex over (r)}(n)−V.sub.ye.sup.jψ.sup.y x(n)|.sup.2, to produce output z(n)=ƒ.sup.T r(n)≈x(n−d), with equalization delay d; train the deep neural network equalizer to determine a set of finite-impulse response channel coefficients custom character of the channel h, having a channel amplitude response V.sub.y, a channel phase response ψ.sub.y, and a channel equalization response ƒ.sup.T; and approximate G′(n) with a Volterra series model z ( n ) = .Math. d = 0 D .Math. k = 0 P g k d r ˆ ( n - d ) .Math. "\[LeftBracketingBar]" r ˆ ( n - d ) .Math. "\[RightBracketingBar]" k - 1 , with g.sub.kd designed such that z(n)≈x(n−custom character) for equalization delay custom character, and coefficients g.sub.kd estimated according to min { g kd } .Math. n = L N .Math. "\[LeftBracketingBar]" x ( n - L ) - .Math. d = 0 D .Math. k = 1 P g k d r ˆ ( n - d ) .Math. "\[LeftBracketingBar]" r ˆ ( n - d ) .Math. "\[RightBracketingBar]" k - 1 .Math. "\[RightBracketingBar]" 2 , with training symbols x(n) and received samples {circumflex over (r)}(n); wherein: custom character is an equalization delay; v(n) is an additive white Gaussian noise component; r(n) is stacked together into M+1 dimensional vectors r(n)=[r(n), . . . , r(n−M)].sup.T, where (⋅).sup.T denotes transpose, such that r(n)=HG(n)×(n)+v(n); H is an (M+1)×(M+L+1) dimensional channel matrix H = [ h o .Math. h L h o .Math. h L ] ; G(n)=diag {V.sub.y(n)e.sup.jψ.sup.y(n), . . . , V.sub.n(y-M-L)e.sup.jψ.sup.y(n-m-L)} is an (M+L+1)×(M+L+1) dimensional diagonal matrix which consists of the nonlinear responses, x(n)=[x(n), . . . , x(n−M−L)].sup.T, and v(n)=[v(n), . . . , v(n−M)].sup.T; f.sup.T=G′(n)[ƒ.sub.0, . . . , ƒ.sub.M], where [ƒ.sub.0, . . . , ƒ.sub.M]H≈[0, . . . , 1, . . . , 0] represents a channel equalization, and G ( n ) 1 V y ( n - d ) e - j ψ y ( n - d ) represents an analog processor equalization; and {circumflex over (r)}(n) is a linearized representation of r(n).

Description

BRIEF DESCRIPTION OF THE DRAWINGS

(1) FIG. 1 shows a system block diagram with nonlinear power amplifier and deep neural network equalizer.

(2) FIG. 2 shows a block diagram of DNN equalizer.

(3) FIGS. 3A-3D show constellations of 16 QAM over a simulated PA. FIG. 3A: received signal. FIG. 3B: Volterra equalizer output. FIG. 3C: time-delayed NN output. FIG. 3D: Volterra+NN output.

(4) FIGS. 4A-4D show constellation of 16 QAM over a real PA. FIG. 4A: received signal. FIG. 4B: Volterra equalizer output. FIG. 4C: time-delayed NN output. FIG. 4D: Volterra+NN output.

(5) FIG. 5 shows a comparison of three equalization methods for 16-QAM under various NLD levels.

(6) FIG. 6: shows a table comparing MSE/SER improvement in percentage for the three equalization methods.

DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS

Volterra-Based DNN Equalizer

(7) The present technology therefore employs deep neural networks to implement the nonlinear equalizer in the receiver, which can mitigate the nonlinear effects of the received signals due to not only PAs but also nonlinear channels and propagations. The architecture of the DNN equalizer is shown in FIG. 2, which shows an input X, which undergoes a series of three 1-d convolutions, am FC dropout, to produce the output Y.

(8) Different from [10], multi-layer convolutional neural networks (CNNs) are employed. Different from conventional neural network predistorters proposed in [6], neural networks are used as equalizers at the receivers. Different from conventional neural network equalizers such as those proposed in [14] [15], in the present DNN equalizer, not only the linear delayed samples r(n), but also the CNN and the input features in X are used. The Volterra series models are applied to create input features.

(9) We can assume that the linear channel H has already been equalized by a linear equalizer, whose output signal is r(n). In fact, this equalization is not required, but simplifies the presentation of the analysis.

(10) According to Volterra series representation of nonlinear functions, the input-output response of the nonlinear equalizer can be written as

(11) z ( n ) = .Math. k = 1 P .Math. d 1 = 0 D .Math. .Math. d k = 0 D f d 1 , .Math. , d k .Math. i = 1 k r ( n - d i ) . ( 20 )

(12) One of major problems is that the number of coefficients ƒ.sub.d.sub.1.sub., . . . ,d.sub.k increases exponentially with the increase of memory length D and nonlinearity order P. There are many different ways to develop more efficient Volterra series representations with reduced number of coefficients. For example, [23], exploits the fact that higher-order terms do not contribute significantly to the memory effects of PAs to reduce the memory depth d when the nonlinearity order k increases.

(13) This technique can drastically reduce the total number of coefficients. In [24] [25] and [26], a dynamic deviation model was developed to reduce the full Volterra series model (20) to the following simplified one:

(14) z ( n ) = z s ( n ) + z d ( n ) = .Math. k = 1 P f k , 0 r k ( n ) + .Math. k = 1 P .Math. j = 1 k r k - j ( n ) .Math. d 1 = 0 D .Math. .Math. d j = d j - 1 D f k , j .Math. i = 1 j r ( n - d i )

(15) where z.sub.s(n) is the static term, and z.sub.d(n) is the dynamic term that includes all the memory effects. We can see that the total number of coefficients can be much reduced by controlling the dynamic order j which is a selectable parameter.

(16) We construct the input features of the DNN based on the model (21). Corresponding to the static term z.sub.s(n), we change it to:

(17) z ^ s ( n ) = .Math. 1 k P f k , 0 r ( n ) .Math. "\[LeftBracketingBar]" r ( n ) .Math. "\[RightBracketingBar]" k - 1 .

(18) The reason that (22) changes r.sup.k(n) to r(n)|r(n)|.sup.k-1 is that only the signal frequency within the valid passband is interested. This means the input feature vector X should include terms r(n)|r(n)|.sup.k-1. Similarly, corresponding to the dynamic term z.sub.d(n), we need to supply

(19) r k - j ( n ) .Math. i = 1 j r ( n - d i )
in the features where half of the terms r(n) and r(n−d.sub.i) should be conjugated. For simplicity, in the DNN equalizer, the vector X includes r(n−q)|r(n−q)|.sup.k-1 for some q and k.

(20) By applying Volterra series components directly as features of the input X, the DNN can develop more complex nonlinear functions with a fewer number of hidden layers and a fewer number of neurons. This will also make the training procedure converge much faster with much less training data.

(21) In FIG. 2, the input X is a tensor formed by the real and imaginary parts of r(n−q)|r(n−q)|.sup.k-1 with appropriate number of delays q and nonlinearities k. There are three single dimension convolutional layers, each with 20 or 10 feature maps. After a drop-out layer for regularization, this is followed by a fully connected layer with 20 neurons. Finally, there is a fully-connected layer to form the output tensor Y which has two dimensions. The output Y is used to construct the complex z(n), where z(n)={circumflex over (x)}(n−d) for some appropriate delay d. All the convolutional layers and the first fully connected layer use the sigmoid activation function, while the output layer uses the linear activation function. The mean square error loss function L.sup.loss=E[|x(n−d)−z(n)|.sup.2] is used, where z(n) is replaced by Y and x(n−d) is replaced by training data labels.

Experiment Evaluations

(22) Experiments are presented on applying the Volterra series based DNN equalizer (Volterra+NN) for nonlinear PA equalization. The (Volterra+NN) scheme with the following equalization methods: a Volterra series-based equalizer (Volterra) and a conventional time-delay neural network equalizer (NN). The performance metrics are mean square error (MSE)
√{square root over (E[|z(n)−x(n−d)|.sup.2]/E[|x(n−d)|.sup.2])}

(23) and symbol error rate (SER).

(24) Both simulated signals and real measurement signals were employed. To generate simulated signals, a Doherty nonlinear PA model consisting of 3rd and 5th order nonlinearities was employed. Referring to (2), the coefficients b.sub.k,q were
b.sub.0,0:2={1.0513+0.0904j,−0.068−0.0023j,0.0289−0.0054j}
b.sub.2,0:2={−0.0542−0.29j,0.2234+0.2317j,−0.0621−0.0932j}
b.sub.4,0:2={−0.9657−0.7028j,−0.2451−0.3735j,0.1229+0.1508j},

(25) which was used in [5] to simulate a 5th order dominant nonlinear distortion derived from PA devices used in the satellite industry. For real measurement, our measurement signals were obtained from PA devices used in the cable TV (CATV) industry, which are typically dominated by 3.sup.rd order nonlinear distortion (NLD). Various levels of nonlinear distortion, in terms of dBc, were generated by adjusting the PAs.

(26) For the Volterra equalizer, the approximate response of the nonlinear equalizer with delays including 8 pre- and post-main taps and with nonlinearities including even and odd order nonlinearity up to the 5th order was employed. To determine the values of the Volterra coefficients, N=4; 096 training symbols were transmitted through the PA and then collected the noisy received samples r(n).

(27) For the conventional time-delay NN equalizer, a feedforward neural network with an 80-dimensional input vector X and 5 fully-connected hidden layers with 20, 20, 10, 10, 10 neurons, respectively, was applied.

(28) FIG. 3 shows the constellation and MSE of the equalizer's outputs. It can be seen that the proposed scheme provides the best performance.

(29) FIG. 4 shows the constellation of 16 QAM equalization over the real PA. The corresponding SER were 0.0067, 0.0027, 0.00025, respectively. It can be seen that the Volterra+NN scheme has the best performance.

(30) FIG. 5 provides MSE measurements for 16-QAM under various nonlinear distortion level dBc. For each 1 dB increase in NLD, the resultant MSE is shown for the “Measured”, “Volterra”, “NN”, and the proposed “Volterra+NN” cases. MSE reduction diminishes appreciably as modulation order increases from QPSK to 64-QAM, but small improvements in MSE have been observed lead to appreciable SER improvement, especially for more complex modulation orders. The 4,096 symbol sample sizes have limited the measurements to a minimum measurable 0.000244 SER, which represents 1 symbol error out of 4,096 symbols.

(31) FIG. 6 summarizes equalization performance, which shows the averaged percent reduction/improvement in MSE and SER from the NLD impaired data for multiple modulation orders. Note that 0% SER improvement for QPSK was because the received signal's SER was already very low.

(32) The nonlinear equalization scheme presented by integrating the Volterra series non-linear model with deep neural networks yields superior results over conventional nonlinear equalization approaches in mitigating nonlinear power amplifier distortions. It finds application for many 5G communication scenarios.

(33) The technology may be implemented as an additional component in a receiver, or within the digital processing signal chain of a modern radio. A radio is described in US 20180262217, expressly incorporated herein by reference.

(34) In an implementation, a base station may include a SDR receiver configured to allow the base station to operate as an auxiliary receiver. In an example implementation, the base station may include a wideband receiver bank and a digital physical/media access control (PHY/MAC) layer receiver. In this example, the SDR receiver may use a protocol analyzer to determine the protocol used by the source device on the uplink to the primary base station, and then configure the digital PHY/MAC layer receiver for that protocol when operating as art auxiliary receiver. Also, the digital PHY/MAC layer receiver may be configured to operate according to another protocol when operating as a primary base station. In another example, the base station may include a receiver hank for a wireless system, for example, a fifth Generation (5G) receiver bank, and include an additional receiver having SDR configurable capability. The additional receiver may be, for example, a digital Wi-Fi receiver configurable to operate according to various Wi-Fi protocols. The base station may use a protocol analyzer to determine the particular Wi-Fi protocol used by the source device on the uplink to the primary base station. The base station may then configure the additional receiver as the auxiliary receiver for that Wi-Fi protocol.

(35) Depending on the hardware configuration, a receiver may be used to flexibly provide uplink support in systems operating according to one or more protocols such as the various IEEE 802.11 Wi-Fi protocols, 3.sup.rd Generation Cellular (3G), 4.sup.th Generation Cellular (4G) wide band code division multiple access (WCDMA), Long Term Evolution (LTE) Cellular, and 5.sup.th generation cellular (5G).

(36) See, 5G References, infra.

(37) Processing unit may comprise one or more processors, or other control circuitry or any combination of processors and control circuitry that provide, overall control according to the disclosed embodiments. Memory may be implemented as any type of as any type of computer readable storage media, including non-volatile and volatile memory.

(38) The example embodiments disclosed herein may be described in the general context of processor-executable code or instructions stored on memory that may comprise one or more computer readable storage media (e.g., tangible non-transitory computer-readable storage media such as memory). As should be readily understood, the terms “computer-readable storage media” or “non-transitory computer-readable media” include the media for storing of data, code and program instructions, such as memory, and do not include portions of the media for storing transitory propagated or modulated data communication signals.

(39) While the functionality disclosed herein has been described by illustrative example using descriptions of the various components and devices of embodiments by referring to functional blocks and processors or processing units, controllers, and memory including instructions and code, the functions and processes of the embodiments may be implemented and performed using any type of processor, circuit, circuitry or combinations of processors and or circuitry and code. This may include, at least in part, one or more hardware logic components. For example, and without limitation, illustrative types of hardware logic components that can be used include field programmable gate arrays (FPGAs), application specific integrated circuits (ASICs), application specific standard products (ASSPs), system-on-a-chip systems (SOCs), complex programmable logic devices (CPLDs), etc. Use of the term processor or processing unit in this disclosure is mean to include all such implementations.

(40) The disclosed implementations include a receiver, one or more processors in communication with the receiver, and memory in communication with the one or more processors, the memory comprising code that, when executed, causes the one or more processors to control the receiver to implement various features and methods according to the present technology.

(41) Although the subject matter has been described in language specific to structural features and/or methodological acts, it is to be understood that the subject matter defined in the appended claims is not necessarily limited to the specific features or acts described above. Rather, the specific features and acts described above are disclosed as example embodiments, implementations, and forms of implementing the claims and these example configurations and arrangements may be changed significantly without departing from the scope of the present disclosure. Moreover, although the example embodiments have been illustrated with reference to particular elements and operations that facilitate the processes, these elements, and operations may be combined with or, be replaced by, any suitable devices, components, architecture or process that achieves the intended functionality of the embodiment. Numerous other changes, substitutions, variations, alterations, and modifications may be ascertained to one skilled in the art and it is intended that the present disclosure encompass all such changes, substitutions, variations, alterations, and modifications a falling within the scope of the appended claims.

REFERENCES

(42) [1] J.-A. Lucciardi, P. Potier, G. Buscarlet, F. Barrami, and G. Mesnager, “Non-linearized amplifier and advanced mitigation techniques: Dvbs-2x spectral efficiency improvement,” in GLOBECOM 2017-2017 IEEE Global Communications Conference. IEEE, 2017, pp. 1-7. [2] J. Wood, Behavioral modeling and linearization of RF power amplifiers. Artech House, 2014. [3] C.-L. Wang and Y. Ouyang, “Low-complexity selected mapping schemes for peak-to-average power ratio reduction in ofdm systems,” IEEE Transactions on signal processing, vol. 53, no. 12, pp. 4652-4660, 2005. [4] J. Kim and K. Konstantinou, “Digital predistortion of wideband signals based on power amplifier model with memory,” Electronics Letters, vol. 37, no. 23, pp. 1417-1418, 2001. [5] L. Ding, G. T. Zhou, D. R. Morgan, Z. Ma, J. S. Kenney, J. Kim, and C. R. Giardina, “A robust digital baseband predistorter constructed using memory polynomials,” IEEE Transactions on communications, vol. 52, no. 1, pp. 159-165, 2004. [6] M. Rawat, K. Rawat, and F. M. Ghannouchi, “Adaptive digital predistortion of wireless power amplifiers/transmitters using dynamic realvalued focused time-delay line neural networks,” IEEE Transactions on Microwave Theory and Techniques, vol. 58, no. 1, pp. 95-104, 2010. [7] S. Dimitrov, “Non-linear distortion cancellation and symbol-based equalization in satellite forward links,” IEEE Trans Wireless Commun, vol. 16, no. 7, pp. 4489-4502, 2017. [8] D. J. Sebald and J. A. Bucklew, “Support vector machine techniques for non-linear equalization,” IEEE Transactions on Signal Processing, vol. 48, no. 11, pp. 3217-3226, 2000. [9] S. Chen, B. Mulgrew, and P. M. Grant, “A clustering technique for digital communications channel equalization using radial basis function networks,” IEEE Transactions on neural networks, vol. 4, no. 4, pp. 570-590, 1993. [10] B. Li, C. Zhao, M. Sun, H. Zhang, Z. Zhou, and A. Nallanathan, “A bayesian approach for nonlinear equalization and signal detection in millimeter-wave communications,” IEEE Transactions on Wireless Communications, vol. 14, no. 7, pp. 3794-3809, 2015. [11] F. Mkadem and S. Boumaiza, “Physically inspired neural network model for rf power amplifier behavioral modeling and digital predistortion,” IEEE Transactions on Microwave Theory and Techniques, vol. 59, no. 4, pp. 913-923, 2011. [12] T. Liu, S. Boumaiza, and F. M. Ghannouchi, “Dynamic behavioral modeling of 3g power amplifiers using real-valued time-delay neural networks,” IEEE Transactions on Microwave Theory and Techniques, vol. 52, no. 3, pp. 1025-1033, 2004. [13] M. Ibnkahla, “Applications of neural networks to digital communications—a survey,” Signal processing, vol. 80, no. 7, pp. 1185-1215, 2000. [14] D.-C. Park and T.-K. J. Jeong, “Complex-bilinear recurrent neural network for equalization of a digital satellite channel,” IEEE Transactions on Neural Networks, vol. 13, no. 3, pp. 711-725, 2002. [15] A. Uncini, L. Vecci, P. Campolucci, and F. Piazza, “Complex-valued neural networks with adaptive spline activation function for digital-radio-links nonlinear equalization,” IEEE Transactions on Signal Processing, vol. 47, no. 2, pp. 505-514, 1999. [16] M. S. Sim, M. Chung, D. Kim, J. Chung, D. K. Kim, and C.-B. Chae, “Non-linear self-interference cancellation for full-duplex radios: From link-level and system-level performance perspectives,” IEEE Communications Magazine, vol. 55, no. 9, pp. 158-167, 2017. [17] I. Yoffe and D. Wulich, “Predistorter for mimo system with nonlinear power amplifiers,” IEEE Transactions on Communications, vol. 65, no. 8, pp. 3288-3301, 2017. [18] M. Abdelaziz, L. Anttila, and M. Valkama, “Reduced-complexity digital predistortion for massive mimo,” in Acoustics, Speech and Signal Processing (ICASSP), 2017 IEEE International Conference on. IEEE, 2017, pp. 6478-6482. [19] H. Yan and D. Cabric, “Digital predistortion for hybrid precoding architecture in millimeter-wave massive mimo systems,” in Acoustics, Speech and Signal Processing (ICASSP), 2017 IEEE International Conference on. IEEE, 2017, pp. 3479-3483. [20] C. Mollén, E. G. Larsson, and T. Eriksson, “Waveforms for the massive mimo downlink: Amplifier efficiency, distortion, and performance,” IEEE Transactions on Communications, vol. 64, no. 12, pp. 5050-5063, 2016. [21] A. Cheaito, M. Crussière, J.-F. Hélard, and Y. Louët, “Quantifying the memory effects of power amplifiers: Evm closed-form derivations of multicarrier signals.” IEEE Wireless Commun. Letters, vol. 6, no. 1, pp. 34-37, 2017. [22] K. Simons, Technical Handbook for CATV Systems, 3rd Edition. Jerrod Publication No. 436-001-01, 1968. [23] J. Staudinger, J.-C. Nanan, and J. Wood, “Memory fading volterra series model for high power infrastructure amplifiers,” in Radio and Wireless Symposium (RWS), 2010 IEEE. IEEE, 2010, pp. 184-187. [24] A. Zhu, J. C. Pedro, and T. J. Brazil, “Dynamic deviation reduction-based volterra behavioral modeling of rf power amplifiers,” IEEE Transactions on microwave theory and techniques, vol. 54, no. 12, pp. 4323-4332, 2006. [25] A. Zhu, P. J. Draxler, J. J. Yan, T. J. Brazil, D. F. Kimball, and P. M. Asbeck, “Open-loop digital predistorter for rf power amplifiers using dynamic deviation reduction-based volterra series,” IEEE Transactions on Microwave Theory and Techniques, vol. 56, no. 7, pp. 1524-1534, 2008. [26] L. Guan and A. Zhu, “Simplified dynamic deviation reduction-based volterra model for doherty power amplifiers,” in Integrated Nonlinear Microwave and Millimetre-Wave Circuits (INMMIC), 2011 Workshop on. IEEE, 2011, pp. 1-4. [27] Schetzen, M. The Volterra and Wiener Theories of Non-linear Systems. (1980) Wiley & Sons. [28] Black, H. S. [October, 1928] Translating system. U.S. Pat. No. 1,686,792. [29] Black, H. S. [December, 1937] Wave translating system. U.S. Pat. No. 2,102,671. [30] Mitchell, A. F. [November, 1979] A 135 MHz feedback amplifier. IEEE Colloq. Broadband High Frequency Amplifiers. [31] Arthanayake, T. and Wood, H. B. [8 Apr. 1971] Linear amplification using envelope feedback. Elec. Lett. [32] Chadwick, P. [1986] Wideband Amplifier Applications Book, Edition 2, Plessey Semiconductor.

AMPLIFIER REFERENCES

(43) Aghvami, A. H. and Robertson, I. D. [April, 1993] Power limitation and high-power amplifier non linearities in on-board satellite communications systems. Electron. and Comm. Engin. J. Arthanayake, T. and Wood, H. B. [8 Apr. 1971] Linear amplification using envelope feedback. Elec. Lett. Bennet, T. J. and Clements, R. F. [May, 1974] Feedforward—An alternative approach to amplifier linearisation. Radio and Electron. Engin. Bhargava, V. K. et al. [1981] Digital Communications by Satellite, John Wiley and Sons. Black, H. S. [December, 1937] Wave translating system. U.S. Pat. No. 2,102,671. Black, H. S. [October, 1928] Translating system. U.S. Pat. No. 1,686,792. Bond F. E. and Meyer, H. F. [April, 1970] Intermodulation effects in limiter amplifier repeaters. IEEE Trans. Comm., Vol. COM-18, p. 127-135. Chadwick, P. [1986] Wideband Amplifier Applications Book, Edition 2, Plessey Semiconductor. Cole, R. A. [December, 1989] Linearisation of a power amplifier using Cartesian Loop feedback. Report No. 72/89/R/451/C. Roke Manor Res. ETSI [August, 1994] Standard ETR 132. Radio broadcasting systems; Code of practice for site engineering VHF FM sound broadcasting transmitters. European Telecommunications Standards Institute, Sophia Antipolis, F-06291, Valbonne Cedex, France. ETSI [January, 1995] European Standard ETS 300 384. Radio broadcasting systems; Very high frequency (VHF), frequency modulated, sound broadcasting transmitters. European Telecommunications Standards Institute, Sophia Antipolis, F-06291, Valbonne Cedex, France. ETSI [June, 1998] Standard ETR 053 Ed 3—Radio site engineering for equipment and systems in the mobile service. European Telecommunications Standards Institute, Sophia Antipolis, F-06291, Valbonne Cedex, France. ETSI [March, 1997] European Standard ETS 300 113. Radio equipment and systems (RES); Land mobile service; Technical characteristics and test conditions for radio equipment intended for the transmission of data (and speech) and having an antenna connector. European Telecommunications Standards Institute, Sophia Antipolis, F-06291, Valbonne Cedex, France. Gray, L. F. [1980] Application of broadband linearisers to satellite transponders. IEEE Conf. Proc. ICC′80. Heathman, A. C. [1989] Methods for intermodulation prediction in communication systems. Ph. D. Thesis, University of Bradford, United Kingdom. IESS [November, 1996] IESS-401 (Rev. 4). Performance requirements for intermodulation products transmitted from INTELSAT earth stations. Intelsat Earth Station Standard (IESS). Kaeadar, K. [December, 1986] Gaussian white-noise generation for digital signal synthesis. IEEE Trans. Inst. and Meas., Vol. IM 35, 4. Kahn, L. R. [July, 1952] SSB transmission by envelope elimination and restoration. Proc. IRE. Mitchell, A. F. [November, 1979] A 135 MHz feedback amplifier. IEEE Colloq. Broadband High Frequency Amplifiers. Pavliouk, A. [1977] Unification of measurement procedures for out-of-band emission spectra and peak envelope power of single-sideband radio transmitter measurements. Proc. of the NIIR, 4 (in Russian). Petrovic, V. and Gosling, W. [10 May 1979] Polar loop transmitter. Elec. Lett. Pye Telecom [November, 1978] Intermodulation in VHF and UHF radio systems—locating and minimizing the effects. Engineering Notes, Pub. Ref. No. TSP480/1, United Kingdom. Radiocommunications Agency [April, 1987] Code of practice for radio site engineering. MPT 1331. Radiocommunications Agency (RA), Flyde Microsystems Ltd. United Kingdom. Saleh, A. M. [May 1982] Intermodulation analysis of FDMA satellite systems employing compensated and uncompensated TWT‘s’. IEEE Trans. Comm., Vol. COM-30, 5. Schetzen, M. [1980] The Volterra and Wiener Theories of Non-linear Systems. Wiley & Sons. Schetzen, M. [1980] The Volterra and Wiener Theories of Non-linear Systems. Wiley & Sons. Shahid, M., Shepherd, S. J., Lin, B., Khairruddin, I., and Barton, S. K. [December, 1996] Study of methods of measuring multi-carrier intermodulation performance Report No. 581, Purchase Order No. 142379 d'ESA, with University of Bradford, United Kingdom. Shimbo, O. [February, 1971] Effects of intermodulation, AM-PM conversion, and additive noise in multicarrier TWT systems. Proc. IEEE, Vol. 59, p. 230-238. Smith, C. N. [1986] Application of the polar loop technique to UHF SSB transmitters. Ph.D. Thesis, University of Bath. Smith, C. N. and PETROVIC, V. [1982] Cartesian loop transmitter. Internal Research Report, University of Bath, School of Electrical and Electronic Engineering. Tondryk, W. [1991] Intermodulation testing of the INMARSAT payload—Response to system PDR. Marconi Space Systems Ltd. Wassermann, M. et al. [1983] Study and breadboarding of an L-band high power linearized TWT amplifier. Final Report, ESTEC contract No. 5459/83/NL/GM. Wood, A. [October 1998] Radio interference: Sources and solutions. LPRA NEWS, p. 21.

VOLTERRA SERIES REFERENCES

(44) Bohm, D. The Special Theory of Relativity, Benjamin, 1965. Censor, D., & Melamed, T, 2002, Volterra differential constitutive operators and locality considerations in electromagnetic theory, PIER—Progress in Electromagnetic Research, 36: 121-137 Censor, D., 2000, A quest for systematic constitutive formulations for general field and wave systems based on the Volterra differential operators, PIER—Progress In Electromagnetics Research, (25): 261-284 Censor, D., 2001, Constitutive relations in inhomogeneous systems and the particle-field conundrum, PIER—Progress In Electromagnetics Research, (30): 305-335 Schetzen, M., 1980, The Volterra and Wiener Theorems of Nonlinear Systems, New York, Chichester, Brisbane and Toronto: John Wiley and Sons Sonnenschein, M & Censor, D., 1998, Simulation of Hamiltonian light beam propagation in nonlinear media, JOSA—Journal of the Optical Society of America B, (15): 1335-1345

FILTER REFERENCES

(45) Akaiwa, Y. Introduction to Digital Mobile Communication. New York: Wiley, 1997. Altera Corporation. Digital Predistortion Reference Design. Application Note 314, 2003. Aysal, Tuncer C., and Kenneth E. Barner, “Myriad-Type Polynomial Filtering”, IEEE Transactions on Signal Processing, vol. 55, no. 2, February 2007. Barner, Kenneth E., and Tuncer Can Aysal, “Polynomial Weighted Median Filtering”, IEEE Transactions on Signal Processing, vol. 54, no. 2, February 2006. Biglieri, Ezio, Sergio Barberis, and Maurizio Catena, “Analysis and Compensation of Nonlinearities in Digital Transmission Systems”, IEEE Journal on selected areas in Communications, vol. 6, no. 1, January 1988. Budura, Georgeta, and Corina Botoca, “Efficient Implementation of the Third Order RLS Adaptive Volterra Filter”, FACTA Universitatis (NIS) Ser.: Elec. Energ. vol. 19, no. 1, April 2006. Ding, L., et al. “A Robust Digital Baseband Predistorter Constructed Using Memory Polynomials,” IEEE Transactions on Communications, Vol. 52, No. 1, June 2004. Fang, Yang-Wang, Li-Cheng Jiao, Xian-Da Zhang and Jin Pan, “On the Convergence of Volterra Filter Equalizers Using a Pth-Order Inverse Approach”, IEEE Transactions on Signal Processing, vol. 49, no. 8, August 2001. Guérin, Alexandre, Gérard Faucon, and Régine Le Bouquin-Jeannès, “Nonlinear Acoustic Echo Cancellation Based on Volterra Filters”, IEEE Transactions on Speech and Audio Processing, vol. 11, no. 6, November 2003. Haykin, Simon, “Adaptive Filter Theory”, Fourth Edition, Pearson Education, 2008. Kamiya, N., and F. Maehara. “Nonlinear Distortion Avoidance Employing Symbol-wise Transmit Power Control for OFDM Transmission,” Proc. of Int'l. OFDM Workshop, Hamburg, 2009. Kim, J., and K. Konstantinou. “Digital predistortion of wideband signals based on power amplifier model with memory,” Electronic Letters, Vol. 37, No. 23, November 2001. Krall, Christoph, Klaus Witrisal, Geert Leus and Heinz Koeppl, “Minimum Mean-Square Error Equalization for Second-Order Volterra Systems”, IEEE Transactions on Signal Processing, vol. 56, no. 10, October 2008. Leis, John, “Adaptive Filter Lecture Notes & Examples”, Nov. 1, 2008 www.usq.edu.au/users/leis/notes/sigproc/adfilt.pdf. López-Valcarce, Roberto, and Soura Dasgupta, “Second-Order Statistical Properties of Nonlinearly Distorted Phase-Shift Keyed (PSK) Signals”, IEEE Communications Letters, vol. 7, no. 7, July 2003. Lozhkin, Alexander N. “Turbo Linearizer for High Power Amplifier.” In 2011 IEEE 73rd Vehicular Technology Conference (VTC Spring), pp. 1-5. IEEE, 2011. Mathews, V. John, “Adaptive Polynomial Filters,” IEEE Signal Processing Magazine, Vol. 8, No. 3, July 1991. Park, Dong-Chul, and Tae-Kyun Jung Jeong, “Complex-Bilinear Recurrent Neural Network for Equalization of a Digital Satellite Channel”, IEEE Transactions on Neural Networks, vol. 13, no. 3, May 2002. Rai, Amrita, and Amit Kumar Kohli. “Analysis of Adaptive Volterra Filters with LMS and RLS Algorithms.” AKGEC Journal of Technology 2, no. 1 (2011). Therrien, Charles W., W. Kenneth Jenkins, and Xiaohui Li, “Optimizing the Performance of Polynomial Adaptive Filters: Making Quadratic Filters Converge Like Linear Filters”, IEEE Transactions on Signal Processing, vol. 47, no. 4, April 1999. Tsimbinos John, and Langford B. White, “Error Propagation and Recovery in Decision-Feedback Equalizers for Nonlinear Channels”, IEEE Transactions on Communications, vol. 49, no. 2, February 2001. Woo, Young Yun, et al. “Adaptive Digital Feedback Predistortion Technique for Linearizing Power Amplifiers,” IEEE Transactions on Microwave Theory and Techniques, Vol. 55, No. 5, May 2007. Zaknich, A., “Principal of Adaptive Filter and Self Learning System”, Springer Link 2005.

VOLTERRA SERIES PATENTS

(46) U.S. Patent and Published Patent Application Nos.: U.S. Pat. Nos. 4,615,038; 4,669,116; 4,870,371; 5,038,187; 5,309,481; 5,329,586; 5,424,680; 5,438,625; 5,539,774; 5,647,023; 5,692,011; 5,694,476; 5,744,969; 5,745,597; 5,790,692; 5,792,062; 5,815,585; 5,889,823; 5,924,086; 5,938,594; 5,991,023; 6,002,479; 6,005,952; 6,064,265; 6,166,599; 6,181,754; 6,201,455; 6,201,839; 6,236,837; 6,240,278; 6,288,610; 6,335,767; 6,351,740; 6,381,212; 6,393,259; 6,406,438; 6,408,079; 6,438,180; 6,453,308; 6,504,885; 6,510,257; 6,512,417; 6,532,272; 6,563,870; 6,600,794; 6,633,208; 6,636,115; 6,668,256; 6,687,235; 6,690,693; 6,697,768; 6,711,094; 6,714,481; 6,718,087; 6,775,646; 6,788,719; 6,812,792; 6,826,331; 6,839,657; 6,850,871; 6,868,380; 6,885,954; 6,895,262; 6,922,552; 6,934,655; 6,940,790; 6,947,857; 6,951,540; 6,954,476; 6,956,433; 6,982,939; 6,992,519; 6,999,201; 6,999,510; 7,007,253; 7,016,823; 7,061,943; 7,065,511; 7,071,797; 7,084,974; 7,092,043; 7,113,037; 7,123,663; 7,151,405; 7,176,757; 7,209,566; 7,212,933; 7,236,156; 7,236,212; 7,239,301; 7,239,668; 7,251,297; 7,268,620; 7,272,594; 7,286,009; 7,295,961; 7,304,591; 7,305,639; 7,308,032; 7,333,559; 7,348,844; 7,400,807; 7,403,884; 7,412,469; 7,423,699; 7,436,883; 7,443,326; 7,489,298; 7,512,900; 7,542,518; 7,551,668; 7,570,856; 7,571,401; 7,576,606; 7,589,725; 7,590,518; 7,602,240; 7,606,539; 7,610,183; 7,657,405; 7,720,232; 7,720,236; 7,728,658; 7,729,446; 7,733,177; 7,746,955; 7,755,425; 7,760,887; 7,773,692; 7,774,176; 7,795,858; 7,796,960; 7,808,315; 7,812,666; 7,821,337; 7,821,581; 7,822,146; 7,826,624; 7,847,631; 7,852,913; 7,853,443; 7,864,881; 7,873,172; 7,885,025; 7,885,797; 7,889,007; 7,894,788; 7,895,006; 7,899,416; 7,902,925; 7,903,137; 7,924,942; 7,929,375; 7,932,782; 7,970,150; 7,970,151; 7,979,837; 7,991,073; 7,991,167; 7,995,674; 8,005,858; 8,023,668; 8,031,882; 8,039,871; 8,045,066; 8,046,199; 8,065,060; 8,089,689; 8,105,270; 8,139,630; 8,148,983; 8,149,950; 8,160,191; 8,165,854; 8,170,508; 8,185,853; 8,193,566; 8,195,103; 8,199,399; 8,213,880; 8,244,787; 8,260,732; 8,265,583; 8,270,530; 8,294,605; 8,295,790; 8,306,488; 8,310,312; 8,315,970; 8,331,511; 8,331,879; 8,345,348; 8,346,692; 8,346,693; 8,346,711; 8,346,712; 8,351,876; 8,354,884; 8,355,684; 8,358,169; 8,364,095; 8,369,447; 8,369,595; 8,380,773; 8,390,375; 8,390,376; 8,396,693; 8,410,843; 8,410,850; 8,412,133; 8,421,534; 8,432,220; 8,437,513; 8,463,582; 8,467,438; 8,477,581; 8,483,343; 8,483,450; 8,487,706; 8,489,047; 8,494,463; 8,498,369; 8,509,347; 8,509,712; 8,519,440; 8,532,215; 8,532,964; 8,538,039; 8,564,368; 8,565,343; 8,577,311; 8,587,375; 8,599,050; 8,605,814; 8,605,819; 8,611,190; 8,611,459; 8,611,820; 8,615,208; 8,619,905; 8,620,631; 8,626,089; 8,649,743; 8,675,925; 8,704,595; 8,705,166; 8,712,345; 8,718,178; 8,718,209; 8,724,857; 8,737,937; 8,737,938; 8,744,141; 8,744,377; 8,758,271; 8,761,409; 8,766,917; 8,767,869; 8,780,693; 8,787,628; 8,798,559; 8,804,807; 8,804,871; 8,811,532; 8,823,452; 8,831,074; 8,831,133; 8,831,135; 8,838,218; 8,843,088; 8,843,089; 8,849,611; 8,855,175; 8,855,234; 8,867,601; 8,874,411; 8,885,765; 8,886,341; 8,891,701; 8,896,471; 8,897,351; 8,903,192; 8,909,176; 8,909,328; 8,933,752; 8,934,573; 8,958,470; 8,964,901; 8,964,996; 8,971,834; 8,976,896; 8,994,657; 8,995,571; 8,995,835; 9,008,153; 9,014,299; 9,019,643; 9,020,454; 9,025,607; 9,031,168; 9,036,734; 9,048,865; 9,048,900; 9,071,313; 9,077,508; 9,088,472; 9,094,036; 9,094,151; 9,104,921; 9,106,304; 9,130,628; 9,137,492; 9,143,274; 9,160,280; 9,160,310; 9,160,687; 9,166,610; 9,166,635; 9,166,698; 9,171,534; 9,184,784; 9,185,529; 9,189,458; 9,191,041; 9,191,049; 9,199,860; 9,209,753; 9,209,841; 9,214,968; 9,214,969; 9,225,295; 9,225,501; 9,231,530; 9,231,647; 9,231,801; 9,236,996; 9,246,525; 9,246,731; 9,252,798; 9,252,821; 9,253,608; 9,257,943; 9,258,156; 9,261,978; 9,264,153; 9,265,461; 9,270,304; 9,270,512; 9,271,123; 9,276,602; 9,294,113; 9,304,501; 9,306,606; 9,311,535; 9,312,892; 9,314,623; 9,322,906; 9,337,781; 9,337,783; 9,352,155; 9,361,681; 9,361,936; 9,362,869; 9,362,942; 9,363,068; 9,369,093; 9,369,255; 9,369,541; 9,397,516; 9,404,950; 9,413,516; 9,419,722; 9,431,972; 9,438,178; 9,438,356; 9,439,597; 9,451,920; 9,460,246; 9,461,597; 9,461,676; 9,473,077; 9,479,322; 9,509,331; 9,509,350; 9,517,030; 9,531,475; 9,536,539; 9,537,759; 9,544,126; 9,559,831; 9,564,876; 9,571,312; 9,575,570; 9,590,664; 9,590,668; 9,595,920; 9,595,982; 9,607,003; 9,607,628; 9,608,676; 9,608,718; 9,614,554; 9,628,119; 9,628,120; 9,646,116; 9,647,717; 9,654,211; 9,654,216; 9,659,120; 9,660,593; 9,660,730; 9,665,510; 9,667,292; 9,674,368; 9,680,423; 9,680,497; 9,697,845; 9,705,477; 9,706,296; 9,712,179; 9,712,233; 9,713,010; 9,722,646; 9,722,691; 9,726,701; 9,727,677; 9,735,741; 9,735,800; 9,735,811; 9,735,876; 9,737,258; 9,742,599; 9,746,506; 9,749,161; 9,755,691; 9,762,268; 9,768,891; 9,778,902; 9,780,869; 9,780,881; 9,787,459; 9,794,000; 9,800,437; 9,800,734; 9,820,311; 9,831,899; 9,837,970; 9,843,346; 9,859,845; 9,866,183; 9,877,265; 9,882,648; 9,887,862; 9,900,088; 9,912,435; 9,913,194; 9,923,524; 9,923,640; 9,923,714; 9,928,212; 9,935,590; 9,935,645; 9,935,715; 9,935,761; 9,940,938; 9,941,963; 9,953,656; 9,954,384; 9,960,794; 9,960,804; 9,960,900; 9,971,920; 9,973,279; 9,974,957; 9,983,243; 9,998,223; 9,998,406; 9,999,780; 10,008,218; 10,009,050; 10,009,109; 10,009,259; 10,013,515; 10,015,593; 10,033,413; 10,033,568; 10,050,636; 10,050,710; 10,050,714; 10,063,265; 10,063,364; 10,075,201; 10,095,927; 10,097,273; 10,097,939; 10,101,370; 10,108,858; 10,110,315; 10,116,390; 10,128,955; 10,141,944; 10,142,754; 10,147,431; 10,148,417; 10,153,793; 10,181,825; 10,224,970; 20010036334; 20010051871; 20020041210; 20020060827; 20020075918; 20020126604; 20020146993; 20020161539; 20020161542; 20020169585; 20020178133; 20020181521; 20020186874; 20030046045; 20030057963; 20030063854; 20030071684; 20030142832; 20030195706; 20030223507; 20040019443; 20040044489; 20040130394; 20040136423; 20040155707; 20040179629; 20040208242; 20040258176; 20050021266; 20050021319; 20050031117; 20050031131; 20050031132; 20050031133; 20050031134; 20050031137; 20050031138; 20050031139; 20050031140; 20050049838; 20050100065; 20050141637; 20050141659; 20050174167; 20050177805; 20050180526; 20050226316; 20050237111; 20050243061; 20050253806; 20050270094; 20050271216; 20050273188; 20060039498; 20060052988; 20060083389; 20060093128; 20060095236; 20060104395; 20060104451; 20060133536; 20060209982; 20060222128; 20060239443; 20060256974; 20060262942; 20060262943; 20060264187; 20060269074; 20060269080; 20060274904; 20070005326; 20070018722; 20070030076; 20070033000; 20070063770; 20070080841; 20070133713; 20070133719; 20070136018; 20070136045; 20070152750; 20070160221; 20070168100; 20070190952; 20070229154; 20070237260; 20070247425; 20070252651; 20070252813; 20070276610; 20080001947; 20080032642; 20080129379; 20080130787; 20080130788; 20080130789; 20080152037; 20080158154; 20080158155; 20080180178; 20080240325; 20080261541; 20080283882; 20080285640; 20080293372; 20090003134; 20090027117; 20090027118; 20090058521; 20090067643; 20090072901; 20090075610; 20090094304; 20090146740; 20090153132; 20090185613; 20090256632; 20090287624; 20090289706; 20090291650; 20090302938; 20090302940; 20090318983; 20100007489; 20100033180; 20100060355; 20100090762; 20100093290; 20100094603; 20100097714; 20100114813; 20100135449; 20100148865; 20100152547; 20100156530; 20100183106; 20100194474; 20100199237; 20100254450; 20100283540; 20100292602; 20100292752; 20100311361; 20100312495; 20110003570; 20110025414; 20110028859; 20110037518; 20110054354; 20110054355; 20110064171; 20110069749; 20110081152; 20110085678; 20110087341; 20110096865; 20110102080; 20110103455; 20110110473; 20110121897; 20110125684; 20110125685; 20110125686; 20110125687; 20110140779; 20110144961; 20110149714; 20110177956; 20110181360; 20110204975; 20110211842; 20110268226; 20110270590; 20110293051; 20120007153; 20120007672; 20120027070; 20120029663; 20120086507; 20120093376; 20120098481; 20120098596; 20120119810; 20120140860; 20120147993; 20120154040; 20120154041; 20120158384; 20120165633; 20120176190; 20120176609; 20120217557; 20120229206; 20120256687; 20120259600; 20120263256; 20120306573; 20120328128; 20130005283; 20130009702; 20130015917; 20130030239; 20130034188; 20130040587; 20130044791; 20130044836; 20130093676; 20130113559; 20130114762; 20130166259; 20130170842; 20130176153; 20130207723; 20130222059; 20130243119; 20130243122; 20130243135; 20130257530; 20130271212; 20130272367; 20130285742; 20130301487; 20130303103; 20130315291; 20130321078; 20130330082; 20130336377; 20140009224; 20140029658; 20140029660; 20140030995; 20140031651; 20140036969; 20140044318; 20140044319; 20140044320; 20140044321; 20140072074; 20140077981; 20140081157; 20140086356; 20140086361; 20140095129; 20140107832; 20140126670; 20140126675; 20140133848; 20140140250; 20140161207; 20140167704; 20140172338; 20140198959; 20140213919; 20140225451; 20140226828; 20140229132; 20140247906; 20140266431; 20140269857; 20140269970; 20140269989; 20140269990; 20140270405; 20140278303; 20140279778; 20140292406; 20140292412; 20140294119; 20140294252; 20140313946; 20140314176; 20140314181; 20140314182; 20140317163; 20140323891; 20140333376; 20140372091; 20150003625; 20150005902; 20150016567; 20150018632; 20150025328; 20150031317; 20150031969; 20150032788; 20150043678; 20150051513; 20150061911; 20150070089; 20150077180; 20150078484; 20150092830; 20150098710; 20150104196; 20150131757; 20150156003; 20150156004; 20150162881; 20150172081; 20150180495; 20150193565; 20150193666; 20150194989; 20150202440; 20150214987; 20150215937; 20150223748; 20150241996; 20150249889; 20150256216; 20150270856; 20150270865; 20150288375; 20150295643; 20150311927; 20150311973; 20150311985; 20150322647; 20150326190; 20150333781; 20150357975; 20150358042; 20150358191; 20150381216; 20150381220; 20150381821; 20160005419; 20160022161; 20160028433; 20160034421; 20160036472; 20160036528; 20160065311; 20160079933; 20160087604; 20160087657; 20160099776; 20160111110; 20160117430; 20160124903; 20160126903; 20160127113; 20160132735; 20160134380; 20160156375; 20160162042; 20160173117; 20160191020; 20160218752; 20160225385; 20160241277; 20160248531; 20160259960; 20160261241; 20160269210; 20160287871; 20160308619; 20160309042; 20160316283; 20160329927; 20160334466; 20160336762; 20160352427; 20160359552; 20160373212; 20160380661; 20160380700; 20170012585; 20170012709; 20170014032; 20170032184; 20170033809; 20170041124; 20170043166; 20170047899; 20170061045; 20170063312; 20170077944; 20170077945; 20170078023; 20170078027; 20170093497; 20170095195; 20170104503; 20170108943; 20170117854; 20170141807; 20170141938; 20170163465; 20170170999; 20170180061; 20170195053; 20170207934; 20170214468; 20170214470; 20170222717; 20170244582; 20170245054; 20170245079; 20170255593; 20170272283; 20170304625; 20170322243; 20170324421; 20170338841; 20170338842; 20170339569; 20170346510; 20170366209; 20170366259; 20170373647; 20170373759; 20180013456; 20180013495; 20180026586; 20180026673; 20180041219; 20180062674; 20180070394; 20180102850; 20180131502; 20180167042; 20180167092; 20180167093; 20180180420; 20180191448; 20180219566; 20180254769; 20180262370; 20180269988; 20180279197; 20180294879; 20180294884; 20180302111; 20180309465; 20180316320; 20180331814; 20180333580; 20180367219; 20190007075; 20190013867; 20190013874; 20190013991; 20190020415; 20190028131; 20190030334; 20190036622; and 20190042536.

5G REFERENCES

(47) U.S. Pat. Nos. 6,675,125; 6,778,966; 7,027,981; 7,190,292; 7,206,420; 7,212,640; 7,558,391; 7,865,177; 8,085,943; 8,599,014; 8,725,706; 8,776,625; 8,898,567; 8,989,762; 9,160,579; 9,203,654; 9,235,268; 9,401,823; 9,432,564; 9,460,617; 9,531,427; 9,544,006; 9,564,927; 9,565,045; 9,613,408; 9,621,387; 9,660,851; 9,680,670; 9,686,112; 9,712,238; 9,712,350; 9,712,354; 9,713,019; 9,722,318; 9,729,281; 9,729,378; 9,742,521; 9,749,083; 9,774,476; 9,859,981; 9,871,679; 9,876,530; 9,877,206; 9,882,608; 9,893,919; 9,899,182; 9,900,048; 9,900,122; 9,900,123; 9,900,190; 9,912,436; 9,929,755; 9,942,074; 9,998,172; 9,998,187; 10,003,364; 10,027,397; 10,027,427; 10,027,523; 10,033,107; 10,033,108; 10,050,815; 10,051,483; 10,051,488; 10,062,970; 10,063,354; 10,069,467; 10,069,535; 10,079,652; 10,084,562; 10,090,594; 10,096,883; 10,103,777; 10,123,217; 10,129,057; 10,135,145; 10,148,016; 10,148,360; 10,168,501; 10,170,840; 10,171,158; 10,191,376; 10,198,582; 10,200,106; 10,205,212; 10,205,231; 10,205,482; 10,205,655; 10,211,855; 10,212,014; 10,218,405; 10,224,634; 20020051546; 20020085725; 20020103619; 20020172374; 20020172376; 20020172378; 20030035549; 20030055635; 20030098805; 20030112088; 20090221257; 20110238690; 20110249024; 20110252320; 20110288457; 20120112908; 20130110974; 20130201316; 20140226035; 20150146805; 20150146806; 20150230105; 20150280945; 20150310739; 20160093029; 20160149665; 20160149731; 20160197642; 20160218891; 20160226681; 20160352361; 20160352362; 20160352419; 20170012862; 20170018851; 20170018852; 20170019131; 20170026095; 20170032129; 20170033465; 20170033466; 20170033953; 20170033954; 20170063430; 20170078400; 20170085003; 20170085336; 20170093693; 20170104617; 20170110795; 20170110804; 20170111805; 20170134205; 20170201288; 20170229782; 20170230083; 20170245157; 20170269481; 20170271117; 20170288917; 20170295048; 20170311307; 20170317781; 20170317782; 20170317783; 20170317858; 20170318482; 20170331899; 20180013452; 20180034912; 20180048497; 20180054232; 20180054233; 20180054234; 20180054268; 20180062886; 20180069594; 20180069731; 20180076947; 20180076979; 20180076982; 20180076988; 20180091195; 20180115040; 20180115058; 20180123256; 20180123257; 20180123749; 20180123836; 20180123856; 20180123897; 20180124181; 20180131406; 20180131541; 20180145411; 20180145412; 20180145414; 20180145415; 20180151957; 20180152262; 20180152330; 20180152925; 20180159195; 20180159196; 20180159197; 20180159228; 20180159229; 20180159230; 20180159232; 20180159240; 20180159243; 20180159615; 20180166761; 20180166784; 20180166785; 20180166787; 20180167105; 20180167148; 20180175892; 20180175978; 20180198668; 20180205399; 20180205481; 20180227158; 20180248592; 20180254754; 20180254924; 20180262243; 20180278693; 20180278694; 20180294897; 20180301812; 20180302145; 20180309206; 20180323826; 20180324005; 20180324006; 20180324021; 20180324601; 20180331413; 20180331720; 20180331721; 20180331871; 20180343304; 20180351687; 20180358678; 20180359126; 20180375940; 20190013577; 20190013837; 20190013838; 20190020530; 20190036222; 20190052505; 20190074563; 20190074564; 20190074565; 20190074568; 20190074580; 20190074584; 20190074597; 20190074598; 20190074864; 20190074865; and 20190074878.