Integrating volterra series model and deep neural networks to equalize nonlinear power amplifiers
US PatentContact
11855813 ยท 2023-12-26
Assignee
Inventors
Cpc classification
H03F1/32
ELECTRICITY
H03F3/20
ELECTRICITY
International classification
H03F1/32
ELECTRICITY
H03F3/20
ELECTRICITY
H03H7/40
ELECTRICITY
H03K5/159
ELECTRICITY
Abstract
The nonlinearity of power amplifiers (PAs) has been a severe constraint in performance of modern wireless transceivers. This problem is even more challenging for the fifth generation (5G) cellular system since 5G signals have extremely high peak to average power ratio. Nonlinear equalizers that exploit both deep neural networks (DNNs) and Volterra series models are provided to mitigate PA nonlinear distortions. The DNN equalizer architecture consists of multiple convolutional layers. The input features are designed according to the Volterra series model of nonlinear PAs. This enables the DNN equalizer to effectively mitigate nonlinear PA distortions while avoiding over-fitting under limited training data. The non-linear equalizers demonstrate superior performance over conventional nonlinear equalization approaches.
Claims
1. A distortion-compensating processor, comprising: at least one automated processor configured to decompose a non-linearly distorted signal derived from an information signal, received from a channel having a channel non-linear distortion into a truncated series expansion of at least third order with memory comprising a series of terms, each term representing incremental non-linearity order and associated delay; an adaptive multi-layer feedforward deep neural network comprising a plurality of hidden layers, and at least one dropout layer, receiving as inputs the series of terms, and producing an equalized output signal; and an output port configured to present the equalized output signal, the multi-layer feedforward deep neural network being trained with respect to the channel non-linear distortion associated with communication of a series of symbols using training data comprising the series of terms, to equalize the signal, the multi-layer feedforward deep neural network being configured to receive the respective terms associated with incremental non-linearity orders and associated delay values, and to selectively produce the equalized output signal, representing the information signal wherein the channel non-linear distortion is reduced.
2. The distortion-compensating processor according to claim 1, wherein the training data comprises a set of small amplitude training signals to estimate a channel response and a set of large amplitude training signals to estimate a power amplifier non-linearity.
3. The distortion-compensating processor according to claim 1, wherein the information signal is distorted by amplification by a radio frequency power amplifier and transmission through a radio frequency communication channel, wherein the non-linearly distorted signal is received by the at least one automated processor from a radio receiver.
4. The distortion-compensating processor according to claim 1, wherein the series expansion of at least third order with memory comprises a Volterra series expansion.
5. The distortion-compensating processor according to claim 4, wherein the terms of the Volterra series expansion are defined by:
6. The distortion-compensating processor according to claim 1, wherein: x(n) is a signal sequence representing information in the non-linearly distorted signal y(n) distorted by an analog process within a channel h; .sub. is a set of finite-impulse response channel coefficients;
is an equalization delay; v(n) is an additive white Gaussian noise component signal sequence; H is an (M+1)(M+L+1) dimensional channel matrix
) for some equalization delay
; g.sub.kd are coefficients determined according to
, analog process responses V.sub.y, .sub.y, and the channel equalizer .sup.T.
7. The distortion-compensating processor according to claim 6, wherein:
8. The distortion-compensating processor according to claim 1, wherein the truncated series expansion of at least third order with memory comprises at least fifth order terms, and the deep multi-layer feedforward neural network has at least two convolutional network layers.
9. The distortion-compensating processor according to claim 1, wherein the deep multi-layer feedforward neural network comprises at least three hidden layers, each hidden layer comprising at least 10 feature maps, and a fully connected layer subsequent to the at least three hidden layers.
10. The distortion-compensating processor according to claim 1, wherein the non-linearly distorted signal comprises a frequency division multiplexed radio frequency modulated set of signals representing the information signal distorted by a radio frequency power amplifier; further comprising a frequency division multiplexed radio frequency signal demodulator configured to demodulate the equalized output signal as the set of symbols.
11. A method of compensating for a distortion, comprising: decomposing a non-linearly distorted signal received from a channel having a channel non-linear distortion into a truncated series expansion of at least third order with memory, based on an information signal communicated through the channel, the decomposition comprising a series of terms, each term representing incremental non-linearity order and associated delay, using at least one automated processor; equalizing the non-linearly distorted signal with an automated equalizer comprising a multi-layer feedforward deep neural network comprising a plurality of hidden layers and at least one dropout layer, by receiving coefficients of respective terms associated with respective incremental non-linearity orders and associated delay by the automated equalizer, and producing selectively from the equalizer an output signal representing the information signal wherein the channel non-linear distortion is reduced; and updating the multi-layer feedforward deep neural network to reduce error of the output signal with respect to the information signal.
12. The method according to claim 11, wherein the non-linearly distorted signal comprises a frequency division multiplexed signal amplified and distorted by a radio frequency power amplifier, received though a radio frequency receiver, further comprising demodulating information contained within the frequency division multiplexed signal from the output signal.
13. The method according to claim 11, further comprising outputting the output signal, wherein the truncated series expansion of at least third order with memory comprises at least fifth order terms and the multi-layer feedforward deep neural network comprises at least two hidden layers.
14. The method according to claim 11, wherein the series expansion of at least third order with memory comprises a Volterra series expansion having terms defined by:
15. The method according to claim 11, wherein: x(n) is a signal sequence representing information in the signal y(n) distorted by an analog process within a channel h; .sub. is a set of finite-impulse response channel coefficients;
is an equalization delay; v(n) is an additive white Gaussian noise component signal sequence; H is an (M+1)(M+L+1) dimensional channel matrix
) for some equalization delay
; g.sub.kd are coefficients determined according to
, analog process responses V.sub.y, .sub.y, and the channel equalizer .sup.T.
16. The method according to claim 11, wherein the multi-layer feedforward deep neural network comprises at least one convolutional layer, followed by a fully connected layer with dropout comprising the at least one dropout layer, and an output layer.
17. The method according to claim 11, wherein the multi-layer feedforward deep neural network comprises at least three one-dimensional convolutional layers each comprising at least 10 feature maps, followed by a first fully-connected layer with the at least one dropout layer provided for regularization, and a second fully-connected layer which produces the output with respect to a delay, the multi-layer feedforward deep neural network producing an output tensor having two dimensions, wherein the at three one-dimensional convolutional layers and the first fully connected layer use a sigmoid activation function, and the fully connected output layer uses a linear activation function.
18. A non-linear distortion-compensating processor, comprising: an input signal processor configured to decompose a received non-linearly distorted signal based on an information signal communicated through a channel, into a truncated Volterra series expansion, the truncated Volterra series expansion comprising a series of terms, each term comprising a sum of multidimensional convolutions of at least third order each with an associated time delay component; and a multi-layer feedforward deep neural network comprising a plurality of hidden neural network layers comprising at least one convolutional neural network layer and at least one dropout layer, trained with respect to a non-linear distortion of the information signal represented in the non-linearly distorted signal, and updated dependent on equalizer error, to receive the series of terms of the truncated Volterra series expansion, and to selectively produce an output signal representing the non-linearly distorted signal at least partially compensated for the non-linear distortion.
19. The non-linear distortion-compensating processor according to claim 18, further comprising a demodulator configured to demodulate information modulated in the non-linearly distorted signal, wherein the non-linearly distorted signal comprises a frequency division multiplexed signal distorted by a power amplifier and a communication channel.
20. The non-linear distortion-compensating processor according to claim 18, wherein the multi-layer feedforward deep neural network comprises a plurality of convolutional layers, each convolutional layer comprising at least 10 feature maps, followed by a first fully connected layer with the at least one dropout layer provided for regularization, and a second fully connected layer for output, wherein the plurality of convolutional layers and the first fully connected layer use a non-linear activation function, and the second fully connected layer uses a linear activation function.
21. A method of compensating for a distortion, comprising: decomposing a signal received from a channel having a channel non-linear distortion into a truncated series expansion of at least third order with memory, using at least one automated processor; and equalizing the signal with an automated equalizer comprising a multi-layer feedforward deep neural network trained having at least one convolutional layer, followed by a fully connected layer with dropout, and an output layer, with respect to the channel non-linear distortion, by receiving terms of the truncated series expansion of at least third order with memory by the automated equalizer, and producing selectively from the equalizer an output representing the signal wherein the channel non-linear distortion is reduced.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
(1)
(2)
(3)
(4)
(5)
DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS
Volterra-Based DNN Equalizer
(6) The present technology therefore employs deep neural networks to implement the nonlinear equalizer in the receiver, which can mitigate the nonlinear effects of the received signals due to not only PAs but also nonlinear channels and propagations. The architecture of the DNN equalizer is shown in
(7) Different from [10], multi-layer convolutional neural networks (CNNs) are employed. Different from conventional neural network predistorters proposed in [6], neural networks are used as equalizers at the receivers. Different from conventional neural network equalizers such as those proposed in [14] [15], in the present DNN equalizer, not only the linear delayed samples r(n), but also the CNN and the input features in X are used. The Volterra series models are applied to create input features.
(8) We can assume that the linear channel H has already been equalized by a linear equalizer, whose output signal is r(n). In fact, this equalization is not required, but simplifies the presentation of the analysis.
(9) According to Volterra series representation of nonlinear functions, the input-output response of the nonlinear equalizer can be written as
(10)
(11) One of major problems is that the number of coefficients .sub.d.sub.
(12) This technique can drastically reduce the total number of coefficients. In [24] [25] and [26], a dynamic deviation model was developed to reduce the full Volterra series model (20) to the following simplified one:
(13)
(14) where .sub.s(n) is the static term, and
.sub.d(n) is the dynamic term that includes all the memory effects. We can see that the total number of coefficients can be much reduced by controlling the dynamic order j which is a selectable parameter.
(15) We construct the input features of the DNN based on the model (21). Corresponding to the static term .sub.s(n), we change it to:
(16)
(17) The reason that (22) changes r.sup.k(n) to r(n)|r(n)|.sup.k-1 is that only the signal frequency within the valid passband is interested. This means the input feature vector X should include terms r(n)|r(n)|.sup.k-1. Similarly, corresponding to the dynamic term .sub.d(n), we need to supply r.sup.k-j(n).sub.i1.sup.jr(nd.sub.i) in the features where half of the terms r(n) and r(nd.sub.i) should be conjugated. For simplicity, in the DNN equalizer, the vector X includes r(nq)|r(nq)|.sup.k-1 for some q and k.
(18) By applying Volterra series components directly as features of the input X, the DNN can develop more complex nonlinear functions with a fewer number of hidden layers and a fewer number of neurons. This will also make the training procedure converge much faster with much less training data.
(19) In (n), where
(n)={circumflex over (x)}(nd) for some appropriate delay d. All the convolutional layers and the first fully connected layer use the sigmoid activation function, while the output layer uses the linear activation function. The mean square error loss function L.sup.loss=E[|x(nd)
(n)|.sup.2] is used, where
(n) is replaced by Y and x(nd) is replaced by training data labels.
Experiment Evaluations
(20) Experiments are presented on applying the Volterra series based DNN equalizer (Volterra+NN) for nonlinear PA equalization. The (Volterra+NN) scheme with the following equalization methods: a Volterra series-based equalizer (Volterra) and a conventional time-delay neural network equalizer (NN). The performance metrics are mean square error (MSE)
{square root over (E[|z(n)x(nd)|.sup.2]/E[|x(nd)|.sup.2])}
(21) and symbol error rate (SER).
(22) Both simulated signals and real measurement signals were employed. To generate simulated signals, a Doherty nonlinear PA model consisting of 3rd and 5th order nonlinearities was employed. Referring to (2), the coefficients b.sub.k,q were
b.sub.0,0:2={1.0513+0.0904j,0.0680.0023j,0.02890.0054j}
b.sub.2,0:2={0.05420.29j,0.2234+0.2317j,0.06210.0932j}
b.sub.4,0:2={0.96570.7028j,0.24510.3735j,0.1229+0.1508j},
(23) which was used in [5] to simulate a 5th order dominant nonlinear distortion derived from PA devices used in the satellite industry. For real measurement, our measurement signals were obtained from PA devices used in the cable TV (CATV) industry, which are typically dominated by 3.sup.rd order nonlinear distortion (NLD). Various levels of nonlinear distortion, in terms of dBc, were generated by adjusting the PAs.
(24) For the Volterra equalizer, the approximate response of the nonlinear equalizer with delays including 8 pre- and post-main taps and with nonlinearities including even and odd order nonlinearity up to the 5th order was employed. To determine the values of the Volterra coefficients, N=4; 096 training symbols were transmitted through the PA and then collected the noisy received samples r(n).
(25) For the conventional time-delay NN equalizer, a feedforward neural network with an 80-dimensional input vector X and 5 fully-connected hidden layers with 20, 20, 10, 10, 10 neurons, respectively, was applied.
(26)
(27)
(28)
(29) Table 2 summarizes equalization performance, which shows the averaged percent reduction/improvement in MSE and SER from the NLD impaired data for multiple modulation orders. Note that 0% SER improvement for QPSK was because the received signal's SER was already very low.
(30) TABLE-US-00002 TABLE 2 EQ-MSE/SER improvement in percentage over measured NLD Volterra NN Volterra + NN MSE SER MSE SER MSE SER 64-QAM 16% 26% 10% 25% 42% 44% 16-QAM 41% 2% 35% 6% 85% 28% QPSK 57% 0% 100% 0% 100% 0% Average 38% 9% 48% 10% 76% 24%
(31) The nonlinear equalization scheme presented by integrating the Volterra series nonlinear model with deep neural networks yields superior results over conventional nonlinear equalization approaches in mitigating nonlinear power amplifier distortions. It finds application for many 5G communication scenarios.
(32) The technology may be implemented as an additional component in a receiver, or within the digital processing signal chain of a modern radio. A radio is described in US 20180262217, expressly incorporated herein by reference.
(33) In an implementation, a base station may include a SDR receiver configured to allow the base station to operate as an auxiliary receiver. In an example implementation, the base station may include a wideband receiver bank and a digital physical/media access control (PHY/MAC) layer receiver. In this example, the SDR receiver may use a protocol analyzer to determine the protocol used by the source device on the uplink to the primary base station, and then configure the digital PHY/MAC layer receiver for that protocol when operating as art auxiliary receiver. Also, the digital PHY/MAC layer receiver may be configured to operate according to another protocol when operating as a primary base station. In another example, the base station may include a receiver hank for a wireless system, for example, a fifth Generation (5G) receiver bank, and include an additional receiver having SDR configurable capability. The additional receiver may be, for example, a digital Wi-Fi receiver configurable to operate according to various Wi-Fi protocols. The base station may use a protocol analyzer to determine the particular Wi-Fi protocol used by the source device on the uplink to the primary base station. The base station may then configure the additional receiver as the auxiliary receiver for that Wi-Fi protocol.
(34) Depending on the hardware configuration, a receiver may be used to flexibly provide uplink support in systems operating according to one or more protocols such as the various IEEE 802.11 Wi-Fi protocols, 3.sup.rd Generation Cellular (3G), 4.sup.th Generation Cellular (4G) wide band code division multiple access (WCDMA), Long Term Evolution (LTE) Cellular, and 5.sup.th generation cellular (5G).
(35) See, 5G References, infra.
(36) Processing unit may comprise one or more processors, or other control circuitry or any combination of processors and control circuitry that provide, overall control according to the disclosed embodiments. Memory may be implemented as any type of as any type of computer readable storage media, including non-volatile and volatile memory.
(37) The example embodiments disclosed herein may be described in the general context of processor-executable code or instructions stored on memory that may comprise one or more computer readable storage media (e.g., tangible non-transitory computer-readable storage media such as memory). As should be readily understood, the terms computer-readable storage media or non-transitory computer-readable media include the media for storing of data, code and program instructions, such as memory, and do not include portions of the media for storing transitory propagated or modulated data communication signals.
(38) While the functionality disclosed herein has been described by illustrative example using descriptions of the various components and devices of embodiments by referring to functional blocks and processors or processing units, controllers, and memory including instructions and code, the functions and processes of the embodiments may be implemented and performed using any type of processor, circuit, circuitry or combinations of processors and or circuitry and code. This may include, at least in part, one or more hardware logic components. For example, and without limitation, illustrative types of hardware logic components that can be used include field programmable gate arrays (FPGAs), application specific integrated circuits (ASICs), application specific standard products (ASSPs), system-on-a-chip systems (SOCs), complex programmable logic devices (CPLDs), etc. Use of the term processor or processing unit in this disclosure is mean to include all such implementations.
(39) The disclosed implementations include a receiver, one or more processors in communication with the receiver, and memory in communication with the one or more processors, the memory comprising code that, when executed, causes the one or more processors to control the receiver to implement various features and methods according to the present technology.
(40) Although the subject matter has been described in language specific to structural features and/or methodological acts, it is to be understood that the subject matter defined in the appended claims is not necessarily limited to the specific features or acts described above. Rather, the specific features and acts described above are disclosed as example embodiments, implementations, and forms of implementing the claims and these example configurations and arrangements may be changed significantly without departing from the scope of the present disclosure. Moreover, although the example embodiments have been illustrated with reference to particular elements and operations that facilitate the processes, these elements, and operations may be combined with or, be replaced by, any suitable devices, components, architecture or process that achieves the intended functionality of the embodiment. Numerous other changes, substitutions, variations, alterations, and modifications may be ascertained to one skilled in the art and it is intended that the present disclosure encompass all such changes, substitutions, variations, alterations, and modifications a falling within the scope of the appended claims.
REFERENCES
(41) [1] J.-A. Lucciardi, P. Potier, G. Buscarlet, F. Barrami, and G. Mesnager, Non-linearized amplifier and advanced mitigation techniques: Dvbs-2 spectral efficiency improvement, in GLOBECOM 2017-2017 IEEE Global Communications Conference. IEEE, 2017, pp. 1-7. [2] J. Wood, Behavioral modeling and linearization of RF power amplifiers. Artech House, 2014. [3] C.-L. Wang and Y. Ouyang, Low-complexity selected mapping schemes for peak-to-average power ratio reduction in ofdm systems, IEEE Transactions on signal processing, vol. 53, no. 12, pp. 4652-4660, 2005. [4] J. Kim and K. Konstantinou, Digital predistortion of wideband signals based on power amplifier model with memory, Electronics Letters, vol. 37, no. 23, pp. 1417-1418, 2001. [5] L. Ding, G. T. Zhou, D. R. Morgan, Z. Ma, J. S. Kenney, J. Kim, and C. R. Giardina, A robust digital baseband predistorter constructed using memory polynomials, IEEE Transactions on communications, vol. 52, no. 1, pp. 159-165, 2004. [6] M. Rawat, K. Rawat, and F. M. Ghannouchi, Adaptive digital predistortion of wireless power amplifiers/transmitters using dynamic realvalued focused time-delay line neural networks, IEEE Transactions on Microwave Theory and Techniques, vol. 58, no. 1, pp. 95-104, 2010. [7] S. Dimitrov, Non-linear distortion cancellation and symbol-based equalization in satellite forward links, IEEE Trans Wireless Commun, vol. 16, no. 7, pp. 4489-4502, 2017. [8] D. J. Sebald and J. A. Bucklew, Support vector machine techniques for nonlinear equalization, IEEE Transactions on Signal Processing, vol. 48, no. 11, pp. 3217-3226, 2000. [9] S. Chen, B. Mulgrew, and P. M. Grant, A clustering technique for digital communications channel equalization using radial basis function networks, IEEE Transactions on neural networks, vol. 4, no. 4, pp. 570-590, 1993. [10] B. Li, C. Zhao, M. Sun, H. Zhang, Z. Zhou, and A. Nallanathan, A bayesian approach for nonlinear equalization and signal detection in millimeter-wave communications, IEEE Transactions on Wireless Communications, vol. 14, no. 7, pp. 3794-3809, 2015. [11] F. Mkadem and S. Boumaiza, Physically inspired neural network model for rf power amplifier behavioral modeling and digital predistortion, IEEE Transactions on Microwave Theory and Techniques, vol. 59, no. 4, pp. 913-923, 2011. [12] T. Liu, S. Boumaiza, and F. M. Ghannouchi, Dynamic behavioral modeling of 3 g power amplifiers using real-valued time-delay neural networks, IEEE Transactions on Microwave Theory and Techniques, vol. 52, no. 3, pp. 1025-1033, 2004. [13] M. Ibnkahla, Applications of neural networks to digital communications-a survey, Signal processing, vol. 80, no. 7, pp. 1185-1215, 2000. [14] D.-C. Park and T.-K. J. Jeong, Complex-bilinear recurrent neural network for equalization of a digital satellite channel, IEEE Transactions on Neural Networks, vol. 13, no. 3, pp. 711-725, 2002. [15] A. Uncini, L. Vecci, P. Campolucci, and F. Piazza, Complex-valued neural networks with adaptive spline activation function for digital-radio-links nonlinear equalization, IEEE Transactions on Signal Processing, vol. 47, no. 2, pp. 505-514, 1999. [16] M. S. Sim, M. Chung, D. Kim, J. Chung, D. K. Kim, and C.-B. Chae, Nonlinear self-interference cancellation for full-duplex radios: From link-level and system-level performance perspectives, IEEE Communications Magazine, vol. 55, no. 9, pp. 158-167, 2017. [17] I. Yoffe and D. Wulich, Predistorter for mimo system with nonlinear power amplifiers, IEEE Transactions on Communications, vol. 65, no. 8, pp. 3288-3301, 2017. [18] M. Abdelaziz, L. Anttila, and M. Valkama, Reduced-complexity digital predistortion for massive mimo, in Acoustics, Speech and Signal Processing (ICASSP), 2017 IEEE International Conference on. IEEE, 2017, pp. 6478-6482. [19] H. Yan and D. Cabric, Digital predistortion for hybrid precoding architecture in millimeter-wave massive mimo systems, in Acoustics, Speech and Signal Processing (ICASSP), 2017 IEEE International Conference on. IEEE, 2017, pp. 3479-3483. [20] C. Molln, E. G. Larsson, and T. Eriksson, Waveforms for the massive mimo downlink: Amplifier efficiency, distortion, and performance, IEEE Transactions on Communications, vol. 64, no. 12, pp. 5050-5063, 2016. [21] A. Cheaito, M. Crussire, J.-F. Helard, and Y. Lout, Quantifying the memory effects of power amplifiers: Evm closed-form derivations of multicarrier signals. IEEE Wireless Commun. Letters, vol. 6, no. 1, pp. 34-37, 2017. [22] K. Simons, Technical Handbook for CATV Systems, 3rd Edition. Jerrod Publication No. 436-001-01, 1968. [23] J. Staudinger, J.-C. Nanan, and J. Wood, Memory fading volterra series model for high power infrastructure amplifiers, in Radio and Wireless Symposium (RWS), 2010 IEEE. IEEE, 2010, pp. 184-187. [24] A. Zhu, J. C. Pedro, and T. J. Brazil, Dynamic deviation reduction-based volterra behavioral modeling of rf power amplifiers, IEEE Transactions on microwave theory and techniques, vol. 54, no. 12, pp. 4323-4332, 2006. [25] A. Zhu, P. J. Draxler, J. J. Yan, T. J. Brazil, D. F. Kimball, and P. M. Asbeck, Open-loop digital predistorter for rf power amplifiers using dynamic deviation reduction-based volterra series, IEEE Transactions on Microwave Theory and Techniques, vol. 56, no. 7, pp. 1524-1534, 2008. [26] L. Guan and A. Zhu, Simplified dynamic deviation reduction-based volterra model for doherty power amplifiers, in Integrated Nonlinear Microwave and Millimetre-Wave Circuits (INMMIC), 2011 Workshop on. IEEE, 2011, pp. 1-4. [27] Schetzen, M. The Volterra and Wiener Theories of Non-linear Systems. (1980) Wiley & Sons. [28] Black, H. S. [October, 1928] Translating system. U.S. Pat. No. 1,686,792. [29] Black, H. S. [December, 1937] Wave translating system. U.S. Pat. No. 2,102,671. [30] Mitchell, A. F. [November, 1979] A 135 MHz feedback amplifier. IEEE Colloq. Broadband High Frequency Amplifiers. [31] Arthanayake, T. and Wood, H. B. [8 Apr. 1971] Linear amplification using envelope feedback. Elec. Lett. [32] Chadwick, P. [1986] Wideband Amplifier Applications Book, Edition 2, Plessey Semiconductor.
Amplifier References
(42) Aghvami, A. H. and Robertson, I. D. [April, 1993] Power limitation and high-power amplifier non linearities in on-board satellite communications systems. Electron. and Comm. Engin. J. Arthanayake, T. and Wood, H. B. [8 Apr. 1971] Linear amplification using envelope feedback. Elec. Lett. Bennet, T. J. and Clements, R. F. [May, 1974] FeedforwardAn alternative approach to amplifier linearisation. Radio and Electron. Engin. Bhargava, V. K. et al. [1981] Digital Communications by Satellite, John Wiley and Sons. Black, H. S. [December, 1937] Wave translating system. U.S. Pat. No. 2,102,671. Black, H. S. [October, 1928] Translating system. U.S. Pat. No. 1,686,792. Bond F. E. and Meyer, H. F. [April, 1970] Intermodulation effects in limiter amplifier repeaters. IEEE Trans. Comm., Vol. COM-18, p. 127-135. Chadwick, P. [1986] Wideband Amplifier Applications Book, Edition 2, Plessey Semiconductor. Cole, R. A. [December, 1989] Linearisation of a power amplifier using Cartesian Loop feedback. Report No. 72/89/R/451/C. Roke Manor Res. ETSI [August, 1994] Standard ETR 132. Radio broadcasting systems; Code of practice for site engineering VHF FM sound broadcasting transmitters. European Telecommunications Standards Institute, Sophia Antipolis, F-06291, Valbonne Cedex, France. ETSI [January, 1995] European Standard ETS 300 384. Radio broadcasting systems; Very high frequency (VHF), frequency modulated, sound broadcasting transmitters. European Telecommunications Standards Institute, Sophia Antipolis, F-06291, Valbonne Cedex, France. ETSI [June, 1998] Standard ETR 053 Ed 3Radio site engineering for equipment and systems in the mobile service. European Telecommunications Standards Institute, Sophia Antipolis, F-06291, Valbonne Cedex, France. ETSI [March, 1997] European Standard ETS 300 113. Radio equipment and systems (RES); Land mobile service; Technical characteristics and test conditions for radio equipment intended for the transmission of data (and speech) and having an antenna connector. European Telecommunications Standards Institute, Sophia Antipolis, F-06291, Valbonne Cedex, France. Gray, L. F. [1980] Application of broadband linearisers to satellite transponders. IEEE Conf. Proc. ICC'80. Heathman, A. C. [1989] Methods for intermodulation prediction in communication systems. Ph. D. Thesis, University of Bradford, United Kingdom. IESS [November, 1996] IESS-401 (Rev. 4). Performance requirements for intermodulation products transmitted from INTELSAT earth stations. Intelsat Earth Station Standard (IESS). Kaeadar, K. [December, 1986] Gaussian white-noise generation for digital signal synthesis. IEEE Trans. Inst. and Meas., Vol. IM 35, 4. Kahn, L. R. [July, 1952] SSB transmission by envelope elimination and restoration. Proc. IRE. Mitchell, A. F. [November, 1979] A 135 MHz feedback amplifier. IEEE Colloq. Broadband High Frequency Amplifiers. Pavliouk, A. [1977] Unification of measurement procedures for out-of-band emission spectra and peak envelope power of single-sideband radio transmitter measurements. Proc. of the NIIR, 4 (in Russian). Petrovic, V. and Gosling, W. [10 May 1979] Polar loop transmitter. Elec. Lett. Pye Telecom [November, 1978] Intermodulation in VHF and UHF radio systemslocating and minimizing the effects. Engineering Notes, Pub. Ref. No. TSP480/1, United Kingdom. Radiocommunications Agency [April, 1987] Code of practice for radio site engineering. MPT 1331. Radiocommunications Agency (RA), Flyde Microsystems Ltd. United Kingdom. Saleh, A. M. [May 1982] Intermodulation analysis of FDMA satellite systems employing compensated and uncompensated TWTs. IEEE Trans. Comm., Vol. COM-30, 5. Schetzen, M. [1980] The Volterra and Wiener Theories of Non-linear Systems. Wiley & Sons. Schetzen, M. [1980] The Volterra and Wiener Theories of Non-linear Systems. Wiley & Sons. Shahid, M., Shepherd, S. J., Lin, B., Khairruddin, I., and Barton, S. K. [December, 1996] Study of methods of measuring multi-carrier intermodulation performance Report No. 581, Purchase Order No. 142379 dESA, with University of Bradford, United Kingdom. Shimbo, O. [February, 1971] Effects of intermodulation, AM-PM conversion, and additive noise in multicarrier TWT systems. Proc. IEEE, Vol. 59, p. 230-238. Smith, C. N. [1986] Application of the polar loop technique to UHF SSB transmitters. Ph.D. Thesis, University of Bath. Smith, C. N. and PETROVIC, V. [1982] Cartesian loop transmitter. Internal Research Report, University of Bath, School of Electrical and Electronic Engineering. Tondryk, W. [1991] Intermodulation testing of the INMARSAT payloadResponse to system PDR. Marconi Space Systems Ltd. Wassermann, M. et al. [1983] Study and breadboarding of an L-band high power linearized TWT amplifier. Final Report, ESTEC contract No. 5459/83/NL/GM. Wood, A. [October 1998] Radio interference: Sources and solutions. LPRA NEWS, p. 21.
Volterra Series References
(43) Bohm, D. The Special Theory of Relativity, Benjamin, 1965. Censor, D., & Melamed, T, 2002, Volterra differential constitutive operators and locality considerations in electromagnetic theory, PIERProgress in Electromagnetic Research, 36: 121-137 Censor, D., 2000, A quest for systematic constitutive formulations for general field and wave systems based on the Volterra differential operators, PIERProgress In Electromagnetics Research, (25): 261-284 Censor, D., 2001, Constitutive relations in inhomogeneous systems and the particle-field conundrum, PIERProgress In Electromagnetics Research, (30): 305-335 Schetzen, M., 1980, The Volterra and Wiener Theorems of Nonlinear Systems, New York, Chichester, Brisbane and Toronto: John Wiley and Sons Sonnenschein, M & Censor, D., 1998, Simulation of Hamiltonian light beam propagation in nonlinear media, JOSAJournal of the Optical Society of America B, (15): 1335-1345
Filter References
(44) Akaiwa, Y. Introduction to Digital Mobile Communication. New York: Wiley, 1997. Altera Corporation. Digital Predistortion Reference Design. Application Note 314, 2003. Aysal, Tuncer C., and Kenneth E. Barner, Myriad-Type Polynomial Filtering, IEEE Transactions on Signal Processing, vol. 55, no. 2, February 2007. Barner, Kenneth E., and Tuncer Can Aysal, Polynomial Weighted Median Filtering, IEEE Transactions on Signal Processing, vol. 54, no. 2, February 2006. Biglieri, Ezio, Sergio Barberis, and Maurizio Catena, Analysis and Compensation of Nonlinearities in Digital Transmission Systems, IEEE Journal on selected areas in Communications, vol. 6, no. 1, January 1988. Budura, Georgeta, and Corina Botoca, Efficient Implementation of the Third Order RLS Adaptive Volterra Filter, FACTA Universitatis (NIS) Ser.: Elec. Energ. vol. 19, no. 1, April 2006. Ding, L., et al. A Robust Digital Baseband Predistorter Constructed Using Memory Polynomials, IEEE Transactions on Communications, Vol. 52, No. 1, June 2004. Fang, Yang-Wang, Li-Cheng Jiao, Xian-Da Zhang and Jin Pan, On the Convergence of Volterra Filter Equalizers Using a Pth-Order Inverse Approach, IEEE Transactions on Signal Processing, vol. 49, no. 8, August 2001. Gurin, Alexandre, Grard Faucon, and Rgine Le Bouquin-Jeanns, Nonlinear Acoustic Echo Cancellation Based on Volterra Filters, IEEE Transactions on Speech and Audio Processing, vol. 11, no. 6, November 2003. Haykin, Simon, Adaptive Filter Theory, Fourth Edition, Pearson Education, 2008. Kamiya, N., and F. Maehara. Nonlinear Distortion Avoidance Employing Symbol-wise Transmit Power Control for OFDM Transmission, Proc. of Int'l. OFDM Workshop, Hamburg, 2009. Kim, J., and K. Konstantinou. Digital predistortion of wideband signals based on power amplifier model with memory, Electronic Letters, Vol. 37, No. 23, November 2001. Krall, Christoph, Klaus Witrisal, Geert Leus and Heinz Koeppl, Minimum Mean-Square Error Equalization for Second-Order Volterra Systems, IEEE Transactions on Signal Processing, vol. 56, no. 10, October 2008. Leis, John, Adaptive Filter Lecture Notes & Examples, Nov. 1, 2008 www.usq.edu.au/users/leis/notes/sigproc/adfilt.pdf. Lpez-Valcarce, Roberto, and Soura Dasgupta, Second-Order Statistical Properties of Nonlinearly Distorted Phase-Shift Keyed (PSK) Signals, IEEE Communications Letters, vol. 7, no. 7, July 2003. Lozhkin, Alexander N. Turbo Linearizer for High Power Amplifier. In 2011 IEEE 73rd Vehicular Technology Conference (VTC Spring), pp. 1-5. IEEE, 2011. Mathews, V. John, Adaptive Polynomial Filters, IEEE Signal Processing Magazine, Vol. 8, No. 3, July 1991. Park, Dong-Chul, and Tae-Kyun Jung Jeong, Complex-Bilinear Recurrent Neural Network for Equalization of a Digital Satellite Channel, IEEE Transactions on Neural Networks, vol. 13, no. 3, May 2002. Rai, Amrita, and Amit Kumar Kohli. Analysis of Adaptive Volterra Filters with LMS and RLS Algorithms. AKGEC Journal of Technology 2, no. 1 (2011). Therrien, Charles W., W. Kenneth Jenkins, and Xiaohui Li, Optimizing the Performance of Polynomial Adaptive Filters: Making Quadratic Filters Converge Like Linear Filters, IEEE Transactions on Signal Processing, vol. 47, no. 4, April 1999. Tsimbinos John, and Langford B. White, Error Propagation and Recovery in Decision-Feedback Equalizers for Nonlinear Channels, IEEE Transactions on Communications, vol. 49, no. 2, February 2001. Woo, Young Yun, et al. Adaptive Digital Feedback Predistortion Technique for Linearizing Power Amplifiers, IEEE Transactions on Microwave Theory and Techniques, Vol. 55, No. 5, May 2007. Zaknich, A., Principal of Adaptive Filter and Self Learning System, Springer Link 2005.
Volterra Series Patents
(45) U.S. Pat. Nos. and Published Patent Application Nos.: U.S. Pat. Nos. 4,615,038; 4,669,116; 4,870,371; 5,038,187; 5,309,481; 5,329,586; 5,424,680; 5,438,625; 5,539,774; 5,647,023; 5,692,011; 5,694,476; 5,744,969; 5,745,597; 5,790,692; 5,792,062; 5,815,585; 5,889,823; 5,924,086; 5,938,594; 5,991,023; 6,002,479; 6,005,952; 6,064,265; 6,166,599; 6,181,754; 6,201,455; 6,201,839; 6,236,837; 6,240,278; 6,288,610; 6,335,767; 6,351,740; 6,381,212; 6,393,259; 6,406,438; 6,408,079; 6,438,180; 6,453,308; 6,504,885; 6,510,257; 6,512,417; 6,532,272; 6,563,870; 6,600,794; 6,633,208; 6,636,115; 6,668,256; 6,687,235; 6,690,693; 6,697,768; 6,711,094; 6,714,481; 6,718,087; 6,775,646; 6,788,719; 6,812,792; 6,826,331; 6,839,657; 6,850,871; 6,868,380; 6,885,954; 6,895,262; 6,922,552; 6,934,655; 6,940,790; 6,947,857; 6,951,540; 6,954,476; 6,956,433; 6,982,939; 6,992,519; 6,999,201; 6,999,510; 7,007,253; 7,016,823; 7,061,943; 7,065,511; 7,071,797; 7,084,974; 7,092,043; 7,113,037; 7,123,663; 7,151,405; 7,176,757; 7,209,566; 7,212,933; 7,236,156; 7,236,212; 7,239,301; 7,239,668; 7,251,297; 7,268,620; 7,272,594; 7,286,009; 7,295,961; 7,304,591; 7,305,639; 7,308,032; 7,333,559; 7,348,844; 7,400,807; 7,403,884; 7,412,469; 7,423,699; 7,436,883; 7,443,326; 7,489,298; 7,512,900; 7,542,518; 7,551,668; 7,570,856; 7,571,401; 7,576,606; 7,589,725; 7,590,518; 7,602,240; 7,606,539; 7,610,183; 7,657,405; 7,720,232; 7,720,236; 7,728,658; 7,729,446; 7,733,177; 7,746,955; 7,755,425; 7,760,887; 7,773,692; 7,774,176; 7,795,858; 7,796,960; 7,808,315; 7,812,666; 7,821,337; 7,821,581; 7,822,146; 7,826,624; 7,847,631; 7,852,913; 7,853,443; 7,864,881; 7,873,172; 7,885,025; 7,885,797; 7,889,007; 7,894,788; 7,895,006; 7,899,416; 7,902,925; 7,903,137; 7,924,942; 7,929,375; 7,932,782; 7,970,150; 7,970,151; 7,979,837; 7,991,073; 7,991,167; 7,995,674; 8,005,858; 8,023,668; 8,031,882; 8,039,871; 8,045,066; 8,046,199; 8,065,060; 8,089,689; 8,105,270; 8,139,630; 8,148,983; 8,149,950; 8,160,191; 8,165,854; 8,170,508; 8,185,853; 8,193,566; 8,195,103; 8,199,399; 8,213,880; 8,244,787; 8,260,732; 8,265,583; 8,270,530; 8,294,605; 8,295,790; 8,306,488; 8,310,312; 8,315,970; 8,331,511; 8,331,879; 8,345,348; 8,346,692; 8,346,693; 8,346,711; 8,346,712; 8,351,876; 8,354,884; 8,355,684; 8,358,169; 8,364,095; 8,369,447; 8,369,595; 8,380,773; 8,390,375; 8,390,376; 8,396,693; 8,410,843; 8,410,850; 8,412,133; 8,421,534; 8,432,220; 8,437,513; 8,463,582; 8,467,438; 8,477,581; 8,483,343; 8,483,450; 8,487,706; 8,489,047; 8,494,463; 8,498,369; 8,509,347; 8,509,712; 8,519,440; 8,532,215; 8,532,964; 8,538,039; 8,564,368; 8,565,343; 8,577,311; 8,587,375; 8,599,050; 8,605,814; 8,605,819; 8,611,190; 8,611,459; 8,611,820; 8,615,208; 8,619,905; 8,620,631; 8,626,089; 8,649,743; 8,675,925; 8,704,595; 8,705,166; 8,712,345; 8,718,178; 8,718,209; 8,724,857; 8,737,937; 8,737,938; 8,744,141; 8,744,377; 8,758,271; 8,761,409; 8,766,917; 8,767,869; 8,780,693; 8,787,628; 8,798,559; 8,804,807; 8,804,871; 8,811,532; 8,823,452; 8,831,074; 8,831,133; 8,831,135; 8,838,218; 8,843,088; 8,843,089; 8,849,611; 8,855,175; 8,855,234; 8,867,601; 8,874,411; 8,885,765; 8,886,341; 8,891,701; 8,896,471; 8,897,351; 8,903,192; 8,909,176; 8,909,328; 8,933,752; 8,934,573; 8,958,470; 8,964,901; 8,964,996; 8,971,834; 8,976,896; 8,994,657; 8,995,571; 8,995,835; 9,008,153; 9,014,299; 9,019,643; 9,020,454; 9,025,607; 9,031,168; 9,036,734; 9,048,865; 9,048,900; 9,071,313; 9,077,508; 9,088,472; 9,094,036; 9,094,151; 9,104,921; 9,106,304; 9,130,628; 9,137,492; 9,143,274; 9,160,280; 9,160,310; 9,160,687; 9,166,610; 9,166,635; 9,166,698; 9,171,534; 9,184,784; 9,185,529; 9,189,458; 9,191,041; 9,191,049; 9,199,860; 9,209,753; 9,209,841; 9,214,968; 9,214,969; 9,225,295; 9,225,501; 9,231,530; 9,231,647; 9,231,801; 9,236,996; 9,246,525; 9,246,731; 9,252,798; 9,252,821; 9,253,608; 9,257,943; 9,258,156; 9,261,978; 9,264,153; 9,265,461; 9,270,304; 9,270,512; 9,271,123; 9,276,602; 9,294,113; 9,304,501; 9,306,606; 9,311,535; 9,312,892; 9,314,623; 9,322,906; 9,337,781; 9,337,783; 9,352,155; 9,361,681; 9,361,936; 9,362,869; 9,362,942; 9,363,068; 9,369,093; 9,369,255; 9,369,541; 9,397,516; 9,404,950; 9,413,516; 9,419,722; 9,431,972; 9,438,178; 9,438,356; 9,439,597; 9,451,920; 9,460,246; 9,461,597; 9,461,676; 9,473,077; 9,479,322; 9,509,331; 9,509,350; 9,517,030; 9,531,475; 9,536,539; 9,537,759; 9,544,126; 9,559,831; 9,564,876; 9,571,312; 9,575,570; 9,590,664; 9,590,668; 9,595,920; 9,595,982; 9,607,003; 9,607,628; 9,608,676; 9,608,718; 9,614,554; 9,628,119; 9,628,120; 9,646,116; 9,647,717; 9,654,211; 9,654,216; 9,659,120; 9,660,593; 9,660,730; 9,665,510; 9,667,292; 9,674,368; 9,680,423; 9,680,497; 9,697,845; 9,705,477; 9,706,296; 9,712,179; 9,712,233; 9,713,010; 9,722,646; 9,722,691; 9,726,701; 9,727,677; 9,735,741; 9,735,800; 9,735,811; 9,735,876; 9,737,258; 9,742,599; 9,746,506; 9,749,161; 9,755,691; 9,762268; 9,768,891; 9,778,902; 9,780,869; 9,780,881; 9,787,459; 9,794,000; 9,800,437; 9,800,734; 9,820311; 9,831,899; 9,837,970; 9,843,346; 9,859,845; 9,866,183; 9,877,265; 9,882,648; 9,887,862; 9,900088; 9,912,435; 9,913,194; 9,923,524; 9,923,640; 9,923,714; 9,928,212; 9,935,590; 9,935,645; 9,935715; 9,935,761; 9,940,938; 9,941,963; 9,953,656; 9,954,384; 9,960,794; 9,960,804; 9,960,900; 9,971920; 9,973,279; 9,974,957; 9,983,243; 9,998,223; 9,998,406; 9,999,780; 10,008,218; 10,009,050; 10,009,109; 10,009,259; 10,013,515; 10,015,593; 10,033,413; 10,033,568; 10,050,636; 10,050,710; 10,050,714; 10,063,265; 10,063,364; 10,075,201; 10,095,927; 10,097,273; 10,097,939; 10,101,370; 10,108,858; 10,110,315; 10,116,390; 10,128,955; 10,141,944; 10,142,754; 10,147,431; 10,148,417; 10,153,793; 10,181,825; 10,224,970; 20010036334; 20010051871; 20020041210; 20020060827; 20020075918; 20020126604; 20020146993; 20020161539; 20020161542; 20020169585; 20020178133; 20020181521; 20020186874; 20030046045; 20030057963; 20030063854; 20030071684; 20030142832; 20030195706; 20030223507; 20040019443; 20040044489; 20040130394; 20040136423; 20040155707; 20040179629; 20040208242; 20040258176; 20050021266; 20050021319; 20050031117; 20050031131; 20050031132; 20050031133; 20050031134; 20050031137; 20050031138; 20050031139; 20050031140; 20050049838; 20050100065; 20050141637; 20050141659; 20050174167; 20050177805; 20050180526; 20050226316; 20050237111; 20050243061; 20050253806; 20050270094; 20050271216; 20050273188; 20060039498; 20060052988; 20060083389; 20060093128; 20060095236; 20060104395; 20060104451; 20060133536; 20060209982; 20060222128; 20060239443; 20060256974; 20060262942; 20060262943; 20060264187; 20060269074; 20060269080; 20060274904; 20070005326; 20070018722; 20070030076; 20070033000; 20070063770; 20070080841; 20070133713; 20070133719; 20070136018; 20070136045; 20070152750; 20070160221; 20070168100; 20070190952; 20070229154; 20070237260; 20070247425; 20070252651; 20070252813; 20070276610; 20080001947; 20080032642; 20080129379; 20080130787; 20080130788; 20080130789; 20080152037; 20080158154; 20080158155; 20080180178; 20080240325; 20080261541; 20080283882; 20080285640; 20080293372; 20090003134; 20090027117; 20090027118; 20090058521; 20090067643; 20090072901; 20090075610; 20090094304; 20090146740; 20090153132; 20090185613; 20090256632; 20090287624; 20090289706; 20090291650; 20090302938; 20090302940; 20090318983; 20100007489; 20100033180; 20100060355; 20100090762; 20100093290; 20100094603; 20100097714; 20100114813; 20100135449; 20100148865; 20100152547; 20100156530; 20100183106; 20100194474; 20100199237; 20100254450; 20100283540; 20100292602; 20100292752; 20100311361; 20100312495; 20110003570; 20110025414; 20110028859; 20110037518; 20110054354; 20110054355; 20110064171; 20110069749; 20110081152; 20110085678; 20110087341; 20110096865; 20110102080; 20110103455; 20110110473; 20110121897; 20110125684; 20110125685; 20110125686; 20110125687; 20110140779; 20110144961; 20110149714; 20110177956; 20110181360; 20110204975; 20110211842; 20110268226; 20110270590; 20110293051; 20120007153; 20120007672; 20120027070; 20120029663; 20120086507; 20120093376; 20120098481; 20120098596; 20120119810; 20120140860; 20120147993; 20120154040; 20120154041; 20120158384; 20120165633; 20120176190; 20120176609; 20120217557; 20120229206; 20120256687; 20120259600; 20120263256; 20120306573; 20120328128; 20130005283; 20130009702; 20130015917; 20130030239; 20130034188; 20130040587; 20130044791; 20130044836; 20130093676; 20130113559; 20130114762; 20130166259; 20130170842; 20130176153; 20130207723; 20130222059; 20130243119; 20130243122; 20130243135; 20130257530; 20130271212; 20130272367; 20130285742; 20130301487; 20130303103; 20130315291; 20130321078; 20130330082; 20130336377; 20140009224; 20140029658; 20140029660; 20140030995; 20140031651; 20140036969; 20140044318; 20140044319; 20140044320; 20140044321; 20140072074; 20140077981; 20140081157; 20140086356; 20140086361; 20140095129; 20140107832; 20140126670; 20140126675; 20140133848; 20140140250; 20140161207; 20140167704; 20140172338; 20140198959; 20140213919; 20140225451; 20140226828; 20140229132; 20140247906; 20140266431; 20140269857; 20140269970; 20140269989; 20140269990; 20140270405; 20140278303; 20140279778; 20140292406; 20140292412; 20140294119; 20140294252; 20140313946; 20140314176; 20140314181; 20140314182; 20140317163; 20140323891; 20140333376; 20140372091; 20150003625; 20150005902; 20150016567; 20150018632; 20150025328; 20150031317; 20150031969; 20150032788; 20150043678; 20150051513; 20150061911; 20150070089; 20150077180; 20150078484; 20150092830; 20150098710; 20150104196; 20150131757; 20150156003; 20150156004; 20150162881; 20150172081; 20150180495; 20150193565; 20150193666; 20150194989; 20150202440; 20150214987; 20150215937; 20150223748; 20150241996; 20150249889; 20150256216; 20150270856; 20150270865; 20150288375; 20150295643; 20150311927; 20150311973; 20150311985; 20150322647; 20150326190; 20150333781; 20150357975; 20150358042; 20150358191; 20150381216; 20150381220; 20150381821; 20160005419; 20160022161; 20160028433; 20160034421; 20160036472; 20160036528; 20160065311; 20160079933; 20160087604; 20160087657; 20160099776; 20160111110; 20160117430; 20160124903; 20160126903; 20160127113; 20160132735; 20160134380; 20160156375; 20160162042; 20160173117; 20160191020; 20160218752; 20160225385; 20160241277; 20160248531; 20160259960; 20160261241; 20160269210; 20160287871; 20160308619; 20160309042; 20160316283; 20160329927; 20160334466; 20160336762; 20160352427; 20160359552; 20160373212; 20160380661; 20160380700; 20170012585; 20170012709; 20170014032; 20170032184; 20170033809; 20170041124; 20170043166; 20170047899; 20170061045; 20170063312; 20170077944; 20170077945; 20170078023; 20170078027; 20170093497; 20170095195; 20170104503; 20170108943; 20170117854; 20170141807; 20170141938; 20170163465; 20170170999; 20170180061; 20170195053; 20170207934; 20170214468; 20170214470; 20170222717; 20170244582; 20170245054; 20170245079; 20170255593; 20170272283; 20170304625; 20170322243; 20170324421; 20170338841; 20170338842; 20170339569; 20170346510; 20170366209; 20170366259; 20170373647; 20170373759; 20180013456; 20180013495; 20180026586; 20180026673; 20180041219; 20180062674; 20180070394; 20180102850; 20180131502; 20180167042; 20180167092; 20180167093; 20180180420; 20180191448; 20180219566; 20180254769; 20180262370; 20180269988; 20180279197; 20180294879; 20180294884; 20180302111; 20180309465; 20180316320; 20180331814; 20180333580; 20180367219; 20190007075; 20190013867; 20190013874; 20190013991; 20190020415; 20190028131; 20190030334; 20190036622; and 20190042536.
5G References
(46) U.S. Pat. Nos. 6,675,125; 6,778,966; 7,027,981; 7,190,292; 7,206,420; 7,212,640; 7,558,391; 7,865,177; 8,085,943; 8,599,014; 8,725,706; 8,776,625; 8,898,567; 8,989,762; 9,160,579; 9,203,654; 9,235,268; 9,401,823; 9,432,564; 9,460,617; 9,531,427; 9,544,006; 9,564,927; 9,565,045; 9,613,408; 9,621,387; 9,660,851; 9,680,670; 9,686,112; 9,712,238; 9,712,350; 9,712,354; 9,713,019; 9,722,318; 9,729,281; 9,729,378; 9,742,521; 9,749,083; 9,774,476; 9,859,981; 9,871,679; 9,876,530; 9,877,206; 9,882,608; 9,893,919; 9,899,182; 9,900,048; 9,900,122; 9,900,123; 9,900,190; 9,912,436; 9,929,755; 9,942,074; 9,998,172; 9,998,187; 10,003,364; 10,027,397; 10,027,427; 10,027,523; 10,033,107; 10,033,108; 10,050,815; 10,051,483; 10,051,488; 10,062,970; 10,063,354; 10,069,467; 10,069,535; 10,079,652; 10,084,562; 10,090,594; 10,096,883; 10,103,777; 10,123,217; 10,129,057; 10,135,145; 10,148,016; 10,148,360; 10,168,501; 10,170,840; 10,171,158; 10,191,376; 10,198,582; 10,200,106; 10,205,212; 10,205,231; 10,205,482; 10,205,655; 10,211,855; 10,212,014; 10,218,405; 10,224,634; 20020051546; 20020085725; 20020103619; 20020172374; 20020172376; 20020172378; 20030035549; 20030055635; 20030098805; 20030112088; 20090221257; 20110238690; 20110249024; 20110252320; 20110288457; 20120112908; 20130110974; 20130201316; 20140226035; 20150146805; 20150146806; 20150230105; 20150280945; 20150310739; 20160093029; 20160149665; 20160149731; 20160197642; 20160218891; 20160226681; 20160352361; 20160352362; 20160352419; 20170012862; 20170018851; 20170018852; 20170019131; 20170026095; 20170032129; 20170033465; 20170033466; 20170033953; 20170033954; 20170063430; 20170078400; 20170085003; 20170085336; 20170093693; 20170104617; 20170110795; 20170110804; 20170111805; 20170134205; 20170201288; 20170229782; 20170230083; 20170245157; 20170269481; 20170271117; 20170288917; 20170295048; 20170311307; 20170317781; 20170317782; 20170317783; 20170317858; 20170318482; 20170331899; 20180013452; 20180034912; 20180048497; 20180054232; 20180054233; 20180054234; 20180054268; 20180062886; 20180069594; 20180069731; 20180076947; 20180076979; 20180076982; 20180076988; 20180091195; 20180115040; 20180115058; 20180123256; 20180123257; 20180123749; 20180123836; 20180123856; 20180123897; 20180124181; 20180131406; 20180131541; 20180145411; 20180145412; 20180145414; 20180145415; 20180151957; 20180152262; 20180152330; 20180152925; 20180159195; 20180159196; 20180159197; 20180159228; 20180159229; 20180159230; 20180159232; 20180159240; 20180159243; 20180159615; 20180166761; 20180166784; 20180166785; 20180166787; 20180167105; 20180167148; 20180175892; 20180175978; 20180198668; 20180205399; 20180205481; 20180227158; 20180248592; 20180254754; 20180254924; 20180262243; 20180278693; 20180278694; 20180294897; 20180301812; 20180302145; 20180309206; 20180323826; 20180324005; 20180324006; 20180324021; 20180324601; 20180331413; 20180331720; 20180331721; 20180331871; 20180343304; 20180351687; 20180358678; 20180359126; 20180375940; 20190013577; 20190013837; 20190013838; 20190020530; 20190036222; 20190052505; 20190074563; 20190074564; 20190074565; 20190074568; 20190074580; 20190074584; 20190074597; 20190074598; 20190074864; 20190074865; and 20190074878.
