APPARATUS AND METHOD FOR CONTROLLING NON-LINEAR EFFECT OF POWER AMPLIFIER
20230138959 · 2023-05-04
Inventors
- Pedamalli SAIKRISHNA (Bangalore, IN)
- Ankur GOYAL (Bangalore, IN)
- Ashok Kumar Reddy Chavva (Bangalore, IN)
- Ashwini Kumar (Bangalore, IN)
- Suhwook KIM (Suwon-si, KR)
- Sangho Lee (Suwon-si, KR)
Cpc classification
G06F18/214
PHYSICS
H03F2201/3227
ELECTRICITY
International classification
Abstract
Embodiments herein disclose a method for controlling a non-linear effect of a power amplifier by an apparatus. The method includes acquiring an input data of the power amplifier of the apparatus and an output data of the power amplifier. Further, the method includes determining an inverse function using a neural network. The inverse function maps normalized output data of the PA to the input data of the PA, where the neural network comprises at least one sub-network for at least one memory tap from a plurality of memory taps in the neural network. Further, the method includes modifying the input data based on the determined inverse function value by dynamically changing a usage of the at least one memory tap from the plurality of memory taps. Further, the method includes compensating the non-linear effect in the output data of the power amplifier.
Claims
1. A method for controlling a non-linear effect of a power amplifier, comprising: training, by an apparatus, a neural network (NN) based Digital Pre-distorter (DPD) of the apparatus, wherein the NN based DPD comprises at least one sub-network from a plurality of sub-networks; and placing, by the apparatus, the trained NN based DPD before the power amplifier of the apparatus to control the non-linear effect of the output data of the power amplifier.
2. The method as claimed in claim 1, wherein the method further comprises transmitting, by the apparatus, a linearly amplified signal comprising the non-linear effect of the output data of the power amplifier.
3. The method as claimed in claim 2, wherein the method further comprises: continuously monitoring, by the apparatus, at least one of an error vector magnitude (EVM) parameter associated with the linearly amplified signal and an adjacent channel leakage ratio (ACLR) parameter associated with the linearly amplified signal; determining, by the apparatus, whether at least one of the EVM parameter and the ACLR parameter meets a specified threshold; and performing, by the apparatus, at least one of: retaining at least one NN parameter in the trained NN based DPD in response determining that at least one of the EVM parameter and the ACLR parameter meets the specified threshold, and retraining the trained NN based DPD to modify at least one NN parameter in the trained NN based DPD in response determining at least one of the EVM parameter and the ACLR parameter does not meet the specified threshold.
4. The method as claimed in claim 1, wherein a number of an output node of the sub-network corresponds to a current input data, to at least one previous input data and at least one previous output data.
5. The method as claimed in claim 1, wherein the NN based DPD comprises: training a fully connected part of the network, wherein the fully connected part captures a non-linear parameter, wherein the non-linear parameter comprises a memory information and temperature information; and training a partially connected part of the plurality of sub-networks based on at least one previous input data and at least one previous output data, wherein the fully connected part and the partially connected part are trained separately, wherein the fully connected part is trained in real time, wherein the fully connected part of the network and the partially connected part of the plurality of sub-networks are trained together in an initial stage, and based on there being further non linearities in the apparatus, the apparatus is configured to train only the fully connected part while retaining the previous parameters for the partially connected part in the real time.
6. The method as claimed in claim 1, wherein the NN based DPD is trained by all sub-networks corresponding to at least one previous input data and at least one previous output data with at least one sub-network parameter, wherein the at least one sub-network parameter comprises an output node, weight and biases.
7. A method for controlling a non-linear effect of a power amplifier, comprising: acquiring, by an apparatus, an input data of the power amplifier (PA) of the apparatus and an output data of the power amplifier; determining, by the apparatus, an inverse function using a neural network, wherein the inverse function maps normalized output data of the PA to the input data of the PA, wherein the neural network comprises at least one sub-network for at least one memory tap from a plurality of memory taps in the neural network; modifying, by the apparatus, the input data based on the determined inverse function value by dynamically changing a usage of the at least one memory tap from the plurality of memory taps; and compensating, by the apparatus, the non-linear effect in the output data of the power amplifier.
8. The method as claimed in claim 7, wherein the inverse function value is learned by reducing an error between a current input of the power amplifier and an estimated input of the power amplifier from a normalized power amplifier output over a period of time.
9. An apparatus configured to control a non-linear effect of a power amplifier, comprising: a neural network (NN) based Digital Pre-distorter (DPD), wherein the NN based DPD comprises at least one sub-network from a plurality of sub-networks; and the trained NN based DPD is disposed before the power amplifier and configured to control the non-linear effect of the output data of the power amplifier.
10. The apparatus as claimed in claim 9, wherein the power amplifier is configured to transmit a linearly amplified signal comprising the non-linear effect of the output data of the power amplifier.
11. The apparatus as claimed in claim 9, wherein the trained NN based DPD is configured to: continuously monitor at least one of an error vector magnitude (EVM) parameter associated with the linearly amplified signal and an adjacent channel leakage ratio (ACLR) parameter associated with the linearly amplified signal; determine whether at least one of the EVM parameter and the ACLR parameter meets a specified threshold; and perform at least one of: retaining at least one NN parameter in the trained NN based DPD in response determining that at least one of the EVM parameter and the ACLR parameter meets the specified threshold, and retraining the trained NN based DPD to modify at least one NN parameter in the trained NN based DPD in response determining at least one of the EVM parameter and the ACLR parameter does not meet the specified threshold.
12. The apparatus as claimed in claim 9, wherein a number of an output node of the sub-network corresponds to a current input data, to at least one previous input data and at least one previous output data, wherein the number of output nodes corresponding to the sub-networks for current and past samples are different.
13. The apparatus as claimed in claim 9, wherein the NN based DPD is configured to be trained by: training a fully connected part of the networks, wherein the fully connected part captures a non-linear parameter, wherein the non-linear parameter comprises a memory information and temperature information; and training a partially connected part of the plurality of sub-networks based on at least one previous input data and at least one previous output data, wherein the fully connected part and the partially connected part are trained separately, wherein the fully connected part is trained in real time, wherein the fully connected part of the network and the partially connected part of the plurality of sub-networks are trained together in an initial stage, and based on there being further non linearities in the apparatus, the apparatus is configured to train only the fully connected part while retaining the previous parameters for the partially connected part in the real time.
14. The apparatus as claimed in claim 9, wherein the NN based DPD is configured to be trained by all sub-networks corresponding to at least one previous input data and at least one previous output data with at least one sub-network parameter, wherein the at least one sub-network parameter comprises an output node, weight and biases.
15. An apparatus configured to control a non-linear effect of a power amplifier, comprising: a neural network (NN) based Digital Pre-distorter (DPD), wherein the NN based DPD comprises at least one sub-network from a plurality of sub-networks, wherein the NN based DPD is configured to: acquire an input data of the power amplifier (PA) of the apparatus and an output data of the power amplifier; determine an inverse function value, wherein the inverse function is configured to map normalized output data of the PA to the input data of the PA wherein the NN based DPD comprises at least one sub-network for at least one memory tap from a plurality of memory taps in the neural network; modifying the input data based on the determined inverse function value by dynamically changing a usage of the at least one memory tap from the plurality of memory taps; and compensating for the non-linear effect in the output data of the power amplifier.
16. The apparatus as claimed in claim 15, wherein the inverse function value is learned by reducing an error between a current input of the power amplifier and an estimated input of the power amplifier from a normalized power amplifier output over a period of time.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
[0021] The above and other aspects, features and advantages of certain embodiments of the present disclosure will be more apparent from the following detailed description, taken in conjunction with the accompanying drawings, in which:
[0022]
[0023]
[0024]
[0025]
[0026]
[0027]
[0028]
[0029]
[0030]
[0031]
DETAILED DESCRIPTION
[0032] The various example embodiments herein and the various features and advantageous details thereof are explained more fully with reference to the non-limiting embodiments that are illustrated in the accompanying drawings and described in the following description. Descriptions of well-known components and processing techniques may be omitted so as to not unnecessarily obscure the description of the various example embodiments herein. The various embodiments described herein are not necessarily mutually exclusive, as various embodiments can be combined with one or more other embodiments to form new embodiments. The term “or” as used herein, refers to a non-exclusive or, unless otherwise indicated. The examples used herein are intended merely to facilitate an understanding of ways in which the embodiments herein can be practiced and to further enable those skilled in the art to practice the embodiments herein. Accordingly, the examples should not be construed as limiting the scope of the embodiments herein.
[0033] Embodiments may be described and illustrated in terms of blocks which carry out a described function or functions. These blocks, which may be referred to herein as managers, units, modules, hardware components or the like, are physically implemented by analog and/or digital circuits such as logic gates, integrated circuits, microprocessors, microcontrollers, memory circuits, passive electronic components, active electronic components, optical components, hardwired circuits and the like, and may optionally be driven by firmware and software. The circuits may, for example, be embodied in one or more semiconductor chips, or on substrate supports such as printed circuit boards and the like. The circuits of a block may be implemented by dedicated hardware, or by a processor (e.g., one or more programmed microprocessors and associated circuitry), or by a combination of dedicated hardware to perform some functions of the block and a processor to perform other functions of the block. Each block of the embodiments may be physically separated into two or more interacting and discrete blocks without departing from the scope of the disclosure. Likewise, the blocks of the embodiments may be physically combined into more complex blocks without departing from the scope of the disclosure.
[0034] Accordingly, various example embodiments provide a method for controlling a non-linear effect of a power amplifier. The method includes acquiring, by an apparatus, an input data of the power amplifier of the apparatus and an output data of the power amplifier. Further, the method includes determining, by the apparatus, an inverse function using a neural network, wherein the inverse function maps normalized output data of the PA to the input data of the PA, wherein the neural network comprises at least one sub-network for at least one memory tap from a plurality of memory taps in the neural network. Further, the method includes modifying, by the apparatus, the input data based on the determined inverse function value by dynamically changing a usage of the at least one memory tap from the plurality of memory taps. Further, the method includes compensating, by the apparatus, the non-linear effect in the output data of the power amplifier.
[0035] There are various NN based DPD architectures in the existing method but none of them have use separate sub-networks for separate memory taps and training only a part of the network online while using parameters trained offline for the majority of the network. The disclosed method has separate sub networks for each of the memory tap. This helps in having output nodes of the sub network flexible to chronology of the samples. In the disclosed method, the separate subnetworks for memory taps provides with a lot of flexibility with respect to number of memory taps to be used. The disclosed NN has a fixed and trainable part during online training. This adapts to the changes in real time with minimal training overhead.
[0036] In the disclosed method, the inference cost along with the training cost will reduce drastically based on the number of past samples used. The disclosed method takes very short time to adapt to the changes compared to ML techniques and has better performance than the conventional approaches. Having separate sub-networks for each memory tap, the disclosed architecture can provide importance to the samples based on chronology and with fixed sub-network for some past samples, the method can further reduce the training and inference complexity.
[0037] Referring now to the drawings and more particularly to
[0038]
[0039] As shown in
[0040] In an embodiment, the trained NN based DPD (504) continuously monitors at least one of an error vector magnitude (EVM) parameter associated with the linearly amplified signal and an adjacent channel leakage ratio (ACLR) parameter associated with the linearly amplified signal. Further, the trained NN based DPD (504) determines whether at least one of the EVM parameter and the ACLR parameter meets a predefined threshold. In an embodiment, in response determining that at least one of the EVM parameter and the ACLR parameter meets the predefined threshold, the trained NN based DPD (504) retain at least one NN parameter in the trained NN based DPD. In an embodiment, the trained NN based DPD (504) retrain the trained NN based DPD to modify at least one NN parameter in the trained NN based DPD in response determining at least one of the EVM parameter and the ACLR parameter does not meet the predefined threshold.
[0041] In an example, for input x(n), let the output of PA (502), y(n) is given by—
y(n)=ƒ.sub.PA[x(n)]
[0042] Then there is a need to find the function ƒ.sub.dpd such that—
y.sub.c(n)=ƒ.sub.PA[ƒ.sub.DPD[x(n)]]=G*x(n)
[0043] Different ways are proposed in the existing methods to learn the inverse function of the PA non linearity, which then acts as the DPD (504). Here the disclosed method makes use of the indirect learning architecture (ILA) principal to achieve this.
[0044] Here, the apparatus (500) tries to learn the PA inverse function, which tries to estimate the PA input given its output normalized with gain. This inverse function is learned by minimizing/reducing the error between the actual input of the PA (502) and the estimated PA input using the learned function from the normalized PA output. Simply put, this inverse function learns to provide x(n) given y(n)/G.
[0045]
[0046]
[0047] Learning the inverse function (DPD)—training phase: The x(n), y(n) input, output pairs are obtained from the PA (502). The real PAs also have some memory effects because of the usage of active elements such as capacitors. This leads to the current samples output being affected by past samples.
[0048] During the training phase, the network is passed normalized output
or current and past samples as inputs and the x(n) as output. This network then learn to minimize and/or reduce the loss between the actual input samples x(n) and the estimated input samples x{circumflex over ( )}(n).
[0049] Implementation—testing phase: during the testing phase, the x(n) of current and previous samples are passed through the network with the previously trained weights and biases (from training phase) to obtain the digital predistorter output z(n), which is then passed through PA to obtain a linearly amplified version of x(n) as shows in
[0050] In the disclosed architecture, in partially connected part of the network, the apparatus (500) have separate densely connected networks (the apparatus (500) refers to them as sub-networks) for each of the current and previous samples for compensating the nonlinear behaviour of the PA with respect to that sample alone. The fully connected part of the network handles the nonlinear memory effect, where the outputs of each of the sub-network are concatenated and are fed to another set of densely connected layers.
[0051] Having separate sub-networks for each sample (current and past), the apparatus (100) can decide the importance that needs to be given for each sample by tweaking the output nodes of those samples alone appropriately. The apparatus can also have a provision, where the apparatus need not train the sub-network part of the architecture for a different memory length requirements on the fly (online). The sub networks can be added/removed for the memory elements to be added/removed using the sub network parameters trained offline. This will help in reducing the training time on the device eventually.
[0052] The neural network based method to model the digital predistorter, which is trained to minimize/reduce the error between estimated and actual PA inputs using current and previous PA outputs, employing separate sub-network for each of the memory sample with their number of output nodes designed based on the chronology, and later the fully connected part of the network handles for the nonlinear caused by the memory and other external effects. Using this we can have the sub-networks trained offline to be fixed and train only the later part of the network online.
[0053] Also, for the same PA make, even though the general characteristics remain similar there could be subtle variations. Here, we can have a large dataset collected from the PAs of a same make and train the network offline and online, we can train only the fully connected part of the network to address the subtle variation for each PA along with other non-linearity causing factors online.
[0054]
[0055]
[0056]
[0057]
[0058]
[0059] As shown in
[0060] As shown in
[0061] As shown in
[0062] As shown in
[0063] As shown in
[0064] As shown in
[0065] As shown in
[0066] As shown in
[0067]
[0068] The various actions, acts, blocks, steps, or the like in the flow charts (S1100-S1300) may be performed in the order presented, in a different order or simultaneously. Further, in various embodiments, some of the actions, acts, blocks, steps, or the like may be omitted, added, modified, skipped, or the like without departing from the scope of the disclosure.
[0069] While the disclosure has been illustrated and described with reference to various example embodiments, it will be understood that the various example embodiments are intended to be illustrative, not limiting. It will be further understood by those skilled in the art that various changes in form and detail may be made without departing from the true spirit and full scope of the disclosure including the appended claims and their equivalents. It will also be understood that any of the embodiment(s) described herein may be used in conjunction with any other embodiment(s) described herein.