Apparatus and method for deprecoding a signal for a machine learning model
12484035 ยท 2025-11-25
Assignee
Inventors
Cpc classification
H04W72/0453
ELECTRICITY
H04L27/26526
ELECTRICITY
H04L27/26536
ELECTRICITY
International classification
H04W72/0453
ELECTRICITY
H04L25/02
ELECTRICITY
Abstract
An apparatus for a receiver is provided, said apparatus comprising means for receiving an input signal, based on a frequency domain signal, means for performing deprecoding on the received input signal and means for providing the deprecoded signal as an input for a machine learning model, wherein the output of the machine learning model is a likelihood ratio.
Claims
1. An apparatus for a receiver, the apparatus comprising: a processor; and a memory storing instructions, which when executed by the processor, cause the apparatus to: receive an input signal, based on a frequency domain signal; provide the input signal to a first machine learning model for processing the input signal; perform deprecoding on the processed input signal output by the first machine learning model; provide the deprecoded signal as an input for a second machine learning model, wherein the output of the second machine learning model is a likelihood ratio.
2. The apparatus according to claim 1, wherein performing deprecoding includes using an inverse Discrete Fourier Transform, iDFT, deprecoding function.
3. The apparatus according to claim 1, wherein the input signal includes a data branch.
4. The apparatus according to claim 3, wherein the data branch comprises frequency-time grid symbols N_fN_t, where N_f comprises the number of subcarriers and N_t comprises a number of orthogonal frequency-division multiplexing, OFDM, symbols.
5. The apparatus according to claim 3, wherein, when executed by the processor, instructions cause the apparatus to concatenate the data branch of the input signal with a raw channel estimate branch of the frequency domain signal and provide the concatenated signal as the input for the first machine learning model.
6. The apparatus according to claim 1, wherein the output of the first machine learning model is 2N_fN_t, where N_f comprises iDFT size and N_t comprises a number of orthogonal frequency-division multiplexing, OFDM, symbols.
7. The apparatus according to claim 6, wherein the output of the further machine learning model further comprises a constellation monitoring output.
8. The apparatus according to claim 1, wherein, when executed by the processor, instructions cause the apparatus to cause a trigger to be sent to a second apparatus for training the second machine learning model and the first machine learning model; and receive an acknowledgment from the second apparatus in response to the request.
9. The apparatus according to claim 8, wherein, when executed by the processor, instructions cause the apparatus to provide training configuration information to the second apparatus; receive training data from the second apparatus in response to the configuration information; and provide the training data to an entity for training the second machine learning model and the first machine learning model.
10. The apparatus according to claim 1, wherein, when executed by the processor, instructions cause the apparatus to receive deployment configuration information from a transmitter of the input signal; and determine whether to deploy the second machine learning model and the first machine learning model based on the deployment configuration information.
11. The apparatus according to claim 10, wherein the deployment configuration information comprises at least one of an indication to perform the deprecoding on the frequency domain signal, subcarrier number information, modulation and coding scheme, rank and a fallback indicator.
12. The apparatus according to claim 11, wherein, when executed by the processor, instructions cause the apparatus to determine not to deploy the second machine learning model and the first machine learning model based on the fallback indicator.
13. A method comprising, at a receiver: receiving an input signal, based on a frequency domain signal; providing the input signal to a first machine learning model for processing the input signal; performing deprecoding on an output signal of the first machine learning model; and providing the deprecoded signal as an input for a second machine learning model, wherein the output of the second machine learning model is a likelihood ratio.
14. The method according to claim 13, wherein performing deprecoding includes using an inverse Discrete Fourier Transform, iDFT, deprecoding function.
15. The method according to claim 13, wherein the input signal includes a data branch.
16. The method according to claim 15, wherein the data branch comprises frequency-time grid symbols N_fN_t, where N_f comprises the number of subcarriers and N_t comprises a number of orthogonal frequency-division multiplexing, OFDM, symbols.
17. An apparatus comprising: at least one processor and at least one memory including a computer program code, the at least one memory and computer program code configured to, with the at least one processor, cause the apparatus at least to, at a receiver: receive an input signal, based on a frequency domain signal; provide the input signal to a first machine learning model for processing the input signal; perform deprecoding on an output of the first machine learning model; and provide the deprecoded signal as an input for a second machine learning model, wherein the output of the second machine learning model is a likelihood ratio.
18. The apparatus according to claim 17, wherein performing deprecoding includes using an inverse Discrete Fourier Transform, iDFT, deprecoding function.
19. The apparatus according to claim 17, wherein the input signal includes a data branch.
20. The apparatus according to claim 19, wherein the data branch comprises frequency-time grid symbols N_fN_t, where N_f comprises the number of subcarriers and N_t comprises a number of orthogonal frequency-division multiplexing, OFDM, symbols.
Description
DESCRIPTION OF FIGURES
(1) Embodiments will now be described, by way of example only, with reference to the accompanying Figures in which:
(2)
(3)
(4)
(5)
(6)
(7)
(8)
(9)
(10)
(11)
(12)
(13)
(14)
(15)
(16)
(17)
(18)
(19)
(20)
DETAILED DESCRIPTION
(21) Before explaining in detail the examples, certain general principles of a wireless communication system and mobile communication devices are briefly explained with reference to
(22) An example of a suitable communications system is the 5G System (5GS). Network architecture in 5GS may be similar to that of LTE-advanced. Base stations of NR systems may be known as next generation Node Bs (gNBs). Changes to the network architecture may depend on the need to support various radio technologies and finer QoS support, and some on-demand requirements for example QoS levels to support QoE of user point of view. Also network aware services and applications, and service and application aware networks may bring changes to the architecture. Those are related to Information Centric Network (ICN) and User-Centric Content Delivery Network (UC-CDN) approaches. NR may use multiple input-multiple output (MIMO) antennas, many more base stations or nodes than the LTE (a so-called small cell concept), including macro sites operating in co-operation with smaller stations and perhaps also employing a variety of radio technologies for better coverage and enhanced data rates.
(23) 5G networks may utilise network functions virtualization (NFV) which is a network architecture concept that proposes virtualizing network node functions into building blocks or entities that may be operationally connected or linked together to provide services. A virtualized network function (VNF) may comprise one or more virtual machines running computer program codes using standard or general type servers instead of customized hardware. Cloud computing or data storage may also be utilized. In radio communications this may mean node operations to be carried out, at least partly, in a server, host or node operationally coupled to a remote radio head. It is also possible that node operations will be distributed among a plurality of servers, nodes or hosts. It should also be understood that the distribution of labour between core network operations and base station operations may differ from that of the LTE or even be non-existent.
(24)
(25) An example 5G core network (CN) comprises functional entities. The 5GCN 106 may comprise one or more access and mobility management functions (AMF) 112, one or more session management functions (SMF) 114, an authentication server function (AUSF) 116, a unified data management (UDM) 118, one or more user plane functions (UPF) 120, a unified data repository (UDR) 122 and/or a network exposure function (NEF) 124. The UPF is controlled by the SMF (Session Management Function) that receives policies from a PCF (Policy Control Function).
(26) The CN is connected to a terminal device via the radio access network (RAN). The 5GRAN may comprise one or more gNodeB (GNB) distributed unit functions connected to one or more gNodeB (GNB) centralized unit functions. The RAN may comprise one or more access nodes.
(27) A UPF (User Plane Function) whose role is called PSA (Protocol Data Unit (PDU) Session Anchor) may be responsible for forwarding frames back and forth between the DN (data network) and the tunnels established over the 5G towards the UE(s) exchanging traffic with the DN.
(28) A possible mobile communication device will now be described in more detail with reference to
(29) A mobile device is typically provided with at least one data processing entity 201, at least one memory 202 and other possible components 203 for use in software and hardware aided execution of tasks it is designed to perform, including control of access to and communications with access systems and other communication devices. The data processing, storage and other relevant control apparatus can be provided on an appropriate circuit board and/or in chipsets. This feature is denoted by reference 204. The user may control the operation of the mobile device by means of a suitable user interface such as key pad 205, voice commands, touch sensitive screen or pad, combinations thereof or the like. A display 208, a speaker and a microphone can be also provided. Furthermore, a mobile communication device may comprise appropriate connectors (either wired or wireless) to other devices and/or for connecting external accessories, for example hands-free equipment, thereto.
(30) The mobile device 200 may receive signals over an air or radio interface 207 via appropriate apparatus for receiving and may transmit signals via appropriate apparatus for transmitting radio signals. In
(31)
(32) Technology known as DeepRx (deep learning receiver) aims to leverage customized AI/ML techniques to augment capacity on the physical layer (PHY). Non-negligible performance gain over conventional methods in terms of reduced transmission error under the same radio frequency (RF) channel condition has been demonstrated in studies of DeepRx based on both simulation and real-world data. Meanwhile, superiority in terms of robustness against diverse channel condition and Doppler effects has also been shown.
(33)
(34)
(35) Discrete Fourier transform (DFT)-Spread OFDM (DFTS-OFDM) has been adopted to enlarge power amplifier (PA) efficiency on the device side in the uplink transmission. It is designed with coverage-challenged situations and supports flexible discrete Fourier transform (DFT) sizes in implementation. Support for DFTS-OFDM is mandatory in the device and the network can therefore configure DFT-precoding for a particular device whenever necessary.
(36)
(37) As described in 3GPP specifications, a UE determines whether or not to apply transform (DFT) precoding and the UE shall consider the transform (DFT) precoding either enabled or disabled according to the higher layer configured parameter.
(38) Therefore, compatibility with a DFTS-OFDM system is a requirement for future implementations of DeepRx, especially when it is deployed in base station (BS) side.
(39) Success of DeepRx on Single Input Single Output (SISO) or Multiple Input Multiple Output (MIMO) OFDM systems has been shown. Nevertheless, nonnegligible performance drop or difficulties in DeepRx training has been observed when it comes to single-carrier OFDM scenarios, e.g., DFTS-OFDM
(40) In other words, current DeepRx is not compatible with the format of DFTS-OFDM, which is a mandatory feature for beyond 5G devices to support as mentioned above. Incompatibility of DeepRx to the DFTS-OFDM system may be due to the fact that the CNN structured DeepRx may not be able to learn to mimic the iDFT deprecoding operation which spans across the whole OFDM frame (shown in
(41)
(42) In S1, the method comprises receiving an input signal based on a frequency domain signal.
(43) In S2, the method comprises performing deprecoding on the received input signal
(44) In S3, the method comprises providing the deprecoded signal as an input for a machine learning model, wherein the output of the machine learning model is a likelihood ratio.
(45) The frequency domain signal may be a single carrier frequency domain signal or DFTS-OFDM signal.
(46) Signal deprecoding is the inverse operation of signal precoding. If the signal precoding were realized by means of DFT precoding then iDFT deprecoding would be the inverse operation. The deprecoding may be performed by an iDFT deprecoding function.
(47) The method may provide two schemes to augment DeepRx's compatibility with DFTS-OFDM signal detection.
(48) In a first example embodiment, the input signal is the frequency domain signal. The method may comprise performing the deprecoding on a data branch of the input signal. The data branch may comprise frequency-time grid symbols N.sub.fN.sub.t, where N.sub.f comprises the number of subcarriers and N.sub.t comprises number of orthogonal frequency-division multiplexing, OFDM, symbols. The method may comprise concatenating the deprecoded signal with a raw channel estimate branch of the frequency domain signal and providing the concatenated signal as the input for the machine learning model. In this example embodiment, the machine learning model may be referred to as DeepRx.
(49) In a first option, where the deprecoding is performed on the frequency domain signal, a pre-executed de-precoding function block is introduced on a received data branch before being concatenated with a raw channel estimation branch, and then fed to legacy DeepRx.
(50)
(51)
(52) Taking 1T1R transmission, the input data includes Branch 1 and Branch 2 which comprise the received symbols after FFT which transforms them to the frequency-time grid symbols with the size of N.sub.fN.sub.t, where N.sub.f refers to the number of subcarriers and N.sub.t refers to the number of OFDM symbols in one slot (which is 14 according to 3GPP definition). Branch 1 and 2 are the real part and imaginary part of the received symbols respectively. The DMRS occupied resource elements (RE) are set to be 0.
(53) The input data also includes Branch 3 and 4 which comprise the raw channel estimates branches where non-demodulation reference signals (DMRS) resource elements (RE) are set to be 0. One branch size is also N.sub.fN.sub.t. Branch 3 and 4 are also the real part and imaginary part respectively.
(54) All the branches are concatenated before being fed to the neural network (NN) 803.
(55) The output data comprises bit LLRs with the dimension of N.sub.modN.sub.fN.sub.t, where N.sub.mod refers the number of bits per RE and depends on the modulation scheme.
(56) In the training phase, the output will be used to compute a loss function with transmitted bits for model training. In the inference phase (application), the output will be fed to decoder.
(57) The deprecoding is only applied to received symbols before being separated to real part and imaginary part. The deprecoding is a fixed operation following the conventional operation on frequency domain.
(58) In the example structure shown in
(59) In a second example embodiment, the method may comprise providing the input signal, prior to deprecoding, to a further machine learning model for processing the input signal and performing deprecoding on an output of the further machine learning mode. In this embodiment, the input signal is a processed frequency domain signal (i.e., the frequency domain signal which has been processed by the further machine learning model).
(60) In this example embodiment, the machine learning model may be referred to as PostDeepRx and the further machine learning model may be referred to as DeepRx.
(61) In the second example embodiment, a cascaded design may be provided with a DeepRx-Deprecoding-PostDeepRx structure where a function block of deprecoding is embedded between the DeepRx and PostDeepRx whose parameters are jointly learnt and optimized.
(62) The output of the further machine learning model in this example embodiment is 2N.sub.fN.sub.t, where N.sub.f comprises iDFT size and N.sub.t comprises the number of orthogonal frequency-division multiplexing, OFDM, symbols.
(63) The output of the further machine learning model may further comprise a constellation monitoring output.
(64) The constellation monitoring output may be provided for further analysis. The output of iDFT deprecoding has semantic meaning if the input and output of the iDFT deprecoding block is well designed. With further expert or machine type analysis, one can learn the hardware and channel conditions, or even DeepRx itself. In other words, this is one step towards making the DeepRx receiver explainable and consequently more trustworthy.
(65)
(66)
(67) The input data in this example embodiment comprises a data branch and a raw estimate branch as described with reference to the first example embodiment. However, in this example, no pre-deprecoding is applied to the received symbols.
(68) The output data of the machine learning model 1003a comprises bit LLRs with the dimension of N.sub.modN.sub.fN.sub.t and the output of the further machine learning mode (1003b) comprises post equalised constellation-like pattern with the dimension of 2N.sub.fN.sub.t.
(69) In the training phase, the Bit LLR output will be used to compute loss function with transmitted bits for model training. In the inference phase (application), the Bit LLR output will be fed to decoder, and the constellation-like pattern may be recorded for further analysis.
(70) In the second example embodiment, the deprecoding function block 1002 is embedded between the NNs, which are described below.
(71) The embedded deprecoding is a fixed operation and not learnable in back propagation. PreDeepRx and Post DeepRx may be jointly trained together.
(72) In the second example embodiment, the DeepRx comprises two separate NNs as shown in
(73) The first NN, referred to as DeepRx NN 1003b, includes 2 convolutional layers and 11 ResNet Blocks. The kernel number of output convolutional layer is set to be 2 with output data dimension of 2N.sub.fN.sub.t.
(74) The second NN, referred to as PostDeepRx 1003a, includes 2 convolutional layers and 1 ResNet in between. The kernel number of the input convolutional layer is 2 with input data dimension of 2N.sub.fN.sub.t. The kernel number of output convolutional layer depends on the modulation scheme.
(75) Any other suitable NN structures may be used in place of the examples described with reference to the Figures.
(76) The design of the output of DeepRx 1003b and input of PostDeepRx 1003a (of 2N.sub.fN.sub.t) may have two advantages.
(77) One is to force the output of DeepRx 1003b to accord with the iDFT deprecoding, whose output can be sequentially coupled to PostDeepRx 1003a. DeepRx 1003b thus opens up a window for constellation surveillance at the output of iDFT deprecoding block 1002. The intermediate constellation derived at the 2nd output branch may be perceived by human or machine type analysis for channel condition (e.g., nonlinearities, Dopplers) surveillance.
(78) The other is to conduct PostDeepRx to learn a nonlinear demodulation detection, leaving DeepRx to serve as joint nonlinear channel estimator and equalizer. Thus, to some degree, it allows elaborate design of DeepRx and PostDeepRx according to the intermediate outcome of learned constellation patterns.
(79) In both the first and second example embodiments as described above, the binary cross entropy may be used as a loss function. The formula is:
(80)
(81) Where y.sub.i represent binary bits 0/1 and P(y.sub.i) is the corresponding probability which is computed by feeding output LLRs to a sigmoid function. N is the total number of bits after channel encoding.
(82) Non-Data RE positions are ignored for back propagation in training.
(83) The deprecoding function block is fixed and not learned when training both ML models. The machine learning model may be trained either by Online Mode or Offline Mode. The offline training could be performed in other devices where model is not deployed.
(84) The online training requires signaling and actions between both transmitter and receiver.
(85) The devices for training the model may include functioning units in receivers (performing data set construction, data processing, model training, updating, deployment), transmitters (preparing training data and send it over the air) and other platforms, which can perform all the related processes.
(86) The method may comprise causing a trigger to be sent to an apparatus (e.g., a transmitter such as a user equipment) for training the machine learning mode (and, if present, the further machine learning mode) and receiving an acknowledgment from the apparatus in response to the request.
(87) The method may comprise providing training configuration information (e.g., but not limited to, carrier frequency, number of subcarriers, MCS, rank, transform precoding enabling indication (DFT precoding), sequence random seed, and number of transmission time intervals (TTI)) to the apparatus, receiving training data from the apparatus in response to the configuration information and providing the training data to an entity for training the machine learning model (and, if present, the further machine learning model). The entity for training the machine learning model and further machine learning model may be located at a receiver, another base station such a gNB or the cloud. The entity for training the machine learning model and the further machine learning model may be a Local Data Pool and Computing Unit (LDP&CU).
(88)
(89) In a first step 1201, the receiver (e.g., BS for uplink transmission) triggers the online training or fine-tuning.
(90) In a second step 1202, the transmitter responds with acknowledges of training request.
(91) In a third step 1203, the receiver sends the training configurations, which may include carrier frequency, number of subcarriers, MCS, rank, transform precoding enabling indication (DFT precoding), sequence random seed, number of TTI required for training, etc.
(92) In a fourth step 1204, the transmitter sends the training sequence over the air. The training sequence should be known to receiver for computing loss.
(93) In a fifth step 1205, the receiver sends the data to a Local Data Pool and Computing Unit (LDP&CU).
(94) In a sixth step 1206, the model is trained in Computing Unit with training data in LDP.
(95) In a seventh step 1207, the CU sends the trained model to the Receiver user plane.
(96) In an eighth step 1208, the receiver deploys the model.
(97) The model updating and finetuning may be viewed as part of online training process, which may be triggered periodically and non-periodically, by events. The steps for model updating and finetuning are as described above with respect to
(98) The receiver side may determine to deploy the machine learning model. For example, the method may comprise receiving deployment configuration information from a transmitter of the input signal and determining whether to deploy the machine learning model (and, if present, the further machine learning model) based on the deployment configuration information.
(99) In an example embodiment, to apply the machine learning model for wireless transmission, a transmitter notifies the receiver with corresponding configurations. The receiver should determine to activate corresponding functioning blocks (such as deprecoding block) and select the corresponding neural network (NN) based on the configuration, before using DeepRx for signal processing.
(100) The deployment configuration information may comprise at least one of an indication to perform the deprecoding on the frequency domain signal, subcarrier number information, modulation and coding scheme, rank and a fallback indicator.
(101)
(102) In step 1 1301, the transmitter sends configurations to the receiver, including transform-precoding enabling information, subcarrier number, rank, MCS, etc.
(103) In step 2 1302, the receiver schedules if the DeepRx should be used. If not, fall back to conventional methods, If yes, go to Step 3
(104) In step 3 1303, based on the configuration, the receiver determines to enable the deprecoding function block and selects the network for current rank, MCS, etc.
(105) In step 4 1304, the transmitter sends user data to the receiver.
(106) In step 5 1305, the receiver processes the data with DeepRx
(107) To schedule the DeepRx, RRM scheduler may notify the user plane of the receiver with information for DeepRx including, DFT-precoding enabling information, PRB numbers, Rank number, MCS, Fallback indicator, etc.
(108) The scheduling of DeepRx model for the first option and the second option is shown in
(109) In a first step, the RRM scheduler 1401 notifies DeepRx if DFT-precoding is enabled. In a second step, if DFT precoding is enabled, the deprecoding block 1402 for DeepRx is activated and if not the deprecoding function block 1402 is deactivated.
(110) The scheduler may also schedule the Model Switch between OFDM and DFTS-OFDM, as well as Model Fallback to Conventional Receiver in corresponding scenarios. The method may comprise determining not to deploy the machine learning model (and, if present, the further machine learning model) based on the fallback indicator. For example, falling back to a conventional receiver may be triggered by the receiver side under conditions where the performance of DeepRx may deteriorate, e.g., continuous errors in transmission, etc.
(111) The model falling back to conventional receiver may also trigger the Model Training (updating or finetuning as described above with reference to
(112) The receiver determines whether to fall back to conventional methods.
(113) If the receiver determines to fall back to conventional methods, the RRM plane scheduler notifies the User plane to fall back to conventional methods and the receiver applies conventional methods. If not, the RRM plane scheduler notifies the User plane to apply DeepRx and the receiver applies DeepRx (i.e. deploys the machine learning model).
(114) Both of the proposed schemes have been tested with a real-life proof of concept, the results showing the benefits of invented receiver architecture under realistic circumstances. According to these proof-of-concept evaluations, the second option may enable a faster convergence in training.
(115) The proposed DeepRx schemes were evaluated with data from a testbed for wireless transmission. The hardware nonlinearity exists and the DFT-precoding is enabled in data collection phase in order to lower the PAPR in transmission. The Tx signal propagated over the air and was captured by receiver at a distance of about 4 feet.
(116) The Tx/Rx antenna numbers are both 1. 2 DMRS symbols are used to estimate the channel. The transmission modulation scheme is 64 QAM. Table 1 shows a list of Configuration in Evaluation.
(117) TABLE-US-00001 TABLE 1 Parameters Configuration Tx/Rx 1/1 Modulation 64QAM Code Rate 873/1024 Subcarrier spacing ~30 MHz FFT size 512 Subcarrier number 300 (25RB) DMRS symbol 3, 12 DFT precoding Enabled
(118) The proposed schemes were tested under the conditions with DFT-precoding enabled and compare them with conventional LMMSE receiver under configuration of 25 RB (300 subcarriers), code rate of 873/1024. There are 4000 subframes of data for training, and another 1000 subframes for inference.
(119)
(120) The x axis is defined by E.sub.RE/N where E.sub.RE is the denoised signal power estimation and N is the noise estimation. In training phase, only the data at E.sub.RE/N of 23.15 dB (last E.sub.RE/N value in the figure) was employed, while in testing phase data from more E.sub.RE/N were used for inference.
(121) It is clearly observed that, under the 10% uncoded BER level, the first option DeepRx shows 0.6 dB gain over Practical LMMSE receiver, while the second option DeepRx outperforms the conventional LMMSE receiver by roughly 1.2 dB. The uncoded BER is decreased to 1/10 at training E.sub.RE/N. In Coded BER curves, the first option shows possible 1 dB gain over LMMSE, while the second option DeepRx may achieve>2 dB gain under 100% coded BER level. The second option achieves the error-free performance at training E.sub.RE/N while that from LMMSE stay at 1%2% level.
(122) There is also a faster convergence from the second option in the training process as shown by
(123) A constellation-like pattern 1900 as illustrated in
(124) Methods and apparatus as described with reference to
(125) The second example embodiment may provide a second constellation like output, which can unveil the nonlinearity in hardware etc, as well as make DeepRx more explainable.
(126) An apparatus for a receiver may comprise means for receiving an input signal, based on a frequency domain signal, means for performing precoding on the received input signal and means for providing the deprecoded signal as an input for a machine learning model, wherein the output of the machine learning model is a likelihood ratio.
(127) It should be understood that the apparatuses may comprise or be coupled to other units or modules etc., such as radio parts or radio heads, used in or for transmission and/or reception. Although the apparatuses have been described as one entity, different modules and memory may be implemented in one or more physical or logical entities.
(128) It is noted that whilst some embodiments have been described in relation to 5G networks, similar principles can be applied in relation to other networks and communication systems. Therefore, although certain embodiments were described above by way of example with reference to certain example architectures for wireless networks, technologies and standards, embodiments may be applied to any other suitable forms of communication systems than those illustrated and described herein.
(129) It is also noted herein that while the above describes example embodiments, there are several variations and modifications which may be made to the disclosed solution without departing from the scope of the present invention.
(130) In general, the various embodiments may be implemented in hardware or special purpose circuitry, software, logic or any combination thereof. Some aspects of the disclosure may be implemented in hardware, while other aspects may be implemented in firmware or software which may be executed by a controller, microprocessor or other computing device, although the disclosure is not limited thereto. While various aspects of the disclosure may be illustrated and described as block diagrams, flow charts, or using some other pictorial representation, it is well understood that these blocks, apparatus, systems, techniques or methods described herein may be implemented in, as non-limiting examples, hardware, software, firmware, special purpose circuits or logic, general purpose hardware or controller or other computing devices, or some combination thereof.
(131) As used in this application, the term circuitry may refer to one or more or all of the following: (a) hardware-only circuit implementations (such as implementations in only analog and/or digital circuitry) and (b) combinations of hardware circuits and software, such as (as applicable): (i) a combination of analog and/or digital hardware circuit(s) with software/firmware and (ii) any portions of hardware processor(s) with software (including digital signal processor(s)), software, and memory(ies) that work together to cause an apparatus, such as a mobile phone or server, to perform various functions) and (c) hardware circuit(s) and or processor(s), such as a microprocessor(s) or a portion of a microprocessor(s), that requires software (e.g., firmware) for operation, but the software may not be present when it is not needed for operation.
(132) This definition of circuitry applies to all uses of this term in this application, including in any claims. As a further example, as used in this application, the term circuitry also covers an implementation of merely a hardware circuit or processor (or multiple processors) or portion of a hardware circuit or processor and its (or their) accompanying software and/or firmware. The term circuitry also covers, for example and if applicable to the particular claim element, a baseband integrated circuit or processor integrated circuit for a mobile device or a similar integrated circuit in server, a cellular network device, or other computing or network device.
(133) The embodiments of this disclosure may be implemented by computer software executable by a data processor of the mobile device, such as in the processor entity, or by hardware, or by a combination of software and hardware. Computer software or program, also called program product, including software routines, applets and/or macros, may be stored in any apparatus-readable data storage medium and they comprise program instructions to perform particular tasks. A computer program product may comprise one or more computer-executable components which, when the program is run, are configured to carry out embodiments. The one or more computer-executable components may be at least one software code or portions of it.
(134) Further in this regard it should be noted that any blocks of the logic flow as in the Figures may represent program steps, or interconnected logic circuits, blocks and functions, or a combination of program steps and logic circuits, blocks and functions. The software may be stored on such physical media as memory chips, or memory blocks implemented within the processor, magnetic media such as hard disk or floppy disks, and optical media such as for example DVD and the data variants thereof, CD. The physical media is a non-transitory media.
(135) The memory may be of any type suitable to the local technical environment and may be implemented using any suitable data storage technology, such as semiconductor based memory devices, magnetic memory devices and systems, optical memory devices and systems, fixed memory and removable memory. The data processors may be of any type suitable to the local technical environment, and may comprise one or more of general purpose computers, special purpose computers, microprocessors, digital signal processors (DSPs), application specific integrated circuits (ASIC), FPGA, gate level circuits and processors based on multi core processor architecture, as non-limiting examples.
(136) Embodiments of the disclosure may be practiced in various components such as integrated circuit modules. The design of integrated circuits is by and large a highly automated process. Complex and powerful software tools are available for converting a logic level design into a semiconductor circuit design ready to be etched and formed on a semiconductor substrate.
(137) The scope of protection sought for various embodiments of the disclosure is set out by the independent claims. The embodiments and features, if any, described in this specification that do not fall under the scope of the independent claims are to be interpreted as examples useful for understanding various embodiments of the disclosure.
(138) The foregoing description has provided by way of non-limiting examples a full and informative description of the exemplary embodiment of this disclosure. However, various modifications and adaptations may become apparent to those skilled in the relevant arts in view of the foregoing description, when read in conjunction with the accompanying drawings and the appended claims. However, all such and similar modifications of the teachings of this disclosure will still fall within the scope of this invention as defined in the appended claims. Indeed, there is a further embodiment comprising a combination of one or more embodiments with any of the other embodiments previously discussed.