METHODS AND DEVICES FOR DEFEATING BUFFER OVERFLOW PROBLEMS IN MULTI-CORE PROCESSORS
20230315463 · 2023-10-05
Inventors
Cpc classification
International classification
Abstract
Disclosed herein are methods and devices for defeating buffer overflow problems in multicore processors. In one embodiment, a processor implemented within a multicore processor integrated circuit (IC) is disclosed. The processor includes an instruction register and selection circuitry including a hardware latch operable to thwart a buffer overflow attack. The selection circuitry is electrically coupled with the instruction register. The selection circuitry is configured for: providing decrypted instructions to the instruction register when the hardware latch is in a first state and providing un-decrypted instructions to the instruction register when the hardware latch is in a second state. The coupling of the selection circuitry can be directly to the instruction register of a processor core, or indirectly by directing the output of the selection circuitry to cache memory inside the processor IC so that the instruction register only receives decrypted instructions from the cache memory.
Claims
1. A processor implemented within a multicore processor integrated circuit (IC), the processor comprising: an instruction register; and selection circuitry comprising a hardware latch operable to thwart a buffer overflow attack, wherein: the selection circuitry is electrically coupled with the instruction register; and the selection circuitry is configured for: providing decrypted instructions to the instruction register when the hardware latch is in a first state; and providing un-decrypted instructions to the instruction register when the hardware latch is in a second state.
2. The processor of claim 1, wherein the hardware latch is set to the first state upon receiving a decrypt command.
3. The processor of claim 2, wherein the hardware latch is set to the second state upon the processor exiting a reset.
4. The processor of claim 1, wherein selection circuitry further comprises: a multiplexor having a first input for receiving decrypted instructions; a second input for receiving un-decrypted instructions; and an output electrically coupled with the instruction register.
5. The processor of claim 1 further comprising a memory interface and the memory interface is configured for coupling to one or more memories, wherein the one or more memories are configured to store boot code instructions, unencrypted instructions, and encrypted instructions.
6. The processor of claim 5, wherein un-decrypted instructions include at least one of the boot code instructions and the unencrypted instructions.
7. The processor of claim 5, wherein the selection circuitry is further configured to receive the un-decrypted instructions from the memory interface.
8. The processor of claim 7 further comprising encryption/decryption circuitry, wherein: the encryption/decryption circuitry is electrically coupled between the memory interface and the selection circuitry; and the encryption/decryption circuitry is configured for: receiving the encrypted instructions from the memory interface; and decrypting the encrypted instructions to provide the decrypted instructions to the selection circuitry.
9. The processor of claim 8, wherein the encryption/decryption circuitry is further configured for: receiving the unencrypted instructions from the memory interface; and encrypting the unencrypted instructions to provide the encrypted instructions to the one or more memories via the memory interface.
10. The processor of claim 9, wherein encrypting the unencrypted instructions is based on a seed value and a built-in algorithm.
11. The processor of claim 10, wherein the data encryption and decryption circuits used to implement the encryption and decryption of instructions for the processor core inside a processor IC are removed from all the processor cores in a multicore processor IC and placed in the External Memory Interface or inside the interface between the cache memories closest to each processor core and the next level up, the purpose for which is to: reduce the frequency with which these circuits occur inside the processor IC, to reduce the heat generated by their presence; reduce the frequency with which these circuits occur inside the processor IC, to reduce the number of gates consumed by the processor IC; allow the internal cache of the processor to only have decrypted instructions residing in it so that an intelligent cache controller, such as a cache and predictive branch controller, is able to go through instructions stored in cache that are frequently being executed by a processor core and recognize conditional branch instructions and attempt to place instructions starting at any location the conditional branch could direct the processor core's program counter could go to, the goal of which is to minimize cache misses, that is, asking for an instruction that is not in cache and therefore requires the processor core to endure wait states until the instruction is brought to it from elsewhere; and further since the decryption circuits are only used when bringing in new instructions from a memory external to the processor IC or from a cache more distant from the processor core than the closest cache, whenever a processor core executes the same set of already decrypted instructions over and again while executing instruction loops out of cache, as the instructions are already decrypted in the cache memory or closest cache memory, the decryption circuits can remain idle and thus generate even less heat than if the circuits were inside each processor core.
12. The processor of claim 11, wherein the cache and predictive branch controller cannot tell which memory location is used to store the return address of the program counter, if said external memory location is unencrypted then by optionally disabling 8 and 16 bit instructions in 32 bit or larger processor cores, then optionally disabling 32 bit instructions in processor cores using instruction widths that are 64 bits or larger, these smaller instructions being the target of malicious users who attempt to change the return address to point to a series of smaller instructions that are incidentally embedded as parts of larger instructions in commonly used programs loaded in memory that can compromise a computing system, the attempt by a malicious user to execute these instructions is thwarted, preventing the compromising of the security of a computer system.
13. The processor of claim 12, when the program counter is stored on a stack during the execution of a subroutine call or the acknowledgement of an interrupt, the value stored on the system stack is not encrypted as the cache and predictive branch controller, which controls all processor core accesses to external memory, is unable to determine which memory location in external memory pointed to by a stack pointer contains the program counter return address.
14. The processor of claim 12, wherein the cache and predictive branch controller cannot tell which memory location is used to store the return address of the program counter, if said memory location is encrypted, then when restoring a processor core's program counter during a return from interrupt or return from subroutine, the processor core informs the cache and predictive branch controller that it will not accept any content that happens to reside in cache or cache sub frame, due to impurity caused by the undecrypted program counter return address, thus faulting and invalidating the cache sub frame, or the entire cache, depending on the cache's implementation choice and style, thereby forcing the cache and predictive branch controller to go to the external memory interface to access the return address, and that when the external memory location is accessed, its contents shall be decrypted before being sent to the processor core.
15. A method implemented on a processor comprising an instruction register and selection circuitry comprising a hardware latch, the method comprising: providing decrypted instructions to the instruction register from the selection circuitry when the hardware latch is in a first state; and providing un-decrypted instructions to the instruction register from the selection circuitry when the hardware latch is in a second state, wherein: the hardware latch is operable to thwart a buffer overflow attack on the processor; and the processor is implemented within a multicore processor integrated circuit (IC).
16. The method of claim 15, wherein the hardware latch is set to the first state upon receiving a decrypt command.
17. The method of claim 16, wherein the hardware latch is set to the second state upon the processor exiting a reset.
18. The method of claim 15, wherein selection circuitry further comprises: a multiplexor having a first input for receiving decrypted instructions; a second input for receiving un-decrypted instructions; and an output electrically coupled with the instruction register.
19. The method of claim 15, wherein: the processor further comprises a memory interface and the memory interface is configured for coupling to one or more memories; and the one or more memories are configured to store boot code instructions, unencrypted instructions, and encrypted instructions.
20. The method of claim 19, wherein the undecrypted instructions include at least one of the boot code instructions and the unencrypted instructions.
21. The method of claim 19, wherein the selection circuitry is further configured to receive the un-decrypted instructions from the memory interface.
22. The method of claim 21, wherein the processor further comprises encryption/decryption circuitry, wherein: the encryption/decryption circuitry is electrically coupled between the memory interface and the selection circuitry; and the encryption/decryption circuitry is configured for: receiving the encrypted instructions from the memory interface; decrypting the encrypted instructions to provide the decrypted instructions to the selection circuitry; receiving the unencrypted instructions from the memory interface; and encrypting the unencrypted instructions to provide the encrypted instructions to the one or more memories via the memory interface.
23. The method of claim 22, wherein the data encryption and decryption circuits used to implement the encryption and decryption of instructions for the processor core inside a processor IC are removed from all the processor cores in a multicore processor IC and placed in the external memory interface, the purpose for which is to: reduce the frequency with which these circuits occur inside the processor IC, to reduce the heat generated by their presence; reduce the frequency with which these circuits occur inside the processor IC, to reduce the number of gates consumed by the processor IC; allow the internal cache of the processor to only have decrypted instructions residing in it so that an intelligent cache controller, such as a cache and predictive branch controller, is able to go through instructions stored in cache that are frequently being executed by a processor core and recognize conditional branch instructions and attempt to place instructions starting at any location the conditional branch could direct the processor core's program counter could go to, the goal of which is to minimize cache misses, that is, asking for an instruction that is not in cache and therefore requires the processor core to endure wait states until the instruction is brought to it from elsewhere; and further since the decryption circuits are only used when bringing in new instructions from a memory external to the processor IC, whenever a processor core executes the same set of already decrypted instructions over and again while executing instruction loops out of cache, as the instructions are already decrypted in the cache memory, the decryption circuits can remain idle and thus generate even less heat than if the circuits were inside each processor core.
24. The method of claim 23, wherein the cache and predictive branch controller cannot tell which memory location is used to store the return address of the program counter, if said external memory location is unencrypted, then by optionally disabling 8 and 16 bit instructions in 32 bit or larger processor cores, and optionally disabling 32 bit instructions in processor cores using instruction widths of 64 bits or larger, these smaller instructions being the target of malicious users who attempt to change the return address to point to a series of smaller instructions that are incidentally embedded as parts of larger instructions in commonly used programs loaded in memory that can compromise a computing system, the attempt by a malicious user to execute these instructions is thwarted, preventing the compromising of the security of a computer system.
25. The method of claim 24, when the program counter is stored on a stack during the execution of a subroutine call or the acknowledgement of an interrupt, the value stored on the system stack is not encrypted as the cache and predictive branch controller, which controls all processor core accesses to external memory, is unable to determine which memory location in external memory pointed to by a stack pointer contains the program counter return address.
26. The method of claim 23, wherein the cache and predictive branch controller cannot tell which memory location is used to store the return address of the program counter, if said memory location is encrypted, then when restoring a processor core's program counter during a return from interrupt or return from subroutine, the processor core informs the cache and predictive branch controller that it will not accept any content that happens to reside in cache or cache sub frame, due to impurity caused by the undecrypted program counter return address, thus faulting and invalidating the cache sub frame, or the entire cache, depending on the cache's implementation choice and style, thereby forcing the cache and predictive branch controller to go to the external memory interface to access the return address, and that when the external memory location is accessed, its contents shall be decrypted before being sent to the processor core.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
[0026] The previous summary and the following detailed descriptions are to be read in view of the drawings, which illustrate particular exemplary embodiments and features as briefly described below. The summary and detailed descriptions, however, are not limited to only those embodiments and features explicitly illustrated.
[0027]
[0028]
[0029]
[0030]
[0031]
[0032]
DETAILED DESCRIPTION
[0033] These descriptions are presented with sufficient details to provide an understanding of one or more particular embodiments of broader inventive subject matters. These descriptions expound upon and exemplify particular features of those particular embodiments without limiting the inventive subject matters to the explicitly described embodiments and features. Considerations in view of these descriptions will likely give rise to additional and similar embodiments and features without departing from the scope of the inventive subject matters. Although the term “step” may be expressly used or implied relating to features of processes or methods, no implication is made of any particular order or sequence among such expressed or implied steps unless an order or sequence is explicitly stated.
[0034] Functional implementations according to one or more embodiments are illustrated in the drawings. The following definitions may be used in the drawings and in these descriptions:
[0035] Boot Code Instructions executed by a processor when it first comes out of reset. Boot Code 101b has the privilege of always being stored in a non-volatile memory that cannot be modified by malicious users (in a properly designed processor), which always allows a processor to come out of reset in a known state.
[0036] Encryption Algorithm Specially designed hardware logic or a sequence of processor instructions that modifies the contents of a new instruction being stored to memory so that when it is decrypted it will be returned to its original value.
[0037] Decryption Algorithm Specially designed hardware logic or a sequence of processor instructions that modifies the contents of an encrypted instruction so that it is returned to its original value. Note that decryption step does not involve writing decrypted instructions back to memory, so the memory contents remain encrypted even after being read. The sequence of processor instructions that modifies an encrypted instruction so that it is returned to its original value will typically be disabled by people developing and debugging code, whereas the specially designed hardware logic would do so automatically and in real time during normal processor execution to provide the actual protection from a BOA.
[0038] Seed Value A randomly generated number that determines how an encryption algorithm is used to encrypt instructions, and how a decryption algorithm is used to decrypt instructions.
[0039] Non-volatile Memory Memory whose contents are preserved when power is removed.
[0040] Volatile Memory Memory whose contents are not preserved when power is removed.
[0041] Cache A small memory that is usually internal to the processor IC, but is much, much faster to access than most external volatile or non-volatile memory. Usually cache is located inside the same integrated circuit that the processor is located in. Because of its high access speed, Cache will cost more. However, due to its small size, the cost impact is trivial. Special logic is used to control Cache so that its contents will mirror the contents of the most commonly accessed portions of Main Memory 101a or Boot Code 101b. When a memory access to Main Memory 101a or Boot Code 101b is to a section that is mirrored in the Cache, the Cache is used rather than the Main Memory 101a or Boot Code 101b, reducing the wait time by the processor and speeding it up. This is often times called a cache hit. When a memory access to Main Memory 101a or Boot Code 101b is to a location that is not mirrored in Cache, then the processor must wait while the Main Memory 101a or Boot Code 101b responds. This is often times called a cache miss. During a cache miss, the logic managing the Cache will determine which part of Cache has been used the least in recent accesses and overwrite it with the contents of the Main Memory 101a's or Boot Code 101b's latest access to increase the chances of more cache hits in the future. When Cache contents are declared invalid, then they must be reloaded from Main Memory 101a or Boot Code 101b to be considered valid again.
[0042] Cache Hit An instruction or data needed by the processor is in a cache 101c and therefore can get into the processor itself sooner, increasing processing throughput and speed of operation.
[0043] Cache Miss An instruction or data needed by the process is not in a cache 101c, therefore it must be brought into the cache by a more remote cache or Main Memory 101a, which takes longer and thus the processor has to wait, reducing processing throughput and speed of operation.
[0044] Main Memory The bulk of a processor's memory, usually located outside the integrated circuit the processor is located in.
[0045] Read Only Memory A non-volatile memory whose contents cannot be modified.
[0046] Inter-Integrated Circuit A protocol that uses a minimum number of pins to transfer data between a master device such as a processor and a slave device such as a memory chip.
[0047] Exception An interrupt to a processor caused by an undefined or illegal instruction, or unauthorized access to a memory location. Properly written code will not generate exceptions. Malicious code that was decrypted and as a result is turned into random, chaotic instructions will eventually create an exception.
[0048] Indexed Address An address pointing to a location in memory that uses a Processor Register 106 to provide a base value. As the Processor Register 106 is incremented or decremented after each access, the memory location for the next access changes without having to modify the instruction itself. This is useful for reading or writing data from or to adjacent memory locations, such as in a temporary data buffer.
[0049] Extended Address An address that points to a location in memory that is not referenced to a Processor Register 106. This is useful for accessing the start of instructions in a Boot Code 101b, or for input and output devices such as disk drives, whose addresses do not change.
[0050] Immediate Data Data that is part of an instruction. For example, assume a certain command must be written to a disk drive in order for it to spin up before files can be read from or written to it. As an example, an immediate data value will be loaded into a Processor Register 106 by an instruction, followed by another instruction which writes the Processor Register 106 containing the immediate data to the disk drive controller. The immediate data will contain a command that tells the disk drive to spin up so it can be accessed.
[0051] The following acronyms may be used in drawings and in these descriptions: [0052] ALU Arithmetic Logic Unit [0053] BOA Buffer Overflow Attack [0054] EDC Encryption and Decryption Circuitry [0055] CER Command Encryption Register [0056] EDC Encryption and Decryption Circuitry [0057] IC Integrated Circuit [0058] I.sup.2C Inter-Integrated Circuit [0059] IEC Instruction Execution Circuitry [0060] JOP Jump On Program [0061] NVM Non-Volatile Memory [0062] PCB Printed Circuit Board [0063] PMI Processor Memory Interface [0064] OS Operating System [0065] RNG Random Number Generator [0066] ROP Return On Program [0067] RTS Return from Subroutine [0068] RTI Return from Interrupt
[0069] Instructions are read from Memory 101 (see
[0070] The arrangement shown in
[0071] The sequencing of the commands from the IEC 102 implements the instruction and provides for the desired outcome of the instruction. Succeeding instructions are read sequentially from Memory 101 and executed in sequence, providing a deterministic outcome that can repeat itself over and over again with a very, very high degree of reliability. This high reliability and repeatability has lead to the use of processors to control and implement much of the more tedious and boring tasks in society, as well as provide new features that a generation or two ago were inconceivable.
[0072] At least one embodiment (see
[0073] Once enough of the operating system has been stored in Main Memory 101a for it to take over, the Boot Code 101b will simultaneously 1) instruct the processor to start executing code from Main Memory 101a where the operating system has been stored and 2) sends the Decrypt Command 206 to a Latch 205 (see
[0074] Once instruction decryption begins, Latch 205 cannot be switched back to selecting un-decrypted instructions except by a processor reset 207. This is necessary as all instructions in Main Memory 101a are now encrypted and must be decrypted each time they are read of out of Main Memory 101a before being sent to the IR 103, as decrypted commands are not written back out to the Main Memory 101a. By also making it impossible to turn decryption off by a command, malicious code would be unable to change the processor back over to its more vulnerable state where a BOA could become successful.
[0075] The method of how the processor selects decrypted instructions or un-decrypted instructions must be such that when the Latch 205 is set it always selects decrypted instructions, and when Latch 205 is not set (that is, it is in a clear state after a reset), it always selects un-decrypted instructions. As an example of how this is accomplished, in
[0076] Because the instructions stored in Main Memory 101a are now encrypted and the Seed Value is unknown to the outside world, malicious users will have to guess at what the Seed Value is, and perhaps even the encryption algorithm. If the malicious user guesses wrong, then when the malicious code placed in Main Memory 101a is decrypted, it isn't turned into the desired instructions. Instead it is turned into random, unpredictable values. The unpredictable instructions produce chaotic results. Because the results are chaotic and do not produce a deterministic result, the processor will not be taken over by the malicious user. Eventually the random, chaotic results will generate an ‘exception’, which is an interrupt to the processor caused by misbehaving code. The exception handler code in the processor will know what part of Main Memory 101a the code was being executed out of when the exception occurs, and will compare its contents (after decrypting it) with what should be there. If there is a difference, the processor will assume it has suffered a BOA and either 1) stop the process that resided in the compromised block of Main Memory 101a, and reload it, or 2) reset itself.
[0077] Note that each reset should generate a different random number for the Seed Value. Hence the malicious user will not know if a previously unsuccessful guess would have actually been the new Seed Value; in other words, after a processor reset, the malicious user will have to start all over again trying to guess what the Seed Value is. Frequently the malicious user will also be unaware of when a processor targeted by the malicious user is reset, further adding to the uncertainty facing the malicious user.
[0078] Since the feedback mechanism between implementing the BOA and determining if the results are successful is extremely slow, an encryption algorithm that implements a reasonably large number of different permutations would take many decades for the malicious user to successfully guess at the correct algorithm. The net result is that the malicious user will tire of their efforts to take control of the processor and stop their BOA attacks. Further, by resetting the processor on a periodic basis or after several unsuccessful BOA attacks have been detected, any past record of known guesses as to what the Seed Value is, but now known to not be the correct seed value, by a malicious user are rendered useless because after a reset the Seed Value will be different, and in fact could be that one of those past attempts would now be the new Seed Value. The malicious user would have to start over again, but due to the nature of their being unsuccessful in implementing a BOA attack, they would have difficulty even knowing that their targeted processor was reset and thus requiring them to start over again, further frustrating their efforts.
[0079] In at least one embodiment, the Encryption and Decryption Circuitry 301 (EDC) in
[0080] The encryption algorithm may actually be one of several different algorithms, not all of which are used in any one processor. To select which algorithm(s) is used can be done by a number of means. In a typical example shown in
[0081] In at least one embodiment shown in
[0082] Decryption algorithms should minimize any delay, or have no delay placed on the flow of an instruction from Memory 101 to the IR 103. As there may be some delay in the decoding logic, it may be necessary to ‘pipeline’ the instruction and use an additional stage of registers.
[0083] During the instruction debugging phase, it may be desirable to disable the EDC 301 so that it does not modify any instruction passing through it. An external pin (not shown in the drawings) on the processor may be used to force the Seed Value in the CER 302 or CER 202 to assume a state that does not encrypt or decrypt instructions. By allowing the signal to float when encryption is to be enabled, or connecting the pin to a low voltage signal such as the ground return signal when encryption is to be disabled, the option to enable or disable encryption is implemented. An optional resistor that is taken out of the bill of materials of a PCB design for production PCBs will provide the needed connection to the ground return line during the debugging phase in a laboratory setting. But by not being inserted for PCBs delivered to customers, the missing resistor ensures that the encryption to stop BOA will be implemented. This is an example of how encryption/decryption can be disabled for code troubleshooting but enabled for production PCBs, however, this method of selectively enabling or disabling encryption by a hardware means does not limit the scope of the claimed subject matter to just this one method.
[0084] Two suggested encryption and decryption algorithms are 1) using the Seed Value, invert selected bits in the instruction, and 2) taking groups of four bits in each instruction, use the Seed Value to swap their positions around. Neither algorithm depends on the state of a bit in the instruction to determine the final outcome of another bit in the instruction. Both algorithms preserve the uniqueness of every bit in the instruction so that the instruction can be faithfully reconstructed during decryption, and both algorithms minimize the amount of logic needed to implement them. It will take one bit of a Seed Value for each bit in the instruction to implement the inversion algorithm, and it will take five bits of a Seed Value for each four bits in the instruction to implement the suggested bit swapping algorithm to provide any of the 24 possible combinations when swapping four bits around. For a 32 bit instruction, the two algorithms provide 2.sup.32 and 2.sup.48, respectively, different permutations; combined they provide over 4.7×10.sup.20 permutations. Larger instructions will involve even larger numbers of permutations. Due to the slow speed by which feedback back to the malicious user on the success or failure of a particular guess is, the number of permutations from a 32 bit instruction alone will be adequate to discourage all future BOA attacks. For 64 bit instructions, the odds are that the processor itself will wear out long before a malicious hacker could ever stumble across the correct Seed Value and algorithm even if the processor is never reset again.
[0085]
[0086]
[0087]
[0088]
[0089]
[0090] A novel concept is implemented to modify the bit arrangement and bit states of instructions for a processor with the goal rendering a malicious user unable to execute a successful BOA. In at least one example, the modification technique used can provide more than 4.7×10.sup.20 permutations on the changes to the bit arrangement and bit states. Given the slow rate with which a malicious user would get feedback on the success or failure of each attempted BOA, it would take many decades for the malicious user to eventually come to the correct permutation. Each time a processor is reset, a different permutation is typically used. This renders all previous failed attempts of a BOA, which the malicious user would use to indicate the permutations that are invalid, mute, as the new permutation after a reset could be one of those permutations the user previously tried and determined were incorrect.
[0091] In some embodiments, all processor instructions written to Main Memory 101a are to be encrypted with the selected permutation, so that when an encrypted instruction is read from Main Memory 101a and decrypted, the instruction will be restored to its original value. To enable this to happen, after reset the processor will not decrypt any instructions while it executes said instructions from a special memory called Boot Code 101b. Boot Code 101b are instructions stored in a non-volatile memory, and having a further attribute that Boot Code 101b is not intended to be changed, unlike code written to a modifiable non-volatile memory such as a disk drive.
[0092] The Boot Code 101b will bring the processor and a minimum set of its input/output components to a known operating state after each reset. In one embodiment it will generate a Seed Value for instruction encryption and decryption. The Boot Code 101b will load the instructions for the processor's operating system into Main Memory 101a, encrypting the instructions prior to writing them to Main Memory 101a.
[0093] After enough of the operating system has been written to Main Memory 101a for the Boot Code 101b to transfer code execution to Main Memory 101a, the Boot Code 101b executes a command that simultaneously starts executing instructions out of the Main Memory 101a and enables instruction decryption to occur.
[0094] Many processors have a special, internal memory called ‘Cache’, which is a volatile memory that is accessed a lot more quickly than Main Memory 101a or Boot Code 101b. The purpose of Cache 101c is to hold the most recently used instructions and data inside the same integrated circuit the processor is in so it can operate faster, as well as freeing up the integrated circuit's External Memory Interface so data can flow into and out of the processor without being slowed down by accesses to frequently used instructions. As such, Cache 101c will contain a copy of the contents of Main Memory 101a or Boot Code 101b that was recently read from or written to.
[0095] Prior to executing the instruction to start decryption, much of the Boot Code 101b may be stored in Cache 101c. As this Boot Code 101b in Cache 101c is unencrypted, it must be ‘flushed’ or declared invalid so there will be no further attempt to use it once instruction decryption starts. If decryption starts without doing so, any Boot Code 101b that is accidentally executed will be changed to unintelligible instructions by the decryption process. That could cause the processor to behave erratically, so the Cache 101c contents must be declared invalid to prevent them from being accessed after decryption starts. Only after Cache 101c is re-loaded with encrypted instructions from Main Memory 101a can its contents be reclassified as valid. If the processor operating system deems that it must execute more Boot Code 101b, it must read the Boot Code 101b, encrypt it and then store it in Main Memory 101a for execution just like it would do so for its operating system or any other code that it reads from a disk drive.
[0096] If a BOA attack occurs on the processor, the malicious code that will be executed will be rendered unintelligible by the decryption process. Unintelligible code will quickly result in an error event called an ‘exception’. An exception can include errors such as accessing non-existent memory, a lower priority operating state accessing memory or input/output devices reserved for a higher priority state or another process, attempting to write to memory that is write protected, executing an unimplemented instruction, dividing by zero, etc. Once one of these errors occurs, the processor will save its register contents for later analysis and then jump to a higher priority operating state. From this higher priority state the processor will examine the instructions in the Main Memory 101a where the exception occurred and compare them with what should be in that location by reading what was loaded there from the disk drive. If it finds a mismatch, the processor should assume it has suffered a BOA attack and shut down the process that uses that portion of Main Memory 101a and reload it, or if it determines it has suffered multiple BOA attacks or cannot safely shut down that process, the processor resets itself.
[0097] Additional instructions need to be added to the processor to enable the encryption and decryption process to occur. One instruction will be the previously mentioned instruction of beginning to execute encrypted code, which involves transferring program control to another part of memory, turning on the decryption process, and for processors with Cache 101c, declaring the entire Cache 101c (which are all of Cache 610, 611, 612 in
[0098] In an enhanced embodiment, another instruction will be to store an unencrypted value in a register associated with the EDC 301 and read out an encrypted version of it. Another instruction will be to write an encrypted value in a register associated with the EDC 301 and reading out the unencrypted value. These instructions will ease encryption and debugging, and for systems with a Seed Value the processor is not allowed to read, provide the only means of encrypting instructions and examining an area of memory where an exception occurred to determine if the processor has suffered a BOA.
[0099] An enhanced embodiment will provide a means of generating a Seed Value for the encryption and decryption process that cannot be read by the processor. This enhances security in that the Seed Value cannot be accidentally disclosed. Note that for debugging purposes it may be necessary to suppress the Seed Value so that there is no encryption or decryption, therefore, the voltage level on an input pin into the processor can allow or deny the processor the ability to use its Seed Value.
[0100] Another enhanced embodiment will decrypt not just actual instructions, but any data in the instruction stream such as immediate data, indexed addressing values or extended addresses. This enhanced version does not require the processor to seek out instructions meant only for the IR 103 in the instruction stream to be encrypted while leaving any addressing information or immediate data unencrypted; all can be encrypted.
[0101] Another enhanced embodiment will have an encryption and Decryption Circuitry possessing multiple different possible algorithms, with the actual algorithms that will be used by the processor randomly selected during the processor's PCB manufacturing. By assigning a different set of algorithms to each PCB in a PCB lot, it will not be possible for someone intimately familiar with the manufacturing process to be able to sell information as to which algorithms were used for a particular lot of PCBs.
Solving the Multi-Core Processor BOA
[0102] As a further embodiment, of the previous disclosure (including
[0103] Alternately, the cache & predictive branch controller shares the seed value used to encrypt instructions so that it can read the contents of Level One Cache 612, or possibly also Level Two Cache 611, or possibly also Level Three Cache 610 while they are still encrypted. This enables the cache & predictive branch controller to still seek out both destinations of a branch instruction and place those addresses in the Cache 101c, hopefully even the Level One Cache 612 that is closest to the Processor Core 613, to minimize the number of cache misses that occur.
[0104] However, for Return on Program (ROP) or Jump on Program (JOP) BOA, the use of Cache 101c and a cache & predictive branch controllers cannot be used to assist the processor IC in protecting itself against ROP and JOP attacks.
[0105] To protect multi-core processor ICs from ROP and JOP attacks, the smaller instructions that are the target of ROP and JOP attacks can be disabled on a per process basis, which shuts down ROP and JOP. Alternately, when a Processor Core does a Return from Interrupt (RTI) or Return from Subroutine (RTS) instruction, the Processor Core will refuse to use whatever is stored in Cache 101c as the cache controller cannot tell which data pulled from stack memory is the return address, so it won't know which one to decrypt. By forcing the cache & predictive branch controller to use Main Memory 101a only for RTS and RTI, or only those Cache 101c where the contents are still encrypted, the external memory controller or the cache and a cache & predictive branch controller can be told which address contains encrypted return address of the Program Counter and decrypts it coming into the processor, which then provides the decrypted return address for the program counter.
[0106] User data that is privileged is protected from unauthorized accesses while stored in mass storage devices by a powerful encrypted algorithms such as AES-256, its equivalent, or better. But when it is loaded into a computer's memory that protection goes away. To protect privileged information in Main Memory 101a (and possibly any cache where privileged data may still reside) the same techniques used to encrypt instructions are used to encrypt user data. Each process handling privileged data will have its own encryption key that is passed to an encryption and decryption circuit used to encrypt data being written to memory, or decrypt it when it is read into the processor the process is running on. When the process is through using the data it will delete the encryption key that is unique to the data, instruct the cache and a cache & predictive branch controller to declare those parts of Cache holding such data to declare those locations invalid, and instruct the MMU to overwrite such data in Main Memory 101a.
Background for the Multi-Core Processor Solution
[0107] As previously described, said invention describes the problem of how buffer overflow attacks (BOAs) threaten to install malicious software disguised as data into the Main Memory 101a of a processor with the intent get the code to run and thus compromise the processor). Main Memory 101a is read/writable memory (often called Random Access Memory, or RAM) where both data and instructions co-exist. The BOA attempts to write so much data into the processor that it overflows the boundaries set up for the data buffer in Main Memory 101a, with the malicious data subsequently overflowing the buffer boundaries set up for it and ends up being written into Main Memory 101a set aside for instructions that the processor executes.
[0108] Attempts to use protective software alone to solve this problem have failed to completely mitigate the problem. Today it is a multi-hundred million dollar industry attempting to stay ahead of all the different ways malicious users take control of processors. And it consumes a significant percentage of a processor's computing capacity to run this still ineffective software. This increases data costs in terms of more processors needed to provide the computing power to do the job, power consumption needed to provide the additional processors plus the additional energy required of air conditioners to cool the additional processors, and floor space for the additional processors which translates into bigger building costs in multiple categories.
[0109] The solution to this problem is to install additional hardware in the processor that disguises the software the processor executes, encrypting it with a simple algorithm that is difficult to guess for hackers because of a near complete lack of feedback on the success of their attempts to do so (referred to as “zero feedback”), and the sheer size of the encryption key. As processor instructions were read from Main Memory 101a they were decrypted just before execution. The decryption process is not written back to memory, which means the same instructions were kept encrypted in memory and had to be decrypted each time they were executed.
[0110] This was by design, as once the decryption process was started, nothing short of a processor reset would turn it off. The goal was to make the processor as secure as possible, which means it must continue decrypting instructions indefinitely once started.
[0111] This solution works extremely well for processors with a single core, or a small handful of cores. However, each Processor Core has to have an implementation of the encryption and decryption logic within it. When large numbers of cores exist inside a processor, to place a copy of this circuitry inside each core increases the power consumption of the Processor Core, and the complexity (and hence the costs) of the same processor IC. Clearly, a better way of encrypting and decrypting instructions inside a processor IC with a large number of cores is needed.
[0112] Another problem arises with any processor Integrated Circuit (IC) containing cache memory that uses predictive branching. A cache memory is a small local memory placed inside the processor IC that can be accessed more swiftly than Main Memory 101a, reducing the number of processor clock cycles that the processor must delay, or “wait” for the instruction to reach the processor. The higher end processors with faster clock speeds and more cores will have multiple levels of cache, with the closest cache to a Processor Core, typically called the Level One Cache, is usually the smallest, but most quickly accessible local cache memory. Typically each Processor Core will have its own dedicated Level One Cache or shares the Level One Cache with a very small number of other cores. In general, a Level One Cache can be accessed with zero wait states.
[0113] Larger numbers of cores will share a larger but slightly slower Level Two Cache, typically with more wait states than a Level One Cache but still fewer than Main Memory 101a. And in the largest processor ICs, there would be an even larger, but also slower, Level Three Cache typically (but not always) accessible by all cores in the processor, or a few Level Three Cache such that all cores in the processor have access to a Level Three Cache. It will typically have more wait states than a Level Two Cache, but still fewer wait states than having to access the same instructions from Main Memory 101a. Other implementations may have more levels of cache or fewer levels of cache, what is described herein is intended as an example only and not intended to limit the scope of the claims.
[0114] The job of a cache memory is to carry an identical copy of an instruction or data kept in Main Memory 101a. This creates management complexities, in that when a Main Memory 101a content changes, then the copy carried inside the cache in the processor IC is no longer valid and must be refreshed. This problem has already been solved by various means and is beyond the scope of this disclosure, but is only mentioned here to ascertain that the inventor is aware of the problem and that it has been solved.
[0115] Among other things, predictive branching is the process that a cache memory controller performs by going through the contents of the cache that the Processor Core is executing out of, looking for the next “conditional branch” instruction. A conditional branch is an instruction that goes to one section of memory if a test condition is true, or else it continues with the next instruction in sequence if the test condition is not true. For example, if the previous mathematical operation resulted in a value of zero, then a “branch if zero” instruction will cause the processor to execute code from a different area of memory rather than the next instruction in sequence. Predictive branching will attempt to determine where the new location the processor jumps to will be and ensure its contents are in cache, as well as the next instructions past the conditional branch instruction, so that no matter whether the processor branches or continues with the next instruction in sequence, both will be in the cache and therefore the processor will not have to wait for Main Memory 101a to be accessed to continue.
[0116] If the instructions stored in Cache 101c remain encrypted, the controller that executes the predictive branching session will be unable to identify the branch instruction, and thus not be able to make attempts to keep the potential destinations of the branch resident in cache.
Description of the Multi-Core Processor BOA Solution
[0117] A way to get around these problems is to place both the encryption and Decryption Circuitry of the invention in the External Memory Interface(s) of the processor IC rather than in each Processor Core inside the processor IC, or to place it between the Level One Cache and the Level Two Cache, or between the Level Two Cache and the Level Three Cache, such that the contents of at least the Level One Cache are not encrypted.
[0118] There should be more than one instance of the Decryption Circuitry and at least one instance of the encryption circuitry at each External Memory Interface. Data rates for External Memory Interfaces may be too fast for a single instance of the decryption circuit 201b to be able to keep up with the incoming data. Therefore, multiple instances of each circuit need to exist in the memory interface. These multiple instances of the Decryption Circuitry would accept a single data transfer on a sequential basis; that is, one decryption circuit 201b would accept one data transfer, the next decryption circuit 201b in line gets the next data transfer, etc. until the first one has had time to complete the decryption process and pass it on to a buffer that can present it to the rest of the processor IC. Once the decrypted instruction is passed from the Decryption Circuitry, the first decryption circuit 201b is ready to accept a new data transfer, starting the process all over again. Control of these decryption circuits can be done with a token passing scheme or a round-robin controller that continuously cycles through all of the decryption circuits. If no data is available the selected decryption circuit 201b stays idle and doesn't accept anything off of the external data bus. If available, the decryption circuit 201b accepts the data and however many clock cycles later that it takes to complete the decryption process presents the decrypted data to the rest of the processor IC.
[0119] Staying with each data word that is decrypted is the fetch completion information. The fetch completion information is not changed as the data word is decrypted. The use of the fetch completion information may change from one processor IC architecture to another, but in general it contains information on which Processor Core, and if multi-threaded, which processor thread within the core, requested the data.
[0120] This is true whether there are only a few Processor Cores or many Processor Cores (even over a hundred Processor Cores) in the processor IC.
[0121] It is not necessary for each memory interface to share the same random number used to encrypt and decrypt the instructions with each other. As each External Memory Interface is physically and electrically separated from each other, there will not be the problem of one memory interface attempting to decrypt a memory location encrypted by a different encryption key from another memory interface. Therefore, each memory interface may have its own Random Number Generator (RNG) if that is more expedient in adding the encryption and decryption circuit 201b to each one when there are more than one memory interfaces on a processor IC. Alternately, if it is easier to install an RNG common to several or all External Memory Interfaces the designer(s) of the processor IC may do so.
[0122] Any Main Memory 101c location accessed by the program counter 106a of a Processor Core 613, or by the predictive branching operation, must be decrypted before it enters the cache or the Processor Core that needs it. Thus, the External Memory Interface 602 will be notified by the Processor Core 613 or by the cache & predictive branch controller that the memory location needs to be decrypted before being sent to the core or the cache memory. This way all instructions inside the processor IC are decrypted when they reach the core itself, and the Decryption Circuitry does not have to reside in any of the Processor Cores 613 anymore.
[0123] This solution also requires that the encryption circuitry reside in the External Memory Interface 602. When a processor wants to load a new program into Main Memory 101c (for example, a web browser, a word processor, or a spreadsheet) instead of writing the instructions directly to Main Memory 101c, the processor would write them to an Input/Output (IO) port at the External Memory Interface 602, along with the address where the instruction is to be stored in Main Memory 101c. The IO port encrypts the instruction, and then the External Memory Interface 602 does the actual writing of the encrypted instruction into Main Memory 101c.
[0124] An alternate solution is to place the Decryption Circuitry in the cache & predictive branch controller Interface 609b between Level One Cache 612 and Level Two Cache 611, or in the cache & predictive branch controller Interface 609a between Level Two Cache 611 and Level Three Cache 610. The encryption of instructions being stored in Main Memory 101c would still reside in the External Memory Interface 602.
ROP and JOP Vulnerabilities
[0125] However, this process makes it difficult to handle the problem of a BOA known as Return on Program (ROP) and Jump on Program (JOP). ROP and JOP attacks are attacks created by malicious users who go through commonly used set of 32 bit or 64 bit instructions found in most processors such as the operating systems (OS) or mathematical libraries. The malicious users' goal is to find 8 bit or 16 bit instructions that were unintentionally created as part of a 32 bit or 64 bit instruction. While not intended to be instructions of their own by their original authors, they will be interpreted as such if the program counter were to jump to the 2nd or later byte of a 32 bit or 64 bit instruction. When they are strung together they create new programs that can compromise processor integrity. The malicious user then has to trick the processor into jumping directly to the address of these 8 and 16 bit instructions rather than the address of 32 bit or 64 bit instruction they were part of.
[0126] Normal processor operations will never access these memory locations as 8 or 16 bit instructions. However, when a processor executes a subroutine or an interrupt, the program counter value is pushed, or stored in data memory in what is known as the stack, which is a section of memory pointed by a stack pointer for the storage of temporary values useful to the program execution. Some buffer overflow attacks will overwrite data stored on the stack, and this includes the return address of the processor. This will point it to a ROP or JOP address, compromising the processor as a result. To mitigate this, the existing invention also encrypts all writes of the program counter to Main Memory 101a, such as during an interrupt or subroutine call, and decrypts the return address before restoring it to the program counter. If the location on the system stack where the program counter was stored is over-written by a malicious user during a buffer overflow attack to point to where a ROP or JOP attack can take place, it will be modified by the decryption process and thus not point to the offending location anymore.
[0127] The problem with placing the decryption process at the External Memory Interface 602 of a processor IC, instead of inside each Processor Core itself, is that the cache controller cannot figure out which location on a stack is the return address of the program counter. As such it cannot tell the External Memory Interface to decrypt what should be the return address. Any return address will have to be placed in cache memory as is, making a multi-core processor IC vulnerable to ROP and JOP attacks again.
ROP and JOP Solutions
[0128] There are two possible solutions to this problem.
[0129] A solution to the ROP and JOP attacks is to disable all 8 and 16 bit instructions in a 32 bit or 64 bit processor, and if processors progress beyond 64 bits, possibly disable 32 bit instructions separately as well. There should be a means of disabling 8 and 16 bit instructions, and another means of disabling 32 bit instructions when processors progress beyond 64 bit data busses. However, such a drastic solution will be strongly opposed by the software community as many of the 8 and 16 bit instructions are quicker to execute and take less memory space than a 32 bit or larger instruction meant to do the same thing. A possible compromise would be to allow each processor session to be able to re-enable these shorter instructions while the session is running, and when the session is not running, they are disabled again until another session (or the same session starts again) and enables them for itself. Note that enabling the smaller instructions does leave the core vulnerable to ROP and JOP attacks and must be done only when essential.
[0130] When this option is implemented, a flag needs to be saved for the session that enables these instructions. This way if the session gets interrupted, these instructions will be disabled for the interrupt. They are then automatically re-enabled when the session continues from an interrupt and the flag is pulled back into the process' control register re-enabling them. Note that at the moment the interrupt is acknowledged the 8 & 16 bit opcodes must be disabled, as the interrupt procedure will assume this and thus not realize it would be vulnerable to ROP and JOP if that were not the case.
The Other Way of Solving the Problem is this:
[0131] Any time the program counter is pushed onto the system stack, which is what happens during a subroutine call or during an interrupt, the Processor Core informs the External Memory Interface 602 that it will use the encrypting IO port to do so. The program counter's current value is sent to the encrypting IO port along with the address of the location on the system stack where it will be stored. The IO port then proceeds to write the encrypted program counter value to Main Memory 101a. When the return from interrupt (RTI) or return from subroutine (RTS) instruction is then executed, the Processor Core 613 informs the External Memory Interface 602 that it will not accept a cached value for the return address of the program counter, therefore the External Memory Interface 602 will have to fetch the return address from Main Memory 101a and decrypt it, or have the cache & predictive branch controller do so if that is where the Decryption Circuitry exists. This will slow the processor down, but these instructions are not executed very frequently at all, unlike instructions in a loop that can be executed a dozen to hundreds or even thousands of times and can be executed quickly when pulled from cache.
[0132] This second option is the most secure, and requires different hardware to implement than the option to disable 8 & 16 bit opcodes. Both options could co-exist in a processor IC, although if option 2 exists it makes no sense at all to be able to implement option 1. Option 2 will slow the core down slightly, but since this is rare event in code execution, the impact will be insignificant. And the time consumed having to pull a return address from Main Memory 101a may be partially made up by not having to execute an instruction to enable or disable 8 and 16 bit instructions.
[0133] Note that the option to not accept a cached value for a certain memory location read is already implemented in many processor ICs so they can deal with deterministic processes. There is a certain variability in fetching something from Cache 101c vs. Main Memory 101a that can make deterministic processes less deterministic. By not accepting a cached value, the variability is reduced. Therefore, in most instances of processor IC designs, very little additional hardware gates (multiple transistors combined together correctly correspond to a gate) will have to be added to implement the 2nd ROP & JOP protection option, as most of the gates already exists for other reasons.
Implementation Multi-Core Processor BOA Solution
[0134] Notes for the following sections in this disclosure: Main Memory 101a, Decryption Circuitry 301b, and Encryption Circuitry 301a reside in
[0135] Referring to
[0136]
[0137] In the processor IC 601 with multiple Processor Cores 613 in it, encryption takes place when any Processor Core 613 writes to the encryption circuitry 301a. The encryption process will also take place automatically if the processor IC design includes the option of writing any storage of the program counter 106a on a stack during a subroutine call or the acknowledgement of an interrupt to the encryption circuitry 301a by any of the Processor Cores.
[0138] The encryption circuitry 301a encrypts the data written to it by the Processor Core 613 and then the encryption circuitry 301a writes the data to the Main Memory 101a over the external memory data bus 604 and control and address bus 605.
[0139] In most processor IC 601 implementations the write data bus 604 and the read data bus 603 are the same bus, but for easier illustration they are shown separately. The claims of this disclosure are applicable to either, however. Both will use the control and address bus 605 to access external memory.
[0140] The description in the implementation described herein consists of cache memories 101c (see
[0141] The method of initially loading encrypted instructions in Main Memory 101a is the same as previously described in the invention's claims 1 through 18. After a processor IC 601 reset, the random number generator 606 develops a seed number that is used by the encryption circuitry 301a in the External Memory Interface 602 to encrypt the instructions as they are written to Main Memory 101a, and the Decryption Circuitry 301b uses the same seed number to decrypt the instruction. Once decryption is enabled by a Decrypt Command 205, the decrypt latch 206 is set and from this point forward the cache & predictive branch controller will always inform the External Memory Interface 602 to pass instructions through the Decryption Circuitry 301b before storing in any local cache 101c, which consists of a Level One Cache 612, possibly a Level Two Cache 611, and possibly a Level Three Cache 610, or even higher levels of cache, which could exist but are not in
[0142] How the processor IC 601 tells the difference between writes to Main Memory 101a that must be encrypted and those that bypass the encryption is dependent upon the internal implementation of the processor IC 601 designers. In this example it will be assumed that a “Pre Opcode” instruction executed inside each Processor Core 613 precedes the actual write instruction, which means the data to be written will be encrypted by the encryption circuitry 301a, and when the write instruction is not preceded by said “Pre Opcode” instruction, the data to be written passes through (or it may bypass) the encryption circuitry 301a without being encrypted. Other methods of using or bypassing the encryption circuitry 301a are also valid, this invention is not limited in scope to using a “Pre Opcode” as the only mechanism to do so. Other valid methods are to have a separate set of instructions for this particular purpose or setting a bit in a control register, but again this does not limit the scope of this invention.
The Purpose of Cache Memory
[0143] In the example, the Processor Core 613 does not directly fetch an instruction from Main Memory 101a. Instead, it presents to the cache & predictive branch controller the address of the needed instruction. If a copy of the instruction exists in the Level One Cache 612 then the instruction passes directly to the Processor Core 613 for execution.
[0144] If the instruction doesn't reside in the Level One Cache 612, but it exists in the Level Two Cache 611 that can feed the Level One Cache 612 associated with the Processor Core 613 accessing it, then it is fetched from the Level Two Cache 611. Over the direct connection bus 609b the instruction is sent directly to the Processor Core 613 over the bus 609c through the cache & predictive branch controller. A copy of the instruction is also saved in the Level One Cache 612 as well.
[0145] If a copy of the instruction doesn't exist in the Level Two Cache 611 but does in the Level Three Cache 610, a similar process occurs, only it involves all the cache 610, 611, 612 and the busses 609a, 609b, 609c that feed instructions or data to the Processor Core, and typically with an even longer delay to get the instruction to the Processor Core 613.
[0146] If the instruction does not reside in any cache 610, 611, 612, then the cache & predictive branch controller sends a command to the External Memory Interface 606 to fetch the instruction from Main Memory 101a while also informing the External Memory Interface 602 that it must decrypt the instruction before sending it to the cache & predictive branch controller by passing it through the Decryption Circuitry 301b in the External Memory Interface 602. Included in the command to the External Memory Interface will be the fetch information, such as the Processor Core 613 requesting the data and the thread within said Processor Core if such a thread exists.
The Quantity of Decryption Circuits 301b
[0147] During the decryption process, the number of decryption circuits 301b needed will vary with the speed with which data can be read from the External Memory Interface 602. Depending on the complexity of the encryption and decryption algorithms, the time needed for an encrypted instruction to be decrypted by the Decryption Circuitry 301b may exceed the arrival time of subsequent encrypted instructions from Main Memory 101a. To prevent overflowing the decryption process, multiple decryption circuits 301b shall be present. Processor ICs 601 where power consumption is at a premium, such as those used for laptop computers, will generally contain slower and fewer Main Memory 101a interfaces 602, and because the Main Memory 101a interface 602 will be slower, it will contain fewer decryption circuits 301b that a faster Main Memory 101a interface 602. In general, slower External Memory Interfaces 602 will consume less power than faster External Memory Interfaces 602.
[0148] With multiple idle decryption circuits 301b, when presented with an encrypted word, one will accept a token and begin the decryption process, and with the next data to be decrypted arriving before the now active decryption circuit 301b is completed with the existing data's decryption. Therefore, the token generated for the next encrypted word arriving will be accepted by the next decryption circuit 301b that is idle. Once a decryption circuit 301b is complete with the decryption process and has passed on the decrypted word along with the fetch completion information for said word, it can accept a token again for the next incoming word needing decryption. In lieu of tokens, a simple controller will sequence from decryption circuit 301b to decryption circuit 301b, one at a time. If the selected decryption circuit 301b finds data from the Main Memory 101a is present, it will accept the data and begin the decryption process. By the time the sequencer gets around to it again the decryption circuit 301b will have completed the decryption process and be ready to accept the next data to be decrypted.
Predictive Branching
[0149] As all instructions residing in the cache 612, and possibly 611, 610 are decrypted, the cache & predictive branch controller is able to monitor instruction requests from each Processor Core 613 over the bus interface 609c and search ahead for branch instructions. Alternately, if the instructions saved in Cache 101c remain encrypted, then the cache & predictive branch controller will be able to decrypt them for the purposes of searching for conditional branches. When it finds one, it determines all of the ways the instruction stream can go and attempts to store in the associated Level One Cache 612 instructions from any branch decision so as to minimize the waiting, otherwise known as a cache miss, that the Processor Core 613 must endure before receiving the next instruction.
Implementing the ROP and JOP Solution
[0150] As mentioned in the ROP and JOP Solution section above, there are two possible solutions this invention may use to solve the ROP and JOP vulnerabilities caused by moving the encryption and Decryption Circuitry 301 out of the Processor Core 613.
[0151] To implement the 1st solution, a bit is added to the flags 108. This additional bit controls whether the 8 & 16 bit instructions are enabled or disabled. These instructions default to being disabled after decryption instruction command 206 begins, and must be enabled by a command executed by the Processor Core 613. Once enabled, the instructions are only enabled until the next interrupt or until the process disables them again. The flags 108 with the bit controlling the 8 & 16 bit instructions being enabled or not is saved on a system stack where the different processes are kept. After saving the flags 108, the 8 & 16 bit instructions are automatically disabled again. If the interrupt needs them enabled it will do so by setting the enable bit for them. Upon completion of the interrupt the Return from Interrupt (RTI) command is executed. The flags 108 for the process are restored, including the bit that enables or disables the 8 & 16 bit opcodes. If Processor Cores 613 larger than 64 bits becomes a reality, then the same bit or a separate bit can enable or disable 32 bit instructions as well, with the default being disabled after decryption is started.
[0152] The second solution is more secure. Upon detecting an attempt to load the program counter 106a while executing a Return from Interrupt (RTI) or Return from Subroutine (RTS), the Processor Core 613 will inform the cache & predictive branch controller that it will not accept a value stored in any cache 101c, whether it is in Level One Cache 612, Level Two Cache 611, Level Three Cache 613 or even beyond if more than three levels of cache 101c exist in the processor IC 601. The cache & predictive branch controller will go to the External Memory Interface 602 and request the return address be read from Main Memory 101a and get it decrypted. After decryption this value is sent to the Processor Core 613 to be loaded into its program counter 106a, completing the RTI or RTS instruction. The ability to tell the cache & predictive branch controller to not accept a value stored in any cache 101c already exists in processors since that helps making certain delays more deterministic.
Acknowledgements
[0153] The inventor wishes to acknowledge Jack Arnold Shulman of Westfield, New Jersey for engineering contributions to the wording of the background section in paragraph [0005] as well as in claim 26, which while conceived by the inventor and his engineer was originally not as comprehensive as it could have been, and Edward Reed Brooks of Plano, Texas for his gracious support, facilitation, and administrative assistance in this effort.