Patent classifications
G06F7/49942
TECHNIQUES, DEVICES, AND INSTRUCTION SET ARCHITECTURE FOR EFFICIENT MODULAR DIVISION AND INVERSION
Disclosed are apparatuses, systems, and techniques to perform and facilitate fast and efficient modular computational operations, such as modular division and modular inversion, using shared platforms, including hardware accelerator engines.
FLOATING-POINT CONVERSION WITH DENORMALIZATION
A method, computer program, and computer system are provided for floating-point conversion with denormalization in a single clock cycle. An input floating-point number corresponding to an input data type is received. An exponent value and a fraction value are extracted from the received input floating-point number. A biasing constant associated with converting the received input floating-point number from the input data type to an output data type is determined. The exponent value is biased based on the biasing constant. The fraction value is converted to the output data type based on a denormalization constant associated with the extracted exponent value and the determined biasing constant. Biasing the exponent value and converting the fraction value occurs in a single clock cycle based on performing these actions in parallel. A floating-point number is output in the output data type corresponding to the converted fraction value and the biased exponent value.
Circuit and method for binary flag determination
The present disclosure relates to a circuit and method for determining a sign indicator bit of a binary datum including a step for processing of the binary datum masked with a masking operation, and not including any processing step of the binary datum.
ACCELERATION CIRCUITRY
Systems, apparatuses, and methods related to acceleration circuitry are described. The acceleration circuitry may be deployed in a memory device and can include a memory resource and/or logic circuitry. The acceleration circuitry can perform operations on data to convert the data between one or more numeric formats, such as floating-point and/or universal number (e.g., posit) formats. The acceleration circuitry can perform arithmetic and/or logical operations on the data after the data has been converted to a particular format. For instance, the memory resource can receive data comprising a bit string having a first format that provides a first level of precision. The logic circuitry can receive the data from the memory resource and convert the bit string to a second format that provides a second level of precision that is different from the first level of precision.
Approximation of samples of a digital signal reducing a number of significant bits
The invention relates to the representation of digital signals. In order to improve the perception by a user of the quality of a digital signal, a first sample of first digital signal is approximated to a second sample of a second digital signal having a second number of significant bits lower than the first number of significant bits of the first sample. The second number of significant bits is also lower than a number of significant bits allowing the second digital signal, or a signal derived therefrom to match an expected bit depth of a processing unit said second digital signal, or a signal derived therefrom is to be sent to.
Method for determining a value of an integer scaling in a linking of input sets to output sets, and computer program product
The invention relates to a method for determining a value of an integer scaling in a linking of input sets to output sets, wherein the linking comprises operators, each of which has operator inputs and operator outputs that are at least partially linked to one another or to the input sets or to the output sets, by using a computer device having a processing unit, a memory unit, and an output unit. Representations of set objects are used to efficiently carry out rescaling operations within the linking, with up to infinitely large resolution sets. This procedure makes it possible to calculate resource-conserving integer scalings for a target system while taking secondary conditions into account.
Hexadecimal exponent alignment for binary floating point unit
Examples of techniques for hexadecimal exponent alignment for a binary floating point unit (BFU) of a computer processor are described herein. An aspect includes receiving, by the BFU, a first operand comprising a first fraction and a first exponent, and a second operand comprising a second fraction and a second exponent. Another aspect includes, based on the first operand and the second operand being in a first floating point format, multiplying each of the first exponent and the second exponent by a factor corresponding to a number of bits in a digit in the first floating point format.
Semiconductor device including multiplier circuit
A semiconductor device including a multiplier circuit is provided. A first cell, a second cell, and a first circuit are included. The first cell includes a first transistor. The second cell includes a second transistor. The first circuit includes a third transistor, a fourth transistor, a fifth transistor, a sixth transistor, a first capacitor, a second capacitor, and a first switch.
Information processing device and method, and recording medium for determining a variable data type for a neural network
An information processing device, includes a memory; and a processor coupled to the memory and configured to: calculate a quantization error when a variable to be used in a neural network is quantized, generate a threshold value based on reference information related to a first recognition rate obtained by past learning of the neural network and a second recognition rate that is obtained by calculation of the neural network, determine a variable of data type to be quantized among variables to be used for calculation of the neural network based on the calculated quantization error and the generated threshold value, and execute the calculation of the neural network by using the variable of data type.
METHOD AND NON-TRANSITORY COMPUTER READABLE MEDIUM FOR COMPUTE-IN-MEMORY MACRO ARRANGEMENT, AND ELECTRONIC DEVICE APPLYING THE SAME
A method and a non-transitory computer readable medium for CIM arrangement, and an electronic device applying the same are proposed. The method for CIM arrangement includes to obtain information of the number of CIM macros and information of the dimension of each of the CIM micros, to obtain information of the number of input channels and the number of output channels of a designated convolutional layer of a designate neural network, and to determine a CIM macro arrangement for arranging the CIM macros according to the number of the CIM macros, the dimension of each of the CIM macros, the number of the input channels and the number of the output channels of the designated convolutional layer of the designated neural network, for applying convolution operation to the input channels to generate the output channels.