Patent classifications
G06N3/06
Hardware architecture for a neural network accelerator
Examples herein describe hardware architecture for processing and accelerating data passing through layers of a neural network. In one embodiment, a reconfigurable integrated circuit (IC) for use with a neural network includes a digital processing engine (DPE) array, each DPE having a plurality of neural network units (NNUs). Each DPE generates different output data based on the currently processing layer of the neural network, with the NNUs parallel processing different input data sets. The reconfigurable IC also includes a plurality of ping-pong buffers designed to alternate storing and processing data for the layers of the neural network.
Hardware architecture for a neural network accelerator
Examples herein describe hardware architecture for processing and accelerating data passing through layers of a neural network. In one embodiment, a reconfigurable integrated circuit (IC) for use with a neural network includes a digital processing engine (DPE) array, each DPE having a plurality of neural network units (NNUs). Each DPE generates different output data based on the currently processing layer of the neural network, with the NNUs parallel processing different input data sets. The reconfigurable IC also includes a plurality of ping-pong buffers designed to alternate storing and processing data for the layers of the neural network.
Redundant memory access for rows or columns containing faulty memory cells in analog neural memory in deep learning artificial neural network
Numerous embodiments are disclosed for accessing redundant non-volatile memory cells in place of one or more rows or columns containing one or more faulty non-volatile memory cells during a program, erase, read, or neural read operation in an analog neural memory system used in a deep learning artificial neural network.
PARALLEL COMPUTING SCHEME GENERATION FOR NEURAL NETWORKS
A device receives a computation graph and transforms the computation graph into a dataflow graph comprising recursive subgraphs. Each recursive subgraph comprises a tuple of another recursive subgraph and an operator node, or an empty graph. The device determines a number of partitioning recursions based on a number of parallel computing devices. For each partitioning recursion, the device determines costs corresponding to operator nodes, determines a processing order of the recursive subgraphs, and processes the recursive subgraphs. To process a recursive subgraph, the device selects a partitioning axis for tensors associated with an operator node of the recursive subgraph. The device outputs a partitioning scheme comprising partitioning axes for each tensor associated with the operator nodes.
PARALLEL COMPUTING SCHEME GENERATION FOR NEURAL NETWORKS
A device receives a computation graph and transforms the computation graph into a dataflow graph comprising recursive subgraphs. Each recursive subgraph comprises a tuple of another recursive subgraph and an operator node, or an empty graph. The device determines a number of partitioning recursions based on a number of parallel computing devices. For each partitioning recursion, the device determines costs corresponding to operator nodes, determines a processing order of the recursive subgraphs, and processes the recursive subgraphs. To process a recursive subgraph, the device selects a partitioning axis for tensors associated with an operator node of the recursive subgraph. The device outputs a partitioning scheme comprising partitioning axes for each tensor associated with the operator nodes.
OSCILLATION CIRCUIT AND INFORMATION PROCESSING DEVICE
An oscillation circuit includes a first oscillation circuit that includes: a first diode that has a first negative differential resistance; a first composite inductor in which a first inductor and a second inductor are connected in series, is connected to the first diode in series; a second diode that has a second negative differential resistance and is connected to the first inductor in parallel; and a third diode that has a third negative differential resistance, is connected to the first diode in series, and is connected to the first composite inductor in parallel, wherein a burst pulse is output from a common connection point of the first inductor, the second inductor, and the second diode.
OSCILLATION CIRCUIT AND INFORMATION PROCESSING DEVICE
An oscillation circuit includes a first oscillation circuit that includes: a first diode that has a first negative differential resistance; a first composite inductor in which a first inductor and a second inductor are connected in series, is connected to the first diode in series; a second diode that has a second negative differential resistance and is connected to the first inductor in parallel; and a third diode that has a third negative differential resistance, is connected to the first diode in series, and is connected to the first composite inductor in parallel, wherein a burst pulse is output from a common connection point of the first inductor, the second inductor, and the second diode.
METHODS OF CHEMICAL COMPUTATION
The invention provides methods for computing with chemicals by encoding digital data into a plurality of chemicals to obtain a dataset; translating the dataset into a chemical form; reading the data set; querying the dataset by performing an operation to obtain a perceptron; and analyzing the perceptron for identifying chemical structure and/or concentration of at least one of the chemicals, thereby developing a chemical computational language. The invention demonstrates a workflow for representing abstract data in synthetic metabolomes. Also presented are several demonstrations of kilobyte-scale image data sets stored in synthetic metabolomes, recovered at >99% accuracy.
Systems and Methods for Generating Motion Forecast Data for Actors with Respect to an Autonomous Vehicle and Training a Machine Learned Model for the Same
Systems and methods for generating motion forecast data for actors with respect to an autonomous vehicle and training a machine learned model for the same are disclosed. The computing system can include an object detection model and a graph neural network including a plurality of nodes and a plurality of edges. The computing system can be configured to input sensor data into the object detection model; receive object detection data describing the location of the plurality of the actors relative to the autonomous vehicle as an output of the object detection model; input the object detection data into the graph neural network; iteratively update a plurality of node states respectively associated with the plurality of nodes; and receive, as an output of the graph neural network, the motion forecast data with respect to the plurality of actors.
Systems and Methods for Generating Motion Forecast Data for Actors with Respect to an Autonomous Vehicle and Training a Machine Learned Model for the Same
Systems and methods for generating motion forecast data for actors with respect to an autonomous vehicle and training a machine learned model for the same are disclosed. The computing system can include an object detection model and a graph neural network including a plurality of nodes and a plurality of edges. The computing system can be configured to input sensor data into the object detection model; receive object detection data describing the location of the plurality of the actors relative to the autonomous vehicle as an output of the object detection model; input the object detection data into the graph neural network; iteratively update a plurality of node states respectively associated with the plurality of nodes; and receive, as an output of the graph neural network, the motion forecast data with respect to the plurality of actors.