CONFIGURING A NEURAL NETWORK FOR EQUIVARIANT OR INVARIANT BEHAVIOR
20230050283 · 2023-02-16
Inventors
- Elise van der Pol (Amsterdam, NL)
- Frans A. Oliehoek (Delft, NL)
- Herke van Hoof (Diemen, NL)
- Max Welling (Bussum, NL)
- Michael Herman (Sindelfingen, DE)
Cpc classification
G06N7/01
PHYSICS
G06N3/006
PHYSICS
B25J9/161
PERFORMING OPERATIONS; TRANSPORTING
International classification
Abstract
A method for configuring a neural network which is designed to map measured data to one or more output variables. The method includes: transformation(s) of the measured data is/are specified which when applied to the measured data, is/are meant to induce the output variables supplied by the neural network to exhibit an invariant or equivariant behavior; at least one equation is set up which links a condition that the desired invariance or equivariance be given with the architecture of the neural network; by solving the at least one equation a feature is obtained that characterizes the desired architecture and/or a distribution of weights of the neural network in at least one location of this architecture; a neural network is configured in such a way that its architecture and/or its distribution of weights in at least one location of this architecture has/have all of the features ascertained in this way.
Claims
1. A method for configuring a neural network which is configured to map measured data to one or more output variables, the method comprising the following steps: specifying one or more transformations of the measured data which, when applied to the measured data, is meant to induce output variables supplied by the neural network to exhibit a desired invariant or equivariant behavior; setting up at least one equation which links a condition that the desired invariance or equivariance be given with an architecture of the neural network; obtaining, by solving the at least one equation, at least one feature that characterizes the architecture and/or a distribution of weights of the neural network in at least one location of the architecture; and configuring the neural network in such a way that its architecture and/or the distribution of weights in at least one location of the architecture, has all of the ascertained at least one feature.
2. The method as recited in claim 1, wherein the at least one equation includes: a function ϕ.sub.u, which describes a further development of features of layers of the neural network during a transition from one layer to the next, and/or a function ϕ.sub.m, which describes an information flow within the neural network as a function of the architecture of the neural network.
3. The method as recited in claim 2, wherein the neural network is a graph in which nodes occupied by features h.sub.i.sup.l are connected by edges e.sub.ij.
4. The method as recited in claim 3, wherein the function ϕ.sub.u links features h.sub.i.sup.l+1 of the i.sup.th node in a layer l+1 with features h.sub.i.sup.l of the i.sup.th node in a layer I and with the information flow m.sub.i.sup.l received in total by the node.
5. The method as recited in claim 4, wherein the function ϕ.sub.m links the information flow m.sub.j.fwdarw.i.sup.l from node j to node i in a layer l with an edge e.sub.ij between nodes I and j and also with features h.sub.j.sup.l of the j.sup.th node in the layer I.
6. The method as recited in claim 1, wherein, in the specifying step, at least one group of transformations is specified for which the desired invariance or equivariance of the output variables is to apply.
7. The method as recited in claim 1, wherein the at least one equation is expressed in hyperparameters which characterize the architecture of the neural network, and the solving of the at least one equation leads to values of the hyperparameters as features.
8. The method as recited in claim 1, wherein: observations of multiple agents of a centralized or decentralized Markov decision process are the measured data, and a reward to be expected when a predefined action is performed in a certain state of the Markov decision process and/or a policy for at least one agent mapping a predefined state to an action to be performed is selected as an output variable of the output variables.
9. The method as recited in claim 8, wherein the observations include positions of agents.
10. The method as recited in claim 8, further comprising: ascertaining, from the reward to be expected, a control signal for at least one robot and/or for at least one vehicle and/or for at least one unmanned flying device and/or from the policy; and controlling, using the control signal, the robot and/or the vehicle and/or the unmanned flying device (.
11. The method as recited in claim 1, wherein the neural network to be configured is a classifier network, which maps the measured data to classification scores with regard to one or more classes of a predefined classification.
12. The method as recited in claim 11, further comprising: ascertaining, from the classification scores, a control signal for at least one robot and/or for at least one vehicle and/or for at least one unmanned flying device; and controlling, using the control signal, the robot and/or the vehicle and/or the unmanned flying device.
13. A non-transitory machine-readable data carrier on which is stored a computer program for configuring a neural network which is configured to map measured data to one or more output variables, the computer program, when executed by one or more computers, causing the one or more computers to perform the following steps: specifying one or more transformations of the measured data which, when applied to the measured data, is meant to induce output variables supplied by the neural network to exhibit a desired invariant or equivariant behavior; setting up at least one equation which links a condition that the desired invariance or equivariance be given with an architecture of the neural network; obtaining, by solving the at least one equation, at least one feature that characterizes the architecture and/or a distribution of weights of the neural network in at least one location of the architecture; and configuring the neural network in such a way that its architecture and/or the distribution of weights in at least one location of the architecture, has all of the ascertained at least one feature.
14. One or more computers configured to configure a neural network which is configured to map measured data to one or more output variables, the one or more computers being configured to: specify one or more transformations of the measured data which, when applied to the measured data, is meant to induce output variables supplied by the neural network to exhibit a desired invariant or equivariant behavior; set up at least one equation which links a condition that the desired invariance or equivariance be given with an architecture of the neural network; obtain, by solving the at least one equation, at least one feature that characterizes the architecture and/or a distribution of weights of the neural network in at least one location of the architecture; and configure the neural network in such a way that its architecture and/or the distribution of weights in at least one location of the architecture, has all of the ascertained at least one feature.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
[0050]
[0051]
[0052]
DETAILED DESCRIPTION OF EXAMPLE EMBODIMENTS
[0053]
[0054] In step 105, observations of multiple agents of a centralized or decentralized Markov decision process are able to be selected as measured data 2. A reward to be expected when carrying out a predefined action in a certain state of the Markov decision process and/or a policy for at least one agent, which maps a predefined state to an action to be performed, may then be selected as output variable 3 of neural network 1 in step 6.
[0055] In step 107, a classifier network which maps measured data 2 to classification scores with regard to one or more classes(s) of a predefined classification can then be selected as neural network 1 to be configured.
[0056] In step 110, one or more transformation(s) 2a of measured data 2 is/are specified, which when applied to measured data 2, are to induce output variables 3 supplied by neural network 1 to exhibit an invariant or equivariant response.
[0057] According to block 111, at least one group G of transformations 2a to which the desired invariance or equivariance of the output variables 3 is meant to apply may be specified in this context.
[0058] In step 120, at least one equation 4 is set up, which links a condition to the effect that the desired invariance or equivariance be given, with architecture la of neural network 1.
[0059] According to block 121, this at least one equation 4 may particularly include a function ϕ.sub.u, for example, which describes a further development of the features of layers of neural network 1 during the transition from one layer to the next.
[0060] Alternatively or in combination therewith, the at least one equation 4 according to block 122 may particularly include a function ϕ.sub.m which describes an information flow within neural network 1 as a function of architecture la of neural network 1, for example.
[0061] According to block 121a, if the neural network is developed as a graph in which nodes occupied by features h.sub.i.sup.l are linked by edges e.sub.ij, function ϕ.sub.u is able to link in particular features h.sub.i.sup.l+1 of the i.sup.th node in layer l+1 with features h.sub.i.sup.l of the i.sup.th node in layer I and also with the information flow m.sub.i.sup.l received in total by this node.
[0062] According to block 122a, alternatively or also in combination therewith, the function ϕ.sub.m is able to link information flow m.sub.j.fwdarw.i.sup.l from node j to node i in layer I with an edge e.sub.ij between nodes i and j and also with features h.sub.j.sup.l of the j.sup.th node in layer I.
[0063] According to block 123, the at least one equation 4 is able to be expressed in hyperparameters which characterize architecture 1a of neural network 1. For instance, the hyperparameters may include the number, size and types of layers of neural network 1 and/or of neurons or other processing units that make up these layers.
[0064] In step 130, at least one feature 5 is obtained by solving the at least one equation 4, which characterizes the desired architecture 1a and/or a distribution of weights of neural network 1 in at least one location in this architecture 1a.
[0065] If equation 4 depends on hyperparameters, then the solving of the at least one equation 4 may lead to values of the hyperparameters as features 5 according to block 131.
[0066] In step 140, at least one neural network 1 is configured in such a way that its architecture la and/or its distribution of weights has/have all of the features 5 ascertained in step 130 in at least one location in this architecture.
[0067] In step 150, following intermediate training of the configured neural network 1 and the populating of this network 1 with measured data 2, for example, a control signal 6 for at least one robot 50 and/or for at least one vehicle 60 and/or for at least one unmanned flying vehicle 11-13 is able to be ascertained as output variable 3 from a reward to be expected and/or from a policy within the framework of a Markov decision process.
[0068] In step 160, for instance following intermediate training of configured neural network 1 and the populating of this network 1 with measured data 2, it is likewise possible to ascertain a control signal 6 as output variables 3 for at least one robot 50 and/or for at least one vehicle 60 and/or for at least one unmanned flying device 11-13 from classification scores for measured data 2 ascertained by the neural network 1.
[0069] Regardless of the source from which control signal 6 is obtained, robot 50 and/or vehicle 60 and/or unmanned flying device 11-13 is/are able to be controlled by this control signal 6 in step 170.
[0070] In
[0071] Since the cameras are perpendicularly pointing down, drones 11-13 are unable to see one another. They can transmit merely their respective current observations from respective acquisition range 11a-13a. If the Markov decision process is controlled in a centralized manner, then these observations may conveyed to a central instance which coordinates the deployment of drones 11-13. If the Markov decision process is controlled in a decentralized manner, the observations of a drone 11-13 are able to be transmitted to other drones 11-13 in a limited physical environment.
[0072]
[0073]