Classification of Image Data with Adjustment of the Degree of Granulation
20230230335 · 2023-07-20
Inventors
Cpc classification
G06V10/454
PHYSICS
G06V20/58
PHYSICS
G06V20/56
PHYSICS
G06V40/103
PHYSICS
G06V10/774
PHYSICS
G06N5/01
PHYSICS
International classification
G06V20/56
PHYSICS
Abstract
A device for classifying image data includes a trainable pre-processing unit configured to retrieve, from a trained context, and based on the image data, at least one specification in terms of how a degree of granulation of the image data is to be reduced, and to reduce the degree of granulation of the image data in accordance with the at least one specification. The device further includes a trainable classifier configured to map the granulation-reduced image data onto an assignment to one or more classes of a specified classification.
Claims
1. A device for classification of image data, comprising: a trainable pre-processing unit configured to retrieve, from a trained relationship, and based on the image data, at least one specification regarding an extent to which a level of detail of the image data is to be reduced, and to reduce the level of detail of the image data in accordance with the at least one specification; and a trainable classifier configured to map the detail-reduced image data onto an assignment to one or more classes of a specified classification.
2. The device according to claim 1, wherein the pre-processing unit and the classifier are configured designed as a common artificial neural network (“ANN”).
3. The device according to claim 1, wherein: the pre-processing unit is connected via a dedicated broadband connection to at least one image source carried by a vehicle, the pre-processing unit is connected to the classifier via a bus system of the vehicle, and the bus system is also used by further on-board systems of the vehicle.
4. The device according to claim 1, wherein the pre-processing unit is configured to: transform the image data into a representation in a working space, and reduce the level of detail of the this representation.
5. The device according to claim 4, wherein the pre-processing unit is configured to: determine the representation in the working space as a linear combination of basic functions of the working space characterized by a set of coefficients, retrieve the at least one specification, based on the image data, from the trained relationship regarding which of the coefficients are to be reduced in terms of magnitude or eliminated, and reduce or eliminate the coefficients of the representation according to the at least one specification.
6. The device according to claim 5, wherein: the pre-processing unit is configured to determine the representation in the working space as a linear combination of wavelets; and the wavelets are characterized by a set of wavelet coefficients.
7. The device according to claim 5, wherein the pre-processing unit is configured to retrieve, as the at least one specification, a numerical fraction of the coefficients which are to be reduced in terms of magnitude or eliminated.
8. The device according to claim 7, wherein the pre-processing unit is configured to determine the numerical fraction of the coefficients which are to be reduced in terms of magnitude or eliminated, using entropy of the image data.
9. The device according to claim 4, wherein the pre-processing unit is configured to: retrieve a dimensionality for a latent space of an autoencoder as a working space using the image data from the trained relationship, and transform the image data into the working space using the autoencoder.
10. A method for training a device, comprising: providing learning image data and associated learning assignments onto which the device is to nominally map the learning image data; defining a specification for reducing a level of detail that is sought on average; and optimizing parameters characterizing a behavior of the trainable pre-processing unit of the device to aims that the device maps the learning image data onto the learning assignments and, at the same time, the reduction in the level of detail which the pre-processing unit performs on the learning image data corresponds, on average, to the specification.
11. The method according to claim 10, wherein parameters characterizing a behavior of the classifier of the device are additionally optimized to an aim that the device maps the learning image data onto the learning assignments.
12. The method according to claim 11, wherein the optimization of the parameters which characterize the behavior of the classifier is additionally also directed to an aim that the level of detail of the image data used by the classifier be as low as possible.
13. The method according to claim 10, wherein a computer program contains machine-readable instructions which, when executed on one or more computers, upgrade the computer or computers to the device and/or cause the method to be carried out.
14. A non-transitory machine-readable data carrier comprising the computer program according to claim 13.
15. A computer comprising the computer program according to claim 13.
Description
EMBODIMENTS
[0045] In the drawings:
[0046]
[0047]
[0048]
[0049]
[0050]
[0051] The pre-processing unit 11 receives image data 2 from any source and in a block 111 determines on the basis of these image data 2 at least one specification 3 regarding the extent to which the level of detail of the image data 2 is to be reduced. In block 112, the level of detail of the image data 2 is reduced according to this specification 3, such that detail-reduced image data 4 result. These detail-reduced image data 4 are mapped onto an assignment 5 to one or more classes of a specified classification.
[0052]
[0053]
[0054] All these details are not relevant to the important recognition that this is a pedestrian. Certain details could even distract a classifier 12. Thus, for example, certain features in the face 71, or a tattered state of the t-shirt 72, could cause the classifier 12 to incorrectly classify the pedestrian 7 as a scarecrow. Likewise, the shoulder pads 74 could cause the classifier 12 to classify the pedestrian 7 as a display dummy. Both cases would be disadvantageous for the pedestrian 7, because a system for at least partially automated driving would assume that only slight material damage occurs in the event of a collision with a scarecrow or a display dummy, and in the case of doubt would give preference to this collision over a collision with another vehicle. The same could happen if the incorrect classification is deliberately brought about by a manipulative “adversarial” pattern 75 in the image data 2, for example by a semi-transparent sticker on the camera lens.
[0055]
[0056]
[0059] In addition, in step 140 in this embodiment, parameters 12* which characterize the behavior of the classifier 12 of the device 1 are also optimized for the aim of the device 1 mapping the learning image data 2a onto the learning assignments 5a. This training is dovetailed with the training 130 of the parameters 11* of the pre-processing unit 11, 11a-11d.
[0060] According to block 141, optimization 140 of the parameters 12* of the classifier 12 is additionally also directed at the aim of the level of detail of the image data 2 used by the classifier 12 being as low as possible.