Method for categorizing a scene comprising a sub-scene with machine learning
11681950 · 2023-06-20
Assignee
Inventors
Cpc classification
G06N7/01
PHYSICS
G06F18/214
PHYSICS
G06F18/217
PHYSICS
G06V10/25
PHYSICS
G06V20/52
PHYSICS
International classification
G06F18/21
PHYSICS
G06F18/214
PHYSICS
G06V10/25
PHYSICS
G06V20/52
PHYSICS
Abstract
A method for identifying a scene, comprising a computing device receiving a plurality of data points corresponding to a scene; the computing device determining one or more subsets of data points from the plurality of data points that are indicative of at least one sub-scene in said scene, said at least one sub-scene displayed on a display device that is part of said scene, wherein said at least one sub-scene does not represent said scene; the computing device categorizing said scene, disregarding said at least one sub-scene, wherein the categorizing includes interpreting said scene by a computer vision system such that said at least one sub-scene is not taken into account in the categorizing of said scene.
Claims
1. A method for categorizing a scene, comprising: a computing device receiving a plurality of data points corresponding to said scene; the computing device determining one or more subsets of data points from the plurality of data points, wherein said one or more subsets of data points are indicative of at least one sub-scene in said scene, said at least one sub-scene displayed on a display device that is part of said scene, wherein said at least one sub-scene does not represent said scene; the computing device categorizing said scene, disregarding said at least one sub-scene, wherein the categorizing includes interpreting said scene by a computer vision system such that said at least one sub-scene is not taken into account in the categorizing of said scene.
2. The method of claim 1, wherein said scene is an indoor scene.
3. The method of claim 1, wherein said scene is an outdoor scene.
4. The method of claim 1, wherein said scene comprises a series of subsequent scenes defining said scene.
5. The method of claim 1, wherein said scene comprises a traffic scene from a viewpoint inside a vehicle looking out of said vehicle.
6. A device comprising an AI system for categorizing a scene, said AI system comprising a computing device running a computer program performing: receiving a plurality of data points corresponding to said scene; determining one or more subsets of data points from the plurality of data points, wherein said one or more subsets of data points are indicative of at least one sub-scene in said scene, said at least one sub-scene displayed on a display device that is part of said scene, wherein said at least one sub-scene does not represent said scene; categorizing said scene, said computer program disregarding said at least one sub-scene, wherein the categorizing includes interpreting said scene by a computer vision system such that said at least one sub-scene is not taken into account in the categorizing of said scene.
7. A non-transitory computer readable medium having stored thereon computer program instructions that, when executed by a processor in a computing device, configure the computing device to perform: receiving a plurality of data points corresponding to a scene; determining one or more subsets of data points from the plurality of data points, wherein said one or more subsets of data points are indicative of at least one sub-scene in said scene, said at least one sub-scene displayed on a display device that is part of said scene, wherein said at least one sub-scene does not represent said scene; categorizing said scene, said computer program instructions disregarding said at least one sub-scene, wherein the categorizing includes interpreting said scene by a computer vision system such that said at least one sub-scene is not taken into account in the categorizing of said scene.
8. An AI system comprising a computing device executing the computer program instructions of claim 7.
9. An apparatus comprising the AI system of claim 8, wherein said scene comprises a representation of a surrounding of said apparatus comprising said scene, said AI system providing instructions to adjust at least one physical parameter of said apparatus based upon said categorizing of said scene.
10. The apparatus of claim 9, selected from a vehicle and a robot system.
11. A monitoring system comprising the AI system of claim 8, wherein said scene comprises a representation of a surrounding of said monitoring system, said AI system providing a signal based upon said categorizing of said scene.
12. A surveillance system comprising the monitoring system of claim 11.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
(1) Embodiments of the invention will now be described, by way of example only, with reference to the accompanying schematic drawings in which corresponding reference symbols indicate corresponding parts, and in which:
(2)
(3)
(4)
(5)
(6) The drawings are not necessarily to scale.
DESCRIPTION OF PREFERRED EMBODIMENTS
(7) The following detailed description describes various features and functions of the disclosed systems and methods with reference to the accompanying figures. In the figures, similar symbols identify similar components, unless context dictates otherwise.
(8)
(9) In
(10) In
(11) In
(12) In
(13) In another method a categorized scene 20 is deducted from one or more categorized actions (21). For example, a box match scene with various billboards can be categorized directly or can be categorized by the activity or series of actions by boxers fighting in a ring.
(14) In yet another method a categorized scene 20 is deducted from one or more categorized subjects (22). For example, a box match scene with various billboards can be categorized directly or can be categorized by a one or more subjects such as a boxing ring, boxers, trainers, crowd and various attributes in scene 10.
(15) The methods (1, 1′, 1″ and 1′″) may include one or more operations, functions, or actions as depicted in
(16) In addition, for the methods (1, 1′, 1″ and 1′″) and other processes and methods disclosed herein, the flow charts show functionality and operation of possible implementations of embodiments. In this regard, each method may represent a module, a segment, or a portion of program code, which includes one or more instructions executable by a processor for implementing specific logical functions or steps in the process. The program code may be stored on any type of computer readable medium or memory, for example, such as a storage device including a disk or hard drive. The computer readable medium may include a non-transitory computer readable medium, for example, such as computer-readable media that stores data for short periods of time like register memory, processor cache and random-access memory (RAM). The computer readable medium may also include non-transitory media or memory, such as secondary or persistent long-term storage, like read only memory (ROM), optical or magnetic disks, compact-disc read only memory (CD-ROM), for example. The computer readable media may also be any other volatile or non-volatile storage systems. The computer readable medium may be considered a computer readable storage medium, a tangible storage device, or other article of manufacture, for example.
(17) In addition, for the methods (1, 1′, 1″ and 1′″) and other processes and methods disclosed herein, computing device 3 may represent circuitry that is wired to perform the specific logical functions in the process. For the sake of example, the methods (1, 1′, 1″ and 1′″) shown in
(18)
(19) In another application computing device is categorizing, within scene 10, an action, a pose, a subject or a combination thereof.
(20)
(21) In this application, computing device 3, when categorizing the people 8 on the square, will, by disregarding the sub-scene on wide screen 2, deduct that the number of people on the square in view of camera 4 is nine. For instance, such information can be used for monitoring and controlling a crowd in an open space.
(22) Additional in this application, computing device 3, when categorizing the houses 9 on the square, will, by disregarding the sub-scene on display device 2′, deduct that the number of houses in view of camera 4 is three and by doing so, it will also increase the correct categorization for merchandise wagon 7 since computing device 3 is not misled by the display device 2′.
(23) Would this application serve as a surveillance system then the system of
(24)
(25) The billboard 2 can be a traditional poster, a digital billboard or a screen configured to display a static image, a (time) series of images, or a video movie.
(26) Further, an example system may take the form of a non-transitory computer-readable medium, which has program instructions stored thereon that are executable by at least one processor to provide the functionality described herein.
(27) An example system may take the form of any vehicle or a subsystem of any vehicle that includes such a non-transitory computer-readable medium having such program instructions stored thereon. Therefore, the terms “computing device” and “autonomous vehicle” can be interchangeable herein. However, in some examples, the computing device may be configured to control the vehicle in an autonomous or semi-autonomous operation mode.
(28) In yet another application, an embodiment is built into a robot so the robot will correctly interpreter its surrounding and the scene wherein the robot is operating.
(29) It may be readily understood that certain aspects of the disclosed systems and methods can be arranged and combined in a wide variety of different configurations, all of which are contemplated herein.
(30) It will also be clear that the above description and drawings are included to illustrate some embodiments of the invention, and not to limit the scope of protection. Starting from this disclosure, many more embodiments will be evident to a skilled person. These embodiments are within the scope of protection and the essence of this invention and are obvious combinations of prior art techniques and the disclosure of this patent.