Patent classifications
G06T7/292
MULTI-VIEW MULTI-TARGET ACTION RECOGNITION
Implementations generally perform robust multi-view multi-target action recognition using reconstructed 3-dimensional (3D) poses. In some implementations, a method includes obtaining a plurality of videos of a plurality of subjects in an environment, where at least one target subject of the plurality of subjects performs one or more actions in the environment. The method further includes tracking the at least one target subject across at least two cameras. The method further includes reconstructing a 3-dimensional (3D) model of the at least one target subject based on the plurality of videos and the tracking of the at least one target subject. The method further includes recognizing the one or more actions of the at least one target subject based on the reconstructing of the 3D model.
MULTI-VIEW MULTI-TARGET ACTION RECOGNITION
Implementations generally perform robust multi-view multi-target action recognition using reconstructed 3-dimensional (3D) poses. In some implementations, a method includes obtaining a plurality of videos of a plurality of subjects in an environment, where at least one target subject of the plurality of subjects performs one or more actions in the environment. The method further includes tracking the at least one target subject across at least two cameras. The method further includes reconstructing a 3-dimensional (3D) model of the at least one target subject based on the plurality of videos and the tracking of the at least one target subject. The method further includes recognizing the one or more actions of the at least one target subject based on the reconstructing of the 3D model.
VEHICULAR ACCESS CONTROL BASED ON VIRTUAL INDUCTIVE LOOP
Methods, systems, and apparatus, including computer programs encoded on computer storage media, for monitoring events using a Virtual Inductive Loop system. In some implementations, image data is obtained from cameras. A region depicted in the obtained image data is identified, the region comprising lines spaced by a distance that satisfies a distance threshold. For each line included in the region: an object depicted crossing the line is determined whether to satisfy a height criteria indicating that the line is activated. In response to determining that an object depicted crossing the line satisfies the height criteria, an event is determined to have likely occurred using data indicating (i) which lines of the lines were activated and (ii) an order in which each of the lines were activated. In response to determining that an event likely occurred, actions are performed using at least some of the data.
System and method for visually tracking persons and imputing demographic and sentiment data
A visual tracking system for tracking and identifying persons within a monitored location, comprising a plurality of cameras and a visual processing unit, each camera produces a sequence of video frames depicting one or more of the persons, the visual processing unit is adapted to maintain a coherent track identity for each person across the plurality of cameras using a combination of motion data and visual featurization data, and further determine demographic data and sentiment data using the visual featurization data, the visual tracking system further having a recommendation module adapted to identify a customer need for each person using the sentiment data of the person in addition to context data, and generate an action recommendation for addressing the customer need, the visual tracking system is operably connected to a customer-oriented device configured to perform a customer-oriented action in accordance with the action recommendation.
System and method for visually tracking persons and imputing demographic and sentiment data
A visual tracking system for tracking and identifying persons within a monitored location, comprising a plurality of cameras and a visual processing unit, each camera produces a sequence of video frames depicting one or more of the persons, the visual processing unit is adapted to maintain a coherent track identity for each person across the plurality of cameras using a combination of motion data and visual featurization data, and further determine demographic data and sentiment data using the visual featurization data, the visual tracking system further having a recommendation module adapted to identify a customer need for each person using the sentiment data of the person in addition to context data, and generate an action recommendation for addressing the customer need, the visual tracking system is operably connected to a customer-oriented device configured to perform a customer-oriented action in accordance with the action recommendation.
Multi-spatial scale analytics
Systems, methods, and computer-readable for multi-spatial scale object detection include generating one or more object trackers for tracking at least one object detected from on one or more images. One or more blobs are generated for the at least one object based on tracking motion associated with the at least one object. One or more tracklets are generated for the at least one object based on associating the one or more object trackers and the one or more blobs, the one or more tracklets including one or more scales of object tracking data for the at least one object. One or more uncertainty metrics are generated using the one or more object trackers and an embedding of the one or more tracklets. A training module for detecting and tracking the at least one object using the embedding and the one or more uncertainty metrics is generated using deep learning techniques.
Multi-spatial scale analytics
Systems, methods, and computer-readable for multi-spatial scale object detection include generating one or more object trackers for tracking at least one object detected from on one or more images. One or more blobs are generated for the at least one object based on tracking motion associated with the at least one object. One or more tracklets are generated for the at least one object based on associating the one or more object trackers and the one or more blobs, the one or more tracklets including one or more scales of object tracking data for the at least one object. One or more uncertainty metrics are generated using the one or more object trackers and an embedding of the one or more tracklets. A training module for detecting and tracking the at least one object using the embedding and the one or more uncertainty metrics is generated using deep learning techniques.
TEMPORAL CODING OF MARKERS FOR OBJECT TRACKING
There is provided a method of motion tracking comprising arranging one or more active marker devices on an object, the active marker devices being configured to emit light and each having an associated temporally repeating pattern comprising a plurality of time frames, controlling the one or more active marker devices to emit light according to their respective temporally repeating patterns, wherein the temporally repeating patterns are such that the active marker device does not emit light during at least one time frame of the plurality of time frames, detecting light emitted by the one or more active marker devices using one or more cameras, and determining a spatial configuration of the object using the light detected by the one or more cameras.
System and method to simultaneously track multiple organisms at high resolution
A microscopy includes multiple cameras working together to capture image data of a sample having a group of organisms distributed over a wide area, under the influence of an excitation instrument. A first processor is coupled to each camera to process the image data captured by the camera. Outputs from the multiple first processors are aggregated and streamed serially to a second processor for tracking the organisms. The presence of the multiple cameras capturing images from the sample, configured with 50% or more overlap, can allow 3D tracking of the organisms through photogrammetry.
System and method to simultaneously track multiple organisms at high resolution
A microscopy includes multiple cameras working together to capture image data of a sample having a group of organisms distributed over a wide area, under the influence of an excitation instrument. A first processor is coupled to each camera to process the image data captured by the camera. Outputs from the multiple first processors are aggregated and streamed serially to a second processor for tracking the organisms. The presence of the multiple cameras capturing images from the sample, configured with 50% or more overlap, can allow 3D tracking of the organisms through photogrammetry.