Patent classifications
G06T2207/10016
Device, method and system for estimating elevation in images from camera devices
A device, method and system for estimating elevation in images from camera devices is provided. The device detects humans at respective positions in images from a camera device, the camera device having a fixed orientation and fixed focal length. The device estimates, for the humans in the images, respective elevations of the humans, relative to the camera device, at the respective positions based at least on camera device parameters defining the fixed orientation and the fixed focal length. The device associates the respective elevations with the respective positions in the images. The device determines, using the respective elevations associated with the respective positions, a function that estimates elevation in an image from the camera device, using a respective image position coordinate as an input. The device provides the function to a video analytics engine to determine relative real-world positions in subsequent images from the camera device.
Depth estimation using biometric data
Method of generating depth estimate based on biometric data starts with server receiving positioning data from first device associated with first user. First device generates positioning data based on analysis of a data stream comprising images of second user that is associated with second device. Server then receives a biometric data of second user from second device. Biometric data is based on output from a sensor or a camera included in second device. Server then determines a distance of second user from first device using positioning data and biometric data of the second user. Other embodiments are described herein.
Apparatus and method for displaying contents on an augmented reality device
A system for displaying contents on an augmented reality (AR) device comprises a capturing module configured to capture a field of view of a user, a recording module configured to record the captured field of view, a user input controller configured to track a vision of the user towards one or more objects and a server. The server comprises a determination module, an identifier, and an analyser. The determination module is configured to determine at least one object of interest. The identifier is configured to identify a frame containing disappearance of the determined object of interest. The analyser is configured to analyse the identified frame based on at least one disappearance of the object of interest, and generate analysed data. The display module is configured to display a content of the object of interest on the AR device.
System and method for visually tracking persons and imputing demographic and sentiment data
A visual tracking system for tracking and identifying persons within a monitored location, comprising a plurality of cameras and a visual processing unit, each camera produces a sequence of video frames depicting one or more of the persons, the visual processing unit is adapted to maintain a coherent track identity for each person across the plurality of cameras using a combination of motion data and visual featurization data, and further determine demographic data and sentiment data using the visual featurization data, the visual tracking system further having a recommendation module adapted to identify a customer need for each person using the sentiment data of the person in addition to context data, and generate an action recommendation for addressing the customer need, the visual tracking system is operably connected to a customer-oriented device configured to perform a customer-oriented action in accordance with the action recommendation.
Automatic correction method for onboard camera and onboard camera device
There is provided an automatic correction method for an onboard camera and an onboard camera device. The automatic correction method includes the following steps: obtaining a lane image with the onboard camera and a current extrinsic parameter matrix, and identifying two lane lines in the lane image; converting the lane image into a top-view lane image, and obtaining two projected lane lines in the top-view lane image for the two lane lines; calculating a plurality of correction parameter matrices corresponding to the current extrinsic parameter matrix according to the two projected lane lines; and correcting the current extrinsic parameter matrix according to the plurality of correction parameter matrices. This can be applied in situations where the vehicle is stationary or travelling for automatic correction on the extrinsic parameter matrix of the onboard camera.
Video analysis for obtaining optical properties of a face
Disclosed is a system and method for obtaining optical properties of skin on a human face through face video analysis. Video of the face is captured, landmarks on the face and tracked, regions-of-interest are defined and tracked using the landmarks, some measurements/optical properties are obtained, the time-based video is transformed into an angular domain, and additional measurements/optical properties are obtained. Such optical properties can be measured using video in real-time or video that has been pre-recorded.
Method and apparatus for detecting abnormal objects in video
Disclosed are a method and an apparatus for detecting abnormal objects in a video. The method for detecting abnormal objects in a video reconstructs a restored batch by applying each input batch to which an inpainting pattern is applied to a trained auto-encoder model, and fuses a time domain reconstruction error using time domain restored frames output by extracting and restoring a time domain feature point by applying a spatial domain reconstruction error and a plurality of successive frames using a restored frame output by combining the reconstructed restoring batch to a trained LSTM auto-encoder model to estimate an area where an abnormal object is positioned.
Automated honeypot creation within a network
Systems and methods for managing Application Programming Interfaces (APIs) are disclosed. Systems may involve automatically generating a honeypot. For example, the system may include one or more memory units storing instructions and one or more processors configured to execute the instructions to perform operations. The operations may include receiving, from a client device, a call to an API node and classifying the call as unauthorized. The operation may include sending the call to a node-imitating model associated with the API node and receiving, from the node-imitating model, synthetic node output data. The operations may include sending a notification based on the synthetic node output data to the client device.
Virtual teach and repeat mobile manipulation system
A method for controlling a robotic device is presented. The method includes positioning the robotic device within a task environment. The method also includes mapping descriptors of a task image of a scene in the task environment to a teaching image of a teaching environment. The method further includes defining a relative transform between the task image and the teaching image based on the mapping. Furthermore, the method includes updating parameters of a set of parameterized behaviors based on the relative transform to perform a task corresponding to the teaching image.
Mobile robot system and method for generating map data using straight lines extracted from visual images
A mobile robot is configured to navigate on a sidewalk and deliver a delivery to a predetermined location. The robot has a body and an enclosed space within the body for storing the delivery during transit. At least two cameras are mounted on the robot body and are adapted to take visual images of an operating area. A processing component is adapted to extract straight lines from the visual images taken by the cameras and generate map data based at least partially on the images. A communication component is adapted to send and receive image and/or map data. A mapping system includes at least two such mobile robots, with the communication component of each robot adapted to send and receive image data and/or map data to the other robot. A method involves operating such a mobile robot in an area of interest in which deliveries are to be made.