Patent classifications
G06V10/806
SYSTEM AND METHOD FOR ONLINE SHOPPING BASED ON FACIAL EMOTIONAL STATE ANALYSIS
An online shopping system based on facial emotional state analysis and a method thereof is provided. The system includes: an online shopping module configured for providing an online shopping interactive interface for a user and collecting facial image data and interactive behavior data of the user in an online shopping process; a facial expression recognition module configured for recognizing an emotional state of the user according to the collected facial image data in the online shopping process of the user; a shopping intention analysis module configured for deciding a shopping intention of the user according to the recognized emotional state and the interactive behavior data of the user; and a shopping recommendation adjustment module configured for dynamically adjusting a commodity recommendation strategy for the user according to the decided shopping intention of the user.
SYSTEM AND METHODS FOR IMPLEMENTING PRIVATE IDENTITY
In various embodiments, a fully encrypted private identity based on biometric and/or behavior information can be used to securely identify any user efficiently. According to various aspects, once identification is secure and computationally efficient, the secure identity/identifier can be used across any number of devices to identify a user an enable functionality on any device based on the underlying identity, and even switch between identified users seamlessly all with little overhead. In some embodiments, devices can be configured to operate with function sets that transition seamlessly between the identified users, even, for example, as they pass a single mobile device back and forth. According to some embodiments, identification can extend beyond the current user of any device, into identification of actors responsible for activity/content on the device.
SYSTEM AND METHODS FOR IMPLEMENTING PRIVATE IDENTITY
In various embodiments, a fully encrypted private identity based on biometric and/or behavior information can be used to securely identify any user efficiently. According to various aspects, once identification is secure and computationally efficient, the secure identity/identifier can be used across any number of devices to identify a user an enable functionality on any device based on the underlying identity, and even switch between identified users seamlessly all with little overhead. In some embodiments, devices can be configured to operate with function sets that transition seamlessly between the identified users, even, for example, as they pass a single mobile device back and forth. According to some embodiments, identification can extend beyond the current user of any device, into identification of actors responsible for activity/content on the device.
SYSTEM AND METHODS FOR IMPLEMENTING PRIVATE IDENTITY
In various embodiments, a fully encrypted private identity based on biometric and/or behavior information can be used to securely identify any user efficiently. According to various aspects, once identification is secure and computationally efficient, the secure identity/identifier can be used across any number of devices to identify a user an enable functionality on any device based on the underlying identity, and even switch between identified users seamlessly all with little overhead. In some embodiments, devices can be configured to operate with function sets that transition seamlessly between the identified users, even, for example, as they pass a single mobile device back and forth. According to some embodiments, identification can extend beyond the current user of any device, into identification of actors responsible for activity/content on the device.
CAMERA SYSTEM, MOBILE TERMINAL, AND THREE-DIMENSIONAL IMAGE ACQUISITION METHOD
A camera system, a mobile terminal, and a three-dimensional image acquisition method are disclosed. The camera system may include, a first photographing device, a second photographing device, a photographing assistance device, and a processor; the photographing assistance device is configured to emit a first feature light to an object; the first photographing device is configured to collect a second feature light reflected by the object; the second photographing device includes a main camera configured to collect a first image of the object and a secondary camera configured to collect a second image of the object; and the processor is configured to acquire depth information of the object according to the second feature light, and perform feature fusion on the first and second images, and perform stereo registration on a result of feature fusion and the depth information, to acquire a 3D image of the object.
PROGRESSIVE LOCALIZATION METHOD FOR TEXT-TO-VIDEO CLIP LOCALIZATION
A progressive localization method for text-to-video clip localization. The method comprises: first, respectively extracting features of two modes, namely a video mode and a text mode by using different feature extraction methods; then progressively selecting different step sizes, and learning the correlation between the video and the text in multiple stages; and finally, training a model in an end-to-end manner based on the correlation loss of each stage. Moreover, the fine time granularity stage is fused with information of the coarse time granularity stage by means of a condition feature update module and up-sampling connection, such that different stages are mutually promoted. Different stages can pay attention to clips with different time granularities, and the model can cope with the situation that the length of a target clip is obviously changed based on the interrelation between the stages.
Multi-modal Segmentation Network for Enhanced Semantic Labeling in Mapping
Provided are methods for enhanced semantic labeling in mapping with a semantic labeling system, which can include receiving, from a LiDAR sensor of a vehicle, LiDAR point cloud information including at least one raw point feature for a point, receiving, from a camera of the vehicle, image data associated with an image captured using the camera, generating at least one rich point feature for the point based on the image data, predicting, using a LiDAR segmentation neural network and based on the at least one raw point feature and the at least one rich point feature, a point-level semantic label for the point, and providing the point-level semantic label to a mapping engine to generate a map based on the point-level semantic label Systems and computer program products are also provided.
Inpainting via an encoding and decoding network
An image processing apparatus including: at least one memory; and at least one processor coupled to the at least one memory and configured to implement: an image acquisition module configured to acquire an input image including an object region; a mask image generation module configured to generate a mask image based on the input image; and an image inpainting module configured to extract a fusion feature map corresponding to the input image using an encoding network according to the input image and the mask image, and to inpaint the object region in the input image using a decoding network based on the fusion feature map, to obtain an inpainting result.
Systems and methods for utilizing models to identify a vehicle accident based on vehicle sensor data and video data captured by a vehicle device
A device may receive sensor data and video data associated with a vehicle, and may process the sensor data, with a rule-based detector model, to determine whether a probability of a vehicle accident satisfies a first threshold. The device may preprocess acceleration data of the sensor data to generate calibrated acceleration data, and may process the calibrated acceleration data, with an anomaly detector model, to determine whether the calibrated acceleration data includes anomalies. The device may filter the sensor data to generate filtered sensor data, and may process the filtered sensor data and anomaly data, with a decision model, to determine whether the probability of the vehicle accident satisfies a second threshold. The device may process the filtered sensor data, the anomaly data, and the video data, with a machine learning model, to determine whether the vehicle accident has occurred, and may perform one or more actions.
CAMERA-RADAR DATA FUSION FOR EFFICIENT OBJECT DETECTION
A method includes obtaining, by a processing device, input data derived from a set of sensors associated with an autonomous vehicle (AV), extracting, by the processing device from the input data, a plurality of sets of features, generating, by the processing device using the plurality of sets of features, a fused bird's-eye view (BEV) grid. The fused BEV grid is generated based on a first BEV grid having a first scale and a second BEV grid having a second scale different from the first scale. The method further includes providing, by the processing device, the fused BEV grid for object detection.