Patent classifications
G06V40/174
Visual-based security compliance processing
Multiple cameras capture videos within a secure room. When individuals are detected as entering the room, identities of the individuals are resolved. When an asset is exposed in a field of view of one of the cameras, the individuals' eye and head movements are tracked from the videos with respect to one another and the asset. Additionally, touches made by any of the individuals on the asset are tracked from the videos. The eye and head movements are correlated with the touches or lack of touches according to a security policy for the asset. Any violations of the security policy are written to a secure audit log for the room and the asset.
SERVICE MANAGEMENT USING USER EXPERIENCE METRICS
A method, system, and computer usable program product to determine a first mood of the user, where the first mood is based on a characteristic of the user at a first time during the providing of an online service and to determine a second mood of the user, where the second mood is based on a characteristic of the user at a second time during the providing of the online service. The first mood of the user and the second mood of the user are compared to determine a delta or change in mood of the user.
METHOD AND APPARATUS FOR RECOMMENDING AN INTERFACE THEME
A method, and an apparatus for recommending an interface theme are provided. An exemplary embodiment of the method includes: obtaining a target image which includes an image of a target person; obtaining characteristic information of the target person based on the target image; obtaining a selection list of recommended themes, wherein the recommended themes are interface themes that match the characteristic information of the target person; and outputting the selection list of recommended themes.
Systems and Methods for Assessing Viewer Engagement
A system for quantifying viewer engagement with a video playing on a display includes at least one camera to acquire image data of a viewing area in front of the display. A microphone acquires audio data emitted by a speaker coupled to the display. The system also includes a memory to store processor-executable instructions and a processor. Upon execution of the processor-executable instructions, the processor receives the image data and the audio data and determines an identity of the video displayed on the display based on the audio data. The processor also estimates a first number of people present in the viewing area and a second number of people engaged with the video. The processor further quantifies the viewer engagement of the video based on the first number of people and the second number of people.
Display Screen Front Panel of HMD for Viewing by Users Viewing the HMD Player
Method for providing image of HMD user to a non-HMD user includes, receiving a first image of a user including the user's facial features captured by an external camera when the user is not wearing a head mounted display (HMD). A second image capturing a portion of the facial features of the user when the user is wearing the HMD is received. An image overlay data is generated by mapping contours of facial features captured in the second image with contours of corresponding facial features captured in the first image. The image overlay data is forwarded to the HMD for rendering on a second display screen that is mounted on a front face of the HMD.
COLLECTION OF MACHINE LEARNING TRAINING DATA FOR EXPRESSION RECOGNITION
Apparatus, methods, and articles of manufacture for implementing crowdsourcing pipelines that generate training examples for machine learning expression classifiers. Crowdsourcing providers actively generate images with expressions, according to cues or goals. The cues or goals may be to mimic an expression or appear in a certain way, or to “break” an existing expression recognizer. The images are collected and rated by same or different crowdsourcing providers, and the images that meet a first quality criterion are then vetted by expert(s). The vetted images are then used as positive or negative examples in training machine learning expression classifiers.
Information-processing device, vehicle, computer-readable storage medium, and information-processing method
An information-processing device includes a first feature-value information acquiring unit for acquiring an acoustic feature-value vector and a language feature-value vector extracted from a user's spoken voice. The information-processing device includes a second feature-value information acquiring unit for acquiring an image feature-value vector extracted from the user's facial image. The information-processing device includes an emotion estimating unit including a learned model including: a first attention layer using, as inputs, a first vector generated from the acoustic feature-value vector and a second vector generated from the image feature-value vector; and a second attention layer using, as an input, an output vector from the first attention layer and a third vector generated from the language feature-value vector, wherein the emotion estimating unit is for estimating the user's emotion based on the output vector from the second attention layer.
Systems, methods, devices and apparatuses for detecting facial expression
A system, method and apparatus for detecting facial expressions according to EMG signals.
Spoof detection based on challenge response analysis
Methods, systems, and computer-readable storage media for determining that a subject is a live person include capturing a set of images of a subject instructed to perform a facial expression. A region of interest for the facial expression is determined in a first image of the set, the first image representing a first facial state that includes the facial expression. A set of facial features is identified in the region of interest, the facial features being indicative of interaction between facial muscles and skin of the subject due to the subject performing the facial expression. A determination is made, based on the facial features, that the first image substantially matches a template image of the facial expression of the subject. Responsive to determining that the first image substantially matches the template image, identifying the subject as a live person.
Message delivery apparatus and methods
The present disclosure provides a more adapt and accessible messaging system. In some aspects, the present disclosure relates to a messaging system that allows users to prerecord messages for future or real time use. Allowing users to communicate emotional messages, both visually and audially to relay feedback to both users. In some embodiments, the system may be useful to help with mental issues, self-esteem problems and other personal issues, non-limiting examples. In some implementations, the device may provide instant messages from one message device to another message device.