G06V40/164

System and method for generating a mixed reality experience

An electronic device includes an image sensor, a projector, an adjustable mount, a processor, and a memory. The memory stores instructions executable by the processor to: receive at least one image of an environment around the electronic device from the image sensor; determine a face pose of a viewer based on the at least one image; determine characteristics of a projection surface based on the at least one image; control the projector to project a plurality of images onto the projection surface, the images determined based in part on the face pose of the viewer and the characteristics of the projection surface, wherein the images are configured to be perceived as a three-dimensional (3D) object image when viewed through 3D glasses; and control the adjustable mount to adjust a position or an orientation of the projector based in part on a change in the face pose of the viewer.

Systems and methods to adapt and optimize human-machine interaction using multimodal user-feedback

Systems and methods for human-machine interaction. An adaptive behavioral control system of a human-machine interaction system controls an interaction sub-system to perform a plurality of actions for a first action type in accordance with a computer-behavioral policy, each action being a different alternative action for the action type. The adaptive behavioral control system detects a human reaction of an interaction participant to the performance of each action of the first action type from data received from a human reaction detection sub-system. The adaptive behavioral control system stores information indicating each detected human reaction in association with information identifying the associated action. In a case where stored information indicating detected human reactions for the first action type satisfy an update condition, the adaptive behavioral control system updates the computer-behavioral policy for the first action type.

Method and System for Verifying Uploaded Customer Photographs
20220392258 · 2022-12-08 ·

A method and system are provided for automatically verifying an image uploaded to a computer system by a customer. The method includes generating an upload prompt to request a customer to upload an image of a headshot of the customer, a front of an identification card of the customer, or a back of an identification card of the customer. The uploaded image is automatically processed with a computer-implemented classification model to attempt to classify the uploaded image as a headshot, an identification card front, or an identification card back. An uploaded headshot is processed with a face detection model to return face bounding box for the image. An uploaded identification card front is processed with an identification card detection model to generate an identification card front bounding box for the uploaded image, and with the face detection model to return an identification card face bounding box for the identification card front. An uploaded identification card back is processed with the identification card detection model to return an identification card back bounding box for the uploaded image.

Methods and systems for correcting, based on speech, input generated using automatic speech recognition
11521608 · 2022-12-06 · ·

Methods and systems for correcting, based on subsequent second speech, an error in an input generated from first speech using automatic speech recognition, without an explicit indication in the second speech that a user intended to correct the input with the second speech, include determining that a time difference between when search results in response to the input were displayed and when the second speech was received is less than a threshold time, and based on the determination, correcting the input based on the second speech. The methods and systems also include determining that a difference in acceleration of a user input device, used to input the first speech and second speech, between when the search results in response to the input were displayed and when the second speech was received is less than a threshold acceleration, and based on the determination, correcting the input based on the second speech.

Face quality of captured images
11514713 · 2022-11-29 · ·

The disclosure pertains to techniques for image processing. One such technique comprises a method for image selection, comprising: obtaining a sequence of images, detecting a first face in one or more images of the sequence of images, determining a first location for the detected first face in each of the images having the detected first face, generating a heat map based on the first location of the detected first face in each of the images of the sequences of images, determining a face quality score for the detected first face for each of the one or more images having the detected first face, determining a peak face quality score for the detected first face based in part on the face quality score and the generated heat map, and selecting a first image of the sequence of images, corresponding with the peak face quality score for the detected first face.

AUTOMATIC MEDIA CAPTURE USING BIOMETRIC SENSOR DATA
20220373791 · 2022-11-24 ·

Systems and methods herein describe a media capture system that receives sensor data from biometric sensors coupled to a head-wearable apparatus, detects a trigger event corresponding to a user of the head-wearable apparatus based on the sensor data, captures images using a camera coupled to the head-wearable apparatus, and transmits the captured images to a client device.

Facial recognition system

Various embodiments of a facial recognition system are provided. In one embodiment, a processor determines a value for a lighting parameter associated with a captured facial image, determines whether any previously obtained images in a biometric database includes a similar value for the lighting parameter and, if not, stores the newly captured image in the database along with the lighting parameter value. In another embodiment, the processor calculates a score indicative of the likelihood that the face in the captured facial image is identical to the face of a previously obtained image in the database, determines whether the score exceeds a threshold value and, if so, generates a signal indicating a match. The processor adjusts the threshold based on one or more parameter values.

Information processing apparatus and information processing method
11503162 · 2022-11-15 · ·

Provided is an information processing apparatus that includes an acquisition unit and a generation control unit. The acquisition unit acquires input information including at least one of an image or audio of a first user. The generation control unit controls, on the basis of request information including a request to output information including at least one of an image or audio and the acquired input information, generation of output information related to the first user to be output by a terminal of a second user who is a communication partner of the first user.

Occupant monitoring device, occupant monitoring method, and occupant monitoring program

An occupant monitoring device includes: an acquisition unit that acquires a captured image obtained by imaging a region in which there is a probability that a face of an occupant is present in a vehicle; a determination unit that determines whether the captured image acquired by the acquisition unit corresponds to a first image including the face a part of which is hidden by an accessory or a second image including the face a part of which is hidden by a non-accessory object other than the accessory; and a processing unit that detects face information regarding the face of the occupant based on the captured image in different modes according to a determination result in the determination unit, and monitors a state change of the face of the occupant based on a detection result.

Systems and methods for performing object detection and motion detection on video information

Systems and methods for converting video information into electronic output files are disclosed. Exemplary implementations may: obtain video information defining one or more videos; select a subset of the visual content included in the video content of the particular video; perform object detection to detect objects in the selected subset; perform motion detection to detect a movement for one or more of the detected objects in the selected subset, responsive to the object detection detecting one or more detected objects; and generate and store an electronic output file, responsive to the detection of the movement.