Patent classifications
G06V20/41
Reducing bandwidth requirements of virtual collaboration sessions
A computer-implemented method, a computer system and a computer program product reduce bandwidth requirements of a virtual collaboration session. The method includes capturing session data from a virtual collaboration session. The session data is selected from a group consisting of video data, audio data, an image of a screen of a connected device and text data. The method also includes connecting to a live blog platform. The method further includes transmitting a text transcription of the virtual collaboration session to the live blog platform. The text transcription is generated by scanning the audio data using a speech-to-text algorithm. In addition, the method includes classifying a topic in the virtual collaboration session based on importance. Lastly, the method includes transmitting a multimedia file related to the topic to the live blog platform in response to the topic being classified as important. The multimedia file is extracted from the session data.
Intelligence-based editing and curating of images
In one embodiment, a method includes accessing a plurality of image frames captured by one or more cameras, classifying one or more first objects detected in one or more first image frames of the plurality of image frames as undesirable, applying a pixel filtering to the one or more first image frames to replace one or more first pixel sets associated with the one or more first objects with pixels from one or more second image frames of the plurality of image frames to generate a final image frame, providing the final image frame for display.
Augmented reality object manipulation
A processing system having at least one processor may detect a first object in a first video of a first user and detect a second object in a second video of a second user, where the first video and the second video are part of a visual communication session between the first user and the second user. The processing system may further detect a first action in the first video relative to the first object, detect a second action in the second video relative to the second object, detect a difference between the first action and the second action, and provide a notification indicative of the difference.
Managing virtual surveillance windows for video surveillance
Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for managing virtual surveillance windows for video surveillance. The methods, systems, and apparatus include actions of obtaining an original video, generating a downscaled video from the original video, detecting a first event at a location from the downscaled video using a first classifier, generating a windowed video from the original video based on the location, detecting a second event from the windowed video, and performing an action in response to detecting the second event.
DEEP LEARNING-BASED MARINE OBJECT CLASSIFICATION USING 360-DEGREE IMAGES
Marine object detection, localization and classification systems and related techniques include an imaging system configured capture a stream of panoramic images of the water surrounding a mobile structure, including a view of the horizon. The images may include a 360-degree view from the mobile structure. The system is configured to analyze the stream of images using a marine video analytics system and/or a convolutional neural network to detect a region of interest comprising an object on the surface of the water, classify the detected object and relay the results to the user and/or a processing system. The analysis may include determining a horizon in a captured image, defining tiles across the horizon, and detecting objects in each tile.
METHOD AND APPARATUS FOR IMAGE SEGMENTATION MODEL TRAINING AND FOR IMAGE SEGMENTATION
A method for training an image segmentation model includes: acquiring target category feature information that represents category features of a training sample and a prediction sample, and associated scene feature information thereof; performing splicing processing on the target category feature information and the associated scene feature information; inputting first spliced feature information obtained by the splicing processing into an initial generation network to perform image synthesis processing; inputting a first synthesized image obtained by the synthesis processing into an initial determination network to determine authenticity; inputting the first synthesized image into a classification network of an initial image segmentation model to perform image segmentation, to obtain a first image segmentation result; and training the classification network of the initial image segmentation model based on a first image determination result, the first image segmentation result, and the target category feature information, so as to obtain a target image segmentation model.
METHOD AND APPARATUS FOR GENERATING VIDEO WITH 3D EFFECT, METHOD AND APPARATUS FOR PLAYING VIDEO WITH 3D EFFECT, AND DEVICE
A method and an apparatus for generating a video with a three-dimensional (3D) effect, a method and an apparatus for playing a video with a 3D effect, and a device are provided. The method includes: obtaining an original video; segmenting at least one frame of raw image of the original video to obtain a foreground image sequence including a moving object, the foreground image sequence including at least one frame of foreground image; determining, based on the foreground image sequence, a target raw image in which a target occlusion image is to be placed and an occlusion method of the target occlusion image in the target raw image; adding the target occlusion image to the target raw image based on the occlusion method to obtain a final image; and generating a target video with a 3D effect based on the final image and the original video.
DETECTING EMOTIONAL STATE OF A USER BASED ON FACIAL APPEARANCE AND VISUAL PERCEPTION INFORMATION
A method for detecting an emotional state of a user includes obtaining a first data stream indicative of facial appearance and gaze direction of the user as the user is viewing a scene, determining, based on the first data stream, facial expression feature information indicative of emotional facial expression of the user, obtaining a second data stream indicative of visual content in a field of view of the user, determining, based on the second data stream, visual feature information indicative of visual content in the scene, determining emotional state information based on analyzing the facial expression feature information determined based on the first data stream and the visual feature information determined based on the second data stream, and performing an operation with respect to the emotional state information, wherein the emotional state information is indicative of the emotional state of the user.
APPLIED BEHAVIORAL THERAPY APPARATUS AND METHOD
An apparatus for providing automated analysis and monitoring of an ABT session is presented herein. The apparatus may include a display configured to present material for the ABT session to a patient, at least one video capture device configured to capture video data for the ABT session related to at least one of first facial features of the patient, second facial features of a therapist, or a response to the material presented on the display, at least one audio capture device configured to capture audio data for the ABT session related to at least one of a first voice of the patient or a second voice of the therapist, and at least one processor configured to analyze, for the ABT session, data regarding the material presented on the display, the captured video data, and the captured audio data to produce an analysis of the ABT session.
INTELLIGENT BIRD FEEDING METHOD, ELECTRONIC DEVICE AND BIRD FEEDER
An intelligent bird feeding method may include: shooting video information of a bird in a preset area through the camera component, transmitting the video information to an electronic device to indicate the electronic device to determine a category of the bird and a state of the bird based on the video information; determining whether a bird food needs to be fed and a category of the bird food to be fed based on the category of the bird and the state of the bird; and under the circumstance that the bird food needs to be fed, selecting a bird food of a corresponding category for feeding according to the category of the bird food to be fed.