Patent classifications
G06V20/41
Video streaming user interface with data from multiple sources
A management server receives a request to view video content; access a stream of video content corresponding to the request; access and store, in one or more databases, metadata associated with the video content, wherein the metadata includes a plurality of detections and was generated by the video gateway device based on an analysis of the video content and previously uploaded to a remote computing device; generating and causing display of a user interface that includes: the stream of the video content, a seek bar configured to be manipulated by a user so that the user can navigate through the video content, and the metadata overlaid on the seek bar such that the plurality of detections are overlaid at locations on the seek bar corresponding to when the plurality of detections were detected.
Assigning case identifiers to video streams
A process mining system performs process mining using visual logs generated from video streams of worker devices. Specifically, for a given worker device, the process mining system obtains a series of images capturing a screen of a worker device while the worker device processes one or more tasks related to an operation process. The process mining system determines activity labels for a plurality of images. An activity label for an image may indicate an activity performed on the worker device when the image was captured. The activity label is determined by extracting information from pixels of the image and inferring the activity of the worker device from the extracted information. The process mining system generates event logs from the visual logs of worker devices and uses the event logs for process mining.
Method and program for producing and providing reactive video
The inventive concept relates to a method for producing a multi-reactive video and providing a multi-reactive video service, and a program using the same. It is possible to grasp a user's reaction to a video by recording manipulation details for a specific user's multi-reactive video. For example, it is possible to grasp the object of interest and the degree of interest of a user and to grasp a user interest in the user, by grasping the number of touch manipulations to the user's multi-reactive video, a frame in which a touch manipulation has been performed, and an object in the frame, or the like.
PERFORMANCE AGENT TRAINING METHOD, AUTOMATIC PERFORMANCE SYSTEM, AND PROGRAM
A performance agent training method realized by at least one computer includes observing a first performance of a musical piece by a performer, generating, by a performance agent, performance data of a second performance to be performed in parallel with the first performance, outputting the performance data such that the second performance is performed in parallel with the first performance of the performer, acquiring a degree of satisfaction of the performer with respect to the second performance performed based on the output performance data, and training the performance agent by reinforcement learning, using the degree of satisfaction as a reward.
Unmanned aerial vehicle (UAV) data collection and claim pre-generation for insured approval
Systems and methods are described for using data collected by unmanned aerial vehicles (UAVs) to generate insurance claim estimates that an insured individual may quickly review, approve, or modify. When an insurance-related event occurs, such as a vehicle collision, crash, or disaster, one or more UAVs are dispatched to the scene of the event to collect various data, including data related to vehicle or real property (insured asset) damage. With the insured's permission or consent, the data collected by the UAVs may then be analyzed to generate an estimated insurance claim for the insured. The estimated insurance claim may be sent to the insured individual, such as to their mobile device via wireless communication or data transmission, for subsequent review and approval. As a result, insurance claim handling and/or the online customer experience may be enhanced.
A SYSTEM AND METHOD FOR PROVIDING ASSISTANCE DURING BOTTLE-FEEDING
A system provides assistance during bottle-feeding. Video images of a subject bottle-feeding an infant are captured and displayed. Using image analysis, a reorientation of the bottle and/or infant is determined that is required in order to reach a desired bottle orientation and/or infant orientation. Reorientation instructions are provided in combination with the video images to assist the subject in reorienting the bottle and/or the infant to achieve the desired bottle orientation.
DISPLAY APPARATUS AND METHOD FOR PERSON RECOGNITION AND PRESENTATION
Provided are a display apparatus and a person recognition and presentation method. The display apparatus includes a display and a controller that is in communication with the display. The controller is configured to: associated information of a display interface of the display and generate a scenario image for recognition in response to a user command; obtain facial feature information for recognition in the scenario image; obtain similar facial feature information when a matching confidence level of pre-stored facial feature information in a database with the facial feature information for recognition does not exceed a preset confidence level; obtain average-person recognition data; generate a sharing control uniquely matching with the facial feature information for recognition; and control the display to present the average-person recognition data and the sharing control on a current display interface.
Method and System for Scene-Aware Audio-Video Representation
Embodiments disclose a method and system for a scene-aware audio-video representation of a scene. The scene-aware audio video representation corresponds to a graph of nodes connected by edges. A node in the graph is indicative of the video features of an object in the scene. An edge in the graph connecting two nodes indicates an interaction of the corresponding two objects in the scene. In the graph, at least one or more edges are associated with audio features of a sound generated by the interaction of the corresponding two objects. The graph of the audio-video representation of the scene may be used to perform a variety of different tasks. Examples of the tasks include one or a combination of an action recognition, an anomaly detection, a sound localization and enhancement, a noisy-background sound removal, and a system control.
INFORMATION PUSHING METHOD IN VEHICLE DRIVING SCENE AND RELATED APPARATUS
This disclosure relates to an information pushing method in a vehicle driving scene. The method may include receiving push information in the vehicle driving scene and obtaining driving scene image information collected by an in-vehicle image collection device. The method may further include identifying scene category identification information based on the driving scene image information. The scene category identification information is for indicating a category of the environmental information. The method may further include pushing, in response to the scene category identification information satisfying a push condition, the push information in the vehicle driving scene.
RECOMMENDATION OF AUDIO BASED ON VIDEO ANALYSIS USING MACHINE LEARNING
An electronic device and method for recommendation of audio based on video analysis is provided. The electronic device receives one or more frames of a first scene of a plurality of scenes of a video. The first scene includes a set of objects. The electronic device applies a trained neural network model on the received one or more frames to detect the set of objects. The electronic device determines an impact score of each object of the detected set of objects of the first scene based on the application of the trained neural network model on the set of objects. The electronic device further selects at least one first object from the set of objects based on the impact score of each object, and recommends one or more first audio tracks as a sound effect for the first scene based on the selected at least one first object.