Patent classifications
H04N21/439
Device and method for visually displaying speaker's voice in 360-degree video
An electronic device is disclosed. In addition, various embodiments identified through the specification are possible. The electronic device includes a display, a processor, and a memory storing instructions that, when executed by the processor, cause the processor to display, when a video supporting a plurality of orientation regions is played, a screen of a first orientation region among the plurality of orientation regions and a first text corresponding to a voice of a first speaker in the screen, and display, in response to a user input of selecting a voice of a second speaker located in a second orientation region, a screen of the second orientation region.
Methods and apparatus to detect spillover
Methods and apparatus to detect spillover are disclosed. An example apparatus includes at least one memory, instructions in the apparatus, and processor circuitry to execute the instructions to: identify a quantity of first durations of loudness in an audio signal of media; calculate a ratio of the quantity of the first durations of loudness to a quantity of second durations of loudness in the audio signal of the media, the quantity of the second durations of loudness including the quantity of the first durations of loudness; and in response to a detection of the audio signal being spillover, store data denoting the media as un-usable to credit a media exposure when the ratio does not satisfy a loudness ratio threshold, the storing of the data to improve an accuracy of media exposure credits by not crediting spillover media.
Methods and systems for recommending content in context of a conversation
A media guidance application may monitor a conversation among users, and identify keywords in the conversation, without the use of wakewords. The keywords are used to search for media content that is relevant to the on-going conversation. Accordingly, the media guidance application presents relevant content to the users, during the conversation, to more actively engage the users. A conversation monitoring window may be used to present conversation information as well as relevant content. A listening mode may be used to manage when the media guidance application processes speech from a conversation. The media guidance application may access user profiles for keywords, select content types, select content sources, and determine relevancy of media content, to provide content in context of a conversation.
Methods and apparatus to identify media based on watermarks across different audio streams and/or different watermarking techniques
Example apparatus disclosed herein are to detect a first watermark embedded in an audio stream associated with media, the first watermark embedded and detected based on a first watermarking technique; and detect a second watermark embedded in the audio stream, the second watermark embedded and detected based on a second watermarking technique. Disclosed example apparatus are also to assign the first watermark to a first monitoring track and to a second monitoring track, the first monitoring track limited to watermarks embedded in the audio stream based on the first watermarking technique, the second monitoring track limited to watermarks embedded in the audio stream based on any of the first or second watermarking techniques; group the first and second watermarks to form a media detection event when the second watermark is assigned to the second monitoring track; and cause transmission of the media detection event to a data collection facility.
ADAPTIVE VOLUME CONTROL FOR MEDIA OUTPUT DEVICES AND SYSTEMS
Various arrangements for performing dynamic volume control are provided. Audio characteristics of audio content being output to a user may be identified. Adjustments made to an audio volume setting by the user while the audio content is being output to the user can be monitored. A machine learning model can be trained based on the adjustments made to the audio volume setting by the user that are mapped with the audio characteristics of the audio content. After the machine learning model is trained, the audio volume setting can be adjusted based at least in part on the trained machine learning model analyzing audio content.
SYSTEM FOR MANAGING A VIRTUAL MEETING
A system for managing a virtual meeting (e.g., video conference) includes memory storing a video conference application and at least one processor to execute the video conference application to generate a virtual meeting view for a first attendee including multiple attendee video streams arranged according to a virtual attendee arrangement specifying positions of the attendee video streams relative to each other in the virtual meeting view, receive second attendee audio data associated with a second attendee video stream, identify a particular video stream position specified by the virtual attendee arrangement, determine differential stereo effect data corresponding with the particular video stream position, and apply the differential stereo effect data to the second attendee audio data to provide differential audio signals on different audio channels output to the first attendee to create a stereo sound effect corresponding with the particular video stream position.
SYSTEM FOR MANAGING A VIRTUAL MEETING
A system for managing a virtual meeting (e.g., video conference) includes memory storing a video conference application and at least one processor to execute the video conference application to generate a virtual meeting view for a first attendee including multiple attendee video streams arranged according to a virtual attendee arrangement specifying positions of the attendee video streams relative to each other in the virtual meeting view, receive second attendee audio data associated with a second attendee video stream, identify a particular video stream position specified by the virtual attendee arrangement, determine differential stereo effect data corresponding with the particular video stream position, and apply the differential stereo effect data to the second attendee audio data to provide differential audio signals on different audio channels output to the first attendee to create a stereo sound effect corresponding with the particular video stream position.
METHOD AND DATA PROCESSING APPARATUS
A method of generating an emotion descriptor icon includes receiving input content comprising video information, and performing analysis on the input content to produce information representing the video information with respect to a plurality of characteristics. The method also includes determining, based on a comparison of the information representing the video information at a temporal position in the video information and a set of information items respectively representing an emotion state, a relative likelihood of association between the input content and at least some of a plurality of emotion states, selecting an emotion state based on the outcome of the determination, and outputting an emotion descriptor icon selected from an emotion descriptor icon set comprising a plurality of emotion descriptor icons. The outputted emotion descriptor icon is associated with the selected emotion state.
Methods and apparatus to identify and credit media using ratios of media characteristics
Apparatus, systems, articles of manufacture, and methods to identify and credit media using ratios of media characteristics are disclosed herein. Example apparatus to identify media include at least one memory, instructions, and at least one processor to execute the instructions to: determine a first ratio based on a first time interval and a second time interval of a monitored media signal; determine a second ratio based on the second time interval and a third time interval of the monitored media signal; generate a first ratio signature based on the first and second ratios; and initiate transmission of the first ratio signature to a recipient that is to compare the first signature with a second ratio signature to identify the media.
ELECTRONIC DEVICE FOR PERFORMING SYNCHRONIZATION OF VIDEO DATA AND AUDIO DATA, AND CONTROL METHOD THEREFOR
An electronic device for use with an external electronic device includes a touchscreen display, at least one speaker, and at least one processor. The at least one processor may obtain a user input for outputting video data of a first medium while audio data of the first medium is output through the at least one speaker, identify a point of time when the audio data is output through the at least one speaker, based on the obtained user input, determine a point of time when the video data is to be output through the touchscreen display or an external electronic device, by a delay time calculated at least based on the identified point of time, and control the touchscreen display or the external electronic device such that the video data is output through the touchscreen display or the external electronic device at the determined point of time.