Patent classifications
H04N5/278
Communication transfer between devices
A method may include obtaining an indicator that a first device is in a location of a second device and in response to obtaining the indicator, sending a redirect request to a communication service provider of the first device to direct, to the second device, incoming communication requests handled by the communication service provider that are directed to the first device. The method may further include after sending the redirect request and after a communication request to a communication session is directed to the first device, obtaining, at the second device, a communication indication to participate in the communication session. The method may further include directing audio of the communication session to a transcription system and obtaining, at the second device, the transcription of the audio from the transcription system. The method may also include presenting, by the second device, the audio and the transcription.
ADJUSTING AUDIO AND NON-AUDIO FEATURES BASED ON NOISE METRICS AND SPEECH INTELLIGIBILITY METRICS
Some implementations involve determining a noise metric and/or a speech intelligibility metric and determining a compensation process corresponding to the noise metric and/or the speech intelligibility metric. The compensation process may involve altering a processing of audio data and/or applying a non-audio-based compensation method. In some examples, altering the processing of the audio data does not involve applying a broadband gain increase to the audio signals. Some examples involve applying the compensation process in an audio environment. Other examples involve determining compensation metadata corresponding to the compensation process and transmitting an encoded content stream that includes encoded compensation metadata, encoded video data and encoded audio data from a first device to one or more other devices.
ADJUSTING AUDIO AND NON-AUDIO FEATURES BASED ON NOISE METRICS AND SPEECH INTELLIGIBILITY METRICS
Some implementations involve determining a noise metric and/or a speech intelligibility metric and determining a compensation process corresponding to the noise metric and/or the speech intelligibility metric. The compensation process may involve altering a processing of audio data and/or applying a non-audio-based compensation method. In some examples, altering the processing of the audio data does not involve applying a broadband gain increase to the audio signals. Some examples involve applying the compensation process in an audio environment. Other examples involve determining compensation metadata corresponding to the compensation process and transmitting an encoded content stream that includes encoded compensation metadata, encoded video data and encoded audio data from a first device to one or more other devices.
Subtitle presentation based on volume control
Systems and methods are provided for presenting subtitles. The systems and methods include accessing, by a user device, a video discovery graphical user interface that includes a plurality of videos; receiving a user input that gradually reduces volume of the user device; determining that the volume of the user device has gradually been reduced by the user input until a mute state has been reached in which audio output of the user device is disabled; and in response to determining that the volume of the user device has gradually been reduced until the mute state has been reached, automatically causing subtitles of a first video of the plurality of videos to be displayed during playback of the first video.
Subtitle presentation based on volume control
Systems and methods are provided for presenting subtitles. The systems and methods include accessing, by a user device, a video discovery graphical user interface that includes a plurality of videos; receiving a user input that gradually reduces volume of the user device; determining that the volume of the user device has gradually been reduced by the user input until a mute state has been reached in which audio output of the user device is disabled; and in response to determining that the volume of the user device has gradually been reduced until the mute state has been reached, automatically causing subtitles of a first video of the plurality of videos to be displayed during playback of the first video.
SYSTEMS AND METHODS OF PRESENTING VIDEO OVERLAYS
Systems and methods are provided for relocating an overlay overlapping information in content. The systems and methods may comprise receiving a content item, the content item comprising a video image, and determining a first screen position of an information box (e.g., a score box) in the video image. Determining may be performed with image analysis and/or a machine learning model. The system receives an overlay image (e.g., a channel logo) with a second screen position and determines if the second screen position (e.g., for the logo) overlaps the first screen position (e.g., for the score). In response to determining the second screen position (e.g., of the logo) overlaps the first screen position (e.g., the score), the system modifies the second screen position (e.g., for the logo). Then the system generates for display the overlay image on the video in the modified screen position. The system may not relocate the overlay if the overlay is a high priority.
Caption modification and augmentation systems and methods for use by hearing assisted user
A system and method for facilitating communication between an assisted user (AU) and a hearing user (HU) includes receiving an HU voice signal as the AU and HU participate in a call using AU and HU communication devices, transcribing HU voice signal segments into verbatim caption segments, processing each verbatim caption segment to identify an intended communication (IC) intended by the HU upon uttering an associated one of the HU voice signal segments, for at least a portion of the HU voice signal segments (i) using an associated IC to generate an enhanced caption different than the associated verbatim caption, (ii) for each of a first subset of the HU voice signal segments, presenting the verbatim captions via the AU communication device display for consumption, and (iii) for each of a second subset of the HU voice signal segments, presenting enhanced captions via the AU communication device display for consumption.
Caption modification and augmentation systems and methods for use by hearing assisted user
A system and method for facilitating communication between an assisted user (AU) and a hearing user (HU) includes receiving an HU voice signal as the AU and HU participate in a call using AU and HU communication devices, transcribing HU voice signal segments into verbatim caption segments, processing each verbatim caption segment to identify an intended communication (IC) intended by the HU upon uttering an associated one of the HU voice signal segments, for at least a portion of the HU voice signal segments (i) using an associated IC to generate an enhanced caption different than the associated verbatim caption, (ii) for each of a first subset of the HU voice signal segments, presenting the verbatim captions via the AU communication device display for consumption, and (iii) for each of a second subset of the HU voice signal segments, presenting enhanced captions via the AU communication device display for consumption.
DISPLAY METHOD, METHOD OF TRAINING SEMANTIC UNIT DETECTION MODEL, ELECTRONIC DEVICE, AND STORAGE MEDIUM
A display method, a method of training a semantic unit detection model, an electronic device, and a storage medium, which relate to a field of artificial intelligence technology, in particular to fields of natural language processing and machine translation technologies. The display method includes: acquiring a language sequence to be displayed; dividing the language sequence to be displayed into a plurality of semantic units with semantics; and converting the plurality of semantic units into subtitles for display one by one.
DISPLAY METHOD, METHOD OF TRAINING SEMANTIC UNIT DETECTION MODEL, ELECTRONIC DEVICE, AND STORAGE MEDIUM
A display method, a method of training a semantic unit detection model, an electronic device, and a storage medium, which relate to a field of artificial intelligence technology, in particular to fields of natural language processing and machine translation technologies. The display method includes: acquiring a language sequence to be displayed; dividing the language sequence to be displayed into a plurality of semantic units with semantics; and converting the plurality of semantic units into subtitles for display one by one.