G10L25/54

DOCUMENT CREATION AND EDITING VIA AUTOMATED ASSISTANT INTERACTIONS
20230049562 · 2023-02-16 ·

Implementations set forth herein relate to an automated assistant that allows a user to create, edit, and/or share documents without directly interfacing with a document editing application. The user can provide an input to the automated assistant in order to cause the automated assistant to interface with the document editing application and create a document. In order to identify a particular action to perform with respect to a document, and/or identify a particular subsection within the document to direct the action, the automated assistant can rely on semantic annotations. As a user continues to interact with the automated assistant to edit a document, the semantic annotations can be updated according to how the document is changing and/or how the user refers to the document. This can allow the automated assistant to more readily fulfill document-related requests that may lack express details.

DOCUMENT CREATION AND EDITING VIA AUTOMATED ASSISTANT INTERACTIONS
20230049562 · 2023-02-16 ·

Implementations set forth herein relate to an automated assistant that allows a user to create, edit, and/or share documents without directly interfacing with a document editing application. The user can provide an input to the automated assistant in order to cause the automated assistant to interface with the document editing application and create a document. In order to identify a particular action to perform with respect to a document, and/or identify a particular subsection within the document to direct the action, the automated assistant can rely on semantic annotations. As a user continues to interact with the automated assistant to edit a document, the semantic annotations can be updated according to how the document is changing and/or how the user refers to the document. This can allow the automated assistant to more readily fulfill document-related requests that may lack express details.

AUDIO FILE ANNOTATION
20230094828 · 2023-03-30 ·

Text-to-speech translation is used to generate a transcript for an audio file. Text segments are associated with time segments in the transcript. A trained machine learning model determines, based on the text in the transcript, one or more topics for the audio file. The transcript is modified to include the determined one or more topics. A user interface may be presented that allows a user to search for portions of an audio file that relate to a particular topic. In response to the selected or entered topic, the user interface presents segments having a matching topic. The user may use voice or other user interface commands to modify the annotation of the audio file. User commands may also be used to extract data from the transcript and copy the data to a clipboard or to another application.

SYSTEMS AND METHODS FOR INSERTING CONTEXTUAL ADVERTISEMENTS INTO A VIRTUAL ENVIRONMENT
20230032565 · 2023-02-02 ·

Systems and methods are described herein for inserting contextual advertisements into a virtual environment. While presenting a virtual environment, a media device receives speech input. The media device selects an advertisement from an advertisement database based on the speech. The media device then determines the current environment of the virtual environment and generates a virtual object that is related to the current environment. The media device overlays the selected advertisement over the generated virtual object and displays the generated virtual object in the virtual environment.

SYSTEMS AND METHODS FOR INSERTING CONTEXTUAL ADVERTISEMENTS INTO A VIRTUAL ENVIRONMENT
20230032565 · 2023-02-02 ·

Systems and methods are described herein for inserting contextual advertisements into a virtual environment. While presenting a virtual environment, a media device receives speech input. The media device selects an advertisement from an advertisement database based on the speech. The media device then determines the current environment of the virtual environment and generates a virtual object that is related to the current environment. The media device overlays the selected advertisement over the generated virtual object and displays the generated virtual object in the virtual environment.

NATURAL-LANGUAGE PROCESSING ACROSS MULTIPLE LANGUAGES
20230096070 · 2023-03-30 ·

A method includes obtaining a query in a base language and translating the query to generate one or more translated queries each in a respective target language. The method also includes searching one or more sets of electronic files based on the one or more translated queries to generate target-language search results, where each translated query is used to search one or more electronic files that include content in the respective target language of the translated query. The method also includes, based on the target-language search results, scheduling one or more electronic files of the one or more sets of electronic files for at least partial translation to the base language.

Interacting with a virtual assistant to receive updates

Technologies are disclosed for interacting with a virtual assistant to request updates associated with one or more events and/or perform actions. According to some examples, a user may use their voice to interact with a virtual assistant to receive updates relating to events occurring during a certain period of time. For example, a user may request an update associated with one or more events occurring that day. The system may access data sources (e.g., calendar services, email services, etc.) to obtain data associated with the events, tag the events according to one or more conditions indicated by the data, and/or rank the events according to the tags. In addition, to resolve conditions associated with the events, the virtual assistant may also include options in the update to perform certain actions and/or to provide response data. The virtual assistant may generate the update and audibly provide the update to the user.

Interacting with a virtual assistant to receive updates

Technologies are disclosed for interacting with a virtual assistant to request updates associated with one or more events and/or perform actions. According to some examples, a user may use their voice to interact with a virtual assistant to receive updates relating to events occurring during a certain period of time. For example, a user may request an update associated with one or more events occurring that day. The system may access data sources (e.g., calendar services, email services, etc.) to obtain data associated with the events, tag the events according to one or more conditions indicated by the data, and/or rank the events according to the tags. In addition, to resolve conditions associated with the events, the virtual assistant may also include options in the update to perform certain actions and/or to provide response data. The virtual assistant may generate the update and audibly provide the update to the user.

Guidance query for cache system

A device may be configured to determine whether an audio file is a first type of audio file that is capable of being processed to recognize the voice query based on a characteristic of the audio file itself or a second type of audio file that may require speech recognition processing in order to recognize the voice query associated with the audio file. In determining whether the audio file is a first type of audio file or a second type of audio file, a query filter associated with the device may be configured to access one or more guidance queries. Using the one or more guidance queries, the device may classify the audio file as a first type of audio file or a second type of audio file based on receiving only a portion of the audio file, thereby improving the speed at which the audio file can be processed.

Guidance query for cache system

A device may be configured to determine whether an audio file is a first type of audio file that is capable of being processed to recognize the voice query based on a characteristic of the audio file itself or a second type of audio file that may require speech recognition processing in order to recognize the voice query associated with the audio file. In determining whether the audio file is a first type of audio file or a second type of audio file, a query filter associated with the device may be configured to access one or more guidance queries. Using the one or more guidance queries, the device may classify the audio file as a first type of audio file or a second type of audio file based on receiving only a portion of the audio file, thereby improving the speed at which the audio file can be processed.