VOICE ASSISTANT SPEECH LANGUAGE PATHOLOGIST (VA SLP), SYSTEMS AND METHODS
20210104174 · 2021-04-08
Inventors
Cpc classification
G10L15/22
PHYSICS
A61B5/4088
HUMAN NECESSITIES
A61B5/4082
HUMAN NECESSITIES
A61B5/4803
HUMAN NECESSITIES
International classification
A61B5/00
HUMAN NECESSITIES
Abstract
There is provided herein a method and system for assisting speech/language therapy practice utilizing a voice interactive artificial intelligence-powered virtual assistant system.
Claims
1. A voice assistant speech language pathologist (VA SLP) based method for assisting speech language therapy practice, the method comprising: utilizing a voice interactive artificial intelligence-powered virtual assistant system, initiating conversation with a user, wherein initiating conversation with the user is triggered in response to the user's command or triggered by the virtual assistant system, wherein initiating conversation with a user comprises: identifying the user and/or uploading a personal speech therapy practice protocol personalized to the user's speech/lingual pathology; based on the personalized practice protocol, requesting the user to perform a task which comprises saying one or more words associated with the user's speech/lingual pathology; if the user's speech is determined to be at or above a threshold, rewarding the user with a positive game feature.
2. The method of claim 1, wherein if the user's speech is determined to be below the threshold, the virtual assistant system penalizes the user with a negative game feature or a lack of a positive game feature.
3. The method of claim 1, wherein the step of requesting the user to say one or more words associated with the user's speech/lingual pathology comprises: providing to the user a set of words and requesting the user to repeat them one or more times, providing a user a set of words and requesting the user to re-order them to form a meaningful sentence, providing to the user a set of words and requesting the user to repeat them, playing a sound and asking the user what object/subject produces such sound, describing an object and asking the user to name it, naming and object and asking the user to describe it, projecting a visual image and/or video clips and asking the user to name/describe it or any combination thereof.
4. The method of claim 1, wherein the step of determining if the speech is at or above a threshold comprises analyzing the user's speech quality.
5. The method of claim 4, wherein analyzing the user's speech quality is performed locally, at a remote server or partially locally and partially at a remote server.
6. The method of claim 4, wherein the speech quality is at least partially determined by the level of similarity between the user's speech and an expected speech.
7. The method of claim 6, wherein the level of similarity between the user's speech and the expected speech is determined based on a number of words which were as expected, a use of synonyms or homonyms, use of words from the same category or any combination thereof.
8. The method of claim 4, wherein analyzing the user's speech quality comprises determining, evaluating and/or measuring reaction time, number of attempts, order of words, stuttering, omission of words, mispronunciation of words/syllables, length of response time, rate of speech, “swallowing” of words, ratio between mispronounced and correctly pronounced words, speech fluency, use of correct word types, grammar correctness, use of key words, number of correct attempts, length of utterance, pitch of speech, intensity of speech or any combination thereof.
9. The method of claim 1, wherein the user's speech/lingual pathology is related to speech/language behavioral, developmental, rehabilitation and/or degenerative related conditions/diseases.
10. The method of claim 9, wherein the conditions/diseases are selected from a group consisting of aphasia, Parkinson, Alzheimer's, ALS, LISP speech disorder and stuttering.
11. The method of claim 1, wherein the user identification is achieved by recognizing the user's voice, by obtaining a predetermined voice command from the user, by a predefined code/PIN, by a command provided by an independent device or by any combination thereof.
12. The method of claim 1, wherein if the user's speech is determined to be at or above the threshold a predetermined number of times, the method further comprised a step of increasing a level of difficulty of a next task presented to the user.
13. The method of claim 1, wherein if the user's speech is determined to be below the threshold a predetermined number of times, the method further comprised a step of decreasing a level of difficulty of a next task presented to the user.
14. The method of claim 1, wherein initiating conversation with the user is triggered in response to the user's voice command and/or by the virtual assistant system.
15. The method of claim 1, wherein the voice interactive artificial intelligence-powered virtual assistant system is selected from a group consisting of Alexa or Google Assistant Siri and Bixby.
16. The method of claim 1, wherein the personalized speech therapy practice protocol comprises content, which varies between different users having different speech/lingual pathologies and/or wherein the personalized speech therapy practice protocol comprises content, which provides different game experience for users having different speech/lingual pathologies.
17. An interactive artificial intelligence-powered virtual voice assistant speech language pathologist (VA SLP) system for assisting speech language therapy practice, the system comprising: interactive artificial intelligence-powered virtual one or more processors configured to: initiate conversation with a user, wherein initiating conversation with the user is triggered in response to the user's command or triggered by the virtual assistant system, wherein initiating conversation with a user comprises: identifying the user and/or uploading a personal speech therapy practice protocol personalized to the user's speech/lingual pathology; trigger, based on the personalized practice protocol, a request to the user to perform a task which comprises saying one or more words associated with the user's speech/lingual pathology; determine if the user's speech is at, above or below a threshold, wherein if the user's speech is determined to be at or above a threshold, the processor is configured to reward the user with a positive game feature.
18. The system of claim 19, wherein if the user's speech is determined to be below the threshold, the processor is configured to penalizes the user with a negative game feature or a lack of a positive game feature.
19. The system of claim 19, wherein the step of requesting the user to say one or more words associated with the user's speech/lingual pathology comprises: providing to the user a set of words and requesting the user to repeat them one or more times, providing a user a set of words and requesting the user to re-order them to form a meaningful sentence, providing to the user a set of words and requesting the user to repeat them, playing a sound and asking the user what object/subject produces such sound, describing an object and asking the user to name it, naming and object and asking the user to describe it, projecting a visual image and asking the user to name/describe it or any combination thereof.
20. The system of claim 19, wherein the step of determining if the speech is at or above a threshold comprises analyzing the user's speech quality.
Description
BRIEF DESCRIPTION OF THE FIGURES
[0031] Some embodiments of the disclosure are described herein with reference to the accompanying FIGURES. The description, together with the FIGURES, makes apparent to a person having ordinary skill in the art how some embodiments may be practiced. The figures are for the purpose of illustrative description and no attempt is made to show structural details of an embodiment in more detail than is necessary for a fundamental understanding of the disclosure. For the sake of clarity, some objects depicted in the figures are not to scale.
In the FIGURES:
[0032]
DETAILED DESCRIPTION
[0033] The principles, uses and implementations of the teachings herein may be better understood with reference to the accompanying description and FIGURES. Upon perusal of the description and FIGURES present herein, one skilled in the art will be able to implement the teachings herein without undue effort or experimentation. In the FIGURES, same reference numerals refer to same parts throughout.
[0034] In the description and claims of the application, the words “include” and “have”, and forms thereof, are not limited to members in a list with which the words may be associated.
[0035] Reference is now made to
[0036] The user's speech is recorded, and the system analyzes the user's vocal response to the tasks and will determine the user's speech quality (step 108). It is noted that, according to some embodiments, the speech analysis may be performed in a remote server. According to other embodiments, the speech analysis may be performed locally, for example in a processor of the voice interactive artificial intelligence-powered virtual assistant. According to some embodiments, the speech analysis may be partially performed in a remote server and partially performed locally, for example in a processor of the voice interactive artificial intelligence-powered virtual assistant. According to some embodiments, the speech may be assigned a score. The score represents the speech quality. The speech quality may be evaluated based on various parameters, such as but not limited to, reaction time, number of attempts, order of words, stuttering, omission of words, mispronunciation of words/syllables, length of response time, rate of speech, “swallowing” of words, ratio between mispronounced and correctly pronounced words, speech fluency, use of correct word types, grammar correctness, use of key words, number of correct attempts, length of utterance, pitch of speech, intensity of speech or any combination thereof. Each possibility is a separate embodiment. The speech quality may also be evaluated based on the level of similarity between the user's speech and the expected speech as determined, for example, based on a number of words which were as expected, a use of synonyms or homonyms, use of words from the same category or any combination thereof. Each possibility is a separate embodiment.
[0037] Once the speech quality was evaluated, for example, assigned a score, the system compares the user's speech quality to a predetermined threshold (step 110). If the user's speech quality is above/at the threshold, the system rewards the user with a positive game feature (step 112). Optionally, if the user's speech quality is below the threshold, the system may penalize the user with a negative game feature (step 114). Such reward system underlines the different game interactions, as disclosed herein. This unique approach may for example include using an Avatar, a living organism (such as a plant, pet, baby, etc.) who is nourished by the reward(s) obtain during practice. Practice adherence and success will lead to a prosperity of the Avatar ecosystem. On the other hand, failure to practice and/or to make progress will lead to it diminish. The users are expected to care about their Avatars (such as, but not limited to, Tamaguchi and Furby) and would not want to let them down. Advantageously, this will encourage users to adhere to their speech/language therapy practice protocol and to make a progress in their training.
[0038] It is appreciated that certain features of the disclosure, which are, for clarity, described in the context of separate embodiments, may also be provided in combination in a single embodiment. Conversely, various features of the disclosure, which are, for brevity, described in the context of a single embodiment, may also be provided separately or in any suitable sub-combination or as suitable in any other described embodiment of the disclosure. No feature described in the context of an embodiment is to be considered an essential feature of that embodiment, unless explicitly specified as such.
[0039] Although steps of methods according to some embodiments may be described in a specific sequence, methods of the disclosure may include some or all of the described steps carried out in a different order. A method of the disclosure may include a few of the steps described or all of the steps described. No particular step in a disclosed method is to be considered an essential step of that method, unless explicitly specified as such.
[0040] Although the disclosure is described in conjunction with specific embodiments thereof, it is evident that numerous alternatives, modifications and variations that are apparent to those skilled in the art may exist. Accordingly, the disclosure embraces all such alternatives, modifications and variations that fall within the scope of the appended claims. It is to be understood that the disclosure is not necessarily limited in its application to the details of construction and the arrangement of the components and/or methods set forth herein. Other embodiments may be practiced, and an embodiment may be carried out in various ways.
[0041] The phraseology and terminology employed herein are for descriptive purpose and should not be regarded as limiting. Citation or identification of any reference in this application shall not be construed as an admission that such reference is available as prior art to the disclosure. Section headings are used herein to ease understanding of the specification and should not be construed as necessarily limiting.