Patent classifications
G10L15/183
Discrete Three-Dimensional Processor
A discrete three-dimensional (3-D) processor comprises stacked first and second dice. The first die comprises three-dimensional memory (3D-M) arrays, whereas the second die comprises at least a portion of a logic/processing circuit and an off-die peripheral-circuit component of the 3D-M array(s). The preferred 3-D processor can be used to compute non-arithmetic function/model. In other applications, the preferred 3-D processor may also be a 3-D configurable computing array, a 3-D pattern processor, or a 3-D neuro-processor.
Discrete Three-Dimensional Processor
A discrete three-dimensional (3-D) processor comprises stacked first and second dice. The first die comprises three-dimensional memory (3D-M) arrays, whereas the second die comprises at least a portion of a logic/processing circuit and an off-die peripheral-circuit component of the 3D-M array(s). The preferred 3-D processor can be used to compute non-arithmetic function/model. In other applications, the preferred 3-D processor may also be a 3-D configurable computing array, a 3-D pattern processor, or a 3-D neuro-processor.
Natural language processing routing
Devices and techniques are generally described for a speech processing routing architecture. In various examples, first data comprising a first feature definition is received. The first feature definition may include a first indication of first source data and first instructions for generating feature data using the first source data. In various examples, the feature data may be generated according to the first feature definition. In some examples, a speech processing system may receive a first request to process a first utterance. The feature data may be retrieved from a non-transitory computer-readable memory. The speech processing system may determine a first skill for processing the first utterance based at least in part on the feature data.
Natural language processing routing
Devices and techniques are generally described for a speech processing routing architecture. In various examples, first data comprising a first feature definition is received. The first feature definition may include a first indication of first source data and first instructions for generating feature data using the first source data. In various examples, the feature data may be generated according to the first feature definition. In some examples, a speech processing system may receive a first request to process a first utterance. The feature data may be retrieved from a non-transitory computer-readable memory. The speech processing system may determine a first skill for processing the first utterance based at least in part on the feature data.
SPEECH SEPARATION AND RECOGNITION METHOD FOR CALL CENTERS
The present invention provides a method for speech separation and recognition. The present invention overcomes the disadvantages of the existing techniques by providing automatic speech recognition and separation that helps managers see what their service agents and customers are saying. From there, quickly and objectively knowing the wishes and concerns of customers as well as whether their service agents can give accurate and correct advice. In addition, the system is constantly updated based on the semi-supervised training mechanism, which means that the system can self-learn from actual data during operation, thereby helping to improve the system's accuracy.
Electronic device and method for controlling the electronic device thereof
An electronic device is provided. The electronic device includes a memory configured to store a speech translation model and at least one processor electronically connected with the memory. The at least one processor is configured to train the speech translation model based on first information related to conversion between a speech in a first language and a text corresponding to the speech in the first language, and second information related to conversion between a text in the first language and a text in a second language corresponding to the text in the first language, and the speech translation model is trained to convert a speech in the first language into a text in the second language and output the text.
Electronic device and method for controlling the electronic device thereof
An electronic device is provided. The electronic device includes a memory configured to store a speech translation model and at least one processor electronically connected with the memory. The at least one processor is configured to train the speech translation model based on first information related to conversion between a speech in a first language and a text corresponding to the speech in the first language, and second information related to conversion between a text in the first language and a text in a second language corresponding to the text in the first language, and the speech translation model is trained to convert a speech in the first language into a text in the second language and output the text.
Method and apparatus for determining output token
A method for determining an output token includes predicting a first probability of each of candidate output tokens of a first model, predicting a second probability of each of the candidate output tokens of a second model interworking with the first model, adjusting the second probability of each of the candidate output tokens based on the first probability, and determining the output token among the candidate output tokens based on the first probability and the adjusted second probability.
Contextual biasing of neural language models using metadata from a natural language understanding component and embedded recent history
Techniques for implementing a chatbot that utilizes context embeddings are described. An exemplary method includes determining a next turn by: applying a language model to the utterance to determine a probability of a sequence of words, generating a context embedding for the utterance based at least on one or more of: a dialog act as defined by a chatbot definition of the chatbot, a topic vector identifying a domain of the chatbot, a previous chatbot response, and one or more slot options; performing neural language model rescoring using the determined probability of a sequence of words as a word embedding and the generated context embedding to predict an hypothesis; determining at least a name of a slot and type to be fulfilled based at least in part on the hypothesis and the chatbot definition; and determining a next turn based at least in part on the chatbot definition, any previous state, and the name of the slot and type to be fulfilled.
Contextual biasing of neural language models using metadata from a natural language understanding component and embedded recent history
Techniques for implementing a chatbot that utilizes context embeddings are described. An exemplary method includes determining a next turn by: applying a language model to the utterance to determine a probability of a sequence of words, generating a context embedding for the utterance based at least on one or more of: a dialog act as defined by a chatbot definition of the chatbot, a topic vector identifying a domain of the chatbot, a previous chatbot response, and one or more slot options; performing neural language model rescoring using the determined probability of a sequence of words as a word embedding and the generated context embedding to predict an hypothesis; determining at least a name of a slot and type to be fulfilled based at least in part on the hypothesis and the chatbot definition; and determining a next turn based at least in part on the chatbot definition, any previous state, and the name of the slot and type to be fulfilled.