Patent classifications
G05B2219/32334
Cleanup support system, cleanup support method, and recording medium
A cleanup support system that supports a cleanup behavior includes: a first obtaining unit configured to obtain first information indicating a level of interest of a target person in cleanup; a second obtaining unit configured to obtain second information indicating a level of achievement of the cleanup performed by the target person; a determination unit configured to determine a content of control corresponding to the first information obtained and the second information obtained, with reference to a rule which associates the level of interest in the cleanup and the level of achievement of the cleanup with a content of control performed on a device; and a control unit configured to control the device according to the content of control determined.
Reinforcement learning for chatbots
A computer-implemented method for generating and deploying a reinforced learning model to train a chatbot. The method includes selecting a plurality of conversations, wherein each conversation includes an agent and a user. The method includes identifying, in each of the conversations, a set of turns and on or more topics. The method further includes associating one or more topics to each turn of the set of turns. The method includes, generating a conversation flow for each conversation, wherein the conversation flow identifies a sequence of the topics. The method includes applying an outcome score to each conversation. The method includes creating a reinforced learning (RL) model, wherein the RL model includes a Markov is based on the conversation flow of each conversation and the outcome score of each conversation. The method includes deploying the RL model, wherein the deploying includes sending the RL model to a chatbot.
ADAPTIVE-LEARNING INTELLIGENT SCHEDULING UNIFIED COMPUTING FRAME AND SYSTEM FOR INDUSTRIAL PERSONALIZED CUSTOMIZED PRODUCTION
The present invention discloses an adaptive-learning intelligent scheduling unified computing frame and system for industrial personalized customized production. Based on a deep neural network and reinforcement learning, a most suitable optimization algorithm is selected by automatic decision-making for a global customized production task with an industrial big data module at the bottom as an information basis, and a global optimal static scheduling plane is generated; a current dynamic event in a factory are monitored in real time; if no dynamic event requiring dynamic scheduling optimization is monitored, the global optimal static plan is executed sequentially; when a dynamic event impact requiring dynamic scheduling optimization is monitored, information of the current dynamic event is interpreted and classified, and corresponding optimization algorithms are automatically selected for dynamic scheduling optimization according to different types of dynamic events; and a dynamic scheduling scheme is evaluated by a subsequent module, an optimization scheme is regenerated or a most suitable optimization algorithm is automatically decided based on the scheme according to an evaluation result, and an equipment deployment sequence is generated for an automatic deployment. Considering the features of complicated procedures, a large amount of customization information and the high frequency of diversified dynamic events in personalized customized production, the present invention provides the adaptive-learning intelligent scheduling unified computing frame and system for industrial personalized customized production, that adopt two steps in the three aspects of static scheduling planning, dynamic scheduling planning and equipment deployment based on deep learning, that is, targeted optimization is performed after classification, thus improving the optimization efficiency and effect; and the system better fits the features of personalized customized production, and can effectively improve the efficiency of personalized customized production and minimize manual decision-making costs.
Machine learning device, robot system, and machine learning method for learning operation program of robot
A machine learning device, which learns an operation program of a robot, includes a state observation unit which observes as a state variable at least one of a shaking of an arm of the robot and a length of an operation trajectory of the arm of the robot; a determination data obtaining unit which obtains as determination data a cycle time in which the robot performs processing; and a learning unit which learns the operation program of the robot based on an output of the state observation unit and an output of the determination data obtaining unit.
SYSTEMS AND METHODS FOR TRAINING A REINFORCEMENT LEARNING SYSTEM FOR PALLET ROUTING IN A MANUFACTURING ENVIRONMENT
A method includes obtaining state information from one or more sensors of a digital twin. The method includes determining an action at the first routing control location based on the state information, where the action includes one of a pallet merging operation and a pallet splitting operation, and determining a consequence state based on the action. The method includes calculating a transient production value based on the consequence state and a transient objective function, calculating a steady state production value based on the consequence state and a steady state objective function, and selectively adjusting one or more reinforcement parameters of the reinforcement learning system based on the transient production value and the steady state production value.
PRODUCTION SYSTEM THAT SETS DETERMINATION VALUE OF VARIABLE RELATING TO ABNORMALITY OF PRODUCT
A production system includes a cell control apparatus that is connected to a machine control apparatus and an inspection control apparatus. The cell control apparatus includes a storage part that stores data on a state of the manufacturing machine, data on an environment state, and data on an inspection result of a product. The cell control apparatus includes a correlation analysis part that selects a variable relating to an abnormality based on a correlation between the inspection result of the product, and the data on the state of the manufacturing machine and the data on the environment state when the abnormality occurs in the inspection result, and a determination value setting part that sets a determination value of the variable relating to the abnormality.
A Controlling Method and Device for an Industrial Device
Various embodiments include methods for controlling an industrial device. Some embodiments include: obtaining a state input characterizing a current state of the industrial device; processing the state input to generate an action output characterizing an expected action to be performed by the industrial device for the current state, based on a machine learning model trained based on states of the industrial device, actions each performed for each state of the industrial device and results each obtained by performing each action; and generating a control signal for the industrial device based on the action output.
Reinforcement learning and simulation based dispatching method in a factory, and an apparatus thereof
Provided is a dispatching method in a factory based on reinforcement learning. The dispatching method in a factory based on reinforcement learning may comprise: constructing a Markov decision process (MDP) for dispatching actions of a dispatcher in the factory and resulting rewards and states of the factory; performing learning by applying reinforcement learning (RL) to the constructed MDP; and as a result of said RL, selecting a job that maximizes a weighted sum of a plurality of scored dispatching rules.
METHODS AND DEVICES FOR A COLLABORATION OF AUTOMATED AND AUTONOMOUS MACHINES
A device may include a processor configured to determine a layout for a plurality of automated machine clusters to be deployed in an environment based on a plurality of operation policies and an input task, wherein each operation policy provides a policy to operate one or more automated machines of one of the plurality of automated machine clusters for a trained task based on one or more policy parameters. The processor may further be configured to adjust the one or more policy parameters of at least one of the plurality of operation policies based on the determined layout in the environment and the input task.
System and method of determining processing condition
A system for determining a processing procedure including a plurality of processes for controlling an object, the system includes a learning unit for performing a learning process for determining a processing condition of each of a plurality of processes, and the learning unit acquires a physical quantity correlating with a state of the object on which a process has been performed under a predetermined processing condition, from a device for controlling the object on the basis of the processing procedure, calculates a pseudo state corresponding to the state of the object on the basis of the physical quantity, performs a learning process using a value function, and determines a processing condition of each of the plurality of processes to achieve a target state of the object.