Predictive robotic controller apparatus and methods
IPC분류정보
국가/구분
United States(US) Patent
등록
국제특허분류(IPC7판)
G05B-019/18
B25J-009/16
출원번호
US-0918620
(2013-06-14)
등록번호
US-9314924
(2016-04-19)
발명자
/ 주소
Laurent, Patryk
Passot, Jean-Baptiste
Sinyavskiy, Oleg
Ponulak, Filip
Gabardos, Borja Ibarz
Izhikevich, Eugene
출원인 / 주소
Brain Corporation
대리인 / 주소
Gazdzinski & Associates, PC
인용정보
피인용 횟수 :
10인용 특허 :
83
초록▼
Robotic devices may be trained by a user guiding the robot along target action trajectory using an input signal. A robotic device may comprise an adaptive controller configured to generate control signal based on one or more of the user guidance, sensory input, performance measure, and/or other info
Robotic devices may be trained by a user guiding the robot along target action trajectory using an input signal. A robotic device may comprise an adaptive controller configured to generate control signal based on one or more of the user guidance, sensory input, performance measure, and/or other information. Training may comprise a plurality of trials, wherein for a given context the user and the robot's controller may collaborate to develop an association between the context and the target action. Upon developing the association, the adaptive controller may be capable of generating the control signal and/or an action indication prior and/or in lieu of user input. The predictive control functionality attained by the controller may enable autonomous operation of robotic devices obviating a need for continuing user guidance.
대표청구항▼
1. A method of training a computerized robotic apparatus having motorized operational elements to perform a target action based on a sensory context, the training being effectuated via a plurality of iterations, the method comprising: during a first portion of the plurality of iterations: causing th
1. A method of training a computerized robotic apparatus having motorized operational elements to perform a target action based on a sensory context, the training being effectuated via a plurality of iterations, the method comprising: during a first portion of the plurality of iterations: causing the apparatus to perform a first action based on a first control signal generated based on a sensory context and a target action indication provided by a user; andcausing the apparatus to adjust a controller learning parameter based on a first performance value determined based on the first action and the target action; andduring a second portion of the plurality of iterations, the second portion being subsequent to the first portion: causing the apparatus to generate an action signal based on the sensory context and the adjusted controller learning parameter, the action signal generation occurring prior to a subsequent provision of the target action indication associated with a subsequent iteration; andcausing the apparatus to perform a second action based on a second control signal generated based on the sensory context and the action signal but absent the provision of the target action indication by the user; and wherein:the action performance based on the second control signal is characterized by a second performance value, where the second performance value is improved compared to the first performance value;where the subsequent provision of the target action indication by the user during the subsequent iteration of the plurality of iterations is configured to cause execution of another action characterized by a third performance value;the third performance value is lower than the second performance value;where the target action, the first action, the second action, and the another action, each comprise rotating of the motorized operational elements of the computerized robotic apparatus; andwhere the first, second, and third performance are determined based at least in part on the sensory context. 2. The method of claim 1, wherein: the third performance value is determined by a controller; andthe third performance value is lower than the first performance value. 3. The method of claim 1, wherein: the controller learning parameter adjustment is adjusted based on a supervised learning process configured based on the sensory context and a combination of a control signal and a user input; andthe control signal comprises the first control signal or the second control signal. 4. The method of claim 3, wherein the user input comprises the target action indication. 5. The method of claim 1, wherein: the sensory context comprises an object representation; andthe target action comprises at least one of an object approach maneuver or an object avoidance maneuver. 6. The method of claim 1, wherein: individual ones of the first portion of the plurality of iterations are characterized by a time interval between an onset of the sensory context and the provision of the user input; anda time period between the onset of the sensory context during the second portion of the plurality of iterations and the action signal generation is no greater than a minimum value of the time interval. 7. The method of claim 1, wherein: individual ones of the first portion of the plurality of iterations are characterized by a time interval between an onset of the sensory context and the provision of the user input; anda time period between the onset of the sensory context during the second portion of the plurality of iterations and the action signal generation is no less than at least one of a mean value of the time interval or a median value of the time interval. 8. The method of claim 1, wherein: a concluding iteration of the first portion of the plurality of iterations is characterized by a concluding target action signal being provided by the user, the concluding iteration being the last iteration of the first portion of the plurality of iterations, the concluding target action being a target action associated with the concluding iteration; andthe generation of the action signal is characterized by the absence of another target action signal being provided by the user within a time period subsequent to time of the concluding target action signal provision and a time of the action signal generation. 9. The method of claim 1, wherein: the first control signal comprises a combination of the target action indication provided by the user and a predicted control signal generated by the apparatus;an action performance during individual ones of the first portion of the plurality of iterations is configured based on the combination of the target action indication provided by the user and the predicted control signal generated by the apparatus; andperforming the action based solely on the target action indication provided by the user or the predicted control signal is characterized by a fourth performance value and a fifth performance value, respectively, the fourth performance value and the fifth performance value being lower than the first performance value. 10. The method of claim 1, wherein: the first performance value is determined based on a first proximity measure between the first action and a target action; andthe second action is characterized by the second performance value determined based on a second proximity measure between the second action and the target action. 11. The method of claim 1, further comprising: analyzing individual ones of a plurality of first control signals comprises determining a deviation between the first action and the target action;wherein: the target action corresponds to operation of the apparatus based on the user; andthe analysis is configured to cause modification of a controller state in accordance with a learning process, the learning process being configured based on a performance measure. 12. The method of claim 1, further comprising: determining a predicted control output based on a characteristic of the sensory context the sensory context conveying information associated with one or both of an environment of the robotic apparatus and a platform state;wherein: the first control signal and the second control signal are configured based on the sensory signal; andthe first action is configured based on the predicted control output. 13. The method of claim 12, further comprising: providing a table configured to store a plurality of teaching inputs, a plurality of sensory signal characteristics, and a plurality of predicted control outputs; andselecting a given predicted control output based on a match between a given characteristic of the sensory context and an individual one of the plurality of sensory signal characteristics. 14. The method of claim 13, wherein the given predicted control output is configured based on a search of the table, the search being configured based on the plurality teaching inputs, individual ones of the plurality of teaching inputs comprising a target action indication provided by the user. 15. The method of claim 12, further comprising determining a combined output based on the given predicted control output and the target action indication provided by the user the combined output being characterized by a transform function configured to combine the predicted control output and the target action indication provided by the user via one or more operations including an additive operation. 16. The method of claim 15, wherein the transform function is configured to combine the predicted control output and the target action indication provided by the user via one or more operations including a union operation. 17. The method of claim 15, wherein: the adjusting the controller learning parameter is configured based on a supervised learning process configured based on the combined output;the supervised learning process is configured to be updated at a plurality of time intervals; andthe adjusting the controller learning parameter comprises determining an error measure between (i) the predicted control output generated at a first time instance and (ii) the target action indication provided by the user at second time instance subsequent to the first time instance, the first time instance and the second time instance separated by one of the time intervals. 18. The method of claim 1, wherein: the user comprises a human trainer;the second control signal is configured to cause a corrective action by the apparatus, the corrective action being characterized by a lower deviation from the target action;the corrective action is effectuated based on a cooperative interaction with the human, the cooperative interaction being characterized by a plurality of iterations;the first control signal corresponds to a first given iteration of the plurality of iterations; andthe second control signal corresponds to a second given iteration of the plurality of iterations, the second given iteration occurring subsequent to the first given iteration. 19. The method of claim 18, further comprising: determining a plurality of predicted control outputs based on a characteristic of a sensory signal and a given user input;wherein: the sensory signal is configured to convey information about one or both of an environment of the robotic apparatus and a plant state;the first action is configured based on a first predicted control output of the plurality of predicted control outputs, the first predicted control output corresponding to the first target action indication by the user;the corrective action is configured based on a second predicted control output of the plurality of predicted control outputs, the second predicted control output corresponding to the second target action indication by the user;the corrective action being characterized by an improved performance as compared to the first action and the target action;the improved performance being quantified based on a lower deviation of the corrected action from the target action compared to a deviation between the first action and the target action; andthe target action being based solely on the target action indication by the user absent a predicted control output of the plurality of predicted control outputs. 20. The method of claim 1, wherein the second control signal is generated by the controller responsive to a confirmation provided by the user.
연구과제 타임라인
LOADING...
LOADING...
LOADING...
LOADING...
LOADING...
이 특허에 인용된 특허 (83)
Ito, Masato; Minamino, Katsuki; Yoshiike, Yukiko; Suzuki, Hirotaka; Kawamoto, Kenta, Apparatus and method for embedding recurrent neural networks into the nodes of a self-organizing map.
DeYong Mark R. (Las Cruces NM) Findley Randall L. (Austin TX) Eskridge Thomas C. (Las Cruces NM) Fields Christopher A. (Rockville MD), Asynchronous temporal neural processing element.
Frank D. Francone ; Peter Nordin SE; Wolfgang Banzhaf DE, Computer implemented machine learning method and system including specifically defined introns.
Spoerre Julie K. (Tallahassee FL) Lin Chang-Ching (Tallahassee FL) Wang Hsu-Pin (Tallahassee FL), Machine performance monitoring and fault classification using an exponentially weighted moving average scheme.
Grossberg Stephen (Newton Highlands MA) Kuperstein Michael (Brookline MA), Massively parellel real-time network architectures for robots capable of self-calibrating their operating parameters thr.
Sakaue Shiyuki (Yokohama JPX) Sugimoto Koichi (Hiratsuka JPX) Arai Shinichi (Yokohama JPX), Method and apparatus for controlling a robot hand along a predetermined path.
Peltola Tero (Helsinki FIX) Matakselka Jorma (Vantaa FIX) Harju Esa (Espoo FIX) Salovuori Heikki (Helsinki FIX) Keskinen Jukka (Vantaa FIX) Makinen Kari (Helsinki FIX) Roikonen Olli (Espoo FIX), Method for congestion management in a frame relay network and a node in a frame relay network.
Wilson Charles L. (Darnestown MD) Garris Michael D. (Gaithersburg MD) Wilkinson ; Jr. Robert A. (Hyattstown MD), Object/anti-object neural network segmentation.
Yokono, Jun; Sabe, Kohtaro; Costa, Gabriel; Ohashi, Takeshi, Operational control method, program, and recording media for robot device, and robot device.
Eguchi, Toru; Yamada, Akihiro; Kusumi, Naohiro; Sekiai, Takaaki; Fukai, Masayuki; Shimizu, Satoru, Plant control system and thermal power generation plant control system.
Coenen, Olivier, Proportional-integral-derivative controller effecting expansion kernels comprising a plurality of spiking neurons associated with a plurality of receptive fields.
Hickman, Ryan; Kuffner, Jr., James J.; Bruce, James R.; Gharpure, Chaitanya; Kohler, Damon; Poursohi, Arshan; Francis, Jr., Anthony G.; Lewis, Thor, Shared robot knowledge base for use with cloud computing system.
Blumberg, Bruce; Brooks, Rodney; Buehler, Christopher J.; Deegan, Patrick A.; DiCicco, Matthew; Dye, Noelle; Ens, Gerry; Linder, Natan; Siracusa, Michael; Sussman, Michael; Williamson, Matthew M., Training and operating industrial robots.
Mochizuki, Yoshiyuki; Naka, Toshiya; Asahara, Shigeo, Virtual space control data receiving apparatus,virtual space control data transmission and reception system, virtual space control data receiving method, and virtual space control data receiving prog.
Meier, Philip; Passot, Jean-Baptiste; Ibarz Gabardos, Borja; Laurent, Patryk; Sinyavskiy, Oleg; O'Connor, Peter; Izhikevich, Eugene, Apparatus and methods for control of robot actions based on corrective user inputs.
※ AI-Helper는 부적절한 답변을 할 수 있습니다.