Trainable convolutional network apparatus and methods for operating a robotic vehicle
원문보기
IPC분류정보
국가/구분
United States(US) Patent
등록
국제특허분류(IPC7판)
G05B-019/04
G05B-019/18
B25J-009/16
G06N-003/00
G06N-003/04
G06N-003/08
출원번호
US-0265113
(2014-04-29)
등록번호
US-9346167
(2016-05-24)
발명자
/ 주소
O'Connor, Peter
Izhikevich, Eugene
출원인 / 주소
Brain Corporation
대리인 / 주소
Gazdzinski & Associates, PC
인용정보
피인용 횟수 :
2인용 특허 :
86
초록▼
A robotic vehicle may be operated by a learning controller comprising a trainable convolutional network configured to determine control signal based on sensory input. An input network layer may be configured to transfer sensory input into a hidden layer data using a filter convolution operation. Inp
A robotic vehicle may be operated by a learning controller comprising a trainable convolutional network configured to determine control signal based on sensory input. An input network layer may be configured to transfer sensory input into a hidden layer data using a filter convolution operation. Input layer may be configured to transfer sensory input into hidden layer data using a filter convolution. Output layer may convert hidden layer data to a predicted output using data segmentation and a fully connected array of efficacies. During training, efficacy of network connections may be adapted using a measure determined based on a target output provided by a trainer and an output predicted by the network. A combination of the predicted and the target output may be provided to the vehicle to execute a task. The network adaptation may be configured using an error back propagation method. The network may comprise an input reconstruction.
대표청구항▼
1. A method of operating a robotic device by a computerized neuron network comprising an input layer, an intermediate layer and an output layer of neurons, the method comprising: during one operation of a plurality of operations: causing the robotic device to execute an action along a first trajecto
1. A method of operating a robotic device by a computerized neuron network comprising an input layer, an intermediate layer and an output layer of neurons, the method comprising: during one operation of a plurality of operations: causing the robotic device to execute an action along a first trajectory in accordance with a first control signal determined based on a sensory input;determining, by the output layer, a performance measure based on an evaluation of the first trajectory and indication related to a target trajectory provided by a trainer;conveying information related to the performance measure to the input layer; andupdating one or more learning parameters of the input layer in accordance with the information; andduring a subsequent operation of a plurality of operations: causing the robotic device to execute the action along a second trajectory in accordance with a second control signal determined based on the sensory input;wherein: the execution of the action along the second trajectory is characterized by a second performance measure; andthe updating is configured to displace the second trajectory closer towards the target trajectory relative to the first trajectory. 2. The method of claim 1, wherein: the first control signal is based on a feature detected in the sensory input;one or more first nodes of the input layer process are configured to effectuate the detection of the feature; andthe updating is configured to modify one or more parameters associated with the one or more first nodes of the input layer. 3. The method of claim 2, wherein: one or more second nodes of the output layer are configured to produce the first control signal;the one or more first nodes of the input layer are coupled to the one or more second nodes of the output layer via a connectivity array of efficacies; andthe updating comprises a plasticity operation configured to modify one or more efficacies of the connectivity array of efficacies. 4. The method of claim 3, wherein: the one or more first nodes of the input layer are coupled to the one or more second nodes of the output layer via an-all to all connectivity pattern; andthe connectivity array of efficacies is characterized by a first dimension determined based on a first number of the one or more first nodes of the input layer and a second dimension determined based on a second number of the one or more second nodes of the output layer. 5. The method of claim 1, wherein: the performance measure comprises a first distance between the first trajectory and the target trajectory; andthe second performance measure comprises a second distance between the second trajectory and the target trajectory, the second distance being smaller than the first distance. 6. The method of claim 1, wherein: the performance measure comprises a first probability parameter between the first trajectory and the target trajectory; andthe second performance measure comprises a second probability between the second trajectory and the target trajectory, the second probability being greater than the first probability. 7. The method of claim 1, wherein: the computerized neuron network is configured for operation in accordance with a supervised learning process configured based on a teaching signal; andthe first control signal comprises a combination of the second layer output and a teaching signal provided to the robotic device. 8. A method of generating a predicted control output by an adaptive controller of a robotic apparatus comprising a predictor and a combiner, the method comprising: configuring the adaptive controller apparatus to detect an object in sensory input provided by a sensor of the robotic apparatus, the object detection causing generation of a control output based on a characteristic of the object;configuring the predictor to determine a predicted control output based on the characteristic of the object;configuring the combiner to determine a combined output based on a control input and the predicted control output, the combined output being characterized by a transform function;determining a performance measure based on the predicted control output and the combined output;updating one or more learning parameters of the adaptive controller in accordance with the performance measure; andconfiguring the adaptive controller to provide the combined output to the robotic apparatus, the combined output configured to cause the robotic apparatus to execute a maneuver in accordance with the characteristic of the object. 9. The method of claim 8, wherein: the object detection is effectuated by a first component of the adaptive controller; andthe generation of the control output is effectuated by a second component of the adaptive controller;the one or more learning parameters update uses a back propagation operation configured to convey the performance measure from the second component to the first component. 10. The method of claim 9, further comprising providing the control input by a training entity based on an evaluation of a trajectory associated with the maneuver executed by the robotic apparatus versus a target trajectory. 11. The method of claim 10, wherein the transform function is configured to combine the predicted control output and the control input via one or more operations including a union operation. 12. The method of claim 10, wherein the transform function is configured to combine the predicted control output and the control input via one or more operations including an additive operation. 13. The method of claim 10, wherein: the training entity comprises a computerized apparatus operable in accordance with a reinforcement learning process; andthe predictor is operable in accordance with a supervised learning process configured based on a teaching signal configured based on the combined output. 14. The method of claim 9, wherein: the robotic apparatus comprises a vehicle;the sensory input comprises a video stream of vehicle surroundings;the characteristic of the object comprises a representation of an obstacle or a target within the vehicle surroundings; andthe maneuver comprises one of an obstacle avoidance or a target approach. 15. A method of operating a robotic device using a computerized neuron network having a plurality of layers of neurons, the method comprising: causing the robotic device to execute an action along a first trajectory in accordance with a first control signal, the first signal determined based at least on a sensory input;determining a performance measure based on an evaluation of the first trajectory relative to a target trajectory;updating one or more learning parameters of a first of the plurality of layers in accordance with information relating to the determined performance measure; andcausing the robotic device to execute the action along a second trajectory in accordance with a second control signal, the second signal determined based at least on a sensory input and the updated one or more learning parameters, the second trajectory being closer to the target trajectory than the first trajectory. 16. A computerized neuron network apparatus configured to provide a response based on analysis of visual input frames, the computerized neuron network apparatus comprising: an input component comprising first portion of neurons configured to implement a convolutional operation on the visual input frames using a plurality of filter masks, the operation being configured to produce convolved input frames;an output component comprising at least one output neuron configured to provide an output based on the at least one output neuron reaching a target state;a connection component configured to couple the input component to the at least one output neuron via an efficacy array; anda cost estimation component configured to determine a first similarity measure between a given response and a target response;wherein: the given response is configured based on the output;the first similarity measure determined based on a first analysis of a first frame of the visual input frames is configured to cause an update of the neuron network, the update of the neuron network being configured to increase a second similarity measure determined based on a second analysis of a second frame of the visual input frames subsequent to the first frame;the convolved input frames are configured to enable detection of an object;the output is configured based on the detected object; andthe given response is configured to be provided to the computerized neuron network apparatus, the given response being configured to cause the computerized neuron network apparatus to execute a first action in accordance with the detected object. 17. The computerized neuron network apparatus of claim 16, wherein the update comprises: determination of a discrepancy parameter based on the first similarity measure;modification of one or more efficacies of the efficacy array using the discrepancy parameter;backward propagation of the discrepancy parameter from the output component to the input component using a gradient operation with respect to a filter mask parameter; andmodification of the filter mask parameter based on an outcome of the gradient operation. 18. The computerized neuron network apparatus of claim 17, further comprising: an input reconstruction component configured to produce reconstructed input frames based on the convolved input frames and a de-convolution filter mask;wherein the update further comprises: a second modification of the filter mask parameter based on an evaluation of the convolved input frames and a first reconstructed input frame, the second modification being configured to reduce the discrepancy for another response generated based on a third frame subsequent to the second frame. 19. The computerized neuron network apparatus of claim 18, wherein: a first response generated in an absence of input reconstruction is characterized by a first value of the second similarity measure;a second response generated based on the input reconstruction is characterized by a second value of the second similarity measure, the second value being greater than the first value. 20. The computerized neuron network apparatus of claim 19, wherein the computerized neuron network apparatus is configured to determine that the given response is closer to the target response based at least on the second value being greater than the first value. 21. The computerized neuron network apparatus of claim 20, further comprising a combiner component configured to determine a combined output based at least on the output and a correction signal, the combined output being characterized by a transform function; wherein the determination of the first similarity measure is configured based on the output and the combined output. 22. The computerized neuron network apparatus of claim 21, wherein the first action is selected from the group comprising: object avoidance actions and object approach actions.
연구과제 타임라인
LOADING...
LOADING...
LOADING...
LOADING...
LOADING...
이 특허에 인용된 특허 (86)
Ito, Masato; Minamino, Katsuki; Yoshiike, Yukiko; Suzuki, Hirotaka; Kawamoto, Kenta, Apparatus and method for embedding recurrent neural networks into the nodes of a self-organizing map.
DeYong Mark R. (Las Cruces NM) Findley Randall L. (Austin TX) Eskridge Thomas C. (Las Cruces NM) Fields Christopher A. (Rockville MD), Asynchronous temporal neural processing element.
Kerr Randal H. (Richford NY) Mesnard Robert M. (Endicott NY), Automatic generation of executable computer code which commands another program to perform a task and operator modificat.
Frank D. Francone ; Peter Nordin SE; Wolfgang Banzhaf DE, Computer implemented machine learning method and system including specifically defined introns.
Spoerre Julie K. (Tallahassee FL) Lin Chang-Ching (Tallahassee FL) Wang Hsu-Pin (Tallahassee FL), Machine performance monitoring and fault classification using an exponentially weighted moving average scheme.
Grossberg Stephen (Newton Highlands MA) Kuperstein Michael (Brookline MA), Massively parellel real-time network architectures for robots capable of self-calibrating their operating parameters thr.
Sakaue Shiyuki (Yokohama JPX) Sugimoto Koichi (Hiratsuka JPX) Arai Shinichi (Yokohama JPX), Method and apparatus for controlling a robot hand along a predetermined path.
Peltola Tero (Helsinki FIX) Matakselka Jorma (Vantaa FIX) Harju Esa (Espoo FIX) Salovuori Heikki (Helsinki FIX) Keskinen Jukka (Vantaa FIX) Makinen Kari (Helsinki FIX) Roikonen Olli (Espoo FIX), Method for congestion management in a frame relay network and a node in a frame relay network.
Wilson Charles L. (Darnestown MD) Garris Michael D. (Gaithersburg MD) Wilkinson ; Jr. Robert A. (Hyattstown MD), Object/anti-object neural network segmentation.
Yokono, Jun; Sabe, Kohtaro; Costa, Gabriel; Ohashi, Takeshi, Operational control method, program, and recording media for robot device, and robot device.
Eguchi, Toru; Yamada, Akihiro; Kusumi, Naohiro; Sekiai, Takaaki; Fukai, Masayuki; Shimizu, Satoru, Plant control system and thermal power generation plant control system.
Coenen, Olivier, Proportional-integral-derivative controller effecting expansion kernels comprising a plurality of spiking neurons associated with a plurality of receptive fields.
Hickman, Ryan; Kuffner, Jr., James J.; Bruce, James R.; Gharpure, Chaitanya; Kohler, Damon; Poursohi, Arshan; Francis, Jr., Anthony G.; Lewis, Thor, Shared robot knowledge base for use with cloud computing system.
Blumberg, Bruce; Brooks, Rodney; Buehler, Christopher J.; Deegan, Patrick A.; DiCicco, Matthew; Dye, Noelle; Ens, Gerry; Linder, Natan; Siracusa, Michael; Sussman, Michael; Williamson, Matthew M., Training and operating industrial robots.
Mochizuki, Yoshiyuki; Naka, Toshiya; Asahara, Shigeo, Virtual space control data receiving apparatus,virtual space control data transmission and reception system, virtual space control data receiving method, and virtual space control data receiving prog.
※ AI-Helper는 부적절한 답변을 할 수 있습니다.