Patents by Inventor Kenji Doya
Kenji Doya has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Patent number: 11645574Abstract: A non-transitory, computer-readable recording medium stores therein a reinforcement learning program that uses a value function and causes a computer to execute a process comprising: estimating first coefficients of the value function represented in a quadratic form of inputs at times in the past than a present time and outputs at the present time and the times in the past, the first coefficients being estimated based on inputs at the times in the past, the outputs at the present time and the times in the past, and costs or rewards that corresponds to the inputs at the times in the past; and determining second coefficients that defines a control law, based on the value function that uses the estimated first coefficients and determining input values at times after estimation of the first coefficients.Type: GrantFiled: September 13, 2018Date of Patent: May 9, 2023Assignees: FUJITSU LIMITED KAWASAKI, JAPAN, OKINAWA INSTITUTE OF SCIENCE AND TECHNOLOGY SCHOOL CORPORATIONInventors: Tomotake Sasaki, Eiji Uchibe, Kenji Doya, Hirokazu Anai, Hitoshi Yanami, Hidenao Iwane
-
Patent number: 11573537Abstract: A non-transitory, computer-readable recording medium stores a program of reinforcement learning by a state-value function. The program causes a computer to execute a process including calculating a temporal difference (TD) error based on an estimated state-value function, the TD error being calculated by giving a perturbation to each component of a feedback coefficient matrix that provides a policy; calculating based on the TD error and the perturbation, an estimated gradient function matrix acquired by estimating a gradient function matrix of the state-value function with respect to the feedback coefficient matrix for a state of a controlled object, when state variation of the controlled object in the reinforcement learning is described by a linear difference equation and an immediate cost or an immediate reward of the controlled object is described in a quadratic form of the state and an input; and updating the feedback coefficient matrix using the estimated gradient function matrix.Type: GrantFiled: September 13, 2018Date of Patent: February 7, 2023Assignees: FUJITSU LIMITED, OKINAWA INSTITUTE OF SCIENCE AND TECHNOLOGY SCHOOL CORPORATIONInventors: Tomotake Sasaki, Eiji Uchibe, Kenji Doya, Hirokazu Anai, Hitoshi Yanami, Hidenao Iwane
-
Patent number: 10896382Abstract: A method of inverse reinforcement learning for estimating cost and value functions of behaviors of a subject includes acquiring data representing changes in state variables that define the behaviors of the subject; applying a modified Bellman equation given by Eq. (1) to the acquired data: q(x)+gV(y)?V(x)=?ln{pi(y|x))/(p(y|x)} (1) where q(x) and V(x) denote a cost function and a value function, respectively, at state x, g represents a discount factor, and p(y|x) and pi(y|x) denote state transition probabilities before and after learning, respectively; estimating a density ratio pi(y|x)/p(y|x) in Eq. (1); estimating q(x) and V(x) in Eq. (1) using the least square method in accordance with the estimated density ratio pi(y|x)/p(y|x), and outputting the estimated q(x) and V(x).Type: GrantFiled: August 7, 2015Date of Patent: January 19, 2021Assignee: OKINAWA INSTITUTE OF SCIENCE AND TECHNOLOGY SCHOOL CORPORATIONInventors: Eiji Uchibe, Kenji Doya
-
Patent number: 10896383Abstract: A method of inverse reinforcement learning for estimating reward and value functions of behaviors of a subject includes: acquiring data representing changes in state variables that define the behaviors of the subject; applying a modified Bellman equation given by Eq. (1) to the acquired data: r ? ( x ) + ? ? ? V ? ( y ) - V ? ( x ) = ? ln ? ? ? ? ( y | x ) b ? ( y | x ) , ? ( 1 ) = ? ln ? ? ? ? ( x , y ) b ? ( x , y ) - ln ? ? ? ? ( x ) b ? ( x ) , ? ( 2 ) where r(x) and V(x) denote a reward function and a value function, respectively, at state x, and ? represents a discount factor, and b(y|x) and ?(y|x) denote state transition probabilities before and after learning, respectively; estimating a logarithm of the density ratio ?(x)/b(x) in Eq. (2); estimating r(x) and V(x) in Eq.Type: GrantFiled: February 6, 2017Date of Patent: January 19, 2021Assignee: OKINAWA INSTITUTE OF SCIENCE AND TECHNOLOGY SCHOOL CORPORATIONInventors: Eiji Uchibe, Kenji Doya
-
Publication number: 20190087751Abstract: A non-transitory, computer-readable recording medium stores therein a reinforcement learning program that uses a value function and causes a computer to execute a process comprising: estimating first coefficients of the value function represented in a quadratic form of inputs at times in the past than a present time and outputs at the present time and the times in the past, the first coefficients being estimated based on inputs at the times in the past, the outputs at the present time and the times in the past, and costs or rewards that corresponds to the inputs at the times in the past; and determining second coefficients that defines a control law, based on the value function that uses the estimated first coefficients and determining input values at times after estimation of the first coefficients.Type: ApplicationFiled: September 13, 2018Publication date: March 21, 2019Applicants: FUJITSU LIMITED, Okinawa Institute of Science and Technology School CorporationInventors: Tomotake Sasaki, Eiji Uchibe, Kenji Doya, Hirokazu Anai, Hitoshi Yanami, Hidenao Iwane
-
Publication number: 20190086876Abstract: A non-transitory, computer-readable recording medium stores a program of reinforcement learning by a state-value function. The program causes a computer to execute a process including calculating a TD error based on an estimated state-value function, the TD error being calculated by giving a perturbation to each component of a feedback coefficient matrix that provides a policy; calculating based on the TD error and the perturbation, an estimated gradient function matrix acquired by estimating a gradient function matrix of the state-value function with respect to the feedback coefficient matrix for a state of a controlled object, when state variation of the controlled object in the reinforcement learning is described by a linear difference equation and an immediate cost or an immediate reward of the controlled object is described in a quadratic form of the state and an input; and updating the feedback coefficient matrix using the estimated gradient function matrix.Type: ApplicationFiled: September 13, 2018Publication date: March 21, 2019Applicants: FUJITSU LIMITED, Okinawa Institute of Science and Technology School CorporationInventors: Tomotake Sasaki, Eiji Uchibe, Kenji Doya, Hirokazu Anai, Hitoshi Yanami, Hidenao Iwane
-
Publication number: 20170213151Abstract: A method of inverse reinforcement learning for estimating cost and value functions of behaviors of a subject includes acquiring data representing changes in state variables that define the behaviors of the subject; applying a modified Bellman equation given by Eq. (1) to the acquired data: q(x)+gV(y)?V(x)=?1n{pi(y|x))/(p(y|x)} (1) where q(x) and V(x) denote a cost function and a value function, respectively, at state x, g represents a discount factor, and p(y|x) and pi(y|x) denote state transition probabilities before and after learning, respectively; estimating a density ratio pi(y|x)/p(y|x) in Eq. (1); estimating q(x) and V(x) in Eq. (1) using the least square method in accordance with the estimated density ratio pi(y|x)/p(y|x), and outputting the estimated q(x) and V(x).Type: ApplicationFiled: August 7, 2015Publication date: July 27, 2017Applicant: Okinawa Institute of Science and Technology School CorporationInventors: Eiji UCHIBE, Kenji DOYA
-
Publication number: 20170147949Abstract: A method of inverse reinforcement learning for estimating reward and value functions of behaviors of a subject includes: acquiring data representing changes in state variables that define the behaviors of the subject; applying a modified Bellman equation given by Eq. (1) to the acquired data: r ? ( x ) + ? ? ? V ? ( y ) - V ? ( x ) = ? ln ? ? ? ? ( y | x ) b ? ( y | x ) , ? ( 1 ) = ? ln ? ? ? ? ( x , y ) b ? ( x , y ) - ln ? ? ? ? ( x ) b ? ( x ) , ? ( 2 ) where r(x) and V(x) denote a reward function and a value function, respectively, at state x, and ? represents a discount factor, and b(y|x) and ?(y|x) denote state transition probabilities before and after learning, respectively; estimating a logarithm of the density ratio ?(x)/b(x) in Eq. (2); estimating r(x) and V(x) in Eq.Type: ApplicationFiled: February 6, 2017Publication date: May 25, 2017Applicant: Okinawa Institute of Science and Technology School CorporationInventors: Eiji UCHIBE, Kenji DOYA
-
Patent number: 7170495Abstract: A key inputting device includes a vowel switch for inputting vowels, and consonant switches for inputting consonants. The vowel switch is displacable in five directions, each consonant switch is displacable in three directions, and displacement directions of each switch are allotted to each letter of alphabet corresponding to at least a movement of articulatory organs when pronouncing each letter, that is, a movement or a location of a jaw, a throat, a tongue, lips. The vowel switch is operationable by a thumb, and the consonant switches are operationable by an index finger, a middle finger, a ring finger, and a little finger, respectively.Type: GrantFiled: March 5, 2003Date of Patent: January 30, 2007Assignee: Advanced Telecommunications Research Institute InternationalInventor: Kenji Doya
-
Publication number: 20040178992Abstract: A key inputting device includes a vowel switch for inputting vowels, and consonant switches for inputting consonants. The vowel switch is displacable in five directions, each consonant switch is displacable in three directions, and displacement directions of each switch are allotted to each letter of alphabet corresponding to at least a movement of articulatory organs when pronouncing each letter, that is, a movement or a location of a jaw, a throat, a tongue, lips. The vowel switch is operationable by a thumb, and the consonant switches are operationable by an index finger, a middle finger, a ring finger, and a little finger, respectively.Type: ApplicationFiled: January 5, 2004Publication date: September 16, 2004Inventor: Kenji Doya
-
Patent number: 6529887Abstract: The invention provides a novel highly-adaptive agent learning machine comprising a plurality of learning modules each having a set of reinforcement learning system which works on an environment and determines an action output for maximizing a reward provided as a result thereof and an environment predicting system which predicts a change in the environment, wherein a responsibility signal is calculated such that the smaller a prediction error of the environment predicting system of each of the learning modules, the larger the value thereof, and the action output by the reinforcement learning system is weighted in proportion to the responsibility signal, thereby providing an action with regard to the environment. The machine switches and combines actions optimum to various states or operational modes of an environment without using any specific teacher signal and performs behavior learning flexibly without using any prior knowledge.Type: GrantFiled: May 18, 2000Date of Patent: March 4, 2003Assignees: Agency of Industrial Science and Technology, Advanced Telecommunication Research Institute InternationalInventors: Kenji Doya, Mitsuo Kawato