Patents by Inventor Naoyuki Kanda
Naoyuki Kanda has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Patent number: 11068072Abstract: In a mixed reality display system, a server and a plurality of mixed reality display terminals are connected, and virtual objects are displayed. The virtual objects include a shared virtual object for which a plurality of terminals have an operation authority and a private virtual object for which only a specific terminal has an operation authority. The server has virtual object attribute information for displaying the virtual objects in each terminal, and each terminal has a motion detecting unit that detects a motion of a user for switching between the shared virtual object and the private virtual object. When a detection result by the motion detecting unit is received from the terminal, the server updates the virtual object attribute information depending on whether the virtual object is the shared virtual object or the private virtual object, and transmits data of the virtual object after the update to each terminal.Type: GrantFiled: August 11, 2020Date of Patent: July 20, 2021Assignee: MAXELL, LTD.Inventor: Naoyuki Kanda
-
Patent number: 10909976Abstract: A speech recognition device includes: an acoustic model based on an End-to-End neural network responsive to an observed sequence formed of prescribed acoustic features obtained from a speech signal by feature extracting unit, for calculating probability of the observed sequence being a certain symbol sequence; and a decoder responsive to a symbol sequence candidate, for decoding a speech signal by a WFST based on a posterior probability of each of word sequences corresponding to the symbol sequence candidate, probabilities calculated by the acoustic model for symbol sequences selected based on an observed sequence, and a posterior probability of each of the plurality of symbol sequences.Type: GrantFiled: June 2, 2017Date of Patent: February 2, 2021Assignee: National Institute of Information and Communications TechnologyInventor: Naoyuki Kanda
-
Publication number: 20200371602Abstract: In a mixed reality display system, a server and a plurality of mixed reality display terminals are connected, and virtual objects are displayed. The virtual objects include a shared virtual object for which a plurality of terminals have an operation authority and a private virtual object for which only a specific terminal has an operation authority. The server has virtual object attribute information for displaying the virtual objects in each terminal, and each terminal has a motion detecting unit that detects a motion of a user for switching between the shared virtual object and the private virtual object. When a detection result by the motion detecting unit is received from the terminal, the server updates the virtual object attribute information depending on whether the virtual object is the shared virtual object or the private virtual object, and transmits data of the virtual object after the update to each terminal.Type: ApplicationFiled: August 11, 2020Publication date: November 26, 2020Inventor: Naoyuki KANDA
-
Patent number: 10775897Abstract: In a mixed reality display system, a server and a plurality of mixed reality display terminals are connected, and virtual objects are displayed. The virtual objects include a shared virtual object for which a plurality of terminals have an operation authority and a private virtual object for which only a specific terminal has an operation authority. The server has virtual object attribute information for displaying the virtual objects in each terminal, and each terminal has a motion detecting unit that detects a motion of a user for switching between the shared virtual object and the private virtual object. When a detection result by the motion detecting unit is received from the terminal, the server updates the virtual object attribute information depending on whether the virtual object is the shared virtual object or the private virtual object, and transmits data of the virtual object after the update to each terminal.Type: GrantFiled: June 6, 2017Date of Patent: September 15, 2020Assignee: Maxell, Ltd.Inventor: Naoyuki Kanda
-
Publication number: 20200241647Abstract: In a mixed reality display system, a server and a plurality of mixed reality display terminals are connected, and virtual objects are displayed. The virtual objects include a shared virtual object for which a plurality of terminals have an operation authority and a private virtual object for which only a specific terminal has an operation authority. The server has virtual object attribute information for displaying the virtual objects in each terminal, and each terminal has a motion detecting unit that detects a motion of a user for switching between the shared virtual object and the private virtual object. When a detection result by the motion detecting unit is received from the terminal, the server updates the virtual object attribute information depending on whether the virtual object is the shared virtual object or the private virtual object, and transmits data of the virtual object after the update to each terminal.Type: ApplicationFiled: June 6, 2017Publication date: July 30, 2020Inventor: Naoyuki KANDA
-
Publication number: 20200167003Abstract: A head-mounted display (HMD) 1, which is operated by a gesture operation performed by a user 3, is provided with a distance image acquisition unit 106 that detects a gesture operation, a position information acquisition unit 103 that acquires position information of the HMD 1, and a communication unit 2 that performs communication with another HMD 1?. A control unit 205 sets and displays an operating space 600 where a gesture operation performed by the user 3 is valid, exchanges position information and operating space information of the host HMD 1 and the other HMD 1 therebetween by the communication unit 2, and adjusts the operating space of the host HMD so that the operating space 600 and an operating space 600? of the other HMD 1 do not overlap each other.Type: ApplicationFiled: August 24, 2017Publication date: May 28, 2020Inventors: Yo NONOMURA, Naoyuki KANDA
-
Patent number: 10607602Abstract: An object is to provide a speech recognition device with improved recognition accuracy using characteristics of a neural network. A speech recognition device includes: an acoustic model 308 implemented by a RNN (recurrent neural network) for calculating, for each state sequence, the 45 posterior probability of a state sequence in response to an observed sequence consisting of prescribed speech features obtained from a speech; a WFST 320 based on S-1HCLG calculating, for each word sequence, posterior probability of a word sequence in response to a state sequence; and a hypothesis selecting unit 322, performing speech recognition of the speech signal based on a score calculated for each hypothesis of a 50 word sequence corresponding to the speech signal, using the posterior probabilities calculated by the acoustic model 308 and the WFST 320 for the input observed sequence.Type: GrantFiled: May 10, 2016Date of Patent: March 31, 2020Assignee: NATIONAL INSTITUTE OF INFORMATION AND COMMUNICATIONS TECHNOLOGYInventor: Naoyuki Kanda
-
Patent number: 10489451Abstract: Provided is a voice search technology that can efficiently find and check a problematic call. To this end, a voice search system of the present invention includes a call search database that stores, for each of a reception channel and a transmission channel of each of a plurality of pieces of recorded call voice data, voice section sequences in association with predetermined keywords and time information. The call search database is searched based on an input search keyword, so that a voice section sequence that contains the search keyword is obtained.Type: GrantFiled: September 11, 2013Date of Patent: November 26, 2019Assignee: Hitachi, Ltd.Inventors: Yusuke Fujita, Ryu Takeda, Naoyuki Kanda
-
Patent number: 10467525Abstract: [Object] An object is to provide a training method of improving training of a recurrent neural network (RNN) using time-sequential data. [Solution] The training method includes a step 220 of initializing the RNN, and a training step 226 of training the RNN by designating a certain vector as a start position and optimizing various parameters to minimize error function. The training step 226 includes: an updating step 250 of updating RNN parameters through Truncated BPTT using consecutive N (N?3) vectors having a designated vector as a start point and using a reference value of a tail vector as a correct label; and a first repetition step 240 of repeating the process of executing the training step by newly designating a vector at a position satisfying a prescribed relation with the tail of N vectors used at the updating step until an end condition is satisfied. The vector at a position satisfying the prescribed relation is positioned at least two vectors behind the designated vector.Type: GrantFiled: May 10, 2016Date of Patent: November 5, 2019Assignee: National Institute of Information and Communications TechnologyInventor: Naoyuki Kanda
-
Publication number: 20190272828Abstract: A speaker estimation method that estimate the speaker from audio and image includes: inputting audio; extracting a feature quantity representing a voice characteristic from the input audio; inputting an image; detecting person regions of respective persons from the input image; estimating feature quantities representing voice characteristics from the respective detected person regions; Performing a change such that an image taken from another position and with another angle is input when any person is not detected; calculating a similarity between the feature quantity representing the voice characteristic extracted from the audio and the feature quantity representing the voice characteristic estimated from the person region in the image; and estimating a speaker from the calculated similarity.Type: ApplicationFiled: February 26, 2019Publication date: September 5, 2019Applicant: HITACHI, LTD.Inventors: Shota HORIGUCHI, Naoyuki KANDA
-
Publication number: 20190139540Abstract: A speech recognition device includes: an acoustic model based on an End-to-End neural network responsive to an observed sequence formed of prescribed acoustic features obtained from a speech signal by feature extracting unit, for calculating probability of the observed sequence being a certain symbol sequence; and a decoder responsive to a symbol sequence candidate, for decoding a speech signal by a WFST based on a posterior probability of each of word sequences corresponding to the symbol sequence candidate, probabilities calculated by the acoustic model for symbol sequences selected based on an observed sequence, and a posterior probability of each of the plurality of symbol sequences.Type: ApplicationFiled: June 2, 2017Publication date: May 9, 2019Applicant: National Institute of Information and Communications TechnologyInventor: Naoyuki KANDA
-
Publication number: 20180204566Abstract: [Object] An object is to provide a speech recognition device with improved recognition accuracy using characteristics of a neural network. [Solution] A speech recognition device includes: an acoustic model 308 implemented by a RNN (recurrent neural network) for calculating, for each state sequence, the posterior probability of a state sequence in response to an observed sequence consisting of prescribed speech features obtained from a speech; a WFST 320 based on S?1HCLG calculating, for each word sequence, posterior probability of a word sequence in response to a state sequence; and a hypothesis selecting unit 322, performing speech recognition of the speech signal based on a score calculated for each hypothesis of a word sequence corresponding to the speech signal, using the posterior probabilities calculated by the acoustic model 308 and the WFST 320 for the input observed sequence.Type: ApplicationFiled: May 10, 2016Publication date: July 19, 2018Inventor: Naoyuki KANDA
-
Patent number: 9989626Abstract: The present invention pertains to a method for estimating a sound source position in a space with high accuracy using a microphone installed on a robot moving in the space. A mobile robot includes a self-position estimation unit configure to estimate the self-position of the mobile robot, a sound source information obtaining unit configured to obtain direction information of an observed sound source, and a sound source position estimation unit configured to estimate the position of the sound source based on the estimated self-position and the direction information of the sound source.Type: GrantFiled: April 12, 2013Date of Patent: June 5, 2018Assignee: Hitachi, Ltd.Inventors: Takashi Sumiyoshi, Yasunari Obuchi, Naoyuki Kanda, Ryu Takeda
-
Publication number: 20180121800Abstract: [Object] An object is to provide a training method of improving training of a recurrent neural network (RNN) using time-sequential data. [Solution] The training method includes a step 220 of initializing the RNN, and a training step 226 of training the RNN by designating a certain vector as a start position and optimizing various parameters to minimize error function. The training step 226 includes: an updating step 250 of updating RNN parameters through Truncated BPTT using consecutive N (N?3) vectors having a designated vector as a start point and using a reference value of a tail vector as a correct label; and a first repetition step 240 of repeating the process of executing the training step by newly designating a vector at a position satisfying a prescribed relation with the tail of N vectors used at the updating step until an end condition is satisfied. The vector at a position satisfying the prescribed relation is positioned at least two vectors behind the designated vector.Type: ApplicationFiled: May 10, 2016Publication date: May 3, 2018Inventor: Naoyuki KANDA
-
Publication number: 20160171100Abstract: Provided is a voice search technology that can efficiently find and check a problematic call. To this end, a voice search system of the present invention includes a call search database that stores, for each of a reception channel and a transmission channel of each of a plurality of pieces of recorded call voice data, voice section sequences in association with predetermined keywords and time information. The call search database is searched based on an input search keyword, so that a voice section sequence that contains the search keyword is obtained.Type: ApplicationFiled: September 11, 2013Publication date: June 16, 2016Applicant: Hitachi, Ltd.Inventors: Yusuke FUJITA, Ryu TAKEDA, Naoyuki KANDA
-
Publication number: 20160103202Abstract: The present invention pertains to a method for estimating a sound source position in a space with high accuracy using a microphone installed on a robot moving in the space. A mobile robot includes a self-position estimation unit configure to estimate the self-position of the mobile robot, a sound source information obtaining unit configured to obtain direction information of an observed sound source, and a sound source position estimation unit configured to estimate the position of the sound source based on the estimated self-position and the direction information of the sound source.Type: ApplicationFiled: April 12, 2013Publication date: April 14, 2016Inventors: Takashi SUMIYOSHI, Yasunari OBUCHI, Naoyuki KANDA, Ryu TAKEDA
-
Publication number: 20090234854Abstract: An acoustic feature representing speech data provided with meta data is extracted. Next, a group of acoustic features which are extracted only from the speech data containing a specific word in the meta data and not from the other speech data is extracted from obtained sub-groups of acoustic features. The word and the extracted group of acoustic features are associated with each other to be stored. When there is a search key matching the word in the input search keys, the group of acoustic features corresponding to the word is output. Accordingly, the efforts of a user for inputting a key when the user searches for speech data are reduced.Type: ApplicationFiled: November 13, 2008Publication date: September 17, 2009Inventors: Naoyuki Kanda, Takashi Sumiyoshi, Yasunari Obuchi