Patents by Inventor Taira Ashikawa

Taira Ashikawa has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20230132418
    Abstract: According to one embodiment, a subtitle generating apparatus includes processing circuitry and a display. The processing circuitry is configured to sequentially acquire texts from voice recognition results. The processing circuitry is configured to store the texts as archive datasets. The processing circuitry is configured to estimate a split position and a concatenation position of the texts from one or more of the archive datasets, and generate a subtitle text from said one or more of the archive datasets based on the split position and the concatenation position. The processing circuitry is configured to update the archive datasets based on the split position and the concatenation position. The display is configured to display the subtitle text.
    Type: Application
    Filed: August 29, 2022
    Publication date: May 4, 2023
    Applicant: KABUSHIKI KAISHA TOSHIBA
    Inventors: Taira ASHIKAWA, Kenji IWATA
  • Patent number: 11620981
    Abstract: According to one embodiment, a speech recognition error correction apparatus includes a correction network memory and an error correction circuitry. The error correction circuitry calculates a difference between a speech recognition result string of an error correction target, which is a result of performing speech recognition on a new series of speech data, and a correction network, where a speech recognition result string and a correction result by a user for the speech recognition result string are associated, and when a value indicating the difference is equal to or less than a threshold, perform error correction on a speech recognition error portion in the speech recognition result string of the error correction target by using the correction network to generate a speech recognition error correction result string.
    Type: Grant
    Filed: September 4, 2020
    Date of Patent: April 4, 2023
    Assignee: KABUSHIKI KAISHA TOSHIBA
    Inventors: Taira Ashikawa, Hiroshi Fujimura, Kenji Iwata
  • Publication number: 20210280168
    Abstract: According to one embodiment, a speech recognition error correction apparatus includes a correction network memory and an error correction circuitry. The error correction circuitry calculates a difference between a speech recognition result string of an error correction target, which is a result of performing speech recognition on a new series of speech data, and a correction network, where a speech recognition result string and a correction result by a user for the speech recognition result string are associated, and when a value indicating the difference is equal to or less than a threshold, perform error correction on a speech recognition error portion in the speech recognition result string of the error correction target by using the correction network to generate a speech recognition error correction result string.
    Type: Application
    Filed: September 4, 2020
    Publication date: September 9, 2021
    Applicant: KABUSHIKI KAISHA TOSHIBA
    Inventors: Taira ASHIKAWA, Hiroshi FUJIMURA, Kenji IWATA
  • Patent number: 10614265
    Abstract: An apparatus for correcting a character string in a text of an embodiment includes a first converter, a first output unit, a second converter, an estimation unit, and a second output unit. The first converter recognizes a first speech of a first speaker, and converts the first speech to a first text. The first output unit outputs a first caption image indicating the first text. The second converter recognizes a second speech of a second speaker for correcting a character string to be corrected in the first text, and converts the second speech to a second text. The estimation unit estimates the character string to be corrected, based on text matching between the first text and the second text. The second output unit outputs a second caption image indicating that the character string to be corrected is to be replaced with the second text.
    Type: Grant
    Filed: December 21, 2016
    Date of Patent: April 7, 2020
    Assignee: Kabushiki Kaisha Toshiba
    Inventors: Kosei Fume, Taira Ashikawa, Masayuki Ashikawa
  • Patent number: 10587941
    Abstract: According to one embodiment, a microphone cooperation device includes a first device including a first microphone and a first speaker, and a second device including a second microphone and a second speaker. The first device and the second device are capable of a first action of joining a group of devices when the second microphone receives a first sound wave emitted from the first speaker.
    Type: Grant
    Filed: February 28, 2018
    Date of Patent: March 10, 2020
    Assignee: Kabushiki Kaisha Toshiba
    Inventors: Toru Taniguchi, Taira Ashikawa, Terufumi Morishita
  • Patent number: 10540987
    Abstract: A summary generating device includes a featural script extracting unit, a segment candidate generating unit, and a structuring estimating unit. The featural script extracting unit extracts featural script information of the words included in text information. Based on the extracted feature script information, the segment candidate generating unit generates candidates of segments that represent the constitutional units for the display purpose. Based on the generated candidates of segments and based on an estimation model for structuring, the structuring estimating unit estimates structure information containing information ranging from information of a comprehensive structure level to information of a local structure level.
    Type: Grant
    Filed: January 26, 2017
    Date of Patent: January 21, 2020
    Assignee: Kabushiki Kaisha Toshiba
    Inventors: Kosei Fume, Taira Ashikawa, Masayuki Ashikawa, Takashi Masuko
  • Patent number: 10373606
    Abstract: A transliteration support device according to an embodiment includes an acquisition unit, an extraction unit, a generation unit, and a reproduction unit. The acquisition unit acquires a text to be transliterated. The addition unit adds a transliteration tag indicating a transliteration setting of the text to the text. The extraction unit extracts a transliteration pattern in which a frequent appearance transliteration setting frequently appearing in the transliteration settings indicated by the transliteration tags and an applicable condition when the frequent appearance transliteration setting is applied to the text are in association with each other. The generation unit produces a synthesized voice using the transliteration pattern. The reproduction unit reproduces the produced synthesized voice.
    Type: Grant
    Filed: January 27, 2017
    Date of Patent: August 6, 2019
    Assignee: Kabushiki Kaisha Toshiba
    Inventors: Taira Ashikawa, Kosei Fume, Yuka Kuroda, Yoshiaki Mizuoka
  • Patent number: 10304460
    Abstract: According to an embodiment, a conference support system includes a recognizer, a classifier, a first caption controller, a second caption controller, and a display controller. The recognizer is configured to recognize text data corresponding speech from a speech section and configured to distinguish between the speech section and a non-speech section in speech data. The classifier is configured to classify the text data into first utterance data representing a principal utterance and second utterance data representing another utterance. The first caption controller is configured to generate first caption data for displaying the first utterance data without waiting for identification of the first utterance data to finish. The second caption controller is configured to generate second caption data for displaying the second utterance data after identification of the second utterance data finishes. The display controller is configured to control a display of the first caption data and the second caption data.
    Type: Grant
    Filed: February 23, 2017
    Date of Patent: May 28, 2019
    Assignee: Kabushiki Kaisha Toshiba
    Inventors: Taira Ashikawa, Kosei Fume, Masayuki Ashikawa, Hiroshi Fujimura
  • Patent number: 10276150
    Abstract: A correction system of the embodiment includes an interface system, a calculator, a generator, and a display controller. The interface system receives correction information for correcting a voice recognition result. The calculator estimates a part of the voice recognition result to be corrected and calculates a degree of association between the part to be corrected and the correction information. The generator generates corrected display information comprising at least one of the correction information and the part to be corrected using a display format corresponding to the degree of association. The display controller outputs the corrected display information on a display.
    Type: Grant
    Filed: February 23, 2017
    Date of Patent: April 30, 2019
    Assignee: Kabushiki Kaisha Toshiba
    Inventors: Kosei Fume, Taira Ashikawa, Masayuki Ashikawa, Hiroshi Fujimura
  • Patent number: 10255904
    Abstract: According to an embodiment, a reading-aloud information editing device includes an acquirer, an analyzer, a first generator, a second generator, and an extractor. The acquirer is configured to acquire an edit region including a text added with reading-aloud information from a document. The analyzer is configured to analyze a document structure of the edit region. The first generator is configured to generate one or more condition patterns by abstracting the edit region on the basis of the document structure. The second generator is configured to generate an extraction condition that is for extracting a text from the document and includes at least one of the condition patterns. The extractor is configured to extract a text suitable for the extraction condition from the document.
    Type: Grant
    Filed: February 9, 2017
    Date of Patent: April 9, 2019
    Assignee: Kabushiki Kaisha Toshiba
    Inventors: Kosei Fume, Masahiro Morita, Taira Ashikawa
  • Publication number: 20190069062
    Abstract: According to one embodiment, a microphone cooperation device includes a first device including a first microphone and a first speaker, and a second device including a second microphone and a second speaker. The first device and the second device are capable of a first action of joining a group of devices when the second microphone receives a first sound wave emitted from the first speaker.
    Type: Application
    Filed: February 28, 2018
    Publication date: February 28, 2019
    Inventors: Toru TANIGUCHI, Taira ASHIKAWA, Terufumi MORISHITA
  • Patent number: 10102859
    Abstract: According to an embodiment, a conference support apparatus includes a recognizer, a detector, a summarizer, and a subtitle generator. The recognizer is configured to recognize speech in speech data and generate text data. The detector is configured to detect a correction operation on the text data, the correction operation being an operation of correcting character data that has been incorrectly converted. The summarizer is configured to generate a summary relating to the text data subsequent to a part to which the correction operation is being performed, among the text data, when the correction operation is being detected. The subtitle generator is configured to generate subtitle information corresponding to the summary when the correction operation is being detected, and configured to generate subtitle information corresponding to the text data except when the correction operation is being detected.
    Type: Grant
    Filed: December 21, 2016
    Date of Patent: October 16, 2018
    Assignee: KABUSHIKI KAISHA TOSHIBA
    Inventors: Taira Ashikawa, Kosei Fume, Masayuki Ashikawa
  • Publication number: 20180082688
    Abstract: According to an embodiment, a conference support system includes a recognizer, a classifier, a first caption controller, a second caption controller, and a display controller. The recognizer is configured to recognize text data corresponding speech from a speech section and configured to distinguish between the speech section and a non-speech section in speech data. The classifier is configured to classify the text data into first utterance data representing a principal utterance and second utterance data representing another utterance. The first caption controller is configured to generate first caption data for displaying the first utterance data without waiting for identification of the first utterance data to finish. The second caption controller is configured to generate second caption data for displaying the second utterance data after identification of the second utterance data finishes. The display controller is configured to control a display of the first caption data and the second caption data.
    Type: Application
    Filed: February 23, 2017
    Publication date: March 22, 2018
    Inventors: Taira ASHIKAWA, Kosei FUME, Masayuki ASHIKAWA, Hiroshi FUJIMURA
  • Publication number: 20180075839
    Abstract: A correction system of the embodiment includes an interface system, a calculator, a generator, and a display controller. The interface system receives correction information for correcting a voice recognition result. The calculator estimates a part of the voice recognition result to be corrected and calculates a degree of association between the part to be corrected and the correction information. The generator generates corrected display information comprising at least one of the correction information and the part to be corrected using a display format corresponding to the degree of association. The display controller outputs the corrected display information on a display.
    Type: Application
    Filed: February 23, 2017
    Publication date: March 15, 2018
    Inventors: Kosei FUME, Taira ASHIKAWA, Masayuki ASHIKAWA, Hiroshi FUJIMURA
  • Publication number: 20170270949
    Abstract: A summary generating device includes a featural script extracting unit, a segment candidate generating unit, and a structuring estimating unit. The featural script extracting unit extracts featural script information of the words included in text information. Based on the extracted feature script information, the segment candidate generating unit generates candidates of segments that represent the constitutional units for the display purpose. Based on the generated candidates of segments and based on an estimation model for structuring, the structuring estimating unit estimates structure information containing information ranging from information of a comprehensive structure level to information of a local structure level.
    Type: Application
    Filed: January 26, 2017
    Publication date: September 21, 2017
    Inventors: Kosei FUME, Taira ASHIKAWA, Masayuki ASHIKAWA, Takashi MASUKO
  • Publication number: 20170270086
    Abstract: An apparatus for correcting a character string in a text of an embodiment includes a first converter, a first output unit, a second converter, an estimation unit, and a second output unit. The first converter recognizes a first speech of a first speaker, and converts the first speech to a first text. The first output unit outputs a first caption image indicating the first text. The second converter recognizes a second speech of a second speaker for correcting a character string to be corrected in the first text, and converts the second speech to a second text. The estimation unit estimates the character string to be corrected, based on text matching between the first text and the second text. The second output unit outputs a second caption image indicating that the character string to be corrected is to be replaced with the second text.
    Type: Application
    Filed: December 21, 2016
    Publication date: September 21, 2017
    Inventors: Kosei FUME, Taira ASHIKAWA, Masayuki ASHIKAWA
  • Publication number: 20170263238
    Abstract: According to an embodiment, a reading-aloud information editing device includes an acquirer, an analyzer, a first generator, a second generator, and an extractor. The acquirer is configured to acquire an edit region including a text added with reading-aloud information from a document. The analyzer is configured to analyze a document structure of the edit region. The first generator is configured to generate one or more condition patterns by abstracting the edit region on the basis of the document structure. The second generator is configured to generate an extraction condition that is for extracting a text from the document and includes at least one of the condition patterns. The extractor is configured to extract a text suitable for the extraction condition from the document.
    Type: Application
    Filed: February 9, 2017
    Publication date: September 14, 2017
    Inventors: Kosei FUME, Masahiro MORITA, Taira ASHIKAWA
  • Publication number: 20170263265
    Abstract: According to an embodiment, a conference support apparatus includes a recognizer, a detector, a summarizer, and a subtitle generator. The recognizer is configured to recognize speech in speech data and generate text data. The detector is configured to detect a correction operation on the text data, the correction operation being an operation of correcting character data that has been incorrectly converted. The summarizer is configured to generate a summary relating to the text data subsequent to a part to which the correction operation is being performed, among the text data, when the correction operation is being detected. The subtitle generator is configured to generate subtitle information corresponding to the summary when the correction operation is being detected, and configured to generate subtitle information corresponding to the text data except when the correction operation is being detected.
    Type: Application
    Filed: December 21, 2016
    Publication date: September 14, 2017
    Inventors: Taira ASHIKAWA, Kosei FUME, Masayuki ASHIKAWA
  • Publication number: 20170140749
    Abstract: A transliteration support device according to an embodiment includes an acquisition unit, an extraction unit, a generation unit, and a reproduction unit. The acquisition unit acquires a text to be transliterated. The addition unit adds a transliteration tag indicating a transliteration setting of the text to the text. The extraction unit extracts a transliteration pattern in which a frequent appearance transliteration setting frequently appearing in the transliteration settings indicated by the transliteration tags and an applicable condition when the frequent appearance transliteration setting is applied to the text are in association with each other. The generation unit produces a synthesized voice using the transliteration pattern. The reproduction unit reproduces the produced synthesized voice.
    Type: Application
    Filed: January 27, 2017
    Publication date: May 18, 2017
    Applicant: KABUSHIKI KAISHA TOSHIBA
    Inventors: Taira ASHIKAWA, Kosei FUME, Yuka KURODA, Yoshiaki MIZUOKA
  • Patent number: 9489944
    Abstract: According to an embodiment, a memory controller stores, in a memory, character strings in voice text obtained through voice recognition on voice data, a node index, a recognition score, and a voice index. A detector detects reproduction section of the voice data. An obtainer obtains reading of a phrase in a text written down from the reproduced voice data, and obtains insertion position of character strings. A searcher searches for a character string including the reading. A determiner determines whether to perform display based on the recognition score corresponding to the retrieved character string. A history updater stores, in a memory, candidate history data indicating the retrieved character string, the recognition score, and the character insertion position. A threshold updater decides on a display threshold value using the recognition score of the candidate history data and/or the recognition score of the character string selected by a selector.
    Type: Grant
    Filed: December 8, 2014
    Date of Patent: November 8, 2016
    Assignee: KABUSHIKI KAISHA TOSHIBA
    Inventors: Taira Ashikawa, Kouji Ueno