Patents by Inventor Javier Latorre
Javier Latorre has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Patent number: 11443733Abstract: A text-to-speech (TTS) system that is capable of considering characteristics of various portions of text data in order to create continuity between segments of synthesized speech. The system can analyze text portions of a work and create feature vectors including data corresponding to characteristics of the individual portions and/or the overall work. A TTS processing component can then consider feature vector(s) from other portions when performing TTS processing on text of a first portion, thus giving the TTS component some intelligence regarding other portions of the work, which can then result in more continuity between synthesized speech segments.Type: GrantFiled: October 28, 2019Date of Patent: September 13, 2022Assignee: Amazon Technologies, Inc.Inventors: Roberto Barra Chicote, Javier Latorre, Adam Franciszek Nadolski, Viacheslav Klimkov, Thomas Edward Merritt
-
Publication number: 20220096771Abstract: The present invention relates to a nasal oxygen cannula, which has a simple and economic design that reduces drug wastage and allows the effective use time to be measured in order to effectively monitor oxygen therapy treatment in patients with spontaneous breathing. Structurally, the cannula is formed by a mechanical system that is only activated when it comes into physical contact with the patientÃs columella, and an electronic component that allows the real use time of the device to be measured by means of an electric actuator that triggers an electric pulse the moment the cannula is in place on the user.Type: ApplicationFiled: July 17, 2019Publication date: March 31, 2022Inventors: Carlos Javier Latorre Rojas, Maria Lucia Arango Cortes, Fabian Cortes Munoz, Jenny Carolina Sanchez Casas
-
Publication number: 20220054219Abstract: The present invention is for use in the field of instruments and devices adapted for surgery and diagnostics with instruments. Specifically, it relates to a device for enabling the punction of particular anatomical structures in the craniofacial area. The device comprises a support frame in the form of an adjustable headband which may be adapted to the user's anatomy, a sphere with two arms that hold a placement fastener onto the face or cranium of the user, applying pressure to a specific position through the use of screws or rings that allow free movement on the hinges of the arms.Type: ApplicationFiled: April 24, 2019Publication date: February 24, 2022Inventors: Maria Lucia Arango Cortes, Carlos Javier Latorre Rojas, Fabian Cortes Munoz, Celso Ernesto Bohorquez Escobar, Jenny Carolina Sanchez Casas
-
Patent number: 11144597Abstract: A system for emulating a subject, to allow a user to interact with a computer generated talking head with the subject's face and voice; said system comprising a processor, a user interface and a personality storage section, the user interface being configured to emulate the subject, by displaying a talking head which comprises the subject's face and output speech from the mouth of the face with the subject's voice, the user interface further comprising a receiver for receiving a query from the user, the emulated subject being configured to respond to the query received from the user, the processor comprising a dialogue section and a talking head generation section, wherein said dialogue section is configured to generate a response to a query inputted by a user from the user interface and generate a response to be outputted by the talking head, the response being generated by retrieving information from said personality storage section, said personality storage section comprising content created by or aboutType: GrantFiled: March 16, 2018Date of Patent: October 12, 2021Assignee: Kabushiki Kaisha ToshibaInventors: Balakrishna Venkata Jagannadha Kolluru, Vincent Ping Leung Wan, Bjorn Dietmar Rafael Stenger, Roberto Cipolla, Javier Latorre-Martinez, Langzhou Chen, Ranniery Da Silva Maia, Kayoko Yanagisawa, Norbert Braunschweiler, Ioannis Stylianou, Robert Arthur Blokland
-
Publication number: 20200152169Abstract: A text-to-speech (TTS) system that is capable of considering characteristics of various portions of text data in order to create continuity between segments of synthesized speech. The system can analyze text portions of a work and create feature vectors including data corresponding to characteristics of the individual portions and/or the overall work. A TTS processing component can then consider feature vector(s) from other portions when performing TTS processing on text of a first portion, thus giving the TTS component some intelligence regarding other portions of the work, which can then result in more continuity between synthesized speech segments.Type: ApplicationFiled: October 28, 2019Publication date: May 14, 2020Inventors: Roberto Barra Chicote, Javier Latorre, Adam Franciszek Nadolski, Viacheslav Klimkov, Thomas Edward Merritt
-
Patent number: 10475438Abstract: A text-to-speech (TTS) system that is capable of considering characteristics of various portions of text data in order to create continuity between segments of synthesized speech. The system can analyze text portions of a work and create feature vectors including data corresponding to characteristics of the individual portions and/or the overall work. A TTS processing component can then consider feature vector(s) from other portions when performing TTS processing on text of a first portion, thus giving the TTS component some intelligence regarding other portions of the work, which can then result in more continuity between synthesized speech segments.Type: GrantFiled: March 2, 2017Date of Patent: November 12, 2019Assignee: Amazon Technologies, Inc.Inventors: Roberto Barra Chicote, Javier Latorre, Adam Franciszek Nadolski, Viacheslav Klimkov, Thomas Edward Merritt
-
Patent number: 10373604Abstract: An acoustic model is adapted, relating acoustic units to speech vectors. The acoustic model comprises a set of acoustic model parameters related to a given speech factor. The acoustic model parameters enable the acoustic model to output speech vectors with different values of the speech factor. The method comprises inputting a sample of speech which is corrupted by noise; determining values of the set of acoustic model parameters which enable the acoustic model to output speech with said first value of the speech factor; and employing said determined values of the set of speech factor parameters in said acoustic model. The acoustic model parameters are obtained by obtaining corrupted speech factor parameters using the sample of speech, and mapping the corrupted speech factor parameters to clean acoustic model parameters using noise characterization paramaters characterizing the noise.Type: GrantFiled: February 2, 2017Date of Patent: August 6, 2019Assignee: Kabushiki Kaisha ToshibaInventors: Javier Latorre-Martinez, Vincent Ping Leung Wan, Kayoko Yanagisawa
-
Publication number: 20180203946Abstract: A system for emulating a subject, to allow a user to interact with a computer generated talking head with the subject's face and voice; said system comprising a processor, a user interface and a personality storage section, the user interface being configured to emulate the subject, by displaying a talking head which comprises the subject's face and output speech from the mouth of the face with the subject's voice, the user interface further comprising a receiver for receiving a query from the user, the emulated subject being configured to respond to the query received from the user, the processor comprising a dialogue section and a talking head generation section, wherein said dialogue section is configured to generate a response to a query inputted by a user from the user interface and generate a response to be outputted by the talking head, the response being generated by retrieving information from said personality storage section, said personality storage section comprising content created by or aboutType: ApplicationFiled: March 16, 2018Publication date: July 19, 2018Applicant: Kabushiki Kaisha ToshibaInventors: Balakrishna Venkata Jagannadha KOLLURU, Vincent Ping Leung Wan, Bjorn Dietmar Rafael Stenger, Roberto Cipolla, Javier Latorre-Martinez, Langzhou Chen, Ranniery Da Silva Maia, Kayoko Yanagisawa, Norbert Braunschweiler, Ioannis Stylianou, Robert Arthur Blokland
-
Patent number: 9959368Abstract: A system for emulating a subject, to allow a user to interact with a computer generated talking head with the subject's face and voice; said system comprising a processor, a user interface and a personality storage section, the user interface being configured to emulate the subject, by displaying a talking head which comprises the subject's face and output speech from the mouth of the face with the subject's voice, the user interface further comprising a receiver for receiving a query from the user, the emulated subject being configured to respond to the query received from the user, the processor comprising a dialogue section and a talking head generation section, wherein said dialogue section is configured to generate a response to a query inputted by a user from the user interface and generate a response to be outputted by the talking head, the response being generated by retrieving information from said personality storage section, said personality storage section comprising content created by or aboutType: GrantFiled: August 13, 2014Date of Patent: May 1, 2018Assignee: Kabushiki Kaisha ToshibaInventors: Balakrishna Venkata Jagannadha Kolluru, Vincent Ping Leung Wan, Bjorn Dietmar Rafael Stenger, Roberto Cipolla, Javier Latorre-Martinez, Langzhou Chen, Ranniery Da Silva Maia, Kayoko Yanagisawa, Norbert Braunschweiler, Ioannis Stylianou, Robert Arthur Blokland
-
Patent number: 9959657Abstract: A method of animating a computer generation of a head, the head having a mouth which moves in accordance with speech to be output by the head, said method comprising: providing an input related to the speech which is to be output by the movement of the lips; dividing said input into a sequence of acoustic units; selecting expression characteristics for the inputted text; converting said sequence of acoustic units to a sequence of image vectors using a statistical model, wherein said model has a plurality of model parameters describing probability distributions which relate an acoustic unit to an image vector, said image vector comprising a plurality of parameters which define a face of said head; and outputting said sequence of image vectors as video such that the mouth of said head moves to mime the speech associated with the input text with the selected expression, wherein a parameter of a predetermined type of each probability distribution in said selected expression is expressed as a weighted sum of paType: GrantFiled: January 29, 2014Date of Patent: May 1, 2018Assignee: Kabushiki Kaisha ToshibaInventors: Javier Latorre-Martinez, Vincent Ping Leung Wan, Bjorn Stenger, Robert Anderson, Roberto Cipolla
-
Publication number: 20170221479Abstract: A method of adapting an acoustic model relating acoustic units to speech vectors, wherein said acoustic model comprises a set of speech factor parameters related to a given speech factor and which enable the acoustic model to output speech vectors with different values of the speech factor, the method comprising: inputting a sample of speech with a first value of the speech factor; determining values of the set of speech factor parameters which enable the acoustic model to output speech with said first value of the speech factor; and employing said determined values of the set of speech factor parameters in said acoustic model, wherein said sample of speech is corrupted by noise, and wherein said step of determining the values of the set of speech factor parameters comprises: (i) obtaining noise characterization parameters characterising the noise; (ii) performing a speech factor parameter generation algorithm on the sample of speech, thereby generating corrupted values of the set of speech factor paramType: ApplicationFiled: February 2, 2017Publication date: August 3, 2017Applicant: Kabushiki Kaisha ToshibaInventors: Javier LATORRE-MARTINEZ, Vincent Ping Leung Wan, Kayoko Yanagisawa
-
Patent number: 9454963Abstract: A text-to-speech method for simulating a plurality of different voice characteristics includes dividing inputted text into a sequence of acoustic units; selecting voice characteristics for the inputted text; converting the sequence of acoustic units to a sequence of speech vectors using an acoustic model having a plurality of model parameters provided in clusters each having at least one sub-cluster and describing probability distributions which relate an acoustic unit to a speech vector; and outputting the sequence of speech vectors as audio with the selected voice characteristics. A parameter of a predetermined type of each probability distribution is expressed as a weighted sum of parameters of the same type using voice characteristic dependent weighting. In converting the sequence of acoustic units to a sequence of speech vectors, the voice characteristic dependent weights for the selected voice characteristics are retrieved for each cluster such that there is one weight per sub-cluster.Type: GrantFiled: March 13, 2013Date of Patent: September 27, 2016Assignee: KABUSHIKI KAISHA TOSHIBAInventors: Javier Latorre-Martinez, Vincent Ping Leung Wan, Kean Kheong Chin, Mark John Francis Gales, Katherine Mary Knill, Masami Akamine, Byung Ha Chung
-
Patent number: 9361722Abstract: A method of animating a computer generation of a head and displaying the text of an electronic book, such that the head has a mouth which moves in accordance with the speech of the text of the electronic book to be output by the head and a word or group of words from the text is displayed while simultaneously being mimed by the mouth, wherein input text is divided into a sequence of acoustic units, which are converted to a sequence of image vectors and into a sequence of text display indicators. The sequence of image vectors is outputted as video such that the mouth of said head moves to mime the speech associated with the input text with a selected expression, and the sequence of text display indicators is output as video which is synchronized with the lip movement of the head.Type: GrantFiled: August 8, 2014Date of Patent: June 7, 2016Assignee: KABUSHIKI KAISHA TOSHIBAInventors: Javier Latorre-Martinez, Vincent Ping Leung Wan, Balakrishna Venkata Jagannadha Kolluru, Ioannis Stylianou, Robert Arthur Blokland, Norbert Braunschweiler, Kayoko Yanagisawa, Langzhou Chen, Ranniery Maia, Robert Anderson, Bjorn Stenger, Roberto Cipolla, Neil Baker
-
Patent number: 9269347Abstract: A text-to-speech method configured to output speech having a selected speaker voice and a selected speaker attribute, including: inputting text; dividing the inputted text into a sequence of acoustic units; selecting a speaker for the inputted text; selecting a speaker attribute for the inputted text; converting the sequence of acoustic units to a sequence of speech vectors using an acoustic model; and outputting the sequence of speech vectors as audio with the selected speaker voice and a selected speaker attribute. The acoustic model includes a first set of parameters relating to speaker voice and a second set of parameters relating to speaker attributes, which parameters do not overlap. The selecting a speaker voice includes selecting parameters from the first set of parameters and the selecting the speaker attribute includes selecting the parameters from the second set of parameters.Type: GrantFiled: March 15, 2013Date of Patent: February 23, 2016Assignee: Kabushiki Kaisha ToshibaInventors: Javier Latorre-Martinez, Vincent Ping Leung Wan, Kean Kheong Chin, Mark John Francis Gales, Katherine Mary Knill, Masami Akamine
-
Publication number: 20150052084Abstract: A system for emulating a subject, to allow a user to interact with a computer generated talking head with the subject's face and voice; said system comprising a processor, a user interface and a personality storage section, the user interface being configured to emulate the subject, by displaying a talking head which comprises the subject's face and output speech from the mouth of the face with the subject's voice, the user interface further comprising a receiver for receiving a query from the user, the emulated subject being configured to respond to the query received from the user, the processor comprising a dialogue section and a talking head generation section, wherein said dialogue section is configured to generate a response to a query inputted by a user from the user interface and generate a response to be outputted by the talking head, the response being generated by retrieving information from said personality storage section, said personality storage section comprising content created by or aboutType: ApplicationFiled: August 13, 2014Publication date: February 19, 2015Applicant: Kabushiki Kaisha ToshibaInventors: Balakrishna Venkata Jagannadha KOLLURU, Vincent Ping Leung WAN, Bjorn Dietmar Rafael STENGER, Roberto CIPOLLA, Javier LATORRE-MARTINEZ, Langzhou CHEN, Ranniery Da Silva MAIA, Kayoko YANAGISAWA, Norbert BRAUNSCHWEILER, Ioannis STYLIANOU, Robert Arthur BLOKLAND
-
Publication number: 20150042662Abstract: A method of animating a computer generation of a head and displaying the text of an electronic book, such that the head has a mouth which moves in accordance with the speech of the text of the electronic book to be output by the head and a word or group of words from the text is displayed while simultaneously being mimed by the mouth, said method comprising: inputting the text of said book; dividing said input text into a sequence of acoustic units; determining expression characteristics for the inputted text; calculating a duration for each acoustic unit using a duration model; converting said sequence of acoustic units to a sequence of image vectors using a statistical model, wherein said model has a plurality of model parameters describing probability distributions which relate an acoustic unit to an image vector, said image vector comprising a plurality of parameters which define a face of said head; converting said sequence of acoustic units into a sequence of text display indicators using an text disType: ApplicationFiled: August 8, 2014Publication date: February 12, 2015Applicant: KABUSHIKI KAISHA TOSHIBAInventors: Javier Latorre-Martinez, Vincent Ping Leung Wan, Balakrishna Venkata Jagannadha Kolluru, Ioannis Stylianou, Robert Arthur Blokland, Norbert Braunschweiler, Kayoko Yanagisawa, Langzhou Chen, Ranniery MAIA, Robert Anderson, Bjorn Stenger, Roberto Cipolla, Neil Baker
-
Publication number: 20140210830Abstract: A method of animating a computer generation of a head, the head having a mouth which moves in accordance with speech to be output by the head, said method comprising: providing an input related to the speech which is to be output by the movement of the lips; dividing said input into a sequence of acoustic units; selecting expression characteristics for the inputted text; converting said sequence of acoustic units to a sequence of image vectors using a statistical model, wherein said model has a plurality of model parameters describing probability distributions which relate an acoustic unit to an image vector, said image vector comprising a plurality of parameters which define a face of said head; and outputting said sequence of image vectors as video such that the mouth of said head moves to mime the speech associated with the input text with the selected expression, wherein a parameter of a predetermined type of each probability distribution in said selected expression is expressed as a weighted sum of paType: ApplicationFiled: January 29, 2014Publication date: July 31, 2014Applicant: Kabushiki Kaisha ToshibaInventors: Javier LATORRE-MARTINEZ, Vincent Ping Leung Wan, Bjorn Stenger, Robert Anderson, Roberto Cipolla
-
Publication number: 20140210831Abstract: A method of animating a computer generation of a head, the head having a mouth which moves in accordance with speech to be output by the head, said method comprising: providing an input related to the speech which is to be output by the movement of the mouth; dividing said input into a sequence of acoustic units; selecting an expression to be output by said head; converting said sequence of acoustic units to a sequence of image vectors using a statistical model, wherein said model has a plurality of model parameters describing probability distributions which relate an acoustic unit to an image vector for a selected expression, said image vector comprising a plurality of parameters which define a face of said head; and outputting said sequence of image vectors as video such that the mouth of said head moves to mime the speech associated with the input text with the selected expression, wherein the image parameters define the face of a head using an appearance model comprising a plurality of shape modes andType: ApplicationFiled: January 29, 2014Publication date: July 31, 2014Applicant: Kabushiki Kaisha ToshibaInventors: Bjorn Stenger, Robert Anderson, Javier Latorre-Martinez, Vincent Ping Leung Wan, Roberto Cipolla
-
Publication number: 20130262109Abstract: A text-to-speech method for simulating a plurality of different voice characteristics includes dividing inputted text into a sequence of acoustic units; selecting voice characteristics for the inputted text; converting the sequence of acoustic units to a sequence of speech vectors using an acoustic model having a plurality of model parameters provided in clusters each having at least one sub-cluster and describing probability distributions which relate an acoustic unit to a speech vector; and outputting the sequence of speech vectors as audio with the selected voice characteristics. A parameter of a predetermined type of each probability distribution is expressed as a weighted sum of parameters of the same type using voice characteristic dependent weighting. In converting the sequence of acoustic units to a sequence of speech vectors, the voice characteristic dependent weights for the selected voice characteristics are retrieved for each cluster such that there is one weight per sub-cluster.Type: ApplicationFiled: March 13, 2013Publication date: October 3, 2013Inventors: Javier Latorre-Martinez, Vincent Ping Leung Wan, Kean Kheong Chin, Mark John Francis Gales, Katherine Mary Knill, Masami Akamine, Byung Ha Chung
-
Publication number: 20130262119Abstract: A text-to-speech method configured to output speech having a selected speaker voice and a selected speaker attribute, including: inputting text; dividing the inputted text into a sequence of acoustic units; selecting a speaker for the inputted text; selecting a speaker attribute for the inputted text; converting the sequence of acoustic units to a sequence of speech vectors using an acoustic model; and outputting the sequence of speech vectors as audio with the selected speaker voice and a selected speaker attribute. The acoustic model includes a first set of parameters relating to speaker voice and a second set of parameters relating to speaker attributes, which parameters do not overlap. The selecting a speaker voice includes selecting parameters from the first set of parameters and the selecting the speaker attribute includes selecting the parameters from the second set of parameters.Type: ApplicationFiled: March 15, 2013Publication date: October 3, 2013Applicant: Kabushiki Kaisha ToshibaInventors: Javier LATORRE-MARTINEZ, Vincent Ping Leung Wan, Kean Kheong Chin, Mark John Francis Gales, Katherine Mary Knill, Masami Akamine