Patents by Inventor Rakesh Iyer
Rakesh Iyer has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Publication number: 20240404308Abstract: Systems and methods for extracting information from documents are provided. In one example embodiment, a computer-implemented method includes obtaining one or more units of text from an image of a document. The method includes determining one or more annotated values from the one or more units of text and determining a set of candidate labels for each annotated value. The method determines each set of candidate labels by performing a search for the candidate labels based at least in part on a language associated with the document and a location of each annotated value. The method includes determining a canonical label for each annotated value based at least in part on the associated candidate labels, and mapping at least one annotated value to an action that is presented to a user based at least in part on the canonical label associated with the annotated value.Type: ApplicationFiled: May 22, 2024Publication date: December 5, 2024Inventors: Rakesh Iyer, Lisha Ruan
-
Patent number: 12125469Abstract: A method for predicting parametric vocoder parameter includes receiving a text utterance having one or more words, each word having one or more syllables, and each syllable having one or more phonemes. The method also includes receiving, as input to a vocoder model, prosodic features that represent an intended prosody for the text utterance and a linguistic specification. The prosodic features include a duration, pitch contour, and energy contour for the text utterance, while the linguistic specification includes sentence-level linguistic features, word-level linguistic features for each word, syllable-level linguistic features for each syllable, and phoneme-level linguistic features for each phoneme. The method also includes predicting vocoder parameters based on the prosodic features and the linguistic specification.Type: GrantFiled: October 17, 2023Date of Patent: October 22, 2024Assignee: Google LLCInventors: Rakesh Iyer, Vincent Wan
-
Publication number: 20240331681Abstract: A computer generated voice can automatically be adapted to be similar to a user's voice. Various implementations include processing audio data capturing a first language spoken utterance to identify one or more pitch characteristics. For example, the one or more pitch characteristics can include an estimated frequency range of the given user's voice. Additionally or alternatively, the system can process the audio data capturing the first language spoken utterance and a set of candidate computer generated voices using a computer generated voice selection model to select a candidate computer generated voice. Various implementations can include automatically modifying the selected candidate computer generated voice based on the one or more pitch characteristics to change the frequency range of the modified computer generated voice based on the user's voice.Type: ApplicationFiled: March 29, 2023Publication date: October 3, 2024Inventors: Rakesh Iyer, Jeffrey Robert Pitman, Pendar Yousefi, Te I, Tiruvilwamalai Raman
-
Patent number: 12033412Abstract: Systems and methods for extracting information from documents are provided. In one example embodiment, a computer-implemented method includes obtaining one or more units of text from an image of a document. The method includes determining one or more annotated values from the one or more units of text and determining a set of candidate labels for each annotated value. The method determines each set of candidate labels by performing a search for the candidate labels based at least in part on a language associated with the document and a location of each annotated value. The method includes determining a canonical label for each annotated value based at least in part on the associated candidate labels, and mapping at least one annotated value to an action that is presented to a user based at least in part on the canonical label associated with the annotated value.Type: GrantFiled: January 28, 2019Date of Patent: July 9, 2024Assignee: GOOGLE LLCInventors: Rakesh Iyer, Lisha Ruan
-
Publication number: 20240046915Abstract: A method for predicting parametric vocoder parameter includes receiving a text utterance having one or more words, each word having one or more syllables, and each syllable having one or more phonemes. The method also includes receiving, as input to a vocoder model, prosodic features that represent an intended prosody for the text utterance and a linguistic specification. The prosodic features include a duration, pitch contour, and energy contour for the text utterance, while the linguistic specification includes sentence-level linguistic features, word-level linguistic features for each word, syllable-level linguistic features for each syllable, and phoneme-level linguistic features for each phoneme. The method also includes predicting vocoder parameters based on the prosodic features and the linguistic specification.Type: ApplicationFiled: October 17, 2023Publication date: February 8, 2024Applicant: Google LLCInventors: Rakesh Iyer, Vincent Wan
-
Patent number: 11830474Abstract: A method for predicting parametric vocoder parameter includes receiving a text utterance having one or more words, each word having one or more syllables, and each syllable having one or more phonemes. The method also includes receiving, as input to a vocoder model, prosodic features that represent an intended prosody for the text utterance and a linguistic specification. The prosodic features include a duration, pitch contour, and energy contour for the text utterance, while the linguistic specification includes sentence-level linguistic features, word-level linguistic features for each word, syllable-level linguistic features for each syllable, and phoneme-level linguistic features for each phoneme. The method also includes predicting vocoder parameters based on the prosodic features and the linguistic specification.Type: GrantFiled: January 6, 2022Date of Patent: November 28, 2023Assignee: Google LLCInventors: Rakesh Iyer, Vincent Wan
-
Publication number: 20230097338Abstract: Systems and methods for synthesizing speech based on received text and one or more emulated speech parameters. Text is received with one or more emulated speech parameters that indicate one or more features for the synthesized speech. Synthesized speech audio is generated based on the received parameters. The synthesized speech audio data is provided to an emulated microphone component that provides the synthesized audio to an automatic speech recognizer. The automatic speech recognizer utilizes one or more speech recognition models to generate converted text based on the synthesized speech audio data.Type: ApplicationFiled: November 23, 2021Publication date: March 30, 2023Inventors: Nnamdi Kalu, Fernando Fernandes, Uri First, Erwin Jansen, Rakesh Iyer, Lingfeng Yang
-
Publication number: 20230065823Abstract: According to an aspect, a method includes storing messages exchanged on a messaging platform in a non-relational database, obtaining a database snapshot of the non-relational database, executing a database task on the database snapshot, and generating, in response to the database task, an update log, where the update log identifies a first record to be changed or deleted in the non-relational database. The method includes determining whether or not the first record identified in the update log has been updated in the non-relational database after a time instance associated with the database task and applying the change or deletion of the first record in the non-relational database in response to the first record being determined as not updated after the time instance associated with the database task.Type: ApplicationFiled: August 24, 2021Publication date: March 2, 2023Inventor: Rakesh Iyer
-
Publication number: 20220130371Abstract: A method for predicting parametric vocoder parameter includes receiving a text utterance having one or more words, each word having one or more syllables, and each syllable having one or more phonemes. The method also includes receiving, as input to a vocoder model, prosodic features that represent an intended prosody for the text utterance and a linguistic specification. The prosodic features include a duration, pitch contour, and energy contour for the text utterance, while the linguistic specification includes sentence-level linguistic features, word-level linguistic features for each word, syllable-level linguistic features for each syllable, and phoneme-level linguistic features for each phoneme. The method also includes predicting vocoder parameters based on the prosodic features and the linguistic specification.Type: ApplicationFiled: January 6, 2022Publication date: April 28, 2022Applicant: Google LLCInventors: Rakesh Iyer, Vincent Wan
-
Patent number: 11232780Abstract: A method for predicting parametric vocoder parameter includes receiving a text utterance having one or more words, each word having one or more syllables, and each syllable having one or more phonemes. The method also includes receiving, as input to a vocoder model, prosodic features that represent an intended prosody for the text utterance and a linguistic specification. The prosodic features include a duration, pitch contour, and energy contour for the text utterance, while the linguistic specification includes sentence-level linguistic features, word-level linguistic features for each word, syllable-level linguistic features for each syllable, and phoneme-level linguistic features for each phoneme. The method also includes predicting vocoder parameters based on the prosodic features and the linguistic specification.Type: GrantFiled: September 26, 2020Date of Patent: January 25, 2022Assignee: Google LLCInventors: Rakesh Iyer, Vincent Wan
-
Publication number: 20210406451Abstract: Systems and methods for extracting information from documents are provided. In one example embodiment, a computer-implemented method includes obtaining one or more units of text from an image of a document. The method includes determining one or more annotated values from the one or more units of text and determining a set of candidate labels for each annotated value. The method determines each set of candidate labels by performing a search for the candidate labels based at least in part on a language associated with the document and a location of each annotated value. The method includes determining a canonical label for each annotated value based at least in part on the associated candidate labels, and mapping at least one annotated value to an action that is presented to a user based at least in part on the canonical label associated with the annotated value.Type: ApplicationFiled: January 28, 2019Publication date: December 30, 2021Inventors: Rakesh Iyer, Lisha Ruan