Patents by Inventor Anshuman Tripathi
Anshuman Tripathi has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Patent number: 11961515Abstract: A method includes receiving a plurality of unlabeled audio samples corresponding to spoken utterances not paired with corresponding transcriptions. At a target branch of a contrastive Siamese network, the method also includes generating a sequence of encoder outputs for the plurality of unlabeled audio samples and modifying time characteristics of the encoder outputs to generate a sequence of target branch outputs. At an augmentation branch of a contrastive Siamese network, the method also includes performing augmentation on the unlabeled audio samples, generating a sequence of augmented encoder outputs for the augmented unlabeled audio samples, and generating predictions of the sequence of target branch outputs generated at the target branch. The method also includes determining an unsupervised loss term based on target branch outputs and predictions of the sequence of target branch outputs. The method also includes updating parameters of the audio encoder based on the unsupervised loss term.Type: GrantFiled: December 14, 2021Date of Patent: April 16, 2024Assignee: Google LLCInventors: Jaeyoung Kim, Soheil Khorram, Hasim Sak, Anshuman Tripathi, Han Lu, Qian Zhang
-
Publication number: 20230402936Abstract: Disclosed herein is a system for controlling a solid state transformer (SST), the SST comprising an AC-to-DC stage, a DC-to-AC stage, and a DC-to-DC stage coupled between the AC-to-DC stage and the DC-to-AC stage, the DC-to-DC stage comprising one or more DC-to-DC converters. The system comprises a stored energy controller coupled to the AC-to-DC stage, the energy controller configured to control the total amount of stored energy within the capacitors of the SST; a power flow controller coupled to the DC-to-AC stage, the power flow controller configured to control power flow in the SST; and one or more energy balancing controllers each coupled to a corresponding DC-to-DC converter, each energy balancing controller configured to balance energy in the corresponding DC-to-DC converter. In some embodiments, the stored energy controller, the power flow controller and the one or more energy balancing controllers are decoupled from one another.Type: ApplicationFiled: November 3, 2021Publication date: December 14, 2023Inventors: Glen Ghias FARIVAR, Howe Li YEO, Radhika SARDA, Fengjiao CUI, Abishek SETHUPANDI, Haonan TIAN, Madasamy Palvesha THEVAR, Brihadeeswara Sriram VAISAMBHAYANA, Anshuman TRIPATHI
-
Publication number: 20230368779Abstract: A transformer-transducer model for unifying streaming and non-streaming speech recognition includes an audio encoder, a label encoder, and a joint network. The audio encoder receives a sequence of acoustic frames, and generates, at each of a plurality of time steps, a higher order feature representation for a corresponding acoustic frame. The label encoder receives a sequence of non-blank symbols output by a final softmax layer, and generates, at each of the plurality of time steps, a dense representation. The joint network receives the higher order feature representation and the dense representation at each of the plurality of time steps, and generates a probability distribution over possible speech recognition hypothesis. The audio encoder of the model further includes a neural network having an initial stack of transformer layers trained with zero look ahead audio context, and a final stack of transformer layers trained with a variable look ahead audio context.Type: ApplicationFiled: July 24, 2023Publication date: November 16, 2023Applicant: Google LLCInventors: Anshuman Tripathi, Hasim Sak, Han Lu, Qian Zhang, Jaeyoung Kim
-
Patent number: 11741947Abstract: A transformer-transducer model for unifying streaming and non-streaming speech recognition includes an audio encoder, a label encoder, and a joint network. The audio encoder receives a sequence of acoustic frames, and generates, at each of a plurality of time steps, a higher order feature representation for a corresponding acoustic frame. The label encoder receives a sequence of non-blank symbols output by a final softmax layer, and generates, at each of the plurality of time steps, a dense representation. The joint network receives the higher order feature representation and the dense representation at each of the plurality of time steps, and generates a probability distribution over possible speech recognition hypothesis. The audio encoder of the model further includes a neural network having an initial stack of transformer layers trained with zero look ahead audio context, and a final stack of transformer layers trained with a variable look ahead audio context.Type: GrantFiled: March 23, 2021Date of Patent: August 29, 2023Assignee: Google LLCInventors: Anshuman Tripathi, Hasim Sak, Han Lu, Qian Zhang, Jaeyoung Kim
-
Publication number: 20230096805Abstract: A method includes receiving a plurality of unlabeled audio samples corresponding to spoken utterances not paired with corresponding transcriptions. At a target branch of a contrastive Siamese network, the method also includes generating a sequence of encoder outputs for the plurality of unlabeled audio samples and modifying time characteristics of the encoder outputs to generate a sequence of target branch outputs. At an augmentation branch of a contrastive Siamese network, the method also includes performing augmentation on the unlabeled audio samples, generating a sequence of augmented encoder outputs for the augmented unlabeled audio samples, and generating predictions of the sequence of target branch outputs generated at the target branch. The method also includes determining an unsupervised loss term based on target branch outputs and predictions of the sequence of target branch outputs. The method also includes updating parameters of the audio encoder based on the unsupervised loss term.Type: ApplicationFiled: December 14, 2021Publication date: March 30, 2023Applicant: Google LLCInventors: Jaeyoung Kim, Soheil Khorram, Hasim Sak, Anshuman Tripathi, Han Lu, Qian Zhang
-
Publication number: 20230089308Abstract: A method includes receiving an input audio signal that corresponds to utterances spoken by multiple speakers. The method also includes processing the input audio to generate a transcription of the utterances and a sequence of speaker turn tokens each indicating a location of a respective speaker turn. The method also includes segmenting the input audio signal into a plurality of speaker segments based on the sequence of speaker tokens. The method also includes extracting a speaker-discriminative embedding from each speaker segment and performing spectral clustering on the speaker-discriminative embeddings to cluster the plurality of speaker segments into k classes. The method also includes assigning a respective speaker label to each speaker segment clustered into the respective class that is different than the respective speaker label assigned to the speaker segments clustered into each other class of the k classes.Type: ApplicationFiled: December 14, 2021Publication date: March 23, 2023Applicant: Google LLCInventors: Quan Wang, Han Lu, Evan Clark, Ignacio Lopez Moreno, Hasim Sak, Wei Xia, Taral Joglekar, Anshuman Tripathi
-
Publication number: 20230084758Abstract: A method for training a speech recognition model with a loss function includes receiving an audio signal including a first segment corresponding to audio spoken by a first speaker, a second segment corresponding to audio spoken by a second speaker, and an overlapping region where the first segment overlaps the second segment. The overlapping region includes a known start time and a known end time. The method also includes generating a respective masked audio embedding for each of the first and second speakers. The method also includes applying a masking loss after the known end time to the respective masked audio embedding for the first speaker when the first speaker was speaking prior to the known start time, or applying the masking loss prior to the known start time when the first speaker was speaking after the known end time.Type: ApplicationFiled: November 15, 2022Publication date: March 16, 2023Applicant: Google LLCInventors: Anshuman Tripathi, Han Liu, Hasim Sak
-
Patent number: 11521595Abstract: A method for training a speech recognition model with a loss function includes receiving an audio signal including a first segment corresponding to audio spoken by a first speaker, a second segment corresponding to audio spoken by a second speaker, and an overlapping region where the first segment overlaps the second segment. The overlapping region includes a known start time and a known end time. The method also includes generating a respective masked audio embedding for each of the first and second speakers. The method also includes applying a masking loss after the known end time to the respective masked audio embedding for the first speaker when the first speaker was speaking prior to the known start time, or applying the masking loss prior to the known start time when the first speaker was speaking after the known end time.Type: GrantFiled: May 1, 2020Date of Patent: December 6, 2022Assignee: Google LLCInventors: Anshuman Tripathi, Han Lu, Hasim Sak
-
Publication number: 20220310097Abstract: A streaming speech recognition model includes an audio encoder configured to receive a sequence of acoustic frames and generate a higher order feature representation for a corresponding acoustic frame in the sequence of acoustic frames. The streaming speech recognition model also includes a label encoder configured to receive a sequence of non-blank symbols output by a final softmax layer and generate a dense representation. The streaming speech recognition model also includes a joint network configured to receive the higher order feature representation generated by the audio encoder and the dense representation generated by the label encoder and generate a probability distribution over possible speech recognition hypotheses. Here, the streaming speech recognition model is trained using self-alignment to reduce prediction delay by encouraging an alignment path that is one frame left from a reference forced-alignment frame.Type: ApplicationFiled: December 15, 2021Publication date: September 29, 2022Applicant: Google LLCInventors: Jaeyoung Kim, Han Lu, Anshuman Tripathi, Qian Zhang, Hasim Sak
-
Publication number: 20220108689Abstract: A transformer-transducer model for unifying streaming and non-streaming speech recognition includes an audio encoder, a label encoder, and a joint network. The audio encoder receives a sequence of acoustic frames, and generates, at each of a plurality of time steps, a higher order feature representation for a corresponding acoustic frame. The label encoder receives a sequence of non-blank symbols output by a final softmax layer, and generates, at each of the plurality of time steps, a dense representation. The joint network receives the higher order feature representation and the dense representation at each of the plurality of time steps, and generates a probability distribution over possible speech recognition hypothesis. The audio encoder of the model further includes a neural network having an initial stack of transformer layers trained with zero look ahead audio context, and a final stack of transformer layers trained with a variable look ahead audio context.Type: ApplicationFiled: March 23, 2021Publication date: April 7, 2022Applicant: Google LLCInventors: Anshuman Tripathi, Hasim Sak, Han Lu, Qian Zhang, Jaeyoung Kim
-
Publication number: 20210343273Abstract: A method for training a speech recognition model with a loss function includes receiving an audio signal including a first segment corresponding to audio spoken by a first speaker, a second segment corresponding to audio spoken by a second speaker, and an overlapping region where the first segment overlaps the second segment. The overlapping region includes a known start time and a known end time. The method also includes generating a respective masked audio embedding for each of the first and second speakers. The method also includes applying a masking loss after the known end time to the respective masked audio embedding for the first speaker when the first speaker was speaking prior to the known start time, or applying the masking loss prior to the known start time when the first speaker was speaking after the known end time.Type: ApplicationFiled: May 1, 2020Publication date: November 4, 2021Applicant: Google LLCInventors: Anshuman Tripathi, Han Lu, Hasim Sak
-
Patent number: 10554144Abstract: This invention relates to a method of controlling a Solid State Transformer (SST).Type: GrantFiled: April 5, 2019Date of Patent: February 4, 2020Assignee: Nanyang Technological UniversityInventors: Shuyu Cao, Brihadeeswara Sriram Vaisambhayana, Anshuman Tripathi, Fengjiao Cui, Abishek Sethupandi, Hossein Dehghani Tafti
-
Publication number: 20190312521Abstract: This invention relates to a method of controlling a Solid State Transformer (SST).Type: ApplicationFiled: April 5, 2019Publication date: October 10, 2019Inventors: Shuyu Cao, Brihadeeswara Sriram Vaisambhayana, Anshuman Tripathi, Fengjiao Cui, Abishek Sethupandi, Hossein Dehghani Tafti
-
Patent number: 10245959Abstract: There is provided a power converter system including a power bus, a plurality of power converter modules connected to the power bus in parallel, a plurality of energy storage modules, each energy storage module coupled to the power bus via a corresponding one of the plurality of power converter modules, and a controller module configured to control at least one of the power converter modules to operate in one of a plurality of operating modes. In particular, the plurality of operating modes of the power converter module includes a plurality of charging power conversion modes for connecting an input power source to the corresponding energy storage module for charging power to the corresponding energy storage module. There is also provided a corresponding method of manufacturing the power converter system.Type: GrantFiled: October 7, 2015Date of Patent: April 2, 2019Assignee: Nanyang Technological UniversityInventors: Anshuman Tripathi, Satyajit Athlekar, Nishanthi Duraisamy
-
Patent number: 10088838Abstract: A method for diagnostic monitoring of a wind turbine generator system, said wind turbine generator system comprising a generator, a drive train, and a number of sensors for providing signals and a control system. The method comprising the selection of three sets of signals from said sensors. From the three sets of signals three conditioned sets of signals are obtained by high pass filtering said first set of signals, low pass filtering the second set of signals, and forming a moving average value on the third set of signals. Based on each of said first, second and third set of conditioned signals an evaluation is performed in order to determine a fault, where said evaluation comprises comparing the first, second and third set of conditioned signals with reference values. If said comparison indicates a fault an alarm is set.Type: GrantFiled: January 17, 2012Date of Patent: October 2, 2018Assignee: VESTAS WIND SYSTEMS A/SInventors: Shu Yu Cao, Bing Li, Anshuman Tripathi, Hock Heng Thia, Rasool Beevi D-O Mohamed Arif, Kheng Hong Ang
-
Patent number: 9722520Abstract: A method for controlling a variable speed wind turbine generator is disclosed. The generator is connected to a power converter comprising switches. The generator comprises a stator and a set of terminals connected to the stator and to the switches of the power converter. The method comprises: determining a stator flux reference value corresponding to a generator power of a desired magnitude, determining an estimated stator flux value corresponding to an actual generator power, determining a difference between the determined stator flux reference value and the estimated stator flux value, and operating said switches in correspondence to the determined stator flux reference value and the estimated stator flux value to adapt at least one stator electrical quantity to obtain said desired generator power magnitude.Type: GrantFiled: August 28, 2009Date of Patent: August 1, 2017Assignee: VESTAS WIND SYSTEMS A/SInventors: Anshuman Tripathi, Cao Shu Yu, Allan Holm Jörgensen, Lars Helle
-
Patent number: 9590546Abstract: A power dissipating arrangement for dissipating power from a generator in a wind turbine is provided. The generator comprises a plurality of output terminals corresponding to a multi-phase output. The power dissipating arrangement comprises a plurality of dissipating units, a plurality of semiconductor switches, a trigger circuit for switching the semiconductor switches and a control unit for controlling the operation of the trigger circuit, thereby controlling the switching of the semiconductor switches.Type: GrantFiled: November 6, 2015Date of Patent: March 7, 2017Assignee: VESTAS WIND SYSTEMS A/SInventors: Ove Styhm, Anshuman Tripathi, Amit Kumar Gupta
-
Patent number: 9494138Abstract: A variable speed wind turbine is provided. The wind turbine comprises a generator, a power converter for converting at least a portion of electrical power generated by the generator, an energy management arrangement coupled to the power converter, the energy management arrangement comprises an energy storage unit, and a controller. The controller is adapted to detect a power imbalance event and to transfer at least a portion of excess electrical energy generated by the generator to the energy storage unit to be stored therein when the power imbalance event is detected.Type: GrantFiled: August 10, 2010Date of Patent: November 15, 2016Assignee: VESTAS WIND SYSTEMS A/SInventors: Amit Kumar Gupta, Gil Lampong Opina, Jr., Anshuman Tripathi, Yugarajan Karuppanan, Michael Casem Tumabcao
-
Patent number: 9476408Abstract: A method of controlling a wind turbine generator is provided, the wind turbine generator converting mechanical energy to electrical. The method comprises: determining an electromagnetic power reference representing the electromagnetic power generated by the wind turbine generator, wherein the electromagnetic power reference is determined based on a desired output of the wind turbine generator; controlling the electrical power generated by the wind turbine generator using a control signal, wherein the control signal is derived from the electromagnetic power reference and is modified in dependence on an inverse power function of the wind turbine generator by incorporating minimal copper loss constraint and stator voltage limiting constraint such that non-linearity of the wind turbine generator is compensated in the control loop and it operates at its maximum efficiency. One effect of the method is that classical linear control loop design can be employed in spite of the plant being a non-linear identity.Type: GrantFiled: March 12, 2014Date of Patent: October 25, 2016Assignee: VESTAS WIND SYSTEMS A/SInventors: Shu Yu Cao, Anshuman Tripathi, Kim B. Larsen, Gert Karmisholt Andersen, Lars Helle
-
Patent number: 9438155Abstract: Parameters of PM machines, especially for IPM machine, are known to vary by significant amounts. This affects the controllability of such machines, which may lead to reduced power loading capability and increased losses. The present invention relates to a method for PM machine inductance profile identification based on voltage mode stator flux observation which could be easily integrated to the generator start-up process in wind turbine application for both stator flux vector feedback control system and current vector feedback control system.Type: GrantFiled: December 28, 2012Date of Patent: September 6, 2016Assignee: Vestas Wind Systems A/SInventors: Shu Yu Cao, Anshuman Tripathi, Swee Yee Fonn, Ramasamy Anbarasu, Amit Kumar Gupta