Patents by Inventor Qiao Liang

Qiao Liang has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20250095634
    Abstract: A method includes receiving a sequence of acoustic frames characterizing one or more utterances as input to a multilingual automated speech recognition (ASR) model. The method also includes generating a higher order feature representation for a corresponding acoustic frame. The method also includes generating a hidden representation based on a sequence of non-blank symbols output by a final softmax layer. The method also includes generating a probability distribution over possible speech recognition hypotheses based on the hidden representation generated by the prediction network at each of the plurality of output steps and the higher order feature representation generated by the encoder at each of the plurality of output steps. The method also includes predicting an end of utterance (EOU) token at an end of each utterance. The method also includes classifying each acoustic frame as either speech, initial silence, intermediate silence, or final silence.
    Type: Application
    Filed: December 2, 2024
    Publication date: March 20, 2025
    Applicant: Google LLC
    Inventors: Bo Li, Tara N. Sainath, Ruoming Pang, Shuo-yiin Chang, Qiumin Xu, Trevor Strohman, Vince Chen, Qiao Liang, Heguang Liu, Yanzhang He, Parisa Haghani, Sameer Bidichandani
  • Patent number: 12183322
    Abstract: A method includes receiving a sequence of acoustic frames characterizing one or more utterances as input to a multilingual automated speech recognition (ASR) model. The method also includes generating a higher order feature representation for a corresponding acoustic frame. The method also includes generating a hidden representation based on a sequence of non-blank symbols output by a final softmax layer. The method also includes generating a probability distribution over possible speech recognition hypotheses based on the hidden representation generated by the prediction network at each of the plurality of output steps and the higher order feature representation generated by the encoder at each of the plurality of output steps. The method also includes predicting an end of utterance (EOU) token at an end of each utterance. The method also includes classifying each acoustic frame as either speech, initial silence, intermediate silence, or final silence.
    Type: Grant
    Filed: September 22, 2022
    Date of Patent: December 31, 2024
    Assignee: Google LLC
    Inventors: Bo Li, Tara N. Sainath, Ruoming Pang, Shuo-yiin Chang, Qiumin Xu, Trevor Strohman, Vince Chen, Qiao Liang, Heguang Liu, Yanzhang He, Parisa Haghani, Sameer Bidichandani
  • Publication number: 20240363122
    Abstract: Techniques disclosed herein are directed towards streaming keyphrase detection which can be customized to detect one or more particular keyphrases, without requiring retraining of any model(s) for those particular keyphrase(s). Many implementations include processing audio data using a speaker separation model to generate separated audio data which isolates an utterance spoken by a human speaker from one or more additional sounds not spoken by the human speaker, and processing the separated audio data using a text independent speaker identification model to determine whether a verified and/or registered user spoke a spoken utterance captured in the audio data. Various implementations include processing the audio data and/or the separated audio data using an automatic speech recognition model to generate a text representation of the utterance.
    Type: Application
    Filed: July 5, 2024
    Publication date: October 31, 2024
    Inventors: Rajeev Rikhye, Quan Wang, Yanzhang He, Qiao Liang, Ian C. McGraw
  • Patent number: 12033641
    Abstract: Techniques disclosed herein are directed towards streaming keyphrase detection which can be customized to detect one or more particular keyphrases, without requiring retraining of any model(s) for those particular keyphrase(s). Many implementations include processing audio data using a speaker separation model to generate separated audio data which isolates an utterance spoken by a human speaker from one or more additional sounds not spoken by the human speaker, and processing the separated audio data using a text independent speaker identification model to determine whether a verified and/or registered user spoke a spoken utterance captured in the audio data. Various implementations include processing the audio data and/or the separated audio data using an automatic speech recognition model to generate a text representation of the utterance.
    Type: Grant
    Filed: January 30, 2023
    Date of Patent: July 9, 2024
    Assignee: GOOGLE LLC
    Inventors: Rajeev Rikhye, Quan Wang, Yanzhang He, Qiao Liang, Ian C. McGraw
  • Publication number: 20240017274
    Abstract: A centrifuge rotor includes tube retainers for holding two or more sample tubes at a 90 degree angle in a minimized configuration for effective separation of a blood sample using a mini centrifuge. The centrifuge rotor with sample tubes in place has an overall width that is less than two times the length of a sample tube.
    Type: Application
    Filed: July 15, 2023
    Publication date: January 18, 2024
    Inventors: Qiao LIANG, Philip FUNG
  • Publication number: 20230335117
    Abstract: A method includes receiving, as input to a speech recognition model, audio data corresponding to a spoken utterance. The method also includes performing, using the speech recognition model, speech recognition on the audio data by, at each of a plurality of time steps, encoding, using an audio encoder, the audio data corresponding to the spoken utterance into a corresponding audio encoding, and decoding, using a speech recognition joint network, the corresponding audio encoding into a probability distribution over possible output labels. At each of the plurality of time steps, the method also includes determining, using an intended query (IQ) joint network configured to receive a label history representation associated with a sequence of non-blank symbols output by a final softmax layer, an intended query decision indicating whether or not the spoken utterance includes a query intended for a digital assistant.
    Type: Application
    Filed: March 20, 2023
    Publication date: October 19, 2023
    Applicant: Google LLC
    Inventors: Shuo-yiin Chang, Guru Prakash Arumugam, Zelin Wu, Tara N. Sainath, Bo LI, Qiao Liang, Adam Stambler, Shyam Upadhyay, Manaal Faruqui, Trevor Strohman
  • Publication number: 20230326461
    Abstract: An automated speech recognition (ASR) model includes a first encoder, a first encoder, a second encoder, and a second decoder. The first encoder receives, as input, a sequence of acoustic frames, and generates, at each of a plurality of output steps, a first higher order feature representation for a corresponding acoustic frame in the sequence of acoustic frames. The first decoder receives, as input, the first higher order feature representation generated by the first encoder, and generates a first probability distribution over possible speech recognition hypotheses. The second encoder receives, as input, the first higher order feature representation generated by the first encoder, and generates a second higher order feature representation for a corresponding first higher order feature frame. The second decoder receives, as input, the second higher order feature representation generated by the second encoder, and generates a second probability distribution over possible speech recognition hypotheses.
    Type: Application
    Filed: March 13, 2023
    Publication date: October 12, 2023
    Applicant: Google LLC
    Inventors: Shaojin Ding, Yangzhang He, Xin Wang, Weiran Wang, Trevor Strohman, Tara N. Sainath, Rohit Parkash Prabhavalkar, Robert David, Rina Panigrahy, Rami Botros, Qiao Liang, Ian Mcgraw, Ding Zhao, Dongseong Hwang
  • Publication number: 20230298591
    Abstract: A computer-implemented method includes receiving a sequence of acoustic frames corresponding to an utterance and generating a reference speaker embedding for the utterance. The method also includes receiving a target speaker embedding for a target speaker and generating feature-wise linear modulation (FiLM) parameters including a scaling vector and a shifting vector based on the target speaker embedding. The method also includes generating an affine transformation output that scales and shifts the reference speaker embedding based on the FiLM parameters. The method also includes generating a classification output indicating whether the utterance was spoken by the target speaker based on the affine transformation output.
    Type: Application
    Filed: March 17, 2023
    Publication date: September 21, 2023
    Applicant: Google LLC
    Inventors: Shaojin Ding, Rajeev Rikhye, Qiao Liang, Yanzhang He, Quan Wang, Arun Narayanan, Tom O'Malley, Ian McGraw
  • Publication number: 20230169984
    Abstract: Techniques disclosed herein are directed towards streaming keyphrase detection which can be customized to detect one or more particular keyphrases, without requiring retraining of any model(s) for those particular keyphrase(s). Many implementations include processing audio data using a speaker separation model to generate separated audio data which isolates an utterance spoken by a human speaker from one or more additional sounds not spoken by the human speaker, and processing the separated audio data using a text independent speaker identification model to determine whether a verified and/or registered user spoke a spoken utterance captured in the audio data. Various implementations include processing the audio data and/or the separated audio data using an automatic speech recognition model to generate a text representation of the utterance.
    Type: Application
    Filed: January 30, 2023
    Publication date: June 1, 2023
    Inventors: Rajeev Rikhye, Quan Wang, Yanzhang He, Qiao Liang, Ian C. McGraw
  • Publication number: 20230108275
    Abstract: A method includes receiving a sequence of acoustic frames characterizing one or more utterances as input to a multilingual automated speech recognition (ASR) model. The method also includes generating a higher order feature representation for a corresponding acoustic frame. The method also includes generating a hidden representation based on a sequence of non-blank symbols output by a final softmax layer. The method also includes generating a probability distribution over possible speech recognition hypotheses based on the hidden representation generated by the prediction network at each of the plurality of output steps and the higher order feature representation generated by the encoder at each of the plurality of output steps. The method also includes predicting an end of utterance (EOU) token at an end of each utterance. The method also includes classifying each acoustic frame as either speech, initial silence, intermediate silence, or final silence.
    Type: Application
    Filed: September 22, 2022
    Publication date: April 6, 2023
    Applicant: Google LLC
    Inventors: Bo Li, Tara N. Sainath, Ruoming Pang, Shuo-yin Chang, Qiumin Xu, Trevor Strohman, Vince Chen, Qiao Liang, Heguang Liu, Yanzhang He, Parisa Haghani, Sameer Bidichandani
  • Patent number: 11568878
    Abstract: Techniques disclosed herein are directed towards streaming keyphrase detection which can be customized to detect one or more particular keyphrases, without requiring retraining of any model(s) for those particular keyphrase(s). Many implementations include processing audio data using a speaker separation model to generate separated audio data which isolates an utterance spoken by a human speaker from one or more additional sounds not spoken by the human speaker, and processing the separated audio data using a text independent speaker identification model to determine whether a verified and/or registered user spoke a spoken utterance captured in the audio data. Various implementations include processing the audio data and/or the separated audio data using an automatic speech recognition model to generate a text representation of the utterance.
    Type: Grant
    Filed: April 16, 2021
    Date of Patent: January 31, 2023
    Assignee: GOOGLE LLC
    Inventors: Rajeev Rikhye, Quan Wang, Yanzhang He, Qiao Liang, Ian C. McGraw
  • Publication number: 20220335953
    Abstract: Techniques disclosed herein are directed towards streaming keyphrase detection which can be customized to detect one or more particular keyphrases, without requiring retraining of any model(s) for those particular keyphrase(s). Many implementations include processing audio data using a speaker separation model to generate separated audio data which isolates an utterance spoken by a human speaker from one or more additional sounds not spoken by the human speaker, and processing the separated audio data using a text independent speaker identification model to determine whether a verified and/or registered user spoke a spoken utterance captured in the audio data. Various implementations include processing the audio data and/or the separated audio data using an automatic speech recognition model to generate a text representation of the utterance.
    Type: Application
    Filed: April 16, 2021
    Publication date: October 20, 2022
    Inventors: Rajeev Rikhye, Quan Wang, Yanzhang He, Qiao Liang, Ian C. McGraw
  • Publication number: 20220310072
    Abstract: Two-pass automatic speech recognition (ASR) models can be used to perform streaming on-device ASR to generate a text representation of an utterance captured in audio data. Various implementations include a first-pass portion of the ASR model used to generate streaming candidate recognition(s) of an utterance captured in audio data. For example, the first-pass portion can include a recurrent neural network transformer (RNN-T) decoder. Various implementations include a second-pass portion of the ASR model used to revise the streaming candidate recognition(s) of the utterance and generate a text representation of the utterance. For example, the second-pass portion can include a listen attend spell (LAS) decoder. Various implementations include a shared encoder shared between the RNN-T decoder and the LAS decoder.
    Type: Application
    Filed: June 3, 2020
    Publication date: September 29, 2022
    Inventors: Tara N. Sainath, Ruoming Pang, David Rybach, Yanzhang He, Rohit Prabhavalkar, Wei Li, Mirkó Visontai, Qiao Liang, Trevor Strohman, Yonghui Wu, Ian C. McGraw, Chung-Cheng Chiu
  • Patent number: 9030846
    Abstract: A resonant power conversion apparatus including a transformer-based resonant converter and first and second switch control units is provided. The transformer-based resonant converter includes a primary switch circuit and a secondary output circuit configured to provide an output voltage to a load. The first switch control unit is configured to control an ON/OFF operation of the primary switch circuit in response to a status of the load. The second switch control unit is configured to determine whether to activate or inactivate the first switch control unit. When the status of the load is the light-loading or the no-loading, the first switch control unit intermittently controls the ON/OFF operation of the primary switch circuit, and meanwhile, the first switch control unit is inactivated during the primary switch circuit is disabled, so as to substantially reduce the light-loading or no-loading loss of the resonant power conversion apparatus.
    Type: Grant
    Filed: September 28, 2012
    Date of Patent: May 12, 2015
    Assignees: FSP Technology Inc., FSP-Powerland Technology Inc.
    Inventors: Yong-Jiang Bai, Qiao-Liang Chen, Ju-Lu Sun
  • Patent number: 8780597
    Abstract: An AC-to-DC conversion apparatus is provided, and which includes a first switch-element, an output capacitor and a bridgeless power-factor-correction (PFC) circuit. The bridgeless PFC circuit is coupled to an AC input, and includes a first inductor, a second inductor and a bridge circuit constructed by second to fifth switch-elements. The first switch-element is connected between bridgeless PFC circuit and the output capacitor. Under such circuit configuration and suitable control manner, the common-mode interference in the provided AC-to-DC conversion apparatus is lowered and thus reducing the power loss.
    Type: Grant
    Filed: July 12, 2012
    Date of Patent: July 15, 2014
    Assignees: FSP Technology Inc., FSP-Powerland Technology Inc.
    Inventors: Ming Xu, Qiao-Liang Chen
  • Patent number: 8710875
    Abstract: A bootstrap gate driver including a load indication unit, a bootstrap gate-drive unit and a drive-control unit is provided. The load indication unit is configured to generate a load indication signal in response to a state of a load. The bootstrap gate-drive unit is configured to drive a switch-transistor circuit in response to an inputted pulse-width-modulation (PWM) signal, wherein the switch-transistor circuit has a high-side driving path and a low-side driving path. The drive-control unit is coupled to the load indication unit and the bootstrap gate-drive unit, and configured to enable or disable the high-side driving path in response to the load indication signal. In the invention, the operation of the low-side driving path is not affected by enabling or disabling the high-side driving path.
    Type: Grant
    Filed: May 18, 2012
    Date of Patent: April 29, 2014
    Assignees: FSP Technology Inc., FSP-Powerland Technology Inc.
    Inventors: Yong-Jiang Bai, Qiao-Liang Chen, Ning-Bin Wang, Ju-Lu Sun
  • Patent number: 8587963
    Abstract: A resonant converter equipped with a phase shifting output circuit includes a resonant circuit to receive input power and regulate to become at least one resonant power, a switch unit to switch an ON period for the input power to pass through the resonant circuit and a power transformation circuit to regulate the resonant power and output a transformed power. The resonant converter further has a primary output circuit and at least one secondary output circuit. The primary output circuit regulates the transformed power to become a primary output power. A resonant control unit captures a feedback signal from the primary output circuit and generates a resonant control signal. A phase shifting control unit receives the resonant control signal and regulate to become a phase shifting driving signal. The secondary output circuit is controlled by the phase shifting driving signal and provides a secondary output power.
    Type: Grant
    Filed: January 21, 2009
    Date of Patent: November 19, 2013
    Assignee: FSP Technology Inc.
    Inventors: Zhanghe Nan, Ming Xu, Qiao Liang Chen
  • Publication number: 20130083564
    Abstract: A resonant power conversion apparatus including a transformer-based resonant converter and first and second switch control units is provided. The transformer-based resonant converter includes a primary switch circuit and a secondary output circuit configured to provide an output voltage to a load. The first switch control unit is configured to control an ON/OFF operation of the primary switch circuit in response to a status of the load. The second switch control unit is configured to determine whether to activate or inactivate the first switch control unit. When the status of the load is the light-loading or the no-loading, the first switch control unit intermittently controls the ON/OFF operation of the primary switch circuit, and meanwhile, the first switch control unit is inactivated during the primary switch circuit is disabled, so as to substantially reduce the light-loading or no-loading loss of the resonant power conversion apparatus.
    Type: Application
    Filed: September 28, 2012
    Publication date: April 4, 2013
    Inventors: Yong-Jiang Bai, Qiao-Liang Chen, Ju-Lu Sun
  • Publication number: 20130016545
    Abstract: An AC-to-DC conversion apparatus is provided, and which includes a first switch-element, an output capacitor and a bridgeless power-factor-correction (PFC) circuit. The bridgeless PFC circuit is coupled to an AC input, and includes a first inductor, a second inductor and a bridge circuit constructed by second to fifth switch-elements. The first switch-element is connected between bridgeless PFC circuit and the output capacitor. Under such circuit configuration and suitable control manner, the common-mode interference in the provided AC-to-DC conversion apparatus is lowered and thus reducing the power loss.
    Type: Application
    Filed: July 12, 2012
    Publication date: January 17, 2013
    Applicants: FSP-POWERLAND TECHNOLOGY INC., FSP TECHNOLOGY INC.
    Inventors: Ming Xu, Qiao-Liang Chen
  • Publication number: 20120293219
    Abstract: A bootstrap gate driver including a load indication unit, a bootstrap gate-drive unit and a drive-control unit is provided. The load indication unit is configured to generate a load indication signal in response to a state of a load. The bootstrap gate-drive unit is configured to drive a switch-transistor circuit in response to an inputted pulse-width-modulation (PWM) signal, wherein the switch-transistor circuit has a high-side driving path and a low-side driving path. The drive-control unit is coupled to the load indication unit and the bootstrap gate-drive unit, and configured to enable or disable the high-side driving path in response to the load indication signal. In the invention, the operation of the low-side driving path is not affected by enabling or disabling the high-side driving path.
    Type: Application
    Filed: May 18, 2012
    Publication date: November 22, 2012
    Applicants: FSP-POWERLAND TECHNOLOGY INC., FSP TECHNOLOGY INC.
    Inventors: Yong-Jiang Bai, Qiao-Liang Chen, Ning-Bin Wang, Ju-Lu Sun