Patents by Inventor Changliang LIU

Changliang LIU has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20250005339
    Abstract: Representative embodiments disclose machine learning classifiers used in scenarios such as speech recognition, image captioning, machine translation, or other sequence-to-sequence embodiments. The machine learning classifiers have a plurality of time layers, each layer having a time processing block and a depth processing block. The time processing block is a recurrent neural network such as a Long Short Term Memory (LSTM) network. The depth processing blocks can be an LSTM network, a gated Deep Neural Network (DNN) or a maxout DNN. The depth processing blocks account for the hidden states of each time layer and uses summarized layer information for final input signal feature classification. An attention layer can also be used between the top depth processing block and the output layer.
    Type: Application
    Filed: September 10, 2024
    Publication date: January 2, 2025
    Inventors: Jinyu LI, Liang LU, Changliang LIU, Yifan GONG
  • Patent number: 12086704
    Abstract: Representative embodiments disclose machine learning classifiers used in scenarios such as speech recognition, image captioning, machine translation, or other sequence-to-sequence embodiments. The machine learning classifiers have a plurality of time layers, each layer having a time processing block and a depth processing block. The time processing block is a recurrent neural network such as a Long Short Term Memory (LSTM) network. The depth processing blocks can be an LSTM network, a gated Deep Neural Network (DNN) or a maxout DNN. The depth processing blocks account for the hidden states of each time layer and uses summarized layer information for final input signal feature classification. An attention layer can also be used between the top depth processing block and the output layer.
    Type: Grant
    Filed: November 3, 2021
    Date of Patent: September 10, 2024
    Assignee: MICROSOFT TECHNOLOGY LICENSING, LLC
    Inventors: Jinyu Li, Liang Lu, Changliang Liu, Yifan Gong
  • Publication number: 20230401392
    Abstract: A data processing system is implemented for receiving speech data for a plurality of languages, and determining letters from the speech data. The data processing system also implements normalizing the speech data by applying linguistic based rules for Latin-based languages on the determined letters, building a computer model using the normalized speech data, fine-tuning the computer model using additional speech data, and recognizing words in a target language using the fine-tuned computer model.
    Type: Application
    Filed: June 9, 2022
    Publication date: December 14, 2023
    Applicant: Microsoft Technology Licensing, LLC
    Inventors: Kshitiz KUMAR, Jian WU, Bo REN, Tianyu WU, Fahimeh BAHMANINEZHAD, Edward C. LIN, Xiaoyang CHEN, Changliang LIU
  • Patent number: 11676006
    Abstract: According to some embodiments, a universal modeling system may include a plurality of domain expert models to each receive raw input data (e.g., a stream of audio frames containing speech utterances) and provide a domain expert output based on the raw input data. A neural mixture component may then generate a weight corresponding to each domain expert model based on information created by the plurality of domain expert models (e.g., hidden features and/or row convolution). The weights might be associated with, for example, constrained scalar numbers, unconstrained scaler numbers, vectors, matrices, etc. An output layer may provide a universal modeling system output (e.g., an automatic speech recognition result) based on each domain expert output after being multiplied by the corresponding weight for that domain expert model.
    Type: Grant
    Filed: May 16, 2019
    Date of Patent: June 13, 2023
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Amit Das, Jinyu Li, Changliang Liu, Yifan Gong
  • Patent number: 11644974
    Abstract: A time-sharing wave recording method is provided. Firstly, N1 operating variables are selected, and the storage addresses of the N1 operating variables are mapped to N1 index variables. Then, N2 record channels in a sequential relationship are provided, and a mapping relationship between each record channel and the N1 index variables is established. Then, the values of the operating variables are assigned to a first record channel of the N2 record channels in response to a rising edge of a pulse of a clock signal and a start triggering signal, the values of the operating variables are sequentially assigned to the rest of the N2 record channels in response to a rising edge of each pulse of the clock signal and the start triggering signal, and the assigned values of the operating variables for the N2 record channels are recorded to the memory.
    Type: Grant
    Filed: October 16, 2019
    Date of Patent: May 9, 2023
    Assignee: DELTA ELECTRONICS (SHANGHAI) CO., LTD.
    Inventors: Guojin Xu, Guoqiao Shen, Yunfeng Liu, Qingsong Tao, Changliang Liu, Jinfa Zhang
  • Patent number: 11631399
    Abstract: According to some embodiments, a machine learning model may include an input layer to receive an input signal as a series of frames representing handwriting data, speech data, audio data, and/or textual data. A plurality of time layers may be provided, and each time layer may comprise a uni-directional recurrent neural network processing block. A depth processing block may scan hidden states of the recurrent neural network processing block of each time layer, and the depth processing block may be associated with a first frame and receive context frame information of a sequence of one or more future frames relative to the first frame. An output layer may output a final classification as a classified posterior vector of the input signal. For example, the depth processing block may receive the context from information from an output of a time layer processing block or another depth processing block of the future frame.
    Type: Grant
    Filed: May 13, 2019
    Date of Patent: April 18, 2023
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Jinyu Li, Vadim Mazalov, Changliang Liu, Liang Lu, Yifan Gong
  • Patent number: 11445295
    Abstract: A system and method include reception of a first plurality of audio signals, generation of a second plurality of beamformed audio signals based on the first plurality of audio signals, each of the second plurality of beamformed audio signals associated with a respective one of a second plurality of beamformer directions, generation of a first TF mask for a first output channel based on the first plurality of audio signals, determination of a first beamformer direction associated with a first target sound source based on the first TF mask, generation of first features based on the first beamformer direction and the first plurality of audio signals, determination of a second TF mask based on the first features, and application of the second TF mask to one of the second plurality of beamformed audio signals associated with the first beamformer direction.
    Type: Grant
    Filed: November 17, 2020
    Date of Patent: September 13, 2022
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Zhuo Chen, Changliang Liu, Takuya Yoshioka, Xiong Xiao, Hakan Erdogan, Dimitrios Basile Dimitriadis
  • Patent number: 11394221
    Abstract: A method for controlling a DC bus voltage in a DC bus system, the system including a DC bus and an energy storage unit coupled to the DC bus, includes: detecting a DC bus voltage; detecting, by the energy storage unit, a DC bus voltage of the DC bus; determining, by the energy storage unit, a power reference value based on the DC bus voltage; and adjusting, by the energy storage unit, one of output power and absorbing power based on the power reference value.
    Type: Grant
    Filed: June 4, 2019
    Date of Patent: July 19, 2022
    Assignee: Delta Electronics, Inc.
    Inventors: Guoqiao Shen, Guojin Xu, Changliang Liu, Jian Li, Yunfeng Liu, Jinfa Zhang
  • Publication number: 20220058442
    Abstract: Representative embodiments disclose machine learning classifiers used in scenarios such as speech recognition, image captioning, machine translation, or other sequence-to-sequence embodiments. The machine learning classifiers have a plurality of time layers, each layer having a time processing block and a depth processing block. The time processing block is a recurrent neural network such as a Long Short Term Memory (LSTM) network. The depth processing blocks can be an LSTM network, a gated Deep Neural Network (DNN) or a maxout DNN. The depth processing blocks account for the hidden states of each time layer and uses summarized layer information for final input signal feature classification. An attention layer can also be used between the top depth processing block and the output layer.
    Type: Application
    Filed: November 3, 2021
    Publication date: February 24, 2022
    Inventors: Jinyu Li, Liang Lu, Changliang Liu, Yifan Gong
  • Patent number: 11244673
    Abstract: Streaming machine learning unidirectional models is facilitated by the use of embedding vectors. Processing blocks in the models apply embedding vectors as input. The embedding vectors utilize context of future data (e.g., data that is temporally offset into the future within a data stream) to improve the accuracy of the outputs generated by the processing blocks. The embedding vectors cause a temporal shift between the outputs of the processing blocks and the inputs to which the outputs correspond. This temporal shift enables the processing blocks to apply the embedding vector inputs from processing blocks that are associated with future data.
    Type: Grant
    Filed: July 19, 2019
    Date of Patent: February 8, 2022
    Assignee: MICROSOFT TECHNOLOGLY LICENSING, LLC
    Inventors: Jinyu Li, Amit Kumar Agarwal, Yifan Gong, Harini Kesavamoorthy, Changliang Liu, Liang Lu
  • Patent number: 11210565
    Abstract: Representative embodiments disclose machine learning classifiers used in scenarios such as speech recognition, image captioning, machine translation, or other sequence-to-sequence embodiments. The machine learning classifiers have a plurality of time layers, each layer having a time processing block and a depth processing block. The time processing block is a recurrent neural network such as a Long Short Term Memory (LSTM) network. The depth processing blocks can be an LSTM network, a gated Deep Neural Network (DNN) or a maxout DNN. The depth processing blocks account for the hidden states of each time layer and uses summarized layer information for final input signal feature classification. An attention layer can also be used between the top depth processing block and the output layer.
    Type: Grant
    Filed: November 30, 2018
    Date of Patent: December 28, 2021
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Jinyu Li, Liang Lu, Changliang Liu, Yifan Gong
  • Publication number: 20210076129
    Abstract: A system and method include reception of a first plurality of audio signals, generation of a second plurality of beamformed audio signals based on the first plurality of audio signals, each of the second plurality of beamformed audio signals associated with a respective one of a second plurality of beamformer directions, generation of a first TF mask for a first output channel based on the first plurality of audio signals, determination of a first beamformer direction associated with a first target sound source based on the first TF mask, generation of first features based on the first beamformer direction and the first plurality of audio signals, determination of a second TF mask based on the first features, and application of the second TF mask to one of the second plurality of beamformed audio signals associated with the first beamformer direction.
    Type: Application
    Filed: November 17, 2020
    Publication date: March 11, 2021
    Inventors: Zhuo CHEN, Changliang LIU, Takuya YOSHIOKA, Xiong XIAO, Hakan ERDOGAN, Dimitrios Basile DIMITRIADIS
  • Patent number: 10856076
    Abstract: A system and method include reception of a first plurality of audio signals, generation of a second plurality of beamformed audio signals based on the first plurality of audio signals, each of the second plurality of beamformed audio signals associated with a respective one of a second plurality of beamformer directions, generation of a first TF mask for a first output channel based on the first plurality of audio signals, determination of a first beamformer direction associated with a first target sound source based on the first TF mask, generation of first features based on the first beamformer direction and the first plurality of audio signals, determination of a second TF mask based on the first features, and application of the second TF mask to one of the second plurality of beamformed audio signals associated with the first beamformer direction.
    Type: Grant
    Filed: April 5, 2019
    Date of Patent: December 1, 2020
    Assignee: MICROSOFT TECHNOLOGY LICENSING, LLC
    Inventors: Zhuo Chen, Changliang Liu, Takuya Yoshioka, Xiong Xiao, Hakan Erdogan, Dimitrios Basile Dimitriadis
  • Publication number: 20200335119
    Abstract: Embodiments are associated with determination of a first plurality of multi-dimensional vectors, each of the first plurality of multi-dimensional vectors representing speech of a target speaker, determination of a multi-dimensional vector representing a speech signal of two or more speakers, determination of a weighted vector representing speech of the target speaker based on the first plurality of multi-dimensional vectors and on similarities between the multi-dimensional vector and each of the first plurality of multi-dimensional vectors, and extraction of speech of the target speaker from the speech signal based on the weighted vector and the speech signal.
    Type: Application
    Filed: June 7, 2019
    Publication date: October 22, 2020
    Inventors: Xiong XIAO, Zhuo CHEN, Takuya YOSHIOKA, Changliang LIU, Hakan ERDOGAN, Dimitrios Basile DIMITRIADIS, Yifan GONG, James Garnet Droppo, III
  • Publication number: 20200334527
    Abstract: According to some embodiments, a universal modeling system may include a plurality of domain expert models to each receive raw input data (e.g., a stream of audio frames containing speech utterances) and provide a domain expert output based on the raw input data. A neural mixture component may then generate a weight corresponding to each domain expert model based on information created by the plurality of domain expert models (e.g., hidden features and/or row convolution). The weights might be associated with, for example, constrained scalar numbers, unconstrained scaler numbers, vectors, matrices, etc. An output layer may provide a universal modeling system output (e.g., an automatic speech recognition result) based on each domain expert output after being multiplied by the corresponding weight for that domain expert model.
    Type: Application
    Filed: May 16, 2019
    Publication date: October 22, 2020
    Inventors: Amit DAS, Jinyu LI, Changliang LIU, Yifan GONG
  • Publication number: 20200334526
    Abstract: According to some embodiments, a machine learning model may include an input layer to receive an input signal as a series of frames representing handwriting data, speech data, audio data, and/or textual data. A plurality of time layers may be provided, and each time layer may comprise a uni-directional recurrent neural network processing block. A depth processing block may scan hidden states of the recurrent neural network processing block of each time layer, and the depth processing block may be associated with a first frame and receive context frame information of a sequence of one or more future frames relative to the first frame. An output layer may output a final classification as a classified posterior vector of the input signal. For example, the depth processing block may receive the context from information from an output of a time layer processing block or another depth processing block of the future frame.
    Type: Application
    Filed: May 13, 2019
    Publication date: October 22, 2020
    Inventors: Jinyu LI, Vadim MAZALOV, Changliang LIU, Liang LU, Yifan GONG
  • Publication number: 20200322722
    Abstract: A system and method include reception of a first plurality of audio signals, generation of a second plurality of beamformed audio signals based on the first plurality of audio signals, each of the second plurality of beamformed audio signals associated with a respective one of a second plurality of beamformer directions, generation of a first TF mask for a first output channel based on the first plurality of audio signals, determination of a first beamformer direction associated with a first target sound source based on the first TF mask, generation of first features based on the first beamformer direction and the first plurality of audio signals, determination of a second TF mask based on the first features, and application of the second TF mask to one of the second plurality of beamformed audio signals associated with the first beamformer direction.
    Type: Application
    Filed: April 5, 2019
    Publication date: October 8, 2020
    Inventors: Zhuo CHEN, Changliang LIU, Takuya YOSHIOKA, Xiong XIAO, Hakan ERDOGAN, Dimitrios Basile DIMITRIADIS
  • Publication number: 20200175335
    Abstract: Representative embodiments disclose machine learning classifiers used in scenarios such as speech recognition, image captioning, machine translation, or other sequence-to-sequence embodiments. The machine learning classifiers have a plurality of time layers, each layer having a time processing block and a depth processing block. The time processing block is a recurrent neural network such as a Long Short Term Memory (LSTM) network. The depth processing blocks can be an LSTM network, a gated Deep Neural Network (DNN) or a maxout DNN. The depth processing blocks account for the hidden states of each time layer and uses summarized layer information for final input signal feature classification. An attention layer can also be used between the top depth processing block and the output layer.
    Type: Application
    Filed: November 30, 2018
    Publication date: June 4, 2020
    Inventors: Jinyu Li, Liang Lu, Changliang Liu, Yifan Gong
  • Publication number: 20200142592
    Abstract: A time-sharing wave recording method is provided. Firstly, N1 operating variables are selected, and the storage addresses of the N1 operating variables are mapped to N1 index variables. Then, N2 record channels in a sequential relationship are provided, and a mapping relationship between each record channel and the N1 index variables is established. Then, the values of the operating variables are assigned to a first record channel of the N2 record channels in response to a rising edge of a pulse of a clock signal and a start triggering signal, the values of the operating variables are sequentially assigned to the rest of the N2 record channels in response to a rising edge of each pulse of the clock signal and the start triggering signal, and the assigned values of the operating variables for the N2 record channels are recorded to the memory.
    Type: Application
    Filed: October 16, 2019
    Publication date: May 7, 2020
    Inventors: Guojin Xu, Guoqiao Shen, Yunfeng Liu, Qingsong Tao, Changliang Liu, Jinfa Zhang
  • Publication number: 20190372380
    Abstract: A method for controlling a DC bus voltage in a DC bus system, the system including a DC bus and an energy storage unit coupled to the DC bus, includes: detecting a DC bus voltage; detecting, by the energy storage unit, a DC bus voltage of the DC bus; determining, by the energy storage unit, a power reference value based on the DC bus voltage; and adjusting, by the energy storage unit, one of output power and absorbing power based on the power reference value.
    Type: Application
    Filed: June 4, 2019
    Publication date: December 5, 2019
    Applicant: Delta Electronics,Inc.
    Inventors: Guoqiao SHEN, Guojin XU, Changliang LIU, Jian LI, Yunfeng LIU, Jinfa ZHANG