Patents by Inventor Changliang LIU
Changliang LIU has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Publication number: 20250005339Abstract: Representative embodiments disclose machine learning classifiers used in scenarios such as speech recognition, image captioning, machine translation, or other sequence-to-sequence embodiments. The machine learning classifiers have a plurality of time layers, each layer having a time processing block and a depth processing block. The time processing block is a recurrent neural network such as a Long Short Term Memory (LSTM) network. The depth processing blocks can be an LSTM network, a gated Deep Neural Network (DNN) or a maxout DNN. The depth processing blocks account for the hidden states of each time layer and uses summarized layer information for final input signal feature classification. An attention layer can also be used between the top depth processing block and the output layer.Type: ApplicationFiled: September 10, 2024Publication date: January 2, 2025Inventors: Jinyu LI, Liang LU, Changliang LIU, Yifan GONG
-
Patent number: 12086704Abstract: Representative embodiments disclose machine learning classifiers used in scenarios such as speech recognition, image captioning, machine translation, or other sequence-to-sequence embodiments. The machine learning classifiers have a plurality of time layers, each layer having a time processing block and a depth processing block. The time processing block is a recurrent neural network such as a Long Short Term Memory (LSTM) network. The depth processing blocks can be an LSTM network, a gated Deep Neural Network (DNN) or a maxout DNN. The depth processing blocks account for the hidden states of each time layer and uses summarized layer information for final input signal feature classification. An attention layer can also be used between the top depth processing block and the output layer.Type: GrantFiled: November 3, 2021Date of Patent: September 10, 2024Assignee: MICROSOFT TECHNOLOGY LICENSING, LLCInventors: Jinyu Li, Liang Lu, Changliang Liu, Yifan Gong
-
Publication number: 20230401392Abstract: A data processing system is implemented for receiving speech data for a plurality of languages, and determining letters from the speech data. The data processing system also implements normalizing the speech data by applying linguistic based rules for Latin-based languages on the determined letters, building a computer model using the normalized speech data, fine-tuning the computer model using additional speech data, and recognizing words in a target language using the fine-tuned computer model.Type: ApplicationFiled: June 9, 2022Publication date: December 14, 2023Applicant: Microsoft Technology Licensing, LLCInventors: Kshitiz KUMAR, Jian WU, Bo REN, Tianyu WU, Fahimeh BAHMANINEZHAD, Edward C. LIN, Xiaoyang CHEN, Changliang LIU
-
Patent number: 11676006Abstract: According to some embodiments, a universal modeling system may include a plurality of domain expert models to each receive raw input data (e.g., a stream of audio frames containing speech utterances) and provide a domain expert output based on the raw input data. A neural mixture component may then generate a weight corresponding to each domain expert model based on information created by the plurality of domain expert models (e.g., hidden features and/or row convolution). The weights might be associated with, for example, constrained scalar numbers, unconstrained scaler numbers, vectors, matrices, etc. An output layer may provide a universal modeling system output (e.g., an automatic speech recognition result) based on each domain expert output after being multiplied by the corresponding weight for that domain expert model.Type: GrantFiled: May 16, 2019Date of Patent: June 13, 2023Assignee: Microsoft Technology Licensing, LLCInventors: Amit Das, Jinyu Li, Changliang Liu, Yifan Gong
-
Patent number: 11644974Abstract: A time-sharing wave recording method is provided. Firstly, N1 operating variables are selected, and the storage addresses of the N1 operating variables are mapped to N1 index variables. Then, N2 record channels in a sequential relationship are provided, and a mapping relationship between each record channel and the N1 index variables is established. Then, the values of the operating variables are assigned to a first record channel of the N2 record channels in response to a rising edge of a pulse of a clock signal and a start triggering signal, the values of the operating variables are sequentially assigned to the rest of the N2 record channels in response to a rising edge of each pulse of the clock signal and the start triggering signal, and the assigned values of the operating variables for the N2 record channels are recorded to the memory.Type: GrantFiled: October 16, 2019Date of Patent: May 9, 2023Assignee: DELTA ELECTRONICS (SHANGHAI) CO., LTD.Inventors: Guojin Xu, Guoqiao Shen, Yunfeng Liu, Qingsong Tao, Changliang Liu, Jinfa Zhang
-
Patent number: 11631399Abstract: According to some embodiments, a machine learning model may include an input layer to receive an input signal as a series of frames representing handwriting data, speech data, audio data, and/or textual data. A plurality of time layers may be provided, and each time layer may comprise a uni-directional recurrent neural network processing block. A depth processing block may scan hidden states of the recurrent neural network processing block of each time layer, and the depth processing block may be associated with a first frame and receive context frame information of a sequence of one or more future frames relative to the first frame. An output layer may output a final classification as a classified posterior vector of the input signal. For example, the depth processing block may receive the context from information from an output of a time layer processing block or another depth processing block of the future frame.Type: GrantFiled: May 13, 2019Date of Patent: April 18, 2023Assignee: Microsoft Technology Licensing, LLCInventors: Jinyu Li, Vadim Mazalov, Changliang Liu, Liang Lu, Yifan Gong
-
Patent number: 11445295Abstract: A system and method include reception of a first plurality of audio signals, generation of a second plurality of beamformed audio signals based on the first plurality of audio signals, each of the second plurality of beamformed audio signals associated with a respective one of a second plurality of beamformer directions, generation of a first TF mask for a first output channel based on the first plurality of audio signals, determination of a first beamformer direction associated with a first target sound source based on the first TF mask, generation of first features based on the first beamformer direction and the first plurality of audio signals, determination of a second TF mask based on the first features, and application of the second TF mask to one of the second plurality of beamformed audio signals associated with the first beamformer direction.Type: GrantFiled: November 17, 2020Date of Patent: September 13, 2022Assignee: Microsoft Technology Licensing, LLCInventors: Zhuo Chen, Changliang Liu, Takuya Yoshioka, Xiong Xiao, Hakan Erdogan, Dimitrios Basile Dimitriadis
-
Patent number: 11394221Abstract: A method for controlling a DC bus voltage in a DC bus system, the system including a DC bus and an energy storage unit coupled to the DC bus, includes: detecting a DC bus voltage; detecting, by the energy storage unit, a DC bus voltage of the DC bus; determining, by the energy storage unit, a power reference value based on the DC bus voltage; and adjusting, by the energy storage unit, one of output power and absorbing power based on the power reference value.Type: GrantFiled: June 4, 2019Date of Patent: July 19, 2022Assignee: Delta Electronics, Inc.Inventors: Guoqiao Shen, Guojin Xu, Changliang Liu, Jian Li, Yunfeng Liu, Jinfa Zhang
-
Publication number: 20220058442Abstract: Representative embodiments disclose machine learning classifiers used in scenarios such as speech recognition, image captioning, machine translation, or other sequence-to-sequence embodiments. The machine learning classifiers have a plurality of time layers, each layer having a time processing block and a depth processing block. The time processing block is a recurrent neural network such as a Long Short Term Memory (LSTM) network. The depth processing blocks can be an LSTM network, a gated Deep Neural Network (DNN) or a maxout DNN. The depth processing blocks account for the hidden states of each time layer and uses summarized layer information for final input signal feature classification. An attention layer can also be used between the top depth processing block and the output layer.Type: ApplicationFiled: November 3, 2021Publication date: February 24, 2022Inventors: Jinyu Li, Liang Lu, Changliang Liu, Yifan Gong
-
Patent number: 11244673Abstract: Streaming machine learning unidirectional models is facilitated by the use of embedding vectors. Processing blocks in the models apply embedding vectors as input. The embedding vectors utilize context of future data (e.g., data that is temporally offset into the future within a data stream) to improve the accuracy of the outputs generated by the processing blocks. The embedding vectors cause a temporal shift between the outputs of the processing blocks and the inputs to which the outputs correspond. This temporal shift enables the processing blocks to apply the embedding vector inputs from processing blocks that are associated with future data.Type: GrantFiled: July 19, 2019Date of Patent: February 8, 2022Assignee: MICROSOFT TECHNOLOGLY LICENSING, LLCInventors: Jinyu Li, Amit Kumar Agarwal, Yifan Gong, Harini Kesavamoorthy, Changliang Liu, Liang Lu
-
Patent number: 11210565Abstract: Representative embodiments disclose machine learning classifiers used in scenarios such as speech recognition, image captioning, machine translation, or other sequence-to-sequence embodiments. The machine learning classifiers have a plurality of time layers, each layer having a time processing block and a depth processing block. The time processing block is a recurrent neural network such as a Long Short Term Memory (LSTM) network. The depth processing blocks can be an LSTM network, a gated Deep Neural Network (DNN) or a maxout DNN. The depth processing blocks account for the hidden states of each time layer and uses summarized layer information for final input signal feature classification. An attention layer can also be used between the top depth processing block and the output layer.Type: GrantFiled: November 30, 2018Date of Patent: December 28, 2021Assignee: Microsoft Technology Licensing, LLCInventors: Jinyu Li, Liang Lu, Changliang Liu, Yifan Gong
-
Publication number: 20210076129Abstract: A system and method include reception of a first plurality of audio signals, generation of a second plurality of beamformed audio signals based on the first plurality of audio signals, each of the second plurality of beamformed audio signals associated with a respective one of a second plurality of beamformer directions, generation of a first TF mask for a first output channel based on the first plurality of audio signals, determination of a first beamformer direction associated with a first target sound source based on the first TF mask, generation of first features based on the first beamformer direction and the first plurality of audio signals, determination of a second TF mask based on the first features, and application of the second TF mask to one of the second plurality of beamformed audio signals associated with the first beamformer direction.Type: ApplicationFiled: November 17, 2020Publication date: March 11, 2021Inventors: Zhuo CHEN, Changliang LIU, Takuya YOSHIOKA, Xiong XIAO, Hakan ERDOGAN, Dimitrios Basile DIMITRIADIS
-
Patent number: 10856076Abstract: A system and method include reception of a first plurality of audio signals, generation of a second plurality of beamformed audio signals based on the first plurality of audio signals, each of the second plurality of beamformed audio signals associated with a respective one of a second plurality of beamformer directions, generation of a first TF mask for a first output channel based on the first plurality of audio signals, determination of a first beamformer direction associated with a first target sound source based on the first TF mask, generation of first features based on the first beamformer direction and the first plurality of audio signals, determination of a second TF mask based on the first features, and application of the second TF mask to one of the second plurality of beamformed audio signals associated with the first beamformer direction.Type: GrantFiled: April 5, 2019Date of Patent: December 1, 2020Assignee: MICROSOFT TECHNOLOGY LICENSING, LLCInventors: Zhuo Chen, Changliang Liu, Takuya Yoshioka, Xiong Xiao, Hakan Erdogan, Dimitrios Basile Dimitriadis
-
Publication number: 20200335119Abstract: Embodiments are associated with determination of a first plurality of multi-dimensional vectors, each of the first plurality of multi-dimensional vectors representing speech of a target speaker, determination of a multi-dimensional vector representing a speech signal of two or more speakers, determination of a weighted vector representing speech of the target speaker based on the first plurality of multi-dimensional vectors and on similarities between the multi-dimensional vector and each of the first plurality of multi-dimensional vectors, and extraction of speech of the target speaker from the speech signal based on the weighted vector and the speech signal.Type: ApplicationFiled: June 7, 2019Publication date: October 22, 2020Inventors: Xiong XIAO, Zhuo CHEN, Takuya YOSHIOKA, Changliang LIU, Hakan ERDOGAN, Dimitrios Basile DIMITRIADIS, Yifan GONG, James Garnet Droppo, III
-
Publication number: 20200334527Abstract: According to some embodiments, a universal modeling system may include a plurality of domain expert models to each receive raw input data (e.g., a stream of audio frames containing speech utterances) and provide a domain expert output based on the raw input data. A neural mixture component may then generate a weight corresponding to each domain expert model based on information created by the plurality of domain expert models (e.g., hidden features and/or row convolution). The weights might be associated with, for example, constrained scalar numbers, unconstrained scaler numbers, vectors, matrices, etc. An output layer may provide a universal modeling system output (e.g., an automatic speech recognition result) based on each domain expert output after being multiplied by the corresponding weight for that domain expert model.Type: ApplicationFiled: May 16, 2019Publication date: October 22, 2020Inventors: Amit DAS, Jinyu LI, Changliang LIU, Yifan GONG
-
Publication number: 20200334526Abstract: According to some embodiments, a machine learning model may include an input layer to receive an input signal as a series of frames representing handwriting data, speech data, audio data, and/or textual data. A plurality of time layers may be provided, and each time layer may comprise a uni-directional recurrent neural network processing block. A depth processing block may scan hidden states of the recurrent neural network processing block of each time layer, and the depth processing block may be associated with a first frame and receive context frame information of a sequence of one or more future frames relative to the first frame. An output layer may output a final classification as a classified posterior vector of the input signal. For example, the depth processing block may receive the context from information from an output of a time layer processing block or another depth processing block of the future frame.Type: ApplicationFiled: May 13, 2019Publication date: October 22, 2020Inventors: Jinyu LI, Vadim MAZALOV, Changliang LIU, Liang LU, Yifan GONG
-
Publication number: 20200322722Abstract: A system and method include reception of a first plurality of audio signals, generation of a second plurality of beamformed audio signals based on the first plurality of audio signals, each of the second plurality of beamformed audio signals associated with a respective one of a second plurality of beamformer directions, generation of a first TF mask for a first output channel based on the first plurality of audio signals, determination of a first beamformer direction associated with a first target sound source based on the first TF mask, generation of first features based on the first beamformer direction and the first plurality of audio signals, determination of a second TF mask based on the first features, and application of the second TF mask to one of the second plurality of beamformed audio signals associated with the first beamformer direction.Type: ApplicationFiled: April 5, 2019Publication date: October 8, 2020Inventors: Zhuo CHEN, Changliang LIU, Takuya YOSHIOKA, Xiong XIAO, Hakan ERDOGAN, Dimitrios Basile DIMITRIADIS
-
Publication number: 20200175335Abstract: Representative embodiments disclose machine learning classifiers used in scenarios such as speech recognition, image captioning, machine translation, or other sequence-to-sequence embodiments. The machine learning classifiers have a plurality of time layers, each layer having a time processing block and a depth processing block. The time processing block is a recurrent neural network such as a Long Short Term Memory (LSTM) network. The depth processing blocks can be an LSTM network, a gated Deep Neural Network (DNN) or a maxout DNN. The depth processing blocks account for the hidden states of each time layer and uses summarized layer information for final input signal feature classification. An attention layer can also be used between the top depth processing block and the output layer.Type: ApplicationFiled: November 30, 2018Publication date: June 4, 2020Inventors: Jinyu Li, Liang Lu, Changliang Liu, Yifan Gong
-
Publication number: 20200142592Abstract: A time-sharing wave recording method is provided. Firstly, N1 operating variables are selected, and the storage addresses of the N1 operating variables are mapped to N1 index variables. Then, N2 record channels in a sequential relationship are provided, and a mapping relationship between each record channel and the N1 index variables is established. Then, the values of the operating variables are assigned to a first record channel of the N2 record channels in response to a rising edge of a pulse of a clock signal and a start triggering signal, the values of the operating variables are sequentially assigned to the rest of the N2 record channels in response to a rising edge of each pulse of the clock signal and the start triggering signal, and the assigned values of the operating variables for the N2 record channels are recorded to the memory.Type: ApplicationFiled: October 16, 2019Publication date: May 7, 2020Inventors: Guojin Xu, Guoqiao Shen, Yunfeng Liu, Qingsong Tao, Changliang Liu, Jinfa Zhang
-
Publication number: 20190372380Abstract: A method for controlling a DC bus voltage in a DC bus system, the system including a DC bus and an energy storage unit coupled to the DC bus, includes: detecting a DC bus voltage; detecting, by the energy storage unit, a DC bus voltage of the DC bus; determining, by the energy storage unit, a power reference value based on the DC bus voltage; and adjusting, by the energy storage unit, one of output power and absorbing power based on the power reference value.Type: ApplicationFiled: June 4, 2019Publication date: December 5, 2019Applicant: Delta Electronics,Inc.Inventors: Guoqiao SHEN, Guojin XU, Changliang LIU, Jian LI, Yunfeng LIU, Jinfa ZHANG