Patents by Inventor SHIXIONG ZHANG
SHIXIONG ZHANG has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Publication number: 20240072354Abstract: This application relates to the field of battery technology, and in particular, to a box assembly, a battery module, a battery, and an electrical device. The box assembly may include at least a partition piece and two end plates. In the box assembly disclosed herein, a stress relief piece may be disposed at two ends of a main body of the partition piece. The main body may be connected to the end plate by use of the stress relief piece. A corner may be formed between the stress relief piece and the main body. In this way, when the battery module is subjected to stress, at least a part of the stress can be relieved through the corner formed between the stress relief piece and the main body.Type: ApplicationFiled: October 25, 2023Publication date: February 29, 2024Applicant: CONTEMPORARY AMPEREX TECHNOLOGY CO., LIMITEDInventors: Hang DU, Yang ZOU, Shixiong ZHENG, Yaowen HU, Zhihong ZHANG, Pengfei LI
-
Inter-channel feature extraction method, audio separation method and apparatus, and computing device
Patent number: 11908483Abstract: This application relates to a method of extracting an inter channel feature from a multi-channel multi-sound source mixed audio signal performed at a computing device.Type: GrantFiled: August 12, 2021Date of Patent: February 20, 2024Assignee: TENCENT TECHNOLOGY (SHENZHEN) COMPANY LIMITEDInventors: Rongzhi Gu, Shixiong Zhang, Lianwu Chen, Yong Xu, Meng Yu, Dan Su, Dong Yu -
Publication number: 20230402038Abstract: A method for facilitating a remote conference includes receiving a digital video and a computer-readable audio signal. A face recognition machine is operated to recognize a face of a first conference participant in the digital video, and a speech recognition machine is operated to translate the computer-readable audio signal into a first text. An attribution machine attributes the text to the first conference participant. A second computer-readable audio signal is processed similarly, to obtain a second text attributed to a second conference participant. A transcription machine automatically creates a transcript including the first text attributed to the first conference participant and the second text attributed to the second conference participant.Type: ApplicationFiled: May 15, 2023Publication date: December 14, 2023Inventors: Adi DIAMANT, Xuedong HUANG, Karen MASTER BEN-DOR, Eyal KRUPKA, Raz HALALY, Yoni SMOLIN, Ilya GURVICH, Aviv HURVITZ, Lijuan QIN, Wei XIONG, Shixiong ZHANG, Lingfeng WU, Xiong XIAO, Ido LEICHTER, Moshe DAVID, Amit Kumar AGARWAL
-
Publication number: 20230209278Abstract: The technology of this application relates to a laser microphone, including a diaphragm, a laser device, a control circuit, a self-mixing signal obtaining apparatus, and a signal processing circuit. The laser device is configured to emit light to the diaphragm and receive a feedback light signal from the diaphragm, and the feedback light signal interferes with laser in a resonant cavity of the laser device to obtain a self-mixing light signal. A distance between the laser device and the diaphragm ranges from 30 to 300 m. The control circuit is connected to the laser device, and is configured to drive and control the laser device to emit light. The self-mixing signal obtaining apparatus is connected to the laser device, and is configured to obtain a target voltage signal related to the self-mixing light signal.Type: ApplicationFiled: March 3, 2023Publication date: June 29, 2023Inventors: Xiaoke HOU, Shixiong ZHANG, Shengjie RUAN
-
Patent number: 11688399Abstract: A method for facilitating a remote conference includes receiving a digital video and a computer-readable audio signal. A face recognition machine is operated to recognize a face of a first conference participant in the digital video, and a speech recognition machine is operated to translate the computer-readable audio signal into a first text. An attribution machine attributes the text to the first conference participant. A second computer-readable audio signal is processed similarly, to obtain a second text attributed to a second conference participant. A transcription machine automatically creates a transcript including the first text attributed to the first conference participant and the second text attributed to the second conference participant.Type: GrantFiled: December 8, 2020Date of Patent: June 27, 2023Assignee: MICROSOFT TECHNOLOGY LICENSING, LLCInventors: Adi Diamant, Karen Master Ben-Dor, Eyal Krupka, Raz Halaly, Yoni Smolin, Ilya Gurvich, Aviv Hurvitz, Lijuan Qin, Wei Xiong, Shixiong Zhang, Lingfeng Wu, Xiong Xiao, Ido Leichter, Moshe David, Xuedong Huang, Amit Kumar Agarwal
-
Publication number: 20220322511Abstract: A backlight control circuit is provided, which includes a driver circuit and a power conversion circuit. The driver circuit includes a feedback output terminal and at least one channel port. The channel port is coupled to a first terminal of a light string group. The driver circuit is configured to obtain a voltage of each channel port, and enable the feedback output terminal to provide a current feedback signal based on the voltage Vch. The power conversion circuit is coupled to the feedback output terminal and includes a voltage output terminal. The voltage output terminal is configured to provide a supply voltage for a second terminal of each light string group. The power conversion circuit is configured to perform voltage conversion on an input voltage, and increase or decrease the supply voltage based on the current feedback signal.Type: ApplicationFiled: October 12, 2020Publication date: October 6, 2022Inventors: Shixiong ZHANG, SOOYOUNG WOO, Min CHEN
-
Patent number: 11222640Abstract: Computing devices and methods utilizing a joint speaker location/speaker identification neural network are provided. In one example a computing device receives an audio signal of utterances spoken by multiple persons. Magnitude and phase information features are extracted from the signal and inputted into a joint speaker location and speaker identification neural network. The neural network utilizes both the magnitude and phase information features to determine a change in the person speaking. Output comprising the determination of the change is received from the neural network. The output is then used to perform a speaker recognition function, speaker location function, or both.Type: GrantFiled: February 27, 2020Date of Patent: January 11, 2022Assignee: Microsoft Technology Licensing, LLCInventors: Shixiong Zhang, Xiong Xiao
-
INTER-CHANNEL FEATURE EXTRACTION METHOD, AUDIO SEPARATION METHOD AND APPARATUS, AND COMPUTING DEVICE
Publication number: 20210375294Abstract: This application relates to a method of extracting an inter channel feature from a multi-channel multi-sound source mixed audio signal performed at a computing device.Type: ApplicationFiled: August 12, 2021Publication date: December 2, 2021Inventors: Rongzhi Gu, Shixiong Zhang, Lianwu Chen, Yong Xu, Meng Yu, Dan Su, Dong Yu -
Patent number: 11152006Abstract: Examples are disclosed that relate to voice identification enrollment. One example provides a method of voice identification enrollment comprising, during a meeting in which two or more human speakers speak at different times, determining whether one or more conditions of a protocol for sampling meeting audio used to establish human speaker voiceprints are satisfied, and in response to determining that the one or more conditions are satisfied, selecting a sample of meeting audio according to the protocol, the sample representing an utterance made by one of the human speakers. The method further comprises establishing, based at least on the sample, a voiceprint of the human speaker.Type: GrantFiled: June 27, 2018Date of Patent: October 19, 2021Assignee: Microsoft Technology Licensing, LLCInventors: Eyal Krupka, Shixiong Zhang, Xiong Xiao
-
Patent number: 11128435Abstract: This disclosure relates to a cloud-local joint or collaborative data analytics framework that provides data analytics models trained and hosted in backend servers for processing data items preprocessed and encrypted by remote terminal devices. The data analytics models are configured to generate encrypted output data items that are then communicated to the local terminal devices for decryption and post-processing. This framework functions without exposing decryption keys of the local terminal devices to the backend servers and the communication network. The encryption/decryption and data analytics in the backend servers are configured to process and communicate data items efficiently to provide real-time or near real-time system response to requests for data analytics from the remote terminal devices.Type: GrantFiled: July 8, 2019Date of Patent: September 21, 2021Assignee: Tencent America LLCInventors: Shixiong Zhang, Dong Yu
-
Publication number: 20210210097Abstract: A method for facilitating a remote conference includes receiving a digital video and a computer-readable audio signal. A face recognition machine is operated to recognize a face of a first conference participant in the digital video, and a speech recognition machine is operated to translate the computer-readable audio signal into a first text. An attribution machine attributes the text to the first conference participant. A second computer-readable audio signal is processed similarly, to obtain a second text attributed to a second conference participant. A transcription machine automatically creates a transcript including the first text attributed to the first conference participant and the second text attributed to the second conference participant.Type: ApplicationFiled: December 8, 2020Publication date: July 8, 2021Inventors: Adi DIAMANT, Karen MASTER BEN-DOR, Eyal KRUPKA, Raz HALALY, Yoni SMOLIN, Ilya GURVICH, Aviv HURVITZ, Lijuan QIN, Wei XIONG, Shixiong ZHANG, Lingfeng WU, Xiong XIAO, Ido LEICHTER, Moshe DAVID, Xuedong HUANG, Amit Kumar AGARWAL
-
Publication number: 20210014039Abstract: This disclosure relates to a cloud-local joint or collaborative data analytics framework that provides data analytics models trained and hosted in backend servers for processing data items preprocessed and encrypted by remote terminal devices. The data analytics models are configured to generate encrypted output data items that are then communicated to the local terminal devices for decryption and post-processing. This framework functions without exposing decryption keys of the local terminal devices to the backend servers and the communication network. The encryption/decryption and data analytics in the backend servers are configured to process and communicate data items efficiently to provide real-time or near real-time system response to requests for data analytics from the remote terminal devices.Type: ApplicationFiled: July 8, 2019Publication date: January 14, 2021Applicant: Tencent America LLCInventors: Shixiong ZHANG, Dong YU
-
Patent number: 10867610Abstract: A method for facilitating a remote conference includes receiving a digital video and a computer-readable audio signal. A face recognition machine is operated to recognize a face of a first conference participant in the digital video, and a speech recognition machine is operated to translate the computer-readable audio signal into a first text. An attribution machine attributes the text to the first conference participant. A second computer-readable audio signal is processed similarly, to obtain a second text attributed to a second conference participant. A transcription machine automatically creates a transcript including the first text attributed to the first conference participant and the second text attributed to the second conference participant.Type: GrantFiled: June 29, 2018Date of Patent: December 15, 2020Assignee: Microsoft Technology Licensing, LLCInventors: Adi Diamant, Karen Master Ben-Dor, Eyal Krupka, Raz Halaly, Yoni Smolin, Ilya Gurvich, Aviv Hurvitz, Lijuan Qin, Wei Xiong, Shixiong Zhang, Lingfeng Wu, Xiong Xiao, Ido Leichter, Moshe David, Xuedong Huang, Amit Kumar Agarwal
-
Publication number: 20200202867Abstract: Computing devices and methods utilizing a joint speaker location/speaker identification neural network are provided. In one example a computing device receives an audio signal of utterances spoken by multiple persons. Magnitude and phase information features are extracted from the signal and inputted into a joint speaker location and speaker identification neural network. The neural network utilizes both the magnitude and phase information features to determine a change in the person speaking. Output comprising the determination of the change is received from the neural network. The output is then used to perform a speaker recognition function, speaker location function, or both.Type: ApplicationFiled: February 27, 2020Publication date: June 25, 2020Applicant: Microsoft Technology Licensing, LLCInventors: Shixiong ZHANG, Xiong XIAO
-
Patent number: 10621991Abstract: A speaker recognition system includes a previously-trained joint neural network. An enrollment machine of the speaker recognition system is configured to operate the previously-trained joint neural network to enroll a new speaker based on audiovisual data featuring the newly enrolled speaker. A recognition machine of the speaker recognition system is configured to operate the previously-trained joint neural network to recognize a previously-enrolled speaker based on audiovisual data featuring the previously-enrolled speaker.Type: GrantFiled: June 28, 2018Date of Patent: April 14, 2020Assignee: MICROSOFT TECHNOLOGY LICENSING, LLCInventors: Shixiong Zhang, Eyal Krupka
-
Patent number: 10580414Abstract: Computing devices and methods utilizing a joint speaker location/speaker identification neural network are provided. In one example a computing device receives a multi-channel audio signal of an utterance spoken by a user. Magnitude and phase information features are extracted from the signal and inputted into a joint speaker location/speaker identification neural network that is trained via utterances from a plurality of persons. A user embedding comprising speaker identification characteristics and location characteristics is received from the neural network and compared to a plurality of enrollment embeddings extracted from the plurality of utterances that are each associated with an identity of a corresponding person. Based at least on the comparisons, the user is matched to an identity of one of the persons, and the identity of the person is outputted.Type: GrantFiled: June 12, 2018Date of Patent: March 3, 2020Assignee: MICROSOFT TECHNOLOGY LICENSING, LLCInventors: Shixiong Zhang, Xiong Xiao
-
Publication number: 20190341053Abstract: A computerized conference assistant includes a camera and a microphone. A face location machine of the computerized conference assistant finds a physical location of a human, based on a position of a candidate face in digital video captured by the camera. A beamforming machine of the computerized conference assistant outputs a beamformed signal isolating sounds originating from the physical location of the human. A diarization machine of the computerized conference assistant attributes information encoded in the beamformed signal to the human.Type: ApplicationFiled: June 26, 2018Publication date: November 7, 2019Applicant: Microsoft Technology Licensing, LLCInventors: Shixiong ZHANG, Lingfeng WU, Eyal KRUPKA, Xiong XIAO, Yifan GONG
-
Publication number: 20190341055Abstract: Examples are disclosed that relate to voice identification enrollment. One example provides a method of voice identification enrollment comprising, during a meeting in which two or more human speakers speak at different times, determining whether one or more conditions of a protocol for sampling meeting audio used to establish human speaker voiceprints are satisfied, and in response to determining that the one or more conditions are satisfied, selecting a sample of meeting audio according to the protocol, the sample representing an utterance made by one of the human speakers. The method further comprises establishing, based at least on the sample, a voiceprint of the human speaker.Type: ApplicationFiled: June 27, 2018Publication date: November 7, 2019Applicant: Microsoft Technology Licensing, LLCInventors: Eyal KRUPKA, Shixiong ZHANG, Xiong XIAO
-
Publication number: 20190341057Abstract: Computing devices and methods utilizing a joint speaker location/speaker identification neural network are provided. In one example a computing device receives a multi-channel audio signal of an utterance spoken by a user. Magnitude and phase information features are extracted from the signal and inputted into a joint speaker location/speaker identification neural network that is trained via utterances from a plurality of persons. A user embedding comprising speaker identification characteristics and location characteristics is received from the neural network and compared to a plurality of enrollment embeddings extracted from the plurality of utterances that are each associated with an identity of a corresponding person. Based at least on the comparisons, the user is matched to an identity of one of the persons, and the identity of the person is outputted.Type: ApplicationFiled: June 12, 2018Publication date: November 7, 2019Applicant: Microsoft Technology Licensing, LLCInventors: Shixiong ZHANG, Xiong XIAO
-
Publication number: 20190341058Abstract: A speaker recognition system includes a previously-trained joint neural network. An enrollment machine of the speaker recognition system is configured to operate the previously-trained joint neural network to enroll a new speaker based on audiovisual data featuring the newly enrolled speaker. A recognition machine of the speaker recognition system is configured to operate the previously-trained joint neural network to recognize a previously-enrolled speaker based on audiovisual data featuring the previously-enrolled speaker.Type: ApplicationFiled: June 28, 2018Publication date: November 7, 2019Applicant: Microsoft Technology Licensing, LLCInventors: Shixiong ZHANG, Eyal KRUPKA