Patents by Inventor Patrick An-Phu Nguyen
Patrick An-Phu Nguyen has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Patent number: 10276163Abstract: Audio data that encodes an utterance of a user is received. It is determined that the user has been classified as a novice user of a speech recognizer. A speech recognizer setting is selected that is used by the speech recognizer in generating a transcription of the utterance. The selected speech recognizer setting is different than a default speech recognizer setting that is used by the speech recognizer in generating transcriptions of utterances of users that are not classified as novice users. The selected speech recognizer setting results in increased speech recognition accuracy in comparison with the default setting. A transcription of the utterance is obtained that is generated by the speech recognizer using the selected setting.Type: GrantFiled: June 29, 2016Date of Patent: April 30, 2019Assignee: Google LLCInventors: Xin Lei, Patrick An Phu Nguyen
-
Patent number: 10176236Abstract: Various embodiments of the present disclosure can include systems, methods, and non-transitory computer readable media configured to receive at least one database query to be executed. Code corresponding to the at least one database query can be generated. One or more optimizations to the generated code can be performed to produce specialized modular code. The one or more optimizations can include Just-In-Time (JIT) compilation techniques. Respective portions of the code can be distributed to a plurality of distributed computing systems for execution, wherein each of the distributed computing systems is connected to a portion of the plurality of distributed computing systems. A result for the at least one database query can be provided.Type: GrantFiled: July 23, 2015Date of Patent: January 8, 2019Assignee: Microsoft Technology Licensing, LLCInventors: Patrick An Phu Nguyen, David Kryze, Theodore Vassilakis, Apostolos Lerios
-
Patent number: 10169433Abstract: Various embodiments can include systems, methods, and non-transitory computer readable media configured to receive at least one operation to be performed using (i) first data that is managed by a first computing system and (ii) second data that is managed by a second computing system, the operation being received through an interface provided by the computing system, and wherein the operation is based at least in part on a Structured Query Language (SQL). At least one optimization can be performed based at least in part on the operation. The operation can be executed using at least the first data and the second data. A result generated can be provided upon executing the operation through the interface provided by the computing system. The computing system, the first computing system, and the second computing system are each able to concurrently process, access, and create at least a portion of the generated result.Type: GrantFiled: July 27, 2015Date of Patent: January 1, 2019Assignee: Microsoft Technology Licensing, LLCInventors: Apostolos Lerios, Theodore Vassilakis, Patrick An Phu Nguyen, David Kryze, Sreenivasa Viswanadha, Eric Karlson, Laurent An Minh Nguyen
-
Patent number: 9460711Abstract: Methods and systems for processing multilingual DNN acoustic models are described. An example method may include receiving training data that includes a respective training data set for each of two or more or languages. A multilingual deep neural network (DNN) acoustic model may be processed based on the training data. The multilingual DNN acoustic model may include a feedforward neural network having multiple layers of one or more nodes. Each node of a given layer may connect with a respective weight to each node of a subsequent layer, and the multiple layers of one or more nodes may include one or more shared hidden layers of nodes and a language-specific output layer of nodes corresponding to each of the two or more languages. Additionally, weights associated with the multiple layers of one or more nodes of the processed multilingual DNN acoustic model may be stored in a database.Type: GrantFiled: April 15, 2013Date of Patent: October 4, 2016Assignee: Google Inc.Inventors: Vincent Olivier Vanhoucke, Jeffrey Adgate Dean, Georg Heigold, Marc'aurelio Ranzato, Matthieu Devin, Patrick An Phu Nguyen, Andrew William Senior
-
Patent number: 9378733Abstract: Embodiments pertain to automatic speech recognition in mobile devices to establish the presence of a keyword. An audio waveform is received at a mobile device. Front-end feature extraction is performed on the audio waveform, followed by acoustic modeling, high level feature extraction, and output classification to detect the keyword. Acoustic modeling may use a neural network or a vector quantization dictionary and high level feature extraction may use pooling.Type: GrantFiled: April 11, 2013Date of Patent: June 28, 2016Assignee: Google Inc.Inventors: Vincent O. Vanhoucke, Oriol Vinyals, Patrick An Phu Nguyen, Maria Carolina Parada San Martin, Johan Schalkwyk
-
Publication number: 20160034530Abstract: Various embodiments of the present disclosure can include systems, methods, and non-transitory computer readable media configured to receive at least one database query to be executed. At least one computation graph corresponding to the at least one database query is generated. The computation graph is transformed to an optimized computation graph. The respective portions of the optimized computation graph are distributed to a plurality of distributed computing systems for execution. A result for the at least one database query is provided.Type: ApplicationFiled: July 23, 2015Publication date: February 4, 2016Inventors: Patrick An Phu Nguyen, Theodore Vassilakis, Sreenivasa Viswanadha, David Kryze
-
Publication number: 20160034529Abstract: Various embodiments of the present disclosure can include systems, methods, and non-transitory computer readable media configured to receive at least one database query to be executed. Code corresponding to the at least one database query can be generated. One or more optimizations to the generated code can be performed to produce specialized modular code. The one or more optimizations can include Just-In-Time (JIT) compilation techniques. Respective portions of the code can be distributed to a plurality of distributed computing systems for execution, wherein each of the distributed computing systems is connected to a portion of the plurality of distributed computing systems. A result for the at least one database query can be provided.Type: ApplicationFiled: July 23, 2015Publication date: February 4, 2016Inventors: Patrick An Phu Nguyen, David Kryze, Theodore Vassilakis, Apostolos Lerios
-
Publication number: 20160034547Abstract: Various embodiments can include systems, methods, and non-transitory computer readable media configured to receive at least one operation to be performed using (i) first data that is managed by a first computing system and (ii) second data that is managed by a second computing system, the operation being received through an interface provided by the computing system, and wherein the operation is based at least in part on a Structured Query Language (SQL). At least one optimization can be performed based at least in part on the operation. The operation can be executed using at least the first data and the second data. A result generated can be provided upon executing the operation through the interface provided by the computing system. The computing system, the first computing system, and the second computing system are each able to concurrently process, access, and create at least a portion of the generated result.Type: ApplicationFiled: July 27, 2015Publication date: February 4, 2016Inventors: Apostolos Lerios, Theodore Vassilakis, Patrick An Phu Nguyen, David Kryze, Sreenivasa Viswanadha, Eric Karlson, Laurent An Minh Nguyen
-
Publication number: 20150279351Abstract: Embodiments pertain to automatic speech recognition in mobile devices to establish the presence of a keyword. An audio waveform is received at a mobile device. Front-end feature extraction is performed on the audio waveform, followed by acoustic modeling, high level feature extraction, and output classification to detect the keyword. Acoustic modeling may use a neural network or Gaussian mixture modeling, and high level feature extraction may be done by aligning the results of the acoustic modeling with expected event vectors that correspond to a keyword.Type: ApplicationFiled: April 11, 2013Publication date: October 1, 2015Inventors: Patrick An Phu Nguyen, Maria Carolina Parada San Martin, Johan Schalkwyk
-
Publication number: 20140278379Abstract: In one implementation, a computer-implemented method includes receiving, at a computer system, a request to predict a next word in a dialog being uttered by a speaker; accessing, by the computer system, a neural network comprising i) an input layer, ii) one or more hidden layers, and iii) an output layer; identifying the local context for the dialog of the speaker; selecting, by the computer system and using a semantic model, at least one vector that represents the semantic context for the dialog; applying input to the input layer of the neural network, the input comprising i) the local context of the dialog and ii) the values for the at least one vector; generating probability values for at least a portion of the candidate words; and providing, by the computer system and based on the probability values, information that identifies one or more of the candidate words.Type: ApplicationFiled: April 16, 2013Publication date: September 18, 2014Applicant: Google Inc.Inventors: Noah B. Coccaro, Patrick An Phu Nguyen
-
Patent number: 8775177Abstract: A speech recognition process may perform the following operations: performing a preliminary recognition process on first audio to identify candidates for the first audio; generating first templates corresponding to the first audio, where each first template includes a number of elements; selecting second templates corresponding to the candidates, where the second templates represent second audio, and where each second template includes elements that correspond to the elements in the first templates; comparing the first templates to the second templates, where comparing comprises includes similarity metrics between the first templates and corresponding second templates; applying weights to the similarity metrics to produce weighted similarity metrics, where the weights are associated with corresponding second templates; and using the weighted similarity metrics to determine whether the first audio corresponds to the second audio.Type: GrantFiled: October 31, 2012Date of Patent: July 8, 2014Assignee: Google Inc.Inventors: Georg Heigold, Patrick An Phu Nguyen, Mitchel Weintraub, Vincent O. Vanhoucke
-
Patent number: 8401852Abstract: A computer-implemented speech recognition system described herein includes a receiver component that receives a plurality of detected units of an audio signal, wherein the audio signal comprises a speech utterance of an individual. A selector component selects a subset of the plurality of detected units that correspond to a particular time-span. A generator component generates at least one feature with respect to the particular time-span, wherein the at least one feature is one of an existence feature, an expectation feature, or an edit distance feature. Additionally, a statistical speech recognition model outputs at least one word that corresponds to the particular time-span based at least in part upon the at least one feature generated by the feature generator component.Type: GrantFiled: November 30, 2009Date of Patent: March 19, 2013Assignee: Microsoft CorporationInventors: Geoffrey Gerson Zweig, Patrick An-Phu Nguyen, James Garnet Droppo, III, Alejandro Acero
-
Publication number: 20110131046Abstract: A computer-implemented speech recognition system described herein includes a receiver component that receives a plurality of detected units of an audio signal, wherein the audio signal comprises a speech utterance of an individual. A selector component selects a subset of the plurality of detected units that correspond to a particular time-span. A generator component generates at least one feature with respect to the particular time-span, wherein the at least one feature is one of an existence feature, an expectation feature, or an edit distance feature. Additionally, a statistical speech recognition model outputs at least one word that corresponds to the particular time-span based at least in part upon the at least one feature generated by the feature generator component.Type: ApplicationFiled: November 30, 2009Publication date: June 2, 2011Applicant: Microsoft CorporationInventors: Geoffrey Gerson Zweig, Patrick An-Phu Nguyen, James Garnet Droppo, III, Alejandro Acero