Patents Examined by Qi Han
-
Patent number: 9548057Abstract: An object of embodiments of the present invention is to provide an improved gain-shape VQ. This is achieved by determining a number of bits to be allocated to a gain adjustment- and shape-quantizer for a plurality of combinations of a current bit rate and a first signal property. The determined allocated number of bits to the gain adjustment- and shape quantizer should provide a better result for the given bitrate and signal property than using a single fixed allocation scheme. That can be achieved by deriving the bit allocation by using an average of optimal bit allocations for a training data set. Thus by pre-calculating a number of bits to the gain adjustment and the shape quantizers for a plurality of combinations of the bit rate and a first signal property and creating a table indicating the number of bits to be allocated to the gain adjustment- and the shape-quantizers for a plurality of combinations of the bit rate and a first signal property.Type: GrantFiled: October 17, 2011Date of Patent: January 17, 2017Assignee: Telefonaktiebolaget LM Ericsson (publ)Inventor: Erik Norvell
-
Patent number: 9542942Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for designating certain voice commands as hotwords. The methods, systems, and apparatus include actions of receiving a hotword followed by a voice command. Additional actions include determining that the voice command satisfies one or more predetermined criteria associated with designating the voice command as a hotword, where a voice command that is designated as a hotword is treated as a voice input regardless of whether the voice command is preceded by another hotword. Further actions include, in response to determining that the voice command satisfies one or more predetermined criteria associated with designating the voice command as a hotword, designating the voice command as a hotword.Type: GrantFiled: January 21, 2016Date of Patent: January 10, 2017Assignee: Google Inc.Inventor: Matthew Sharifi
-
Patent number: 9536525Abstract: A speaker indexing device extracts a plurality of features from a speech signal on a frame-by-frame basis, models a distribution of first feature sets by a mixture distribution containing as many probability distributions as there are speakers, selects for each probability distribution either first feature sets located within a predetermined distance from the center of the probability distribution or a predetermined number of first feature sets in sequence starting from the first feature set closest to the center of the probability distribution, selects a second feature for the frame corresponding to the selected first feature sets as first training data for the speaker corresponding to the probability distribution and, using the first training data, trains a speaker model to be used to append to each frame identification information for identifying the speaker speaking in the frame.Type: GrantFiled: August 13, 2015Date of Patent: January 3, 2017Assignee: FUJITSU LIMITEDInventor: Shoji Hayakawa
-
Patent number: 9524715Abstract: According to a preferred aspect of the instant invention, there is provided a system and method for content recognition in portable devices. Content, preferably audio content is recorded by the instant invention, preferably a sample with a length between 1 and 10 seconds. A fingerprint will be generated from the recorded sample and automatically, and preferably without further user interaction, prompting, notification, etc. (e.g. invisible to the user), compared with the fingerprints in a fingerprint database that is stored locally in the portable device and the result thereafter presented to the user.Type: GrantFiled: December 27, 2012Date of Patent: December 20, 2016Assignee: BELLEVUE INVESTMENTS GMBH & CO. KGAAInventors: Tilman Herberger, Titus Tost
-
Patent number: 9520144Abstract: A method, an apparatus, and a computer-readable medium configured with instructions that when executed carry out the method for determining a measure of harmonicity. In one embodiment the method includes selecting candidate fundamental frequencies within a range, and for candidate determining a mask or retrieving a pre-calculated mask that has positive value for each frequency that contributed to harmonicity, and negative value for each frequency that contributes to inharmonicity. A candidate harmonicity measure is calculated for each candidate fundamental by summing the product of the mask and the magnitude measure spectrum. The harmonicity measure is selected as the maximum of the candidate harmonicity measures.Type: GrantFiled: March 21, 2013Date of Patent: December 13, 2016Assignee: Dolby Laboratories Licensing CorporationInventors: David Gunawan, Glenn N. Dickins
-
Patent number: 9519635Abstract: The present invention is a method and apparatus for narrative content generation using narrative frameworks by receiving a first phrase variation and a second phrase variation and displaying an error indication when the first phrase variation fails to satisfy a criterion relative to the second phrase variation. If there is an error indication, alternate phrase variations are received and compared against the first phrase variation until an alternate phrase variation is selected that has no error indication. Additionally, multiple sets of operators for updating one or more narrative phrases selected for inclusion in the narrative content framework may be utilized to update selected phrases after inclusion in the narrative framework but prior to finalizing the narrative content to be output.Type: GrantFiled: September 11, 2014Date of Patent: December 13, 2016Assignee: Automated Insights, Inc.Inventors: Robert Allen, Joe Procopio, Robert C Rogers
-
Patent number: 9483462Abstract: A method for generating training data for disambiguation of an entity comprising a word or word string related to a topic to be analyzed includes acquiring sent messages by a user, each including at least one entity in a set of entities; organizing the messages and acquiring sets, each containing messages sent by each user; identifying a set of messages including different entities, greater than or equal to a first threshold value, and identifying a user corresponding to the identified set as a hot user; receiving an instruction indicating an object entity to be disambiguated; determining a likelihood of co-occurrence of each keyword and the object entity in sets of messages sent by hot users; and determining training data for the object entity on the basis of the likelihood of co-occurrence of each keyword and the object entity in the sets of messages sent by the hot users.Type: GrantFiled: August 7, 2015Date of Patent: November 1, 2016Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATIONInventors: Yohei Ikawa, Akiko Suzuki
-
Patent number: 9483768Abstract: A computer-implemented method and an apparatus for modeling customer interaction experiences receives interaction data corresponding to one or more interactions between a customer and a customer support representative. At least one language associated with the interaction data is detected. Textual content in a plurality of languages is generated corresponding to the interaction data based at least in part on translating the interaction data using two or more languages different than the at least one language. At least one emotion score is determined for text corresponding to each language from among the plurality of languages. An aggregate emotion score is determined using the at least one emotion score for the text corresponding to the each language. An interaction experience of the customer is modeled based at least in part on the aggregate emotion score.Type: GrantFiled: August 10, 2015Date of Patent: November 1, 2016Assignee: 24/7 CUSTOMER, INC.Inventor: Bhupinder Singh
-
Patent number: 9484017Abstract: A first speech processing device includes a first speech input unit and a first speech output unit. A second speech processing device includes a second speech input unit and a second speech output unit. In a server therebetween, a speech of a first language sent from the first speech input unit is recognized. The speech recognition result is translated into a second language. The translation result is back translated into the first language. A first speech synthesis signal of the back translation result is sent to the first speech output unit. A second speech synthesis signal of the translation result is sent to the second speech output unit. Duration of the second speech synthesis signal or the first speech synthesis signal is measured. The first speech synthesis signal and the second speech synthesis signal are outputted by synchronizing a start time and an end time thereof, based on the duration.Type: GrantFiled: September 12, 2014Date of Patent: November 1, 2016Assignee: KABUSHIKI KAISHA TOSHIBAInventors: Kazuo Sumita, Akinori Kawamura, Satoshi Kamatani
-
Patent number: 9483582Abstract: A system for verifying factual assertions in natural language. Receiving a text input. Identifying, a verifiable factual statement in the text. Forming a query based on the verifiable factual assertion. Searching a corpus based on the query. Determining the veracity of verifiable factual assertion based on the search results.Type: GrantFiled: September 12, 2014Date of Patent: November 1, 2016Assignee: International Business Machines CorporationInventors: Brian P. Gaucher, Dario Gil, Jeffrey O. Kephart, Jonathan Lenchner, David O. S. Melville, John M. Prager, Valentina Salapura
-
Patent number: 9472185Abstract: An interactive response system directs input to a software-based router, which is able to intelligently respond to the input by drawing on a combination of human agents, advanced recognition and expert systems. The system utilizes human “intent analysts” for purposes of interpreting customer input. Automated recognition subsystems are trained by coupling customer input with IA-selected intent corresponding to the input, using model-updating subsystems to develop the training information for the automated recognition subsystems.Type: GrantFiled: October 10, 2013Date of Patent: October 18, 2016Assignee: Interactions LLCInventors: Yoryos Yeracaris, Larissa Lapshina, Alwin B Carus
-
Patent number: 9471568Abstract: According to one embodiment, a speech of a first language is recognized using a speech recognition dictionary to recognize the first language and a second language, and a source sentence of the first language is generated. The source sentence is translated into a second language, and a translation sentence of the second language is generated. An unknown word included in the translation sentence is detected. The unknown word is not stored in the speech recognition dictionary. A first pronunciation candidate of the unknown word is estimated, from a representation of the unknown word. A second pronunciation candidate of the unknown word is estimated from a pronunciation of an original word included in the source sentence corresponding to the unknown word. The unknown word, the first pronunciation candidate and the second pronunciation candidate, are registered into the speech recognition dictionary correspondingly.Type: GrantFiled: September 12, 2014Date of Patent: October 18, 2016Assignee: KABUSHIKI KAISHA TOSHIBAInventors: Satoshi Kamatani, Kazuo Sumita, Akinori Kawamura
-
Patent number: 9466299Abstract: A method and associated system and computer program product. A sample of speech, for which a source of the sample of speech is to be classified, is received. A frequency clip level of the sample of speech is determined. A higher frequency clip level indicates the source is human and a lower frequency clip level indicates the source is machine generated. A dynamic range of the sample of speech is determined. A lower dynamic range indicates the source is human and a higher dynamic range indicates the source is machine generated. The frequency clip level and the dynamic range are weighted by a respective weighting factor as to whether the source is human or the source is machine generated. The source is classified as human generated or machine generated. The classifying of the source is based on the frequency clip level, the dynamic range, and the respective weighting factors thereof.Type: GrantFiled: November 18, 2015Date of Patent: October 11, 2016Assignee: International Business Machines CorporationInventors: Andrew S. Feltham, Robert S. Smart, Graham White
-
Patent number: 9460714Abstract: In a speech processing apparatus, an acquisition unit is configured to acquire a speech. A separation unit is configured to separate the speech into a plurality of sections in accordance with a prescribed rule. A calculation unit is configured to calculate a degree of similarity in each combination of the sections. An estimation unit is configured to estimate, with respect to the each section, a direction of arrival of the speech. A correction unit is configured to group the sections whose directions of arrival are mutually similar into a same group and correct the degree of similarity with respect to the combination of the sections in the same group. A clustering unit is configured to cluster the sections by using the corrected degree of similarity.Type: GrantFiled: September 12, 2014Date of Patent: October 4, 2016Assignee: KABUSHIKI KAISHA TOSHIBAInventors: Ning Ding, Yusuke Kida, Makoto Hirohata
-
Patent number: 9454966Abstract: Disclosed herein are systems, computer-implemented methods, and computer-readable storage media for generating personalized user models. The method includes receiving automatic speech recognition (ASR) output of speech interactions with a user, receiving an ASR transcription error model characterizing how ASR transcription errors are made, generating guesses of a true transcription and a user model via an expectation maximization (EM) algorithm based on the error model and the respective ASR output where the guesses will converge to a personalized user model which maximizes the likelihood of the ASR output. The ASR output can be unlabeled. The method can include casting speech interactions as a dynamic Bayesian network with four variables: (s), (u), (r), (m), and encoding relationships between (s), (u), (r), (m) as conditional probability tables. At each dialog turn (r) and (m) are known and (s) and (u) are hidden.Type: GrantFiled: June 25, 2013Date of Patent: September 27, 2016Assignee: AT&T Intellectual Property I, L.P.Inventors: Jason Williams, Umar Ali Syed
-
Patent number: 9437217Abstract: A pre-processing apparatus for speech recognition may include: a trailing silence period detection unit configured to detect the length of a trailing silence period contained in a speech signal; a reference trailing silence period storage unit configured to store the length of a reference trailing silence period; and a trailing silence period adjusting unit configured to adjust the length of the trailing silence period contained in the speech signal based on the length of the reference trailing silence period.Type: GrantFiled: September 11, 2014Date of Patent: September 6, 2016Assignee: HYUNDAI MOBIS Co., Ltd.Inventor: Min Ho Kwon
-
Patent number: 9424251Abstract: A method of extracting the semantic distance from the mathematical sentence and classifying the mathematical sentence by the semantic distance, includes: receiving a user query; extracting at least one keyword included in the received user query; and extracting a semantic distance by, indexing one or more of natural language tokens and mathematical equation tokens including semantic information, extracting the semantic distance, between the at least one extracted keyword and the one or more indexed semantic information by referring indexed information, and acquiring a similarity of the received user query and the semantic information.Type: GrantFiled: June 6, 2013Date of Patent: August 23, 2016Assignee: SK TELECOM CO., LTD.Inventors: Keun Tae Park, Yong Gil Park, Hyeongin Choi, Nam Sook Wee, Doo Seok Lee, Jung Kyo Sohn, Haeng Moon Kim, Dong Hahk Lee
-
Patent number: 9418651Abstract: An electronic device includes a microphone that receives an audio signal, and a processor that is electrically coupled to the microphone. The processor detects a trigger phrase in the received audio signal and measure characteristics of the detected trigger phrase. Based on the measured characteristics of the detected trigger phrase, the processor determines whether the detected trigger phrase is valid.Type: GrantFiled: October 10, 2013Date of Patent: August 16, 2016Assignee: Google Technology Holdings LLCInventors: Joel A Clark, Tenkasi V Ramabadran, Mark A. Jasiuk
-
Patent number: 9411785Abstract: A method, apparatus, and system relating to embedding hidden content within a Unicode message and using the hidden content to perform a particular computer action.Type: GrantFiled: April 22, 2015Date of Patent: August 9, 2016Assignee: Pebble Technology, Corp.Inventors: Yoon Kean Wong, Henry Levak, Heiko Behrens
-
Patent number: 9406300Abstract: An automobile to detect a signal, having a security code, from a mobile transponder within range of the automobile is disclosed. The automobile may determine that the security code is valid to process an audible command from a user. The audible command may correspond to an automobile function. The automobile may also determine if the audible command matches a voiceprint of the user and process the automobile function accordingly. The automobile function may provide customized user settings for utilization of the automobile.Type: GrantFiled: February 5, 2014Date of Patent: August 2, 2016Assignee: Tamiras Per Pte. Ltd., LLCInventor: Richard B. Himmelstein