Patents by Inventor Yeon-Jun Kim

Yeon-Jun Kim has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 10146868
    Abstract: Apparatuses, systems, methods, and media for filtering a data stream are provided. The data stream is partitioned into a plurality of data stream segments. An acoustic parameter is measured in each of the data stream segments. It is determined whether the acoustic parameter satisfies a first predetermined condition. The first predetermined condition includes a number of variances, in which the acoustic parameter exceeds a predetermined variance threshold, exceeding a predetermined number threshold. An extraneous portion of the data stream is identified in which the first predetermined condition is satisfied. It is determined whether the extraneous portion satisfies a second predetermined condition in the data stream. The extraneous portion is deleted from the data stream to produce a filtered data stream in response to the second predetermined condition being satisfied.
    Type: Grant
    Filed: June 8, 2017
    Date of Patent: December 4, 2018
    Assignee: AT&T INTELLECTUAL PROPERTY I, L.P.
    Inventors: Yeon-Jun Kim, I. Dan Melamed, Bernard S. Renger, Steven Neil Tischer
  • Patent number: 10002612
    Abstract: Disclosed herein are systems, computer-implemented methods, and tangible computer-readable storage media for captioning a media presentation. The method includes receiving automatic speech recognition (ASR) output from a media presentation and a transcription of the media presentation. The method includes selecting via a processor a pair of anchor words in the media presentation based on the ASR output and transcription and generating captions by aligning the transcription with the ASR output between the selected pair of anchor words. The transcription can be human-generated. Selecting pairs of anchor words can be based on a similarity threshold between the ASR output and the transcription. In one variation, commonly used words on a stop list are ineligible as anchor words. The method includes outputting the media presentation with the generated captions. The presentation can be a recording of a live event.
    Type: Grant
    Filed: November 14, 2016
    Date of Patent: June 19, 2018
    Assignee: AT&T Intellectual Property I, L.P.
    Inventors: Yeon-Jun Kim, David C. Gibbon, Horst J. Schroeter
  • Patent number: 9978360
    Abstract: Disclosed herein are systems, methods, and non-transitory computer-readable storage media for detecting and correcting abnormal stress patterns in unit-selection speech synthesis. A system practicing the method detects incorrect stress patterns in selected acoustic units representing speech to be synthesized, and corrects the incorrect stress patterns in the selected acoustic units to yield corrected stress patterns. The system can further synthesize speech based on the corrected stress patterns. In one aspect, the system also classifies the incorrect stress patterns using a machine learning algorithm such as a classification and regression tree, adaptive boosting, support vector machine, and maximum entropy. In this way a text-to-speech unit selection speech synthesizer can produce more natural sounding speech with suitable stress patterns regardless of the stress of units in a unit selection database.
    Type: Grant
    Filed: February 22, 2016
    Date of Patent: May 22, 2018
    Assignee: NUANCE COMMUNICATIONS, INC.
    Inventors: Yeon-Jun Kim, Mark Charles Beutnagel, Alistair D. Conkie, Ann K. Syrdal
  • Patent number: 9934792
    Abstract: A method of detecting pre-determined phrases to determine compliance quality of an agent includes determining a presence of a predetermined input based on a comparison between stored pre-determined phrases and a received communication, and determining a compliance rating of the agent based on a presence of a pre-determined phrase associated with the predetermined input in the communication.
    Type: Grant
    Filed: January 31, 2017
    Date of Patent: April 3, 2018
    Assignee: AT&T INTELLECTUAL PROPERTY I, L.P.
    Inventors: I. Dan Melamed, Andrej Ljolje, Bernard S. Renger, David J. Smith, Yeon-Jun Kim
  • Publication number: 20170270201
    Abstract: Apparatuses, systems, methods, and media for filtering a data stream are provided. The data stream is partitioned into a plurality of data stream segments. An acoustic parameter is measured in each of the data stream segments. It is determined whether the acoustic parameter satisfies a first predetermined condition. The first predetermined condition includes a number of variances, in which the acoustic parameter exceeds a predetermined variance threshold, exceeding a predetermined number threshold. An extraneous portion of the data stream is identified in which the first predetermined condition is satisfied. It is determined whether the extraneous portion satisfies a second predetermined condition in the data stream. The extraneous portion is deleted from the data stream to produce a filtered data stream in response to the second predetermined condition being satisfied.
    Type: Application
    Filed: June 8, 2017
    Publication date: September 21, 2017
    Applicant: AT&T INTELLECTUAL PROPERTY I, L.P.
    Inventors: Yeon-Jun KIM, I. Dan MELAMED, Bernard S. RENGER, Steven Neil TISCHER
  • Patent number: 9703865
    Abstract: Apparatuses, systems, methods, and media for filtering a data stream are provided. The data stream is analyzed based on an acoustic parameter to determine extraneous portions in which a first predetermined condition is satisfied. When a first extraneous portion is separated from a second extraneous portion by a non-extraneous portion in which the first predetermined condition is not satisfied, it is determined whether the first extraneous portion being separated from the second extraneous portion by the non-extraneous portion satisfies a second predetermined condition. At least one of the first extraneous portion and the second extraneous portion is deleted from the data stream to produce a filtered data stream in response to determining the second predetermined condition is satisfied.
    Type: Grant
    Filed: September 25, 2015
    Date of Patent: July 11, 2017
    Assignee: AT&T INTELLECTUAL PROPERTY I, L.P.
    Inventors: Yeon-Jun Kim, I. Dan Melamed, Bernard S. Renger, Steven Neil Tischer
  • Publication number: 20170140775
    Abstract: A method of detecting pre-determined phrases to determine compliance quality of an agent includes determining a presence of a predetermined input based on a comparison between stored pre-determined phrases and a received communication, and determining a compliance rating of the agent based on a presence of a pre-determined phrase associated with the predetermined input in the communication.
    Type: Application
    Filed: January 31, 2017
    Publication date: May 18, 2017
    Applicant: AT&T INTELLECTUAL PROPERTY I, L.P.
    Inventors: I. Dan MELAMED, Andrej LJOLJE, Bernard S. Renger, David J. Smith, Yeon-Jun Kim
  • Patent number: 9607279
    Abstract: A method of detecting pre-determined phrases to determine compliance quality includes determining whether a precursor event has occurred based on a comparison between stored pre-determined phrases and a received communication, and determining a compliance rating based on a presence of a pre-determined phrase associated with the precursor event in the communication.
    Type: Grant
    Filed: April 15, 2015
    Date of Patent: March 28, 2017
    Assignee: AT&T INTELLECTUAL PROPERTY I, L.P.
    Inventors: I. Dan Melamed, Andrej Ljolje, Bernard Renger, Yeon-Jun Kim, David J. Smith
  • Publication number: 20170061986
    Abstract: Disclosed herein are systems, computer-implemented methods, and tangible computer-readable storage media for captioning a media presentation. The method includes receiving automatic speech recognition (ASR) output from a media presentation and a transcription of the media presentation. The method includes selecting via a processor a pair of anchor words in the media presentation based on the ASR output and transcription and generating captions by aligning the transcription with the ASR output between the selected pair of anchor words. The transcription can be human-generated. Selecting pairs of anchor words can be based on a similarity threshold between the ASR output and the transcription. In one variation, commonly used words on a stop list are ineligible as anchor words. The method includes outputting the media presentation with the generated captions. The presentation can be a recording of a live event.
    Type: Application
    Filed: November 14, 2016
    Publication date: March 2, 2017
    Inventors: Yeon-Jun KIM, David C. GIBBON, Horst J. SCHROETER
  • Patent number: 9564121
    Abstract: Disclosed herein are systems, computer-implemented methods, and computer-readable storage media for unit selection synthesis. The method causes a computing device to add a supplemental phoneset to a speech synthesizer front end having an existing phoneset, modify a unit preselection process based on the supplemental phoneset, preselect units from the supplemental phoneset and the existing phoneset based on the modified unit preselection process, and generate speech based on the preselected units. The supplemental phoneset can be a variation of the existing phoneset, can include a word boundary feature, can include a cluster feature where initial consonant clusters and some word boundaries are marked with diacritics, can include a function word feature which marks units as originating from a function word or a content word, and/or can include a pre-vocalic or post-vocalic feature. The speech synthesizer front end can incorporates the supplemental phoneset as an extra feature.
    Type: Grant
    Filed: August 7, 2014
    Date of Patent: February 7, 2017
    Assignee: AT&T Intellectual Property I, L.P.
    Inventors: Alistair D. Conkie, Mark Beutnagel, Yeon-Jun Kim, Ann K. Syrdal
  • Patent number: 9495964
    Abstract: Disclosed herein are systems, computer-implemented methods, and tangible computer-readable storage media for captioning a media presentation. The method includes receiving automatic speech recognition (ASR) output from a media presentation and a transcription of the media presentation. The method includes selecting via a processor a pair of anchor words in the media presentation based on the ASR output and transcription and generating captions by aligning the transcription with the ASR output between the selected pair of anchor words. The transcription can be human-generated. Selecting pairs of anchor words can be based on a similarity threshold between the ASR output and the transcription. In one variation, commonly used words on a stop list are ineligible as anchor words. The method includes outputting the media presentation with the generated captions. The presentation can be a recording of a live event.
    Type: Grant
    Filed: March 16, 2016
    Date of Patent: November 15, 2016
    Assignee: AT&T Intellectual Property I, L.P.
    Inventors: Yeon-Jun Kim, David C. Gibbon, Horst J. Schroeter
  • Patent number: 9412359
    Abstract: Disclosed herein are systems, methods, and non-transitory computer-readable storage media for generating speech. One variation of the method is from a server side, and another variation of the method is from a client side. The server side method, as implemented by a network-based automatic speech processing system, includes first receiving, from a network client independent of knowledge of internal operations of the system, a request to generate a text-to-speech voice. The request can include speech samples, transcriptions of the speech samples, and metadata describing the speech samples. The system extracts sound units from the speech samples based on the transcriptions and generates an interactive demonstration of the text-to-speech voice based on the sound units, the transcriptions, and the metadata, wherein the interactive demonstration hides a back end processing implementation from the network client. The system provides access to the interactive demonstration to the network client.
    Type: Grant
    Filed: April 13, 2015
    Date of Patent: August 9, 2016
    Assignee: AT&T Intellectual Property I, L.P.
    Inventors: Mark Charles Beutnagel, Alistair D. Conkie, Yeon-Jun Kim, Horst Juergen Schroeter
  • Publication number: 20160198234
    Abstract: Disclosed herein are systems, computer-implemented methods, and tangible computer-readable storage media for captioning a media presentation. The method includes receiving automatic speech recognition (ASR) output from a media presentation and a transcription of the media presentation. The method includes selecting via a processor a pair of anchor words in the media presentation based on the ASR output and transcription and generating captions by aligning the transcription with the ASR output between the selected pair of anchor words. The transcription can be human-generated. Selecting pairs of anchor words can be based on a similarity threshold between the ASR output and the transcription. In one variation, commonly used words on a stop list are ineligible as anchor words. The method includes outputting the media presentation with the generated captions. The presentation can be a recording of a live event.
    Type: Application
    Filed: March 16, 2016
    Publication date: July 7, 2016
    Inventors: Yeon-Jun KIM, David C. GIBBON, Horst J. SCHROETER
  • Publication number: 20160171970
    Abstract: Disclosed herein are systems, methods, and non-transitory computer-readable storage media for detecting and correcting abnormal stress patterns in unit-selection speech synthesis. A system practicing the method detects incorrect stress patterns in selected acoustic units representing speech to be synthesized, and corrects the incorrect stress patterns in the selected acoustic units to yield corrected stress patterns. The system can further synthesize speech based on the corrected stress patterns. In one aspect, the system also classifies the incorrect stress patterns using a machine learning algorithm such as a classification and regression tree, adaptive boosting, support vector machine, and maximum entropy. In this way a text-to-speech unit selection speech synthesizer can produce more natural sounding speech with suitable stress patterns regardless of the stress of units in a unit selection database.
    Type: Application
    Filed: February 22, 2016
    Publication date: June 16, 2016
    Inventors: Yeon-Jun KIM, Mark Charles BEUTNAGEL, Alistair D. CONKIE, Ann K. Syrdal
  • Patent number: 9305552
    Abstract: Disclosed herein are systems, computer-implemented methods, and tangible computer-readable storage media for captioning a media presentation. The method includes receiving automatic speech recognition (ASR) output from a media presentation and a transcription of the media presentation. The method includes selecting via a processor a pair of anchor words in the media presentation based on the ASR output and transcription and generating captions by aligning the transcription with the ASR output between the selected pair of anchor words. The transcription can be human-generated. Selecting pairs of anchor words can be based on a similarity threshold between the ASR output and the transcription. In one variation, commonly used words on a stop list are ineligible as anchor words. The method includes outputting the media presentation with the generated captions. The presentation can be a recording of a live event.
    Type: Grant
    Filed: September 22, 2014
    Date of Patent: April 5, 2016
    Assignee: AT&T Intellectual Property I, L.P.
    Inventors: Yeon-Jun Kim, David C. Gibbon, Horst J. Schroeter
  • Publication number: 20160085858
    Abstract: Apparatuses, systems, methods, and media for filtering a data stream are provided. The data stream is analyzed based on an acoustic parameter to determine extraneous portions in which a first predetermined condition is satisfied. When a first extraneous portion is separated from a second extraneous portion by a non-extraneous portion in which the first predetermined condition is not satisfied, it is determined whether the first extraneous portion being separated from the second extraneous portion by the non-extraneous portion satisfies a second predetermined condition. At least one of the first extraneous portion and the second extraneous portion is deleted from the data stream to produce a filtered data stream in response to determining the second predetermined condition is satisfied.
    Type: Application
    Filed: September 25, 2015
    Publication date: March 24, 2016
    Applicant: AT&T INTELLECTUAL PROPERTY I, L.P.
    Inventors: Yeon-Jun KIM, I. Dan MELAMED, Bernard S. RENGER, Steven Neil TISCHER
  • Patent number: 9269348
    Abstract: Disclosed herein are systems, methods, and non-transitory computer-readable storage media for detecting and correcting abnormal stress patterns in unit-selection speech synthesis. A system practicing the method detects incorrect stress patterns in selected acoustic units representing speech to be synthesized, and corrects the incorrect stress patterns in the selected acoustic units to yield corrected stress patterns. The system can further synthesize speech based on the corrected stress patterns. In one aspect, the system also classifies the incorrect stress patterns using a machine learning algorithm such as a classification and regression tree, adaptive boosting, support vector machine, and maximum entropy. In this way a text-to-speech unit selection speech synthesizer can produce more natural sounding speech with suitable stress patterns regardless of the stress of units in a unit selection database.
    Type: Grant
    Filed: February 23, 2015
    Date of Patent: February 23, 2016
    Assignee: AT&T Intellectual Property I, L.P.
    Inventors: Yeon-Jun Kim, Mark Charles Beutnagel, Alistair D. Conkie, Ann K. Syrdal
  • Patent number: 9183177
    Abstract: Methods, apparatuses, and media for filtering a data stream are provided. The data stream is partitioned into a plurality of data stream segments. An acoustic parameter of each of the data stream segments is measured, and it is determined whether the acoustic parameter of each of the data stream segments satisfies a predetermined condition. Extraneous segments of the data stream segments are identified in which the predetermined condition is satisfied, and it is determined whether the extraneous segments have a predetermined relationship in the data stream. The extraneous segments are deleted from the data stream to produce a filtered data stream in response to the extraneous segments having the predetermined relationship.
    Type: Grant
    Filed: April 22, 2013
    Date of Patent: November 10, 2015
    Assignee: AT&T INTELLECTUAL PROPERTY I, L.P.
    Inventors: Yeon-Jun Kim, I. Dan Melamed, Steven Neil Tischer, Bernard S. Renger
  • Publication number: 20150220870
    Abstract: A method of detecting pre-determined phrases to determine compliance quality includes determining whether a precursor event has occurred based on a comparison between stored pre-determined phrases and a received communication, and determining a compliance rating based on a presence of a pre-determined phrase associated with the precursor event in the communication.
    Type: Application
    Filed: April 15, 2015
    Publication date: August 6, 2015
    Applicant: AT&T INTELLECTUAL PROPERTY I, L.P.
    Inventors: I. Dan MELAMED, Andrej LJOLJE, Bernard RENGER, Yeon-Jun KIM, David J. SMITH
  • Publication number: 20150221298
    Abstract: Disclosed herein are systems, methods, and non-transitory computer-readable storage media for generating speech. One variation of the method is from a server side, and another variation of the method is from a client side. The server side method, as implemented by a network-based automatic speech processing system, includes first receiving, from a network client independent of knowledge of internal operations of the system, a request to generate a text-to-speech voice. The request can include speech samples, transcriptions of the speech samples, and metadata describing the speech samples. The system extracts sound units from the speech samples based on the transcriptions and generates an interactive demonstration of the text-to-speech voice based on the sound units, the transcriptions, and the metadata, wherein the interactive demonstration hides a back end processing implementation from the network client. The system provides access to the interactive demonstration to the network client.
    Type: Application
    Filed: April 13, 2015
    Publication date: August 6, 2015
    Inventors: Mark Charles BEUTNAGEL, Alistair D. CONKIE, Yeon-Jun KIM, Horst Juergen SCHROETER