Patents by Inventor Horst J. Schroeter

Horst J. Schroeter has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20170061986
    Abstract: Disclosed herein are systems, computer-implemented methods, and tangible computer-readable storage media for captioning a media presentation. The method includes receiving automatic speech recognition (ASR) output from a media presentation and a transcription of the media presentation. The method includes selecting via a processor a pair of anchor words in the media presentation based on the ASR output and transcription and generating captions by aligning the transcription with the ASR output between the selected pair of anchor words. The transcription can be human-generated. Selecting pairs of anchor words can be based on a similarity threshold between the ASR output and the transcription. In one variation, commonly used words on a stop list are ineligible as anchor words. The method includes outputting the media presentation with the generated captions. The presentation can be a recording of a live event.
    Type: Application
    Filed: November 14, 2016
    Publication date: March 2, 2017
    Inventors: Yeon-Jun KIM, David C. GIBBON, Horst J. SCHROETER
  • Patent number: 9584500
    Abstract: Methods, systems, and products authenticate users for access to devices, applications, and services. Skills of a user are learned over time, such that an electronic model of random subject matter may be generated. The user is prompted to interpret the random subject matter, such as with an electronic drawing. The user's interpretation is then compared to the electronic model of the random subject matter. If the user's interpretation matches the electronic model, the user may be authenticated.
    Type: Grant
    Filed: June 12, 2015
    Date of Patent: February 28, 2017
    Assignee: AT&T INTELLECTUAL PROPERTY I, L.P.
    Inventors: Kevin A. Li, Troy C. Meuninck, Robert Raymond Miller, II, James H. Pratt, Horst J. Schroeter, Behzad Shahraray
  • Patent number: 9575997
    Abstract: Methods, systems, and products help users recall memories and search for content of those memories. When a user cannot recall a memory, the user is prompted with questions to help recall the memory. As the user answers the questions, a virtual recollection of the memory is synthesized from the answers to the questions. When the user is satisfied with the virtual recollection of the memory, a database of content may be searched for the virtual recollection of the memory. Video data, for example, may be retrieved that matches the virtual recollection of the memory. The video data is thus historical data documenting past events.
    Type: Grant
    Filed: July 12, 2015
    Date of Patent: February 21, 2017
    Assignee: AT&T INTELLECTUAL PROPERTY I, L.P.
    Inventors: Robert Raymond Miller, II, Kevin A. Li, Troy C. Meuninck, James H. Pratt, Horst J. Schroeter, Behzad Shahraray
  • Publication number: 20170046335
    Abstract: Methods, systems, and computer program products provide personalized feedback in a cloud-based environment. A client device routes image data to a server for analysis. The server analyzes the image data to recognize people of interest. Because the server performs image recognition, the client device is relieved of these intensive operations.
    Type: Application
    Filed: October 31, 2016
    Publication date: February 16, 2017
    Applicant: AT&T Intellectual Property I, L.P.
    Inventors: Dimitrios B. Dimitriadis, Horst J. Schroeter
  • Patent number: 9570092
    Abstract: Devices, systems, methods, media, and programs for detecting an emotional state change in an audio signal are provided. A number of segments of the audio signal are analyzed based on separate lexical and acoustic evaluations, and, for each segment, an emotional state and a confidence score of the emotional state are determined. A current emotional state of the audio signal is tracked for each of the number of segments. For a particular segment, it is determined whether the current emotional state of the audio signal changes to another emotional state based on the emotional state and a comparison of the confidence score of the particular segment to a predetermined threshold.
    Type: Grant
    Filed: April 26, 2016
    Date of Patent: February 14, 2017
    Assignee: AT&T INTELLECTUAL PROPERTY I, L.P.
    Inventors: Dimitrios Dimitriadis, Mazin E. Gilbert, Taniya Mishra, Horst J. Schroeter
  • Patent number: 9558735
    Abstract: Disclosed herein are systems, methods, and computer readable-media for providing an automatic synthetically generated voice describing media content, the method comprising receiving one or more pieces of metadata for a primary media content, selecting at least one piece of metadata for output, and outputting the at least one piece of metadata as synthetically generated speech with the primary media content. Other aspects of the invention involve alternative output, output speech simultaneously with the primary media content, output speech during gaps in the primary media content, translate metadata in foreign language, tailor voice, accent, and language to match the metadata and/or primary media content. A user may control output via a user interface or output may be customized based on preferences in a user profile.
    Type: Grant
    Filed: March 18, 2016
    Date of Patent: January 31, 2017
    Assignee: AT&T Intellectual Property I, L.P.
    Inventors: Linda Roberts, Hong Thi Nguyen, Horst J Schroeter
  • Patent number: 9530408
    Abstract: A system for providing an acoustic environment recognizer for optimal speech processing is disclosed. In particular, the system may utilize metadata obtained from various acoustic environments to assist in suppressing ambient noise interfering with a desired audio signal. In order to do so, the system may receive an audio stream including an audio signal associated with a user and including ambient noise obtained from an acoustic environment of the user. The system may obtain first metadata associated with the ambient noise, and may determine if the first metadata corresponds to second metadata in a profile for the acoustic environment. If the first metadata corresponds to the second metadata, the system may select a processing scheme for suppressing the ambient noise from the audio stream, and process the audio stream using the processing scheme. Once the audio stream is processed, the system may provide the audio stream to a destination.
    Type: Grant
    Filed: October 31, 2014
    Date of Patent: December 27, 2016
    Assignee: AT&T INTELLECTUAL PROPERTY I, L.P.
    Inventors: Horst J. Schroeter, Donald J. Bowen, Dimitrios B. Dimitriadis, Lusheng Ji
  • Patent number: 9525767
    Abstract: Disclosed herein are systems, methods, and computer readable-media for answering a communication notification. The method for answering a communication notification comprises receiving a notification of communication from a user, converting information related to the notification to speech, outputting the information as speech to the user, and receiving from the user an instruction to accept or ignore the incoming communication associated with the notification. In one embodiment, information related to the notification comprises one or more of a telephone number, an area code, a geographic origin of the request, caller id, a voice message, address book information, a text message, an email, a subject line, an importance level, a photograph, a video clip, metadata, an IP address, or a domain name. Another embodiment involves notification assigned an importance level and repeat attempts at notification if it is of high importance.
    Type: Grant
    Filed: March 2, 2016
    Date of Patent: December 20, 2016
    Assignee: AT&T Intellectual Property I, L.P.
    Inventor: Horst J. Schroeter
  • Patent number: 9507770
    Abstract: Methods, systems, and computer program products provide personalized feedback in a cloud-based environment. A client device routes image data to a server for analysis. The server analyzes the image data to recognize people of interest. Because the server performs image recognition, the client device is relieved of these intensive operations.
    Type: Grant
    Filed: August 15, 2015
    Date of Patent: November 29, 2016
    Assignee: AT&T INTELLECTUAL PROPERTY I, L.P.
    Inventors: Dimitrios B. Dimitriadis, Horst J. Schroeter
  • Publication number: 20160343379
    Abstract: Disclosed herein are systems, methods, and tangible computer readable-media for detecting synthetic speaker verification. The method comprises receiving a plurality of speech samples of the same word or phrase for verification, comparing each of the plurality of speech samples to each other, denying verification if the plurality of speech samples demonstrate little variance over time or are the same, and verifying the plurality of speech samples if the plurality of speech samples demonstrates sufficient variance over time. One embodiment further adds that each of the plurality of speech samples is collected at different times or in different contexts. In other embodiments, variance is based on a pre-determined threshold or the threshold for variance is adjusted based on a need for authentication certainty. In another embodiment, if the initial comparison is inconclusive, additional speech samples are received.
    Type: Application
    Filed: August 5, 2016
    Publication date: November 24, 2016
    Inventor: Horst J. Schroeter
  • Patent number: 9495964
    Abstract: Disclosed herein are systems, computer-implemented methods, and tangible computer-readable storage media for captioning a media presentation. The method includes receiving automatic speech recognition (ASR) output from a media presentation and a transcription of the media presentation. The method includes selecting via a processor a pair of anchor words in the media presentation based on the ASR output and transcription and generating captions by aligning the transcription with the ASR output between the selected pair of anchor words. The transcription can be human-generated. Selecting pairs of anchor words can be based on a similarity threshold between the ASR output and the transcription. In one variation, commonly used words on a stop list are ineligible as anchor words. The method includes outputting the media presentation with the generated captions. The presentation can be a recording of a live event.
    Type: Grant
    Filed: March 16, 2016
    Date of Patent: November 15, 2016
    Assignee: AT&T Intellectual Property I, L.P.
    Inventors: Yeon-Jun Kim, David C. Gibbon, Horst J. Schroeter
  • Publication number: 20160330394
    Abstract: Network connectivity is used to share relevant visual and other sensory information between vehicles, as well as delivering relevant information provided by network services to create an enhanced view of the vehicle's surroundings. The enhanced view is presented to the occupants of the vehicle to provide an improved driving experience and/or enable the occupants to take proper action (e.g., avoid obstacles, identify traffic delays, etc.). In one example, the enhanced view comprises information that is not visible to the naked eye and/or cannot be currently sensed by the vehicle's sensors (e.g., due to a partial or blocked view, low visibility conditions, hardware capabilities of the vehicle's sensors, position of the vehicle's sensors, etc.).
    Type: Application
    Filed: July 18, 2016
    Publication date: November 10, 2016
    Inventors: Behzad Shahraray, Alicia Abella, David Crawford Gibbon, Mazin E. Gilbert, Michael J. Johnston, Horst J. Schroeter, Jay Gordon Wilpon
  • Patent number: 9484018
    Abstract: Disclosed herein are systems, methods, and non-transitory computer-readable storage media for building an automatic speech recognition system through an Internet API. A network-based automatic speech recognition server configured to practice the method receives feature streams, transcriptions, and parameter values as inputs from a network client independent of knowledge of internal operations of the server. The server processes the inputs to train an acoustic model and a language model, and transmits the acoustic model and the language model to the network client. The server can also generate a log describing the processing and transmit the log to the client. On the server side, a human expert can intervene to modify how the server processes the inputs. The inputs can include an additional feature stream generated from speech by algorithms in the client's proprietary feature extraction.
    Type: Grant
    Filed: November 23, 2010
    Date of Patent: November 1, 2016
    Assignee: AT&T Intellectual Property I, L.P.
    Inventors: Enrico Bocchieri, Dimitrios Dimitriadis, Horst J. Schroeter
  • Publication number: 20160253478
    Abstract: Methods, systems, and products monitor a person's regimen for medicinal and dietary restrictions. When the person's regimen requires a liquid medication or supplement, an oral instrument is commanded to dispense a dosage of fluid. The oral instrument stores a reservoir of the fluid. If the oral instrument is a spoon, for example, the spoon may automatically dispense cough syrup or other medicine. A toothbrush, likewise, may automatically dispense mouthwash. A sensor may confirm presence of the oral instrument in the person's mouth, thus ensuring the dosage of fluid is ingested.
    Type: Application
    Filed: March 21, 2016
    Publication date: September 1, 2016
    Inventors: Kevin Ansia Li, TROY C. MEUNINCK, ROBERT RAYMOND MILLER, II, JAMES H. PRATT, HORST J. SCHROETER, BEHZAD SHAHRARAY
  • Publication number: 20160253988
    Abstract: A system for self-organized acoustic signal cancellation over a network is disclosed. The system may transmit an acoustic sounding signal to an interfering device so that a channel measurement may be performed for a channel between the interfering device and an interferee device. The system may receive the channel measurement for the channel from the interfering device and also receive a digitized version of an audio interference signal associated with the interfering device. Based on the channel measurement and the digital version of the interference signal, the system may calculate a cancellation signal prior to the arrival of the original over-the-air audio interference signal that corresponds to the digital version of audio interference signal. The system may then apply the cancellation signal to an audio signal associated with the interferee device to remove the interference signal from the audio signal.
    Type: Application
    Filed: May 9, 2016
    Publication date: September 1, 2016
    Applicant: AT&T Intellectual Property I, L.P.
    Inventors: Lusheng Ji, Donald J. Bowen, Dimitrios B. Dimitriadis, Horst J. Schroeter
  • Publication number: 20160240214
    Abstract: Devices, systems, methods, media, and programs for detecting an emotional state change in an audio signal are provided. A number of segments of the audio signal are analyzed based on separate lexical and acoustic evaluations, and, for each segment, an emotional state and a confidence score of the emotional state are determined. A current emotional state of the audio signal is tracked for each of the number of segments. For a particular segment, it is determined whether the current emotional state of the audio signal changes to another emotional state based on the emotional state and a comparison of the confidence score of the particular segment to a predetermined threshold.
    Type: Application
    Filed: April 26, 2016
    Publication date: August 18, 2016
    Inventors: DIMITRIOS DIMITRIADIS, MAZIN E. GILBERT, TANIYA MISHRA, HORST J. SCHROETER
  • Patent number: 9412382
    Abstract: Disclosed herein are systems, methods, and tangible computer readable-media for detecting synthetic speaker verification. The method comprises receiving a plurality of speech samples of the same word or phrase for verification, comparing each of the plurality of speech samples to each other, denying verification if the plurality of speech samples demonstrate little variance over time or are the same, and verifying the plurality of speech samples if the plurality of speech samples demonstrates sufficient variance over time. One embodiment further adds that each of the plurality of speech samples is collected at different times or in different contexts. In other embodiments, variance is based on a pre-determined threshold or the threshold for variance is adjusted based on a need for authentication certainty. In another embodiment, if the initial comparison is inconclusive, additional speech samples are received.
    Type: Grant
    Filed: September 21, 2015
    Date of Patent: August 9, 2016
    Assignee: AT&T Intellectual Property I, L.P.
    Inventor: Horst J. Schroeter
  • Patent number: 9403482
    Abstract: Network connectivity is used to share relevant visual and other sensory information between vehicles, as well as delivering relevant information provided by network services to create an enhanced view of the vehicle's surroundings. The enhanced view is presented to the occupants of the vehicle to provide an improved driving experience and/or enable the occupants to take proper action (e.g., avoid obstacles, identify traffic delays, etc.). In one example, the enhanced view comprises information that is not visible to the naked eye and/or cannot be currently sensed by the vehicle's sensors (e.g., due to a partial or blocked view, low visibility conditions, hardware capabilities of the vehicle's sensors, position of the vehicle's sensors, etc.).
    Type: Grant
    Filed: November 22, 2013
    Date of Patent: August 2, 2016
    Assignee: AT&T INTELLECTUAL PROPERTY I, L.P.
    Inventors: Behzad Shahraray, Alicia Abella, David Crawford Gibbon, Mazin E. Gilbert, Michael J. Johnston, Horst J. Schroeter, Jay Gordon Wilpon
  • Publication number: 20160203815
    Abstract: Disclosed herein are systems, methods, and computer readable-media for providing an automatic synthetically generated voice describing media content, the method comprising receiving one or more pieces of metadata for a primary media content, selecting at least one piece of metadata for output, and outputting the at least one piece of metadata as synthetically generated speech with the primary media content. Other aspects of the invention involve alternative output, output speech simultaneously with the primary media content, output speech during gaps in the primary media content, translate metadata in foreign language, tailor voice, accent, and language to match the metadata and/or primary media content. A user may control output via a user interface or output may be customized based on preferences in a user profile.
    Type: Application
    Filed: March 18, 2016
    Publication date: July 14, 2016
    Inventors: LINDA ROBERTS, HONG THI NGUYEN, HORST J. SCHROETER
  • Publication number: 20160198234
    Abstract: Disclosed herein are systems, computer-implemented methods, and tangible computer-readable storage media for captioning a media presentation. The method includes receiving automatic speech recognition (ASR) output from a media presentation and a transcription of the media presentation. The method includes selecting via a processor a pair of anchor words in the media presentation based on the ASR output and transcription and generating captions by aligning the transcription with the ASR output between the selected pair of anchor words. The transcription can be human-generated. Selecting pairs of anchor words can be based on a similarity threshold between the ASR output and the transcription. In one variation, commonly used words on a stop list are ineligible as anchor words. The method includes outputting the media presentation with the generated captions. The presentation can be a recording of a live event.
    Type: Application
    Filed: March 16, 2016
    Publication date: July 7, 2016
    Inventors: Yeon-Jun KIM, David C. GIBBON, Horst J. SCHROETER