Patents by Inventor Richard J. Cartwright
Richard J. Cartwright has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Publication number: 20210035563Abstract: In some embodiments, methods and systems for training an acoustic model, where the training includes a training loop (including at least one epoch) following a data preparation phase. During the training loop, training data are augmented to generate augmented training data. During each epoch of the training loop, at least some of the augmented training data is used to train the model. The augmented training data used during each epoch may be generated by differently augmenting (e.g., augmenting using a different set of augmentation parameters) at least some of the training data. In some embodiments, the augmentation is performed in the frequency domain, with the training data organized into frequency bands. The acoustic model may be of a type employed (when trained) to perform speech analytics (e.g., wakeword detection, voice activity detection, speech recognition, or speaker recognition) and/or noise suppression.Type: ApplicationFiled: July 23, 2020Publication date: February 4, 2021Inventors: Richard J. Cartwright, Christopher Graham Hines
-
Patent number: 10872602Abstract: Computer-implemented methods for training an acoustic model for a far-field utterance processing system are provided. The acoustic model may be configured to map an input audio signal into linguistic or paralinguistic units. The training may involve imparting far-field acoustic characteristics upon near-field training vectors that include a plurality of near-microphone utterance signals. Imparting the far-field acoustic characteristics may involve generating a plurality of simulated room impulse responses, convolving one or more of the simulated room impulse responses with the near-field training vectors, to produce a plurality of simulated far-field utterance signals and saving the results of the training in one or more non-transitory memory devices corresponding with the acoustic model. Generating simulated room impulse responses may involve simulating room reverberation times but not simulating early reflections from room surfaces.Type: GrantFiled: May 2, 2019Date of Patent: December 22, 2020Assignee: Dolby Laboratories Licensing CorporationInventors: Hadis Nosrati, David S. McGrath, Richard J. Cartwright
-
Publication number: 20200396555Abstract: A computer implemented system for rendering captured audio soundfields to a listener comprises apparatus to deliver the audio soundfields to the listener. The delivery apparatus delivers the audio soundfields to the listener with first and second audio elements perceived by the listener as emanating from first and second virtual source locations, respectively, and with the first audio element and/or the second audio element delivered to the listener from a third virtual source location. The first virtual source location and the second virtual source location are perceived by the listener as being located to the front of the listener, and the third virtual source location is located to the rear or the side of the listener.Type: ApplicationFiled: June 22, 2020Publication date: December 17, 2020Applicant: DOLBY LABORATORIES LICENSING CORPORATIONInventors: Richard J. CARTWRIGHT, David S. MCGRATH, Glenn N. DICKINS
-
Patent number: 10812401Abstract: Disclosed is an apparatus and method operative to receive packets of media from a network including a receiver unit operative to receive the packets from the network, a jitter buffer data structure for receiving the packets in an ordered queue, the jitter buffer data structure having a tail into which the packets are input; a plurality of heads defining points in the jitter buffer data structure from which the ordered queue of packets are to be played back, the heads comprise an adjustable actual playback head coupled to an actual playback unit and at least one prototype head, each prototype head having associated therewith a target latency a processor having decision logic operable to determine a cost of achieving the associated target latency for each prototype head, wherein the decision logic compares the costs determined for each prototype head to identify a particular target latency and head location for the actual playback head of the buffer and a playback unit coupled to the processor for actual playbaType: GrantFiled: March 16, 2017Date of Patent: October 20, 2020Assignee: Dolby Laboratories Licensing CorporationInventors: Richard J. Cartwright, Hannes Muesch
-
Patent number: 10728688Abstract: Described herein is a method for creating an object-based audio signal from an audio input, the audio input including one or more audio channels that are recorded to collectively define an audio scene. The one or more audio channels are captured from a respective one or more spatially separated microphones disposed in a stable spatial configuration. The method includes the steps of: a) receiving the audio input; b) performing spatial analysis on the one or more audio channels to identify one or more audio objects within the audio scene; c) determining contextual information relating to the one or more audio objects; d) defining respective audio streams including audio data relating to at least one of the identified one or more audio objects; and e) outputting an object-based audio signal including the audio streams and the contextual information.Type: GrantFiled: May 28, 2019Date of Patent: July 28, 2020Assignee: Dolby Laboratories Licensing CorporationInventors: Glenn N. Dickins, Richard J. Cartwright
-
Patent number: 10708436Abstract: Embodiments are described for a soundfield system that receives a transmitting soundfield, wherein the transmitting soundfield includes a sound source at a location in the transmitting soundfield. The system determines a rotation angle for rotating the transmitting soundfield based on a desired location for the sound source. The transmitting soundfield is rotated by the determined angle and the system obtains a listener's soundfield based on the rotated transmitting soundfield. The listener's soundfield is transmitted for rendering to a listener.Type: GrantFiled: April 18, 2018Date of Patent: July 7, 2020Assignee: Dolby Laboratories Licensing CorporationInventor: Richard J. Cartwright
-
Patent number: 10694305Abstract: A computer implemented system for rendering captured audio soundfields to a listener comprises apparatus to deliver the audio soundfields to the listener. The delivery apparatus delivers the audio soundfields to the listener with first and second audio elements perceived by the listener as emanating from first and second virtual source locations, respectively, and with the first audio element and/or the second audio element delivered to the listener from a third virtual source location. The first virtual source location and the second virtual source location are perceived by the listener as being located to the front of the listener, and the third virtual source location is located to the rear or the side of the listener.Type: GrantFiled: July 22, 2019Date of Patent: June 23, 2020Assignee: Dolby Laboratories Licensing CorporationInventors: Richard J. Cartwright, David S. McGrath, Glenn N. Dickins
-
Publication number: 20200127865Abstract: Some aspects of the present disclosure involve the recording, processing and playback of audio data corresponding to conferences, such as teleconferences. In some teleconference implementations, the audio experience heard when a recording of the conference is played back may be substantially different from the audio experience of an individual conference participant during the original teleconference. In some implementations, the recorded audio data may include at least some audio data that was not available during the teleconference. In some examples, the spatial characteristics of the played-back audio data may be different from that of the audio heard by participants of the teleconference.Type: ApplicationFiled: December 19, 2019Publication date: April 23, 2020Applicant: DOLBY LABORATORIES LICENSING CORPORATIONInventors: Richard J. CARTWRIGHT, Glenn N. DICKINS
-
Publication number: 20200092422Abstract: Teleconference audio data including a plurality of individual uplink data packet streams, may be received during a teleconference. Each uplink data packet stream may corresponding to a telephone endpoint used by one or more teleconference participants. The teleconference audio data may be analyzed to determine a plurality of suppressive gain coefficients, which may be applied to first instances of the teleconference audio data during the teleconference, to produce first gain-suppressed audio data provided to the telephone endpoints during the teleconference. Second instances of the teleconference audio data, as well as gain coefficient data corresponding to the plurality of suppressive gain coefficients, may be sent to a memory system as individual uplink data packet streams. The second instances of the teleconference audio data may be less gain-suppressed than the first gain-suppressed audio data.Type: ApplicationFiled: November 21, 2019Publication date: March 19, 2020Applicant: DOLBY LABORATORIES LICENSING CORPORATIONInventors: Glenn N. DICKINS, Richard J. CARTWRIGHT
-
Patent number: 10567185Abstract: Some aspects of the present disclosure involve the recording, processing and playback of audio data corresponding to conferences, such as teleconferences. In some teleconference implementations, the audio experience heard when a recording of the conference is played back may be substantially different from the audio experience of an individual conference participant during the original teleconference. In some implementations, the recorded audio data may include at least some audio data that was not available during the teleconference. In some examples, the spatial characteristics of the played-back audio data may be different from that of the audio heard by participants of the teleconference.Type: GrantFiled: February 3, 2016Date of Patent: February 18, 2020Assignee: Dolby Laboratories Licensing CorporationInventors: Richard J. Cartwright, Glenn N. Dickins
-
Publication number: 20200021935Abstract: A computer implemented system for rendering captured audio soundfields to a listener comprises apparatus to deliver the audio soundfields to the listener. The delivery apparatus delivers the audio soundfields to the listener with first and second audio elements perceived by the listener as emanating from first and second virtual source locations, respectively, and with the first audio element and/or the second audio element delivered to the listener from a third virtual source location. The first virtual source location and the second virtual source location are perceived by the listener as being located to the front of the listener, and the third virtual source location is located to the rear or the side of the listener.Type: ApplicationFiled: July 22, 2019Publication date: January 16, 2020Applicant: DOLBY LABORATORIES LICENSING CORPORATIONInventors: Richard J. CARTWRIGHT, David S. MCGRATH, Glenn N. DICKINS
-
Patent number: 10522151Abstract: Various disclosed implementations involve processing and/or playback of a recording of a conference involving a plurality of conference participants. Some implementations disclosed herein involve analyzing conversational dynamics of the conference recording. Some examples may involve searching the conference recording to determine instances of segment classifications. The segment classifications may be based, at least in part, on conversational dynamics data. Some implementations may involve segmenting the conference recording into a plurality of segments, each of the segments corresponding with a time interval and at least one of the segment classifications. Some implementations allow a listener to scan through a conference recording quickly according to segments, words, topics and/or talkers of interest.Type: GrantFiled: February 3, 2016Date of Patent: December 31, 2019Assignee: Dolby Laboratories Licensing CorporationInventors: Richard J. Cartwright, Kai Li, Xuejing Sun
-
Patent number: 10516782Abstract: Various disclosed implementations involve processing and/or playback of a recording of a conference involving a plurality of conference participants. Some implementations disclosed herein involve receiving audio data corresponding to a recording of at least one conference involving a plurality of conference participants. The audio data may include conference participant speech data from multiple endpoints, recorded separately and/or conference participant speech data from a single endpoint corresponding to multiple conference participants and including spatial information for each conference participant of the multiple conference participants. A search of the audio data may be based on one or more search parameters. The search may be a concurrent search for multiple features of the audio data. Instances of conference participant speech may be rendered to at least two different virtual conference participant positions of a virtual acoustic space.Type: GrantFiled: February 3, 2016Date of Patent: December 24, 2019Assignee: Dolby Laboratories Licensing CorporationInventors: Richard J. Cartwright, Shen Huang
-
Patent number: 10511718Abstract: Teleconference audio data including a plurality of individual uplink data packet streams, may be received during a teleconference. Each uplink data packet stream may corresponding to a telephone endpoint used by one or more teleconference participants. The teleconference audio data may be analyzed to determine a plurality of suppressive gain coefficients, which may be applied to first instances of the teleconference audio data during the teleconference, to produce first gain-suppressed audio data provided to the telephone endpoints during the teleconference. Second instances of the teleconference audio data, as well as gain coefficient data corresponding to the plurality of suppressive gain coefficients, may be sent to a memory system as individual uplink data packet streams. The second instances of the teleconference audio data may be less gain-suppressed than the first gain-suppressed audio data.Type: GrantFiled: June 15, 2016Date of Patent: December 17, 2019Assignee: Dolby Laboratories Licensing CorporationInventors: Glenn N. Dickins, Richard J. Cartwright
-
Publication number: 20190362711Abstract: Computer-implemented methods for training an acoustic model for a far-field utterance processing system are provided. The acoustic model may be configured to map an input audio signal into linguistic or paralinguistic units. The training may involve imparting far-field acoustic characteristics upon near-field training vectors that include a plurality of near-microphone utterance signals. Imparting the far-field acoustic characteristics may involve generating a plurality of simulated room impulse responses, convolving one or more of the simulated room impulse responses with the near-field training vectors, to produce a plurality of simulated far-field utterance signals and saving the results of the training in one or more non-transitory memory devices corresponding with the acoustic model. Generating simulated room impulse responses may involve simulating room reverberation times but not simulating early reflections from room surfaces.Type: ApplicationFiled: May 2, 2019Publication date: November 28, 2019Applicant: Dolby Laboratories Licensing CorporationInventors: Hadis Nosrati, David S. McGrath, Richard J. Cartwright
-
Patent number: 10439951Abstract: Disclosed is a method and apparatus operative to process packets of media received from a network including a receiver unit operative, a jitter buffer data structure and a playback head defining a point in the jitter buffer data structure from which the ordered queue of packets are to be played back, and at least one prototype head. Each prototype head having a predetermined latency assigned thereto and defining a point in the jitter buffer data structure from which the ordered queue of packets is being played back containing said latency a processor operable to determine a measure of conversational quality associated with the ordered queue of packets being played back by each prototype head. Also described is a head selector operable to compare the measures of conversational quality associated with the ordered queue of packets being played back by each prototype head to select the prototype head with the highest measure of conversational quality and a playback unit coupled to the playback head.Type: GrantFiled: March 16, 2017Date of Patent: October 8, 2019Assignee: Dolby Laboratories Licensing CorporationInventors: Hannes Muesch, Richard J. Cartwright
-
Publication number: 20190281404Abstract: Described herein is a method for creating an object-based audio signal from an audio input, the audio input including one or more audio channels that are recorded to collectively define an audio scene. The one or more audio channels are captured from a respective one or more spatially separated microphones disposed in a stable spatial configuration. The method includes the steps of: a) receiving the audio input; b) performing spatial analysis on the one or more audio channels to identify one or more audio objects within the audio scene; c) determining contextual information relating to the one or more audio objects; d) defining respective audio streams including audio data relating to at least one of the identified one or more audio objects; and e) outputting an object-based audio signal including the audio streams and the contextual information.Type: ApplicationFiled: May 28, 2019Publication date: September 12, 2019Applicant: Dolby Laboratories Licensing CorporationInventors: Glenn N. Dickins, Richard J. Cartwright
-
Patent number: 10362420Abstract: A computer implemented system for rendering captured audio soundfields to a listener comprises apparatus to deliver the audio soundfields to the listener. The delivery apparatus delivers the audio soundfields to the listener with first and second audio elements perceived by the listener as emanating from first and second virtual source locations, respectively, and with the first audio element and/or the second audio element delivered to the listener from a third virtual source location. The first virtual source location and the second virtual source location are perceived by the listener as being located to the front of the listener, and the third virtual source location is located to the rear or the side of the listener.Type: GrantFiled: June 14, 2018Date of Patent: July 23, 2019Assignee: Dolby Laboratories Licensing CorporationInventors: Richard J. Cartwright, David S. McGrath, Glenn N. Dickins
-
Patent number: 10334384Abstract: A method for processing audio data, the method comprising: receiving audio data corresponding to a plurality of instances of audio, including at least one of: (a) audio data from multiple endpoints, recorded separately or (b) audio data from a single endpoint corresponding to multiple talkers and including spatial information for each of the multiple talkers; rendering the audio data in a virtual acoustic space such that each of the instances of audio has a respective different virtual position in the virtual acoustic space; and scheduling the instances of audio to be played back with a playback overlap between at least two of the instances of audio, wherein the scheduling is performed, at least in part, according to a set of perceptually-motivated rules.Type: GrantFiled: February 3, 2016Date of Patent: June 25, 2019Assignee: Dolby Laboratories Licensing CorporationInventors: Xuejing Sun, Richard J. Cartwright, Michael P. Hollier, Michael Eckert
-
Patent number: 10321256Abstract: Systems, methods, and computer program products for creating an object-based audio signal from an audio input are described. The audio input includes one or more audio channels that are recorded to collectively define an audio scene. The one or more audio channels are captured from a respective one or more spatially separated microphones disposed in a stable spatial configuration. A system receives the audio input. The system performs spatial analysis on the one or more audio channels to identify one or more audio objects within the audio scene. The system determines contextual information relating to the one or more audio objects. The system defines respective audio streams including audio data relating to at least one of the identified one or more audio objects. The system then outputs an object-based audio signal including the audio streams and the contextual information.Type: GrantFiled: February 2, 2016Date of Patent: June 11, 2019Assignee: Dolby Laboratories Licensing CorporationInventors: Glenn N. Dickins, Richard J. Cartwright