Audio To Video Patents (Class 348/515)
  • Patent number: 11461072
    Abstract: A movie theater audio distribution system includes a visual display associated with a theater, the visual display having a transmitter; a server to control the visual display; a computer having an audio programing platform to command the server to control the visual display; a headphone device having a transceiver to communicate wirelessly with the server, the headphone device having a control system to receive commands from the server; the audio programming platform provides a way to command the server to transmit an audio to the headphone device correlated to the visual display; and the transmitter is to wirelessly communicate with the transceiver thereby activating the audio associated with the visual display to play through the headphone device when the headphone device is in close proximity to the visual display.
    Type: Grant
    Filed: September 10, 2020
    Date of Patent: October 4, 2022
    Inventor: Stacey Castillo
  • Patent number: 11443739
    Abstract: Coordinated operation of a voice-controlled device and an accessory device in an environment is described. A remote system processes audio data it receives from the voice-controlled device in the environment to identify a first intent associated with a first domain, a second intent associated with a second domain, and a named entity associated with the audio data. The remote system sends, to the voice-controlled device, first information for accessing main content associated with the named entity, and a first instruction corresponding to the first intent. The remote system also sends, to the accessory device, second information for accessing control information or supplemental content associated with the main content, and a second instruction corresponding to the second intent. The first and second instructions, when processed by the devices in the environment, cause coordinated operation of the voice-controlled device and the accessory device.
    Type: Grant
    Filed: October 30, 2019
    Date of Patent: September 13, 2022
    Assignee: Amazon Technologies, Inc.
    Inventors: Derick Deller, Apoorv Naik, Zoe Adams, Aslan Appleman, Link Cornelius, Pete Klein
  • Patent number: 11436780
    Abstract: A method for matching mouth shape and movement in digital video to alternative audio includes deriving a sequence of facial poses including mouth shapes for an actor from a source digital video. Each pose in the sequence of facial poses corresponds to a middle position of each audio sample. The method further includes generating an animated face mesh based on the sequence of facial poses and the source digital video, transferring tracked expressions from the animated face mesh or the target video to the source video, and generating a rough output video that includes transfers of the tracked expressions. The method further includes generating a finished video at least in part by refining the rough video using a parametric autoencoder trained on mouth shapes in the animated face mesh or the target video. One or more computers may perform the operations of the method.
    Type: Grant
    Filed: November 23, 2020
    Date of Patent: September 6, 2022
    Assignee: WARNER BROS. ENTERTAINMENT INC.
    Inventors: Tom David Stratton, Shaun Lile
  • Patent number: 11403676
    Abstract: Provided herein are systems and methods of classifying video content. At least one server can identify a video content item identifying a plurality of segments to play primary video content. The at least one server can identify a set of words from a segment of the plurality of segments by using at least one of a transcription corresponding to the segment or speech recognition on audio content corresponding to the segment. The at least one server can determine a classification for the segment based on the set of words from the segment. The at least one server can store, in one or more data structures, an association between the video content item and the classification to categorize the segment of the video content item.
    Type: Grant
    Filed: June 8, 2020
    Date of Patent: August 2, 2022
    Assignee: GOOGLE LLC
    Inventors: Jason S. Bayer, Ronojoy Chakrabarti, Keval Desai, Manish P Gupta, Jill A Huchital, Willard V T Rusch, II
  • Patent number: 11343552
    Abstract: Systems and methods of the present disclosure provide for dynamic delay equalization of related media signals in a media transport system. Methods include receiving a plurality of related media signals, transporting the related media signals along different media paths, calculating uncorrected propagation delays for the media paths, and delaying each of the related media signals by an amount related to the difference between the longest propagation delay (of the uncorrected propagation delays) and the uncorrected propagation delay of the related media signal/media path. Calculating the uncorrected propagation delays and delaying the related media signals may be performed in response to a change to the propagation delay of at least one of the related media signals/media paths. Additionally or alternatively, calculating the uncorrected propagation delays and delaying the related media signals may be performed while transporting the related media signals.
    Type: Grant
    Filed: September 1, 2020
    Date of Patent: May 24, 2022
    Assignee: Biamp Systems, LLC
    Inventors: Eugene Gurfinkel, Michael K. Davis, Charles H. Van Dusen
  • Patent number: 11342002
    Abstract: An automated solution to determine suitable time ranges or timestamps for captions is described. In one example, a content file includes subtitle data with captions for display over respective timeframes of video. Audio data is extracted from the video, and the audio data is compared against a sound threshold to identify auditory timeframes in which sound is above the threshold. The subtitle data is also parsed to identify subtitle-free timeframes in the video. A series of candidate time ranges is then identified based on overlapping ranges of the auditory timeframes and the subtitle-free timeframes. In some cases, one or more of the candidate time ranges can be merged together or omitted, and a final series of time ranges or timestamps for captions is obtained. The time ranges or timestamps can be used to add additional non-verbal and contextual captions and indicators, for example, or for other purposes.
    Type: Grant
    Filed: December 5, 2018
    Date of Patent: May 24, 2022
    Assignee: AMAZON TECHNOLOGIES, INC.
    Inventors: Prabhakar Gupta, Shaktisingh P Shekhawat, Kumar Keshav
  • Patent number: 11330151
    Abstract: An apparatus, method and computer program product for receiving captured visual information comprising a representation of an object, receiving captured audio information associated with the object, determining a user awareness parameter indicating a level of user comprehension of a context of capturing the visual information and the audio information and selecting, based on the user awareness parameter, a type of synchronization of the captured audio information with respect to the captured visual information.
    Type: Grant
    Filed: April 10, 2020
    Date of Patent: May 10, 2022
    Assignee: Nokia Technologies Oy
    Inventors: Matti Kajala, Antero Tossavainen, Mikko Olavi Heikkinen, Miikka Tapani Vilermo
  • Patent number: 11328011
    Abstract: A method includes computing match scores for each portion of multiple portions of a first audio fingerprint. The match scores are based on a comparison of the portion with each of multiple portions of a second audio fingerprint. The method includes generating a list of runs based on the highest score for each portion of the multiple portions of the first audio fingerprint. The method includes determining, based on the list of runs, an unordered match between a set of consecutive portions of the first audio fingerprint and a set of non-consecutive portions of the second audio fingerprint. The method includes, in response to determining that a position threshold of the unordered match satisfies a position criterion, outputting an indicator that the first audio fingerprint matches the second audio fingerprint.
    Type: Grant
    Filed: October 1, 2020
    Date of Patent: May 10, 2022
    Assignee: iHeartMedia Management Services, Inc.
    Inventor: Dyon Anniballi
  • Patent number: 11271666
    Abstract: A networked system is provided for transporting digital media packets, such as audio and video. The network includes network devices interconnected to send and receive packets. Each network device can receive and transmit media signals from media devices. A master clock generates a system time signal that the network devices use, together with a network time protocol to generate a local clock signal synchronised to the system time signal for both rate and offset. The local clock signal governs both the rate and offset of the received or transmitted media signals. The system, which can be implemented using conventional network equipment enables media signals to be transported to meet quality and timing requirements for high quality audio and video reproduction.
    Type: Grant
    Filed: September 24, 2019
    Date of Patent: March 8, 2022
    Assignee: AUDINATE HOLDINGS PTY LIMITED
    Inventors: Aidan Williams, Varuni Witana
  • Patent number: 11244418
    Abstract: A method of encoding a watermark into a digital image is provided. The method includes partitioning an image into a plurality of blocks of a same size; accumulating the plurality of blocks of the same size into a single block image; performing a Fourier transformation on the single block image to obtain a two-dimensional Fourier spectrum defined by Fourier coefficients at different positions of a Fourier domain; inserting a watermark into a frequency domain of the two-dimensional Fourier spectrum by modifying the two-dimensional Fourier spectrum as a function of watermarking coefficients in the watermark, to obtain a modified Fourier spectrum; performing an inverse Fourier transformation on the modified Fourier spectrum to obtain a watermarked image; copying the watermarked image horizontally and vertically into a plurality of copied watermarked images; and splicing the plurality of copied watermarked images into a reconstituted watermark image.
    Type: Grant
    Filed: March 15, 2019
    Date of Patent: February 8, 2022
    Assignee: BOE Technology Group Co., Ltd.
    Inventor: Xiaojun Tang
  • Patent number: 11232805
    Abstract: The disclosure relates to methods, apparatus and systems for side load processing of packetized media streams. In an embodiment, the apparatus comprises: a receiver for receiving a bitstream, and a splitter for identifying a packet type in the bitstream and splitting, based on the identification of a value of the packet type in the bit stream into a main stream and an auxiliary stream.
    Type: Grant
    Filed: February 22, 2019
    Date of Patent: January 25, 2022
    Assignee: Dolby International AB
    Inventors: Stephan Schreiner, Christof Fersch
  • Patent number: 11234037
    Abstract: A projector includes: a projection unit which projects content in response to a playback instruction to play back the content; and a transmitting unit which transmits the playback instruction to another projector. The playback instruction includes specification information which specifies the content.
    Type: Grant
    Filed: July 18, 2018
    Date of Patent: January 25, 2022
    Assignee: SEIKO EPSON CORPORATION
    Inventors: Kazuyoshi Kitabayashi, Takahiro Otsu
  • Patent number: 11190806
    Abstract: A display apparatus is disclosed. The display apparatus includes a display, a communication interface, a receiver, and a processor configured to decode an encoded video frame and an encoded audio frame, received through the receiver, transmit information on decoding time of the decoded video frame to an audio apparatus through the communication interface, delay the decoded audio frame by a first time, and transmit information on decoding time of the decoded audio frame, information on the first time, and an audio frame delayed by the first time to the audio apparatus through the communication interface, in response to the transmission, receive information on a second time delayed in the audio apparatus to output the audio frame from the audio apparatus through the communication interface, and synchronize an audio frame output from the audio apparatus with a video frame output through the display based on the information on the second time.
    Type: Grant
    Filed: August 5, 2020
    Date of Patent: November 30, 2021
    Assignee: SAMSUNG ELECTRONICS CO., LTD.
    Inventors: Jaecheol Lee, Haejong Kim
  • Patent number: 11119727
    Abstract: Digital tutorial generation techniques and systems are described in which a digital tutorial is generated automatically and without user intervention. History data is generated describing a sequence of user inputs provided as part of user interaction with an application and audio data is received capturing user utterances, e.g., speech, from a microphone of the computing device. A step-identification module of the tutorial generation system identifies a plurality of tutorial steps based on a sequence of user inputs described by the history data. A segmentation module of the tutorial generation system then generates a plurality of audio segments from the audio data corresponding to respective ones of the plurality of tutorial steps. The digital tutorial is then generated by a synchronization module of the tutorial generation system by synchronizing the plurality of audio segments as part of the plurality of tutorial steps, which is then output.
    Type: Grant
    Filed: June 25, 2020
    Date of Patent: September 14, 2021
    Assignee: Adobe Inc.
    Inventors: Subham Gupta, Sudhir Tubegere Shankaranarayana, Jaideep Jeyakar, Ashutosh Dwivedi
  • Patent number: 11108935
    Abstract: In one example embodiment, a camera system includes a plurality of cameras, a camera controller configured to control the plurality of cameras, a control signal line configured to facilitate an exchange of at least one control signal between the camera controller and the plurality of cameras and a synchronization signal line commonly connected to the plurality of cameras, and configured to transmit at least one transmission synchronization signal for synchronizing at least two cameras among the plurality of cameras.
    Type: Grant
    Filed: September 18, 2020
    Date of Patent: August 31, 2021
    Assignee: Samsung Electronics Co., Ltd.
    Inventors: Il Joong Kim, Jeong A Jo
  • Patent number: 11102523
    Abstract: Systems and methods are disclosed herein for selective audio segment compression for accelerated playback of media assets by service providers. A playback speed of the video segment of a media asset is calculated based on the duration of the video segment and a received playback time period. The system receives audio segments and corresponding priority weights. The audio segments with the lowest priority weight are removed from the group of various audio segments. The system then determines whether the duration of the remaining audio segments exceeds the received playback time period. If so, the system modifies the remaining audio segments by removing another audio segment with the lowest priority weight from the remaining audio segments. The system then rechecks whether the received playback time period is exceeded. If not, the system generates for playback the video segment based on the video playback speed and the remaining audio segments.
    Type: Grant
    Filed: March 19, 2019
    Date of Patent: August 24, 2021
    Assignee: Rovi Guides, Inc.
    Inventors: Neeraj Kumar, Vishwas Sharadanagar Panchaksharaiah, Vikram Makam Gupta
  • Patent number: 10992451
    Abstract: An audio and video playback system includes an audio and video playback device having a local audio device, and a secondary audio device. A method for playing audio data includes: allocating a local audio buffer space and a secondary audio buffer space to the local audio device and the secondary audio device, respectively; processing obtained multimedia data to generate local audio data and secondary audio data; writing the local audio data and the secondary audio data to the local audio buffer space and the secondary audio buffer space, respectively; reading the local audio data and the secondary audio data buffered in the local audio buffer space and the secondary audio buffer space to the local audio device and the secondary audio device, to have the local audio device and the secondary audio device play the local audio data and the secondary audio data, respectively.
    Type: Grant
    Filed: November 5, 2018
    Date of Patent: April 27, 2021
    Assignee: MEDIATEK INC.
    Inventor: Fu Jun Zhu
  • Patent number: 10986154
    Abstract: A method or system configured for receiving a first single data stream representing a first multimedia file, the first single data stream including an interleaved sequence of data elements of a plurality of media, and/or transmitting a second single data stream representing a second multimedia file, the second single data stream including an interleaved sequence of data elements of said plurality of media, where the second multimedia file differs from said first multimedia file by at least one data element of a selected medium extracted from said first multimedia file, and/or by at least one data element of a selected medium added to the first multimedia file, and/or by at least one data element of a selected medium added to the first multimedia file being a converted version of the at least one data element of a selected medium extracted from the first multimedia file.
    Type: Grant
    Filed: May 12, 2017
    Date of Patent: April 20, 2021
    Assignee: GLIDE TALK LTD.
    Inventors: Liron Hertz, Roi Ginat
  • Patent number: 10971121
    Abstract: A system for platform-independent visualization of audio content, in particular audio tracks utilizing a central computer system in communication with user devices via a computer network. The central system utilizes various algorithms to identify spoken content from audio tracks and selects visual assets associated with the identified content. Thereafter, a visualized audio track is available for users to listen and view. Audio tracks, for example Podcasts, may be segmented into topical audio segments based upon themes or topics, with segments from disparate podcasts combined into a single listening experience, based upon certain criteria, e.g., topics, themes, keywords, and the like.
    Type: Grant
    Filed: July 9, 2019
    Date of Patent: April 6, 2021
    Assignee: Tree Goat Media, Inc.
    Inventors: Michael Kakoyiannis, Sherry Mills, Christoforos Lambrou, Vladimir Canic, Srdjan Jovanovic
  • Patent number: 10904692
    Abstract: Embodiments are described for an adaptive audio system that processes audio data comprising a number of independent monophonic audio streams. One or more of the streams has associated with it metadata that specifies whether the stream is a channel-based or object-based stream. Channel-based streams have rendering information encoded by means of channel name; and the object-based streams have location information encoded through location expressions encoded in the associated metadata. A codec packages the independent audio streams into a single serial bitstream that contains all of the audio data. This configuration allows for the sound to be rendered according to an allocentric frame of reference, in which the rendering location of a sound is based on the characteristics of the playback environment (e.g., room size, shape, etc.) to correspond to the mixer's intent.
    Type: Grant
    Filed: November 11, 2019
    Date of Patent: January 26, 2021
    Assignee: Dolby Laboratories Licensing Corporation
    Inventors: Charles Q. Robinson, Nicolas R. Tsingos, Christophe Chabanne
  • Patent number: 10880587
    Abstract: Virtual Reality (VR) processing devices and methods are provided for transmitting user feedback information comprising at least one of user position information and user orientation information, receiving encoded audio-video (A/V) data, which is generated based on the transmitted user feedback information, separating the A/V data into video data and audio data corresponding to a portion of a next frame of a sequence of frames of the video data to be displayed, decoding the portion of a next frame of the video data and the corresponding audio data, providing the audio data for aural presentation and controlling the portion of the next frame of the video data to be displayed in synchronization with the corresponding audio data.
    Type: Grant
    Filed: September 5, 2019
    Date of Patent: December 29, 2020
    Assignees: ATI TECHNOLOGIES ULC, ADVANCED MICRO DEVICES, INC.
    Inventors: Lei Zhang, Gabor Sines, Khaled Mammou, David Glen, Layla A. Mah, Rajabali M. Koduri, Bruce Montag
  • Patent number: 10880659
    Abstract: There is provided a system (100) comprising an audio streaming device (102) having an audio streaming device receiver (104) arranged for receiving a first audio signal (106) comprising a first audio content and a second audio signal (108) comprising a second audio content, the system furthermore comprising a memory device (110) arranged for storing a user defined setting (112), a processor (114) arranged for providing an output audio signal (116), said output audio signal comprising a combination of the first audio content, and the second audio content, wherein the output audio signal comprises a ratio of a level of the first audio content and a level of the second audio content, and the ratio is determined based on the user defined setting (112), and wherein the system is further comprising a system transmitter (118) arranged for wirelessly transmitting the output audio signal (116).
    Type: Grant
    Filed: April 10, 2020
    Date of Patent: December 29, 2020
    Assignee: OTICON A/S
    Inventors: Michael Syskind Pedersen, Povl Koch, David Thorn Blix, Matias Tofteby Bach
  • Patent number: 10848801
    Abstract: A reception side can easily recognize that metadata is inserted into an audio stream. A container of a predetermined format including an audio stream into which metadata is inserted is transmitted. Identification information indicating that the metadata is inserted into the audio stream is inserted into a layer of the container. At the reception side, it is possible to easily recognize that the metadata is inserted into the audio stream and acquire the metadata reliably without waste by performing the process of extracting the metadata inserted into the audio stream based on the recognition.
    Type: Grant
    Filed: April 23, 2019
    Date of Patent: November 24, 2020
    Assignee: Saturn Licensing LLC
    Inventor: Ikuo Tsukagoshi
  • Patent number: 10841359
    Abstract: A media application is disclosed. The media application provides a playback of a media item that includes a video portion and an audio portion. The media application stops the playback of the video portion of the media item while continuing to provide the audio portion of the media item. The media application resumes the playback of the video portion of the media item in synchronization with the audio portion being provided.
    Type: Grant
    Filed: September 30, 2019
    Date of Patent: November 17, 2020
    Assignee: GOOGLE LLC
    Inventors: Oliver John Woodman, Matt Doucleff
  • Patent number: 10805663
    Abstract: Systems, methods, and apparatuses are described for detecting synchronization errors between audio and video signals. Scene changes may be detected based on anchor frames. Offsets between a scene change in a video signal and a reduced audio level or burst of high audio level in the audio signal may indicate a synchronization error.
    Type: Grant
    Filed: July 13, 2018
    Date of Patent: October 13, 2020
    Assignee: Comcast Cable Communications, LLC
    Inventor: Michael Rekstad
  • Patent number: 10805665
    Abstract: A device configured to determine a time on a progress bar and to identify a timestamp in the video timing map based on the time on the progress bar. The device is further configured to identify a source scene identifier corresponding with the identified timestamp and to play a video scene corresponding with the identified source scene identifier. The device is further configured to identify a first animation identifier corresponding with the identified timestamp and to play a first animation associated with the first animation identifier. The device is further configured to determine that the first animation identifier is present in the audio sample buffer, to identify an audio sample associated with the first animation identifier, and to play the identified audio sample.
    Type: Grant
    Filed: December 13, 2019
    Date of Patent: October 13, 2020
    Assignee: Bank of America Corporation
    Inventor: Shankar Sangoli
  • Patent number: 10805658
    Abstract: Provided herein are system, apparatus, article of manufacture, method and/or computer program product embodiments, and/or combinations and sub-combinations thereof, for synchronizing playback of audio and video associated with a content, such as a movie or TV show. Also provided herein are system, apparatus, article of manufacture, method and/or computer program product embodiments, and/or combinations and sub-combinations thereof, for coordinating devices in a whole home entertainment system that includes a wireless network, to improve collective utilization of the wireless network and thereby enhance user experience.
    Type: Grant
    Filed: September 12, 2018
    Date of Patent: October 13, 2020
    Assignee: ROKU, INC.
    Inventors: Ilya Asnis, Anthony John Wood
  • Patent number: 10764620
    Abstract: Systems and methods of the present disclosure provide for dynamic delay equalization of related media signals in a media transport system. Methods include receiving a plurality of related media signals, transporting the related media signals along different media paths, calculating uncorrected propagation delays for the media paths, and delaying each of the related media signals by an amount related to the difference between the longest propagation delay (of the uncorrected propagation delays) and the uncorrected propagation delay of the related media signal/media path. Calculating the uncorrected propagation delays and delaying the related media signals may be performed in response to a change to the propagation delay of at least one of the related media signals/media paths. Additionally or alternatively, calculating the uncorrected propagation delays and delaying the related media signals may be performed while transporting the related media signals.
    Type: Grant
    Filed: August 18, 2019
    Date of Patent: September 1, 2020
    Assignee: Biamp Systems, LLC
    Inventors: Eugene Gurfinkel, Michael K. Davis, Charles H. Van Dusen
  • Patent number: 10748390
    Abstract: A method or system that receives input media including at least video data in which a video event within the video data is detected. Related data that is associated with the detected video event is collected and one or more feature parameters are configured based on the collected related data. The type of video event is determining and a set of feature parameters is selected based on the type of video event. A haptic effect is then automatically generated based on the selected set of feature parameters.
    Type: Grant
    Filed: October 12, 2018
    Date of Patent: August 18, 2020
    Assignee: Immersion Corporation
    Inventor: Liwen Wu
  • Patent number: 10694243
    Abstract: Methods and apparatus to identify media based on watermarks across different audio streams and/or different watermarking techniques are disclosed. An example apparatus includes a watermark detector to detect a first watermark embedded in a first audio stream associated with media and to detect a second watermark embedded in a second audio stream associated with the media. The second audio stream is different than the first audio stream. The example apparatus includes a watermark analyzer to compare first media identifying information in the first watermark with second media identifying information in the second watermark. The example apparatus also includes a media detection event controller to associate the first and second watermarks with a media detection event when the first media identifying information is consistent with the second media identifying information. The example apparatus further includes a transmitter to transmit the media detection event to a data collection facility.
    Type: Grant
    Filed: May 31, 2018
    Date of Patent: June 23, 2020
    Assignee: The Nielsen Company (US), LLC
    Inventors: Justin Fahnestock, Ronan Heffernan, Muhammad Amir, Wes Kercher, Scott Barraclough, John Kistenmacher
  • Patent number: 10650862
    Abstract: A system that incorporates teachings of the subject disclosure may include, for example, detecting a first action at a first time during a first presentation of video content of a multimedia stream. The first action is coincident with a visual aspect of an event observable in the video content. A second action is detected at a second time during a second presentation of audio content of an audio stream, wherein the second action is coincident with an audible aspect of the event observable in a the second presentation of the audio content. A time difference is determined between the first time and the second time, wherein the first presentation of the video content and the second presentation of the audio content are synchronized based on the time difference. Other embodiments are disclosed.
    Type: Grant
    Filed: December 9, 2015
    Date of Patent: May 12, 2020
    Assignee: AT&T Intellectual Property I, L.P.
    Inventors: Wayne R. Heinmiller, Carol S. Gruchala, Dianna Tiliks
  • Patent number: 10623789
    Abstract: Methods, systems, and computer programs for measuring quality of multimedia delivery to a client are presented. A method includes operations for embedding video markers in a video stream of a multimedia stream, and embedding audio markers in an audio stream of the multimedia stream. The video stream and the audio stream are then transmitted separately to the client. Further, video markers received at the client are extracted from the transmitted video stream, and audio markers received at the client are extracted from the transmitted audio stream. A measure of the audio-video synchronization quality is obtained by determining a quantifiable time difference between the video stream and the audio stream received at the client, where the quantifiable time difference is calculated based on the extracted video markers and the extracted audio markers.
    Type: Grant
    Filed: May 26, 2017
    Date of Patent: April 14, 2020
    Assignee: VMware, Inc.
    Inventors: Lawrence Andrew Spracklen, Banit Agrawal, Rishi Bidarkar
  • Patent number: 10607479
    Abstract: A remote control framework enables a plurality of target devices to be controlled by a plurality of remote control devices irrespective of bearer types. In a preferred embodiment any target device may also act as a control device and any control device may also act as a target device. The framework also enables any application running on any target device to be controlled by any controller device.
    Type: Grant
    Filed: February 28, 2018
    Date of Patent: March 31, 2020
    Assignee: Conversant Wireless Licensing S.a.r.l.
    Inventors: Sian James, Neal Harris, John Turner, Tim Howes
  • Patent number: 10535371
    Abstract: Techniques are provided for video summarization, based on speaker segmentation and clustering, to identify persons and scenes of interest. A methodology implementing the techniques according to an embodiment includes extracting audio content from a video stream and detecting one or more segments of the audio content that include the voice of a single speaker. The method also includes grouping the one or more detected segments into an audio cluster associated with the single speaker and providing a portion of the audio cluster to a user. The method further includes receiving an indication from the user that the single speaker is a person of interest. Segments of interest are then extracted from the video stream, where each segment of interest is associated with a scene that includes the person of interest. The extracted segments of interest are then combined into a summarization video.
    Type: Grant
    Filed: September 13, 2016
    Date of Patent: January 14, 2020
    Assignee: INTEL CORPORATION
    Inventors: Gokcen Cilingir, Narayan Biswal
  • Patent number: 10499178
    Abstract: There is provided a non-transitory memory storing an executable code, a hardware processor executing the executable code to receive a visualization of a three-dimensional (3D) position for each audio object of a plurality of audio objects in a first mix of an object-based audio of a media content, the visualization corresponding to a timeline of the media content, receive a second mix of the object-based audio of the media content, and play the second mix of the object-based audio of the media content using an audio playback system while displaying the visualization of the 3D position for each of the plurality of audio objects of the first mix of the object-based audio on a display.
    Type: Grant
    Filed: October 14, 2016
    Date of Patent: December 3, 2019
    Assignee: Disney Enterprises, Inc.
    Inventor: Mark Arana
  • Patent number: 10477339
    Abstract: Embodiments are described for an adaptive audio system that processes audio data comprising a number of independent monophonic audio streams. One or more of the streams has associated with it metadata that specifies whether the stream is a channel-based or object-based stream. Channel-based streams have rendering information encoded by means of channel name; and the object-based streams have location information encoded through location expressions encoded in the associated metadata. A codec packages the independent audio streams into a single serial bitstream that contains all of the audio data. This configuration allows for the sound to be rendered according to an allocentric frame of reference, in which the rendering location of a sound is based on the characteristics of the playback environment (e.g., room size, shape, etc.) to correspond to the mixer's intent.
    Type: Grant
    Filed: June 17, 2019
    Date of Patent: November 12, 2019
    Assignee: Dolby Laboratories Licensing Corporation
    Inventors: Charles Q. Robinson, Nicolas R. Tsingos, Christophe Chabanne
  • Patent number: 10469750
    Abstract: A method is provided for embedding motion data of an object collected by an inertial measurement unit that is attached to the object into a video file that includes video frames of the object in motion captured by a video recording device. The video file has a predefined video file format that is configured to include metadata that is storable at predefined time intervals of the video file. The method operates by capturing video frames of an object in motion and simultaneously collecting motion data of the object, storing the captured video frames in the video file, and storing the collected motion data, converting the motion data to the metadata, and inserting the metadata into one or more time intervals of the video file, wherein the metadata in each time interval includes the metadata for a plurality of successive or preceding video frames.
    Type: Grant
    Filed: May 22, 2018
    Date of Patent: November 5, 2019
    Assignee: BioForce Analytics LLC
    Inventors: Eric L. Canfield, Scott J. Soma, Brandon T. Fanti, Vineeth Voruganti, Daniel J. Gao, Aron Sun, Ryan M. LaRue, Saahas S. Yechuri
  • Patent number: 10461872
    Abstract: A networked system is provided for transporting digital media packets, such as audio and video. The network includes network devices interconnected to send and receive packets. Each network device can receive and transmit media signals from media devices. A master clock generates a system time signal that the network devices use, together with a network time protocol to generate a local clock signal synchronized to the system time signal for both rate and offset. The local clock signal governs both the rate and offset of the received or transmitted media signals. The system, which can be implemented using conventional network equipment enables media signals to be transported to meet quality and timing requirements for high quality audio and video reproduction.
    Type: Grant
    Filed: August 31, 2018
    Date of Patent: October 29, 2019
    Assignee: Audinate Pty Limited
    Inventors: Aidan Williams, Varuni Witana
  • Patent number: 10455050
    Abstract: In one approach, a server computer receives a playlist from a first client computer, wherein the playlist identifies a plurality of media assets and includes synchronization information that specifies how to present the plurality of media assets as a synchronized media presentation. The server computer receives a request from the first client computer to share the playlist with a second client computer. The server computer causes the plurality of media assets to be deposited in a client storage accessible to the second client computer. The server computer sends the playlist to the second client computer. The second client computer presents the synchronized media presentation based on the plurality of media assets deposited in the client storage and the synchronization information of the playlist.
    Type: Grant
    Filed: April 24, 2017
    Date of Patent: October 22, 2019
    Assignee: QWIRE INC.
    Inventors: Leigh B. Roberts, Jr., Jonathan Louis Ehrlich, Scott Freiman
  • Patent number: 10425678
    Abstract: A system for providing group performance using a set of client devices. The time frame of reference of individual client devices of the set of client devices can be synchronized to the host time frame of reference. The individual client devices may effectuate the presentation of the content of the group performance predetermined points in time with respect to the host time frame of reference. Audio and/or visual content of the content provided by the individual client devices is based on the real-world location of the individual client devices. The content provided can be monitored to check if the content is presented at the predetermined points in time. The content provided is adjusted such that the content is provided at the more predetermined points in time. The presentation of the content in a synchronized manner results in the group performance.
    Type: Grant
    Filed: August 8, 2018
    Date of Patent: September 24, 2019
    Assignee: Disney Enterprises, Inc.
    Inventors: Taylor Hellam, Mohammad Poswal, Malcolm E. Murdock
  • Patent number: 10390061
    Abstract: Systems and methods of the present disclosure provide for dynamic delay equalization of related media signals in a media transport system. Methods include receiving a plurality of related media signals, transporting the related media signals along different media paths, calculating uncorrected propagation delays for the media paths, and delaying each of the related media signals by an amount related to the difference between the longest propagation delay (of the uncorrected propagation delays) and the uncorrected propagation delay of the related media signal/media path. Calculating the uncorrected propagation delays and delaying the related media signals may be performed in response to a change to the propagation delay of at least one of the related media signals/media paths. Additionally or alternatively, calculating the uncorrected propagation delays and delaying the related media signals may be performed while transporting the related media signals.
    Type: Grant
    Filed: October 30, 2018
    Date of Patent: August 20, 2019
    Assignee: BIAMP SYSTEMS, LLC
    Inventors: Eugene Gurfinkel, Michael K. Davis, Charles H. Van Dusen
  • Patent number: 10390109
    Abstract: An example method comprises receiving, at a first digital device, video data, scanning video content of the video data for visual transitions within the video content between consecutive frames of the video data, each transition indicating significant visual transitions relative to other frames of the video data, timestamping each visual transition and create a first set of temporal video fingerprints, identifying items of metadata to be associated with the video data, identifying a location within the video data using the temporal video fingerprints for the identified items of metadata, generating a metadata index identifying each item of metadata and a location for each item of metadata relative to the video data using at least one of the temporal video fingerprints, and transmitting, at the first digital device, the video data, the first set of temporal video fingerprints, and the metadata index to a different digital device.
    Type: Grant
    Filed: February 20, 2018
    Date of Patent: August 20, 2019
    Assignee: Crystal Computer Corporation
    Inventors: Alan David Young, Dimitri Felixovich Tarassenko, Roger James Franklin
  • Patent number: 10313722
    Abstract: Techniques are described for synchronizing audio content and video content when server-side fragment insertion techniques are used.
    Type: Grant
    Filed: September 27, 2017
    Date of Patent: June 4, 2019
    Assignee: Amazon Technologies, Inc.
    Inventors: Ron Searl, Yongjun Wu
  • Patent number: 10287032
    Abstract: The present application relates to a gimbal configured to mount a photographing device. The gimbal includes a main body, a mounting shaft connected to the main body and a base disposed on the mounting shaft. The base is configured to fasten the photographing device. The gimbal further includes a fixing plate disposed on the base and a circuit board mounted between the fixing plate and the base. The circuit board is provided with a USB input interface and a USB output interface. The USB input interface is configured to form a data transmission connection to the photographing device. The USB output interface is configured to be connected to an external storage device. The present application further relates to an image photographing apparatus including the foregoing gimbal and an unmanned aerial vehicle including the foregoing image photographing apparatus.
    Type: Grant
    Filed: May 4, 2018
    Date of Patent: May 14, 2019
    Assignee: AUTEL ROBOTICS CO., LTD.
    Inventors: Zhengli Zhang, Fazhan Chen
  • Patent number: 10290322
    Abstract: An audio and video synchronizing perceptual model is described that is based on how a person perceives audio and/or video (e.g., how the brain processes sound and/or visual content). The relative emotional impact associated with different audio portions may be employed to determine transition points to facilitate automatic synchronization of audio data to video data to create a production that achieves a particular overall emotional effect on the listener/viewer. Various processing techniques of the perceptual model may utilize perceptual characteristics within the audio portions to determine a transition point for automatic synchronization with video data.
    Type: Grant
    Filed: May 11, 2018
    Date of Patent: May 14, 2019
    Assignee: Adobe Inc.
    Inventor: Peter Merrill
  • Patent number: 10178345
    Abstract: Media device systems and methods synchronize video content with audio content presented by a plurality of wireless audio headsets. In an exemplary embodiment, a first time delay corresponds to a first duration of time between communication of the audio content from the media device and presentation of the audio content by a first wireless audio headset. A second time delay corresponds to a second duration of time between communication of the audio content from the media device and presentation of the audio content by a second wireless audio headset, wherein the first time delay is greater than the second time delay. Video content communicated to a display is delayed by the first time delay. Audio content communicated to the second wireless audio headset is delayed by a time delay difference between the first time delay and the second time delay.
    Type: Grant
    Filed: June 7, 2018
    Date of Patent: January 8, 2019
    Assignee: EchoStar Technologies L.L.C.
    Inventors: Gregory Greene, David Innes
  • Patent number: 10171710
    Abstract: In order to eliminate timing offset between reproduction devices when a content transmitted from a distribution device is received and reproduced by a plurality of reproduction devices, data (SCR) indicating the elapsed time from the start of the content, generated by counting clock pulses, and data (FCR) indicating a frame number generated by counting the number of frames reproduced by a decoder (54) are transmitted by the distribution device, and a clock generation unit (103) in each reproduction device is controlled so that data (STC) indicating the elapsed time and data (FTC) indicating the frame number, which are generated in the same manner by each reproduction device, match the transmitted data (SCR, FCR). Synchronization between reproduction devices can thereby be established even when, in a state in which a content is being reproduced by one reproduction device, another reproduction device subsequently connects to the distribution device.
    Type: Grant
    Filed: March 6, 2013
    Date of Patent: January 1, 2019
    Assignee: MITSUBISHI ELECTRIC CORPORATION
    Inventor: Eiji Matsuo
  • Patent number: 10158906
    Abstract: System and method for flexible video construction, particularly of a personalized video clip which provides instructions to a viewer with regard to health and wellness. An ordered list of video input files is chained together, to create a single output video file using a chosen container. Timestamp values are tracked, to ensure synchronization of multiple joined clips, optionally using adjustments of the audio channel or the video channel. A video construction server utilizes information from multiple sources, to construct the video clip.
    Type: Grant
    Filed: February 12, 2018
    Date of Patent: December 18, 2018
    Assignee: TELESOFIA MEDICAL LTD.
    Inventors: Rami Cohen, Tzvi Rotshtein, Danny Ben Shitrit
  • Patent number: 10142585
    Abstract: System and methods for synchronizing supplemental media content to media content being provided by a different content source. The systems and methods may identify media content from user input, and determine a plurality of associated supplementary media. By monitoring and interpreting a data stream containing information about the primary media content, the systems and methods may interpret indicia of content ceasing and resuming and similarly cease and resume playback of the supplemental media in response, preserving the synchronicity of the multiple content sources.
    Type: Grant
    Filed: April 28, 2015
    Date of Patent: November 27, 2018
    Assignee: Rovi Guides, Inc.
    Inventor: David D. Chung
  • Patent number: 10142679
    Abstract: A content processing apparatus is provided. The content processing apparatus includes output circuitry configured to output a content, communication circuitry configured to communicate with a server, and a processor configured to extract, from the content, first characteristic information and second characteristic information of a different type from the first characteristic information, to control the communication circuitry to transmit the extracted first characteristic information to the server, and in response to receiving a plurality of matching information corresponding to the transmitted first characteristic information, to control the output circuitry to output matching information corresponding to the second characteristic information extracted from the content.
    Type: Grant
    Filed: December 7, 2016
    Date of Patent: November 27, 2018
    Assignee: SAMSUNG ELECTRONICS CO., LTD.
    Inventor: Kwang-hyun Koh