Patents by Inventor Ajay Divakaran

Ajay Divakaran has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20060059120
    Abstract: A method identifies highlight segments in a video including a sequence of frames. Audio objects are detected to identify frames associated with audio events in the video, and visual objects are detected to identify frames associated with visual events. Selected visual objects are matched with an associated audio object to form an audio-visual object only if the selected visual object matches the associated audio object, the audio-visual object identifying a candidate highlight segment. The candidate highlight segments are further refined, using low level features, to eliminate false highlight segments.
    Type: Application
    Filed: August 27, 2004
    Publication date: March 16, 2006
    Inventors: Ziyou Xiong, Regunathan Radhakrishnan, Ajay Divakaran
  • Patent number: 7003038
    Abstract: A method describes activity in a video sequence. The method measures intensity, direction, spatial, and temporal attributes in the video sequence, and the measured attributes are combined in a digital descriptor of the activity of the video sequence.
    Type: Grant
    Filed: August 13, 2002
    Date of Patent: February 21, 2006
    Assignee: Mitsubishi Electric Research Labs., Inc.
    Inventors: Ajay Divakaran, Huifang Sun, Hae-Kwang Kim, Chul-Soo Park, Xinding Sun, Bangalore S. Manjunath, Vinod V. Vasudevan, Manoranjan D. Jesudoss, Ganesh Rattinassababady, Hyundoo Shin
  • Patent number: 7003154
    Abstract: A system and method for temporally processing an input video including input frames. Each frame has an associated frame play time, and the input video has a total input video play time that is a sum of the input frame play times of all of the input frames. Each of the input frames is classified according to a content characteristic of each frames. An output frame play time is allocated to each of the input frames that is based on the classified content characteristic of each of the input frames to generate a plurality of output frames that form an output video.
    Type: Grant
    Filed: November 17, 2000
    Date of Patent: February 21, 2006
    Assignee: Mitsubishi Electric Research Laboratories, Inc.
    Inventors: Kadir A. Peker, Ajay Divakaran, Huifang Sun
  • Publication number: 20050249412
    Abstract: A method detects events in multimedia. Features are extracted from the multimedia. The features are sampled using a sliding window to obtain samples. A context model is constructed for each sample. An affinity matrix is determined from the models and a commutative distance metric between each pair of context models. A second generation eigenvector is determined for the affinity matrix, and the samples are then clustered into events according to the second generation eigenvector.
    Type: Application
    Filed: May 7, 2004
    Publication date: November 10, 2005
    Inventors: Regunathan Radhakrishnan, Ajay Divakaran
  • Publication number: 20050251532
    Abstract: A method detects events in multimedia. Features are extracted from the multimedia. The features are sampled using a sliding window to obtain samples. A context model is constructed for each sample. The context models form a time series. An affinity matrix is determined from the time series models and a commutative distance metric between each pair of context models. A second generalized eigenvector is determined for the affinity matrix, and the samples are then clustered into events according to the second generalized eigenvector.
    Type: Application
    Filed: August 20, 2004
    Publication date: November 10, 2005
    Inventors: Regunathan Radhakrishnan, Isao Otsuka, Ajay Divakaran
  • Patent number: 6956904
    Abstract: A method for summarizing a video first detects audio peaks in a sub-sampled audio signal of the video. Then, motion activity in the video is extracted and filtered. The filtered motion activity is quantized to a continuous stream of digital pulses, one pulse for each frame. If the motion activity is greater than a predetermined threshold the pulse is one, otherwise the pulse is zero. Each quantized pulse is tested with respect to the timing of rising and falling edges. If the pulse meets the condition of the test, then the pulse is selected as a candidate pulse related to an interesting event in the video, otherwise the pulse is discarded. The candidate pulses are correlated, time-wise to the audio peaks, and patterns between the pulses and peaks are examined. The correlation patterns segment the video into uninteresting and interesting portions, which can then be summarized.
    Type: Grant
    Filed: January 15, 2002
    Date of Patent: October 18, 2005
    Assignee: Mitsubishi Electric Research Laboratories, Inc.
    Inventors: Romain Cabasson, Kadir A. Peker, Ajay Divakaran
  • Publication number: 20050198570
    Abstract: A system and method summarizes multimedia stored in a compressed multimedia file partitioned into a sequence of segments, where the content of the multimedia is, for example, video signals, audio signals, text, and binary data. An associated metadata file includes index information and an importance level for each segment. The importance information is continuous over as closed interval. An importance level threshold is selected in the closed interval, and only segments of the multimedia having a particular importance level greater than the importance level threshold are reproduced. The importance level can also be determined for fixed-length windows of multiple segments, or a sliding window. Furthermore, the importance level can be weighted by a factor, such as the audio volume.
    Type: Application
    Filed: January 21, 2005
    Publication date: September 8, 2005
    Inventors: Isao Otsuka, Ajay Divakaran, Masaharu Ogawa, Kazuhiko Nakane
  • Publication number: 20050154973
    Abstract: A system and method summarizes multimedia stored in a compressed multimedia file partitioned into a sequence of segments, where the content of the multimedia is, for example, video signals, audio signals, text, and binary data. An associated metadata file includes index information and an importance level for each segment. The importance information is continuous over as closed interval. An importance level threshold is selected in the closed interval, and only segments of the multimedia having a particular importance level greater than the importance level threshold are reproduced. The importance level can also be determined for fixed-length windows of multiple segments, or a sliding window. Furthermore, the importance level can be weighted by a factor, such as the audio volume.
    Type: Application
    Filed: February 13, 2004
    Publication date: July 14, 2005
    Inventors: Isao Otsuka, Ajay Divakaran, Masaharu Ogawa, Kazuhiko Nakane
  • Publication number: 20050154987
    Abstract: A system and method summarizes multimedia stored in a compressed multimedia file partitioned into a sequence of segments, where the content of the multimedia is, for example, video signals, audio signals, text, and binary data. An associated metadata file includes index information and an importance level for each segment. The importance information is continuous over as closed interval. An importance level threshold is selected in the closed interval, and only segments of the multimedia having a particular importance level greater than the importance level threshold are reproduced.
    Type: Application
    Filed: January 14, 2004
    Publication date: July 14, 2005
    Inventors: Isao Otsuka, Ajay Divakaran, Masaharu Ogawa, Kazuhiko Nakane
  • Publication number: 20050131869
    Abstract: A method learns a structure of a video, in an unsupervised setting, to detect events in the video consistent with the structure. Sets of features are selected from the video. Based on the selected features, a hierarchical statistical model is updated, and an information gain of the hierarchical statistical model is evaluated. Redundant features are then filtered, and the hierarchical statistical model is updated, based on the filtered features. A Bayesian information criteria is applied to each model and feature set pair, which can then be rank ordered according to the criteria to detect the events in the video.
    Type: Application
    Filed: December 12, 2003
    Publication date: June 16, 2005
    Inventors: Lexing Xie, Ajay Divakaran, Shih-Fu Chang
  • Publication number: 20050125223
    Abstract: A method uses probabilistic fusion to detect highlights in videos using both audio and visual information. Specifically, the method uses coupled hidden Markov models (CHMMs). Audio labels are generated using audio classification via Gaussian mixture models (GMMs), and visual labels are generated by quantizing average motion vector magnitudes. Highlights are modeled using discrete-observation CHMMs trained with labeled videos. The CHMMs have better performance than conventional hidden Markov models (HMMs) trained only on audio signals, or only on video frames.
    Type: Application
    Filed: December 5, 2003
    Publication date: June 9, 2005
    Inventors: Ajay Divakaran, Ziyou Xiong, Regunathan Radhakrishnan
  • Patent number: 6865226
    Abstract: A method analyzes a high-level syntax and structure of a continuous compressed video according to a plurality of states. First, a set of hidden Markov models for each of the states is trained with a training video segmented into known states. Then, a set of domain specific features are extracted from a fixed-length sliding window of the continuous compressed video, and a set of maximum likelihoods is determined for each set of domain specific features using the sets of trained hidden Markov models. Finally, dynamic programming is applied to each set of maximum likelihoods to determine a specific state for each fixed-length sliding window of frames of the compressed video.
    Type: Grant
    Filed: December 5, 2001
    Date of Patent: March 8, 2005
    Assignee: Mitsubishi Electric Research Laboratories, Inc.
    Inventors: Lexing Xie, Shih-Fu Chang, Ajay Divakaran, Huifang Sun
  • Publication number: 20050018881
    Abstract: A method plays frames of a video adaptively according to a visual complexity of the video. First a spatial frequency of pixel within frames of the video is measured, as well as a temporal velocity of corresponding pixels between frames of the video. The spatial frequency is multiplied by the temporal velocity to obtain a measure of the visual complexity of the frames of the video. The frames of the video are then played at a speed that corresponds to the visual complexity.
    Type: Application
    Filed: July 10, 2003
    Publication date: January 27, 2005
    Inventors: Kadir Peker, Ajay Divakaran
  • Patent number: 6847680
    Abstract: A method identifies a talking head or principal cast in a compressed video by first segmenting the video into shots. Motion activity descriptors are extracted from each of the shots, and combined into a shot motion activity descriptor. A distance between the shot motion activity descriptor and a template motion activity descriptor is measured. The template motion activity descriptor is obtained from a training video. If the measured distance is less than a predetermined threshold, then the shot is identified as including a talking head.
    Type: Grant
    Filed: December 17, 2001
    Date of Patent: January 25, 2005
    Assignee: Mitsubishi Electric Research Laboratories, Inc.
    Inventors: Ajay Divakaran, Regunathan Radhakrishnan
  • Publication number: 20040268380
    Abstract: A method detects short term, unusual events in a video. First, features are extracted features from the audio and the video portions of the video. Segments of the video are labeled according to the features. A global sliding window is applied to the labeled segments to determine global characteristics over time, while a local sliding window is applied only to the labeled segments of the global sliding window to determine local characteristic over time. The local window is substantially shorter in time than the global window. A distance between the global and local characteristic is measured to determine occurrences of the unusual short time events.
    Type: Application
    Filed: June 30, 2003
    Publication date: December 30, 2004
    Inventors: Ajay Divakaran, Ziyou Xiong, Regunathan Radhakrishnan, Kadir A. Peker, Koji Miyahara
  • Publication number: 20040233987
    Abstract: A method segments a video into objects, without user assistance. An MPEG compressed video is converted to a structure called a pseudo spatial/temporal data using DCT coefficients and motion vectors. The compressed video is first parsed and the pseudo spatial/temporal data are formed. Seeds macro-blocks are identified using, e.g., the DCT coefficients and changes in the motion vector of macro-blocks.
    Type: Application
    Filed: May 21, 2003
    Publication date: November 25, 2004
    Inventors: Fatih M. Porikli, Huifang Sun, Ajay Divakaran
  • Patent number: 6823011
    Abstract: A method detects an unusual event in a video. Motion vectors are extracted from each frame in a video acquired by a camera of a scene. Zero run-length parameters are determined for each frame from the motion vectors. The zero run-length parameters are summed over predetermined time intervals of the video, and a distance is determined between the sum of the zero run-lengths of a current time interval and the sum of the zero run-lengths of a previous time interval. Then, the unusual event is detected if the distance is greater than a predetermined threshold.
    Type: Grant
    Filed: November 19, 2001
    Date of Patent: November 23, 2004
    Assignee: Mitsubishi Electric Research Laboratories, Inc.
    Inventors: Ajay Divakaran, Regunathan Radhakrishnan
  • Patent number: 6813313
    Abstract: A system and method analyzes a compressed video including a sequence of frames. The amount of a dominant feature in each frame of the compressed video is measured. A label is associated with each frame according the measured amount of the dominant feature. Views in the video are identified according to the labels, and the video is segmented into actions according to the views. The video can then be analyzed according to the action to determine significant events in the video.
    Type: Grant
    Filed: April 20, 2001
    Date of Patent: November 2, 2004
    Assignee: Mitsubishi Electric Research Laboratories, Inc.
    Inventors: Peng Xu, Shih-Fu Chang, Ajay Divakaran
  • Publication number: 20040167767
    Abstract: A method extracts highlights from an audio signal of a sporting event. The audio signal can be part of a sports videos. First, sets of features are extracted from the audio signal. The sets of features are classified according to the following classes: applause, cheering, ball hit, music, speech and speech with music. Adjacent sets of identically classified features are grouped. Portions of the audio signal corresponding to groups of features classified as applause or cheering and with a duration greater than a predetermined threshold are selected as highlights.
    Type: Application
    Filed: February 25, 2003
    Publication date: August 26, 2004
    Inventors: Ziyou Xiong, Regunathan Radhakrishnan, Ajay Divakaran
  • Patent number: 6778708
    Abstract: A compressed bit-stream represents a corresponding sequence having intra-coded frames and inter-coded frames. The compressed bit-stream includes bits associated with each of the inter-coded frames representing a displacement from the associated inter-coded frame to a closest matching of the intra-coded frames. A magnitude of the displacement of a first of the inter-coded frames is determined based on the bits in the compressed bit-stream associated with that inter-coded frame. The inter-coded frame is then identified based on the determined displacement magnitude. The inter-coded frame includes macro-blocks. Each macro-block is associated with a respective portion of the inter-coded frame bits which represent the displacement from that macro-block to the closest matching intra-coded frame. The displacement magnitude is an average of the displacement magnitudes of all the macro-blocks associated with the inter-coded frame.
    Type: Grant
    Filed: July 1, 1999
    Date of Patent: August 17, 2004
    Assignee: Mitsubishi Electric Research Laboratories, Inc.
    Inventors: Ajay Divakaran, Huifang Sun