Patents by Inventor Ajay Divakaran

Ajay Divakaran has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20040143434
    Abstract: A method segments and summarizes a news video using both audio and visual features extracted from the video. The summaries can be used to quickly browse the video to locate topics of interest. A generalized sound recognition hidden Markov model (HMM) framework for joint segmentation and classification of the audio signal of the news video is used. The HMM not only provides a classification label for audio segment, but also compact state duration histogram descriptors.
    Type: Application
    Filed: January 17, 2003
    Publication date: July 22, 2004
    Inventors: Ajay Divakaran, Regunathan Radhakrishnan
  • Patent number: 6763069
    Abstract: A method extracts high-level features from a video including a sequence of frames. Low-level features are extracted from each frame of the video. Each frame of the video is labeled according to the extracted low-level features to generate sequences of labels. Each sequence of labels is associated with one of the extracted low-level feature. The sequences of labels are analyzed using learning machine learning techniques to extract high-level features of the video.
    Type: Grant
    Filed: July 6, 2000
    Date of Patent: July 13, 2004
    Assignee: Mitsubishi Electric Research Laboratories, Inc
    Inventors: Ajay Divakaran, Anthony Vetro, Huifang Sun, Peng Xu, Shih-Fu Chang
  • Publication number: 20040086180
    Abstract: A method discovers patterns in unknown content of a video. The video is partitioned into sets of disjoint segments. Each set includes all frames of the video, and each set is partitioned according to a selected low-level feature of the video. The disjoint segments are grouped into corresponding sets of clusters, each cluster including similar segments. The clusters are then labeled, and association rules are identified among the labels to discover high-level patterns in the unknown content of the video.
    Type: Application
    Filed: November 1, 2002
    Publication date: May 6, 2004
    Inventor: Ajay Divakaran
  • Publication number: 20040085339
    Abstract: A method summarizes unknown content of a video. First, low-level features of the video are selected. The video is then partitioned into segments according to the low-level features. The segments are grouped into disjoint clusters where each cluster contains similar segments. The clusters are labeled according to the low-level features, and parameters characterizing the clusters are assigned. High-level patterns among the labels are found, and the these patterns are used to extract frames from the video according to form a content-adaptive summary of the unknown content of the video.
    Type: Application
    Filed: November 1, 2002
    Publication date: May 6, 2004
    Inventors: Ajay Divakaran, Kadir A. Peker
  • Publication number: 20040085323
    Abstract: A method mines unknown content of a video by first selecting one or more low-level features of the video. For each selected feature, or combination of features, time series data is generated. The time series data is then self-correlated to identify similar segments of the video according to the low-level features. The similar segments are grouped into clusters to discover high-level patterns in the unknown content of video.
    Type: Application
    Filed: November 1, 2002
    Publication date: May 6, 2004
    Inventors: Ajay Divakaran, Kadir A. Peker
  • Patent number: 6697523
    Abstract: A method extracts an intensity of motion activity from shots in a compressed video. The method then uses the intensity of motion activity to segment the video into easy and difficult segments to summarize. Easy to summarize segments are represented by any frames selected from the easy to summarize segments, while a color based summarization process extracts generates sequences of frames from each difficult to summarize segment. The selected and generated frames of each segment in each shot are combined to form the summary of the compressed video.
    Type: Grant
    Filed: August 9, 2000
    Date of Patent: February 24, 2004
    Assignee: Mitsubishi Electric Research Laboratories, Inc.
    Inventors: Ajay Divakaran, Kadir A. Peker, Huifang Sun
  • Publication number: 20040008789
    Abstract: A method segments a compressed video by extracting audio and visual features from the compressed video. The audio features are clustered according to K-means clustering in a set of classes, and the compressed video is then partitioned into first segments according to the set of classes. The visual features are then used to partitioning each first segment into second segments using motion analysis. Summaries of the second segments can be provided to assist in the browsing of the compressed video.
    Type: Application
    Filed: July 10, 2002
    Publication date: January 15, 2004
    Inventors: Ajay Divakaran, Regunathan Radhakrishnan, Michael A. Casey
  • Patent number: 6646676
    Abstract: A surveillance and control system includes a feature extraction unit to dynamically extract low-level features from a compressed digital video signal, a description encoder, coupled to the feature extraction unit, to encode the low-level features as content descriptors. An event detector is coupled to the description encoder to detect security events from the content descriptors, and a control signal processor, coupled to the event detector, to generate control signals in response to detecting the security events.
    Type: Grant
    Filed: July 10, 2000
    Date of Patent: November 11, 2003
    Assignee: Mitsubishi Electric Research Laboratories, Inc.
    Inventors: Paul DaGraca, Anthony Vetro, Ajay Divakaran, Huifang Sun
  • Patent number: 6618507
    Abstract: This invention relates to methods of feature extraction from MPEG-2 and MPEG-4 compressed video sequences. The spatio-temporal compression complexity of video sequences is evaluated for feature extraction by inspecting the compressed bitstream and the complexity is used as a descriptor of the spatio-temporal characteristics of the video sequence. The spatio-temporal compression complexity measure is used as a matching criterion and can also be used for absolute indexing. Feature extraction can be accomplished in conjunction with scene change detection techniques and the combination has reasonable accuracy and the advantage of high simplicity since it is based on entropy decoding of signals in compressed form and does not require computationally expensive inverse Discrete Cosine Transformation (DCT).
    Type: Grant
    Filed: January 25, 1999
    Date of Patent: September 9, 2003
    Assignee: Mitsubishi Electric Research Laboratories, Inc
    Inventors: Ajay Divakaran, Huifang Sun, Hiroshi Ito
  • Patent number: 6600784
    Abstract: A method describes motion activity in a video sequence. A motion activity matrix is determined for the video sequence. A threshold for the motion activity matrix is determined. Connected regions of motion vectors at least equal to the threshold are identified and measured for size. A histogram of the distribution of the sizes of the connected areas is constructed for the entire video sequence. The histogram is normalized to characterize the spatial distribution of the video sequence in a motion activity descriptor.
    Type: Grant
    Filed: February 2, 2000
    Date of Patent: July 29, 2003
    Assignee: Mitsubishi Electric Research Laboratories, Inc.
    Inventors: Ajay Divakaran, Kadir A. Peker, Huifang Sun
  • Publication number: 20030133511
    Abstract: A method for summarizing a video first detects audio peaks in a sub-sampled audio signal of the video. Then, motion activity in the video is extracted and filtered. The filtered motion activity is quantized to a continuous stream of digital pulses, one pulse for each frame. If the motion activity is greater than a predetermined threshold the pulse is one, otherwise the pulse is zero. Each quantized pulse is tested with respect to the timing of rising and falling edges. If the pulse meets the condition of the test, then the pulse is selected as a candidate pulse related to an interesting event in the video, otherwise the pulse is discarded. The candidate pulses are correlated, time-wise to the audio peaks, and patterns between the pulses and peaks are examined. The correlation patterns segment the video into uninteresting and interesting portions, which can then be summarized.
    Type: Application
    Filed: January 15, 2002
    Publication date: July 17, 2003
    Inventors: Romain Cabasson, Kadir A. Peker, Ajay Divakaran
  • Publication number: 20030112865
    Abstract: A method identifies a talking head or principal cast in a compressed video by first segmenting the video into shots. Motion activity descriptors are extracted from each of the shots, and combined into a shot motion activity descriptor. A distance between the shot motion activity descriptor and a template motion activity descriptor is measured. The template motion activity descriptor is obtained from a training video. If the measured distance is less than a predetermined threshold, then the shot is identified as including a talking head.
    Type: Application
    Filed: December 17, 2001
    Publication date: June 19, 2003
    Inventors: Ajay Divakaran, Regunathan Radhakrishnan
  • Publication number: 20030103565
    Abstract: A method analyzes a high-level syntax and structure of a continuous compressed video according to a plurality of states. First, a set of hidden Markov models for each of the states is trained with a training video segmented into known states. Then, a set of domain specific features are extracted from a fixed-length sliding window of the continuous compressed video, and a set of maximum likelihoods is determined for each set of domain specific features using the sets of trained hidden Markov models. Finally, dynamic programming is applied to each set of maximum likelihoods to determine a specific state for each fixed-length sliding window of frames of the compressed video.
    Type: Application
    Filed: December 5, 2001
    Publication date: June 5, 2003
    Inventors: Lexing Xie, Shih-Fu Chang, Ajay Divakaran, Huifang Sun
  • Patent number: 6574279
    Abstract: A method for transcoding a compressed video partitions the compressed video into hierarchical levels, and extracts features from each of the hierarchical levels. One of a number of conversion modes of a transcoder is selected dependent on the features extracted from the hierarchical levels. The compressed video is then transcoded according to the selected conversion mode.
    Type: Grant
    Filed: April 11, 2000
    Date of Patent: June 3, 2003
    Assignee: Mitsubishi Electric Research Laboratories, Inc.
    Inventors: Anthony Vetro, Ajay Divakaran, Huifang Sun
  • Publication number: 20030095602
    Abstract: A method detects an unusual event in a video. Motion vectors are extracted from each frame in a video acquired by a camera of a scene. Zero run-length parameters are determined for each frame from the motion vectors. The zero run-length parameters are summed over predetermined time intervals of the video, and a distance is determined between the sum of the zero run-lengths of a current time interval and the sum of the zero run-lengths of a previous time interval. Then, the unusual event is detected if the distance is greater than a predetermined threshold.
    Type: Application
    Filed: November 19, 2001
    Publication date: May 22, 2003
    Inventors: Ajay Divakaran, Regunathan Radhakrishnan
  • Patent number: 6546135
    Abstract: A method for generating a representation of multimedia content by first segmenting the multimedia content spatially and temporally to extract objects. Feature extraction is applied to the objects to produce semantic and syntactic attributes, relations, and a containment set of content entities. The content entities are coded to produce directed acyclic graphs of the content entities, where each directed acyclic graph represents a particular interpretation of the multimedia content.
    Type: Grant
    Filed: August 30, 1999
    Date of Patent: April 8, 2003
    Assignee: Mitsubishi Electric Research Laboratories, INC
    Inventors: I-Jong Lin, Anthony Vetro, Ajay Divakaran, Sun-Yuan Kung
  • Patent number: 6542546
    Abstract: A multi-media delivery system for delivering a compressed bitstream through a network to a user device includes a transcoder and a manager. The transcoder is configured to operate on the bit stream using in any one of a plurality of conversion modes. The manager is configured to selecting a particular one of the plurality of conversion modes dependent on semantic content of the bitstream and network characteristics. The system also includes a content classifier to determine the content characteristics, and a model predicator to determine the network characteristics, and user device characteristics. An integrator of the manager generates an optimal rate-quality function to be used for selecting the particular conversion model for a given available bit rate of the network.
    Type: Grant
    Filed: February 2, 2000
    Date of Patent: April 1, 2003
    Assignee: Mitsubishi Electric Research Laboratories, Inc.
    Inventors: Anthony Vetro, Ajay Divakaran, Huifang Sun
  • Publication number: 20030026340
    Abstract: A method describes activity in a video sequence. The method measures intensity, direction, spatial, and temporal attributes in the video sequence, and the measured attributes are combined in a digital descriptor of the activity of the video sequence.
    Type: Application
    Filed: August 13, 2002
    Publication date: February 6, 2003
    Inventors: Ajay Divakaran, Huifang Sun, Hae-Kwang Kim, Chul-Soo Park, Xinding Sun, Bangalore S. Manjunath, Vinod V. Vasudevan, Manoranjan D. Jesudoss, Ganesh Rattinassababady, Hyundoo Shin
  • Publication number: 20030007555
    Abstract: A method measures an intensity of motion activity in a compressed video. The intensity of the motion activity is used to partition the video into segments of equal cumulative motion activity. Key-frames are then selected from each segments. The selected key-frames are concatenated in temporal order to form a summary of the video.
    Type: Application
    Filed: April 27, 2001
    Publication date: January 9, 2003
    Applicant: Mitsubishi Electric Research Laboratories, Inc.
    Inventors: Ajay Divakaran, Regunathan Radhakrishnan, Kadir A. Peker
  • Patent number: 6490320
    Abstract: In an apparatus for transcoding a compressed video, a generator simulates constraints of a network and constraints of a user device. A classifier is coupled to receive an input compressed video and the constraints. The classifier generates content information from features of the input compressed video. A manager produces a plurality of conversions modes dependent the constraints and content information, and a transcoder produces output compressed videos, one for each of the plurality conversion modes.
    Type: Grant
    Filed: April 11, 2000
    Date of Patent: December 3, 2002
    Assignee: Mitsubishi Electric Research Laboratories Inc.
    Inventors: Anthony Vetro, Ajay Divakaran, Huifang Sun