Patents by Inventor Om D. Deshmukh

Om D. Deshmukh has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20170318013
    Abstract: The disclosed embodiments illustrate methods for voice-based user authentication and content evaluation. The method includes receiving a voice input of a user from a user-computing device, wherein the voice input corresponds to a response to a query. The method further includes authenticating the user based on a comparison of a voiceprint of the voice input and a sample voiceprint of the user. Further, the method includes evaluating content of the response of the user based on the authentication and a comparison between text content and a set of pre-defined answers to the query, wherein the text content is determined based on the received voice input.
    Type: Application
    Filed: April 29, 2016
    Publication date: November 2, 2017
    Inventors: Shourya Roy, Kundan Shrivastava, Om D Deshmukh
  • Publication number: 20170300752
    Abstract: A method and a system are provided for creating a summarized multimedia content. The method extracts one or more frames from a plurality of frames in a multimedia content based on a measure of area occupied by a text content in a portion of each of the plurality of frames. The method selects one or more sentences from an audio content associated with the multimedia content based on at least a weight associated with a plurality of words present in the audio content. The method extracts one or more audio segments from the audio content associated with the multimedia content based on one or more parameters associated with the audio content. The method creates the summarized multimedia content based on the one or more frames, the one or more sentences, and the one or more audio segments.
    Type: Application
    Filed: April 18, 2016
    Publication date: October 19, 2017
    Inventors: Arijit Biswas, Harish Arsikere, Pramod Sankar Kompalli, Kuldeep Yadav, Jagadeesh Chandra Bose Rantham Prabhakara, Kovendhan Ponnavaikko, Om D. Deshmukh, Mohana Prasad Sathya Moorthy
  • Patent number: 9785834
    Abstract: According to embodiments illustrated herein, a method and system is provided for indexing a multimedia content. The method includes extracting, by one or more processors, a set of frames from the multimedia content, wherein the set of frames comprises at least one of a human object and an inanimate object. Thereafter, a body language information pertaining to the human object is determined from the set of frames by utilizing one or more image processing techniques. Further, an interaction information is determined from the set of frames. The interaction information is indicative of an action performed by the human object on the inanimate object. Thereafter, the multimedia content is indexed in a content database based at least on the body language information and the interaction information.
    Type: Grant
    Filed: July 14, 2015
    Date of Patent: October 10, 2017
    Assignee: VIDEOKEN, INC.
    Inventors: Arijit Biswas, Harish Arsikere, Kundan Shrivastava, Om D Deshmukh
  • Publication number: 20170287346
    Abstract: Features are extracted from visual and audio modalities of a video to infer the location of figures/tables/equations/graphs/flow-charts determined as video anchor points which are highlighted on the video timeline to enable quick navigation and provide a quick summary of the video. A voice-based mechanism navigates to a point-of-interest in the video. In case of bandwidth-constrained settings, videos are often played at a very low resolution (quality), and often users need to increase video resolution manually to understand content presented in the figures. Using the automatic identification of these aforementioned anchored points, the resolution can be changed dynamically during streaming a video, which will provide a better viewing experience.
    Type: Application
    Filed: April 1, 2016
    Publication date: October 5, 2017
    Applicant: YEN4KEN INC.
    Inventors: Kuldeep Yadav, Arijit Biswas, Ankit Gandhi, Sumit Negi, Om D. Deshmukh
  • Publication number: 20170242849
    Abstract: According to embodiments illustrated herein, a method and a system is provided for extracting one or more content items from content. The method includes determining, by one or more processors, one or more features associated with each of a plurality of content items in the content. Further, determining, by the one or more processors, a score for each of the plurality of content items based on a weight assigned to each of the one or more features associated with each of the plurality of content items. Thereafter, one or more content items are extracted from the plurality of content items based on the determined score to create at least an index of the content.
    Type: Application
    Filed: February 24, 2016
    Publication date: August 24, 2017
    Inventors: Arijit Biswas, Ankit Gandhi, Om D Deshmukh
  • Publication number: 20170228614
    Abstract: According to embodiments illustrated herein there is provided a method for detecting one or more topic transitions in a multimedia content. The method includes identifying, one or more frames from a plurality of frames of the multimedia content based on a comparison between one or more content items in a first frame of the plurality of frames, and the one or more content items in a first set of frames of the plurality of frames. The method further includes determining at least a first score, and a second score for each of the one or more frames. Additionally, the method includes determining a likelihood for each of the one or more frames based at least on the first score, and the second score, wherein the likelihood is indicative of a topic transition among the one or more frames.
    Type: Application
    Filed: February 4, 2016
    Publication date: August 10, 2017
    Inventors: Ankit Gandhi, Arijit Biswas, Om D. Deshmukh
  • Patent number: 9715540
    Abstract: Systems and associated methods configured to provide user-driven audio content navigation for the spoken web are described. Embodiments allow users to skim audio for content that seems to be of relevance to the user, similar to visual skimming of standard web pages, and mark point of interest within the audio. Embodiments provide techniques for navigating audio content while interacting with information systems in a client-server environment, where the client device can be a simple, standard telephone.
    Type: Grant
    Filed: June 24, 2010
    Date of Patent: July 25, 2017
    Assignee: International Business Machines Corporation
    Inventors: Om D. Deshmukh, Nitendra Rajput
  • Patent number: 9710552
    Abstract: Systems and associated methods configured to provide user-driven audio content navigation for the spoken web are described. Embodiments allow users to skim audio for content that seems to be of relevance to the user, similar to visual skimming of standard web pages, and mark point of interest within the audio. Embodiments provide techniques for navigating audio content while interacting with information systems in a client-server environment, where the client device can be a simple, standard telephone.
    Type: Grant
    Filed: August 28, 2012
    Date of Patent: July 18, 2017
    Assignee: International Business Machines Corporation
    Inventors: Nitendra Rajput, Om D. Deshmukh
  • Patent number: 9620147
    Abstract: Methods, systems, and computer program products for identifying one or more utterances that are likely to carry the intent of a speaker are provided herein. A method includes providing a transcript of utterances to a word weight scoring module to perform inverse document frequency based scoring on each word in the transcript, thereby generating a weight for each word; calculating a weight for each utterance in the transcript to generate weighted utterances by summing the weights or each constituent word in each utterance; comparing at least one weighted utterance to pre-existing example utterances carrying the intent of a speaker to determine a relevancy score for the at least one weighted utterance; and generating a ranked order of the at least one weighted utterance from highest to lowest intent relevancy score, wherein the highest intent relevancy score corresponds to the utterance which is most likely to carry intent of the speaker.
    Type: Grant
    Filed: February 3, 2015
    Date of Patent: April 11, 2017
    Assignee: International Business Machines Corporation
    Inventors: Om D. Deshmukh, Sachindra Joshi, Saket Saurabh, Ashish Verma
  • Publication number: 20170063954
    Abstract: A method and a system are provided for segmenting a multimedia content. The method estimates a count of a plurality of multimedia segments in the multimedia content, and a duration of each of the plurality of multimedia segments in the multimedia content. The method determines a cost function associated with a multimedia segment from the plurality of multimedia segments, based on the count of the plurality of multimedia segments, and the duration of each of the plurality of multimedia segments. The method further determines an updated count of the plurality of multimedia segments, and an updated duration of each of the plurality of multimedia segments until the cost function satisfies a pre-defined criteria. Based on the updated count of the plurality of multimedia segments, and the updated duration of each of the plurality of multimedia segments, the method segments the multimedia content into the plurality of multimedia segments.
    Type: Application
    Filed: September 1, 2015
    Publication date: March 2, 2017
    Inventors: Arijit Biswas, Ankit Gandhi, Ranjeet Kumar, Om D. Deshmukh
  • Publication number: 20170017861
    Abstract: The disclosed embodiments relate to a method for content recommendation. The method includes determining, by one or more processors, one or more features of a segment of a first content being accessed during a presentation of the first content on a user-computing device. The segment of the first content is accessed for a predetermined number of times. The method further includes extracting for a feature from the one or more features, a second content based on the feature, wherein the second content is recommended through the user-computing device.
    Type: Application
    Filed: July 17, 2015
    Publication date: January 19, 2017
    Inventors: Sonal S. Patil, Kundan Shrivastava, Om D. Deshmukh
  • Publication number: 20170017838
    Abstract: According to embodiments illustrated herein, a method and system is provided for indexing a multimedia content. The method includes extracting, by one or more processors, a set of frames from the multimedia content, wherein the set of frames comprises at least one of a human object and an inanimate object. Thereafter, a body language information pertaining to the human object is determined from the set of frames by utilizing one or more image processing techniques. Further, an interaction information is determined from the set of frames. The interaction information is indicative of an action performed by the human object on the inanimate object. Thereafter, the multimedia content is indexed in a content database based at least on the body language information and the interaction information.
    Type: Application
    Filed: July 14, 2015
    Publication date: January 19, 2017
    Inventors: Arijit Biswas, Harish Arsikere, Kundan Shrivastava, Om D. Deshmukh
  • Patent number: 9484032
    Abstract: The disclosed embodiments illustrate methods and systems for processing multimedia content. The method includes extracting one or more words from an audio stream associated with multimedia content. Each word has associated one or more timestamps indicative of temporal occurrences of said word in said multimedia content. The method further includes creating a word cloud of said one or more words in said multimedia content based on a measure of emphasis laid on each word in said multimedia content and said one or more timestamps associated with said one or more words. The method further includes presenting one or more multimedia snippets, of said multimedia content, associated with a word selected by a user from said word cloud. Each of said one or more multimedia snippets corresponds to said one or more timestamps associated with occurrences of said word in said multimedia content.
    Type: Grant
    Filed: October 27, 2014
    Date of Patent: November 1, 2016
    Assignee: Xerox Corporation
    Inventors: Kuldeep Yadav, Kundan Shrivastava, Om D Deshmukh
  • Publication number: 20160307563
    Abstract: The disclosed embodiments illustrate methods and systems for detecting plagiarism in a conversation. The method includes receiving first input corresponding to a query from a first user in said conversation. The first input corresponds to at least a first audio signal received from said first user. The method includes receiving second input corresponding to one or more responses received from a second user in response to said query. The second input corresponds to at least a second audio signal received from said second user. Thereafter, the method includes determining a first score for one or more websites, based on a comparison between said one or more responses and content obtained from said one or more websites in response to said query. The first score is a measure of a similarity between said one or more responses and said content. The method is performed by one or more microprocessors.
    Type: Application
    Filed: April 15, 2015
    Publication date: October 20, 2016
    Inventors: Kundan Shrivastava, Om D. Deshmukh, Geetha Manjunath
  • Publication number: 20160225273
    Abstract: According to embodiments illustrated herein, there is provided a method and a mobile device for evaluating a user on a question. The method includes monitoring one or more inputs of the user, pertaining to at least one step performed for solving the question, by one or more sensors in a mobile device. While the user attempts the question, the one or more sensors also monitor one or more facial expressions of the user, which are analyzed using one or more image processing techniques. Thereafter, one or more processors in the mobile device determine if the at least one step is in accordance with a predefined set of rules deterministic of one or more correct steps involved in solving the question. Further, the one or more processors evaluate the user on the question based on the determination and the analysis of the one or more facial expressions.
    Type: Application
    Filed: January 29, 2015
    Publication date: August 4, 2016
    Inventors: Kaushik Baruah, Kuldeep Yadav, Om D. Deshmukh, William K. Stumbo
  • Publication number: 20160118060
    Abstract: The disclosed embodiments illustrate methods and systems for processing multimedia content. The method includes extracting one or more words from an audio stream associated with multimedia content. Each word has associated one or more timestamps indicative of temporal occurrences of said word in said multimedia content. The method further includes creating a word cloud of said one or more words in said multimedia content based on a measure of emphasis laid on each word in said multimedia content and said one or more timestamps associated with said one or more words. The method further includes presenting one or more multimedia snippets, of said multimedia content, associated with a word selected by a user from said word cloud. Each of said one or more multimedia snippets corresponds to said one or more timestamps associated with occurrences of said word in said multimedia content.
    Type: Application
    Filed: October 27, 2014
    Publication date: April 28, 2016
    Inventors: Kuldeep Yadav, Kundan Shrivastava, Om D. Deshmukh
  • Patent number: 9270825
    Abstract: Embodiments of a system are disclosed for stress assessment of a call center agent while interacting with a customer. The system is for use with a communication network. The system includes a stress assessment device and an agent device that includes an imaging unit. The agent device is configured to capture video of a target region of exposed skin of the agent using the imaging unit, collect customer interaction data based on interaction with a customer device over the communication network, and communicate the captured video and the customer interaction data to the stress assessment device. The stress assessment device is configured to passively estimate agent stress-level based on the received video, and generate feedback to the agent based on correlation between the customer interaction data and the estimated stress-level over a predefined time interval.
    Type: Grant
    Filed: March 18, 2014
    Date of Patent: February 23, 2016
    Assignee: XEROX CORPORATION
    Inventors: Om D Deshmukh, Nischal Murthy Piratla, Lalit K Mestha
  • Publication number: 20150379654
    Abstract: An embodiment is directed to an attendance management system having a client device and a server device. The client device includes a first transceiving module for receiving one or more parameters from a host; a display module for displaying one or more modes of interaction for selection to the host, the modes of interaction being an audio interaction mode and a visual interaction mode; and an attendance module for capturing and assigning an attendance to one or more members based on a selection of the at least one of the modes of interaction. The server device includes a second transceiving module for receiving the assigned attendance, and a database for storing and maintaining the assigned attendance and other information. The client device and the server device are configured to communicate with each other over a network.
    Type: Application
    Filed: June 26, 2014
    Publication date: December 31, 2015
    Inventors: Om D. Deshmukh, Rinku Gajera
  • Publication number: 20150287339
    Abstract: The disclosed embodiments illustrate methods and systems for imparting a spoken language training. The method includes performing a spoken language evaluation of a speech input received from a user on a first training content. Thereafter, the user is categorized based on the spoken language evaluation and a profile of the user. Further, a second training content, comprising one or more tasks, is transmitted to the user based on the categorization and the spoken language evaluation. The user interacts with another user belonging to at least the user group, by comparing a temporal progression of the user with the other user on the one or more tasks, challenging the other user on a task from the one or more tasks, and selecting the task from the one or more tasks based on a difficulty level assessed by the other user.
    Type: Application
    Filed: April 4, 2014
    Publication date: October 8, 2015
    Applicant: Xerox Corporation
    Inventors: Om D. Deshmukh, Kuldeep Yadav, Shourya Roy
  • Publication number: 20150271329
    Abstract: Embodiments of a system are disclosed for stress assessment of a call center agent while interacting with a customer. The system is for use with a communication network. The system includes a stress assessment device and an agent device that includes an imaging unit. The agent device is configured to capture video of a target region of exposed skin of the agent using the imaging unit, collect customer interaction data based on interaction with a customer device over the communication network, and communicate the captured video and the customer interaction data to the stress assessment device. The stress assessment device is configured to passively estimate agent stress-level based on the received video, and generate feedback to the agent based on correlation between the customer interaction data and the estimated stress-level over a predefined time interval.
    Type: Application
    Filed: March 18, 2014
    Publication date: September 24, 2015
    Applicant: Xerox Corporation
    Inventors: Om D. Deshmukh, Nischal Murthy Piratla, Lalit K. Mestha