Patents by Inventor Anirudh Koul

Anirudh Koul has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 11935294
    Abstract: This disclosure describes how to identify objects in an augmented reality environment. More specifically, the various systems and methods described herein describe how an augmented reality device can recognize objects within a real world environment, determine where the object is located, and also identify the various surfaces of the object in real time or substantially real time.
    Type: Grant
    Filed: September 20, 2022
    Date of Patent: March 19, 2024
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Anirudh Koul, Baihan Huang, Sambodhi Chakrabarty
  • Patent number: 11699360
    Abstract: Aspects of the present disclosure relate to systems and methods that aid users with hearing and/or speech impediments to have a conversation with a remote phone number without human assistance. In one aspect, an application residing on a user's device, such as a smartphone, tablet computer, laptop, etc., may be used to initiate a phone call to a recipient. Upon initiating the phone call locally, a service residing on the server may receive a request to initiate a connection to the recipient. Once the recipient answers, the user may converse with the recipient by providing text input to their local app. The text input may be transmitted to the service. The service may use a text to speech converter to translate the received text to speech that can be delivered to the recipient.
    Type: Grant
    Filed: November 30, 2020
    Date of Patent: July 11, 2023
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Anirudh Koul, Meher Kasam, Michael Johnston, Swetha Machanavajhala, Elizabeth Halper
  • Publication number: 20230020093
    Abstract: This disclosure describes how to identify objects in an augmented reality environment. More specifically, the various systems and methods described herein describe how an augmented reality device can recognize objects within a real world environment, determine where the object is located, and also identify the various surfaces of the object in real time or substantially real time.
    Type: Application
    Filed: September 20, 2022
    Publication date: January 19, 2023
    Applicant: Microsoft Technology Licensing, LLC
    Inventors: Anirudh KOUL, Baihan HUANG, Sambodhi CHAKRABARTY
  • Patent number: 11475662
    Abstract: This disclosure describes how to identify objects in an augmented reality environment. More specifically, the various systems and methods described herein describe how an augmented reality device can recognize objects within a real world environment, determine where the object is located, and also identify the various surfaces of the object in real time or substantially real time.
    Type: Grant
    Filed: September 23, 2020
    Date of Patent: October 18, 2022
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Anirudh Koul, Baihan Huang, Sambodhi Chakrabarty
  • Publication number: 20210082315
    Abstract: Aspects of the present disclosure relate to systems and methods that aid users with hearing and/or speech impediments to have a conversation with a remote phone number without human assistance. In one aspect, an application residing on a user's device, such as a smartphone, tablet computer, laptop, etc., may be used to initiate a phone call to a recipient. Upon initiating the phone call locally, a service residing on the server may receive a request to initiate a connection to the recipient. Once the recipient answers, the user may converse with the recipient by providing text input to their local app. The text input may be transmitted to the service. The service may use a text to speech converter to translate the received text to speech that can be delivered to the recipient.
    Type: Application
    Filed: November 30, 2020
    Publication date: March 18, 2021
    Applicant: Microsoft Technology Licensing, LLC
    Inventors: Anirudh KOUL, Meher KASAM, Michael JOHNSTON, Swetha MACHANAVAJHALA, Elizabeth HALPER
  • Publication number: 20210004599
    Abstract: This disclosure describes how to identify objects in an augmented reality environment. More specifically, the various systems and methods described herein describe how an augmented reality device can recognize objects within a real world environment, determine where the object is located, and also identify the various surfaces of the object in real time or substantially real time.
    Type: Application
    Filed: September 23, 2020
    Publication date: January 7, 2021
    Applicant: Microsoft Technology Licensing, LLC
    Inventors: Anirudh Koul, Baihan Huang, Sambodhi Chakrabarty
  • Patent number: 10854110
    Abstract: Aspects of the present disclosure relate to systems and methods that aid users with hearing and/or speech impediments to have a conversation with a remote phone number without human assistance. In one aspect, an application residing on a user's device, such as a smartphone, tablet computer, laptop, etc., may be used to initiate a phone call to a recipient. Upon initiating the phone call locally, a service residing on the server may receive a request to initiate a connection to the recipient. Once the recipient answers, the user may converse with the recipient by providing text input to their local app. The text input may be transmitted to the service. The service may use a text to speech converter to translate the received text to speech that can be delivered to the recipient.
    Type: Grant
    Filed: October 11, 2017
    Date of Patent: December 1, 2020
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Anirudh Koul, Meher Kasam, Michael Johnston, Swetha Machanavajhala, Elizabeth Halper
  • Patent number: 10817724
    Abstract: This disclosure describes how to identify objects in an augmented reality environment. More specifically, the various systems and methods described herein describe how an augmented reality device can recognize objects within a real world environment, determine where the object is located, and also identify the various surfaces of the object in real time or substantially real time.
    Type: Grant
    Filed: May 1, 2017
    Date of Patent: October 27, 2020
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Anirudh Koul, Baihan Huang, Sambodhi Chakrabarty
  • Patent number: 10785510
    Abstract: Architecture that enables the identification of entities such as people and content in live broadcasts (e.g., streaming content (e.g., video) of live events) and non-live presentations (e.g., movies), in realtime, using recognition processes. This can be accomplished by extracting live data related to a live event. With respect to people entities, filtering can be performed to identify the named (people) entities from the extracted live data, and trending topics discovered as relate to the named entities, as associated with the live event. Multiple images of the named entities that capture the named entities under different conditions are captured for the named entities. The images are then processed to extract and learn facial features (train one or more models), and facial recognition is then performed on faces in the video using the trained model(s).
    Type: Grant
    Filed: November 6, 2018
    Date of Patent: September 22, 2020
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Anirudh Koul, Serge-Eric Tremblay
  • Patent number: 10579330
    Abstract: Users' interaction performance with an automated voice system is improved, as is users' efficiency, by visually displaying options audibly presented by the automated voice system, thereby enabling users to interact with the system more quickly and accurately. Options can be obtained from a centralized audible menu database with the communicational identifier utilized to establish a communication connection with the automated voice system. The database is populated from crowdsourced information, provided when users establish communicational connections with portions of automated voice systems whose options have not yet been stored in the database, and then transcribe the options that are audibly presented by the automated voice system. Transcription of audibly presented options likewise serves as a double check to verify options already displayed.
    Type: Grant
    Filed: May 13, 2015
    Date of Patent: March 3, 2020
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Anirudh Koul, Vinay Vyas Vemuri
  • Publication number: 20190296933
    Abstract: A technique is described herein for facilitating the programming and control of a collection of devices. In one manner of operation, the technique involves: receiving signals from the collection of devices that describe a sequence of events that have occurred in operation of the collection of devices; storing the signals; determining a rule associated with the sequence of events using a machine-trained sequence-detection component (SDC), the rule identifying a next event in the sequence of events; determining whether the rule is viable; and, if the rule is determined to be viable, sending control information to at least one device in the collection of devices. The control information instructs the identified device(s) to perform the next event that has been identified.
    Type: Application
    Filed: March 20, 2018
    Publication date: September 26, 2019
    Inventors: Anirudh KOUL, Ranjitha GURUNATH KULKARNI
  • Patent number: 10313845
    Abstract: Non-limiting examples of the present disclosure describe proactive speech detection on behalf of a user and alerting the user when a specific word, name, etc. is detected. Speech detection is actively executed through a computing device, where the speech detection analyzes spoken utterances in association with a dynamic grammar file stored locally on the computing device. An alert is generated that indicates when a spoken word of the utterances matches a word stored in the dynamic grammar file. The alert may be displayed, for example, through the computing device. The alert provides indication that the spoken word is identified in the spoken utterances. In further examples, a buffered window of the spoken utterances is captured which is associated with a detection of the specific word. A live transcription of the content in the buffered window is generated and provided to a computing device of the user.
    Type: Grant
    Filed: June 6, 2017
    Date of Patent: June 4, 2019
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Anirudh Koul, Venkata Naga Vijaya Swetha Machanavajhala, Stéphane Morichère-Matte, David Dai Wei Zhang, Anass Al-Wohoush, Jan Ervin Mickel Takata Clarin, Sheng-Ting Lin, Sitthinon Jinaphant, Shery Sharonjit Sumal
  • Publication number: 20190075341
    Abstract: Architecture that enables the identification of entities such as people and content in live broadcasts (e.g., streaming content (e.g., video) of live events) and non-live presentations (e.g., movies), in realtime, using recognition processes. This can be accomplished by extracting live data related to a live event. With respect to people entities, filtering can be performed to identify the named (people) entities from the extracted live data, and trending topics discovered as relate to the named entities, as associated with the live event. Multiple images of the named entities that capture the named entities under different conditions are captured for the named entities. The images are then processed to extract and learn facial features (train one or more models), and facial recognition is then performed on faces in the video using the trained model(s).
    Type: Application
    Filed: November 6, 2018
    Publication date: March 7, 2019
    Applicant: Microsoft Technology Licensing, LLC
    Inventors: Anirudh Koul, Serge-Eric Tremblay
  • Patent number: 10165307
    Abstract: Architecture that enables the identification of entities such as people and content in live broadcasts (e.g., streaming content (e.g., video) of live events) and non-live presentations (e.g., movies), in realtime, using recognition processes. This can be accomplished by extracting live data related to a live event. With respect to people entities, filtering can be performed to identify the named (people) entities from the extracted live data, and trending topics discovered as relate to the named entities, as associated with the live event. Multiple images of the named entities that capture the named entities under different conditions are captured for the named entities. The images are then processed to extract and learn facial features (train one or more models), and facial recognition is then performed on faces in the video using the trained model(s).
    Type: Grant
    Filed: April 19, 2017
    Date of Patent: December 25, 2018
    Assignee: MICROSOFT TECHNOLOGY LICENSING, LLC
    Inventors: Anirudh Koul, Serge-Eric Tremblay
  • Publication number: 20180352390
    Abstract: Non-limiting examples of the present disclosure describe proactive speech detection on behalf of a user and alerting the user when a specific word, name, etc. is detected. Speech detection is actively executed through a computing device, where the speech detection analyzes spoken utterances in association with a dynamic grammar file stored locally on the computing device. An alert is generated that indicates when a spoken word of the utterances matches a word stored in the dynamic grammar file. The alert may be displayed, for example, through the computing device. The alert provides indication that the spoken word is identified in the spoken utterances. In further examples, a buffered window of the spoken utterances is captured which is associated with a detection of the specific word. A live transcription of the content in the buffered window is generated and provided to a computing device of the user.
    Type: Application
    Filed: June 6, 2017
    Publication date: December 6, 2018
    Inventors: Anirudh Koul, Venkata Naga Vijaya Swetha Machanavajhala, Stéphane Morichère-Matte, David Dai Wei Zhang, Anass Al-Wohoush, Jan Ervin Mickel Takata Clarin, Sheng-Ting Lin, Sitthinon Jinaphant, Shery Sharonjit Sumal
  • Publication number: 20180314888
    Abstract: This disclosure describes how to identify objects in an augmented reality environment. More specifically, the various systems and methods described herein describe how an augmented reality device can recognize objects within a real world environment, determine where the object is located, and also identify the various surfaces of the object in real time or substantially real time.
    Type: Application
    Filed: May 1, 2017
    Publication date: November 1, 2018
    Applicant: Microsoft Technology Licensing, LLC
    Inventors: Anirudh Koul, Baihan Huang, Sambodhi Chakrabarty
  • Publication number: 20180253992
    Abstract: Aspects of the present disclosure relate to systems and methods that aid users with hearing and/or speech impediments to have a conversation with a remote phone number without human assistance. In one aspect, an application residing on a user's device, such as a smartphone, tablet computer, laptop, etc., may be used to initiate a phone call to a recipient. Upon initiating the phone call locally, a service residing on the server may receive a request to initiate a connection to the recipient. Once the recipient answers, the user may converse with the recipient by providing text input to their local app. The text input may be transmitted to the service. The service may use a text to speech converter to translate the received text to speech that can be delivered to the recipient.
    Type: Application
    Filed: October 11, 2017
    Publication date: September 6, 2018
    Applicant: Microsoft Technology Licensing, LLC
    Inventors: Anirudh KOUL, Meher KASAM, Michael JOHNSTON, Swetha MACHANAVAJHALA, Elizabeth HALPER
  • Patent number: 9922095
    Abstract: One or more systems and/or techniques are provided for automatic closed captioning for media content. In an example, real-time content, occurring within a threshold timespan of a broadcast of media content (e.g., social network posts occurring during and an hour before a live broadcast of an interview), may be accessed. A list of named entities, occurring within the social network data, may be generated (e.g., Interviewer Jon, Interviewee Kathy, Husband Dave, Son Jack, etc.). A ranked list of named entities may be created based upon trending named entities within the list of named entities (e.g., a named entity may be ranked higher based upon a more frequent occurrence within the social network posts). A dynamic grammar (e.g., library, etc.) may be built based upon the ranked list of named entities. Speech recognition may be performed upon the broadcast of media content utilizing the dynamic grammar to create closed caption text.
    Type: Grant
    Filed: June 2, 2015
    Date of Patent: March 20, 2018
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Anirudh Koul, Ranjitha Gurunath Kulkarni, Serge-Eric Tremblay
  • Publication number: 20170286383
    Abstract: Systems, apparatuses, services, platforms, and methods are discussed herein that provide assistance for user interface devices. In one example, an assistance application is provided comprising an imaging system configured to capture an image of a scene, an interface system configured to provide data associated with the image to a distributed assistance service that responsively processes the data to recognize properties of the scene and establish feedback for a user based at least on the properties of the scene, and a user interface configured to provide the feedback to the user.
    Type: Application
    Filed: August 22, 2016
    Publication date: October 5, 2017
    Inventors: Anirudh Koul, Ao Li, Elias Haroun, Irene Wen Ling Chen, Shweta Sharma, Christiano Bianchet, Saqib Shaikh, Stéphane Morichère-Matte, Biing Tsyr Lai, Nathan Pak Kei Lam, Wendy Lu
  • Publication number: 20170223392
    Abstract: Architecture that enables the identification of entities such as people and content in live broadcasts (e.g., streaming content (e.g., video) of live events) and non-live presentations (e.g., movies), in realtime, using recognition processes. This can be accomplished by extracting live data related to a live event. With respect to people entities, filtering can be performed to identify the named (people) entities from the extracted live data, and trending topics discovered as relate to the named entities, as associated with the live event. Multiple images of the named entities that capture the named entities under different conditions are captured for the named entities. The images are then processed to extract and learn facial features (train one or more models), and facial recognition is then performed on faces in the video using the trained model(s).
    Type: Application
    Filed: April 19, 2017
    Publication date: August 3, 2017
    Applicant: Microsoft Technology Licensing, LLC
    Inventors: Anirudh Koul, Serge-Eric Tremblay