Patents by Inventor Thomas M. Soemo
Thomas M. Soemo has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Publication number: 20240078682Abstract: Training a multi-object tracking model includes: generating a plurality of training images based at least on scene generation information, each training image comprising a plurality of objects to be tracked; generating, for each training image, original simulated data based at least on the scene generation information, the original simulated data comprising tag data for a first object; locating, within the original simulated data, tag data for the first object, based on at least an anomaly alert (e.g., occlusion alert, proximity alert, motion alert) associated with the first object in the first training image; based at least on locating the tag data for the first object, modifying at least a portion of the tag data for the first object from the original simulated data, thereby generating preprocessed training data from the original simulated data; and training a multi-object tracking model with the preprocessed training data to produce a trained multi-object tracker.Type: ApplicationFiled: November 13, 2023Publication date: March 7, 2024Inventors: Ishani CHAKRABORTY, Jonathan C. HANZELKA, Lu YUAN, Pedro Urbina ESCOS, Thomas M. SOEMO
-
Patent number: 11854211Abstract: Training a multi-object tracking model includes: generating a plurality of training images based at least on scene generation information, each training image comprising a plurality of objects to be tracked; generating, for each training image, original simulated data based at least on the scene generation information, the original simulated data comprising tag data for a first object; locating, within the original simulated data, tag data for the first object, based on at least an anomaly alert (e.g., occlusion alert, proximity alert, motion alert) associated with the first object in the first training image; based at least on locating the tag data for the first object, modifying at least a portion of the tag data for the first object from the original simulated data, thereby generating preprocessed training data from the original simulated data; and training a multi-object tracking model with the preprocessed training data to produce a trained multi-object tracker.Type: GrantFiled: January 26, 2022Date of Patent: December 26, 2023Assignee: Microsoft Technology Licensing, LLC.Inventors: Ishani Chakraborty, Jonathan C. Hanzelka, Lu Yuan, Pedro Urbina Escos, Thomas M. Soemo
-
Patent number: 11335008Abstract: Training a multi-object tracking model includes: generating a plurality of training images based at least on scene generation information, each training image comprising a plurality of objects to be tracked; generating, for each training image, original simulated data based at least on the scene generation information, the original simulated data comprising tag data for a first object; locating, within the original simulated data, tag data for the first object, based on at least an anomaly alert (e.g., occlusion alert, proximity alert, motion alert) associated with the first object in the first training image; based at least on locating the tag data for the first object, modifying at least a portion of the tag data for the first object from the original simulated data, thereby generating preprocessed training data from the original simulated data; and training a multi-object tracking model with the preprocessed training data to produce a trained multi-object tracker.Type: GrantFiled: September 18, 2020Date of Patent: May 17, 2022Assignee: Microsoft Technology Licensing, LLCInventors: Ishani Chakraborty, Jonathan C. Hanzelka, Lu Yuan, Pedro Urbina Escos, Thomas M. Soemo
-
Publication number: 20220148197Abstract: Training a multi-object tracking model includes: generating a plurality of training images based at least on scene generation information, each training image comprising a plurality of objects to be tracked; generating, for each training image, original simulated data based at least on the scene generation information, the original simulated data comprising tag data for a first object; locating, within the original simulated data, tag data for the first object, based on at least an anomaly alert (e.g., occlusion alert, proximity alert, motion alert) associated with the first object in the first training image; based at least on locating the tag data for the first object, modifying at least a portion of the tag data for the first object from the original simulated data, thereby generating preprocessed training data from the original simulated data; and training a multi-object tracking model with the preprocessed training data to produce a trained multi-object tracker.Type: ApplicationFiled: January 26, 2022Publication date: May 12, 2022Inventors: Ishani CHAKRABORTY, Jonathan C. HANZELKA, Lu YUAN, Pedro Urbina ESCOS, Thomas M. SOEMO
-
Publication number: 20220092792Abstract: Training a multi-object tracking model includes: generating a plurality of training images based at least on scene generation information, each training image comprising a plurality of objects to be tracked; generating, for each training image, original simulated data based at least on the scene generation information, the original simulated data comprising tag data for a first object; locating, within the original simulated data, tag data for the first object, based on at least an anomaly alert (e.g., occlusion alert, proximity alert, motion alert) associated with the first object in the first training image; based at least on locating the tag data for the first object, modifying at least a portion of the tag data for the first object from the original simulated data, thereby generating preprocessed training data from the original simulated data; and training a multi-object tracking model with the preprocessed training data to produce a trained multi-object tracker.Type: ApplicationFiled: September 18, 2020Publication date: March 24, 2022Inventors: Ishani CHAKRABORTY, Jonathan C. HANZELKA, Lu YUAN, Pedro Urbina ESCOS, Thomas M. SOEMO
-
Patent number: 10534438Abstract: A multimedia entertainment system combines both gestures and voice commands to provide an enhanced control scheme. A user's body position or motion may be recognized as a gesture, and may be used to provide context to recognize user generated sounds, such as speech input. Likewise, speech input may be recognized as a voice command, and may be used to provide context to recognize a body position or motion as a gesture. Weights may be assigned to the inputs to facilitate processing. When a gesture is recognized, a limited set of voice commands associated with the recognized gesture are loaded for use. Further, additional sets of voice commands may be structured in a hierarchical manner such that speaking a voice command from one set of voice commands leads to the system loading a next set of voice commands.Type: GrantFiled: April 28, 2017Date of Patent: January 14, 2020Assignee: MICROSOFT TECHNOLOGY LICENSING, LLCInventors: Christian Klein, Ali M. Vassigh, Jason S. Flaks, Vanessa Larco, Thomas M. Soemo
-
Patent number: 10368120Abstract: A method and system are disclosed in which a group of people are able to replicate the physical world experience of going with a group of friends to pick a movie, watch the movie together, and provide commentary on the movie itself in the virtual world on a virtual couch while each user is sitting in different physical locations. Additionally, the virtual representation of the destination that the group of people are watching the movie together in can be themed to allow users to watch movies in different locations pivoting on special events or by the users choice.Type: GrantFiled: August 11, 2016Date of Patent: July 30, 2019Assignee: MICROSOFT TECHNOLOGY LICENSING, LLCInventors: Andrew Lawrence Mattingly, Brian Charles Kramp, Thomas M. Soemo, Eddie Mays
-
Patent number: 9945946Abstract: Examples are disclosed herein that relate to depth imaging techniques using ultrasound One example provides an ultrasonic depth sensing system configured to, for an image frame, emit an ultrasonic pulse from each of a plurality of transducers, receive a reflection of each ultrasonic pulse at a microphone array, perform transmit beamforming and also receive beamforming computationally after receiving the reflections, form a depth image, and output the depth image for the image frame.Type: GrantFiled: September 11, 2014Date of Patent: April 17, 2018Assignee: MICROSOFT TECHNOLOGY LICENSING, LLCInventors: Ivan Dokmanic, Ivan J. Tashev, Thomas M. Soemo
-
Publication number: 20170228036Abstract: A multimedia entertainment system combines both gestures and voice commands to provide an enhanced control scheme. A user's body position or motion may be recognized as a gesture, and may be used to provide context to recognize user generated sounds, such as speech input. Likewise, speech input may be recognized as a voice command, and may be used to provide context to recognize a body position or motion as a gesture. Weights may be assigned to the inputs to facilitate processing. When a gesture is recognized, a limited set of voice commands associated with the recognized gesture are loaded for use. Further, additional sets of voice commands may be structured in a hierarchical manner such that speaking a voice command from one set of voice commands leads to the system loading a next set of voice commands.Type: ApplicationFiled: April 28, 2017Publication date: August 10, 2017Applicant: Microsoft Technology Licensing, LLCInventors: Christian Klein, Ali M. Vassigh, Jason S. Flaks, Vanessa Larco, Thomas M. Soemo
-
Publication number: 20170041658Abstract: A method and system are disclosed in which a group of people are able to replicate the physical world experience of going with a group of friends to pick a movie, watch the movie together, and provide commentary on the movie itself in the virtual world on a virtual couch while each user is sitting in different physical locations. Additionally, the virtual representation of the destination that the group of people are watching the movie together in can be themed to allow users to watch movies in different locations pivoting on special events or by the users choice.Type: ApplicationFiled: August 11, 2016Publication date: February 9, 2017Inventors: Andrew Lawrence Mattingly, Brian Charles Kramp, Thomas M. Soemo, Eddie Mays
-
Patent number: 9423945Abstract: A method and system are disclosed in which a group of people are able to replicate the physical world experience of going with a group of friends to pick a movie, watch the movie together, and provide commentary on the movie itself in the virtual world on a virtual couch while each user is sitting in different physical locations. Additionally, the virtual representation of the destination that the group of people are watching the movie together in can be themed to allow users to watch movies in different locations pivoting on special events or by the users choice.Type: GrantFiled: August 24, 2015Date of Patent: August 23, 2016Assignee: Microsoft Technology Licensing, LLCInventors: Andrew Lawrence Mattingly, Brian Charles Kramp, Thomas M. Soemo, Eddie Mays
-
Publication number: 20160077206Abstract: Examples are disclosed herein that relate to depth imaging techniques using ultrasound One example provides an ultrasonic depth sensing system configured to, for an image frame, emit an ultrasonic pulse from each of a plurality of transducers, receive a reflection of each ultrasonic pulse at a microphone array, perform transmit beamforming and also receive beamforming computationally after receiving the reflections, form a depth image, and output the depth image for the image frame.Type: ApplicationFiled: September 11, 2014Publication date: March 17, 2016Inventors: Ivan Dokmanic, Ivan J. Tashev, Thomas M. Soemo
-
Publication number: 20150363099Abstract: A method and system are disclosed in which a group of people are able to replicate the physical world experience of going with a group of friends to pick a movie, watch the movie together, and provide commentary on the movie itself in the virtual world on a virtual couch while each user is sitting in different physical locations. Additionally, the virtual representation of the destination that the group of people are watching the movie together in can be themed to allow users to watch movies in different locations pivoting on special events or by the users choice.Type: ApplicationFiled: August 24, 2015Publication date: December 17, 2015Inventors: Andrew Lawrence Mattingly, Brian Charles Kramp, Thomas M. Soemo, Eddie Mays
-
Patent number: 9118737Abstract: A method and system are disclosed in which a group of people are able to replicate the physical world experience of going with a group of friends to pick a movie, watch the movie together, and provide commentary on the movie itself in the virtual world on a virtual couch while each user is sitting in different physical locations. Additionally, the virtual representation of the destination that the group of people are watching the movie together in can be themed to allow users to watch movies in different locations pivoting on special events or by the users choice.Type: GrantFiled: February 24, 2014Date of Patent: August 25, 2015Assignee: MICROSOFT TECHNOLOGY LICENSING, LLCInventors: Andrew Lawrence Mattingly, Brian Charles Kramp, Thomas M. Soemo, Eddie Mays
-
Publication number: 20140173462Abstract: A method and system are disclosed in which a group of people are able to replicate the physical world experience of going with a group of friends to pick a movie, watch the movie together, and provide commentary on the movie itself in the virtual world on a virtual couch while each user is sitting in different physical locations. Additionally, the virtual representation of the destination that the group of people are watching the movie together in can be themed to allow users to watch movies in different locations pivoting on special events or by the users choice.Type: ApplicationFiled: February 24, 2014Publication date: June 19, 2014Applicant: Microsoft CorporationInventors: Andrew Lawrence Mattingly, Brian Charles Kramp, Thomas M. Soemo, Eddie Mays
-
Patent number: 8661353Abstract: A method and system are disclosed in which a group of people are able to replicate the physical world experience of going with a group of friends to pick a movie, watch the movie together, and provide commentary on the movie itself in the virtual world on a virtual couch while each user is sitting in different physical locations. Additionally, the virtual representation of the destination that the group of people are watching the movie together in can be themed to allow users to watch movies in different locations pivoting on special events or by the users choice.Type: GrantFiled: August 31, 2009Date of Patent: February 25, 2014Assignee: Microsoft CorporationInventors: Andrew Lawrence Mattingly, Brian Charles Kramp, Thomas M. Soemo, Eddie Mays
-
Patent number: 8660847Abstract: A system for integrating local speech recognition with cloud-based speech recognition in order to provide an efficient natural user interface is described. In some embodiments, a computing device determines a direction associated with a particular person within an environment and generates an audio recording associated with the direction. The computing device then performs local speech recognition on the audio recording in order to detect a first utterance spoken by the particular person and to detect one or more keywords within the first utterance. The first utterance may be detected by applying voice activity detection techniques to the audio recording. The first utterance and the one or more keywords are subsequently transferred to a server which may identify speech sounds within the first utterance associated with the one or more keywords and adapt one or more speech recognition techniques based on the identified speech sounds.Type: GrantFiled: September 2, 2011Date of Patent: February 25, 2014Assignee: Microsoft CorporationInventors: Thomas M. Soemo, Leo Soong, Michael H. Kim, Chad R. Heinemann, Dax H. Hawkins
-
Publication number: 20130060571Abstract: A system for integrating local speech recognition with cloud-based speech recognition in order to provide an efficient natural user interface is described. In some embodiments, a computing device determines a direction associated with a particular person within an environment and generates an audio recording associated with the direction. The computing device then performs local speech recognition on the audio recording in order to detect a first utterance spoken by the particular person and to detect one or more keywords within the first utterance. The first utterance may be detected by applying voice activity detection techniques to the audio recording. The first utterance and the one or more keywords are subsequently transferred to a server which may identify speech sounds within the first utterance associated with the one or more keywords and adapt one or more speech recognition techniques based on the identified speech sounds.Type: ApplicationFiled: September 2, 2011Publication date: March 7, 2013Applicant: Microsoft CorporationInventors: Thomas M. Soemo, Leo Soong, Michael H. Kim, Chad R. Heinemann, Dax H. Hawkins
-
Patent number: 8387015Abstract: Scalable empirical testing of media file playback utilizes test hooks in each media player to support simulated human interaction and playback monitoring. A media crawler catalogs media files accumulated in a media file database to create a wordlist. One or more scalable instances of media tester accesses the wordlist to select items of work linked to media files. Work items and/or operating modes of media tester specify test parameters such as performance profiles or further define testing such as specifying repetitious playback on one or more media players. Media files are downloaded to and played by a scalable number of media players. Playback performance is monitored, analyzed and reported. Failure reports are accompanied by instructions to reproduce failures and cross-references to content or source code in media files. Failures can be audited by additional work items for follow-up testing.Type: GrantFiled: January 31, 2008Date of Patent: February 26, 2013Assignee: Microsoft CorporationInventors: Russell D. Christensen, Jun Ma, Thomas M. Soemo
-
Patent number: 8296151Abstract: A multimedia entertainment system combines both gestures and voice commands to provide an enhanced control scheme. A user's body position or motion may be recognized as a gesture, and may be used to provide context to recognize user generated sounds, such as speech input. Likewise, speech input may be recognized as a voice command, and may be used to provide context to recognize a body position or motion as a gesture. Weights may be assigned to the inputs to facilitate processing. When a gesture is recognized, a limited set of voice commands associated with the recognized gesture are loaded for use. Further, additional sets of voice commands may be structured in a hierarchical manner such that speaking a voice command from one set of voice commands leads to the system loading a next set of voice commands.Type: GrantFiled: June 18, 2010Date of Patent: October 23, 2012Assignee: Microsoft CorporationInventors: Christian Klein, Ali M. Vassigh, Jason S. Flaks, Vanessa Larco, Thomas M. Soemo