Patents by Inventor Sergey Chub
Sergey Chub has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Patent number: 10235017Abstract: An integrated media user interface is described. In an implementation, a user interface is output that includes a plurality of representations of media, each of the media rendered by the mobile device through execution of a respective one of a plurality of applications. Responsive to selection of at least one of the representations, a corresponding application is located to render the media represented by the at least one said representation. The media represented by the selected representation is rendered through execution of the located application.Type: GrantFiled: April 19, 2016Date of Patent: March 19, 2019Assignee: Microsoft Technology Licensing, LLCInventors: Kerri I. H. Sheldon, Timothy J. Scudder, Veronica Y. Law, Marc S. Doll, Jonathan D. Wiswall, Megan L. Tedesco, Sergey Chub, Eric J. Hull
-
Publication number: 20160306512Abstract: An integrated media user interface is described. In an implementation, a user interface is output that includes a plurality of representations of media, each of the media rendered by the mobile device through execution of a respective one of a plurality of applications. Responsive to selection of at least one of the representations, a corresponding application is located to render the media represented by the at least one said representation. The media represented by the selected representation is rendered through execution of the located application.Type: ApplicationFiled: April 19, 2016Publication date: October 20, 2016Applicant: Microsoft Technology Licensing, LLCInventors: Kerri I. H. Sheldon, Timothy J. Scudder, Veronica Y. Law, Marc S. Doll, Jonathan D. Wiswall, Megan L. Tedesco, Sergey Chub, Eric J. Hull
-
Patent number: 9335903Abstract: An integrated media user interface is described. In an implementation, a user interface is output that includes a plurality of representations of media, each of the media rendered by the mobile device through execution of a respective one of a plurality of applications. Responsive to selection of at least one of the representations, a corresponding application is located to render the media represented by the at least one said representation. The media represented by the selected representation is rendered through execution of the located application.Type: GrantFiled: July 1, 2013Date of Patent: May 10, 2016Inventors: Kerri I. H. Sheldon, Timothy J. Scudder, Veronica Y. Law, Marc S. Doll, Jonathan D. Wiswall, Megan L. Tedesco, Sergey Chub, Eric J. Hull
-
Patent number: 9223766Abstract: An e-reader application preserves formatting of a selection through a snippet. The application determines a format of a portion of content in response to a user action selecting the portion. The formatted portion is stored in the snippet. The source code of the portion is extracted from the content and stored in the snippet to preserve its format. The snippet is displayed in a navigation pane using the format. The snippet is resized according to the defined format to fit available space in the navigation pane. The application is enabled to navigate to a location of the portion within the content in response to detecting another selection activating the snippet.Type: GrantFiled: November 21, 2012Date of Patent: December 29, 2015Assignee: Microsoft Technology Licensing, LLCInventors: Sergey Chub, Raman Narayanan, Ming Liu, Radoslav Nickolov, Wei Zeng
-
Publication number: 20140143645Abstract: An e-reader application preserves formatting of a selection through a snippet. The application determines a format of a portion of content in response to a user action selecting the portion. The formatted portion is stored in the snippet. The source code of the portion is extracted from the content and stored in the snippet to preserve its format. The snippet is displayed in a navigation pane using the format. The snippet is resized according to the defined format to fit available space in the navigation pane. The application is enabled to navigate to a location of the portion within the content in response to detecting another selection activating the snippet.Type: ApplicationFiled: November 21, 2012Publication date: May 22, 2014Applicant: Microsoft CorporationInventors: Sergey Chub, Raman Narayanan, Ming Liu, Radoslav Nickolov, Wei Zeng
-
Publication number: 20140047332Abstract: E-reader architecture with rich service infrastructure for identity, storage, protection, commercial exchange, and sharing of content is provided. The architecture may enable consumption of content through a set of compatible cross-platform consumption experiences that provide easy access to professional, institutional, collaborative or personal content with auxiliary capabilities such as search, commenting, posting, and similar ones.Type: ApplicationFiled: November 15, 2012Publication date: February 13, 2014Applicant: Microsoft CorporationInventors: Ming Liu, Wei Zeng, Raman Narayanan, Sergey Chub, Radoslav Petrov Nickolov
-
Publication number: 20140047308Abstract: An e-reader application provides a note based annotation of content in an e-reader. A user action to annotate the content initiates a note pane to capture a note. The note includes variety of types including text, ink, snippet, audio recording, and photo. The note is captured in the note pane which is configured according to the note type. The note is stored in a platform-agnostic in a cloud storage. A local cache also maintains the note for offline usage. The note is anchored to the content using an anchor marker placed within proximity to a selection associated with the note in the content.Type: ApplicationFiled: November 21, 2012Publication date: February 13, 2014Applicant: Microsoft CorporationInventors: Sergey Chub, Raman Narayanan, Ming Liu, Radoslav Nickolov, Wei Zeng
-
Publication number: 20130298022Abstract: An integrated media user interface is described. In an implementation, a user interface is output that includes a plurality of representations of media, each of the media rendered by the mobile device through execution of a respective one of a plurality of applications. Responsive to selection of at least one of the representations, a corresponding application is located to render the media represented by the at least one said representation. The media represented by the selected representation is rendered through execution of the located application.Type: ApplicationFiled: July 1, 2013Publication date: November 7, 2013Inventors: Kerri I. H. Sheldon, Timothy J. Scudder, Veronica Y. Law, Marc S. Doll, Jonathan D. Wiswall, Megan L. Tedesco, Sergey Chub, Eric J. Hull
-
Publication number: 20130263052Abstract: A user interface (UI) system calculates movements in a multi-layer graphical user interface. The UI system receives user input corresponding to gestures on a touchscreen. The UI system calculates a movement of a first layer in a first direction (e.g., a horizontal direction) at a first movement rate. The UI system calculates a movement of a second layer substantially parallel to the movement of the first layer, at a second movement rate that differs from the first movement rate. The UI system calculates a movement (e.g., a vertical movement) in a direction substantially orthogonal to the first direction, in a UI element of one of the layers.Type: ApplicationFiled: May 24, 2013Publication date: October 3, 2013Applicant: Microsoft CorporationInventors: Jeffrey Cheng-Yao Fong, Eric J. Hull, Sergey Chub
-
Patent number: 8494590Abstract: An integrated media user interface is described. In an implementation, a user interface is output that includes a plurality of representations of media, each of the media rendered by the mobile device through execution of a respective one of a plurality of applications. Responsive to selection of at least one of the representations, a corresponding application is located to render the media represented by the at least one said representation. The media represented by the selected representation is rendered through execution of the located application.Type: GrantFiled: March 3, 2010Date of Patent: July 23, 2013Assignee: Microsoft CorporationInventors: Kerri I. H. Sheldon, Timothy J. Scudder, Veronica Y. Law, Marc S. Doll, Jonathan D. Wiswall, Megan L. Tedesco, Sergey Chub, Eric J. Hull
-
Patent number: 8473860Abstract: A user interface (UI) system calculates movements in a multi-layer graphical user interface. The UI system receives user input corresponding to gestures on a touchscreen. The UI system calculates a movement of a first layer in a first direction (e.g., a horizontal direction) at a first movement rate. The UI system calculates a movement of a second layer substantially parallel to the movement of the first layer, at a second movement rate that differs from the first movement rate. The UI system calculates a movement (e.g., a vertical movement) in a direction substantially orthogonal to the first direction, in a UI element of one of the layers.Type: GrantFiled: June 25, 2010Date of Patent: June 25, 2013Assignee: Microsoft CorporationInventors: Jeffrey Cheng-Yao Fong, Eric J. Hull, Sergey Chub
-
Publication number: 20120159404Abstract: A processing device and method are provided for capturing images, via an image-capturing component of a processing device, and determining a motion of the processing device. An adaptive search center technique may be employed to determine a search center with respect to multiple equal-sized regions of an image frame, based on previously estimated motion vectors. One of several fast block matching methods may be used, based on one or more conditions, to match a block of pixels of one image frame with a second block of pixels of a second image. Upon matching blocks of pixels, motion vectors of the multiple equal-sized regions may be estimated. The motion may be determined, based on the estimated motion vectors, and an associated action may be performed. Various embodiments may implement techniques to distinguish motion blur from de-focus blur and to determine a change in lighting condition.Type: ApplicationFiled: February 16, 2012Publication date: June 21, 2012Applicant: MICROSOFT CORPORATIONInventors: Srinath Vasireddy, Sergey Chub, Gang Hua, Ting-yi Yang
-
Patent number: 8170901Abstract: A user interface for building a componentized workflow model. Each step of the workflow is modeled as an activity that has metadata to describe design time aspects, compile time aspects, and runtime aspects of the workflow step. A user selects and arranges the activities to create the workflow via the user interface. The metadata associated with each of the activities in the workflow is collected to create a persistent representation of the workflow. Users extend the workflow model by authoring custom activities.Type: GrantFiled: January 31, 2005Date of Patent: May 1, 2012Assignee: Microsoft CorporationInventors: Dharma K. Shukla, Aditya G. Bhandarkar, Akash J. Sagar, Sergey Chub, Mayank Mehta, Dennis Pilarinos, Eli Hisdai, Abhay Vinayak Parasnis
-
Patent number: 8144780Abstract: A processing device and method are provided for capturing images, via an image-capturing component of a processing device, and determining a motion of the processing device. An adaptive search center technique may be employed to determine a search center with respect to multiple equal-sized regions of an image frame, based on previously estimated motion vectors. One of several fast block matching methods may be used, based on one or more conditions, to match a block of pixels of one image frame with a second block of pixels of a second image. Upon matching blocks of pixels, motion vectors of the multiple equal-sized regions may be estimated. The motion may be determined, based on the estimated motion vectors, and an associated action may be performed. Various embodiments may implement techniques to distinguish motion blur from de-focus blur and to determine a change in lighting condition.Type: GrantFiled: September 24, 2007Date of Patent: March 27, 2012Assignee: Microsoft CorporationInventors: Srinath Vasireddy, Sergey Chub, Gang Hua, Ting-yi Yang
-
Patent number: 8090610Abstract: A user interface for building a componentized workflow model. Each step of the workflow is modeled as an activity that has metadata to describe design time aspects, compile time aspects, and runtime aspects of the workflow step. A user selects and arranges the activities to create the workflow via the user interface. The metadata associated with each of the activities in the workflow is collected to create a persistent representation of the workflow. Users extend the workflow model by authoring custom activities.Type: GrantFiled: January 31, 2005Date of Patent: January 3, 2012Assignee: Microsoft CorporationInventors: Dharma K. Shukla, Aditya G. Bhandarkar, Akash J. Sagar, Sergey Chub, Mayank Mehta, Dennis Pilarinos, Eli Hisdai, Abhay Vinayak Parasnis
-
Publication number: 20110199318Abstract: A user interface (UI) system calculates movements in a multi-layer graphical user interface. The UI system receives user input corresponding to gestures on a touchscreen. The UI system calculates a movement of a first layer in a first direction (e.g., a horizontal direction) at a first movement rate. For example, the first movement rate can be substantially equal to the movement rate of a gesture made by a user's finger or other object on the touchscreen. The UI system calculates movements of other layers substantially parallel to the movement of the first layer, at movement rates that differ from the first movement rate.Type: ApplicationFiled: June 25, 2010Publication date: August 18, 2011Applicant: Microsoft CorporationInventors: Jeffrey Cheng-Yao Fong, Eric J. Hull, Sergey Chub
-
Publication number: 20110202837Abstract: A user interface (UI) system calculates movements in a multi-layer graphical user interface. The UI system receives user input corresponding to gestures on a touchscreen. The UI system calculates a movement of a first layer in a first direction (e.g., a horizontal direction) at a first movement rate. The UI system calculates a movement of a second layer substantially parallel to the movement of the first layer, at a second movement rate that differs from the first movement rate. The UI system calculates a movement (e.g., a vertical movement) in a direction substantially orthogonal to the first direction, in a UI element of one of the layers.Type: ApplicationFiled: June 25, 2010Publication date: August 18, 2011Applicant: Microsoft CorporationInventors: Jeffrey Cheng-Yao Fong, Eric J. Hull, Sergey Chub
-
Publication number: 20110190032Abstract: An integrated media user interface is described. In an implementation, a user interface is output that includes a plurality of representations of media, each of the media rendered by the mobile device through execution of a respective one of a plurality of applications. Responsive to selection of at least one of the representations, a corresponding application is located to render the media represented by the at least one said representation. The media represented by the selected representation is rendered through execution of the located application.Type: ApplicationFiled: March 3, 2010Publication date: August 4, 2011Inventors: Kerri I.H. Sheldon, Timothy J. Scudder, Veronica Y. Law, Marc S. Doll, Jonathan D. Wiswall, Megan L. Tedesco, Sergey Chub, Eric J. Hull
-
Publication number: 20090080526Abstract: A processing device and method are provided for capturing images, via an image-capturing component of a processing device, and determining a motion of the processing device. An adaptive search center technique may be employed to determine a search center with respect to multiple equal-sized regions of an image frame, based on previously estimated motion vectors. One of several fast block matching methods may be used, based on one or more conditions, to match a block of pixels of one image frame with a second block of pixels of a second image. Upon matching blocks of pixels, motion vectors of the multiple equal-sized regions may be estimated. The motion may be determined, based on the estimated motion vectors, and an associated action may be performed. Various embodiments may implement techniques to distinguish motion blur from de-focus blur and to determine a change in lighting condition.Type: ApplicationFiled: September 24, 2007Publication date: March 26, 2009Applicant: MICROSOFT CORPORATIONInventors: Srinath Vasireddy, Sergey Chub, Gang Hua, Ting-yi Yang
-
Publication number: 20060288332Abstract: Workflow debugging. A debug engine integrated with an external development tool debugging framework sets breakpoints directly on workflow activities and stops execution of the workflow at each of the breakpoints. The state of the workflow activity at the breakpoint is displayed. The debug engine extracts source code of the workflow for debugging at the source code level.Type: ApplicationFiled: June 21, 2005Publication date: December 21, 2006Applicant: Microsoft CorporationInventors: Akash Sagar, Aditya Bhandarkar, Dennis Pilarinos, Dharma Shukla, Mayank Mehta, Sergey Chub, Vishal Kalra