Patents Assigned to Oblong Industries, Inc.
-
Patent number: 8407725Abstract: Embodiments described herein include mechanisms for encapsulating data that needs to be shared between or across processes. These mechanisms include slawx (plural of “slaw”), proteins, and pools. Generally, slawx provide the lowest-level of data definition for inter-process exchange, proteins provide mid-level structure and hooks for querying and filtering, and pools provide for high-level organization and access semantics. Slawx includes a mechanism for efficient, platform-independent data representation and access. Proteins provide a data encapsulation and transport scheme using slawx as the payload. Pools provide structured and flexible aggregation, ordering, filtering, and distribution of proteins within a process, among local processes, across a network between remote or distributed processes, and via longer term (e.g. on-disk, etc.) storage.Type: GrantFiled: April 24, 2008Date of Patent: March 26, 2013Assignee: Oblong Industries, Inc.Inventors: Kwindla Hultman Kramer, John S. Underkoffler
-
Patent number: 8370383Abstract: A multi-process interactive system is described. The system includes numerous processes running on a processing device. The processes include separable program execution contexts of application programs, such that each application program comprises at least one process. The system translates events of each process into data capsules. A data capsule includes an application-independent representation of event data of an event and state information of the process originating the content of the data capsule. The system transfers the data messages into pools or repositories. Each process operates as a recognizing process, where the recognizing process recognizes in the pools data capsules comprising content that corresponds to an interactive function of the recognizing process and/or an identification of the recognizing process. The recognizing process retrieves recognized data capsules from the pools and executes processing appropriate to contents of the recognized data capsules.Type: GrantFiled: October 14, 2009Date of Patent: February 5, 2013Assignee: Oblong Industries, Inc.Inventors: Kwindla H. Kramer, John S. Underkoffler
-
Publication number: 20100131549Abstract: A multi-process interactive system is described. The system includes numerous processes running on a processing device. The processes include separable program execution contexts of application programs, such that each application program comprises at least one process. The system translates events of each process into data capsules. A data capsule includes an application-independent representation of event data of an event and state information of the process originating the content of the data capsule. The system transfers the data messages into pools or repositories. Each process operates as a recognizing process, where the recognizing process recognizes in the pools data capsules comprising content that corresponds to an interactive function of the recognizing process and/or an identification of the recognizing process. The recognizing process retrieves recognized data capsules from the pools and executes processing appropriate to contents of the recognized data capsules.Type: ApplicationFiled: October 14, 2009Publication date: May 27, 2010Applicant: OBLONG INDUSTRIES, INC.Inventors: Kwindla H. Kramer, John S. Underkoffler
-
Publication number: 20100128062Abstract: A multi-process interactive system is described. The system includes numerous processes running on a processing device. The processes include separable program execution contexts of application programs, such that each application program comprises at least one process. The system translates events of each process into data capsules. A data capsule includes an application-independent representation of event data of an event and state information of the process originating the content of the data capsule. The system transfers the data messages into pools or repositories. Each process operates as a recognizing process, where the recognizing process recognizes in the pools data capsules comprising content that corresponds to an interactive function of the recognizing process and/or an identification of the recognizing process. The recognizing process retrieves recognized data capsules from the pools and executes processing appropriate to contents of the recognized data capsules.Type: ApplicationFiled: October 14, 2009Publication date: May 27, 2010Applicant: OBLONG INDUSTRIES, INC.Inventors: Kwindla H. Kramer, John S. Underkoffler
-
Publication number: 20100127968Abstract: A multi-process interactive system is described. The system includes numerous processes running on a processing device. The processes include separable program execution contexts of application programs, such that each application program comprises at least one process. The system translates events of each process into data capsules. A data capsule includes an application-independent representation of event data of an event and state information of the process originating the content of the data capsule. The system transfers the data messages into pools or repositories. Each process operates as a recognizing process, where the recognizing process recognizes in the pools data capsules comprising content that corresponds to an interactive function of the recognizing process and/or an identification of the recognizing process. The recognizing process retrieves recognized data capsules from the pools and executes processing appropriate to contents of the recognized data capsules.Type: ApplicationFiled: October 14, 2009Publication date: May 27, 2010Applicant: OBLONG INDUSTRIES, INC.Inventors: Kwindla Hultman Kramer, John S. Underkoffler
-
Publication number: 20100090947Abstract: The system provides a gestural interface to various visually presented elements, presented on a display screen or screens. A gestural vocabulary includes ‘instantaneous’ commands, in which forming one or both hands into the appropriate ‘pose’ results in an immediate, one-time action; and ‘spatial’ commands, in which the operator either refers directly to elements on the screen by way of literal ‘pointing’ gestures or performs navigational maneuvers by way of relative or “offset” gestures. The system contemplates the ability to identify the users hands in the form of a glove or gloves with certain indicia provided thereon, or any suitable means for providing recognizable indicia on a user's hands or body parts. A system of cameras can detect the position, orientation, and movement of the user's hands and translate that information into executable commands.Type: ApplicationFiled: October 2, 2009Publication date: April 15, 2010Applicant: OBLONG INDUSTRIES, INC.Inventors: John S. Underkoffler, Kevin T. Parent
-
Publication number: 20100090946Abstract: The system provides a gestural interface to various visually presented elements, presented on a display screen or screens. A gestural vocabulary includes ‘instantaneous’ commands, in which forming one or both hands into the appropriate ‘pose’ results in an immediate, one-time action; and ‘spatial’ commands, in which the operator either refers directly to elements on the screen by way of literal ‘pointing’ gestures or performs navigational maneuvers by way of relative or “offset” gestures. The system contemplates the ability to identify the users hands in the form of a glove or gloves with certain indicia provided thereon, or any suitable means for providing recognizable indicia on a user's hands or body parts. A system of cameras can detect the position, orientation, and movement of the user's hands and translate that information into executable commands.Type: ApplicationFiled: October 2, 2009Publication date: April 15, 2010Applicant: OBLONG INDUSTRIES, INC.Inventors: John S. Underkoffler, Kevin T. Parent
-
Publication number: 20100066676Abstract: Systems and methods are described for controlling a remote system. The controlling of the remote system comprises detecting a gesture of a body from gesture data received via a detector. The gesture data is absolute three-space location data of an instantaneous state of the body at a point in time and physical space. The detecting comprises identifying the gesture using the gesture data. The controlling comprises translating the gesture to a gesture signal, and controlling a component of the remote system in response to the gesture signal.Type: ApplicationFiled: September 10, 2009Publication date: March 18, 2010Applicant: OBLONG INDUSTRIES, INC.Inventors: Kwindla Hultman Kramer, Tom White, Mattie Ruth Kramer
-
Publication number: 20100060576Abstract: Systems and methods are described for navigating through a data space. The navigating comprises detecting a gesture of a body from gesture data received via a detector. The gesture data is absolute three-space location data of an instantaneous state of the body at a point in time and physical space. The detecting comprises identifying the gesture using the gesture data. The navigating comprises translating the gesture to a gesture signal, and navigating through the data space in response to the gesture signal. The data space is a data-representational space comprising a dataset represented in the physical space.Type: ApplicationFiled: September 3, 2009Publication date: March 11, 2010Applicant: Oblong Industries, Inc.Inventors: John S. Underkoffler, Kwindla H. Kramer
-
Publication number: 20100060570Abstract: Systems and methods are described for navigating through a data space. The navigating comprises detecting a gesture of a body from gesture data received via a detector. The gesture data is absolute three-space location data of an instantaneous state of the body at a point in time and physical space. The detecting comprises identifying the gesture using the gesture data. The navigating comprises translating the gesture to a gesture signal, and navigating through the data space in response to the gesture signal. The data space is a data-representational space comprising a dataset represented in the physical space.Type: ApplicationFiled: September 3, 2009Publication date: March 11, 2010Applicant: OBLONG INDUSTRIES, INC.Inventors: John S. Underkoffler, Kwindla H. Kramer
-
Publication number: 20100053304Abstract: Systems and methods are described for navigating through a data space. The navigating comprises detecting a gesture of a body from gesture data received via a detector. The gesture data is absolute three-space location data of an instantaneous state of the body at a point in time and physical space. The detecting comprises identifying the gesture using the gesture data. The navigating comprises translating the gesture to a gesture signal, and navigating through the data space in response to the gesture signal. The data space is a data-representational space comprising a dataset represented in the physical space.Type: ApplicationFiled: September 3, 2009Publication date: March 4, 2010Applicant: OBLONG INDUSTRIES, INC.Inventors: John S. Underkoffler, Kwindla H. Kramer
-
Publication number: 20090278915Abstract: Systems and method for controlling vehicle subsystems using gestures are described. The controlling comprises receiving an image of an object via sensors in a vehicle and outputting gesture data representing a gesture made by the object. The object comprises at least one hand and/or finger of a user. The gesture data is an instantaneous state of the object at a point in time in space. The controlling comprises automatically detecting the gesture of the object from the gesture data by extracting positions of the object, recovering quantized poses and orientation vectors of the object, and translating the gesture data to a gesture signal. The controlling comprises managing user interactions with the vehicle subsystems in response to the gesture signal.Type: ApplicationFiled: June 18, 2009Publication date: November 12, 2009Applicant: Oblong Industries, Inc.Inventors: Kwindla Hultman Kramer, John S. Underkoffler
-
Patent number: 7598942Abstract: The system provides a gestural interface to various visually presented elements, presented on a display screen or screens. A gestural vocabulary includes ‘instantaneous’ commands, in which forming one or both hands into the appropriate ‘pose’ results in an immediate, one-time action; and ‘spatial’ commands, in which the operator either refers directly to elements on the screen by way of literal ‘pointing’ gestures or performs navigational maneuvers by way of relative or “offset” gestures. The system contemplates the ability to identify the users hands in the form of a glove or gloves with certain indicia provided thereon, or any suitable means for providing recognizable indicia on a user's hands or body parts. A system of cameras can detect the position, orientation, and movement of the user's hands and translate that information into executable commands.Type: GrantFiled: February 8, 2006Date of Patent: October 6, 2009Assignee: Oblong Industries, Inc.Inventors: John S. Underkoffler, Kevin T. Parent, Kwindla H. Kramer
-
Publication number: 20090231278Abstract: Systems and methods are described for gesture-based control using three-dimensional information extracted over an extended depth of field. The system comprises a plurality of optical detectors coupled to at least one processor. The optical detectors image a body. At least two optical detectors of the plurality of optical detectors comprise wavefront coding cameras. The processor automatically detects a gesture of the body, wherein the gesture comprises an instantaneous state of the body. The detecting comprises aggregating gesture data of the gesture at an instant in time. The gesture data includes focus-resolved data of the body within a depth of field of the imaging system. The processor translates the gesture to a gesture signal, and uses the gesture signal to control a component coupled to the processor.Type: ApplicationFiled: April 2, 2009Publication date: September 17, 2009Applicant: OBLONG INDUSTRIES, INC.Inventors: Pierre St. Hilaire, John S. Underkoffler