Patents Assigned to Oblong Industries, Inc.
  • Patent number: 8407725
    Abstract: Embodiments described herein include mechanisms for encapsulating data that needs to be shared between or across processes. These mechanisms include slawx (plural of “slaw”), proteins, and pools. Generally, slawx provide the lowest-level of data definition for inter-process exchange, proteins provide mid-level structure and hooks for querying and filtering, and pools provide for high-level organization and access semantics. Slawx includes a mechanism for efficient, platform-independent data representation and access. Proteins provide a data encapsulation and transport scheme using slawx as the payload. Pools provide structured and flexible aggregation, ordering, filtering, and distribution of proteins within a process, among local processes, across a network between remote or distributed processes, and via longer term (e.g. on-disk, etc.) storage.
    Type: Grant
    Filed: April 24, 2008
    Date of Patent: March 26, 2013
    Assignee: Oblong Industries, Inc.
    Inventors: Kwindla Hultman Kramer, John S. Underkoffler
  • Patent number: 8370383
    Abstract: A multi-process interactive system is described. The system includes numerous processes running on a processing device. The processes include separable program execution contexts of application programs, such that each application program comprises at least one process. The system translates events of each process into data capsules. A data capsule includes an application-independent representation of event data of an event and state information of the process originating the content of the data capsule. The system transfers the data messages into pools or repositories. Each process operates as a recognizing process, where the recognizing process recognizes in the pools data capsules comprising content that corresponds to an interactive function of the recognizing process and/or an identification of the recognizing process. The recognizing process retrieves recognized data capsules from the pools and executes processing appropriate to contents of the recognized data capsules.
    Type: Grant
    Filed: October 14, 2009
    Date of Patent: February 5, 2013
    Assignee: Oblong Industries, Inc.
    Inventors: Kwindla H. Kramer, John S. Underkoffler
  • Publication number: 20100131549
    Abstract: A multi-process interactive system is described. The system includes numerous processes running on a processing device. The processes include separable program execution contexts of application programs, such that each application program comprises at least one process. The system translates events of each process into data capsules. A data capsule includes an application-independent representation of event data of an event and state information of the process originating the content of the data capsule. The system transfers the data messages into pools or repositories. Each process operates as a recognizing process, where the recognizing process recognizes in the pools data capsules comprising content that corresponds to an interactive function of the recognizing process and/or an identification of the recognizing process. The recognizing process retrieves recognized data capsules from the pools and executes processing appropriate to contents of the recognized data capsules.
    Type: Application
    Filed: October 14, 2009
    Publication date: May 27, 2010
    Applicant: OBLONG INDUSTRIES, INC.
    Inventors: Kwindla H. Kramer, John S. Underkoffler
  • Publication number: 20100128062
    Abstract: A multi-process interactive system is described. The system includes numerous processes running on a processing device. The processes include separable program execution contexts of application programs, such that each application program comprises at least one process. The system translates events of each process into data capsules. A data capsule includes an application-independent representation of event data of an event and state information of the process originating the content of the data capsule. The system transfers the data messages into pools or repositories. Each process operates as a recognizing process, where the recognizing process recognizes in the pools data capsules comprising content that corresponds to an interactive function of the recognizing process and/or an identification of the recognizing process. The recognizing process retrieves recognized data capsules from the pools and executes processing appropriate to contents of the recognized data capsules.
    Type: Application
    Filed: October 14, 2009
    Publication date: May 27, 2010
    Applicant: OBLONG INDUSTRIES, INC.
    Inventors: Kwindla H. Kramer, John S. Underkoffler
  • Publication number: 20100127968
    Abstract: A multi-process interactive system is described. The system includes numerous processes running on a processing device. The processes include separable program execution contexts of application programs, such that each application program comprises at least one process. The system translates events of each process into data capsules. A data capsule includes an application-independent representation of event data of an event and state information of the process originating the content of the data capsule. The system transfers the data messages into pools or repositories. Each process operates as a recognizing process, where the recognizing process recognizes in the pools data capsules comprising content that corresponds to an interactive function of the recognizing process and/or an identification of the recognizing process. The recognizing process retrieves recognized data capsules from the pools and executes processing appropriate to contents of the recognized data capsules.
    Type: Application
    Filed: October 14, 2009
    Publication date: May 27, 2010
    Applicant: OBLONG INDUSTRIES, INC.
    Inventors: Kwindla Hultman Kramer, John S. Underkoffler
  • Publication number: 20100090947
    Abstract: The system provides a gestural interface to various visually presented elements, presented on a display screen or screens. A gestural vocabulary includes ‘instantaneous’ commands, in which forming one or both hands into the appropriate ‘pose’ results in an immediate, one-time action; and ‘spatial’ commands, in which the operator either refers directly to elements on the screen by way of literal ‘pointing’ gestures or performs navigational maneuvers by way of relative or “offset” gestures. The system contemplates the ability to identify the users hands in the form of a glove or gloves with certain indicia provided thereon, or any suitable means for providing recognizable indicia on a user's hands or body parts. A system of cameras can detect the position, orientation, and movement of the user's hands and translate that information into executable commands.
    Type: Application
    Filed: October 2, 2009
    Publication date: April 15, 2010
    Applicant: OBLONG INDUSTRIES, INC.
    Inventors: John S. Underkoffler, Kevin T. Parent
  • Publication number: 20100090946
    Abstract: The system provides a gestural interface to various visually presented elements, presented on a display screen or screens. A gestural vocabulary includes ‘instantaneous’ commands, in which forming one or both hands into the appropriate ‘pose’ results in an immediate, one-time action; and ‘spatial’ commands, in which the operator either refers directly to elements on the screen by way of literal ‘pointing’ gestures or performs navigational maneuvers by way of relative or “offset” gestures. The system contemplates the ability to identify the users hands in the form of a glove or gloves with certain indicia provided thereon, or any suitable means for providing recognizable indicia on a user's hands or body parts. A system of cameras can detect the position, orientation, and movement of the user's hands and translate that information into executable commands.
    Type: Application
    Filed: October 2, 2009
    Publication date: April 15, 2010
    Applicant: OBLONG INDUSTRIES, INC.
    Inventors: John S. Underkoffler, Kevin T. Parent
  • Publication number: 20100066676
    Abstract: Systems and methods are described for controlling a remote system. The controlling of the remote system comprises detecting a gesture of a body from gesture data received via a detector. The gesture data is absolute three-space location data of an instantaneous state of the body at a point in time and physical space. The detecting comprises identifying the gesture using the gesture data. The controlling comprises translating the gesture to a gesture signal, and controlling a component of the remote system in response to the gesture signal.
    Type: Application
    Filed: September 10, 2009
    Publication date: March 18, 2010
    Applicant: OBLONG INDUSTRIES, INC.
    Inventors: Kwindla Hultman Kramer, Tom White, Mattie Ruth Kramer
  • Publication number: 20100060576
    Abstract: Systems and methods are described for navigating through a data space. The navigating comprises detecting a gesture of a body from gesture data received via a detector. The gesture data is absolute three-space location data of an instantaneous state of the body at a point in time and physical space. The detecting comprises identifying the gesture using the gesture data. The navigating comprises translating the gesture to a gesture signal, and navigating through the data space in response to the gesture signal. The data space is a data-representational space comprising a dataset represented in the physical space.
    Type: Application
    Filed: September 3, 2009
    Publication date: March 11, 2010
    Applicant: Oblong Industries, Inc.
    Inventors: John S. Underkoffler, Kwindla H. Kramer
  • Publication number: 20100060570
    Abstract: Systems and methods are described for navigating through a data space. The navigating comprises detecting a gesture of a body from gesture data received via a detector. The gesture data is absolute three-space location data of an instantaneous state of the body at a point in time and physical space. The detecting comprises identifying the gesture using the gesture data. The navigating comprises translating the gesture to a gesture signal, and navigating through the data space in response to the gesture signal. The data space is a data-representational space comprising a dataset represented in the physical space.
    Type: Application
    Filed: September 3, 2009
    Publication date: March 11, 2010
    Applicant: OBLONG INDUSTRIES, INC.
    Inventors: John S. Underkoffler, Kwindla H. Kramer
  • Publication number: 20100053304
    Abstract: Systems and methods are described for navigating through a data space. The navigating comprises detecting a gesture of a body from gesture data received via a detector. The gesture data is absolute three-space location data of an instantaneous state of the body at a point in time and physical space. The detecting comprises identifying the gesture using the gesture data. The navigating comprises translating the gesture to a gesture signal, and navigating through the data space in response to the gesture signal. The data space is a data-representational space comprising a dataset represented in the physical space.
    Type: Application
    Filed: September 3, 2009
    Publication date: March 4, 2010
    Applicant: OBLONG INDUSTRIES, INC.
    Inventors: John S. Underkoffler, Kwindla H. Kramer
  • Publication number: 20090278915
    Abstract: Systems and method for controlling vehicle subsystems using gestures are described. The controlling comprises receiving an image of an object via sensors in a vehicle and outputting gesture data representing a gesture made by the object. The object comprises at least one hand and/or finger of a user. The gesture data is an instantaneous state of the object at a point in time in space. The controlling comprises automatically detecting the gesture of the object from the gesture data by extracting positions of the object, recovering quantized poses and orientation vectors of the object, and translating the gesture data to a gesture signal. The controlling comprises managing user interactions with the vehicle subsystems in response to the gesture signal.
    Type: Application
    Filed: June 18, 2009
    Publication date: November 12, 2009
    Applicant: Oblong Industries, Inc.
    Inventors: Kwindla Hultman Kramer, John S. Underkoffler
  • Patent number: 7598942
    Abstract: The system provides a gestural interface to various visually presented elements, presented on a display screen or screens. A gestural vocabulary includes ‘instantaneous’ commands, in which forming one or both hands into the appropriate ‘pose’ results in an immediate, one-time action; and ‘spatial’ commands, in which the operator either refers directly to elements on the screen by way of literal ‘pointing’ gestures or performs navigational maneuvers by way of relative or “offset” gestures. The system contemplates the ability to identify the users hands in the form of a glove or gloves with certain indicia provided thereon, or any suitable means for providing recognizable indicia on a user's hands or body parts. A system of cameras can detect the position, orientation, and movement of the user's hands and translate that information into executable commands.
    Type: Grant
    Filed: February 8, 2006
    Date of Patent: October 6, 2009
    Assignee: Oblong Industries, Inc.
    Inventors: John S. Underkoffler, Kevin T. Parent, Kwindla H. Kramer
  • Publication number: 20090231278
    Abstract: Systems and methods are described for gesture-based control using three-dimensional information extracted over an extended depth of field. The system comprises a plurality of optical detectors coupled to at least one processor. The optical detectors image a body. At least two optical detectors of the plurality of optical detectors comprise wavefront coding cameras. The processor automatically detects a gesture of the body, wherein the gesture comprises an instantaneous state of the body. The detecting comprises aggregating gesture data of the gesture at an instant in time. The gesture data includes focus-resolved data of the body within a depth of field of the imaging system. The processor translates the gesture to a gesture signal, and uses the gesture signal to control a component coupled to the processor.
    Type: Application
    Filed: April 2, 2009
    Publication date: September 17, 2009
    Applicant: OBLONG INDUSTRIES, INC.
    Inventors: Pierre St. Hilaire, John S. Underkoffler