Patents by Inventor Walter Rolandi

Walter Rolandi has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 10992491
    Abstract: A smart home interaction system is presented. It is built on a multi-modal, multithreaded conversational dialog engine. The system provides a natural language user interface for the control of household devices, appliances or household functionality. The smart home automation agent can receive input from users through sensing devices such as a smart phone, a tablet computer or a laptop computer. Users interact with the system from within the household or from remote locations. The smart home system can receive input from sensors or any other machines with which it is interfaced. The system employs interaction guide rules for processing reaction to both user and sensor input and driving the conversational interactions that result from such input. The system adaptively learns based on both user and sensor input and can learn the preferences and practices of its users.
    Type: Grant
    Filed: July 29, 2019
    Date of Patent: April 27, 2021
    Assignee: NANT HOLDINGS IP, LLC
    Inventors: Farzad Ehsani, Silke Maren Witt-Ehsani, Walter Rolandi
  • Publication number: 20190349214
    Abstract: A smart home interaction system is presented. It is built on a multi-modal, multithreaded conversational dialog engine. The system provides a natural language user interface for the control of household devices, appliances or household functionality. The smart home automation agent can receive input from users through sensing devices such as a smart phone, a tablet computer or a laptop computer. Users interact with the system from within the household or from remote locations. The smart home system can receive input from sensors or any other machines with which it is interfaced. The system employs interaction guide rules for processing reaction to both user and sensor input and driving the conversational interactions that result from such input. The system adaptively learns based on both user and sensor input and can learn the preferences and practices of its users.
    Type: Application
    Filed: July 29, 2019
    Publication date: November 14, 2019
    Applicant: NANT HOLDINGS IP, LLC
    Inventors: Farzad Ehsani, Silke Maren Witt-Ehsani, Walter Rolandi
  • Patent number: 10367652
    Abstract: A smart home interaction system is presented. It is built on a multi-modal, multithreaded conversational dialog engine. The system provides a natural language user interface for the control of household devices, appliances or household functionality. The smart home automation agent can receive input from users through sensing devices such as a smart phone, a tablet computer or a laptop computer. Users interact with the system from within the household or from remote locations. The smart home system can receive input from sensors or any other machines with which it is interfaced. The system employs interaction guide rules for processing reaction to both user and sensor input and driving the conversational interactions that result from such input. The system adaptively learns based on both user and sensor input and can learn the preferences and practices of its users.
    Type: Grant
    Filed: February 21, 2017
    Date of Patent: July 30, 2019
    Assignee: NANT HOLDINGS IP, LLC
    Inventors: Farzad Ehsani, Silke Maren Witt-Ehsani, Walter Rolandi
  • Publication number: 20170163435
    Abstract: A smart home interaction system is presented. It is built on a multi-modal, multithreaded conversational dialog engine. The system provides a natural language user interface for the control of household devices, appliances or household functionality. The smart home automation agent can receive input from users through sensing devices such as a smart phone, a tablet computer or a laptop computer. Users interact with the system from within the household or from remote locations. The smart home system can receive input from sensors or any other machines with which it is interfaced. The system employs interaction guide rules for processing reaction to both user and sensor input and driving the conversational interactions that result from such input. The system adaptively learns based on both user and sensor input and can learn the preferences and practices of its users.
    Type: Application
    Filed: February 21, 2017
    Publication date: June 8, 2017
    Inventors: FARZAD EHSANI, SILKE MAREN WITT-EHSANI, WALTER ROLANDI
  • Patent number: 9614690
    Abstract: A smart home interaction system is presented. It is built on a multi-modal, multithreaded conversational dialog engine. The system provides a natural language user interface for the control of household devices, appliances or household functionality. The smart home automation agent can receive input from users through sensing devices such as a smart phone, a tablet computer or a laptop computer. Users interact with the system from within the household or from remote locations. The smart home system can receive input from sensors or any other machines with which it is interfaced. The system employs interaction guide rules for processing reaction to both user and sensor input and driving the conversational interactions that result from such input. The system adaptively learns based on both user and sensor input and can learn the preferences and practices of its users.
    Type: Grant
    Filed: November 23, 2015
    Date of Patent: April 4, 2017
    Assignee: NANT HOLDINGS IP, LLC
    Inventors: Farzad Ehsani, Silke Maren Witt-Ehsani, Walter Rolandi
  • Publication number: 20160080165
    Abstract: A smart home interaction system is presented. It is built on a multi-modal, multithreaded conversational dialog engine. The system provides a natural language user interface for the control of household devices, appliances or household functionality. The smart home automation agent can receive input from users through sensing devices such as a smart phone, a tablet computer or a laptop computer. Users interact with the system from within the household or from remote locations. The smart home system can receive input from sensors or any other machines with which it is interfaced. The system employs interaction guide rules for processing reaction to both user and sensor input and driving the conversational interactions that result from such input. The system adaptively learns based on both user and sensor input and can learn the preferences and practices of its users.
    Type: Application
    Filed: November 23, 2015
    Publication date: March 17, 2016
    Inventors: FARZAD EHSANI, SILKE MAREN WITT-EHSANI, WALTER ROLANDI
  • Patent number: 9230560
    Abstract: A smart home interaction system is presented. It is built on a multi-modal, multithreaded conversational dialog engine. The system provides a natural language user interface for the control of household devices, appliances or household functionality. The smart home automation agent can receive input from users through sensing devices such as a smart phone, a tablet computer or a laptop computer. Users interact with the system from within the household or from remote locations. The smart home system can receive input from sensors or any other machines with which it is interfaced. The system employs interaction guide rules for processing reaction to both user and sensor input and driving the conversational interactions that result from such input. The system adaptively learns based on both user and sensor input and can learn the preferences and practices of its users.
    Type: Grant
    Filed: October 8, 2013
    Date of Patent: January 5, 2016
    Assignee: Nant Holdings IP, LLC
    Inventors: Farzad Ehsani, Silke Maren Witt-Ehsani, Walter Rolandi
  • Publication number: 20140108019
    Abstract: A smart home interaction system is presented. It is built on a multi-modal, multithreaded conversational dialog engine. The system provides a natural language user interface for the control of household devices, appliances or household functionality. The smart home automation agent can receive input from users through sensing devices such as a smart phone, a tablet computer or a laptop computer. Users interact with the system from within the household or from remote locations. The smart home system can receive input from sensors or any other machines with which it is interfaced. The system employs interaction guide rules for processing reaction to both user and sensor input and driving the conversational interactions that result from such input. The system adaptively learns based on both user and sensor input and can learn the preferences and practices of its users.
    Type: Application
    Filed: October 8, 2013
    Publication date: April 17, 2014
    Applicant: Fluential, LLC
    Inventors: Farzad Ehsani, Silke Maren Witt-Ehsani, Walter Rolandi
  • Patent number: 8000453
    Abstract: A computer-based virtual assistant includes a virtual assistant application running on a computer capable of receiving human voice communications from a user of a remote user interface and transmitting a vocalization to the remote user interface, the virtual assistant application enabling the user to access email and voicemail messages of the user, the virtual assistant application selecting a responsive action to a verbal query or instruction received from the remote user interface and transmitting a vocalization characterizing the selected responsive action to the remote user interface, and the virtual assistant waiting a predetermined period of time, and if no canceling indication is received from the remote user interface, proceeding to perform the selected responsive action, and if a canceling indication is received from the remote user interface halting the selected responsive action and transmitting a new vocalization to the remote user interface. Also a method of using the virtual assistant.
    Type: Grant
    Filed: March 21, 2008
    Date of Patent: August 16, 2011
    Assignee: Avaya Inc.
    Inventors: Robert S. Cooper, Jeff F. McElroy, Walter Rolandi, Derek Sanders, Richard M. Ulmer, Edward Peebles
  • Publication number: 20080201306
    Abstract: A computer-based virtual assistant includes a virtual assistant application running on a computer capable of receiving human voice communications from a user of a remote user interface and transmitting a vocalization to the remote user interface, the virtual assistant application enabling the user to access email and voicemail messages of the user, the virtual assistant application selecting a responsive action to a verbal query or instruction received from the remote user interface and transmitting a vocalization characterizing the selected responsive action to the remote user interface, and the virtual assistant waiting a predetermined period of time, and if no canceling indication is received from the remote user interface, proceeding to perform the selected responsive action, and if a canceling indication is received from the remote user interface halting the selected responsive action and transmitting a new vocalization to the remote user interface. Also a method of using the virtual assistant.
    Type: Application
    Filed: March 21, 2008
    Publication date: August 21, 2008
    Inventors: Robert S. Cooper, Jeff F. McElroy, Walter Rolandi, Derek Sanders, Richard M. Ulmer, Edward Peebles
  • Patent number: 7415100
    Abstract: A computer-based virtual assistant the behavior of which can be changed by the user, comprising a voice user interface for inputting information into and receiving information from the virtual assistant by speech, a communications network, a virtual assistant application running on a remote computer, the remote computer being electronically coupled to the user interface via the communications network, wherein the behavior of the virtual assistant changes responsive to user input.
    Type: Grant
    Filed: May 4, 2004
    Date of Patent: August 19, 2008
    Assignee: Avaya Technology Corp.
    Inventors: Robert S. Cooper, Jeff F. McElroy, Walter Rolandi, Derek Sanders, Richard M. Ulmer, Edward Peebles
  • Publication number: 20040225650
    Abstract: A computer-based virtual assistant the behavior of which can be changed by the user, comprising a voice user interface for inputting information into and receiving information from the virtual assistant by speech, a communications network, a virtual assistant application running on a remote computer, the remote computer being electronically coupled to the user interface via the communications network, wherein the behavior of the virtual assistant changes responsive to user input.
    Type: Application
    Filed: May 4, 2004
    Publication date: November 11, 2004
    Applicant: Avaya Technology Corp.
    Inventors: Robert S. Cooper, Jeff F. McElroy, Walter Rolandi, Derek Sanders, Richard M. Ulmer, Edward Peebles
  • Patent number: 6757362
    Abstract: A computer-based virtual assistant the behavior of which can be changed by the user, comprising a voice user interface for inputting information into and receiving information from the virtual assistant by speech, a communications network, a virtual assistant application running on a remote computer, the remote computer being electronically coupled to the user interface via the communications network, wherein the behavior of the virtual assistant changes responsive to user input.
    Type: Grant
    Filed: March 6, 2000
    Date of Patent: June 29, 2004
    Assignee: Avaya Technology Corp.
    Inventors: Robert S. Cooper, Jeff F. McElroy, Walter Rolandi, Derek Sanders, Richard M. Ulmer, Edward Peebles
  • Patent number: 6466654
    Abstract: A computer based method for performing a command via a voice user interface on a subset of objects. The subset is selected from a set of objects, each having an object type. At least one taggable field is associated with the object type and has a corresponding value. The set of objects is stored in the computer memory. An utterance is received from the user and includes a command, an object type selection, a taggable field selection, and a value for the taggable field. Responsive to the utterance, at least one object is retrieved from the set of objects, the object of the type selected by the user and having a value in the taggable field selection that matches the taggable field value received from the user. The command is performed on the object. The object consists of text which is converted to voice output.
    Type: Grant
    Filed: March 6, 2000
    Date of Patent: October 15, 2002
    Assignee: Avaya Technology Corp.
    Inventors: Robert S. Cooper, Jeff F. McElroy, Walter Rolandi, Derek Sanders, Richard M. Ulmer