Patents by Inventor James Paul Callaghan

James Paul Callaghan has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 11887264
    Abstract: Embodiments relate to using sensor data and location data from a device to generate augmented reality images. A mobile device pose can be determined (a geographic position, direction and a three dimensional orientation of the device) within a location. A type of destination in the location can be identified and multiple destinations can be identified, with the mobile device receiving queue information about the identified destinations from a server. A first image can be captured. Based on the queue information, one of the identified destinations can be selected. The geographic position of each identified destination can be identified, and these positions can be combined with the mobile device pose to generate a second image. Finally, an augmented reality image can be generated by combining the first image and the second image, the augmented reality image identifying the selected one destination.
    Type: Grant
    Filed: December 16, 2022
    Date of Patent: January 30, 2024
    Assignee: Live Nation Entertainment, Inc.
    Inventor: James Paul Callaghan
  • Publication number: 20230211090
    Abstract: Embodiments relate to using sensor data and location data from a device to generate augmented reality images. A mobile device pose can be determined (a geographic position, direction and a three dimensional orientation of the device) within a location. A type of destination in the location can be identified and multiple destinations can be identified, with the mobile device receiving queue information about the identified destinations from a server. A first image can be captured. Based on the queue information, one of the identified destinations can be selected. The geographic position of each identified destination can be identified, and these positions can be combined with the mobile device pose to generate a second image. Finally, an augmented reality image can be generated by combining the first image and the second image, the augmented reality image identifying the selected one destination.
    Type: Application
    Filed: December 16, 2022
    Publication date: July 6, 2023
    Applicant: Live Nation Entertainment, Inc.
    Inventor: James Paul Callaghan
  • Patent number: 11532131
    Abstract: Embodiments relate to using sensor data and location data from a device to generate augmented reality images. A mobile device pose can be determined (a geographic position, direction and a three dimensional orientation of the device) within a location. A type of destination in the location can be identified and multiple destinations can be identified, with the mobile device receiving queue information about the identified destinations from a server. A first image can be captured. Based on the queue information, one of the identified destinations can be selected. The geographic position of each identified destination can be identified, and these positions can be combined with the mobile device pose to generate a second image. Finally, an augmented reality image can be generated by combining the first image and the second image, the augmented reality image identifying the selected one destination.
    Type: Grant
    Filed: February 24, 2020
    Date of Patent: December 20, 2022
    Assignee: Live Nation Entertainment, Inc.
    Inventor: James Paul Callaghan
  • Publication number: 20200279439
    Abstract: Embodiments relate to using sensor data and location data from a device to generate augmented reality images. A mobile device pose can be determined (a geographic position, direction and a three dimensional orientation of the device) within a location. A type of destination in the location can be identified and multiple destinations can be identified, with the mobile device receiving queue information about the identified destinations from a server. A first image can be captured. Based on the queue information, one of the identified destinations can be selected. The geographic position of each identified destination can be identified, and these positions can be combined with the mobile device pose to generate a second image. Finally, an augmented reality image can be generated by combining the first image and the second image, the augmented reality image identifying the selected one destination.
    Type: Application
    Filed: February 24, 2020
    Publication date: September 3, 2020
    Inventor: James Paul Callaghan
  • Patent number: 10573084
    Abstract: Embodiments relate to using sensor data and location data from a device to generate augmented reality images. A mobile device pose can be determined (a geographic position, direction and a three dimensional orientation of the device) within a location. A type of destination in the location can be identified and multiple destinations can be identified, with the mobile device receiving queue information about the identified destinations from a server. A first image can be captured. Based on the queue information, one of the identified destinations can be selected. The geographic position of each identified destination can be identified, and these positions can be combined with the mobile device pose to generate a second image. Finally, an augmented reality image can be generated by combining the first image and the second image, the augmented reality image identifying the selected one destination.
    Type: Grant
    Filed: October 8, 2018
    Date of Patent: February 25, 2020
    Assignee: Live Nation Entertainment, Inc.
    Inventor: James Paul Callaghan
  • Publication number: 20190108684
    Abstract: Embodiments relate to using sensor data and location data from a device to generate augmented reality images. A mobile device pose can be determined (a geographic position, direction and a three dimensional orientation of the device) within a location. A type of destination in the location can be identified and multiple destinations can be identified, with the mobile device receiving queue information about the identified destinations from a server. A first image can be captured. Based on the queue information, one of the identified destinations can be selected. The geographic position of each identified destination can be identified, and these positions can be combined with the mobile device pose to generate a second image. Finally, an augmented reality image can be generated by combining the first image and the second image, the augmented reality image identifying the selected one destination.
    Type: Application
    Filed: October 8, 2018
    Publication date: April 11, 2019
    Inventor: James Paul Callaghan
  • Patent number: 10096161
    Abstract: Embodiments relate to using sensor data and location data from a device to generate augmented reality images. A mobile device pose can be determined (a geographic position, direction and a three dimensional orientation of the device) within a location. A type of destination in the location can be identified and multiple destinations can be identified, with the mobile device receiving queue information about the identified destinations from a server. A first image can be captured. Based on the queue information, one of the identified destinations can be selected. The geographic position of each identified destination can be identified, and these positions can be combined with the mobile device pose to generate a second image. Finally, an augmented reality image can be generated by combining the first image and the second image, the augmented reality image identifying the selected one destination.
    Type: Grant
    Filed: September 8, 2015
    Date of Patent: October 9, 2018
    Assignee: Live Nation Entertainment, Inc.
    Inventor: James Paul Callaghan
  • Publication number: 20160071325
    Abstract: Embodiments relate to using sensor data and location data from a device to generate augmented reality images. A mobile device pose can be determined (a geographic position, direction and a three dimensional orientation of the device) within a location. A type of destination in the location can be identified and multiple destinations can be identified, with the mobile device receiving queue information about the identified destinations from a server. A first image can be captured. Based on the queue information, one of the identified destinations can be selected. The geographic position of each identified destination can be identified, and these positions can be combined with the mobile device pose to generate a second image. Finally, an augmented reality image can be generated by combining the first image and the second image, the augmented reality image identifying the selected one destination.
    Type: Application
    Filed: September 8, 2015
    Publication date: March 10, 2016
    Inventor: James Paul Callaghan
  • Patent number: 8676615
    Abstract: Described are systems and methods for designing certain aspects of an event venue and for communicating information regarding the event and the event venue to others. Certain embodiments provide a dynamic seat map via which an operator can assign certain characteristics to specific seats and/or seating sections. Certain embodiments generate interactive maps for users, via which information from a plurality of sources may be integrated and visually displayed. The user may specify certain criteria, and the interactive map may identify to the user seats and/or sections that match such criteria. Certain embodiments provide an interactive seat map via which users can select seats and share information.
    Type: Grant
    Filed: November 4, 2011
    Date of Patent: March 18, 2014
    Assignee: Ticketmaster LLC
    Inventors: James Paul Callaghan, Dennis Denker
  • Publication number: 20120323612
    Abstract: Described are systems and methods for designing certain aspects of an event venue and for communicating information regarding the event and the event venue to others. Certain embodiments provide a dynamic seat map via which an operator can assign certain characteristics to specific seats and/or seating sections. Certain embodiments generate interactive maps for users, via which information from a plurality of sources may be integrated and visually displayed. The user may specify certain criteria, and the interactive map may identify to the user seats and/or sections that match such criteria. Certain embodiments provide an interactive seat map via which users can select seats and share information.
    Type: Application
    Filed: November 4, 2011
    Publication date: December 20, 2012
    Applicant: Ticketmaster, LLC
    Inventor: James Paul Callaghan
  • Publication number: 20120323488
    Abstract: Described are systems and methods for designing certain aspects of an event venue and for communicating information regarding the event and the event venue to others. Certain embodiments provide a dynamic seat map via which an operator can assign certain characteristics to specific seats and/or seating sections. Certain embodiments generate interactive maps for users, via which information from a plurality of sources may be integrated and visually displayed. The user may specify certain criteria, and the interactive map may identify to the user seats and/or sections that match such criteria. Certain embodiments provide an interactive seat map via which users can select seats and share information.
    Type: Application
    Filed: November 4, 2011
    Publication date: December 20, 2012
    Applicant: Ticketmaster, LLC.
    Inventor: James Paul Callaghan