Patents by Inventor James Paul Callaghan
James Paul Callaghan has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Patent number: 11887264Abstract: Embodiments relate to using sensor data and location data from a device to generate augmented reality images. A mobile device pose can be determined (a geographic position, direction and a three dimensional orientation of the device) within a location. A type of destination in the location can be identified and multiple destinations can be identified, with the mobile device receiving queue information about the identified destinations from a server. A first image can be captured. Based on the queue information, one of the identified destinations can be selected. The geographic position of each identified destination can be identified, and these positions can be combined with the mobile device pose to generate a second image. Finally, an augmented reality image can be generated by combining the first image and the second image, the augmented reality image identifying the selected one destination.Type: GrantFiled: December 16, 2022Date of Patent: January 30, 2024Assignee: Live Nation Entertainment, Inc.Inventor: James Paul Callaghan
-
Publication number: 20230211090Abstract: Embodiments relate to using sensor data and location data from a device to generate augmented reality images. A mobile device pose can be determined (a geographic position, direction and a three dimensional orientation of the device) within a location. A type of destination in the location can be identified and multiple destinations can be identified, with the mobile device receiving queue information about the identified destinations from a server. A first image can be captured. Based on the queue information, one of the identified destinations can be selected. The geographic position of each identified destination can be identified, and these positions can be combined with the mobile device pose to generate a second image. Finally, an augmented reality image can be generated by combining the first image and the second image, the augmented reality image identifying the selected one destination.Type: ApplicationFiled: December 16, 2022Publication date: July 6, 2023Applicant: Live Nation Entertainment, Inc.Inventor: James Paul Callaghan
-
Patent number: 11532131Abstract: Embodiments relate to using sensor data and location data from a device to generate augmented reality images. A mobile device pose can be determined (a geographic position, direction and a three dimensional orientation of the device) within a location. A type of destination in the location can be identified and multiple destinations can be identified, with the mobile device receiving queue information about the identified destinations from a server. A first image can be captured. Based on the queue information, one of the identified destinations can be selected. The geographic position of each identified destination can be identified, and these positions can be combined with the mobile device pose to generate a second image. Finally, an augmented reality image can be generated by combining the first image and the second image, the augmented reality image identifying the selected one destination.Type: GrantFiled: February 24, 2020Date of Patent: December 20, 2022Assignee: Live Nation Entertainment, Inc.Inventor: James Paul Callaghan
-
Publication number: 20200279439Abstract: Embodiments relate to using sensor data and location data from a device to generate augmented reality images. A mobile device pose can be determined (a geographic position, direction and a three dimensional orientation of the device) within a location. A type of destination in the location can be identified and multiple destinations can be identified, with the mobile device receiving queue information about the identified destinations from a server. A first image can be captured. Based on the queue information, one of the identified destinations can be selected. The geographic position of each identified destination can be identified, and these positions can be combined with the mobile device pose to generate a second image. Finally, an augmented reality image can be generated by combining the first image and the second image, the augmented reality image identifying the selected one destination.Type: ApplicationFiled: February 24, 2020Publication date: September 3, 2020Inventor: James Paul Callaghan
-
Patent number: 10573084Abstract: Embodiments relate to using sensor data and location data from a device to generate augmented reality images. A mobile device pose can be determined (a geographic position, direction and a three dimensional orientation of the device) within a location. A type of destination in the location can be identified and multiple destinations can be identified, with the mobile device receiving queue information about the identified destinations from a server. A first image can be captured. Based on the queue information, one of the identified destinations can be selected. The geographic position of each identified destination can be identified, and these positions can be combined with the mobile device pose to generate a second image. Finally, an augmented reality image can be generated by combining the first image and the second image, the augmented reality image identifying the selected one destination.Type: GrantFiled: October 8, 2018Date of Patent: February 25, 2020Assignee: Live Nation Entertainment, Inc.Inventor: James Paul Callaghan
-
Publication number: 20190108684Abstract: Embodiments relate to using sensor data and location data from a device to generate augmented reality images. A mobile device pose can be determined (a geographic position, direction and a three dimensional orientation of the device) within a location. A type of destination in the location can be identified and multiple destinations can be identified, with the mobile device receiving queue information about the identified destinations from a server. A first image can be captured. Based on the queue information, one of the identified destinations can be selected. The geographic position of each identified destination can be identified, and these positions can be combined with the mobile device pose to generate a second image. Finally, an augmented reality image can be generated by combining the first image and the second image, the augmented reality image identifying the selected one destination.Type: ApplicationFiled: October 8, 2018Publication date: April 11, 2019Inventor: James Paul Callaghan
-
Patent number: 10096161Abstract: Embodiments relate to using sensor data and location data from a device to generate augmented reality images. A mobile device pose can be determined (a geographic position, direction and a three dimensional orientation of the device) within a location. A type of destination in the location can be identified and multiple destinations can be identified, with the mobile device receiving queue information about the identified destinations from a server. A first image can be captured. Based on the queue information, one of the identified destinations can be selected. The geographic position of each identified destination can be identified, and these positions can be combined with the mobile device pose to generate a second image. Finally, an augmented reality image can be generated by combining the first image and the second image, the augmented reality image identifying the selected one destination.Type: GrantFiled: September 8, 2015Date of Patent: October 9, 2018Assignee: Live Nation Entertainment, Inc.Inventor: James Paul Callaghan
-
Publication number: 20160071325Abstract: Embodiments relate to using sensor data and location data from a device to generate augmented reality images. A mobile device pose can be determined (a geographic position, direction and a three dimensional orientation of the device) within a location. A type of destination in the location can be identified and multiple destinations can be identified, with the mobile device receiving queue information about the identified destinations from a server. A first image can be captured. Based on the queue information, one of the identified destinations can be selected. The geographic position of each identified destination can be identified, and these positions can be combined with the mobile device pose to generate a second image. Finally, an augmented reality image can be generated by combining the first image and the second image, the augmented reality image identifying the selected one destination.Type: ApplicationFiled: September 8, 2015Publication date: March 10, 2016Inventor: James Paul Callaghan
-
Patent number: 8676615Abstract: Described are systems and methods for designing certain aspects of an event venue and for communicating information regarding the event and the event venue to others. Certain embodiments provide a dynamic seat map via which an operator can assign certain characteristics to specific seats and/or seating sections. Certain embodiments generate interactive maps for users, via which information from a plurality of sources may be integrated and visually displayed. The user may specify certain criteria, and the interactive map may identify to the user seats and/or sections that match such criteria. Certain embodiments provide an interactive seat map via which users can select seats and share information.Type: GrantFiled: November 4, 2011Date of Patent: March 18, 2014Assignee: Ticketmaster LLCInventors: James Paul Callaghan, Dennis Denker
-
Publication number: 20120323612Abstract: Described are systems and methods for designing certain aspects of an event venue and for communicating information regarding the event and the event venue to others. Certain embodiments provide a dynamic seat map via which an operator can assign certain characteristics to specific seats and/or seating sections. Certain embodiments generate interactive maps for users, via which information from a plurality of sources may be integrated and visually displayed. The user may specify certain criteria, and the interactive map may identify to the user seats and/or sections that match such criteria. Certain embodiments provide an interactive seat map via which users can select seats and share information.Type: ApplicationFiled: November 4, 2011Publication date: December 20, 2012Applicant: Ticketmaster, LLCInventor: James Paul Callaghan
-
Publication number: 20120323488Abstract: Described are systems and methods for designing certain aspects of an event venue and for communicating information regarding the event and the event venue to others. Certain embodiments provide a dynamic seat map via which an operator can assign certain characteristics to specific seats and/or seating sections. Certain embodiments generate interactive maps for users, via which information from a plurality of sources may be integrated and visually displayed. The user may specify certain criteria, and the interactive map may identify to the user seats and/or sections that match such criteria. Certain embodiments provide an interactive seat map via which users can select seats and share information.Type: ApplicationFiled: November 4, 2011Publication date: December 20, 2012Applicant: Ticketmaster, LLC.Inventor: James Paul Callaghan