Patents by Inventor Paul Callaghan
Paul Callaghan has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Patent number: 11887264Abstract: Embodiments relate to using sensor data and location data from a device to generate augmented reality images. A mobile device pose can be determined (a geographic position, direction and a three dimensional orientation of the device) within a location. A type of destination in the location can be identified and multiple destinations can be identified, with the mobile device receiving queue information about the identified destinations from a server. A first image can be captured. Based on the queue information, one of the identified destinations can be selected. The geographic position of each identified destination can be identified, and these positions can be combined with the mobile device pose to generate a second image. Finally, an augmented reality image can be generated by combining the first image and the second image, the augmented reality image identifying the selected one destination.Type: GrantFiled: December 16, 2022Date of Patent: January 30, 2024Assignee: Live Nation Entertainment, Inc.Inventor: James Paul Callaghan
-
Publication number: 20230211090Abstract: Embodiments relate to using sensor data and location data from a device to generate augmented reality images. A mobile device pose can be determined (a geographic position, direction and a three dimensional orientation of the device) within a location. A type of destination in the location can be identified and multiple destinations can be identified, with the mobile device receiving queue information about the identified destinations from a server. A first image can be captured. Based on the queue information, one of the identified destinations can be selected. The geographic position of each identified destination can be identified, and these positions can be combined with the mobile device pose to generate a second image. Finally, an augmented reality image can be generated by combining the first image and the second image, the augmented reality image identifying the selected one destination.Type: ApplicationFiled: December 16, 2022Publication date: July 6, 2023Applicant: Live Nation Entertainment, Inc.Inventor: James Paul Callaghan
-
Patent number: 11532131Abstract: Embodiments relate to using sensor data and location data from a device to generate augmented reality images. A mobile device pose can be determined (a geographic position, direction and a three dimensional orientation of the device) within a location. A type of destination in the location can be identified and multiple destinations can be identified, with the mobile device receiving queue information about the identified destinations from a server. A first image can be captured. Based on the queue information, one of the identified destinations can be selected. The geographic position of each identified destination can be identified, and these positions can be combined with the mobile device pose to generate a second image. Finally, an augmented reality image can be generated by combining the first image and the second image, the augmented reality image identifying the selected one destination.Type: GrantFiled: February 24, 2020Date of Patent: December 20, 2022Assignee: Live Nation Entertainment, Inc.Inventor: James Paul Callaghan
-
Publication number: 20200279439Abstract: Embodiments relate to using sensor data and location data from a device to generate augmented reality images. A mobile device pose can be determined (a geographic position, direction and a three dimensional orientation of the device) within a location. A type of destination in the location can be identified and multiple destinations can be identified, with the mobile device receiving queue information about the identified destinations from a server. A first image can be captured. Based on the queue information, one of the identified destinations can be selected. The geographic position of each identified destination can be identified, and these positions can be combined with the mobile device pose to generate a second image. Finally, an augmented reality image can be generated by combining the first image and the second image, the augmented reality image identifying the selected one destination.Type: ApplicationFiled: February 24, 2020Publication date: September 3, 2020Inventor: James Paul Callaghan
-
Patent number: 10573084Abstract: Embodiments relate to using sensor data and location data from a device to generate augmented reality images. A mobile device pose can be determined (a geographic position, direction and a three dimensional orientation of the device) within a location. A type of destination in the location can be identified and multiple destinations can be identified, with the mobile device receiving queue information about the identified destinations from a server. A first image can be captured. Based on the queue information, one of the identified destinations can be selected. The geographic position of each identified destination can be identified, and these positions can be combined with the mobile device pose to generate a second image. Finally, an augmented reality image can be generated by combining the first image and the second image, the augmented reality image identifying the selected one destination.Type: GrantFiled: October 8, 2018Date of Patent: February 25, 2020Assignee: Live Nation Entertainment, Inc.Inventor: James Paul Callaghan
-
Publication number: 20190108684Abstract: Embodiments relate to using sensor data and location data from a device to generate augmented reality images. A mobile device pose can be determined (a geographic position, direction and a three dimensional orientation of the device) within a location. A type of destination in the location can be identified and multiple destinations can be identified, with the mobile device receiving queue information about the identified destinations from a server. A first image can be captured. Based on the queue information, one of the identified destinations can be selected. The geographic position of each identified destination can be identified, and these positions can be combined with the mobile device pose to generate a second image. Finally, an augmented reality image can be generated by combining the first image and the second image, the augmented reality image identifying the selected one destination.Type: ApplicationFiled: October 8, 2018Publication date: April 11, 2019Inventor: James Paul Callaghan
-
Patent number: 10096161Abstract: Embodiments relate to using sensor data and location data from a device to generate augmented reality images. A mobile device pose can be determined (a geographic position, direction and a three dimensional orientation of the device) within a location. A type of destination in the location can be identified and multiple destinations can be identified, with the mobile device receiving queue information about the identified destinations from a server. A first image can be captured. Based on the queue information, one of the identified destinations can be selected. The geographic position of each identified destination can be identified, and these positions can be combined with the mobile device pose to generate a second image. Finally, an augmented reality image can be generated by combining the first image and the second image, the augmented reality image identifying the selected one destination.Type: GrantFiled: September 8, 2015Date of Patent: October 9, 2018Assignee: Live Nation Entertainment, Inc.Inventor: James Paul Callaghan
-
Publication number: 20160071325Abstract: Embodiments relate to using sensor data and location data from a device to generate augmented reality images. A mobile device pose can be determined (a geographic position, direction and a three dimensional orientation of the device) within a location. A type of destination in the location can be identified and multiple destinations can be identified, with the mobile device receiving queue information about the identified destinations from a server. A first image can be captured. Based on the queue information, one of the identified destinations can be selected. The geographic position of each identified destination can be identified, and these positions can be combined with the mobile device pose to generate a second image. Finally, an augmented reality image can be generated by combining the first image and the second image, the augmented reality image identifying the selected one destination.Type: ApplicationFiled: September 8, 2015Publication date: March 10, 2016Inventor: James Paul Callaghan
-
Patent number: 8676615Abstract: Described are systems and methods for designing certain aspects of an event venue and for communicating information regarding the event and the event venue to others. Certain embodiments provide a dynamic seat map via which an operator can assign certain characteristics to specific seats and/or seating sections. Certain embodiments generate interactive maps for users, via which information from a plurality of sources may be integrated and visually displayed. The user may specify certain criteria, and the interactive map may identify to the user seats and/or sections that match such criteria. Certain embodiments provide an interactive seat map via which users can select seats and share information.Type: GrantFiled: November 4, 2011Date of Patent: March 18, 2014Assignee: Ticketmaster LLCInventors: James Paul Callaghan, Dennis Denker
-
Publication number: 20120323612Abstract: Described are systems and methods for designing certain aspects of an event venue and for communicating information regarding the event and the event venue to others. Certain embodiments provide a dynamic seat map via which an operator can assign certain characteristics to specific seats and/or seating sections. Certain embodiments generate interactive maps for users, via which information from a plurality of sources may be integrated and visually displayed. The user may specify certain criteria, and the interactive map may identify to the user seats and/or sections that match such criteria. Certain embodiments provide an interactive seat map via which users can select seats and share information.Type: ApplicationFiled: November 4, 2011Publication date: December 20, 2012Applicant: Ticketmaster, LLCInventor: James Paul Callaghan
-
Publication number: 20120323488Abstract: Described are systems and methods for designing certain aspects of an event venue and for communicating information regarding the event and the event venue to others. Certain embodiments provide a dynamic seat map via which an operator can assign certain characteristics to specific seats and/or seating sections. Certain embodiments generate interactive maps for users, via which information from a plurality of sources may be integrated and visually displayed. The user may specify certain criteria, and the interactive map may identify to the user seats and/or sections that match such criteria. Certain embodiments provide an interactive seat map via which users can select seats and share information.Type: ApplicationFiled: November 4, 2011Publication date: December 20, 2012Applicant: Ticketmaster, LLC.Inventor: James Paul Callaghan
-
Publication number: 20060097725Abstract: A magnetic assembly for a nuclear magnetic resonance apparatus includes a number of primary permanent magnets 1 disposed in an array about a longitudinal axis, the arrangement and/or characteristics of the plurality of magnets being such so as to create a zone of homogeneous magnetic field at some location along the axis forward of the array (and into the material when provided). A secondary permanent magnet 7 may be located along the longitudinal axis within the array of primary magnets and may be moveable. The primary magnets 1 have a north pole and a south pole with an axis therebetween, and may be arranged such that the axis between the poles is at an angle to the longitudinal axis.Type: ApplicationFiled: July 11, 2003Publication date: May 11, 2006Inventors: Paul Callaghan, Mark Hunter, Robin Dykstra, Craig Eccles
-
Publication number: 20050172459Abstract: A cord retainer (10, 100, 140, 160) for retaining cords such as electrical power leads (32) and pneumatic hoses in an elevated position above the ground or floor. The cord retainer (10, 100, 140, 160) has a flexible web (14, 146, 168) with holes (30, 148) dimensioned to receive a cord (32) and substantially retained it against longitudinal movement with respect to the web (14, 146, 168) to the extent that the cord (32) is substantially prevented from sliding longitudinally in the hole (30, 148) under the force of its own weight or during normal installation of the cord (32), in two spaced apart cord retainers (10, 100, 140, 160). A series of frames 170, 180 and 190 are also provided for attaching the cord retainer (10, 100, 140, 160) structures on building and construction sites.Type: ApplicationFiled: November 28, 2002Publication date: August 11, 2005Inventors: Martin Ryan, Michael Montgomery, Paul Callaghan