COMBINATION VIDEO AND OPTICAL SIGHT
A sight system is provided that includes an optical mode providing an optical image of a field of view and a video display mode providing video of the field of view acquired by an image sensor. The sight system can include optical components configured to present the field of view to the image sensor and to present substantially the field of view to a user through an eyepiece coupled to the sight system. The sight system can include a display configured to display to the user, through the eyepiece, video of the field of view acquired by the image sensor. The sight system can include a user interface feature configured to change from the optical mode to the video display mode when actuated by the user. Thus, the sight system can be configured to provide bore-sighted optical and video data to the user.
This application claims the benefit of priority to U.S. Prov. App'n No. 62/027,697, filed Jul. 22, 2014, entitled “Combination video and optical sight,” which is incorporated by reference herein in its entirety.
BACKGROUND FieldThe present disclosure generally relates to sight systems, and particularly to sight systems for guns.
Description of Related ArtSight systems can be used to aid a user of a gun to locate and aim at a desired target. Sight systems can be attached to the gun to provide visual feedback to the user about a field of view down range of the gun. Sight systems can include reticles and other features to aid in adjusting an aiming point so that the gun will hit the targeted object. To improve the visual feedback, sight systems can present magnified views of the field of view located down range of the gun so as to allow the user to resolve greater detail and fine-tune the aiming point of the gun.
SUMMARYThe systems, methods and devices of the disclosure each have innovative aspects, no single one of which is indispensable or solely responsible for the desirable attributes disclosed herein. Without limiting the scope of the claims, some of the advantageous features will now be summarized.
Some embodiments provide for a sight system that includes an optical mode providing an optical image of a field of view and a video display mode providing video of the field of view acquired by an image sensor. The image sensor can be adapted to detect low light levels and/or infrared radiation. The sight system can include optical components configured to present the field of view to the image sensor and to present substantially the field of view to a user through an eyepiece coupled to the sight system. The sight system can include a display configured to display to the user, through the eyepiece, video of the field of view acquired by the image sensor. The sight system can include a user interface feature configured to change from the optical mode to the video display mode when actuated by the user. Thus, the sight system can be configured to provide bore-sighted optical and video data to the user.
In some embodiments, the image sensor can be configured to be sensitive to low light levels and to short wave infrared radiation. The sight system can be used to provide video imagery from the image sensor that can combine acquired visible and infrared radiation data to increase or enhance displayed contrast between objects. The sight system can provide visual information to the user that increases or enhances contrast during a thermal transition in a diurnal cycle by combining information from the visible radiation and infrared radiation. Thus, the sight system can provide to a user bore-sighted optical images, low light level video, and infrared video.
In some embodiments, the sight system can include modular capabilities such that additional components can be attached to the sight system without obstructing the field of view of the sight system. In some embodiments, the sight system can include a radio frequency transceiver configured to transmit information from the sight system and/or modules coupled to the sight system to devices over a short range. In some embodiments, the sight system provides a magnified image of the field of view to the user. In some embodiments, the sight system includes integrated sensors such as a tilt or inertial sensor. In some embodiments, the sight system can present a reticle to the user. In some embodiments, the sight system can present a reticle with a ballistics-compensated indicator. The ballistics-compensated indicator can be based at least in part on information from integrated sensors in the sight system and a laser range finder coupled to the sight system.
The drawings are provided to illustrate example embodiments described herein and are not intended to limit the scope of the disclosure. Throughout the drawings, reference numbers may be re-used to indicate general correspondence between referenced elements.
Various aspects of the disclosure will now be described with regard to certain examples and embodiments, which are intended to illustrate but not to limit the disclosure. Nothing in this disclosure is intended to imply that any particular feature or characteristic of the disclosed embodiments is essential.
Sight Systems OverviewThe 102 housing can include module rails 108 integrated into the housing for mounting, powering, and connecting to modular add-on components which can be bore-sighted such as, for example, a laser range finder (“LRF”), a thermal imager with close combat or longer range optic, an ultraviolet (“UV”) or short-wave infrared (“SWIR”) imager, a UV or SWIR pointer or illuminator, or the like. The module rails 108 can be configured to be compatible with modules having Picatinny rails, Weaver rails, or the like. The module rails 108 can be configured to provide power to modules connected thereto through inductive means or through electrical contacts. The module rails 108 can be configured to transmit data between modules attached thereto or to send or receive data from attached modules. As described herein, the mount rails 106 or module rails 108 can include data and/or power contacts that provide electrical coupling, optical coupling, or both between the sight system 100, attached modules, and/or the system to which the sight system 100 is mounted. For example, the rails 106 or 108 can include fiber optic non-contact optical connectors to optically couple optical signals between a module and the sight system 100, between a module and another module, between a module and the device to which the sight system 100 is mounted, or between the sight system 100 and the device to which it is mounted. The module rails 108 can be integrated (e.g., cast or machined into) the housing 102 which can result in suitable alignment between the optical connectors and corresponding optical connectors on attached modules.
The sight system 100 can include a front end optical system 110 configured to provide an image of a field of view. The field of view can be at least about 2° and less than or equal to about 20°, at least about 4° and less than or equal to about 15°, at least about 6° and less than or equal to about 10°, at least about 7° and less than or equal to about 9°. The front end optical system 110 can include a reticle at a real image plane of the optical system. The front end optical system can be configured to provide a magnified image of the field of view where the magnification is at least about 1× and less than or equal to about 25×, at least about 2× and less than or equal to about 10×, at least about 3× and less than or equal to about 5×.
The sight system 100 can include an eyepiece 112 configured to provide to a user a direct-view capability where the user sees the optical image of the field of view of the front end optical system 110. The eyepiece 112 can have a field of view that is at least about 15° and less than or equal to about 40°, at least about 20° and less than or equal to about 35°, at least about 30° and less than or equal to about 34°.
The sight system can include an image sensor 114 situated within the housing 102. The image sensor 114 can be any suitable image sensor capable of converting electromagnetic radiation to electrical data. For example, the image sensor 114 can be a focal plane array, such as a CMOS image sensor, a CCD, or the like. The image sensor 114 can be a relatively high resolution (e.g., about 1 megapixel, about 2 megapixels, about 5 megapixels, or greater than about 5 megapixels), electronic-zoom-capable CMOS imager. The image sensor 114 can be configured to see the same bore-sighted image and reticle as the direct view channel (e.g., the view provided to the user by a combination of the front end optical system 110 and the eyepiece 112). The image sensor 114 and associated electronics and modules can be configured to provide greater magnification compared to the direct-view channel (e.g., through an electronic zoom functionality) and/or an image recording functionality.
The sight system 100 can include a display system (not shown) that shares the eyepiece 112 with the front end optical system such that the sight system 100 can provide a direct-view mode where the user sees the image produced by the combination of the front end optical system and the eyepiece, and a video view mode where the user sees the image acquired by the image sensor 114 and presented on the display system through the eyepiece 112. The display system can be, for example, monochrome or color and can conform to a video or resolution standard such as SXGA, VGA, HD720, HD1080, WGA, and the like. The display system can be configured to present magnified imagery from the direct view channel by displaying and magnifying image data acquired by the image sensor 114. The display system can be configured to present imagery or information from any module mounted to the module rails 108 such as a rail-mounted thermal or other spectral band camera. The display system can be configured to present a ballistics-corrected reticle which may be derived from, for example, a mounted LRF.
Thus, the sight system 100 can be configured to provide direct view sighting, video sighting, video identification, video recording, a data interface display, and the like. The sight system 100 can be configured to accept a range of other capabilities by providing a modular attachment system with the module rails 108 using a standardized electrical, data, and mechanical interface. For example, the rails can be similar to power rails manufactured by T. Worx Ventures as described in U.S. Pat. No. 7,627,975, Wilcox Industries' fusion rail system, or NATO's powered rail standard. The sight system 100 can integrate infrared functionality thereby reducing or eliminating a need for using clip-on thermal imagers which can add to the weight of the gun, alter the balance of the gun, and may be misaligned relative to the bore of the gun. In some embodiments, the sight system 100 can provide a direct view channel and a relatively high-sensitivity CMOS channel to provide sighting during diurnal intervals of low thermal contrast.
In some embodiments, the sight system 100 can include a radio frequency (“RF”) communication system 116 situated within the housing 102. The RF communication system 116 can be configured to communicate with external systems such as, for example, visualization systems, night vision goggles, data processing systems, weapons systems, computers, cellular telephones, PDAs, laptops, or other such electrical devices associated with the user or another person. The RF communication system 116 can be utilized to transmit and receive information to these other systems to integrate information from the sight system 100 to other systems. For example, the sight system 100 can be utilized in a rapid target acquisition (“RTA”) system that combines imagery and other information from the sight system 100 with information from a visualization system of a user to provide the user with a video display that shows the pointing direction of the sight system on the display of the visualization system. In this way, the user can quickly adjust an aiming point without looking through the sight system 100. In some embodiments, the RF communication system 116 can communicate using any suitable wireless communication such as through the IEEE 802.11 standard, including IEEE 802.11(a), (b), (g), or (n). In some embodiments, the RF communication system 116 communicates according to BLUETOOTH™ Specification Version 3.0+HS adopted in 2009. In some embodiments, the RF communication system 116 transmits and receives CDMA, GSM, AMPS or other known signals that are used to communicate within a wireless cell phone network. In some embodiments, the RF communication system 116 is an ultra-wide band communication system. In some embodiments, the RF communication system 116 is configured to communicate with devices that are less than about 10 m from the sight system 100. The RF communication system 116 can be configured to have a low probability of interception and/or a low probability of detection and a relatively high bandwidth.
In some embodiments, the sight system 100 can include integrated sensors to provide data to the sight system 100, to attached modules, or to external systems through the RF communication system 116. The integrated sensors can be, for example, tilt sensors, inertial sensors, accelerometers, or the like. In some embodiments, information from the integrated sensors and an attached LRF can be used to derive a ballistics-corrected reticle for display on the display system.
The sight system 100 can include a power source 118 situated within the housing 102. For example, the housing can include one or more batteries to provide power to the electrical components in the sight system 100 and/or to modules attached to the module rails 108.
The sight system 100 can be configured to be relatively light-weight compared to other sight systems providing similar functionality. For example, the sight system 100 can be configured to weigh less than or equal to about 3 lbs., less than or equal to about 2 lbs. or less than or equal to about 1.5 lbs. Based at least in part on additional capabilities provided by the sight system 100 and the significant weight reduction associated with use of rail-mounted imagers or LRF modules in place of stand-alone weapon mounted systems, the total weight of a gun or weapon system incorporating the sight system 100 can be significantly reduced.
The sight system 100 can include a reticle 304 positioned at a real image plane of front end optics 110 such that the reticle is visible to a user in a direct-view mode and is part of video acquired by the image sensor 114 and displayed by the display 310. The sight system 100 can include a reticle adjustment control 302 configured to move the position of the reticle 304.
The display 310 of the sight system 100 can be positioned at a substantially similar distance from the dichroic mirror when it is in the optical path as compared to the reticle 304. Thus, the eyepiece optics 112 can remain static or unchanged to view an image produced by the display 310 or an image produced by the front end optics 110. In some embodiments, the display 310 can be configured to display information from attached modules. The information from attached modules can be combined with the imagery acquired by the image sensor 114 or it can be displayed without any data from the image sensor 114. The modules or peripheral modules can include pointing modules (e.g., lasers or laser pointers), illumination modules (e.g., SWIR illuminators), sensor modules (e.g., accelerometers), imaging modules (e.g., thermal imagers), or the like.
The relay optics 306 of the sight system 100 can be configured to produce an image of the real image plane of the front end optics 110 at the image sensor 114. This can include an image of the reticle 304 such that the video recorded and/or displayed includes the reticle 304. By providing the bore-sighted image and/or reticle 304 to the image sensor 114, add-on modules can use the bore-sighted video, such as, for example, infrared imagers, SWIR imagers, UV imagers, IR pointers, SWIR pointers, UV pointers, LRFs, or any combination of these or other modules. The relay optics 306 can be adjusted to adjust the image plane for the image sensor 114. By changing the relay optics 306, the image plane for the image sensor 114 can be changed as well as the field of view at the image sensor, the f-number, and the like.
The front end optics 110 can have a nominal f-number of about F/3.2, and can range between about F/1.4 to about F/16. In some embodiments, the front end optics 110 are configured to support a SXGA 6.5 micron pixel LLL CMOS or SiOnyx black silicon image sensor to cover thermal transitions in the diurnal cycle. For example, the signal from the visible portion of the spectrum and the signal from the SWIR portion of the spectrum can be multiplexed to provide enhanced or increased contrast during thermal transition periods (e.g., around dawn and dusk) where thermal contrast is not sufficient to effectively distinguish objects in a scene. Thus, the sight system 100 can be used to provide effective bore-sighted visualization capabilities in many light level conditions, e.g. during day using a direct-view or SWIR imaging, during night using LLL imaging or SWIR imaging, and during diurnal thermal transition periods using LLL and/or multiplexed LLL and SWIR imaging. In some embodiments, the front end optics 110 support a commercial CMOS image sensor having greater than or equal to about 3 megapixels, greater than or equal to about 4 megapixels, or greater than or equal to about 5 megapixels. The SiOnyx sensor mentioned above can be configured to be responsive to radiation having wavelengths up to about 1.2 microns. By coupling this sensor with low-noise readouts, sight system 100 can provide about 4× additional signal. By increasing the signal, the front end optics 110 having an f-number of about F/3.2 to reach well into low light level (“LLL”) regions for thermal transition coverage as well as improved identification at quarter moon levels. The image sensor 114 can be selected to provide coverage of the SWIR electromagnetic region, thus allowing a single sight system 100 to provide direct-view capabilities, low light level capabilities, and SWIR imaging capabilities.
SWIR imaging can be useful because the SWIR band is not visible to the human eye but it is reflective like visible light. Thus, a SWIR illuminator module can be used in conjunction with the sight system 100 to cast SWIR light down range of the sight system 100 and the image sensor 114 can be configured to detect the reflected SWIR light. This can provide image data that facilitates object identification through contrast caused by shadows of objects, similar to visible light.
The optical design illustrated in
The optical design illustrated in
The LLLCMOS field of view can be configured by modifying the relay optic 306 design and CMOS package size constraints. The design of the relay optics 306 can be modified and moved to deliver a larger field of view, up to the full field of view provided by the front end optics, by using a compatible focal plane array with an appropriate active image area.
Additional examples of a site system are described herein with reference to
An example module includes a night vision goggle incorporating an electronics module for add-on use with a sight system according to some embodiments. Sight and display modules can be configured to utilize standardized interfaces that allow system capabilities, such as thermal imaging and range-finding, to be added and quickly bore-sighted to the sight system 100 in a modular form. The example night vision goggle can provide a SXGA display and 60 Hz frame rate electronic module designs that can be adapted as the basis for electronics and display to be developed for the sight system 100.
The LRF module 1400 using a version of this transmitter using about 100 micro-Joules and about a 25 mm diameter avalanche photodiode (“APD”) receiver have achieved about 3 km range against typical reflectance targets. Mounting a second element of the transmitter collimating optic in a nominal 35 mm diameter receiver optic can leave sufficient collection area for the laser return signal (including slight blocking by the transmitter folding mirrors). In some embodiments, the LRF module 1400 can be used on the sight system 100 module rails 108.
CONCLUSIONIt should be appreciated that in the above description of embodiments, various features are sometimes grouped together in a single embodiment, figure, or description thereof for the purpose of streamlining the disclosure and aiding in the understanding of one or more of the various inventive aspects. This method of disclosure, however, is not to be interpreted as reflecting an intention that any claim require more features than are expressly recited in that claim. Moreover, any components, features, or steps illustrated and/or described in a particular embodiment herein can be applied to or used with any other embodiment(s). Thus, it is intended that the scope of the inventions herein disclosed should not be limited by the particular embodiments described above, but should be determined only by a fair reading of the claims that follow.
Conditional language used herein, such as, among others, “can,” “could,” “might,” “may,” “e.g.,” and the like, unless specifically stated otherwise, or otherwise understood within the context as used, is generally intended to convey that certain embodiments include, while other embodiments do not include, certain features, elements and/or states. Thus, such conditional language is not generally intended to imply that features, elements and/or states are in any way required for one or more embodiments. As used herein, the terms “comprises,” “comprising,” “includes,” “including,” “has,” “having” or any other variation thereof, are intended to cover a non-exclusive inclusion. For example, a process, method, article, or apparatus that comprises a list of elements is not necessarily limited to only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Also, the term “or” is used in its inclusive sense (and not in its exclusive sense) so that when used, for example, to connect a list of elements, the term “or” means one, some, or all of the elements in the list. Conjunctive language such as the phrase “at least one of X, Y and Z,” unless specifically stated otherwise, is otherwise understood with the context as used in general to convey that an item, term, etc. may be either X, Y or Z. Thus, such conjunctive language is not generally intended to imply that certain embodiments require at least one of X, at least one of Y and at least one of Z each to be present.
In general, the word “module,” as used herein, refers to logic embodied in hardware or firmware, or to a collection of software instructions, possibly having entry and exit points, written in a programming language, such as, for example, Java, C or C++. A software module may be compiled and linked into an executable program, installed in a dynamic link library, or may be written in an interpreted programming language such as, for example, BASIC, Perl, or Python. It will be appreciated that software modules may be callable from other modules or from themselves, and/or may be invoked in response to detected events or interrupts. Software instructions may be embedded in firmware, such as an EPROM. It will be further appreciated that hardware modules may be comprised of connected logic units, such as gates and flip-flops, and/or may be comprised of programmable units, such as programmable gate arrays or processors. The modules described herein are preferably implemented as software modules, but may be represented in hardware or firmware. Generally, the modules described herein refer to logical modules that may be combined with other modules or divided into sub-modules despite their physical organization or storage.
The various illustrative logical blocks, modules, data structures, and processes described herein may be implemented as electronic hardware, computer software, or combinations of both. To clearly illustrate this interchangeability of hardware and software, various illustrative components, blocks, modules, and states have been described above generally in terms of their functionality. However, while the various modules are illustrated separately, they may share some or all of the same underlying logic or code. Certain of the logical blocks, modules, and processes described herein may instead be implemented monolithically.
The various illustrative logical blocks, modules, data structures, and processes described herein may be implemented or performed by a machine, such as a computer, a processor, a digital signal processor (DSP), an application specific integrated circuit (ASIC), a field programmable gate array (FPGA) or other programmable logic device, discrete gate or transistor logic, discrete hardware components, or any combination thereof designed to perform the functions described herein. A processor may be a microprocessor, a controller, a microcontroller, a state machine, combinations of the same, or the like. A processor may also be implemented as a combination of computing devices—for example, a combination of a DSP and a microprocessor, a plurality of microprocessors or processor cores, one or more graphics or stream processors, one or more microprocessors in conjunction with a DSP, or any other such configuration.
The blocks or states of the processes described herein may be embodied directly in hardware, in a software module executed by a processor, or in a combination of the two. For example, each of the processes described above may also be embodied in, and fully automated by, software modules executed by one or more machines such as computers or computer processors. A module may reside in a non-transitory computer-readable storage medium such as RAM memory, flash memory, ROM memory, EPROM memory, EEPROM memory, registers, a hard disk, a removable disk, a CD-ROM, memory capable of storing firmware, or any other form of computer-readable storage medium. An exemplary computer-readable storage medium can be coupled to a processor such that the processor can read information from, and write information to, the computer readable storage medium. In the alternative, the computer-readable storage medium may be integral to the processor. The processor and the computer-readable storage medium may reside in an ASIC.
Depending on the embodiment, certain acts, events, or functions of any of the processes or algorithms described herein can be performed in a different sequence, may be added, merged, or left out altogether. Thus, in certain embodiments, not all described acts or events are necessary for the practice of the processes. Moreover, in certain embodiments, acts or events may be performed concurrently, e.g., through multi-threaded processing, interrupt processing, or via multiple processors or processor cores, rather than sequentially.
Claims
1. A sight system configured to provide a bore-sighted optical view and a bore-sighted video view to a user, the sight system comprising:
- a housing;
- an image sensor supported within the housing;
- a display supported within the housing, the display operably coupled to an output of the image sensor to display video acquired with the image sensor;
- a front end optical system supported within the housing, the front end optical system configured to generate a real image of a field of view at a front end image location in the housing;
- an adjustable reticle positioned at the front end image location in the housing, the adjustable reticle configured to be positionable within the housing;
- an eyepiece optical system configured to provide to a user an optical image of the real image of the field of view of the front end optical system;
- an optical redirection element configured to be positionable between the front end optical system and the eyepiece optical system on an image side of the adjustable reticle;
- a relay optical system configured to: relay the real image of the field of view of the front end optical system to the image sensor, and form an image of the adjustable reticle at the image sensor;
- wherein, in a direct-view configuration, the optical redirection element is positioned so as to not redirect an optical path between the front end optical system and the eyepiece optical system thereby providing an optical view of the real image of the field of view with the adjustable reticle, and
- wherein, in a display-view configuration, the optical redirection element is positioned to redirect at least a portion of light from the front end optical system to the relay optical system so that the image sensor acquires video of the real image of the field of view with the adjustable reticle.
2. The sight system of claim 1, wherein the housing is configured to include a rail system for attaching a peripheral imaging, sensor, pointing, or illuminating module to the sight system.
3. The sight system of claim 2, wherein the rail system further includes a non-contact optical connector configured to couple optical signals between an attached module and the sight system.
4. The sight system of claim 3, wherein the display is further configured to display information from a peripheral module attached to the rail system, where information from the attached peripheral module is received through the non-contact optical connector.
5. The sight system of claim 2, wherein the rail system further includes an electrical connector configured to couple electrical signals between an attached module and the sight system.
6. The sight system of claim 1, wherein the sight system weighs less than or equal to about 2 lbs.
7. The sight system of claim 1, wherein the image sensor is configured to acquire infrared light.
8. The sight system of claim 7, wherein the image sensor is configured to acquire short wave infrared light.
9. The sight system of claim 1, wherein the image sensor comprises a low light level CMOS sensor.
10. The sight system of claim 1, wherein a field of view of the front end optical system is at least about 2 degrees and less than or equal to about 20 degrees.
11. The sight system of claim 10, wherein a field of view of the eyepiece optical system is at least about 15 degrees and less than or equal to about 40 degrees.
12. The sight system of claim 1, wherein the optical redirection element comprises a variable beamsplitter.
13. The sight system of claim 1, wherein, in the display-view configuration, the optical redirection element is further configured to provide a view of the display through the eyepiece optical system to provide a view of video acquired with the image sensor comprising the real image of the field of view with the adjustable reticle.
14. A method of providing a bore-sighted optical view and a bore-sighted video to a user, the method comprising:
- forming a real image of a field of view at a front end image location in a housing using a front end optical system supported within the housing;
- positioning an adjustable reticle at the front end image location in the housing;
- forming an optical image of the field of view of the front end optical system using an eyepiece optical system;
- selectively redirecting an optical axis using an optical redirection element configured to be positionable between the front end optical system and the eyepiece optical system on an image side of the adjustable reticle by: in a direct-view configuration, adjusting a position of the optical redirection element so as to not redirect an optical path between the front end optical system and the eyepiece optical system thereby providing an optical view of the field of view with the adjustable reticle; and in a display-view configuration, adjusting a position of the optical redirection element to redirect at least a portion of light from the front end optical system to the relay optical system so that the image sensor acquires video of the optical image of the field of view with the adjustable reticle; and
- relaying the optical image of the field of view of the front end optical system to an image sensor using a relay optical system.
15. The method of claim 14 further comprising, in the display-view configuration, providing an optical path from the display through the eyepiece optical system using the optical redirection element to provide a view of video acquired with the image sensor comprising the real image of the field of view with the adjustable reticle.
16. The method of claim 14 further comprising receiving optical signals through a non-contact optical connector configured to couple optical signals between an attached module and the sight system.
17. The method of claim 14 further comprising receiving electrical signals through an electrical connector configured to couple electrical signals between an attached module and the sight system.
18. The method of claim 14, wherein the image sensor is configured to acquire infrared light.
19. The method of claim 18, wherein the image sensor is a focal plane array.
20. The method of claim 14, wherein the image sensor is configured to acquire visible light.
Type: Application
Filed: Jun 18, 2018
Publication Date: Oct 18, 2018
Inventors: David Michael Masarik (Newport Beach, CA), Peter Denis Vallianos (Manhattan Beach, CA)
Application Number: 16/011,226