INDICIA-READING SYSTEMS HAVING AN INTERFACE WITH A USER'S NERVOUS SYSTEM

Indicia-reading systems that interface with a user's nervous system include a device with electrodes capable of detecting electromagnetic signals produced in the brain or skeletal muscles of a user. The systems also include a computer with a processor and memory. The computer is configured to monitor the electromagnetic signals that are detected by the electrodes. The computer is also configured to perform operations in response to certain monitored electromagnetic signals. The computer may be an indicia-reading device configured to acquire indicia information in response to certain detected electromagnetic signals. The computer may also be a vehicle-mounted computer configured to sound an alarm in response to certain detected electromagnetic signals.

Skip to: Description  ·  Claims  · Patent History  ·  Patent History
Description
FIELD OF THE INVENTION

The present invention relates to the field of indicia-reading systems and, more specifically, to indicia-reading systems that interface with a user's nervous system.

BACKGROUND

Indicia readers, such as barcode scanners, are typically configured to acquire information from indicia and then decode that information for use in data systems. Traditional indicia-reading systems embrace various kinds of devices used to read indicia, including handheld barcode scanners.

Handheld indicia-reading devices, such as handheld barcode scanners and mobile computers, are currently used in numerous environments for various applications (e.g., warehouses, delivery vehicles, hospitals, etc.). In this regard, a large percentage of retailers, notably grocery stores and general consumer merchandisers, currently rely on barcode technology to improve the efficiency and reliability of the checkout process. Traditionally, a user interacts with a handheld indicia-reading device via a trigger or a touchscreen display.

More recently, wearable computing devices (e.g., GOOGLE GLASS™ from Google, Inc.) have been developed. Wearable computing devices may be used in indicia-reading systems. As these types of devices become more common, the options through which users can interface with these devices and systems will change and expand as the demand for hands-free interface grows stronger.

Current hands-free interface options for computing systems include gesture optical recognition (i.e., mathematical interpretation of human motion by a computing device). Gesture recognition can originate from any bodily motion or state, but commonly originates from the hands. Gesture interface provides a useful building block for a hands-free interface, but does not offer a completely hands-free experience as it is actually more of a touch-free interface that still requires free hands for gesturing.

A technology that does offer the possibility of a completely hands-free and touch-free interface is the brain-computer interface. For example, electroencephalography (EEG) can be used to detect electrical activity in the brain. Traditional EEG testing in a medical or laboratory environment involves flat metal discs (electrodes) attached directly to the scalp to measure the electrical activity of the brain (i.e., to measure brain waves). Traditional EEG testing equipment is inadequate for more mainstream applications, however, because it involves equipment that requires shaving the head, affixing gelled electrodes to the scalp, etc.

Recent advances in EEG, however, open the ability to read electrical signals produced by the brain to more mainstream applications. For instance, companies such as Emotiv Systems, an Australian electronics company, have brought EEG devices to market that do not require shaving a user's head or gels of any kind to measure the electrical activity of the brain. One such device is the EMOTIVE INSIGHT™ from Emotiv Systems.

Another technology that opens the possibility to facilitate a hands-free or touch-free interface without the requirement of optical recognition of gestures is electromyography (EMG). EMG is a technique for evaluating and recording the electrical activity produced by skeletal muscles. EMG is performed using an instrument called an electromyograph to produce a record of activity called an electromyogram.

Recent advances in EMG have opened the ability to read electrical signals produced by skeletal muscles to more mainstream applications. Companies such as Thalmic Labs, Inc. of Ontario Canada have brought commercial EMG devices to market that are unobtrusive for a user to wear. These devices can connect wirelessly (via, for example, BLUETOOTH® protocols) to most modern day devices.

While traditional methods of user interaction with indicia-reading devices (such as via a trigger or touchscreen interface) are generally effective, the effectiveness of such traditional methods is not completely hands-free or touch-free.

Therefore, a need exists for more efficient and effective user interfaces for indicia-reading systems, including but not limited to indicia-reading systems that interface with a user's nervous system.

SUMMARY

Accordingly, in one aspect, the present invention embraces an indicia-reading system having an interface with a user's nervous system. The system may include a headset with electrodes capable of detecting electromagnetic signals produced in the brain of a user. The system may also include a indicia reader in communication with the headset, including a central processing unit and memory, an indicia capturing subsystem for acquiring information about indicia within the indicia-capturing subsystem's field of view, and an indicia-decoding subsystem configured for decoding indicia information acquired by the indicia-capturing subsystem. The indicia reader may be configured to monitor the electromagnetic signals detected by the headset.

In an exemplary embodiment, the indicia-reading system may include an indicia reader configured to perform an operation in response to electromagnetic signals detected by the headset.

In another exemplary embodiment, the indicia reader operation that may be performed in response to electromagnetic signals detected by the headset is acquiring information about indicia within the indicia-capturing subsystem's field of view.

In yet another exemplary embodiment, the indicia reader operation that may be performed in response to electromagnetic signals detected by the headset is placing the indicia reader into a different mode.

In yet another exemplary embodiment, the detected signals produced in the brain of the user may correspond to a facial expression.

In yet another exemplary embodiment, the detected signals produced in the brain of the user may correspond to a wink.

In yet another exemplary embodiment, the detected signals produced in the brain of the user may correspond to mental commands.

In yet another exemplary embodiment, the communication between the headset and the indicia reader may be wireless communication.

In yet another exemplary embodiment, the indicia reader may be a wearable computer.

In another aspect, the present invention may include an indicia-reading system having an interface with a user's nervous system including a band comprising electrodes capable of detecting electromagnetic signals produced in the skeletal muscles of a user. The system may also include a indicia reader in communication with the band, comprising a central processing unit and memory, an indicia capturing subsystem for acquiring information about indicia within the indicia-capturing subsystem's field of view, and an indicia-decoding subsystem configured for decoding indicia information acquired by the indicia-capturing subsystem. The indicia reader may be configured to monitor the electromagnetic signals detected by the band.

In an exemplary embodiment, the indicia reader may be configured to perform an operation in response to electromagnetic signals detected by the band.

In another exemplary embodiment, the indicia reader operation in response to electromagnetic signals detected by the band may be acquiring information about indicia within the indicia-capturing subsystem's field of view.

In yet another exemplary embodiment, the indicia reader operation in response to electromagnetic signals detected by the band may be placing the indicia reader into a different scanning mode.

In yet another exemplary embodiment, the detected signals produced in the skeletal muscles of a user may correspond to arm or hand gestures.

In yet another exemplary embodiment, the detected gesture may be a snap of two fingers.

In yet another exemplary embodiment, the detected gesture may be a clenched fist.

In yet another exemplary embodiment, the detected gesture may be a combination of hand, finger, or arm movements.

In yet another exemplary embodiment, the band may be an arm band configured to be worn on the user's forearm.

In another aspect, the present invention may include a vehicle safety system having an interface with a user's nervous system including a headset with electrodes capable of detecting electromagnetic signals produced in the brain of a user. The system may also include a vehicle computer including a central processing unit and memory in communication with the headset. The vehicle computer may be configured to monitor the electromagnetic signals detected by the headset.

In an exemplary embodiment, the vehicle computer may be configured to perform an operation in response to electromagnetic signals detected by the headset.

The foregoing illustrative summary, as well as other exemplary objectives and/or advantages of the invention, and the manner in which the same are accomplished, are further explained within the following detailed description and its accompanying drawings.

BRIEF DESCRIPTION OF THE DRAWINGS

FIG. 1 graphically depicts a user wearing certain components of an exemplary indicia-reading system according to the present invention.

FIG. 2 graphically depicts a user wearing certain components of another exemplary indicia-reading system according to the present invention.

FIG. 3 is a block diagram illustrating certain components of an exemplary indicia-reading system according to the present invention.

FIG. 4 is a block diagram illustrating certain components of an exemplary system that interfaces with a user's nervous system according to the present invention.

DETAILED DESCRIPTION

The present invention embraces systems that interface with a user's nervous system. In particular, the present invention embraces hands-free indicia-reading systems that interface with a user's nervous system. Although indicia-reading systems are typically referred to herein, a person having skill in the art will recognize that the systems that interact with a user's nervous system may be utilized in other environments as set forth herein (e.g., for use with vehicle safety systems).

The term indicia as used herein is intended to refer broadly to various types of machine-readable indicia, including barcodes, QR codes, matrix codes, 1D codes, 2D codes, RFID tags, characters, etc. The indicia are typically graphical representations of information (e.g., data) such as product numbers, package tracking numbers, or personnel identification numbers. The use of indicia readers to input data into a system, rather than manual data entry, results in generally faster and more reliable data entry.

An exemplary indicia-reading system according to the present invention may include an electroencephalogram in the form of a headset that a user will wear and an indicia-reading device in electronic communication with the headset. When certain brainwave activity is detected by the headset, the system is configured to trigger operations of the indicia reader.

In another exemplary embodiment, an indicia-reading system according to the present invention may include an electromyograph band that a user may wear on their arm and an indicia reading device in electronic communication with the band. When certain skeletal muscle activity is detected by the band, the system is configured trigger operations of the indicia reader.

Non-limiting examples of typical indicia-reading devices may include handheld computers, handheld scanners, wearable computers, and similar products. Preferably, a wearable computer may be used in the exemplary embodiments disclosed herein for ease of user interface. References in the disclosure to particular types of devices are not intended to limit the disclosure to particular devices.

Referring now to the drawings, FIG. 1 depicts a user 101 wearing an exemplary indicia-reading system 100 having an interface with a user's nervous system, specifically an indicia-reading system having an interface with a user's brain.

The exemplary indicia-reading system 100 includes an indicia-reading wearable computer 102 (e.g., GOOGLE GLASS™ from Google, Inc.). Although a certain type of wearable computer 102 is depicted, various types of wearables or other kinds of devices that read indicia may alternatively be used (e.g., hand-held indicia readers such as trigger-type readers and mobile computing devices like smartphones).

The wearable computer 102 of the exemplary indicia-reading system 100 may include an indicia-capturing subsystem 103 (FIG. 1 and FIG. 3). In some instances, indicia-capturing subsystem 103 may include laser scanning subsystems that sweep light beams (e.g., a laser beam) across a scan path (i.e., a field of view), and then receive the optical signals that reflect or scatter off the indicium. Typically, in this type of embodiment the optical signal is received using a photoreceptor (e.g., photodiode) and is converted into an electrical signal. The electrical signal is an electronic representation of the indicia information (e.g., the data represented by the indicia). When in the form of an electrical signal, this information can be processed (e.g., decoded) by an indicia-decoding subsystem 104.

In other instances, the indicia-capturing subsystem 103 (FIG. 1 and FIG. 3) may include an imaging subsystem (e.g., the built-in camera of a smartphone, tablet, or wearable computer such as GOGGLE GLASS™) or some combination of an imaging subsystem and a laser scanning subsystem. The imaging subsystem captures digital images of objects within the subsystem's field of view 124 (FIG. 3) (e.g., 1D, 2D, and Postal barcodes).

When the indicia information takes the form of a digital image, the indicia information is typically processed by an indicia-decoding subsystem 104 (FIG. 3) through the use of image-processing software (e.g., optical character recognition (OCR) technology), which can both identify the presence of indicia in the digital image and decode the indicia. The components of indicia-decoding subsystem 104 are known in the art and may include a storage memory 118 for transmitting the signal to a central processing unit (CPU) 117 for processing digital signals. The exemplary wearable computer 102 may also include random access memory (RAM) 119, a read only memory (ROM) 120, and a mass storage device 121 (e.g., flash memory, a hard drive, etc.) with an operating system 122 and applications programs 123 (FIG. 3).

The exemplary indicia-reading system 100 also includes a EEG headset 105 (e.g., the EMOTIVE INSIGHT™ from Emotiv Systems, NEUROSKY® EEG biosensor from NeuroSky of San Jose, Calif., or similar devices) with a number of electrodes 106 capable of detecting signals produced in the brain of a user 101. For instance, EEG headset 105 is capable of producing a graph measurement of a user's 101 brain waves. The electrodes 106 are, for example, disks that conduct electrical activity, capture it from the brain, and convey it out through an amplifier.

As EEG technology has progressed, researchers (e.g., researchers at Emotiv Systems) have applied the technology to create high-fidelity brain computer interface systems that can read and interpret conscious and non-conscious thoughts as well as emotions. In this regard, the electrodes 106 of the exemplary indicia-reading system 100 can be used to record the resulting brain waves during a user's 101 concentration. Thereafter, the electrical activity of the user's 101 brain waves can be correlated based upon the recorded pattern to, for example, the user's 101 state of mind or, for example, to when the user 101 performs a facial expression (e.g., a wink, a smile, a frown, etc.).

A communication module pair 107A, 107B may be included respectively in the wearable computer 102 and the headset 105 of the exemplary indicia-reading system 100 for data communication. The wireless communication may include, but is not limited to, ZIGBEE® and BLUETOOTH® protocols. Although wireless communication is preferred (e.g., to provide the user with a greater range of motion), a wired connection may also be used.

Through the interface between the headset 105 and the wearable computer 102, EEG brainwave activity can be communicated in near real-time. For example, a software application program 123 running on the wearable computer 102 can monitor the user's 101 brainwave activity. The wearable computer 102 can be configured to trigger a scan event to the indicia-capturing subsystem 103 when the triggering event is detected using a software application program (such as, for example, SWIFTDECODER MOBILE™ barcode decoding software from Honeywell International, Inc.). By way of example, the relevant events to trigger a scan event to the indicia-capturing subsystem 103 may include a facial gesture such as a strong blink by the user 101, or a mental command such as when the user 101 focuses intensely on a particular location or imagines pushing a barcode away.

In addition to a trigger for a scan event to the indicia-capturing subsystem 103, mental commands or gesture commands can also be used to trigger any other operation feature in the indicia-reader 102 such as putting it into a different mode (e.g., presentation scanning), turning on and off the indicia-reader's illumination feature, or any other feature that the indicia-reader supports.

The software programs 123 can also, for example, be configured to recognize the direction a user 101 is looking in order to determine which indicia to return to the indicia-decoding subsystem 104 when multiple indicia are present in the field of view 124 (FIG. 3). The trigger command may be used to initially begin the scan event, and all indicia in the field of view 124 could then be captured by the indicia-capturing subsystem 103 and decoded by the indicia-decoding subsystem 104. If multiple indicia were present in the field of view, the indicia closest to the direction in which the user 101 was viewing can be captured by the indicia-capturing subsystem 103 and returned to the indicia-decoding subsystem 104.

FIG. 2 depicts certain components of another exemplary indicia-reading system 200 according to the present invention. The components of the exemplary system 200 depicted at FIG. 2 also provide an interface with a user's nervous system, but rather than interfacing with a user's brain as with system 100, the indicia-reading system 200 interfaces with a user's skeletal muscle activity.

A person having skill in the art will recognize that the relevant discussion with regard to the interface with a user's brain described above and depicted at FIGS. 1 and 3 is applicable to the indicia-reading system 200 that interacts with a user's skeletal muscle activity. Referring to the indicia reading system 200, the system may have elements 202, 203, 204, 205, 206, 207A, and 207B, which operate in a similar manner as corresponding elements 102, 103, 104, 105, 106, 107A, and 107B of indicia-reading system 100.

The exemplary indicia-reading system 200 includes an electromyography (EMG) band 205 that a user 201 may wear on their forearm (e.g., the MYO™ EEG arm-band developed by Thalmic Labs, Inc. of Ontario Canada or related devices). Similar to the discussion regarding EEG technology above, researchers (e.g., researchers at Thalmic Labs) have applied EMG technology to read the electrical activity of a user's muscles to allow for control of a device. In this regard, the band 205 contains a number of electrodes 206 that can read the electrical activity of a user's muscles.

An indicia-reading device, for example wearable computer 202, is in electronic communication to the band 205. The communication channels may be wired or wireless, but preferably includes wireless communication using a wireless communication module 207A, 207B.

When certain skeletal muscle activity is detected by the band 205, the system 200 is configured trigger operations of an indicia reader 202 using hardware and software programs of the type described above with reference to FIGS. 1 and 3. The trigger event may be based upon the detected electrical activity of the user's muscles such as when the user 202 performs an arm or hand gesture. The trigger may include, for example, the snap of two fingers, the rotation of the arm, clench of the fist, touching of two fingers, or various other combinations of hand, finger, or arm movements.

In addition to a trigger for a scan event, muscular activity commands can also be used to trigger other operations in the indicia-reader (e.g., wearable computer 202) such as putting it into a different mode (e.g., presentation scanning), tuning and off the indicia-reader's illumination feature, or any other feature that the indicia-reader supports. The EMG band 205 could also be used to holster an arm mounted device/computer in addition to providing a gesture recognition system.

In another exemplary embodiment, systems that interface with a user's nervous system may be utilized to control or monitor vehicles such as forklifts, cranes, delivery trucks and similar industrial vehicles (e.g., vehicles used in industrial operations, factory or warehouse settings, and the like). References in the disclosure to particular types of vehicles are not intended to limit the disclosure to particular vehicles.

FIG. 4 is a block diagram illustrating certain components of an exemplary system 300 that interfaces with a user's nervous system and that may be utilized to control or monitor vehicles, such as forklifts, according to the present invention. The exemplary system 300 is related to detecting or preventing a safety related incident quickly before or after it has occurred.

Some vehicle safety systems may use inertial sensors, cameras, or other sensors to detect safety-related events. The exemplary system 300 utilizes a user's brain response to an incident to trigger a notification/alarm or responsive action by the vehicle. Such events may include, but are not limited to, the imminent collision of a forklift and a person, an operator that is losing focus on a particular task at hand, or a driver falling asleep at the wheel. The system 300 is related to detecting these events and their warning signs. Further, system 300 may be utilized to prevent the occurrence of safety incidents.

The exemplary system 300 includes an EEG headset 305 which may be of the type described above with regard to indicia-reading system 100. The headset 305 includes electrodes 306 that conduct electrical activity, capture it from the brain of a user 301, and convey it out through an amplifier. A communication module 307 may be included for data communication.

The system 300 may also include a vehicle computer 320 which may be mounted within the applicable vehicle. Rather than a vehicle-mounted computer, other computing devices may alternatively be used (e.g., wearable or handheld computing devices). Exemplary vehicle computer 320 includes a mass storage device 340 (e.g., a solid state drive, optical drive, removable flash drive or any other component with similar storage capabilities) for storing an operating system 345 (e.g., WINDOWS® 7 and WINDOWS® EMBEDDED COMPACT (i.e., WINDOWS® CE) from MICROSOFT® CORPORATION of Redmond, Wash., and the LINUX® open source operating system) and various application programs 350. The mass storage device 340 may store other types of information as well.

Main memory 330 provides for storage of instructions and information directly accessible by central processing unit (CPU) 325. Main memory 330 may be configured to include random-access memory 332 (RAM) and read-only memory 334 (ROM). The ROM 334 may permanently store firmware or a basic input/output system (BIOS), which provide first instructions to vehicle-mount computer 320 when it is booted. RAM 332 may serve as temporary and immediately accessible storage for operating system 345 and application programs 350.

As illustrated in FIG. 4, computer touch screen 370 may be provided for inputting and displaying information using vehicle-mount computer 320. Computer touch screen 370 is operably connected to, and in communication with, vehicle-mount computer 320. Although touch screen 370 is illustrated in FIG. 4, other input devices (e.g., keyboard or mouse) or display devices may be utilized in connection with vehicle mount computer 320. The vehicle computer 320 may also include speaker 380 or other types of internal or external sound output devices.

As depicted in FIG. 4, the vehicle-mount computer 320 of the exemplary system 300 may also include network interface 365. Network interface 365 is operably connected to communications network 385, enabling vehicle-mount computer 320 to communicate with communications network 385. Communications network 385 may include any collection of computers or communication devices interconnected by communication channels. The communication channels may be wired or wireless (e.g., using BLUETOOTH® protocols). Examples of such communication networks include, without limitation, local area networks, the Internet, and cellular networks.

The connection to the communications network 385 allows vehicle computer 320 to communicate with the headset 305. The vehicle computer 320 may also be in communication with vehicle systems 381 such as a controlled braking system (e.g., wired or wireless communication). As described above with regard to system 100, the EEG headset 305 can monitor the user's 301 EEG activity in near real time and transmit the activity to the vehicle computer 320. The EEG headset 305 allows for the monitoring of attention, focus, engagement, interest, excitement, affinity, relaxation and stress, all of which can be used to make inferences into the activity being performed by the user 301.

In one embodiment, a sudden detection of high excitement from the user 301 could be used to trigger vehicle systems 381 such as a controlled breaking system on a forklift, given that a state of high excitement could be due to someone stepping in front of the vehicle. The predictive breaking system could go into effect before the forklift operator had time to consciously process what has happened and engage in an appropriate response.

In another embodiment, a user 301 might be operating a piece of heavy machinery and start to lose focus on the task at hand. The vehicle system 381 would then either perform a controlled slow down or completely stop the machinery until the operator 301 has given the task of operation their full attention.

In another embodiment, a motor vehicle user 301 could be monitored using the headset 305 for signs of drowsiness, which would sound an alarm through speakers 380, or a vehicle system 381 (e.g., an ignition lock-out system) could prevent the user 301 from operating the vehicle until the state of alertness was improved. A person having skill in the art will recognize that system 300 could be configured for use for multiple different safety/vehicle situations, and system 301 is not limited to the exemplary configurations referenced above.

To supplement the present disclosure, this application incorporates entirely by reference the following commonly assigned patents, patent application publications, and patent applications:

  • U.S. Pat. No. 6,832,725; U.S. Pat. No. 7,128,266;
  • U.S. Pat. No. 7,159,783; U.S. Pat. No. 7,413,127;
  • U.S. Pat. No. 7,726,575; U.S. Pat. No. 8,294,969;
  • U.S. Pat. No. 8,317,105; U.S. Pat. No. 8,322,622;
  • U.S. Pat. No. 8,366,005; U.S. Pat. No. 8,371,507;
  • U.S. Pat. No. 8,376,233; U.S. Pat. No. 8,381,979;
  • U.S. Pat. No. 8,390,909; U.S. Pat. No. 8,408,464;
  • U.S. Pat. No. 8,408,468; U.S. Pat. No. 8,408,469;
  • U.S. Pat. No. 8,424,768; U.S. Pat. No. 8,448,863;
  • U.S. Pat. No. 8,457,013; U.S. Pat. No. 8,459,557;
  • U.S. Pat. No. 8,469,272; U.S. Pat. No. 8,474,712;
  • U.S. Pat. No. 8,479,992; U.S. Pat. No. 8,490,877;
  • U.S. Pat. No. 8,517,271; U.S. Pat. No. 8,523,076;
  • U.S. Pat. No. 8,528,818; U.S. Pat. No. 8,544,737;
  • U.S. Pat. No. 8,548,242; U.S. Pat. No. 8,548,420;
  • U.S. Pat. No. 8,550,335; U.S. Pat. No. 8,550,354;
  • U.S. Pat. No. 8,550,357; U.S. Pat. No. 8,556,174;
  • U.S. Pat. No. 8,556,176; U.S. Pat. No. 8,556,177;
  • U.S. Pat. No. 8,559,767; U.S. Pat. No. 8,599,957;
  • U.S. Pat. No. 8,561,895; U.S. Pat. No. 8,561,903;
  • U.S. Pat. No. 8,561,905; U.S. Pat. No. 8,565,107;
  • U.S. Pat. No. 8,571,307; U.S. Pat. No. 8,579,200;
  • U.S. Pat. No. 8,583,924; U.S. Pat. No. 8,584,945;
  • U.S. Pat. No. 8,587,595; U.S. Pat. No. 8,587,697;
  • U.S. Pat. No. 8,588,869; U.S. Pat. No. 8,590,789;
  • U.S. Pat. No. 8,596,539; U.S. Pat. No. 8,596,542;
  • U.S. Pat. No. 8,596,543; U.S. Pat. No. 8,599,271;
  • U.S. Pat. No. 8,599,957; U.S. Pat. No. 8,600,158;
  • U.S. Pat. No. 8,600,167; U.S. Pat. No. 8,602,309;
  • U.S. Pat. No. 8,608,053; U.S. Pat. No. 8,608,071;
  • U.S. Pat. No. 8,611,309; U.S. Pat. No. 8,615,487;
  • U.S. Pat. No. 8,616,454; U.S. Pat. No. 8,621,123;
  • U.S. Pat. No. 8,622,303; U.S. Pat. No. 8,628,013;
  • U.S. Pat. No. 8,628,015; U.S. Pat. No. 8,628,016;
  • U.S. Pat. No. 8,629,926; U.S. Pat. No. 8,630,491;
  • U.S. Pat. No. 8,635,309; U.S. Pat. No. 8,636,200;
  • U.S. Pat. No. 8,636,212; U.S. Pat. No. 8,636,215;
  • U.S. Pat. No. 8,636,224; U.S. Pat. No. 8,638,806;
  • U.S. Pat. No. 8,640,958; U.S. Pat. No. 8,640,960;
  • U.S. Pat. No. 8,643,717; U.S. Pat. No. 8,646,692;
  • U.S. Pat. No. 8,646,694; U.S. Pat. No. 8,657,200;
  • U.S. Pat. No. 8,659,397; U.S. Pat. No. 8,668,149;
  • U.S. Pat. No. 8,678,285; U.S. Pat. No. 8,678,286;
  • U.S. Pat. No. 8,682,077; U.S. Pat. No. 8,687,282;
  • U.S. Pat. No. 8,692,927; U.S. Pat. No. 8,695,880;
  • U.S. Pat. No. 8,698,949; U.S. Pat. No. 8,717,494;
  • U.S. Pat. No. 8,717,494; U.S. Pat. No. 8,720,783;
  • U.S. Pat. No. 8,723,804; U.S. Pat. No. 8,723,904;
  • U.S. Pat. No. 8,727,223; U.S. Pat. No. D702,237;
  • U.S. Pat. No. 8,740,082; U.S. Pat. No. 8,740,085;
  • U.S. Pat. No. 8,746,563; U.S. Pat. No. 8,750,445;
  • U.S. Pat. No. 8,752,766; U.S. Pat. No. 8,756,059;
  • U.S. Pat. No. 8,757,495; U.S. Pat. No. 8,760,563;
  • U.S. Pat. No. 8,763,909; U.S. Pat. No. 8,777,108;
  • U.S. Pat. No. 8,777,109; U.S. Pat. No. 8,779,898;
  • U.S. Pat. No. 8,781,520; U.S. Pat. No. 8,783,573;
  • U.S. Pat. No. 8,789,757; U.S. Pat. No. 8,789,758;
  • U.S. Pat. No. 8,789,759; U.S. Pat. No. 8,794,520;
  • U.S. Pat. No. 8,794,522; U.S. Pat. No. 8,794,526;
  • U.S. Pat. No. 8,798,367; U.S. Pat. No. 8,807,431;
  • U.S. Pat. No. 8,807,432; U.S. Pat. No. 8,820,630;
  • International Publication No. 2013/163789;
  • International Publication No. 2013/173985;
  • International Publication No. 2014/019130;
  • International Publication No. 2014/110495;
  • U.S. Patent Application Publication No. 2008/0185432;
  • U.S. Patent Application Publication No. 2009/0134221;
  • U.S. Patent Application Publication No. 2010/0177080;
  • U.S. Patent Application Publication No. 2010/0177076;
  • U.S. Patent Application Publication No. 2010/0177707;
  • U.S. Patent Application Publication No. 2010/0177749;
  • U.S. Patent Application Publication No. 2011/0202554;
  • U.S. Patent Application Publication No. 2012/0111946;
  • U.S. Patent Application Publication No. 2012/0138685;
  • U.S. Patent Application Publication No. 2012/0168511;
  • U.S. Patent Application Publication No. 2012/0168512;
  • U.S. Patent Application Publication No. 2012/0193423;
  • U.S. Patent Application Publication No. 2012/0203647;
  • U.S. Patent Application Publication No. 2012/0223141;
  • U.S. Patent Application Publication No. 2012/0228382;
  • U.S. Patent Application Publication No. 2012/0248188;
  • U.S. Patent Application Publication No. 2013/0043312;
  • U.S. Patent Application Publication No. 2013/0056285;
  • U.S. Patent Application Publication No. 2013/0070322;
  • U.S. Patent Application Publication No. 2013/0075168;
  • U.S. Patent Application Publication No. 2013/0082104;
  • U.S. Patent Application Publication No. 2013/0175341;
  • U.S. Patent Application Publication No. 2013/0175343;
  • U.S. Patent Application Publication No. 2013/0200158;
  • U.S. Patent Application Publication No. 2013/0256418;
  • U.S. Patent Application Publication No. 2013/0257744;
  • U.S. Patent Application Publication No. 2013/0257759;
  • U.S. Patent Application Publication No. 2013/0270346;
  • U.S. Patent Application Publication No. 2013/0278425;
  • U.S. Patent Application Publication No. 2013/0287258;
  • U.S. Patent Application Publication No. 2013/0292475;
  • U.S. Patent Application Publication No. 2013/0292477;
  • U.S. Patent Application Publication No. 2013/0293539;
  • U.S. Patent Application Publication No. 2013/0293540;
  • U.S. Patent Application Publication No. 2013/0306728;
  • U.S. Patent Application Publication No. 2013/0306730;
  • U.S. Patent Application Publication No. 2013/0306731;
  • U.S. Patent Application Publication No. 2013/0307964;
  • U.S. Patent Application Publication No. 2013/0308625;
  • U.S. Patent Application Publication No. 2013/0313324;
  • U.S. Patent Application Publication No. 2013/0313325;
  • U.S. Patent Application Publication No. 2013/0341399;
  • U.S. Patent Application Publication No. 2013/0342717;
  • U.S. Patent Application Publication No. 2014/0001267;
  • U.S. Patent Application Publication No. 2014/0002828;
  • U.S. Patent Application Publication No. 2014/0008430;
  • U.S. Patent Application Publication No. 2014/0008439;
  • U.S. Patent Application Publication No. 2014/0025584;
  • U.S. Patent Application Publication No. 2014/0027518;
  • U.S. Patent Application Publication No. 2014/0034734;
  • U.S. Patent Application Publication No. 2014/0036848;
  • U.S. Patent Application Publication No. 2014/0039693;
  • U.S. Patent Application Publication No. 2014/0042814;
  • U.S. Patent Application Publication No. 2014/0049120;
  • U.S. Patent Application Publication No. 2014/0049635;
  • U.S. Patent Application Publication No. 2014/0061305;
  • U.S. Patent Application Publication No. 2014/0061306;
  • U.S. Patent Application Publication No. 2014/0063289;
  • U.S. Patent Application Publication No. 2014/0066136;
  • U.S. Patent Application Publication No. 2014/0067692;
  • U.S. Patent Application Publication No. 2014/0070005;
  • U.S. Patent Application Publication No. 2014/0071840;
  • U.S. Patent Application Publication No. 2014/0074746;
  • U.S. Patent Application Publication No. 2014/0075846;
  • U.S. Patent Application Publication No. 2014/0076974;
  • U.S. Patent Application Publication No. 2014/0078341;
  • U.S. Patent Application Publication No. 2014/0078342;
  • U.S. Patent Application Publication No. 2014/0078345;
  • U.S. Patent Application Publication No. 2014/0084068;
  • U.S. Patent Application Publication No. 2014/0097249;
  • U.S. Patent Application Publication No. 2014/0098792;
  • U.S. Patent Application Publication No. 2014/0100774;
  • U.S. Patent Application Publication No. 2014/0100813;
  • U.S. Patent Application Publication No. 2014/0103115;
  • U.S. Patent Application Publication No. 2014/0104413;
  • U.S. Patent Application Publication No. 2014/0104414;
  • U.S. Patent Application Publication No. 2014/0104416;
  • U.S. Patent Application Publication No. 2014/0104451;
  • U.S. Patent Application Publication No. 2014/0106594;
  • U.S. Patent Application Publication No. 2014/0106725;
  • U.S. Patent Application Publication No. 2014/0108010;
  • U.S. Patent Application Publication No. 2014/0108402;
  • U.S. Patent Application Publication No. 2014/0108682;
  • U.S. Patent Application Publication No. 2014/0110485;
  • U.S. Patent Application Publication No. 2014/0114530;
  • U.S. Patent Application Publication No. 2014/0124577;
  • U.S. Patent Application Publication No. 2014/0124579;
  • U.S. Patent Application Publication No. 2014/0125842;
  • U.S. Patent Application Publication No. 2014/0125853;
  • U.S. Patent Application Publication No. 2014/0125999;
  • U.S. Patent Application Publication No. 2014/0129378;
  • U.S. Patent Application Publication No. 2014/0131438;
  • U.S. Patent Application Publication No. 2014/0131441;
  • U.S. Patent Application Publication No. 2014/0131443;
  • U.S. Patent Application Publication No. 2014/0131444;
  • U.S. Patent Application Publication No. 2014/0131445;
  • U.S. Patent Application Publication No. 2014/0131448;
  • U.S. Patent Application Publication No. 2014/0133379;
  • U.S. Patent Application Publication No. 2014/0136208;
  • U.S. Patent Application Publication No. 2014/0140585;
  • U.S. Patent Application Publication No. 2014/0151453;
  • U.S. Patent Application Publication No. 2014/0152882;
  • U.S. Patent Application Publication No. 2014/0158770;
  • U.S. Patent Application Publication No. 2014/0159869;
  • U.S. Patent Application Publication No. 2014/0160329;
  • U.S. Patent Application Publication No. 2014/0166755;
  • U.S. Patent Application Publication No. 2014/0166757;
  • U.S. Patent Application Publication No. 2014/0166759;
  • U.S. Patent Application Publication No. 2014/0166760;
  • U.S. Patent Application Publication No. 2014/0166761;
  • U.S. Patent Application Publication No. 2014/0168787;
  • U.S. Patent Application Publication No. 2014/0175165;
  • U.S. Patent Application Publication No. 2014/0175169;
  • U.S. Patent Application Publication No. 2014/0175172;
  • U.S. Patent Application Publication No. 2014/0175174;
  • U.S. Patent Application Publication No. 2014/0191644;
  • U.S. Patent Application Publication No. 2014/0191913;
  • U.S. Patent Application Publication No. 2014/0197238;
  • U.S. Patent Application Publication No. 2014/0197239;
  • U.S. Patent Application Publication No. 2014/0197304;
  • U.S. Patent Application Publication No. 2014/0203087;
  • U.S. Patent Application Publication No. 2014/0204268;
  • U.S. Patent Application Publication No. 2014/0214631;
  • U.S. Patent Application Publication No. 2014/0217166;
  • U.S. Patent Application Publication No. 2014/0217180;
  • U.S. patent application Ser. No. 13/367,978 for a Laser Scanning Module Employing an Elastomeric U-Hinge Based Laser Scanning Assembly, filed Feb. 7, 2012 (Feng et al.);
  • U.S. patent application Ser. No. 29/436,337 for an Electronic Device, filed Nov. 5, 2012 (Fitch et al.);
  • U.S. patent application Ser. No. 13/771,508 for an Optical Redirection Adapter, filed Feb. 20, 2013 (Anderson);
  • U.S. patent application Ser. No. 13/852,097 for a System and Method for Capturing and Preserving Vehicle Event Data, filed Mar. 28, 2013 (Barker et al.);
  • U.S. patent application Ser. No. 13/902,110 for a System and Method for Display of Information Using a Vehicle-Mount Computer, filed May 24, 2013 (Hollifield);
  • U.S. patent application Ser. No. 13/902,144, for a System and Method for Display of Information Using a Vehicle-Mount Computer, filed May 24, 2013 (Chamberlin);
  • U.S. patent application Ser. No. 13/902,242 for a System For Providing A Continuous Communication Link With A Symbol Reading Device, filed May 24, 2013 (Smith et al.);
  • U.S. patent application Ser. No. 13/912,262 for a Method of Error Correction for 3D Imaging Device, filed Jun. 7, 2013 (Jovanovski et al.);
  • U.S. patent application Ser. No. 13/912,702 for a System and Method for Reading Code Symbols at Long Range Using Source Power Control, filed Jun. 7, 2013 (Xian et al.);
  • U.S. patent application Ser. No. 29/458,405 for an Electronic Device, filed Jun. 19, 2013 (Fitch et al.);
  • U.S. patent application Ser. No. 13/922,339 for a System and Method for Reading Code Symbols Using a Variable Field of View, filed Jun. 20, 2013 (Xian et al.);
  • U.S. patent application Ser. No. 13/927,398 for a Code Symbol Reading System Having Adaptive Autofocus, filed Jun. 26, 2013 (Todeschini);
  • U.S. patent application Ser. No. 13/930,913 for a Mobile Device Having an Improved User Interface for Reading Code Symbols, filed Jun. 28, 2013 (Gelay et al.);
  • U.S. patent application Ser. No. 29/459,620 for an Electronic Device Enclosure, filed Jul. 2, 2013 (London et al.);
  • U.S. patent application Ser. No. 29/459,681 for an Electronic Device Enclosure, filed Jul. 2, 2013 (Chaney et al.);
  • U.S. patent application Ser. No. 13/933,415 for an Electronic Device Case, filed Jul. 2, 2013 (London et al.);
  • U.S. patent application Ser. No. 29/459,785 for a Scanner and Charging Base, filed Jul. 3, 2013 (Fitch et al.);
  • U.S. patent application Ser. No. 29/459,823 for a Scanner, filed Jul. 3, 2013 (Zhou et al.);
  • U.S. patent application Ser. No. 13/947,296 for a System and Method for Selectively Reading Code Symbols, filed Jul. 22, 2013 (Rueblinger et al.);
  • U.S. patent application Ser. No. 13/950,544 for a Code Symbol Reading System Having Adjustable Object Detection, filed Jul. 25, 2013 (Jiang);
  • U.S. patent application Ser. No. 13/961,408 for a Method for Manufacturing Laser Scanners, filed Aug. 7, 2013 (Saber et al.);
  • U.S. patent application Ser. No. 14/018,729 for a Method for Operating a Laser Scanner, filed Sep. 5, 2013 (Feng et al.);
  • U.S. patent application Ser. No. 14/019,616 for a Device Having Light Source to Reduce Surface Pathogens, filed Sep. 6, 2013 (Todeschini);
  • U.S. patent application Ser. No. 14/023,762 for a Handheld Indicia Reader Having Locking Endcap, filed Sep. 11, 2013 (Gannon);
  • U.S. patent application Ser. No. 14/035,474 for Augmented-Reality Signature Capture, filed Sep. 24, 2013 (Todeschini);
  • U.S. patent application Ser. No. 29/468,118 for an Electronic Device Case, filed Sep. 26, 2013 (Oberpriller et al.);
  • U.S. patent application Ser. No. 14/055,234 for Dimensioning System, filed Oct. 16, 2013 (Fletcher);
  • U.S. patent application Ser. No. 14/053,314 for Indicia Reader, filed Oct. 14, 2013 (Huck);
  • U.S. patent application Ser. No. 14/065,768 for Hybrid System and Method for Reading Indicia, filed Oct. 29, 2013 (Meier et al.);
  • U.S. patent application Ser. No. 14/074,746 for Self-Checkout Shopping System, filed Nov. 8, 2013 (Hejl et al.);
  • U.S. patent application Ser. No. 14/074,787 for Method and System for Configuring Mobile Devices via NFC Technology, filed Nov. 8, 2013 (Smith et al.);
  • U.S. patent application Ser. No. 14/087,190 for Optimal Range Indicators for Bar Code Validation, filed Nov. 22, 2013 (Hejl);
  • U.S. patent application Ser. No. 14/094,087 for Method and System for Communicating Information in an Digital Signal, filed Dec. 2, 2013 (Peake et al.);
  • U.S. patent application Ser. No. 14/101,965 for High Dynamic-Range Indicia Reading System, filed Dec. 10, 2013 (Xian);
  • U.S. patent application Ser. No. 14/150,393 for Indicia-reader Having Unitary Construction Scanner, filed Jan. 8, 2014 (Colavito et al.);
  • U.S. patent application Ser. No. 14/154,207 for Laser Barcode Scanner, filed Jan. 14, 2014 (Hou et al.);
  • U.S. patent application Ser. No. 14/165,980 for System and Method for Measuring Irregular Objects with a Single Camera filed Jan. 28, 2014 (Li et al.);
  • U.S. patent application Ser. No. 14/166,103 for Indicia Reading Terminal Including Optical Filter filed Jan. 28, 2014 (Lu et al.);
  • U.S. patent application Ser. No. 14/200,405 for Indicia Reader for Size-Limited Applications filed Mar. 7, 2014 (Feng et al.);
  • U.S. patent application Ser. No. 14/231,898 for Hand-Mounted Indicia-Reading Device with Finger Motion Triggering filed Apr. 1, 2014 (Van Horn et al.);
  • U.S. patent application Ser. No. 14/250,923 for Reading Apparatus Having Partial Frame Operating Mode filed Apr. 11, 2014, (Deng et al.);
  • U.S. patent application Ser. No. 14/257,174 for Imaging Terminal Having Data Compression filed Apr. 21, 2014, (Barber et al.);
  • U.S. patent application Ser. No. 14/257,364 for Docking System and Method Using Near Field Communication filed Apr. 21, 2014 (Showering);
  • U.S. patent application Ser. No. 14/264,173 for Autofocus Lens System for Indicia Readers filed Apr. 29, 2014 (Ackley et al.);
  • U.S. patent application Ser. No. 14/274,858 for Mobile Printer with Optional Battery Accessory filed May 12, 2014 (Marty et al.);
  • U.S. patent application Ser. No. 14/277,337 for MULTIPURPOSE OPTICAL READER, filed May 14, 2014 (Jovanovski et al.);
  • U.S. patent application Ser. No. 14/283,282 for TERMINAL HAVING ILLUMINATION AND FOCUS CONTROL filed May 21, 2014 (Liu et al.);
  • U.S. patent application Ser. No. 14/300,276 for METHOD AND SYSTEM FOR CONSIDERING INFORMATION ABOUT AN EXPECTED RESPONSE WHEN PERFORMING SPEECH RECOGNITION, filed Jun. 10, 2014 (Braho et al.);
  • U.S. patent application Ser. No. 14/305,153 for INDICIA READING SYSTEM EMPLOYING DIGITAL GAIN CONTROL filed Jun. 16, 2014 (Xian et al.);
  • U.S. patent application Ser. No. 14/310,226 for AUTOFOCUSING OPTICAL IMAGING DEVICE filed Jun. 20, 2014 (Koziol et al.);
  • U.S. patent application Ser. No. 14/327,722 for CUSTOMER FACING IMAGING SYSTEMS AND METHODS FOR OBTAINING IMAGES filed Jul. 10, 2014 (Oberpriller et al,);
  • U.S. patent application Ser. No. 14/327,827 for a MOBILE-PHONE ADAPTER FOR ELECTRONIC TRANSACTIONS, filed Jul. 10, 2014 (Hejl);
  • U.S. patent application Ser. No. 14/329,303 for CELL PHONE READING MODE USING IMAGE TIMER filed Jul. 11, 2014 (Coyle);
  • U.S. patent application Ser. No. 14/333,588 for SYMBOL READING SYSTEM WITH INTEGRATED SCALE BASE filed Jul. 17, 2014 (Barten);
  • U.S. patent application Ser. No. 14/334,934 for a SYSTEM AND METHOD FOR INDICIA VERIFICATION, filed Jul. 18, 2014 (Hejl);
  • U.S. patent application Ser. No. 14/336,188 for METHOD OF AND SYSTEM FOR DETECTING OBJECT WEIGHING INTERFERENCES, Filed Jul. 21, 2014 (Amundsen et al.);
  • U.S. patent application Ser. No. 14/339,708 for LASER SCANNING CODE SYMBOL READING SYSTEM, filed Jul. 24, 2014 (Xian et al.);
  • U.S. patent application Ser. No. 14/340,627 for an AXIALLY REINFORCED FLEXIBLE SCAN ELEMENT, filed Jul. 25, 2014 (Rueblinger et al.);
  • U.S. patent application Ser. No. 14/340,716 for an OPTICAL IMAGER AND METHOD FOR CORRELATING A MEDICATION PACKAGE WITH A PATIENT, filed Jul. 25, 2014 (Ellis);
  • U.S. patent application Ser. No. 14/342,544 for Imaging Based Barcode Scanner Engine with Multiple Elements Supported on a Common Printed Circuit Board filed Mar. 4, 2014 (Liu et al.);
  • U.S. patent application Ser. No. 14/345,735 for Optical Indicia Reading Terminal with Combined Illumination filed Mar. 19, 2014 (Ouyang);
  • U.S. patent application Ser. No. 14/336,188 for METHOD OF AND SYSTEM FOR DETECTING OBJECT WEIGHING INTERFERENCES, Filed Jul. 21, 2014 (Amundsen et al.);
  • U.S. patent application Ser. No. 14/355,613 for Optical Indicia Reading Terminal with Color Image Sensor filed May 1, 2014 (Lu et al.);
  • U.S. patent application Ser. No. 14/370,237 for WEB-BASED SCAN-TASK ENABLED SYSTEM AND METHOD OF AND APPARATUS FOR DEVELOPING AND DEPLOYING THE SAME ON A CLIENT-SERVER NETWORK filed Jul. 2, 2014 (Chen et al.);
  • U.S. patent application Ser. No. 14/370,267 for INDUSTRIAL DESIGN FOR CONSUMER DEVICE BASED SCANNING AND MOBILITY, filed Jul. 2, 2014 (Ma et al.);
  • U.S. patent application Ser. No. 14/376,472, for an ENCODED INFORMATION READING TERMINAL INCLUDING HTTP SERVER, filed Aug. 4, 2014 (Lu);
  • U.S. patent application Ser. No. 14/379,057 for METHOD OF USING CAMERA SENSOR INTERFACE TO TRANSFER MULTIPLE CHANNELS OF SCAN DATA USING AN IMAGE FORMAT filed Aug. 15, 2014 (Wang et al.);
  • U.S. patent application Ser. No. 14/452,697 for INTERACTIVE INDICIA READER, filed Aug. 6, 2014 (Todeschini);
  • U.S. patent application Ser. No. 14/453,019 for DIMENSIONING SYSTEM WITH GUIDED ALIGNMENT, filed Aug. 6, 2014 (Li et al.);
  • U.S. patent application Ser. No. 14/460,387 for APPARATUS FOR DISPLAYING BAR CODES FROM LIGHT EMITTING DISPLAY SURFACES filed Aug. 15, 2014 (Van Horn et al.);
  • U.S. patent application Ser. No. 14/460,829 for ENCODED INFORMATION READING TERMINAL WITH WIRELESS PATH SELECTION CAPABILITY, filed Aug. 15, 2014 (Wang et al.);
  • U.S. patent application Ser. No. 14/462,801 for MOBILE COMPUTING DEVICE WITH DATA COGNITION SOFTWARE, filed on Aug. 19, 2014 (Todeschini et al.);
  • U.S. patent application Ser. No. 14/446,387 for INDICIA READING TERMINAL PROCESSING PLURALITY OF FRAMES OF IMAGE DATA RESPONSIVELY TO TRIGGER SIGNAL ACTIVATION filed Jul. 30, 2014 (Wang et al.);
  • U.S. patent application Ser. No. 14/446,391 for MULTIFUNCTION POINT OF SALE APPARATUS WITH OPTICAL SIGNATURE CAPTURE filed Jul. 30, 2014 (Good et al.);
  • U.S. patent application Ser. No. 29/486,759 for an Imaging Terminal, filed Apr. 2, 2014 (Oberpriller et al.);
  • U.S. patent application Ser. No. 29/492,903 for an INDICIA SCANNER, filed Jun. 4, 2014 (Zhou et al.); and
  • U.S. patent application Ser. No. 29/494,725 for an IN-COUNTER BARCODE SCANNER, filed Jun. 24, 2014 (Oberpriller et al.).

In the specification and/or figures, typical embodiments and environments of the invention have been disclosed. The present invention is not limited to such exemplary embodiments. The use of the term “and/or” includes any and all combinations of one or more of the associated listed items. The figures are schematic representations and so are not necessarily drawn to scale. Unless otherwise noted, specific terms have been used in a generic and descriptive sense and not for purposes of limitation.

Claims

1. An indicia-reading system having an interface with a user's nervous system, comprising:

a headset comprising electrodes capable of detecting electromagnetic signals produced in the brain of a user; and
a indicia reader in communication with the headset, comprising (i) a central processing unit and memory (ii) an indicia capturing subsystem for acquiring information about indicia within the indicia-capturing subsystem's field of view, and (iii) an indicia-decoding subsystem configured for decoding indicia information acquired by the indicia-capturing subsystem;
wherein the indicia reader is configured to monitor the electromagnetic signals detected by the headset.

2. The system according to claim 1, wherein the indicia reader is configured to perform an operation in response to electromagnetic signals detected by the headset.

3. The system according to claim 2, wherein the indicia reader operation is acquiring information about indicia within the indicia-capturing subsystem's field of view.

4. The system according to claim 2, wherein the indicia reader operation is placing the indicia reader into a different mode.

5. The system according to claim 1, wherein the detected signals produced in the brain of the user correspond to a facial expression.

6. The system according to claim 5, wherein the facial expression is a wink.

7. The system according to claim 1, wherein the detected signals produced in the brain of the user correspond to mental commands.

8. The system according to claim 1, wherein the communication between the headset and the indicia reader is wireless communication.

9. The system according to claim 1, wherein the indicia reader is a wearable computer.

10-21. (canceled)

22. A system, comprising:

at least one electrode positioned to detect electromagnetic signals from a user's brain;
a computer in communication with the at least one electrode, the computer including a central processing unit and memory; and
an indicia reader in communication with the computer, the indicia reader including an indicia capturing subsystem and an indicia-decoding subsystem;
wherein the computer monitors the detected electromagnetic signals and correlates the electromagnetic signals to user activities.

23. The system according to claim 22, wherein the user activities comprise facial expressions or mental commands.

24. The system according to claim 23, wherein the facial expression is a blink.

25. The system according to claim 22, wherein the indicia reader is configured to perform an operation in response to the user activities.

26. The system according to claim 25, wherein the indicia reader operation is acquiring information about indicia within the indicia-capturing subsystem's field of view.

27. The system according to claim 25, wherein the indicia reader operation is placing the indicia reader into a different mode.

28. The system according to claim 22, comprising a headset configured to be worn by the user, the headset having a housing that at least partially encloses the at least one electrode.

29. The system according to claim 28, wherein the computer is a wearable computer.

30. The system according to claim 29, wherein the indicia reader is a handheld indicia reader.

31. The system according to claim 30, wherein the headset, the computer, and the indicia reader are in wireless communication.

32. A method, comprising:

detecting electromagnetic signals from a user's brain;
monitoring the detected electromagnetic signals;
correlating the electromagnetic signals to activities performed by the user;
controlling operation of an indicia reader wherein the indicia reader executes an operation when the user performs the user activities.

33. The method according to claim 32, wherein the indicia reader operation is acquiring information about indicia within the indicia reader's field of view.

Patent History
Publication number: 20160364586
Type: Application
Filed: Jun 10, 2015
Publication Date: Dec 15, 2016
Inventor: Erik Todeschini (Camillus, NY)
Application Number: 14/735,717
Classifications
International Classification: G06K 7/016 (20060101); B60W 40/09 (20060101);