Increasing reliability and reducing latency in a wireless network

A mesh communication network for use in, for example, process control plants includes a plurality of network devices transmitting and receiving data according to a network schedule defined as a set of concurrent overlapping superframes, and along a set of graphs defining communication paths between pairs of network devices. A network manager residing in or outside the communication network develops a routing scheme for the network by analyzing the topology of the network and defining a set of graphs for use in routing or transmitting data between various nodes of the network, each graph including one or more communication paths between pairs of network devices. Concurrently or consequently, the network manager defines the network schedule in view of at least transmission requirements, power availability, and signal quality at each network device. If desired, the network manager may begin to define the network schedule upon completing the definition of the graphs of the communication network, so that the network manager may define the network schedule in view both the defined graphs and the transmission, power, etc. parameters associated with each network device.

Skip to: Description  ·  Claims  ·  References Cited  · Patent History  ·  Patent History
Description
CROSS-REFERENCE TO RELATED APPLICATIONS

This application is based on and claims the benefit of priority to U.S. Provisional Application No. 60/911,795, entitled “Routing, Scheduling, Reliable and Secure Operations in a Wireless Communication Protocol” filed Apr. 13, 2007, the entire disclosure of which is hereby expressly incorporated herein by reference.

FIELD OF TECHNOLOGY

The present patent relates generally to wireless communications and, more particularly, to increasing reliability in a wireless network.

BACKGROUND

Communication protocols rely on various routing techniques to transfer data between communication endpoints on a communication network. Communication or network protocols and the corresponding routing strategies are typically selected in view of such factors as knowledge of network topology, size of the network, type of medium used as a signal carrier, security and reliability requirements, tolerable transmission delays, and types of devices forming the network. Due to a large number of such factors, a typical routing technique meets some of the design objectives at the expense of the others. For example, a certain routing technique may provide a high level of reliability in data delivery but may also require a relatively high overhead. Thus, while there are many known approaches to routing and many protocols compatible with these routing methods, there remain communication networks with the specific requirements that are not fully satisfied by any of the available routing methods and protocols. Moreover, as new types of communication networks, with the increasing demands for efficiency, throughput, and reliability, emerge in various industrial and commercial applications, the architects and developers frequently encounter new problems which are not easily addressed by the existing protocols and the associated routing techniques.

Generally speaking, a communication network includes nodes which are the senders and recipients of data sent over communication paths (either hardwired or wireless communication paths) connecting the nodes. Additionally, communication networks typically include dedicated routers responsible for directing traffic between nodes, and, optionally, include dedicated devices responsible for configuring and managing the network. Some or all of the nodes may be also adapted to function as routers so as to direct traffic sent between other network devices. Network devices may be inter-connected in a wired or wireless manner, and network devices may have different routing and transfer capabilities than certain nodes within the network. For example, dedicated routers may be capable of high volume transmissions while some nodes may only be capable of sending and receiving relatively little traffic over the same period of time. Additionally, the connections between nodes on a network may have different throughput capabilities and different attenuation characteristics. A fiber-optic cable, for example, may be capable of providing a bandwidth several orders of magnitude higher than a wireless link because of the difference in the inherent physical limitations of the medium.

In order for a node to send data to another node on a typical network, either the complete path from the source to the destination or the immediately relevant part of the path must be known. For example, the World Wide Web (WWW) allows pairs of computer hosts to communicate over large distances without either host knowing the complete path prior to sending the information. Instead, hosts are configured with the information about their assigned gateways and dedicated routers. In particular, the Internet Protocol (IP) provides network layer connectivity to the WWW. The EP defines a sub-protocol known as Address Resolution Protocol (ARP) which provides a local table at each host specifying the routing rules. Thus, a typical host connected to the WWW or a similar Wide Area Network (WAN) may know to route all packets with the predefined addresses matching a pre-configured pattern to host A and route the rest of the packets to host B. Similarly, the intermediate hosts forwarding the packets, or “hops,” also execute partial routing decisions and typically direct data in the general direction of the destination.

In most network protocols, most or all network devices are assigned sufficiently unique addresses to enable hosts to exchange information in an unambiguous manner. At least in case of unicast (one-to-one) transmissions, the destination address must be specified at the source. For this reason, network protocols typically define a rigid addressing scheme. As one of ordinary skill in the art will recognize, modifying or expanding addressing schemes is a complicated and expensive process. For example, the transition from version 4 of the IP protocol (IPv4) to version 6 (IPv6) requires significant updates to much of the infrastructure supporting IPv4. On the other hand, defining addressing schemes with large capability for small networks creates an unnecessary overhead. Thus, a network protocol ideally suited for a particular application offers a sufficient number of possible addresses without an excessive overhead in data transmission.

In short, a number of factor influence the implementation of particular protocols in particular industries. In the process control industry, it is known to use standardized communication protocols to enable devices made by different manufacturers to communicate with one another in an easy to use and implement manner. One such well known communication standard used in the process control industry is the Highway Addressable Remote Transmitter (HART) Communication Foundation protocol, referred to generally as the HART® protocol. Generally speaking, the HART® protocol supports a combined digital and analog signal on a dedicated wire or set of wires, in which on-line process signals (such as control signals, sensor measurements, etc.) are provided as an analog current signal (e.g., ranging from 4 to 20 milliamps) and in which other signals, such as device data, requests for device data, configuration data, alarm and event data, etc., are provided as digital signals superimposed or multiplexed onto the same wire or set of wires as the analog signal. However, the HART protocol currently requires the use of dedicated, hardwired communication lines, resulting in significant wiring needs within a process plant.

There has been a move, in the past number of years, to incorporate wireless technology into various industries including, in some limited manners, the process control industry. However, there are significant hurdles in the process control industry that limit the full scale incorporation, acceptance and use of wireless technology. In particular, the process control industry requires a completely reliable process control network because loss of signals can result in the loss of control of a plant, leading to catastrophic consequences, including explosions, the release of deadly chemicals or gases, etc. For example, Tapperson et al., U.S. Pat. No. 6,236,334 discloses the use of a wireless communications in the process control industry as a secondary or backup communication path or for use in sending non-critical or redundant communication signals. Moreover, there have been many advances in the use of wireless communication systems in general that may be applicable to the process control industry, but which have not yet been applied to the process control industry in a manner that allows or provides a reliable, and in some instances completely wireless, communication network within a process plant. U.S. Patent Application Publication Numbers 2005/0213612, 2006/0029060 and 2006/0029061 for example disclose various aspects of wireless communication technology related to a general wireless communication system.

Similar to wired communications, wireless communication protocols are expected to provide efficient, reliable and secure methods of exchanging information. Of course, much of the methodology developed to address these concerns on wired networks does not apply to wireless communications because of the shared and open nature of the medium. Further, in addition to the typical objectives behind a wired communication protocol, wireless protocols face other requirements with respect to the issues of interference and co-existence of several networks that use the same part of the radio frequency spectrum. Moreover, some wireless networks operate in the part of the spectrum that is unlicensed, or open to the public. Therefore, protocols servicing such networks must be capable of detecting and resolving issues related to frequency (channel) contention, radio resource sharing and negotiation, etc.

In the process control industry, developers of wireless communication protocols face additional challenges, such as achieving backward compatibility with wired devices, supporting previous wired versions of a protocol, providing transition services to devices retrofitted with wireless communicators, and providing routing techniques which can ensure both reliability and efficiency. Meanwhile, there remains a wide number of process control applications in which there are few, if any, in-place measurements. Currently these applications rely on observed measurements (e.g. water level is rising) or inspection (e.g. period maintenance of air conditioning unit, pump, fan, etc) to discover abnormal situations. In order to take action, operators frequently require face-to-face discussions. Many of these applications could be greatly simplified if measurement and control devices were utilized. However, current measurement devices usually require power, communications infrastructure, configuration, and support infrastructure which simply is not available.

In yet another aspect, the process control industry requires that the communication protocol servicing a particular process control network be able to accommodate field devices with different data transmission requirements, priorities, and power capabilities. In particular, some process control systems may include measurement devices that frequently (such as several times per second) report measurements to a centralized controller or to another field device. Meanwhile, another device in the same system may report measurements, alarms, or other data only once per hour. However, both devices may require that the respective measurement reports propagate to a destination host, such as a controller, a workstation, or a peer field device, with as little overhead in time and bandwidth as possible.

SUMMARY OF THE DISCLOSURE

A mesh communication network for use in, for example, process control plants includes a plurality of network devices transmitting and receiving data according to a network schedule defined as a set of concurrent overlapping superframes, and along a set of graphs defining communication paths between pairs of network devices. In some embodiments, a network manager residing in or outside the communication network develops a routing scheme for the network by analyzing the topology of the network and defining a set of graphs for use in routing or transmitting data between various nodes of the network, each graph including one or more communication paths between pairs of network devices. Concurrently or consequently, the network manager defines the network schedule in view of at least transmission requirements, power availability, and signal quality at each network device. If desired, the network manager may begin to define the network schedule upon completing the definition of the graphs of the communication network, so that the network manager may define the network schedule in view both the defined graphs and the transmission, power, etc. parameters associated with each network device.

In some embodiments, the mesh communication network is a wireless network, the graphs are directed graphs and, accordingly, the communication paths are unidirectional communication paths. The network devices of mesh communication network may originate and route data on behalf of other network devices. In a still further embodiment, the mesh communication network conforms to a star mesh topology, in which some network devices can only receive data or originate data and some network devices can receive data, originate data, and relay data between other network devices.

In another aspect, each of the superframes includes several communication timeslots of a predetermined time duration and each superframe repeats immediately after the occurrence of all communication timeslots in the superframe. Generally speaking, the total number of timeslots in each superframe defines the length of the superframe and each particular timeslot has a relative timeslot number that corresponds to the number of timeslots existing in the superframe prior to the particular timeslot. In one embodiment, each timeslot includes a transmission time segment during which one or more network devices transmit data and an acknowledgement segment during which one or more network devices send an acknowledgement corresponding to the transmitted data. In some embodiments, a network device further performs Clear Channel Assessment to ascertain whether a particular shared timeslot is available for transmission.

Upon defining the set of graphs and the communication schedule, the network manager communicates the relevant routing and scheduling information to some or all network devices (nodes) so that a packet sent from one network device to another network device can be properly routed according to graphs defined by the network manager and within the timeslots allocated by the network manager. In another embodiment, the function of analyzing the network and obtaining the topology of the network is distributed among at least several network devices so that one or more network devices participate in defining unidirectional or bidirectional graphs.

The length of each superframe may correspond to a transmission requirement of a particular network device. In this case, a dedicated service defines superframes and allocates timeslots within each of the superframes according to the needs of network devices and of external hosts communicating with the network devices. In one embodiment, the dedicated service associates a network device with one or more timeslots of a particular superframe so that the network device may transmit or receive data during each occurrence of the timeslot. If desired, a network device may participate in multiple superframes to transmit data specific to the network device and to forward data between other network devices.

If desired, the dedicated service may dynamically create and destroy superframes in view of changes in network conditions such as data bursts, congestion, block transfers, and network devices entering or leaving the network. Moreover, a network device or the dedicated service may efficiently deactivate a superframe without destroying the superframe by issuing a particular command. If desired, the dedicated service may be a software entity running on a dedicated physical host, or the dedicated service may run on a gateway device connecting the wireless mesh network to an external network or host.

The network schedule may include multiple communication channels and, in some embodiments, each communication channel may correspond to a unique carrier radio frequency. Each network device may have an individual schedule that includes relative timeslot numbers and communication channel identifiers and the individual schedule may specify the individually scheduled timeslots that the network device uses to transmit process data, route data originated from another network device, receive device-specific data, or receive broadcast data. In some embodiments, the individual schedule for a network device may specify a timeslot associated with several distinct communication channels during different superframe cycles, so that the network device transmits or receives data over different communication channels within a timeslot having the same relative slot number of a particular superframe. In some of these embodiments, the network device, in accordance with a corresponding individual schedule, iterates through several channels associated with a particular timeslot in a predefined manner. In other embodiments, the network schedule does not allocate different communication channels to the same timeslot.

If desired, the dedicated service may create long superframes for a transient device that periodically wakes up according to a predefined schedule, thereby eliminating the need for the transient device to negotiate resources for each transmission session. In yet another embodiment, the transient device conserves power by transmitting data only according to the necessary update rate of the transient device.

In a still further embodiment, at least some of the network devices are field devices performing a measurement or control functions in a process control environment. Each of these field devices is provisioned with a particular update rate, or frequency of communicating process data to another network device. In this case, the dedicated service defines superframes according to the update rates of the field devices. Moreover, if desired, a field device can negotiate a temporary change in timeslot allocation due to an occurrence of a transient condition requiring a higher-than-normal or lower-than-normal bandwidth.

In an embodiment, the management entity responsible, in part, for defining unidirectional graphs is a dedicated network manager and may be implemented as a software module run on a host residing in or outside of the network. In another embodiment, the network manager may be a dedicated physical device communicatively coupled to the network. In yet another embodiment, the network manager may be distributed among several devices residing in or outside of the network.

DETAILED DESCRIPTION OF THE DRAWINGS

FIG. 1 schematically illustrates a wireless network, connected to a plant automation network via a gateway device, providing communication between field devices and router devices and utilizing the routing techniques of the present disclosure.

FIG. 2 is a schematic representation of the layers of a wireless HART protocol which may be used in the wireless network illustrated in FIG. 1.

FIG. 3 is a block diagram that illustrates segments of a communication timeslot defined in accordance with one of the embodiments discussed herein.

FIG. 4 is a block diagram that illustrates an exemplary association of timeslots of a three-slot superframe with several communicating devices.

FIG. 5 schematically illustrates association of a timeslot of an exemplary superframe with several communication channels.

FIG. 6 is a block diagram that schematically illustrates an exemplary superframe definition including several concurrent superframes of different lengths.

FIG. 7 schematically illustrates a star network topology to which a wireless network such as the network illustrated in FIG. 1 or FIG. 3 may conform.

FIG. 8 schematically illustrates a mesh network topology to which a wireless network such as the network illustrated in FIG. 1 or FIG. 3 may conform.

FIG. 9 schematically illustrates a star mesh network topology to which a wireless network such as the network illustrated in FIG. 1 or FIG. 3 may conform.

FIG. 10 is a block diagram illustrating redundant upstream data paths in a wireless network utilizing some of the graph routing techniques of the present disclosure.

FIG. 11 is a block diagram illustrating redundant downstream data paths in a wireless network utilizing some of the graph routing techniques of the present disclosure.

FIG. 12 is a block diagram illustrating downstream and upstream data paths in a wireless network utilizing some of the source routing techniques of the present disclosure.

FIG. 13 schematically illustrates a wireless network having several network devices transmitting data according to a device-specific schedule and utilizing some of the routing and scheduling techniques of the present disclosure.

FIG. 14 illustrates example superframe definitions for two devices illustrated in FIG. 13.

FIGS. 15 and 16 illustrate timeslot allocation for several types of data which may be transmitted in a management superframe of the network illustrated in FIG. 13.

FIGS. 17-20 illustrate an example communication schedule definitions for each of the devices illustrated in FIG. 13.

DETAILED DESCRIPTION

FIG. 1 illustrates an exemplary network 10 in which the scheduling and routing techniques described herein may be used. In particular, the network 10 may include a plant automation network 12 connected to a wireless communication network 14. The plant automation network 12 may include one or more stationary workstations 16 and one or more portable workstations 18 connected over a communication backbone 20 which may be implemented using Ethernet, RS-485, Profibus DP, or using other suitable communication hardware and protocol. The workstations and other equipment forming the plant automation network 12 may provide various control and supervisory functions to plant personnel, including access to devices in the wireless network 14. The plant automation network 12 and the wireless network 14 may be connected via a gateway device 22. More specifically, the gateway device 22 may be connected to the backbone 20 in a wired manner and may communicate with the plant automation network 12 using any suitable (e.g., known) communication protocol. The gateway device 22, which may be implemented in any other desired manner (e.g., as a standalone device, a card insertable into an expansion slot of the host workstations 16 or 18, as a part of the input/output (IO) subsystem of a PLC-based or DCS-based system, etc.), may provide applications that are running on the network 12 with access to various devices of the wireless network 14. In addition to protocol and command conversion, the gateway device 22 may provide synchronized clocking used by time slots and superframes (sets of communication time slots spaced equally in time) of a scheduling scheme associated with a wireless protocol (referred to herein as a WirelessHART protocol) implemented in the network 14.

In some configurations, the network 10 may include more than one gateway device 22 to improve the efficiency and reliability of the network 10. In particular, multiple gateway devices 22 may provide additional bandwidth for the communication between the wireless network 14 and the plant automation network 12, as well as the outside world. On the other hand, the gateway 22 device may request bandwidth from the appropriate network service according to the gateway communication needs within the wireless network 14. A network manager software module 27, which may reside in the gateway device 22, may further reassess the necessary bandwidth while the system is operational. For example, the gateway device 22 may receive a request from a host residing outside of the wireless network 14 to retrieve a large amount of data. The gateway device 22 may then request the network manager 27 to allocate additional bandwidth to accommodate this transaction. For example, the gateway device 22 may issue an appropriate service request. The gateway device 22 may then request the network manager 27 to release the bandwidth upon completion of the transaction.

With continued reference to FIG. 1, the wireless network 14 may include one or more field devices 30-36. In general, process control systems, like those used in chemical, petroleum or other process plants, include field devices such as valves, valve positioners, switches, sensors (e.g., temperature, pressure and flow rate sensors), pumps, fans, etc. Generally speaking, field devices perform physical control functions within the process such as opening or closing valves or take measurements of process parameters. In the wireless communication network 14, field devices 30-36 are producers and consumers of wireless communication packets.

The devices 30-36 may communicate using a wireless communication protocol that provides the functionality of a similar wired network, with similar or improved operational performance. In particular, this protocol may enable the system to perform process data monitoring, critical data monitoring (with the more stringent performance requirements), calibration, device status and diagnostic monitoring, field device troubleshooting, commissioning, and supervisory process control. The applications performing these functions, however, typically require that the protocol supported by the wireless network 14 provide fast updates when necessary, move large amounts of data when required, and support network devices which join the wireless network 14, even if only temporarily for commissioning and maintenance work.

If desired, the network 14 may include non-wireless devices. For example, a field device 38 of FIG. 1 may be a legacy 4-20 mA device and a field device 40 may be a traditional wired HART device. To communicate within the network 14, the field devices 38 and 40 may be connected to the WirelessHART network 14 via a WirelessHART adapter (WHA) 50A, 50B, respectively. Additionally, each of the WHAs 50A, 50B may support other communication protocols such as Foundation® Fieldbus, PROFIBUS, DeviceNet, etc. In these embodiments, each of the WHAs 50A, 50B supports protocol translation on a lower layer of the protocol stack. Additionally, it is contemplated that a single WHA 50A or 50B may also function as a multiplexer and may support multiple HART or non-HART devices.

In general, the network manager 27 may be responsible for adapting the wireless network 14 to changing conditions and for scheduling communication resources. As network devices join and leave the network, the network manager 27 may update its internal model of the wireless network 14 and use this information to generate communication schedules and communication routes. Additionally, the network manager 27 may consider the overall performance of the wireless network 14 as well as the diagnostic information to adapt the wireless network 14 to changes in topology and communication requirements. Once the network manager 27 has generated the overall communication schedule, all or respective parts of the overall communication schedule may be transferred through a series of commands from the network manager 27 to the network devices.

To further increase bandwidth and improve reliability, the gateway device 22 may be functionally divided into a virtual gateway 24 and one or more network access points 25, which may be separate physical devices in wired communication with the gateway device 22. However, while FIG. 1 illustrates a wired connection 26 between the physically separate gateway device 22 and the access points 25, it will be understood that the elements 22-26 may also be provided as an integral device. Because the network access points 25 may be physically separated from the gateway device 22, the access points 25 may be strategically placed in several different locations with respect to the network 14. In addition to increasing the bandwidth, multiple access points 25 can increase the overall reliability of the network 14 by compensating for a potentially poor signal quality at one access point 25 using the other access point 25. Having multiple access points 25 also provides redundancy in case of a failure at one or more of the access points 25.

In addition to allocating bandwidth and otherwise bridging the networks 12 and 14, the gateway device 22 may perform one or more managerial functions in the wireless network 14. As illustrated in FIG. 1, a network manager software module 27 and a security manager software module 28 may be stored in and executed in the gateway device 22. Alternatively, the network manager 27 and/or the security manager 28 may run on one of the hosts 16 or 18 in the plant automation network 12. For example, the network manager 27 may run on the host 16 and the security manager 28 may run on the host 18. The network manager 27 may be responsible for configuration of the network 14, scheduling communication between wireless devices, managing routing tables associated with the wireless devices, monitoring the overall health of the wireless network 14, reporting the health of the wireless network 14 to the workstations 16 and 18, as well as other administrative and supervisory functions. Although a single active network manager 27 may be sufficient in the wireless network 14, redundant network managers 27 may be similarly supported to safeguard the wireless network 14 against unexpected equipment failures. Meanwhile, the security manager 28 may be responsible for protecting the wireless network 14 from malicious or accidental intrusions by unauthorized devices. To this end, the security manager 28 may manage authentication codes, verify authorization information supplied by devices attempting to join the wireless network 14, update temporary security data such as expiring secret keys, and perform other security functions.

With continued reference to FIG. 1, the wireless network 14 may include one or more field devices 30-36. In general, process control systems, like those used in chemical, petroleum or other process plants, include such field devices as valves, valve positioners, switches, sensors (e.g., temperature, pressure and flow rate sensors), pumps, fans, etc. Field devices perform physical control functions within the process such as opening or closing valves or take measurements of process parameters. In the wireless communication network 14, field devices 30-36 are producers and consumers of wireless communication packets.

The devices 30-36 may communicate using a wireless communication protocol that provides the functionality of a similar wired network, with similar or improved operational performance. In particular, this protocol may enable the system to perform process data monitoring, critical data monitoring (with the more stringent performance requirements), calibration, device status and diagnostic monitoring, field device troubleshooting, commissioning, and supervisory process control. The applications performing these functions, however, typically require that the protocol supported by the wireless network 14 provide fast updates when necessary, move large amounts of data when required, and support network devices which join the wireless network 14, even if only temporarily for commissioning and maintenance work.

In one embodiment, the wireless protocol supporting network devices 30-36 of the wireless network 14 is an extension of the known wired HART protocol, a widely accepted industry standard, that maintains the simple workflow and practices of the wired environment. In this sense, the network devices 30-36 may be considered WirelessHART devices. The same tools used for wired HART devices may be easily adapted to wireless devices 30-36 with a simple addition of new device description files. In this manner, the wireless protocol may leverage the experience and knowledge gained using the wired HART protocol to minimize training and simplify maintenance and support. Generally speaking, it may be convenient to adapt a protocol for wireless use so that most applications running on a device do not “notice” the transition from a wired network to a wireless network. Clearly, such transparency greatly reduces the cost of upgrading networks and, more generally, reduces the cost associated with developing and supporting devices that may be used with such networks. Some of the additional benefits of a wireless extension of the well-known HART protocol include access to measurements that were difficult or expensive to obtain with wired devices and the ability to configure and operate instruments from system software that can be installed on laptops, handhelds, workstations, etc. Another benefit is the ability to send diagnostic alerts from wireless devices back through the communication infrastructure to a centrally located diagnostic center. For example, every heat exchanger in a process plant could be fitted with a WirelessHART device and the end user and supplier could be alerted when a heat exchanger detects a problem. Yet another benefit is the ability to monitor conditions that present serious health and safety problems. For example, a WirelessHART device could be placed in flood zones on roads and be used to alert authorities and drivers about water levels. Other benefits include access to a wide range of diagnostics alerts and the ability to store trended as well as calculated values at the WirelessHART devices so that, when communications to the device are established, the values can be transferred to a host. In this manner, the WirelessHART protocol can provide a platform that enables host applications to have wireless access to existing HART-enabled field devices and the WirelessHART protocol can support the deployment of battery operated, wireless only HART-enabled field devices. The WirelessHART protocol may be used to establish a wireless communication standard for process applications and may further extend the application of HART communications and the benefits that this protocol provides to the process control industry by enhancing the basic HART technology to support wireless process automation applications.

Referring again to FIG. 1, the field devices 30-36 may be WirelessHART field devices, each provided as an integral unit and supporting all layers of the WirelessHART protocol stack. For example, in the network 14, the field device 30 may be a WirelessHART flow meter, the field devices 32 may be WirelessHART pressure sensors, the field device 34 may be a WirelessHART valve positioner, and the field device 36 may a WirelessHART pressure sensor. Importantly, the wireless devices 30-36 may support all of the HART features that users have come to expect from the wired HART protocol. As one of ordinary skill in the art will appreciate, one of the core strengths of the HART protocol is its rigorous interoperability requirements. In some embodiments, all WirelessHART equipment includes core mandatory capabilities in order to allow equivalent device types (made by different manufacturers, for example) to be interchanged without compromising system operation. Furthermore, the WirelessHART protocol is backward compatible to HART core technology such as the device description language (DDL). In the preferred embodiment, all of the WirelessHART devices should support the DDL, which ensures that end users immediately have the tools to begin utilizing the WirelessHART protocol.

If desired, the network 14 may include non-wireless devices. For example, a field device 38 of FIG. 1 may be a legacy 4-20 mA device and a field device 40 may be a traditional wired HART device. To communicate within the network 14, the field devices 38 and 40 may be connected to the WirelessHART network 14 via a WirelessHART adapter (WHA) 50A and 50B, respectively. Additionally, each of the WHAs 50A and 50B may support other communication protocols such as FOUNDATION® Fieldbus, PROFIBUS, DeviceNet, etc. In these embodiments, each of the WHAs 50A and 50B supports protocol translation on a lower layer of the protocol stack. Additionally, it is contemplated that a single WHA 50A or 50B may also function as a multiplexer and may support multiple HART or non-HART devices.

Plant personnel may additionally use handheld devices for installation, control, monitoring, and maintenance of network devices. Generally speaking, handheld devices are portable equipment that can connect directly to the wireless network 14 or through the gateway devices 22 as a host on the plant automation network 12. As illustrated in FIG. 1, a WirelessHART-connected handheld device 55 may communicate directly with the wireless network 14. When operating with a formed wireless network 14, the handheld device 55 may join the network 14 as just another WirelessHART field device. When operating with a target network device that is not connected to a WirelessHART network, the handheld device 55 may operate as a combination of the gateway device 22 and the network manager 27 by forming its own wireless network with the target network device.

A plant automation network-connected handheld device (not shown) may be used to connect to the plant automation network 12 through known networking technology, such as Wi-Fi. This device communicates with the network devices 30-40 through the gateway device 22 in the same fashion as external plant automation servers (not shown) or the workstations 16 and 18 communicate with the devices 30-40.

Additionally, the wireless network 14 may include a router device 60 which is a network device that forwards packets from one network device to another network device. A network device that is acting as a router device uses internal routing tables to conduct routing, i.e., to decide to which network device a particular packet should be sent. Standalone routers such as the router 60 may not be required in those embodiments where all of the devices on the wireless network 14 support routing. However, it may be beneficial (e.g. to extend the network, or to save the power of a field device in the network) to add one or more dedicated routers 60 to the network 14.

All of the devices directly connected to the wireless network 14 may be referred to as network devices. In particular, the wireless field devices 30-36, the adapters 50A and 50B, the routers 60, the gateway devices 22, the access points 25, and the wireless handheld device 55 are, for the purposes of routing and scheduling, network devices, each of which forms a node of the wireless network 14. In order to provide a very robust and an easily expandable wireless network, all of the devices in a network may support routing and each network device may be globally identified by a substantially unique address, such as a HART address, for example. The network manager 27 may contain a complete list of network devices and may assign each device a short, network unique 16-bit (for example) nickname. Additionally, each network device may store information related to update (or “scan”) rates, connection sessions, and device resources. In short, each network device maintains up-to-date information related to routing and scheduling within the wireless network 14. The network manager 27 may communicate this information to network devices whenever new devices join the network or whenever the network manager 27 detects or originates a change in topology or scheduling of the wireless network 14.

Further, each network device may store and maintain a list of neighbor devices that the network device has identified during listening operations. Generally speaking, a neighbor of a network device is another network device of any type potentially capable of establishing a connection with the network device in accordance with the standards imposed by a corresponding network. In case of the WirelessHART network 14, the connection is a direct wireless connection. However, it will be appreciated that a neighboring device may also be a network device connected to the particular device in a wired manner. As will be discussed later, network devices may promote their discovery by other network devices through advertisement, or special messages sent out during designated periods of time. Network devices operatively connected to the wireless network 14 have one or more neighbors which they may choose according to the strength of the advertising signal or to some other principle.

In the example illustrated in FIG. 1, each of a pair of network devices connected by a direct wireless connection 65 recognizes the other as a neighbor. Thus, network devices of the wireless network 14 may form a large number of inter-device connections 65. The possibility and desirability of establishing a direct wireless connection 65 between two network devices is determined by several factors, such as the physical distance between the nodes, obstacles between the nodes (devices), signal strength at each of the two nodes, etc. In general, each wireless connection 65 is characterized by a large set of parameters related to the frequency of transmission, the method of access to a radio resource, etc. One of ordinary skill in the art will recognize that, in general, wireless communication protocols may operate on designated frequencies, such as the ones assigned by the Federal Communications Commission (FCC) in the United States, or in the unlicensed part of the radio spectrum (e.g., 2.4 GHz). While the system and method discussed herein may be applied to a wireless network operating on any designated frequency or range of frequencies, the example embodiment discussed below relates to the wireless network 14 operating in the unlicensed, or shared part of the radio spectrum. In accordance with this embodiment, the wireless network 14 may be easily activated and adjusted to operate in a particular unlicensed frequency range as needed.

With continued reference to FIG. 1, two or more direct wireless connections 65 may form a communication path between nodes that cannot form a direct wireless connection 65. For example, the direct wireless connection 65A between the WirelessHART hand-held device 55 and WirelessHART device 36, along with the direct wireless connection 65B between the WirelessHART device 36 and the router 60, may form a communication path between the devices 55 and 60. As discussed in greater detail below, at least some of the communication paths may be directed communication paths (i.e., permitting or defining data transfer in only one direction between a pair of devices). Meanwhile, the WirelessHART device 36 may directly connect to each of the network devices 55, 60, 32, and to the network access points 25A and 25B. In general, network devices operating in the wireless network 14 may originate data packets, relay data packets sent by other devices, or perform both types of operations. As used herein, the term “end device” refers to a network device that does not relay data packets sent by other devices and term “routing device” refers to a network device that relays data packets traveling between other network devices. Of course, a routing device may also originate its own data or in some cases be an end device. One or several end devices and routing devices, along with several direct connections 65, may thus form a part of a mesh network.

Because a process plant may have hundreds or even thousands of field devices, the wireless network 14 operating in the plant may include a large number of nodes and, in many cases, an even larger number of direct connections 65 between pairs of nodes. As a result, the wireless network 14 may have a complex mesh topology, and some pairs of devices that do not share a direct connection 65 may have to communicate through many intermediate hops to perform communications between these devices. Thus, a data packet may sometimes need to travel along many direct connections 65 after leaving a source device but before reaching a destination device, and each direct connection 65 may add a delay to the overall delivery time of the data packet. Moreover, some of these intermediate devices may be located at an intersection of many communication paths of a mesh network. As such, these devices may be responsible for relaying a large number of packets originated by many different devices, possibly in addition to originating its own data. Consequently, a relatively busy intermediate device may not forward a transient data packet immediately, and instead may queue the packet for a relatively significant amount of time prior to sending the packet to a next node in the corresponding communication path. When the data packet eventually reaches the destination device, the destination device may reply with an acknowledgement packet which may also encounter similar delays. During the time the packet travels to the destination device and the corresponding acknowledgment packet travels back to the originating device from the destination device, the originating node may not know whether the data packet has successfully reached the destination device. Moreover, devices may leave the wireless network 14 due to scheduled maintenance and upgrades or due to unexpected failures, thus changing the topology of the mesh network and destroying some of the communication paths. Similarly, the devices may join the wireless network 14, adding additional direct connections 65. These and other changes to the topology of the wireless network 14 may significantly impact data transmissions between pairs of nodes if not processed in an efficient and timely manner.

Importantly, however, the efficiency of delivering data packets may largely determine the reliability, security, and the overall quality of plant operations. For example, a data packet including measurements indicative of an excessive temperature of a reactor should quickly and reliably reach another node, such as the hand-held device 55 or even the workstation 16, so that the operator or a controller may immediately take the appropriate action and address a dangerous condition if necessary. To efficiently utilize the available direct wireless connections 65 and properly adjust to the frequently changing network topology, the network manager 27 may maintain a complete network map 68, define a routing scheme that connects at least some pairs of network devices 30-50A-B, and communicate the relevant parts of the routing scheme to each network device that participates in the routing scheme.

In particular, the network manager 27 may define a set of directed graphs including one or more unidirectional communication paths, assign a graph identifier to each defined directed graph, and may communicate a relevant part of each graph definition to each corresponding network device, which may then update the device-specific, locally stored connection table 69A, 69B. As explained in more detail below, the network devices 30-50A-B may then route data packets based on the graph identifier included in the headers, trailers, etc. of the data packets. If desired, each connection table 69A, 69B may only store routing information directly related to the corresponding network device, so that the network device does not know the complete definition of a directed graph which includes the network device. In other words, the network device may not “see” the network beyond its immediate neighbors and, in this sense, the network device may be unaware of the complete topology of the wireless network 14. For example, the router device 60 illustrated in FIG. 1 may store a connection table 69B, which may only specify the routing information related to the neighboring network devices 32, 36, 50B, and 34. Meanwhile, the device 34 may store a connection table 69A, which accordingly may specify the routing information related to the neighbors of the device 34.

In some cases, the network manager 27 may define duplicate communication paths between pairs of network devices to ensure that a data packet may still reach the destination device along the secondary communication path if one of the direct connections 65 of the primary communication path becomes unavailable. However, some of the direct connections 65 may be shared between the primary and the secondary path of a particular pair of network devices. Moreover, the network manager 27 may, in some cases, communicate the entire communication path to be used to a certain network device, which may then originate a data packet and include the complete path information in the header or the trailer of the data packet. Preferably, network devices use this method of routing for data which does not have stringent latency requirements. As discussed in detail below, this method (referred to herein as “source routing”) may not provide the same degree of reliability and flexibility and, in general, may be characterized by longer delivery delays.

Another one of the core requirements of a wireless network protocol (and particularly of a wireless network operating in an unlicensed frequency band) is the minimally disruptive coexistence with other equipment utilizing the same band. Coexistence generally defines the ability of one system to perform a task in a shared environment in which other systems can similarly perform their tasks while conforming to the same set of rules or to a different (and possibly unknown) set of rules. One requirement of coexistence in a wireless environment is the ability of the protocol to maintain communication while interference is present in the environment. Another requirement is that the protocol should cause as little interference and disruption as possible with respect to other communication systems.

In other words, the problem of coexistence of a wireless system with the surrounding wireless environment has two general aspects. The first aspect of coexistence is the manner in which the system affects other systems. For example, an operator or developer of the particular system may ask what impact the transmitted signal of one transmitter has on other radio system operating in proximity to the particular system. More specifically, the operator may ask whether the transmitter disrupts communication of some other wireless device every time the transmitter turns on or whether the transmitter spends excessive time on the air effectively “hogging” the bandwidth. Ideally, each transmitter should be a “silent neighbor” that no other transmitter notices. While this ideal characteristic is rarely, if ever, attainable, a wireless system that creates a coexistence environment in which other wireless communication systems may operate reasonably well may be called a “good neighbor.” The second aspect of coexistence of a wireless system is the ability of the system to operate reasonably well in the presence of other systems or wireless signal sources. In particular, the robustness of a wireless system may depend on how well the wireless system prevents interference at the receivers, on whether the receivers easily overload due to proximate sources of RF energy, on how well the receivers tolerate an occasional bit loss, and similar factors. In some industries, including the process control industry, there are a number of important potential applications in which the loss of data is frequently not allowable. A wireless system capable of providing reliable communications in a noisy or dynamic radio environment may be called a “tolerant neighbor.”

Effective coexistence (i.e., being a good neighbor and a tolerant neighbor) relies in part on effectively employing three aspects of freedom: time, frequency and distance. Communication can be successful when it occurs 1) at a time when the interference source (or other communication system) is quiet; 2) at a different frequency than the interference signal; or 3) at a location sufficiently removed from the interference source. While a single one of these factors could be used to provide a communication scheme in the shared part of the radio spectrum, a combination of two or all three of these factors can provide a high degree of reliability, security and speed.

Still referring to FIG. 1, the network manager 27 or another application or service running on the network 14 or 12 may define a master network schedule 67 for the wireless communication network 14 in view of the factors discussed above. The master network schedule 67 may specify the allocation of resources such as time segments and radio frequencies to the network devices 25 and 30-55. In particular, the master network schedule 67 may specify when each of the network devices 25 and 30-55 transmits process data, routes data on behalf of other network devices, listens to management data propagated from the network manager 27, and transmits advertisement data for the benefit of devices wishing to join the wireless network 14. To allocate the radio resources in an efficient manner, the network manager 27 may define and update the master network schedule 67 in view of the topology of the wireless network 14. More specifically, the network manager 27 may allocate the available resources to each of the nodes of the wireless network 14 (i.e., wireless devices 30-36, 50A-B, and 60) according to the direct wireless connections 65 identified at each node. In this sense, the network manager 27 may define and maintain the network schedule 67 in view of both the transmission requirements and of the routing possibilities at each node.

The master network schedule 67 may partition the available radio sources into individual communication channels, and further measure transmission and reception opportunities on each channel in such units as Time Division Multiple Access (TDMA) communication timeslots, for example. In particular, the wireless network 14 may operate within a certain frequency band which, in most cases, may be safely associated with several distinct carrier frequencies, so that communications at one frequency may occur at the same time as communications at another frequency within the band. One of ordinary skill in the art will appreciate that carrier frequencies in a typical application (e.g., public radio) are sufficiently spaced apart to prevent interference between the adjacent carrier frequencies. For example, in the 2.4 GHz band, IEEE assigns frequency 2.455 to channel number 21 and frequency 2.460 to channel number 22, thus allowing the spacing of 5 KHz between two adjacent segments of the 2.4 GHz band. The master network schedule 67 may thus associate each communication channel with a distinct carrier frequency, which may be the center frequency in a particular segment of the band.

Meanwhile, as typically used in the industries utilizing TDMA technology, the term “timeslot” refers to a segment of a specific duration into which a larger period of time is divided to provide a controlled method of sharing. For example, a second may be divided into 10 equal 100 millisecond timeslots. Although the master network schedule 67 preferably allocates resources as timeslots of a single fixed duration, it is also possible to vary the duration of the timeslots, provided that each relevant node of the wireless network 14 is properly notified of the change. To continue with the example definition of ten 100-millisecond timeslots, two devices may exchange data every second, with one device transmitting during the first 100 ms period of each second (i.e., the first timeslot), the other device transmitting during the fourth 100 ms period of each second (i.e., the fourth timeslot), and with the remaining timeslots being unoccupied. Thus, a node on the wireless network 14 may identify the scheduled transmission or reception opportunity by the frequency of transmission and the timeslot during which the corresponding device may transmit or receive data.

As part of defining an efficient and reliable network schedule 67, the network manager 27 may logically organize timeslots into cyclically repeating sets, or superframes. As used herein, a superframe may be more precisely understood as a series of equal superframe cycles, each superframe cycle corresponding to a logical grouping of several adjacent time slots forming a contiguous segment of time. The number of time slots in a given superframe defines the length of the superframe and determines how often each time slot repeats. In other words, the length of a superframe, multiplied by the duration of a single timeslot, specifies the duration of a superframe cycle. Additionally, the timeslots within each frame cycle may be sequentially numbered for convenience. To take one specific example, the network manager 27 may fix the duration of a timeslot at 10 milliseconds and may define a superframe of length 100 to generate a one-second frame cycle (i.e., 10 milliseconds multiplied by 100). In a zero-based numbering scheme, this example superframe may include timeslots numbered 0, 1, . . . 99.

As discussed in greater detail below, the network manager 27 reduces latency and otherwise optimizes data transmissions by including multiple concurrent superframes of different sizes in the network schedule 67. Moreover, some or all of the superframes of the network schedule 67 may span multiple channels, or carrier frequencies. Thus, the master network schedule 67 may specify the association between each timeslot of each superframe and one of the available channels.

Thus, the master network schedule 67 may correspond to an aggregation of individual device schedules. For example, a network device, such as the valve positioner 34, may have an individual device schedule 67A. The device schedule 67A may include only the information relevant to the corresponding network device 34. Similarly, the router device 60 may have an individual device schedule 67B. Accordingly, the network device 34 may transmit and receive data according to the device schedule 67A without knowing the schedules of other network devices such as the schedule 67B of the device 60. To this end, the network manager 27 may manage both the overall network schedule 67 and each of the individual device schedules 67 (e.g., 67A and 67B) and communicate the individual device schedules 67A and 67 B to the corresponding devices when necessary. Of course the device schedules 67A and 67B are subsets of and are derived from the overall or master network schedule 67. In other embodiments, the individual network devices 25 and 35-50A-B may at least partially define or negotiate the device schedules 67A and 67 B and report these schedules to the network manager 27. According to this embodiment, the network manager 27 may assemble the network schedule 67 from the received device schedules 67A-B while checking for resource contention and resolving potential conflicts.

To optimally utilize the available wireless resources and to guarantee efficient and reliable delivery of data, the network manager 27 may further optimize scheduling decisions in view of routing or, alternatively, optimize routing in view of the scheduling decisions. In some particularly useful embodiments, the network manager 27 may conceptually combine routing concepts such as edges of directed graphs with scheduling resources such as timeslots and superframes to define links. The links may further include several distinct types such as dedicated links associated with a known pair of devices, shared links where least one of a transmitter or a receiver is not pre-assigned, broadcast and multicast links used for addressing multiple devices from a single source, etc. In these embodiments, the network manager 27 may efficiently use link resources by analyzing the topology of the network 14, forming a set of directed graphs to specify communication paths between pairs of network devices, defining superframes based on, in part, the update rates at the network devices, and sequentially allocating timeslots within the defined superframes to the directed connections between the devices to define a set of links specifying both the direction and the time of each data transmission. Additionally, each link may specify the channel on which a particular transmission occurs in those embodiments where the network 14 operates on several wireless channels. As discussed in detail below, the network manager 27 may thereby ensure that the network devices communicate both efficiently and reliably. For example, the network manager 27 may ensure that on a multi-hop communication path, a data packet spends as little time as possible before a transfer to the next hop in the path.

The communication protocol supporting the wireless network 14 generally described above is referred to herein as the WirelessHART protocol 70, and the operation of this protocol is discussed in more detail with respect to FIG. 2. As will be understood, each of the direct wireless connections 65 may transfer data according to the physical and logical requirements of the WirelessHART protocol 70. Meanwhile, the WirelessHART protocol 70 may efficiently support communications within timeslots and on the carrier frequencies associated with the superframes defined by the device-specific schedules 67A, 67B.

FIG. 2 schematically illustrates the layers of one example embodiment of the WirelessHART protocol 70, approximately aligned with the layers of the well-known ISO/OSI 7-layer model for communications protocols. By way of comparison, FIG. 2 additionally illustrates the layers of the existing “wired” HART protocol 72. It will be appreciated that the WirelessHART protocol 70 need not necessarily have a wired counterpart. However, as will be discussed in detail below, the WirelessHART protocol 70 can significantly improve the convenience of its implementation by sharing one or more upper layers of the protocol stack with an existing protocol. As indicated above, the WirelessHART protocol 70 may provide the same or greater degree of reliability and security as the wired protocol 72 servicing a similar network. At the same time, by eliminating the need to install wires, the WirelessHART protocol 70 may offer several important advantages, such as the reduction of cost associated with installing network devices, for example. It will be also appreciated that although FIG. 2 presents the WirelessHART protocol 70 as a wireless counterpart of the HART protocol 72, this particular correspondence is provided herein by way of example only. In other possible embodiments, one or more layers of the WirelessHART protocol 70 may correspond to other protocols or, as mentioned above, the WirelessHART protocol 70 may not share even the uppermost application layer with any existing protocols. It is also possible to use the WirelessHART protocol stack as a network layer under other communication standards such as Foundation Fieldbus, Profinet, Modbus TCP, and Internet IP. In these cases, the WirelessHART protocol 70 would be responsible for transferring real-time data, alarms, alerts, trends, and other information consistent with the HART communication standard.

As illustrated in FIG. 2, the wireless expansion of HART technology may add at least one new physical layer (e.g., the IEEE 802.15.4 radio standard) and two data-link layers (e.g., wired and wireless mesh) to the known wired HART implementation. In general, the WirelessHART protocol 70 may be a secure, wireless mesh networking technology operating in the 2.4 GHz ISM radio band (block 74). In one embodiment, the WirelessHART protocol 70 may utilize IEEE 802.15.4b compatible direct sequence spread spectrum (DSSS) radios with channel hopping on a transaction by transaction basis. This WirelessHART communication may be arbitrated using TDMA to schedule link activity (block 76). As such, all communications are preferably performed within a designated time slot. One or more source and one or more destination devices may be scheduled to communicate in a given slot, and each slot may be dedicated to communication from a single source device, or the source devices may be scheduled to communicate using a CSMA/CA-like shared communication access mode. Source devices may send messages to one or more specific target devices or may broadcast messages to all of the destination devices assigned to a slot.

Because the WirelessHART protocol 70 described herein allows deployment of mesh topologies, a significant network layer 78 may be specified as well. In particular, the network layer 78 may enable establishing direct wireless connections 65 between individual devices and routing data between a particular node of the wireless network 14 (e.g., the device 34) and the gateway 22 via one or more intermediate hops. In some embodiments, pairs of network devices 30-50A-B may establish communication paths including one or several hops while in other embodiments, all data may travel either upstream to the gateway device 22 or downstream from the gateway device 22 to a particular node.

To enhance reliability, the WirelessHART protocol 70 may combine TDMA with a method of associating multiple radio frequencies with a single communication resource, e.g., channel hopping. Channel hopping provides frequency diversity which minimizes interference and reduces multi-path fading effects. In particular, the data link 76 may create an association between a single superframe and multiple carrier frequencies which the data link layer 76 cycles through in a controlled and predefined manner. For example, the available frequency band of a particular instance of the WirelessHART network 14 may have carrier frequencies F1, F2, . . . Fn. A relative frame R of a superframe S may be scheduled to occur at a frequency F1 in the cycle Cn+2, at a frequency F5 in the following cycle Cn+1, at a frequency F2 in the cycle Cn+2, and so on. The network manager 27 may configure the relevant network devices with this information so that the network devices communicating in the superframe S may adjust the frequency of transmission or reception according to the current cycle of the superframe S.

The data link layer 76 of the WirelessHART protocol 70 may offer an additional feature of channel blacklisting, which restricts the use of certain channels in the radio band by the network devices. The network manager 27 may blacklist a radio channel in response to detecting excessive interference or other problems on the channel. Further, operators or network administrators may blacklist channels in order to protect a wireless service that uses a fixed portion of the radio band that would otherwise be shared with the WirelessHART network 14. In some embodiments, the WirelessHART protocol 70 controls blacklisting on a superframe basis so that each superframe has a separate blacklist of prohibited channels.

In one embodiment, the network manager 27 is responsible for allocating, assigning, and adjusting time slot resources associated with the data link layer 76. If a single instance of the network manager 27 supports multiple WirelessHART networks 14, the network manager 27 may create an overall schedule for each instance of the WirelessHART network 14. The schedule may be organized into superframes containing time slots numbered relative to the start of the superframe. Additionally, the network manager 27 may maintain a global absolute slot count which may reflect the total of number of time slots scheduled since the start-up of the WirelessHART network 14. This absolute slot count may be used for synchronization purposes.

The WirelessHART protocol 70 may further define links or link objects in order to logically unite scheduling and routing. In particular, a link may be associated with a specific network device, a specific superframe, a relative slot number, one or more link options (transmit, receive, shared), and a link type (normal, discovery, broadcast, join). As illustrated in FIG. 2, the data link layer 76 may be frequency-agile. More specifically, a channel offset may be used to calculate the specific radio frequency used to perform communications. The network manager 27 may define a set of links in view of the communication requirements at each network device. Each network device may then be configured with the defined set of links. The defined set of links may determine when the network device needs to wake up, and whether the network device should transmit, receive, or both transmit/receive upon waking up.

With continued reference to FIG. 2, the transport layer 80 of the WirelessHART protocol 70 allows efficient, best-effort communication and reliable, end-to-end acknowledged communications. As one skilled in the art will recognize, best-effort communications allow devices to send data packets without an end-to-end acknowledgement and no guarantee of data ordering at the destination device. User Datagram Protocol (UDP) is one well-known example of this communication strategy. In the process control industry, this method may be useful for publishing process data. In particular, because devices propagate process data periodically, end-to-end acknowledgements and retries have limited utility, especially considering that new data is generated on a regular basis. In contrast, reliable communications allow devices to send acknowledgement packets. In addition to guaranteeing data delivery, the transport layer 80 may order packets sent between network devices. This approach may be preferable for request/response traffic or when transmitting event notifications. When the reliable mode of the transport layer 80 is used, the communication may become synchronous.

Reliable transactions may be modeled as a master issuing a request packet and one or more slaves replying with a response packet. For example, the master may generate a certain request and can broadcast the request to the entire network. In some embodiments, the network manager 27 may use reliable broadcast to tell each network device in the WirelessHART network 14 to activate a new superframe. Alternatively, a field device such as the sensor 30 may generate a packet and propagate the request to another field device such as to the portable HART communicator 55. As another example, an alarm or event generated by the field device 34 may be transmitted as a request directed to the gateway device 22. In response to successfully receiving this request, the gateway device 22 may generate a response packet and send the response packet to the device 34, acknowledging receipt of the alarm or event notification.

Referring again to FIG. 2, the session layer 82 may provide session-based communications between network devices. End-to-end communications may be managed on the network layer by sessions. A network device may have more than one session defined for a given peer network device. If desired, almost all network devices may have at least two sessions established with the network manager 27: one for pairwise communication and one for network broadcast communication from the network manager 27. Further, all network devices may have one or several gateway session keys. The sessions may be distinguished by the network device addresses assigned to them. Each network device may keep track of security information (encryption keys, nonce counters) and transport information (reliable transport sequence numbers, retry counters, etc.) for each session in which the device participates.

Finally, both the WirelessHART protocol 70 and the wired HART protocol 72 may support a common HART application layer 84. The application layer of the WirelessHART protocol 70 may additionally include a sub-layer 86 supporting auto-segmented transfer of large data sets. By sharing the application layer 84, the protocols 70 and 72 allow for a common encapsulation of HART commands and data and eliminate the need for protocol translation in the uppermost layer of the protocol stack.

Scheduling Communications

FIGS. 3-6 provide a more detailed illustration of channel and timeslot resource allocation supported by the data link layer 76 and the network layer 78 of the WirelessHART protocol 70. As discussed above in reference to FIG. 1, the network manager 27 may manage the definition of one or more superframes and may associate individual timeslots within each of the defined superframes with one of the available channels (e.g., carrier frequencies). By way of one specific example, FIG. 3 illustrates a possible communication scheme within an individual timeslot, while FIG. 4 illustrates an example data exchange between several devices using the timeslots of a certain superframe. Next, FIG. 5 illustrates a possible association between an example timeslot and several available channels, and FIG. 6 is a schematic representation of several concurrent superframes which include the timeslots illustrated in FIGS. 3-5.

Referring specifically to FIG. 3, two or mode network devices may exchange data in a timeslot 100, which may be a dedicated timeslot shared by one transmitting device and one receiving device or a shared timeslot having more than one transmitter and/or one or more receivers. In either case, the timeslot 100 may have a transmit schedule 102 and a receive schedule 104. In other words, one or more transmitting devices may communicate within the timeslot 100 according to the transmit timeslot schedule 102 while one or more receiving devices may communicate within the timeslot 100 according to the receive timeslot schedule 104. Of course, the timeslot schedules 102 and 104 are substantially precisely synchronized and begin at the same relative time 106. Over the course of the timeslot 100, a transmitting network device sends a predetermined amount of data over a communication channel such as that defined by a particular carrier radio frequency. In some cases, the transmitting network device may also expect to receive a positive or negative acknowledgement within the same timeslot 100.

Thus, as illustrated in FIG. 3, the transmit timeslot schedule 102 may include a transmit segment 110 for transmitting outbound data, preceded by a pre-transmission segment 112, and may include a receive segment 122 for receiving an acknowledgement for the data transmitted during the segment 110. The transmit segment 110 may be separated from the receive segment 122 by a transition segment 116, during which the corresponding network device may adjust the hardware settings, for example. Meanwhile, the receive schedule 104 may include segments for performing functions complementary to those carried out in the segments 112-122, as discussed below.

In particular, the transmitting device may send out the entire packet or stream segment associated with a capacity of the timeslot 100 during the segment 110. As mentioned above, the network schedule 67 may include shared timeslots which do not exclusively belong to an individual device schedule 67A or 67B of one of the network devices 25A-B and 30-55. For example, a shared timeslot may have a dedicated receiver such as the gateway device 22 but no single dedicated transmitter. When necessary, one of the network devices 25-60 may transmit unscheduled information, such as a request for additional bandwidth, over the shared timeslot. In these cases, the potentially transmitting device may check whether the shared timeslot is available by performing Clear Channel Assessment (CCA) in a pre-transmission segment 112. In particular, the transmitting network device may listen to signals propagated over the communication channel associated with the timeslot 100 for the duration of the pre-transmission segment 112 to confirm that no other network device is attempting to use the timeslot 100.

On the receiving end of the timeslot 100, the receiving device may receive the entire packet associated with the timeslot 100 within a packet receive segment 114. As illustrated in FIG. 3, the packet receive segment 114 may begin at an earlier point in time than the transmit segment 110. If desired, the packet receive segment 114 may extend beyond the transmit segment 110 (not shown) to allow for slight timing mismatches. Next, the transmit timeslot schedule 102 requires that the transmitting device transition the radio mode in a transition segment 116. Similarly, the receive timeslot schedule 104 includes a transition segment 118. However, the segment 116 may be shorter than the segment 118 because the transmitting device may start listening for acknowledgement data early to avoid missing a beginning of an acknowledgement.

Still further, the transmit schedule 102 may include an acknowledgement receive segment 122 during which the transmitting device receives an acknowledgement transmitted during an acknowledgement transmit segment 124 associated with the receive schedule 104. The acknowledgement receive segment 122 may start prior to the acknowledgement transmit segment 124 and may, if desired, end later than the acknowledgement transmit segment 124 to reduce missing acknowledgments. The transmitting device may delete the packet transmitted during the transmit segment 110 from an associated transmit queue upon receiving a positive acknowledgement. On the other hand, the transmitting device may attempt to re-transmit the packet in the next scheduled dedicated timeslot or in the next available shared timeslot if no acknowledgement arrives or if the acknowledgement is negative.

Several timeslots 100 discussed above may be organized into a superframe 140, as schematically illustrated in FIG. 4. In particular, the superframe 140 may include a (typically) infinite series of superframe cycles 150-154, each cycle including a set of timeslots, illustrated in FIG. 4 as a timeslot 142 with a relative timeslot number 0 (TS0), a timeslot 144 with a relative timeslot number 1 (TS1), and a timeslot 146 with a relative timeslot number 2 (TS2). Accordingly, the size of the superframe 140 of FIG. 4 is three timeslots. In other words, each of the timeslots 142-146 of the superframe 140 repeats in time at an interval of two intermediate timeslots. Thus, for a 10 millisecond timeslot, the interval between the end of a timeslot with a particular relative slot number and the beginning of a next timeslot with the same relative slot number is 20 milliseconds. Conceptually, the timeslots 142-146 may be further grouped into superframe cycles 150-154. As illustrated in FIG. 4, each superframe cycle corresponds to a new instance of a sequence of timeslots 142-146.

The master network schedule 67 may associate transmission and reception opportunities of some of the network devices participating in the wireless network 14 with particular timeslots of the superframe 140. Referring again to FIG. 4, a network fragment 160 schematically illustrates a partial communication scheme implemented between the network devices 34, 60, and 36 of FIG. 1. To simplify the illustration of the superframe 140, the network devices 34, 60, and 36 are additionally designed in FIG. 4 as nodes A, B, and C, respectively. Thus, according to FIG. 4, the node A transmits data to the node B which, in turn, transmits data to the node C. As discussed above, each of the nodes A-C includes a device schedule 67A-C, which specifies the timeslots and channels (e.g., radio carrier frequencies) for transmitting and receiving data at the corresponding device. The master network schedule 67 may include part of all of the data information stored in the individual device schedules 67A-C. More specifically, the network manager 27 may maintain the master network schedule 67 as an aggregate of the schedules associated with each of the network devices 30-50A-B, including the device schedules 67A-67C.

In this example, the duration of the timeslot 100 (FIG. 3) may be 10 milliseconds and the network device A may report data to the device C every 30 milliseconds. Accordingly, the network manager 27 may set the length of the superframe 140 at three timeslots specifically in view of the update rate of the network device A. Further, the network manager 27 may assign the timeslot 142 with a relative number 0 (TS0) to the network devices A and B, with the device A as the transmitter and the device B as the receiver. The network manager 27 may further allocate the next available timeslot 144, having the relative slot number 1 (TS1), to be associated with the transmission from the device B to the device C. Meanwhile, the timeslot 146 remains unassigned. In this manner, the superframe 140 provides a scheme according to which the network manager 27 may allocate resources in the network fragment 160 for the transmission of data from the device A to the device C in view of the available wireless connections between the devices A, B, and C.

In the example illustrated in FIG. 4, the network device at node A may store information related to the timeslot 142 as part of its device schedule 67A. Similarly, the network device at node B may store information related to the timeslots 142 (receive) and 144 (transmit) as part of its device schedule 67B. Finally, the network device C may store information related to the timeslot 144 in the device schedule 67C. In at least some of the embodiments, the network manager 27 stores information about the entire superframe 140, including an indication that the timeslot 146 is available.

Importantly, the superframe 140 need not be restricted to a single radio frequency or other single communication channel. In other words, the individual timeslots 142-146 defining the superframe 140 may be associated with different radio frequencies on a permanent or floating basis. Moreover, the frequencies used by the various devices need not always be adjacent in the electromagnetic spectrum. In one embodiment, for example, the timeslot 142 of each of the superframe cycles 150-154 may be associated with a carrier frequency F1 and the timeslot 144 of each of the superframe cycles 150-154 may be associated with a carrier frequency F2, with the frequencies F1 and F2 being adjacent or non-adjacent in the electromagnetic spectrum.

In another embodiment, at least some of the timeslots 142-146 may move about the allocated frequency band in a predefined manner. FIG. 5 illustrates an example association of the timeslot 144 of FIG. 4 with channels 172-179 (corresponding to frequency sub-bands F1-F5) in the available frequency band 170. In particular, each of the channels 172-179 may correspond to one of the center frequencies F1, F2, . . . F5 which preferably differ from their respective neighbors by the same offset. The channels 172-179 may form a continuous section of the spectrum covering the entire available frequency band 170, although the channels 172-179 need be contiguous or form a continuous band in all embodiments. The superframe 140 may use at least a portion of the frequency band 170, so that one or more of the timeslots 142-146 are scheduled on different carrier frequencies in at least two consecutive cycles.

As illustrated in FIG. 5, the timeslot 144 may use the channel 176 (frequency F3) during the frame cycle 150, may use the channel 174 (frequency F4) during the frame cycle 152, and may use the channel 178 (frequency F2) during the frame cycle 154. The timeslot 144 may then “return” to the channel 176 in the next superframe cycle 150A, which may similar to the cycle 150. Each of the specific associations of the timeslot 144 with one of the channels 172-179 is illustrated as a timeslot/channel tuple 144A-C. For example, the tuple 144A specifies the timeslot 2 scheduled, in the cycle 150, on the channel 176 associated with the center frequency F3. Similarly, the tuple 144B specifies the timeslot 2 scheduled, in the cycle 152, on the channel 174 associated with the center frequency F4. Meanwhile, the channel 172 associated with the center frequency F5 may not be assigned to the timeslot 2 during any of the cycles 150-152. However, a different timeslot of the superframe 140 such as the timeslot 146, for example, may be associated with the channel 172 during one or more of the cycles 150-152.

In this example, the frequency assignment associated with the superframe cycle 150 may repeat immediately following the cycle 154 (illustrated as a cycle 150A in the FIG. 5), and the timeslot 144 may again correspond to the tuple 144A after two cycles of the superframe 140. Thus, the timeslot 144 may regularly cycle through the channels 176, 174, and 178. It will be appreciated that the timeslot 144 may similarly cycle through a greater or smaller number of channels irrespective of the length of the superframe 140, provided, of course, that enough channels are available in the frequency band 170. The association of a single timeslot with multiple channels during different superframe cycles, discussed above with respect to FIG. 5 and referred to herein as “channel hopping,” significantly increases the reliability of the wireless network 14. In particular, channel hopping reduces the probability that a pair of devices, scheduled to communicate in a particular timeslot of a certain superframe, fail to transmit and receive data when a certain channel is jammed or otherwise unavailable. Thus, for example, the failure of the channel 174 prevents the devices using the timeslot 144 from communicating in the frame cycle 152 but not during the frame cycles 150 or 154.

Referring again to FIG. 4, the device schedules 67B and 67C may include the information regarding each of the tuples 144A-144C discussed above in reference to FIG. 5. In particular, each of the device schedules 67B and 67C may store an assignment of the timeslot 144 to one of the channels 172-179 within each of the cycles 150-152. The master network schedule 67 (FIG. 1) may similarly include this information. Meanwhile, the device schedule 67A need not necessarily include the information related to the timeslot 144 because the corresponding node A (the device 34) does not communicate during the timeslot 144 of the superframe 140. In operation, the devices 60 and 36 corresponding to the nodes B and C may prepare for data transmission and reception, respectively, at the beginning of each timeslot 144. To determine whether the timeslot 144 currently corresponds to the tuple 144A, 144B, or 144C, the devices 60 and 36 may apply the global absolute slot count to determine whether the timeslot 144 is currently in the frame cycle 150, 152, or 154.

In the process of defining the network schedule 67, the network manager 27 may define multiple concurrent superframes in view of the update rates of the network devices 25 and 35-50A-B. As illustrated in FIG. 6, the network schedule 67 may include the superframe 140 of length three as well superframes 190 and 192. The superframe 190 may be a five-slot superframe and the superframe 192 may be a four-slot superframe, although the different superframes may have a different number of slots and various different superframes may have the same number of slots. As illustrated in FIG. 6, the superframes need not necessarily align with respect to the relative slot numbers. In particular, at a particular time 194, the superframe 190 may schedule the timeslot with the relative number two (TS2) while the superframes 140 and 192 may schedule the timeslots with the relative number one (TS1). Preferably, the superframes 140, 190, and 192 are time-synchronized on that each transition to a new timeslot within each of these superframes occurs at the same time.

Each of the superframes 140, 190 and 192 may be primarily associated with, or “belong to” an individual one of or a subset of the network devices 30-50A-B. For example, the superframe 140 illustrated in FIG. 4 may belong to the node A (i.e., the network device 34), and the length of the superframe 140 may be advantageously selected an that the node A sends out measurement data to the node B during the timeslot 142 (TS0) once during each of the cycles 150-154. In case the wireless network 14 defines 10 millisecond timeslot, the node A sends data to the node B once every 30 milliseconds. If, however, the node A is reconfigured to report measurements once every 50 milliseconds, the network manager 27, alone or in cooperation with the node A, may reconfigure the frame 140 to have a length of five timeslots instead. In other words, the length of each superframe may reflect a particular transmission requirement of a particular network device 30-50A-B.

On the other hand, more than one network device 30-50A-B may use a superframe for transmitting or receiving data. Referring again to FIG. 4, the node B (the network device 60) may regularly transmit data to the node C (the network device 36) in the timeslot 144 of the superframe 140, although the superframe 140 may be primarily associated with the node A. Thus, different timeslots of a particular superframe may be used by different network devices to originate, route, or receive data. In a sense, the timeslots of each superframe may be understood as a resource allocated to different devices, with a particular priority assigned to the device that “owns” the superframe. Further, it will be appreciated that each network device may participate in multiple superframes. For example, the network device 34 in FIG. 4 may route data on behalf of other network devices (e.g., the network device 32 illustrated in FIG. 1), in addition to propagating its own data via the router device 60. Preferably, a device participating in multiple superframes does not schedule simultaneous communications in different superframes. While only three superframes are illustrated in FIG. 6, the wireless network 14 of FIG. 1 may include any number of superframes, with each of the different superframes having any desired or useful length based on the types and frequencies of communication being performed in or between particular devices and set of devices.

The methods described above with respect to FIGS. 1-6 may be applied to a process control system, for example, in which various devices report measurements or other data according to individual device schedules, as well during occasional, frequently unpredictable, data “bursts.”

Routing Techniques

As generally discussed above in reference to FIG. 1, it is important to consider the location of the network devices 30-50A-B so that the wireless network 14 can establish itself in an efficient and reliable form. In some cases, it may be necessary to add routers 60 in those locations where plant equipment could block or seriously affect a wireless connection or where network devices are spaced far apart from one another. Thus, in this and in similar situations, it is desirable that the wireless network 14 be “self-healing,” i.e., capable of automatically addressing at least some of the delivery failures. To meet this and other design requirements, the wireless network 14 may define redundant paths and schedules so that in response to detecting a failure of one or more direct wireless connections 65, the network 14 may route data via an alternate route. Moreover, the paths may be added and deleted without shutting down or restarting the wireless network 14. Because some of the obstructions or interference sources in many industrial environments may be temporary or mobile, the wireless network 14 may be capable of automatically reorganizing itself. More specifically, in response to one or more predetermined conditions, pairs of network devices (e.g., 32 and 34, 60 and 36, etc.) may recognize each other as neighbors and thus create anew direct wireless connection 65 or, conversely, dissolve the existing direct wireless connections 65. Additionally, the network manager 27 may create, delete, or temporarily suspend paths between non-neighboring devices.

Irrespective of whether a particular network configuration is permanent or temporary, the wireless network 14 generally requires a fast and reliable method of routing data between nodes. In one possible embodiment, the network manager 27 may analyze the information regarding the layout of the network, the transmission capability and update rate of each network device 32-50A-B, as well as other relevant information. The network manager 27 may then define routes and schedules in view of these factors. When defining routes and schedules, the network manager 27 may recognize the wireless network 14 as conforming to one of several network topologies compatible with the routing and scheduling techniques of the present disclosure.

FIGS. 7-9 schematically illustrate some of these network topologies. For the sake of clarity, each of FIGS. 7-9 illustrates bidirectional connections between pairs of devices. However, it will be appreciated that each of the topologies illustrated in FIGS. 7-9 is also compatible with unidirectional connections or mixed bidirectional and unidirectional connections (i.e., including both bidirectional and unidirectional connections). Moreover, each connection illustrated in FIGS. 7-9 may support several unidirectional connections in one or both directions, with each unidirectional connection associated with a particular time of transmission, for example. Referring specifically to FIG. 7, a network 350 may have a star network topology. The star network 350 includes a routing device 352 and one or more end devices 254. The routing device 352 may be a network device arranged to route data while the end device 254 may be a network device arranged to send data only on its own behalf and to only receive (or decode) data addressed to the end device 254. Of course, the routing device 352 may also be a recipient and originator of data and may perform routing functions in addition to other tasks. As illustrated in FIG. 7, end devices 254 may have a direct connection 265 to the routing device 352 but end devices 254 cannot be connected directly in a star topology. The direct connection 265 may be a direct wireless connection 65 or a wired connection.

The end device 254 may be the same type of physical device as the routing device 352 and may be physically capable of routing data. The routing capability of the end device 254 may be disabled during the installation of the end device 254 or in operation of a corresponding network (such as the WirelessHART network 14). Moreover, the routing capability of the end device 254 may be disabled by the end device 254 itself or by a dedicated service such as the network manager 27. In some sense, the star network 350 corresponds to the simplest of possible topologies. It may be appropriate for small applications that require low power consumption and low latency. Additionally, it will be noted that the star network 350 is deterministic because there is only one possible route between the routing device 352 and a particular end device 254.

Now referring to FIG. 8, a network 270 is arranged in a mesh network topology. Each network device of the mesh network 270 is a routing device 352. Mesh networks provide a robust network with multiple paths between various devices. In wireless applications, mesh networks are better able to adapt to changing radio environments. For example, the device 274 of the network 270 may send data to the device 276 via an intermediate hop 278 or an intermediate hop 380, provided that the corresponding paths 382-388 allow transmission in this direction. As illustrated in FIG. 8, both a path 382 and a path 384 enable the routing device 274 to send data to the routing device 276, providing redundancy and thus improved reliability to the network 270.

Another type of network topology is illustrated in FIG. 9. The network 390 incorporates elements of both star and mesh topologies. In particular, the star mesh network 390 includes several routing devices 352 (labeled “R”) and end devices 254 (labeled “E”). The routing devices 352 may be connected in a mesh format and may support redundant paths. The selection of a particular topology may be performed automatically by a network component, such as the network manager 27, or by a user configuring the network. In particular, the user may choose to override the topology selected by the network manager 27 or the default topology associated with the WirelessHART protocol 70. It is contemplated that in most applications, mesh topology may be the default topology because of the inherent reliability, efficiency, and redundancy of this topology. Clearly, because WirelessHART devices may act as router devices, several different configurations may be compatible with the same physical disposition of field devices and routers.

Both source routing and graph routing may be applied to the topologies discussed in reference to FIGS. 7-9. Although both types of routing may be equally useful in different situations, graph routing will be discussed first. Generally, in mathematical theories and applications, a graph is a set of vertices (nodes such as 252 or 254) and edges (direct connections 65 or 265). The WirelessHART protocol 70 or another protocol servicing the wireless network 14 may use graphs to configure paths connecting communication endpoints such as the device 30 to the gateway 22 illustrated in FIG. 1, for example. In some embodiments, graphs and the associated paths are configured by the network manager 27. The network manager 27 may also configure individual network devices such as field devices 30-40, routers 60, etc. with partial graph and path information, which may be stored in the connection tables 69. The wireless network 14 may contain multiple graphs, some of which may overlap. Further, a certain network device may have paths of multiple graphs going through the device, and some of the paths may direct data to the same neighbor of the device. Preferably, every graph in a network is associated with a unique graph identifier.

The protocol servicing the wireless network 14 (such as the WirelessHART protocol 70) may be configured to operate with a number of different topologies to support various application requirements. As a result, the wireless network 14 may concurrently support several methods of routing, such as unidirectional graph routing and source routing, for example. Although the forthcoming examples of a wireless network support these two approaches, it will be appreciated that the wireless network 14 may additionally support bidirectional graph routing, or may route data using only one of these techniques. However, irrespective of a type and number of concurrent routing techniques, each device on the wireless network 14 may be assigned a unique network address. Once every potential receiver of data acquires some form of unambiguous identification with respect to other network elements, decisions related to routing may be made by individual devices such as field devices 30-40, by a centralized dedicated service such as the network manager 27, or by individual devices acting in cooperation with the centralized service. As indicated above, at least one possible implementation of the wireless network 14 may rely on the network manager 27 to carry out most or all of the routing decisions and to communicate the relevant data to the network devices 30-50A-B to be stored in the connection tables 69. Further, routing decisions can be made at the originating point (i.e. at the source of a data packet) or at a centralized location. Moreover, routing decisions can be adjusted at each intermediate stop, or “hop,” in the path of the packet from the source to a destination.

In the examples discussed below, a wireless network provides at least two approaches to routing that may be selected according to the specific requirements and conditions of a given system, such as the physical layout of the network elements that make up the system, the number of elements, the expected amount of data to be transmitted to and from each element, etc. Moreover, the two approaches may be used by the wireless network at the same time and each may be selectively applied to a particular type of data or to a particular host or a set of hosts in view of certain aspects of performance of each of the two approaches. For example, a measurement of a process variable or a command to open a valve may tolerate a relatively small delay in delivery and the wireless network 14 may accordingly apply the faster and the more reliable of the two methods. Meanwhile, a device configuration command or a response may tolerate a longer delay and may be suitable for the other approach.

As briefly indicated above, it is common for a certain distributed control networks and, in particular, to networks connecting devices in the process control industry, to direct data to a certain device for management, diagnostic, log collection, and other purposes. FIGS. 10-12 illustrate several perspectives of a wireless network 300 which implements data transfer in two general directions: toward a gateway 305A or 305B (referred to herein as the “upstream” direction) and away from the 305A or 305B (referred to herein as the “downstream” direction). For security reasons, the network 300 does not allow direct data transfer between peer field devices although the technique described herein could be used in such a situation if so desired.

FIG. 10 illustrates upstream routing in the network 300. In particular, the network manager 302A (or the stand-by network manager 302B) may define several directed graphs, each graph including either the gateway device 305A or a second gateway device 305B as the terminal node. In other words, the paths of each graph in the exemplary network 300 lead to and terminate at one of the two gateway devices 305A or 305B. Specifically, a graph 310 (shown in solid bold arrows) may include network devices 312, 314, 316, 318, and the gateway 305A wherein the paths associated with the graph 310 may include direct wireless connections 320, 322, 324, 326, and 328. A graph 340 (shown in dotted bold arrows) may include network devices 312, 316, 318, 342, and the gateway 305B, with a path that includes direct wireless connections 344, 346, 348, 350, and 352. In the directed graph 310, the network device 312 may be called the head of the directed graph 310 and the gateway device 305A may be called the tail of the directed graph 310. Similarly, the network device 312 is the head of the directed graph 340 and the gateway device 305B is the tail of the directed graph 340. The network manager 302A or, under certain operating conditions, a backup network manager 302B may define the graphs 310 and 340 and may communicate complete or partial definitions of these graphs 310 and 340 to the network devices 312-318 and 342. As discussed above in reference to FIG. 1, the network devices 312-318 and 342 may maintain up-to-date versions of the connection tables 69 storing these partial path definitions. In some embodiments, the gateway devices 305A-B may not require the information regarding the graphs 310 and 340 if the corresponding communication path terminates at one of the gateway devices 305A-B. However, it will be appreciated that the gateway devices 305A-B may also originate data and may store information regarding one or more graphs with paths originating at the gateway devices 305A-B. It will be further noted that, in general, a path of a certain graph may traverse the gateway device 305A or 305B as an intermediate node. However, the exemplary network 300 defines paths that always either originate or terminate at one of the gateway devices 305A or 305B.

To send a data packet along a certain graph, a source network device may include an identifier of the graph in the header or trailer of the data packet. The data packet may travel via the paths corresponding to the graph identifier until it either reaches its destination or is discarded. To be able to route packets in the graph 310, for example, a connection table 69 of each network device that belongs to the graph 310 may contain entries that include the graph identifier and address of a neighbor network device which (1) belongs to the same graph, and (2) is one hop closer to the destination. For example, the network device 316 may store the following connection table:

GRAPH IDENTIFIER NODE GRAPH_310 318 GRAPH_340 318 GRAPH_340 342

while the network device 342 may store the following information in the connection table:

GRAPH IDENTIFIER NODE GRAPH_340 305B

While the exemplary connection tables above simply list the devices associated with a particular entry, it will be noted that the NODE column of the connection table may store the address of the neighboring device as defined in the addressing scheme of the network 300 or WirelessHART network 14.

In another embodiment, the NODE column may store the nickname of the neighboring device, an index into an array storing full or short addresses of the neighbors, or any other means of unambiguously identifying a network device. Alternatively, the connection table may store graph identifier/wireless connection tuples as illustrated below:

GRAPH IDENTIFIER CONNECTION GRAPH_310 326 GRAPH_340 346 GRAPH_340 348

In other words, the connection table may list one or more direct wireless connections 65 corresponding to a particular graph. The network device 316 may, for example, consult the connection table and transmit a packet carrying the graph identifier 340 via the direct wireless connection 346 or 348.

As illustrated in FIG. 10 and in the tables above, redundant paths may be set up by having more than one neighbor associated with the same graph identifier. Thus, a data packet arriving at the network device 316 and containing the graph identifier 340 in the header or trailer may be routed to either the network device 318 or to the network device 342. While executing a routing operation, the network device 316 may perform a lookup in the connection table by the graph identifier 340, and send the packet to either (or both) of the network devices 318 or 342. Moreover, the routing selection between two or more possible hops may be random or may be carried out according to a predefined algorithm. For example, the selection may be made in consideration of a load balancing objective or in view of the delivery statistics. Thus, the network device 316 may learn, through a peer network device or from the network manager 27, that selecting the network device 318 as the next hop while routing packets along the graph 340 has a lower probability of delivering the packet successfully or has a longer expected or average delay in delivery. The network device 316 may then attempt to route more or possibly all of the packets associated with the graph 340 to the network device 342.

In one embodiment, a neighbor device acknowledges the receipt of a data packet by sending a confirmation packet. In the example above, once the neighboring network device 318 or 342 acknowledges receipt of the packet, the network device 316 may immediately release it. If, on the other hand, the acknowledgement is not received within a predefined time period, the network device 316 may attempt to resend the packet or to route the packet via the alternate hop or path. Additionally, the network device 316 may collect statistics of both successful delivery attempts and of failed delivery attempts. The subsequent routing decisions, such as selecting between the hops 318 and 342, may include or be based on the adjusted statistical data. Of course, the network device 316 may apply the statistics related to network devices 318 and 342 to other relevant graphs and may also communicate the statistics to other network devices, either directly or via the network manager 27.

As discussed above, in the graph routing approach, a network device sends packets with a graph identifier in a network header along a set of paths to the destination. Importantly, a graph identifier alone is sufficient for routing packets and, while other routing information may be also included in the header, each packet can be properly delivered based solely on the graph identifier. All network devices on the way (i.e., on the path) to the destination may be pre-configured with graph information that specifies the neighbors to which the packets may be forwarded. Because graph routing requires pre-configuration of intermediate network devices for each potential destination, graph routing may be better suited for communications from a network device to a gateway and from a gateway to a network device.

Now referring to FIG. 11, the network manager 302A or 302B may also support routing downstream with respect to one or both of the gateways 305A-B. In particular, a graph 380 (shown in solid bold arrows) may include the nodes 305A, 315, 314, and 312, and the direct wireless connections 382-386. The gateway device 305A is the head of the graph 380 and wireless device 312 is the tail of the graph 380. Meanwhile, a graph 390 (shown in dotted bold arrows) may similarly connect the gateway device 305A to the wireless device 312, with the gateway device 305A as the head of the graph 390. However, the graph 390 may include the nodes 305A, 318, 342, 316, and 312, and the direct connections 392-298. Thus, to send a data packet to the wireless device 312, the gateway device 305A may include a graph identifier in the header or the trailer of the data packet which corresponds to either the graph 380 or 390. It will be appreciated that each of the graphs 380 or 390 may also include duplicate connection paths to ensure reliability and that, in general, the network manager 302A or 302B may use techniques similar to those discussed above in reference to FIG. 10. Also, it will be noted that the connection table 69 of each of the wireless devices 312-318 and 342 may include graph route information related to both downstream and upstream graphs used for routing purposes.

As illustrated in FIG. 12, the wireless network 300 may additionally use source routing. In source routing, pre-configuration of the relaying devices is not necessary. To send a packet to its destination using source routing, the source network device may include, in the header of a data packet, for example, an ordered list of devices through which the data packet must travel. The ordered list of devices may effectively define a communication path for the data packet. As the packet traverses the specified path, each routing device may extract the next node address from the packet to determine where the data packet should travel next, i.e., where the next data packet should be sent in the next hop. Consequently, source routing requires advance knowledge of the topology of the network 14. If, however, a certain network device does not find itself on the routing list, the network device may send the packet back to the first device specified in the source routing list. Source routing allows packets to go to an arbitrary destination without an explicit or preconfigured setup of intermediate devices.

For example, the network device 312 may send a packet to the gateway device 305A by specifying the complete path in the packet header or the packet trailer. As illustrated in FIG. 12, the network device 312 may generate a list 410 containing the addresses of network devices 314, 315, and 305A and send the list 410 along with the packet to the first hop or device on the list, i.e., the network device 314. The network device 314 may then traverse the list 410, locate the identity of the network device 314, extract this field from the list 410, identify the network device 315 as the next hop for the received packet, and finally send the data packet to the network device 315. The source routing list may reside in the optional area of the network header, and may be of variable size depending on number of hops to the destination. Similarly, the network device 315 may traverse the list 410, locate its own address or identity, and send the data packet to the next hop or device in the list 410 (in this case, the gateway device 305A).

In general, only those network devices that have obtained full network information from the network manager 27 or 302A-B use source routing because only the network manager 27 or 302A-302B knows the complete topology of the network. An additional limitation of source routing is that it provides no redundancy at intermediate network devices because each packet is originated with a header or a trailer that explicitly specifies each intermediate hop and does not provide any routing alternatives. Thus, if one of the intermediate network devices fails to relay the packet as specified by the packet header or trailer, the delivery of the packet fails. It is therefore the responsibility of the network manager 27 or 302A-302B to detect the failure and reprogram or reconfigure the source with an alternate route. To facilitate the detection of such error cases, the wireless network 14 or 300 may require network devices to send a routing failure notification to the network manager 27 or 302A-302B. Accordingly, a protocol such as the WirelessHART protocol 70 may provide a message type or an information element in the protocol definition for reporting this and other types of delivery failures. In another embodiment, the routing list 410 (referring to FIG. 12) may specify alternate routes in addition to the route selected by the sender. In yet another embodiment, primary and one or more alternate routes may be partially merged to avoid duplication of common parts of the path in the packet header or trailer.

Referring generally to FIGS. 1, 3, and 10-12, the network manager 27 or 302A-302B may maintain a list of all devices in the network. The network manager 27 or 302A-B may also contain the overall network topology including a complete graph of the network and the up-to-date portions of the graph that have been communicated to each device. The network manager 27 or 302A-302B may generate the route and connection information using the information that the network manager 27 receives from the network devices 30-40, 50A-B, 60, 55, etc. The network manager 27 or 302A-302B may then build the graph of the network from the list of network devices and the neighbors reported by each network device. Referring back to FIG. 1, for example, the network device 50B may report “seeing” the neighbor devices 60 and 34. The network manager may be also responsible for generating and maintaining all of the route information for the network. In one embodiment, there is always one complete network route and several special purpose routes which are used to send setpoint and other settings from the gateways 305A or 305B to the final control commands (FIGS. 7-9). Further, broadcast routes (which flow through most or all of the devices in the network) may be used to send broadcast messages from the network manager 27, 114, or 302A-B to all of the devices of the network 14 or 300. Still further, the network manager 27, 111, or 302A-302B may also carry out the scheduling of network resources once the routing information and burst mode update rates are known.

When devices are initially added to the network 14 or 300, the corresponding network manager may store all neighbor entries as reported from each network device. The network manager 27 or 302A-302B may use this information to build an initial complete network graph 68 and to revise the graphs during operation. The network graph 68 is put together optimizing several properties including hop count, reporting rates, power usage, and overall traffic flow as reflected by the statistics gathering discussed above. One key aspect of the topology is the list of connections that connect devices together. Because the presence and health of individual connections may change over time, the network manager 27 or 302A-302B may be additionally programmed or configured to update the overall topology, which may include adding and deleting information in each network device. In some embodiments, only the network manager 27 or 302A-302B and the gateway 22 or 302A-302B may know enough information to use source routing. More specifically, it may be desirable to prevent peer-to-peer communication between any two arbitrary devices for security purposes. Moreover, while graph and source routing have been described herein as occurring between a field device and a gateway device, both of these types of routing can be used between any two devices in the network, including, for example, between any two field or network devices in the network any two gateway devices, etc.

In short, graph routing may direct traffic both upstream and downstream with respect to the network manager 27 or gateway 22 and both graph and source routes can be optimized to satisfy applications with low latency requirements, which includes measurement information that is transferred from network devices to the gateway and control information that is transferred from gateway devices to final control commands such as regulating valves, on-off valves, pumps, fans, dampers, as well as motors used in many other ways.

In some embodiments, path redundancy may be a matter of policy of the network manager 27 or 302A-302B rather than a coincidental overlap of graphs. In other words, the network manager 27 or 302A-302B may attempt to define at least two neighbors for each device. Thus, the network manager 27 or 302A-302B may be configured to actively pursue a mesh or a star mesh topology. The supporting protocol, such as the WirelessHART protocol 70, may thus provide a very high end-to-end data reliability. From the physical perspective, each field device or other network device should be within communication range of at least two other devices that can receive messages from the field device and forward these messages.

The network manager 27 or 302A-302B may additionally verify each graph definition in order to ensure that no loops have been formed. In those embodiments where the network manager 27 or 302A-302B actively pursues path redundancy and defines many graphs of various size, a communication path may be sometimes erroneously defined to direct data packets from a source back to the same source. In accordance with such faulty graph definition, a packet may be routed back to the source directly from the source or may visit one or more intermediate hops prior to arriving back at the source. Loop verification may be performed each time the topology of the associated network changes, such as due to an addition or removal of a device, or whenever the network manager 27 adjusts the routing graphs and schedules for any reason. Alternatively, the network manager 27 may perform loop checking periodically as a background task.

Combining Routing and Scheduling Decisions

In a wireless network such as the wireless network 14 or 300, the same graph route may be used with several schedules. In particular, a packet may be transmitted along the same route while the network schedule changes and the time at which the packet is sent from and to a certain node is updated. In this sense, routing and scheduling may be conceptually and functionally separate to facilitate network management. On the other hand, however, the network manager 27 may execute routing and scheduling substantially in parallel to achieve robustness and improve the performance and reliability of the wireless network 14 or 300. More specifically, the network manager 27 may make at least some scheduling decisions in view of the relevant routing constraints and, conversely, make routing decisions in view of the scheduling constraints. In some particularly useful embodiments, the network manager 27 may analyze the topology of the network 14 or 300 first, construct the network graph 68, and then proceed to build a network schedule 67 in view of the network graph 68 and device-specific parameters such as transmission rate, power capability, etc.

When making scheduling decisions in view of a network topology, the network manager 27 may conceptually combine a timeslot in a particular superframe with an edge of a particular direct graph (which is a direct connection 65 in the examples discussed herein) to define a convenient combination space and time unit, i.e., a link. In particular, a link may be associated with a direct connection 65 between two communication endpoints as well as with a time during which these two communication endpoints exchange data.

Further, the network manager 27 may associate slots with several types corresponding to different principles of allocating and using the slots. In particular, a dedicated unicast link may be shared by a particular pair of network devices 32-50A-B, so that one of the pair of network devices transmits information to the other one of the pair of network devices with a specified timeslot. Of course, as discussed above with respect to FIG. 3, a network device transmitting certain information during a timeslot may also receive a corresponding acknowledgement from the device receiving this information and in this sense, each of the pair of network devices acts as a transmitter and a receiver in the course of a single timeslot. However, for the sake of simplicity, the device transmitting a block of information in a timeslot is referred to herein as a “talker” and the device receiving this block of information accordingly is referred to as a “listener.”

In contrast to a dedicated link, a shared link may have more than one talker but only one listener. In a sense, a shared link is still a unicast link because this link has only one listener. On the other hand, broadcast and multicast links may have one talker and many listeners. Further, directed links have one talker and a limited set of listeners.

In another aspect, a particular network device may view a dedicated, shared, directed, or broadcast link as a transmit link or as a receive link. Referring back to FIG. 10, for example, the wireless devices 312 and 314 may share a dedicated link associated with some particular timeslot and the direct wireless connection 320, so that the wireless device 312 transmits data via this link and the wireless device 314 receives data via this link. Accordingly, the wireless device 312 may consider this link a transmit link whereas the wireless device 314 may consider the same link a receive link. Thus, classification of links into transmit and receive links is a matter of a device perspective. It will be also noted that some links may be defined and used as transmit/receive links if the devices at either end of a graph edge associated with the link transmit information at some point during the corresponding timeslot.

As indicated above, each link may be associated with a certain timeslot irrespective of a link type. In operation of the wireless network 14 or 300, the network manager 27 or 302A-302B may assign links to network devices 30-50A-B, 312, 314, 316, 318, etc. Depending on the link type, the network device associated with the link may transmit a packet, receive a packet, or remain idle. A network device that operates with a transmit link or a transmit/receive link may send a packet within the appropriate segment of the link if the destination of the packet matches the one or more neighbors on the other end of the link. On other hand, a network device that has a receive link, or a transmit/receive link with no packet to send, listens for one or more incoming packets within the timeslot associated with the link. In those embodiments that also use shared links, a device preferably performs a clear channel assessment (CCA) or another method of preventing resource contention before the device begins to transmit via a shared link. In at least some embodiments, all devices that participate in either dedicated or shared links must be awake and listening during the timeslots associated with these dedicated or shared links.

As explained above in reference to FIG. 3, in the course of a single timeslot, one network device may send a data packet and another device may reply with an acknowledgement. Similarly, a communication session via a link may include transmission of a data packet and of an acknowledgement which may be positive (“ACK”) or negative (“NACK”). In general, a positive acknowledgement may indicate that the receiver has successfully received the data packet and has assumed ownership of the data packet for further routing if the receiver is not the final destination for this packet. Meanwhile, a negative acknowledgement may indicate that the receiver cannot accept the data packet at this time but has detected the packet with no errors. Further, both ACK's and NACK's may carry timing information so that the corresponding devices may properly maintain network synchronization. In some embodiments, packets sent to a unicast network device address may require a link-layer acknowledgement within the same timeslot while packets sent to a broadcast network device address (e.g., 0xFFFF) may not require an acknowledgement.

When the network manager 27, for example, defines direct connections 65 between network devices 30-50A-B, these network devices receive link assignments. The devices may accordingly define and maintain the respective device schedules 67A, 67B (see FIGS. 1 and 4). A link assignment may specify, in part, how the network device should use a certain slot in a superframe. Thus, each link may include exactly one timeslot, type assignment (i.e., transmit and/or receive), neighbor information or other data identifying the edge of a bidirectional or unidirectional graph with which the link is associated, and additional transmit and/or receive attributes.

In some embodiments, the device schedule 67A, 67B of each network device 30-50A-B may maintain additional flags or indicators to properly maintain links of various types. For example, a device schedule 67A, 67B may set a shared flag for each shared link so that the corresponding network device 30-50A-B may properly access this link for transmissions. Further with respect to shared links, the network 14 or 300 may use the well-known slotted Aloha contention management algorithm to define the lifecycle of shared link. Accordingly, the network devices 30-50A-B, 305A-305B, 312, 314, 316, 318, etc. may use a collision-avoidance scheme with a backoff (delay) in collision situations. In some embodiments, the delay may be implemented as a time measure unrelated to the duration of a single timeslot. In other particularly useful embodiments, the backoff may be implemented as a delay measured in an integer number of timeslots. In particular, a device that has encountered a collision may back off by the duration of a single timeslot, two timeslots, etc. starting at the beginning of the next scheduled timeslot. By synchronizing backoff intervals with timeslots, devices may optimize the retry mechanism and ensure that retry attempts happen only when there is a possibility of transmission. Using shared links may be desirable when bandwidth requirements of devices are low, and/or traffic is irregular or comes in bursts. In some situations, using shared links may decrease latency because the network device does not need to wait for dedicated links, although this generally true only when chances of collisions are relatively low.

Forming Efficient Schedules and Routing Schemes

In addition to optimizing routing by analyzing the network topology, the network manager 27 may define graphs and allocate resources during scheduling also in view of the type of data that a particular network device may transmit and, for each type of data, the expected frequency of transmission at each particular device. More specifically, the WirelessHART protocol 70 may support several types of network communication traffic. Both the existing HART protocol 72 and the WirelessHART protocol 70 support exchanging request/response data, publishing of process data, sending broadcast messages, and block data transfer of large data files. The WirelessHART protocol 70 may also support transmission of management data, such as network configuration data, and device communications, such as periodic measurements reported by field devices, using the same protocol and the same pool of resources, thus allowing for greater efficiency in scheduling and routing.

The network manager may allocate communication resources to each network device according to the amount of data which the network device may publish per unit of time. For example, the WirelessHART flowmeter 30 in the WirelessHART network 14 may have an update rate of four seconds while the WirelessHART pressure sensor 32 may have an update rate of ten seconds. An operator may configure the network devices 30 and 32 with these values according to the specific needs of the process control system in which the WirelessHART network 14 is implemented. As indicated above, multiple superframes may be used to define a different communication schedule for various network devices or groups of network devices. Initially, the network manager 27 may reserve a particular superframe for all network manager requests. Upon accepting network devices such as the flowmeter 30 and the pressure sensor 32, the network manager 27 may allocate additional superframes for the communication rates of four and ten seconds and assign the additional superframes to the network devices 30 and 32. The network manager 27 may also define superframes for alarms and network events prior to or after adding the network devices 30 and 32 to the WirelessHART network 14, respectively. The network devices 30 and 32 may, but are not required to, participate in one or more superframes simultaneously. By configuring a certain network device to participate in multiple overlapping superframes of different sizes, it is possible to establish multiple communication schedules and connectivity matrices which may work concurrently without scheduling conflicts. Moreover, because some key applications such as asset management and device specific applications often require considerable bandwidth for short durations of time, the network manager 27 may also create additional temporary superframes as needed. For example, a user may issue a request to view or change the configuration of a device or create a diagnostic screen. The network manager 27 may support this temporary increase in demand for communication slots by defining additional superframes with additional timeslots which may stay active for, by way of example only, several minutes.

The network manager 27 may consider both the update rates of the network devices and the topology of the WirelessHART network 27 when creating directed graphs. However, the network manager 27 may also carry out graph routing decisions separately from scheduling. For example, the network manager may add, delete, or update graphs while keeping the network schedule intact. More specifically, the network schedule may have available timeslots in the defined superframes which the network manager 27 may use as resources when defining new graphs or updating the existing graphs. In this sense, the WirelessHART protocol 70 allows graph configuration decisions to proceed separately from scheduling decisions. This feature of the WirelessHART protocol 70 can make the WirelessHART network 14 more responsive to changes in the environment and in the operational states of the network devices because the WirelessHART protocol 70 may quickly and non-intrusively adjust the routing by changing only part of the existing configuration.

It will be appreciated that some of the methods discussed above need not be restricted to data packets and may be applied to other communication techniques. For example, a network may use a circuit-switched approach and instead of traveling in packets of a finite size, the data may be transmitted as a stream over a dedicated channel between communication endpoints.

By way of a simplified example, FIGS. 13-20 and the accompanying description illustrate the formation of an efficient routing scheme and of a corresponding communication schedule applying the concepts discussed above to a four-node mesh network. In particular, FIG. 13 schematically illustrates a network 400, several direct wireless connections formed between nodes A, B, C, and D (or network devices 402-408) of the network 400, and the edges 410-420 of several directed graphs which may be formed in the network 400 in view of such as factors as network topology, signal strength and/or quality, latency requirements, etc. To further simplify this example, only upstream (i.e., device-to-gateway) data propagation is discussed in reference to FIGS. 13-20. For the sake of clarity, the network nodes are referred to below as nodes A-D whereas the devices corresponding to these nodes are referred to below as the gateway device 402 or the field devices 404-408. As further illustrated in FIG. 13, the nodes A-D may store, maintain, and update device-specific connection tables 422-428. Next, FIGS. 14-16 illustrate the formation of several superframes to support data exchange between the nodes A-D. Finally, FIGS. 17-20 illustrate example device-specific communication schedules 432-438 which may correspond to the nodes A-D of FIG. 13.

Referring specifically to FIG. 13, the wireless network 400 may include the gateway device 402 operating in the network 400 as the node A and connecting the wireless network 400 to an external network (not shown) or a standalone external device (not shown). Because of the typical traffic requirements, the gateway device may be a powered device, i.e., connected to a substantially unlimited power source via a power cable or wire. On the other hand, the nodes B-D may be battery-operated field devices. Of course, some or all of the nodes B-D may also be connected to power lines or other sources of energy in other embodiments. As also illustrated in FIG. 13, each of the field devices 404-408 may have a particular update rate at which the device sends data to another node such as the gateway device 402. For example, the field devices 404 and 406 may originate outbound data every second and the field device 408 may originate outbound data every four seconds.

In general, the nodes B-D may also correspond to any type of devices used for communication or equipped with communication capabilities. It is contemplated, for example, that the nodes B-D may be personal computers operating in a home network. However, because the routing and scheduling techniques therein are particularly useful in process control and in sensor mesh networks, the nodes B-D in this particular example are field devices operating in a process control environment and performing various sensing and measurement functions (e.g., temperature, pressure, flow, acidity, valve actuator position, etc.) or control functions (actuation, positioning, etc.). The field devices 404-408 may exchange the measurement and control data over any protocol supporting the routing and scheduling techniques discussed above in reference to FIGS. 3-12. In one embodiment, the protocol supporting these functions may be the WirelessHART protocol 70 illustrated in FIG. 12.

A network manager 440 may be a software module operating within the node A. Similar to the network manager 27 illustrated in FIG. 1 or the network manager 302A-302B illustrated in FIGS. 10-12, the network manger 440 may be responsible for making routing and scheduling decisions in the network 440. In particular, the network manager 440 may initially collect device and signaling information from each of the nodes B-D, as well as the information related to its own host node A, and define an initial network graph 421 for the network 400 in view of these factors. More specifically, each of the nodes A-D may report, for each of its potential neighbors, a received signal strength indication (RSSI) value specifying the energy of a signal. Alternatively, the nodes A-D may report the measured quality of the signal or any other measurement which may be used to assess relative signal quality. Additionally, each of the nodes A-D may report to the network 400 such parameters as power capability (e.g., battery-powered, power line-powered, etc.), projected or actual transmission requirements (e.g., measurement update every second, every four seconds, etc.; no regular automatic updates scheduled; updates only in response to a request; etc.) and other information regarding the operation of the field device in the network 400.

Upon collecting the device information from the nodes A-D, the network manager 440 may establish the topology of the network 400 by selecting direct wireless connections between pairs of neighboring devices to define one or more directed graphs between pairs of nodes A-D. In the particular example illustrated in FIG. 13, each of the directed graphs has node A as either a head or a tail. The network manager 440 may thereby enable the flow of data in the network upstream or downstream relative to the gateway device 402.

When defining direct wireless connections, the network devices 402-404 or, if desired, the network manager 440 may compare the strength or quality of a wireless signal, transmitted by one of the nodes A-D and measured by another one of the nodes A-D, to a threshold value to decide this signal is sufficient to support a direct wireless connection. In at least some embodiments, each of the network devices 402-404 autonomously determine with which of the other network devices 402-404 the network device may establish a direct connection to thereby consider the corresponding device its neighbor. To this end, each of the network devices 402-404 may execute the same routine for collecting and analyzing signal measurements. Referring to FIG. 13, for example, the network device 408 may determine that the signal emitted by a transceiver of the gateway device 402 cannot provide a sufficiently strong direct wireless connection between these devices or, in other words, between the nodes A and D. On the other hand, the network device 404 at the node B may measure the same signal from the node A to obtain a better (e.g., greater strength or higher quality) measurement due to the closer proximity of the device 404 to the device 402, fewer obstacles between the devices 402 and 404, or other factors. The network device 404 may accordingly determine that there is a potential direct wireless connection between the nodes A and B. Preferably but not necessarily, the network device 402 makes a similar determination and also decides that there is a potential direct wireless connection between the nodes A and B.

Next, each of the network devices 402-408 may report the collected measurements and the potential direct connections to the network manager 440. Alternatively, the network 400 may form iteratively so that node A initially forms a complete and operational single-node network, the nodes B and C then join the network 400 to form a complete and operational three-node network, and the node D finally joins the network 400 as a neighbor of the nodes B and C to form a complete and operational four-node network. In some cases, and particularly in large networks having numerous network devices, a certain node could establish direct connections with many potential neighbors. According to some embodiments, the network manager 400 may direct the network devices to disable some of these superfluous direct connections based on the relative quality of these direct connections, for example.

In the formed network 400 illustrated in FIG. 13, each of the nodes B and C may have a direct wireless connection to the node A (supporting the directed connections 410 and 412, respectively). In other words, each of the field devices 404 and 406 may be one hop away from the gateway device 402. Moreover, the nodes B and C may also share a direct wireless connection to support a directed B-to-C connection 414 and a directed C-to-B connection 416. Meanwhile, the node D may be connected to the node A only through the node B, node C, or both. The node D may thus be two (D to B to A) or three (D to B to C to A) hops away from the node A.

To arrive at the definition of directed connections 410-420 and further of the directed routing graphs of the network 400, the network manager 440 may consider several factors in addition to the set of available direct wireless connections between the neighboring network devices. In particular, the network manager 440 may attempt to define directed graphs from the nodes B-D and terminating in the node A by using the least possible number of hops. As a result, the directed graph from the node B to the node A includes a single directed connection 410. Similarly, the directed graph from the node C to the node A includes a single directed connection 412. A directed graph from the node D to the node A requires at least a single intermediate node B or C but may also include two intermediate nodes. Applying the hop-number minimization principle stated above, the network manager 440 may define a directed graph connecting D to A including a sequence of the connections 418 and 410, for example. It will be noted that a directed graph connecting A to D need not include the same intermediate hop B (as discussed in greater detail above in reference to FIGS. 10-12).

Further, the network manager 440 may prefer powered devices as intermediate nodes. If, for example, the node B were battery-powered while the node C had an unlimited source of power, the network manager 440 would define the graph from the node D to the node A via the node C if other factors were the same. However, in the example illustrated in FIG. 13, only the node A is coupled to an unlimited source of power.

In some embodiments, the network manager 440 may also compare the relative signal strengths of the available direct connections, as reported by the network devices, to choose between similar paths. Assuming for example that the node D reported a stronger signal from the node B than from the node C, the network manager 440 may prefer a directed path from the node D to the node A which includes the node B as an intermediate hop.

With continued reference to FIG. 13, the network 400 may thus include the following directed graph definitions for upstream (i.e., device-to-gateway) data propagation:

Identifier Source Destination Definition G1 B A B -> A G2 C A C -> A G3 D A D -> B -> A G4 D A D -> C -> A

In accordance with the principles discussed above, each of the device-specific connection tables 422-428 may store a corresponding part of these graph definitions.

The network manager 440 may apply further principles to define efficient directed routing graphs for the network 400. In some embodiments, the network manager 440 may limit the number of hops in a graph to three or less. Of course, other limits are also possible. In some embodiments, the network manager 440 may also enforce a limit on the number of neighbors of a network device. If, for example, a network device could detect signals transmitted by ten neighboring devices, the network manager 440 may prune this number to allow the network device to establish direct connections with only three or four network devices, selected on the basis of the signal strength or some other principle.

Upon defining the directed graphs between some of the nodes A-D, the network manager 440 may proceed to define a network schedule based on the graph decisions made above. In this example, the network manager 440 may enforce the following additional limitations and/or principles: 1) limiting the number of concurrent active communication channels to 16; 2) not allowing any of the network devices 402-408 to listen twice within the same timeslot; 3) allowing more than one device to transmit to the same target device (e.g., allowing the broadcast links and dedicated links to coexist in the same network schedule; 4) scheduling early hops prior to later hops on multi-hop paths; and 5) supporting update rates that conform to the formula update rate=2x, where x is a positive or a negative integer values (i.e., making the update rate selections of 1 sec., 2 sec., 4 sec., 8 sec., 16 sec., 32 sec., etc. possible). It will be also noted that in the example discussed in reference to FIGS. 13-20, the network manager 440 is aware of the “raw” network topology (i.e., of all the measurements of signal from the potential neighbors reported from every network device), that each of the network devices 402-408 stores and maintains a connection table 422-428, and that the network manager 440 knows the data update rate of each of the network devices 402-408. As an additional requirement, the network 400 may implement a redundancy scheme whereby each transmission is configured with one retry on one path and one retry on another path. Of course, other update rates, routing restrictions, redundancy schemes, etc. are also possible, and it will be appreciated that the principles listed above are provided by way of example only.

In general, the network manager 440 may use the following strategy when defining superframes: 1) set the length of a superframe according to the data update rate of a device with which the superframe is associated; 2) allocate timeslots starting with the fastest update rate; 3) staring from a non-remote device most distant from the gateway 402, allocate one link for each intermediate network device in the path to the gateway 402; 4) schedule each transmission with a retry on the primary link and, whenever possible, schedule another retry on the redundant link; 5) schedule each network device to receive at most once in a timeslot; and 6) start with faster update rates when assigning timeslots.

In addition to device-specific superframes, the network manager 440 may define a network management superframe to propagate network management through the network 400 and receive service requests from the network devices 402-408. In particular, the network manager 440 may set the network management superframe length to be equal to the longest device superframe. In some embodiments, it may be desirable to further enforce a low management superframe length limit such as 600 timeslots, for example (i.e., one minute in a ten-millisecond timeslot embodiment). Further, the network manager 440 may reserve a set of timeslots for advertisement an that potential network devices wishing to join the network 400 may transmit join request messages during these advertisement timeslots. Next, the network manager 440 may traverse the network graph 421 according to the breath-first search starting from the gateway device 402 and number each encountered device N0, N1, . . . , Nn.

As illustrated in FIG. 14, the network manager 440 may define a 100-timeslot superframe 450 for the one-second update of the nodes B and C. It will be noted that, in this example embodiment, the network manager 440 creates a single superframe for two nodes having the same update rate. However, in other embodiments, each node may be associated with a separate superframe. To improve the reliability of transmissions, the superframe 450 may operate over several channels (as discussed with respect to FIG. 5, for example), illustrated as offsets 0-3 in column 452. Meanwhile, each of the columns 454 specifies a particular timeslot within the 100-slot superframe 450. Each cell in the table illustrating the superframe 450 accordingly specifies one or more talkers and one or more listeners for each offset/timeslot tuple.

With continued reference to FIG. 14, the network manager 440 may begin to assign timeslots for the directed graph terminating at the node A by starting with the node B because the node B has the fastest update rate of one second. (Of course, the node C in this example has the same update rate and is also located one hop away from the node A. In this and similar cases, the network manager 440 may apply any desired tie-braking technique to choose between the two or more equal candidates). Because the node B has a direct connection 410 to the node A as part of the directed graph G1, the network manager 440 may allocate the timeslots T0 and T1 on the O-offset channel to the node B for transmission of the periodic update data to the node A. In particular, the node B may use the timeslot T0 for the scheduled transmission of data and the timeslot T1 for a retry. FIG. 14 illustrates these channel-timeslot-directed connection assignments as dedicated links 460 and 462. In view of the discussion provided above, it will also be appreciated that the node B will view the dedicated links 460 and 462 as dedicated transmit links while the node A will view these links as dedicated receive links.

It will be appreciated that although the timeslots T0 and T1 are associated with the same channel in this particular example, the retry may also be allocated to a different channel (e.g., offset 1) to diversify channel usage. Alternatively, the network manager 440 may instruct each of the nodes A-B to perform channel hopping according to a predefined order and with respect to a network schedule that does not directly specify frequency hopping. For example, each of the nodes A-B may always shift the channel offset by one with each superframe cycle, thus ensuring that the relative order specified by the network schedule stays the same.

Next, the network manager 440 may allocate the timeslots TS3 and TS4 on the 1-offset channel to the directed connection 412 from the node C to the node A. Similar to the links 460 and 462, the dedicated links 464 and 466 may correspond to the same scheduled update in a redundant manner. Also, as part of diversifying channel usage, the links 464 and 466 in this example are allocated on a different channel than the links 460 and 462.

To continue with this example, the network manager 440 may then define a 400-timeslot superframe 470 for the four-second update of the node D. Because the superframes 450 and 470 may start at the same time, the timeslots TS0 and TS1 on the O-offset channel may be considered occupied and therefore unavailable. Also, because the node D may transfer its update data to the node A via the node B along the graph G3, the timeslots T0 and T1 may not be allocated to the directed connection 418 at all, irrespective of the channel offset. This limitation can be easily understood by observing that during the timeslots T0 and T1, the node B already engages in communications, i.e., transmitting data in the links 460 and 462. For this reason, the earliest timeslots that the network manager 440 may allocate to the directed connection 418 are TS2 and TS3. Of course, the 1-offset channel is already occupied in the timeslots TS2 and TS3 by the communicating nodes A and C, and the directed connection 418 may therefore reserve the still-available channel with the offset 0 (links 472 and 474 in FIG. 14).

As the next step, the network manager 440 may schedule the directed connection 420 associated with the secondary path from the node D to the node A (illustrated above as graph G4). The dedicated link 476 may accordingly receive the assignment of the timeslot TS4 on the O-offset channel. In the same timeslot TS4, the link 478 associated with the directed connection 410 may reserve the adjacent channel (offset 1). It will be noted that the node B may use the link 478 to forward the data to the node A which the node B received in the timeslot TS2 or TS3 from the node D. In other words, the link 478 may be associated with the graph G3. Applying similar principles, the network manager 440 may then define the links 480-484 also illustrated in FIG. 14.

Now referring to FIG. 15, the network manager 440 may also define a 400-timeslot management superframe 500 in accordance with the length of the longest data superframe 470. Unlike the superframes 450 and 470, the management superframe 500 may include several shared links 502-508 reserved for advertisement, several dedicated links 510-516 reserved for propagating join requests to the network manager 440 residing at the node A, and several dedicated links 520-526 reserved for propagating join responses from the network manager 440 residing at the node A to the end nodes B-D of the network 400.

Specifically with respect to the shared links 502-508, it will be noted that each of the nodes A-D expects to receive data from any potential devices via these shared links except from one of the existing nodes of the network 400. Thus, in at least some of the embodiments, the links 502-508 provide an opportunity for the external devices to submit requests to join the network 400 and to eventually become new network nodes. Meanwhile, the links 510-516 service the respective paths for delivering these requests from the end nodes B-D to the node A.

FIG. 16 illustrates other links which the network manager 440 may reserve in the management superframe 500. More specifically, the links 530-536 may transfer command data from the nodes B-D to the network manager 440 at the node A. The links 540-546 may accordingly support the transfer of command responses in the opposite (i.e., downstream) direction with respect to the node A.

Additionally, the network manager 440 may assign unique superframe identifiers to each of the superframes 450, 470, and 500. To simplify maintenance work, debugging, and visual presentation, the network manager 440 may assign the identifier ‘0’ to the management superframe, the identifier ‘1’ to the one-second superframe, and the identifier ‘4’ to the four-second superframe. In this manner, the frame identifier may conveniently convey at least some information about the frame to the engineer or operator.

Generally with respect to the superframes 450, 470, and 500, it will be appreciated that the particular manner in which the network manager 440 assigns timeslots within these superframes prevents overlap conflicts in the frame cycles following the first cycle. In particular, the superframe 450 goes through four frame cycles during a single cycle of the superframe 470 or the superframe 500. Therefore, if all superframes initially start at the same time, the timeslot TS0 of the superframe 450 will coincide with the timeslots TS0, TS100, TS200, and TS300 of the superframes 450 and 500. The method of allocating timeslots described above may ensure a conflict-free superposition of multiple superframes such as 400, 450, and 500 by making the participating network devices conform to the 2x update rate which provides for optimal divisibility of the superframe lengths.

Thus, in the manner described above, the network manager 440 may generate several directed graphs G1-G4 to efficiently route data in the network 400, define several data superframes 450 and 470 for data updates, define at least one management superframe 500 for propagating advertisement information, join requests and responses, commands and command responses, and other types of network management data, and subsequently schedule dedicated, shared, and possibly other types of links based on the directed graphs G1-G4 and on the schedules superframes 450, 470, and 500. It will be appreciated that the allocation of timeslots illustrated above minimizes the amount of time a transient data packet spends between the arrival at an intermediate node and departure for the next node in the communication path. For example, a packet leaving the node D in the direction of the node A via the node B in the timeslot TS2 may travel from the node B to the node A as early as in the timeslot TS4.

To enable the operation of every node A-D in accordance with these definitions, the network manager 440 may then distribute the graph and scheduling definitions to each node A-D. FIGS. 17-20 illustrate several example device-specific schedules 432-438 (see FIG. 13) consistent with the definitions illustrated in FIGS. 14-16.

Referring to FIG. 17, the node A device-specific schedule 432 may specify, for each entry, a frame identifier 602, a time slot number 604, a channel offset 606, a peer device identifier 608, link options 610, and a link type 612. In other embodiments, the schedule 432 may include alternative or additional information such as, for example, a MAC address of a neighboring device. In this particular example, each line of the schedule 432 may specifically indicate which action, if any, the device 402 performs in a particular timeslot. To simplify readability, FIG. 17 lists the entries of the schedule 432 in the ascending order of timeslots. However, each of the nodes A-D can store the corresponding schedule in any suitable manner, e.g., an array, a linked list, a binary tree, etc.

Referring by way of example to the entry 620 of the schedule 432, the node A may know that in the timeslot TS4, the node A must switch its transceiver to the receive mode, tune to the channel with the offset 1, and expect to receive data from the node B via a normal, dedicated link. Moreover, the node A may know that this transmission is associated with the superframe 4 (i.e., the 400-slot superframe). To take another example, the entry 626 may correspond to the timeslot 8 and may specify the association with the management superframe. Because the node A is scheduled to receive advertisement from candidate network devices, the node A may not know the sender of the data (marked by an asterisk).

It will be noted that the example schedule 432 only illustrates the allocation of timeslots 0-21. However, the superframes 450 may have timeslots TS0, TS1, . . . TS99 and the superframes 470 and 500 may have timeslots TS0, TS1, . . . TS399. Although FIGS. 17-20 do not illustrate each of the unassigned timeslots, it will be appreciated that node A, to take one example, neither receives nor transmits data for the duration of timeslots TS22 through TS99 of the shorter superframe 450.

FIGS. 18-20 illustrate the respective device-specific schedules 434, 436, and 438. As can be seen from these illustrations, each of the schedules 432-438 specifies only the scheduling information relevant to a particular node A-D. Similarly, each of the connection tables 422-428 may store relatively little information which may nevertheless specify all routing actions relevant to the corresponding device. Thus, the node B need not know the routing or scheduling information at the node C, to take one example. This approach may significantly reduce the amount of management traffic transported over the network because the network manager 440 does not need to share the entire formed topology or the complete network schedule with every device. Moreover, the network 400 may become more secure because none of the nodes A-D may know where or when the rest of the network 400 transfers information (with a possible exception of the node A which may, at least in this example, host the network manager 440).

While the description of the scheduling and routing performed in the wireless networks described herein are described in detail, for the sake of simplicity, for a four node wireless network, the principles described herein are applicable to any size of network, having any number of nodes, graphs, etc. Likewise, the graphs may be of any desired or needed size or length.

Although the forgoing text sets forth a detailed description of numerous different embodiments, it should be understood that the scope of the patent is defined by the words of the claims set forth at the end of this patent. The detailed description is to be construed as exemplary only and does not describe every possible embodiment because describing every possible embodiment would be impractical, if not impossible. Numerous alternative embodiments could be implemented, using either current technology or technology developed after the filing date of this patent, which would still fall within the scope of the claims.

Claims

1. A method of reliably transferring data in a communication network having a plurality of network devices and operating in a process control environment, wherein the method is executed in a computer that manages the communication network, the method comprising:

(i) generating a first routing graph having a multiplicity of nodes and a set of edges based on a topology of the communication network, including:
associating each of the multiplicity of nodes with a respective one of the plurality of network devices; and
associating each edge in the set of edges with a respective direct connection between two of the plurality of network devices;
(ii) defining a communication schedule of the communication network based on at least the first routing graph, including:
defining a communication timeslot of a fixed predetermined duration; defining a superframe with a number of communication timeslots, a first portion of the number of communication timeslots of the superframe used for delivery of first process data using the first routing graph, and a second portion of the number of communication timeslots of the superframe used for delivery of second process data using a second routing graph, wherein at least one of the first process data or the second process data is generated based on a physical control function performed by a field device to control a process in the process control environment;
defining a superframe length based on a predetermined periodic update rate of a first one of the plurality of network devices; and
assigning a respective communication timeslot associated with the superframe to the each edge in the set of edges of the first routing graph according to a predetermined order,
wherein at least one of the generating the first routing graph or the defining the communication schedule is based on a set of requirements corresponding to delivery of the first process data from a source network device to a destination network device via at least one intermediate network device, the source network device, the destination network device and the at least one intermediate network device included in the plurality of network devices; and
(iii) communicating, to the first one of the plurality of network devices and to another one of the plurality of network devices, a respective portion of the first routing graph and respective one or more communication timeslots included in the first portion of the number of communication timeslots of the superframe,
wherein the first one of the plurality of network devices transmits, via the communication network and using the respective portion of the first routing graph and the respective one or more communication timeslots, the first process data to the another one of the plurality of network devices and third process data to a gateway in connection with another communication network operating in the process control environment.

2. The method of claim 1, wherein generating the first routing graph includes generating a directed graph defining a unidirectional flow of data from exactly one source node to a destination node; and wherein the first one of the plurality of network devices is one of the exactly one source node or the destination node of the directed graph.

3. The method of claim 2, wherein assigning the respective communication timeslot to the each edge in the set of edges of the first routing graph according to the predetermined order includes allocating the respective communication timeslot to the each edge in the set of edges of the first routing graph in an order of traversal of the first routing graph from the exactly one source node toward the destination node.

4. The method of claim 2, wherein generating the first routing graph further includes selecting a graph node from a set of candidate nodes based on a type of power supply available at each node in the set candidate nodes; and wherein each node in the set candidate nodes can serve as an intermediate node between the exactly one source node and the destination node.

5. The method of claim 2,

wherein the first routing graph is a first directed routing graph, the multiplicity of nodes is a first multiplicity of nodes, the set of edges is a first set of edges, the exactly one source node is a first exactly one source node, and the destination node is a first destination node;
wherein the second routing graph is a second directed routing graph having a second multiplicity of nodes and a second set of edges based on the topology of the communication network, the second directed routing graph having a second exactly one source node and a second destination node and one or more intermediate nodes; and
wherein one of the second exactly one source node or the second destination node of the second directed routing graph is the same as a corresponding one of the first exactly one source node or the first destination node of the first directed routing graph; and wherein defining the communication schedule of the communication network is based on at least the first directed routing graph and the second directed routing graph.

6. The method of claim 1, wherein the communication network is a wireless network; and wherein associating the each edge in the set of edges with the respective connection between the two of the plurality of network devices includes selecting a connection from a set of candidate connections based on a signal quality associated with the connection.

7. The method of claim 1, wherein the superframe is a first superframe, the superframe length is a first superframe length, the number of communication timeslots is a first number of communication timeslots, and wherein defining the communication schedule of the communication network further includes:

defining a second superframe with a second number of communication timeslots defining a second superframe length based on a predetermined periodic update rate at which a second one of the plurality of network devices generates fourth process data to be transmitted to the another one of the plurality of network devices or receives the fourth process data from the another one of the plurality of network devices; wherein the second superframe length is not equal to the first superframe length.

8. The method of claim 1, wherein the superframe is a first superframe, the superframe length is a first superframe length, the number of communication timeslots is a first number of communication timeslots, and wherein the method further comprises defining a second superframe with a second number of communication timeslots defining a second superframe length; wherein defining the communication schedule of the communication network further includes associating the second superframe with network management data.

9. The method of claim 1, wherein defining the communication schedule of the communication network includes defining a plurality of superframes, each of the plurality of superframes including a repeating sequence of cycles of a number of consecutively scheduled communication timeslots, wherein the number can be expressed as 2x such that x is an integer number.

10. The method of claim 1, wherein defining the communication schedule of the communication network further includes assigning a primary communication timeslot and a secondary communication timeslot to the each edge; wherein a transmitting network device associated with a particular edge transmits data to a receiving network device associated with the particular edge during the primary communication timeslot to generate a transmission result; and conditionally transmits data to the receiving network device during the secondary communication timeslot based on the transmission result.

11. The method of claim 1, further comprising:

allocating one of a plurality of channels to each edge to which a communication timeslot is allocated.

12. The method of claim 11, wherein each of the plurality of channels is a radio band associated with a respective central frequency.

13. The method of claim 1, further comprising:

distributing device specific schedule information to at least some of the plurality of network devices, including: specifying, for each direct connection to a different network device of the plurality of network devices, a timeslot, one of a plurality of communication channels, and one of a transmit mode or a receive mode.

14. A method of increasing reliability of a wireless mesh network in a process control system including a plurality of nodes, wherein the method is executed in a computer that manages the wireless mesh network, the method comprising:

(i) establishing a plurality of direct connections, wherein each of the plurality of direct connections is a unidirectional wireless connection having a transmitting node and a receiving node;
(ii) generating a plurality of directed graphs defining communication paths between pairs of nodes based on a topology of the wireless network, wherein:
each directed graph includes at least two of the plurality direct connections,
each directed graph has exactly one source node, a destination node, and one or more intermediate nodes,
the source node of a particular directed graph is associated with only one or more outbound communication connections on the particular directed graph, and
the destination node of the particular directed graph is associated with only one or more inbound communication connections on the particular directed graph;
(iii) defining a multiplicity of concurrent superframes as repeating cycles of consecutively scheduled communication timeslots of a single predefined duration, wherein:
a first portion of the communication timeslots of a first superframe included in the multiplicity of concurrent superframes is used for delivery of a first set of data packets over a first directed graph included in the plurality of directed graphs,
a second portion of the communication timeslots of the first superframe included in the multiplicity of concurrent superframes is used for delivery of a second set of data packets over a second directed graph included in the plurality of directed graphs,
the first set of data packets and the second set of data packets include process control data generated by or provided to control a set of physical control functions performed by a set of field devices to control a process in the process control system,
a number of timeslots in each of the multiplicity of concurrent superframes defines a length of the superframe, and
defining the multiplicity of concurrent superframes includes assigning at least one of the consecutively scheduled communications timeslots based on a predetermined periodic update rate of at least one of the plurality nodes;
(iv) defining a plurality of primary links to generate a communication schedule of the wireless mesh network including, for each directed graph:
associating each primary link with one of the plurality of direct connections of the each directed graph, wherein the one of the plurality of direct connections is associated with a primary fixed communication path from the source node to the destination node, wherein at least one intermediate node of the each directed graph, in response to a failure to deliver a data packet along the primary link, transmits the data packet along a secondary link associated with another one of the plurality of direct connections, and
allocating a different individual timeslot associated with one of the plurality of superframes to each primary link of the each directed graph so that the allocated individual timeslots are scheduled in the order in which packets travel through the each directed graph,
wherein at least one of the generating the plurality of directed graphs or the defining the plurality of primary links to generate the communication schedule is based on a set of requirements corresponding to delivery of process control data from a particular source node to a particular destination node; and
(v) communicating, to each source node, each intermediate node, and each destination node, one or more portions of one or more respective directed graphs and one or more respective different individual timeslots corresponding to one or more respective primary links,
wherein a first node of the plurality of nodes included in the wireless mesh network transmits, in accordance with a respective one or more portions of the one or more respective directed graphs and the one or more respective different individual timeslots, respective process control data to a second node of the plurality of nodes and other process control data to a gateway in connection with another communication network operating in the process control system.

15. The method of claim 14, wherein generating the plurality of directed graphs includes:

designating a first one of the plurality of nodes as a head node of the wireless mesh network;
generating at least one inbound graph defining a flow of data in an inbound direction of the head node along a first subset of direct connections; and
generating at least one outbound graph defining a flow of data in an outbound direction away from the head node along a second subset of direct connections.

16. The method of claim 14, wherein:

the predetermined periodic update rate is a scan rate;
a first at least a portion of the plurality of nodes of the wireless mesh network are field devices each performing a respective control function in the process control system;
a second at least a portion of the plurality of nodes of the wireless mesh network are field devices each performing a respective measurement function in the process control system; and
nodes included in the first at least the portion of the plurality of nodes and nodes included in the second at least the portion of the plurality of nodes each publish report data at respective scan rates; and
wherein defining the multiplicity of concurrent superframes includes: defining a first superframe according to a scan rate of a first field device having a first number of timeslots corresponding to a first length; and defining a second superframe according to a scan rate of a second field device having a second number of timeslots corresponding to a second length; wherein the first length is smaller than the second length.

17. The method of claim 16, including allocating timeslots within the first superframe prior to allocating time slots within the second superframe.

18. The method of claim 14, further comprising:

allocating a plurality of carrier radio frequencies for use by the wireless network, wherein each of the plurality of carrier radio frequencies is identified by a unique channel offset; and
wherein defining the plurality of primary links to generate the communication schedule further includes associating each link with one of the channel offsets.

19. The method of claim 14, further comprising:

defining a plurality of secondary links as part of generating the communication schedule, wherein at least one of the plurality of direct connections included in at least one of the plurality of directed graphs is associated with a primary link and a secondary link, and wherein distinct timeslots in a particular superframe are allocated to the primary link and the secondary link.

20. The method of claim 19, further comprising allocating consecutive timeslots to one of the primary links and to one of the secondary links.

21. The method of claim 19, wherein generating the plurality of directed graphs includes defining duplicate communication paths between pairs of nodes including a primary and secondary communication path, wherein the primary communication path and the secondary communication path differ in at least one direct connection.

22. The method of claim 14, wherein establishing the set of direct connections includes establishing unidirectional wireless connections between two nodes according to a strength of a radio signal detected by a first of the two nodes and transmitted by a second of the two nodes.

23. The method of claim 14, wherein each directed graph defines a fixed primary communication path to be taken by communication packets when traveling along the each directed graph from the source node to the destination node of the each directed graph.

24. The method of claim 14, wherein defining the communication schedule of the wireless mesh network includes assigning communication timeslots to each edge in the set of edges of the first directed graph so that the communication timeslots are associated with the set of edges according to a predetermined order of traversal of the first directed graph from the source node to the destination node.

25. A method of reliably transferring data in a wireless communication network operating in a process control environment and including a plurality of network devices, wherein the method is executed in a computer-based centralized network manager that manages the wireless communication network, the method comprising:

(i) generating a routing scheme of the wireless communication network, including defining a plurality of routing graphs based on a topology of the wireless communication network, each of the plurality of routing graphs having a plurality of nodes and a set of edges, each of the plurality of nodes corresponding to a respective one of the plurality of network devices;
(ii) defining a multiplicity of superframes each having a set of consecutively numbered communication timeslots of a predetermined duration based on predetermined regular communication requirements of at least some of the plurality of network devices, wherein a first portion of the communication timeslots of a first superframe included in the multiplicity of superframes is used for delivery of first process data using a first routing graph included in the plurality of routing graphs, and a second portion of the communication timeslots of the first superframe included in the multiplicity of superframes is used for delivery of second process data using a second routing graph included in the plurality of routing graphs, and wherein at least one of the first process data or the second process data is generated by or provided to control a physical control function performed by a field device to control a process in the process control environment; and
(iii) combining the routing scheme with the definition of the multiplicity of superframes to define a communication schedule of the wireless communication network, including, for at least the first superframe included in the multiplicity of superframes and a respective particular routing graph in the plurality of routing graphs:
assigning communication timeslots within the at least the first superframe to edges of the respective particular routing graph so that the communication timeslots are associated with the set of edges of the respective particular routing graph according to a predetermined order from the source node to the destination node of the respective particular routing graph, and
communicating, to each of the plurality of network devices corresponding to respective nodes included in the respective particular routing graph, one or more portions of the respective particular routing graph and one or more respective communication timeslots,
wherein:
at least a portion of at least one of the defining the plurality of routing graphs or the defining the multiplicity of superframes is based on one or more requirements corresponding to delivery of the first process data from a source node to a destination node by way of at least one intermediate node, the source node, the destination node and the at least one intermediate node included in the plurality of nodes, and
a first network device transmits, via the wireless communication network and according to a respective one or more portions of the respective particular routing graph and the one or more respective communication timeslots, respective process data to a second network device of the plurality of network devices to control the process and other process data to a gateway in connection with another communication network operating in the process control environment.

26. The method of claim 25, wherein defining the plurality of routing graphs includes:

defining a first unidirectional routing graph connecting a first source node to a first destination node; and
defining a second unidirectional routing graph connecting a second source node to a second destination node; wherein the first source node and the second source node correspond to a same one of the plurality of network devices.

27. The method of claim 25, further comprising:

updating the definition of the at least the first superframe while the wireless communication network is active, including adjusting the length of the at least the first superframe; and
assigning communication timeslots within at least the first superframe to edges of the particular routing graph without updating the definition of the particular routing graph.

28. A method of reliably transferring data in a communication network operating in a process control environment, wherein the method is executed in a computer-based centralized network manager that manages the communication network, the method comprising:

(i) generating a first routing graph having a plurality of nodes and a set of edges based on a topology of the communication network, including:
associating each of the plurality of nodes with a respective one of a plurality of network devices participating in the communication network;
associating each edge in the set of edges with a respective connection between two of the plurality of network devices participating in the communication network;
associating a first one of the plurality of nodes with a source of the first routing graph; and
associating a second one of the plurality of nodes with a destination of the first routing graph, wherein the source of the first routing graph and the destination of the first routing graph are separated by at least one intermediate node; and
(ii) generating a communication schedule of the communication network, including:
(a) selecting a superframe of a particular length for transmitting data from the source of the first routing graph, wherein:
the superframe includes a number of consecutively scheduled communication timeslots defining the length of the superframe,
a first portion of timeslots in the superframe are used for delivery of first process control data using the first routing graph,
a second portion of communication timeslots in the superframe are used for delivery of second process control data using a second routing graph,
at least one of the first process control data or the second process control data is generated based on a physical control function performed by a field device to control a process in the process control environment, and
the length of the superframe is selected based on a criterion associated with one of the source or the destination of the first routing graph or one of the source of the destination of the second routing graph; and
(b) assigning at least some of the communication timeslots of the superframe to the set of edges of the first routing graph according to a predetermined order of traversal of the first routing graph from the source to the destination,
wherein at least one of the generating the first routing graph or the generating the communication schedule is based on a set of requirements corresponding to delivery of the first process control data from a source network device via one or more intermediate network devices to a destination network device, the source network device, the destination network device and the one or more intermediate network devices included in the plurality of network devices; and
(iii) communicating, to network devices corresponding to the plurality of nodes of the first routing graph, respective one or more portions of the first routing graph and one or more respective communication timeslots,
wherein a first network device transmits, via the communication network and according to a respective one or more portions of the first routing graph and the one or more respective communication timeslots, respective process control data to a second network device node for performing a control function in the process control environment and other respective process control data to a gateway in connection with another communication network operating in the process control environment.

29. The method of claim 28, wherein the criterion is a regularly scheduled transmission requirement of the one of the source or the destination of the first routing graph.

30. The method of claim 28, wherein selecting the superframe of the particular length includes selecting the superframe in view of the first routing graph so that each edge of the first routing graph can be accommodated within the superframe.

31. A multi-node mesh communication network operating in a process control environment, comprising:

a plurality of field devices to perform respective control functions in the process control environment and to define respective nodes of the multi-node mesh communication network;
a network device operating in the multi-node mesh communication network to define a node of the multi-node mesh communication network;
first computer-executable instructions stored on a computer-readable memory and executable to generate and store a routing scheme of the multi-node mesh communication network and a communication schedule of the multi-node mesh communication network, at least one of the routing scheme or the communication schedule based on a set of requirements corresponding to delivery of process data from a source field device to a destination field device,
the routing scheme including a multiplicity of directed graphs, each of the multiplicity of directed graphs connecting a respective pair of nodes of the multi-node mesh communication network via one or more intermediate nodes, wherein each direct connection between two nodes associated with a particular directed graph define an edge of the graph; and
the communication schedule including a superframe including a number of consecutively scheduled communication timeslots defining a length of the superframe, a first portion of the communication timeslots of the superframe used for delivery of first process data using a first directed graph in the multiplicity of directed graphs, and a second portion of the communication timeslots of the superframe used for delivery of second process data using a second directed graph in the multiplicity of directed graphs, wherein the length of the superframe is based on a predetermined periodic update rate of one of the plurality of field devices, and wherein at least one of the first process data or the second process data is generated based on a physical control function performed by a field device to control a process in the process control environment; and
second computer-executable instructions stored on the computer-readable memory and executable to communicate, to the one of the plurality of field devices, respective portions of a particular directed graph with which the one of the plurality of field devices is associated and respective one or more corresponding communication timeslots of the superframe for the one of the plurality of field devices to use for transmitting respective process data to a gateway in connection with another communication network operating in the process control environment.

32. The multi-node mesh communication network of claim 31, wherein the network device is a gateway device that provides an external interface to the multi-node mesh communication network; and wherein the multiplicity of directed graphs includes:

a first unidirectional graph connecting a first one of the plurality of field devices defining a source of the first unidirectional graph to the gateway device defining a destination of the first unidirectional graph; and
a second unidirectional graph connecting a second one of the plurality of field devices defining a source of the second unidirectional graph to the gateway device defining a destination of the second unidirectional graph.

33. The method of claim 1, wherein:

one of the communication network or the another communication network is a wireless communication network, and the other one of the communication network or the another communication network is a wired communication network.

34. The method of claim 1, wherein a first protocol used in the communication network and a second protocol used in the another communication network are different protocols and are each based on a common, standard process control industry protocol.

35. The method of claim 34, wherein the common, standard process control industry protocol is the Highway Addressable Remote Transmitter (HART) Communication Foundation protocol.

36. The method of claim 14, wherein the process control system includes a plurality of field devices including at least one sensing device, at least one valve device, and at least one switch device; and wherein each field device corresponds to a different node of the plurality of nodes.

37. The method of claim 25, wherein the first network device transmits the respective process data to the second network device using a protocol based on the Highway Addressable Remote Transmitter (HART) Communication Foundation protocol.

38. The method of claim 28, wherein:

the communication network is a first communication network; and
one of the first communication network or the other communication network is a wireless communication network and the other one of the first communication network or the other communication network is a wired communication network.

39. The method of claim 1, wherein the set of requirements includes a latency requirement.

40. The method of claim 1, wherein the set of requirements includes a delay tolerance requirement.

41. The method of claim 1, further comprising dynamically and automatically performing superframe management based on a change in the communication network, performing the superframe management including at least one of:

destroying the superframe;
deactivating the superframe without destroying the superframe;
modifying the superframe;
creating a new superframe;
modifying the communication schedule; or
modifying the first routing graph.

42. The method of claim 1,

wherein the source network device is a first source network device and the destination network device is a first destination network device; and
wherein the set of requirements further corresponds to transmitting the second process data from a second source network device directly to a second destination network device, the second source network device and the second destination network device included in the plurality of network devices.

43. The method of claim 14, wherein the set of requirements includes at least one of a latency requirement, a delay tolerance requirement, a transmission requirement, a signal strength requirement, a signal quality requirement, or a power availability requirement.

44. The method of claim 14, further comprising modifying at least one of the plurality of directed graphs or the plurality of primary links based on a changed network condition.

45. The method of claim 44, wherein the changed network condition includes at least one of a data burst, congestion, a block transfer, or a change in network topology.

46. The method of claim 25, further comprising modifying at least one of the multiplicity of superframes based on a detected change in the wireless communication network, the detected change included in a set of changes to the wireless communication network comprising a data burst, congestion, a block transfer, a particular network device entering or leaving the network, or a change in network topology.

47. The method of claim 25, wherein the one or more requirements include at least one of a latency requirement, a delay tolerance requirement, a transmission requirement, a signal strength requirement, a signal quality requirement, or a power availability requirement.

48. The method of claim 28, further comprising modifying the superframe based on a detected change in the communication network, the detected change included in a set of changes to the communication network comprising a data burst, a detected congestion, a block transfer, a particular network device entering or leaving the network, or a change in a topology of the communication network.

49. The method of claim 28, wherein the set of requirements includes at least one of a latency requirement, a delay tolerance requirement, a transmission requirement, a signal strength requirement, a signal quality requirement, or a power availability requirement.

50. The multi-node mesh communication network of claim 31, wherein the set of requirements includes at least one of a latency requirement, a delay tolerance requirement, a transmission requirement, a signal strength requirement, a signal quality requirement, or a power availability requirement.

51. The multi-node mesh communication network of claim 31, further comprising third computer-executable instructions stored on the computer-readable memory and executable to dynamically and automatically perform superframe management based on a change in the multi-node mesh communication network, wherein superframe management includes at least one of:

destroying the superframe;
deactivating the superframe without destroying the superframe;
modifying the superframe;
creating a new superframe; or
modifying at least one of the multiplicity of directed graphs.
Referenced Cited
U.S. Patent Documents
4792944 December 20, 1988 Takahashi et al.
5159592 October 27, 1992 Perkins
5719859 February 17, 1998 Kobayashi et al.
5926531 July 20, 1999 Petite
6028522 February 22, 2000 Petite
6198751 March 6, 2001 Dorsey et al.
6218953 April 17, 2001 Petite
6233327 May 15, 2001 Petite
6236334 May 22, 2001 Tapperson et al.
6298377 October 2, 2001 Hartikainen et al.
6360277 March 19, 2002 Ruckley et al.
6430268 August 6, 2002 Petite
6437692 August 20, 2002 Petite et al.
6522974 February 18, 2003 Sitton
6578047 June 10, 2003 Deguchi
6594530 July 15, 2003 Brett et al.
6618578 September 9, 2003 Petite
6628764 September 30, 2003 Petite
6678474 January 13, 2004 Masuda et al.
6747557 June 8, 2004 Petite et al.
6801777 October 5, 2004 Rusch
6836737 December 28, 2004 Petite et al.
6891838 May 10, 2005 Petite et al.
6914533 July 5, 2005 Petite
6914893 July 5, 2005 Petite
6920171 July 19, 2005 Souissi et al.
6959356 October 25, 2005 Packwood et al.
6970909 November 29, 2005 Schulzrinne
6996100 February 7, 2006 Haartsen
7002958 February 21, 2006 Basturk et al.
7053767 May 30, 2006 Petite et al.
7079810 July 18, 2006 Petite et al.
7103511 September 5, 2006 Petite
7110380 September 19, 2006 Shvodian
7137550 November 21, 2006 Petite
7209840 April 24, 2007 Petite et al.
7263073 August 28, 2007 Petite et al.
7292246 November 6, 2007 Goldschmidt
7295128 November 13, 2007 Petite
7346463 March 18, 2008 Petite et al.
7375594 May 20, 2008 Lemkin et al.
7397907 July 8, 2008 Petite
7420980 September 2, 2008 Pister et al.
7468661 December 23, 2008 Petite et al.
7529217 May 5, 2009 Pister et al.
7602741 October 13, 2009 Tapperson et al.
7650425 January 19, 2010 Davis et al.
7675935 March 9, 2010 Samudrala et al.
7680033 March 16, 2010 Khan et al.
7697492 April 13, 2010 Petite
7848827 December 7, 2010 Chen
7853221 December 14, 2010 Rodriguez et al.
7961664 June 14, 2011 Pister et al.
7978059 July 12, 2011 Petite et al.
8013732 September 6, 2011 Petite et al.
8064412 November 22, 2011 Petite
20010030957 October 18, 2001 McCann et al.
20010041591 November 15, 2001 Carroll
20020007414 January 17, 2002 Inoue et al.
20020013679 January 31, 2002 Petite
20020031101 March 14, 2002 Petite et al.
20020067693 June 6, 2002 Kodialam et al.
20020111169 August 15, 2002 Vanghi
20020120671 August 29, 2002 Daffner et al.
20030014535 January 16, 2003 Mora
20030026268 February 6, 2003 Navas
20030040897 February 27, 2003 Murphy et al.
20030074489 April 17, 2003 Steger et al.
20030169722 September 11, 2003 Petrus et al.
20030198220 October 23, 2003 Gross et al.
20030236579 December 25, 2003 Hauhia et al.
20040011716 January 22, 2004 Sandt et al.
20040028023 February 12, 2004 Mandhyan et al.
20040053600 March 18, 2004 Chow et al.
20040095951 May 20, 2004 Park
20040117497 June 17, 2004 Park
20040148135 July 29, 2004 Balakrishnan et al.
20040174904 September 9, 2004 Kim et al.
20040183687 September 23, 2004 Petite et al.
20040203973 October 14, 2004 Khan
20040257995 December 23, 2004 Sandy et al.
20040259533 December 23, 2004 Nixon et al.
20050013253 January 20, 2005 Lindskog et al.
20050018643 January 27, 2005 Neilson et al.
20050025129 February 3, 2005 Meier
20050030968 February 10, 2005 Rich et al.
20050049727 March 3, 2005 Tapperson et al.
20050078672 April 14, 2005 Caliskan et al.
20050085928 April 21, 2005 Shani
20050114517 May 26, 2005 Maffeis
20050125085 June 9, 2005 Prasad et al.
20050160138 July 21, 2005 Ishidoshiro
20050164684 July 28, 2005 Chen et al.
20050190712 September 1, 2005 Lee et al.
20050192037 September 1, 2005 Nanda et al.
20050213612 September 29, 2005 Pister et al.
20050228509 October 13, 2005 James
20050239413 October 27, 2005 Wiberg et al.
20050249137 November 10, 2005 Todd et al.
20050276233 December 15, 2005 Shepard et al.
20050282494 December 22, 2005 Kossi et al.
20060002341 January 5, 2006 Bejerano et al.
20060007927 January 12, 2006 Lee et al.
20060029060 February 9, 2006 Pister
20060029061 February 9, 2006 Pister et al.
20060045016 March 2, 2006 Dawdy et al.
20060062192 March 23, 2006 Payne
20060067280 March 30, 2006 Howard et al.
20060077917 April 13, 2006 Brahmajosyula et al.
20060120384 June 8, 2006 Boutboul et al.
20060174017 August 3, 2006 Robertson
20060182076 August 17, 2006 Wang
20060192671 August 31, 2006 Isenmann et al.
20060213612 September 28, 2006 Perron et al.
20060215582 September 28, 2006 Castagnoli et al.
20060245440 November 2, 2006 Mizukoshi
20060253584 November 9, 2006 Dixon et al.
20070016724 January 18, 2007 Gaither et al.
20070070943 March 29, 2007 Livet et al.
20070074489 April 5, 2007 Erhardt et al.
20070076600 April 5, 2007 Ekl et al.
20070078995 April 5, 2007 Benard et al.
20070109592 May 17, 2007 Parvathaneni et al.
20070118604 May 24, 2007 Costa Requena
20070121531 May 31, 2007 Lee et al.
20070140245 June 21, 2007 Anjum et al.
20070143392 June 21, 2007 Choe et al.
20070161371 July 12, 2007 Dobrowski et al.
20070237094 October 11, 2007 Bi et al.
20070243879 October 18, 2007 Park et al.
20070280144 December 6, 2007 Hodson et al.
20070280286 December 6, 2007 Hodson et al.
20070282463 December 6, 2007 Hodson et al.
20070283030 December 6, 2007 Deininger et al.
20080075007 March 27, 2008 Mehta et al.
20080082294 April 3, 2008 Pihlaja et al.
20080082636 April 3, 2008 Hofmann et al.
20080084852 April 10, 2008 Karschnia
20080098226 April 24, 2008 Zokumasui
20080112384 May 15, 2008 Seo et al.
20080117836 May 22, 2008 Savoor et al.
20080120676 May 22, 2008 Morad et al.
20080130570 June 5, 2008 Jung et al.
20080148296 June 19, 2008 Chen et al.
20080192812 August 14, 2008 Naeve et al.
20080198860 August 21, 2008 Jain et al.
20080215773 September 4, 2008 Christison et al.
20080259847 October 23, 2008 Chou et al.
20080285582 November 20, 2008 Pister
20090059855 March 5, 2009 Nanda et al.
20090068947 March 12, 2009 Petite
20090097502 April 16, 2009 Yamamoto
20090154481 June 18, 2009 Han et al.
20100194582 August 5, 2010 Petite
20100312881 December 9, 2010 Davis et al.
20110264324 October 27, 2011 Petite et al.
20110309953 December 22, 2011 Petite et al.
20110320050 December 29, 2011 Petite et al.
Foreign Patent Documents
2324563 September 1999 CA
1170464 January 1998 CN
1804744 July 2006 CN
1169690 January 2002 EP
1236075 September 2002 EP
1293853 March 2003 EP
1370958 December 2003 EP
1376939 January 2004 EP
2388708 November 2011 EP
2 403 043 December 2004 GB
2005143001 June 2005 JP
1020010076781 August 2001 KR
1020040048245 June 2004 KR
1020050028737 March 2005 KR
1020060066580 June 2006 KR
1020050016891 September 2006 KR
102006011131 October 2006 KR
102007002660 March 2007 KR
WO-00/55825 September 2000 WO
WO-01/35190 May 2001 WO
WO-02/05199 January 2002 WO
WO-02/13036 February 2002 WO
WO-02/13412 February 2002 WO
WO-02/13413 February 2002 WO
WO-02/13414 February 2002 WO
WO-02/075565 September 2002 WO
WO-2005079026 August 2005 WO
WO-2005/096722 October 2005 WO
WO-2006121114 November 2006 WO
Other references
  • International Preliminary Report on Patentability from corresponding International Application No. PCT/US08/04777 (Oct. 13, 2009).
  • International Search Report and Written Opinion for corresponding International Application No. PCT/US08/004777 (Jun. 23, 2009).
  • European Search Report for EP 08826233 dated Jul. 23, 2010.
  • Wong, “A Fuzzy-Decision-Based Routing Protocol for Mobile Ad Hoc Networks,” Networks, 2002. ICON 2002. 10th IEEE International Conference on Aug. 27-30, 2002, Piscataway, NJ, USA, IEEE, Aug. 27, 2002, pp. 317-322.
  • Alandjani et al., “Fuzzy Routing in Ad Hoc Networks,” Conference Proceedings fo the 2003 IEEE International Performance, Computing, and Communications Conference. Phoenix, AZ, Apr. 9-11, 2003, IEEE, vol. Conf. 22, Apr. 9, 2003, pp. 525-530.
  • Thomas et al., “Anthoc-QoS: Quality of 1-7 Service Routing in Mobile Ad Hoc Networks Using Swarm Intelligence” Mobile Technology, Applications and Systems, 2005 2nd International Conference on Guangzhou, China Nov. 15-17, 2005, Piscathaway, NJ, USA, IEEE, Piscathaway, NJ, USA Nov. 15, 2005, pp. 1-8.
  • Kastner et al., “Communication Systems for Building Automation and Control,” Proceedings of the IEEE, vol. 93, No. 6, Jun. 1, 2005, pp. 1178-1203.
  • Thomesse, J., “Fieldbus Technology in Industrial Automation,” Proceedings of the IEEE, vol. 93, No. 6, Jun. 1, 2005, pp. 1073-1101.
  • Cleveland, F., “IEC TC57 Security Standards for the Power System's Information Infrastructure—Beyond Simple Encryption,” May 21, 2006, pp. 1079-1087.
  • Matkurbanov et al., “A Survey and Analysis of Wireless Fieldbus for Industrial Environments,” SICE-ICCAS 2006 International Joint Conference, 5555-5561 (2006).
  • Lopez et al., “Wireless communications deployment in industry: a review of issues, options and technologies,” Computers in Industry, Elsevier Science, 56:29-53 (2005).
  • Shen et al., “Wireless Sensor Networks for Industrial Applications,” WCICA, 15-19 (2004).
  • “A Survey and Analysis of Wireless Field bus for Industrial Environments”, Pulat Matkurbanov, SeungKi Lee, Dong-Sung Kim; Dept. of Electron. Eng., Kumoh Nat. Inst. of Technol., Gumi. This paper appears in: SICE-ICASE, 2006. International Joint Conference: Issue Date: Oct. 18-21, 2006, on pp. 55555-5561; Print ISBN: 89-950038-4-7.
  • “D1100 SmartMesh Network Release 1.0 Engineering Software Specifications,” Dust Networks, Dec. 17, 2011, 36 pages.
  • “D1100 SmartMesh Network Release 1.0 Engineering Software Specifications,” Dust Networks, 36 pages.
  • “Multiple Interpenetrating MultiDiGraphs,” Dust Incorporated, 12 pages. (Powerpoint).
  • “SmartMesh-XT CLI Commands Guide,” Dust Networks, Inc., Jun. 27, 2007, 36 pages.
  • SmartMesh-XT KT1030/KT2135/KT2030 Evaluation Kit Guide, Dust Networks, Inc., Nov. 2, 2007, 58 pages.
  • “SmartMesh-XT M2135-2, M2030-2 2.4 GHz Wireless Analog/Digital/Serial Motes,” Dust Networks, Inc., Mar. 28, 2007, 33 pages.
  • SmartMesh-XT Manager XML API Guide, Dust Networks, Inc., Apr. 4, 2007, 148 pages.
  • “System Description for Security Review SmartMesh Alba,” Dust Networks, 36 pages.
  • Berlemann, Software Defined Protocols Based on Generic Protocol Functions for Wired and Wireless Networks, Nov. 2033, RWTH Aachen University.
  • Qu et al., “A web-enabled distributed control application platform for industrial automation”, Emerging Technologies and Factory Automation, Proceedings, 2:129-32 (Sep. 16, 2003).
  • Willig (ed.), “An architecture for wireless extension of PROFIBUS”, IECON 2003—Proceedings of the 29th Annual Conference of the IEEE Industrial Electronics Society, New York, vol. 3, pp. 2369-2375 (Nov. 2-6, 2003).
  • Zheng, “ZigBee wireless sensor network in industrial applications”, SICE-ICCAS 2006 International Joint Conference, IEEE, New Jersey, pp. 1067-1070 (Oct. 1, 2006).
  • Deering et al., “Internet Protocol, Version 6 (IPv6) Specification; rfc1883.txt”, IETF Standard, Internet Engineering Task Force, IETF, CH, Dec. 1, 1995, XP015007667.
  • Almeroth et al., “A lightweight protocol for interconnecting heterogeneous devices in dynamic environments”, Multimedia Computing and Systems, 1999, IEEE International Conference on Florence, Italy Jun. 7-11, 1999, Los Alamitos, CA, USA, IEEE Comput. Soc, US, vol. 2, Jun. 7, 1999, pp. 420-425, XP010519427.
  • Egea-Lopez et al., A wireless sensor networks MAC protocol for real-time applications, Pers Ubiquit Comput, 12:111-112 (2008).
  • Kwon et al., Traffic adaptive IEEE 802.15.4 MAC for Wireless Sensor Networks, Lecture Notes in Computer Science (LNCS), 4096:864-873 (2006).
Patent History
Patent number: 8798084
Type: Grant
Filed: Apr 10, 2008
Date of Patent: Aug 5, 2014
Patent Publication Number: 20080279204
Assignee: Hart Communication Foundation (Austin, TX)
Inventors: Wallace A. Pratt, Jr. (Pflugerville, TX), Mark J. Nixon (Round Rock, TX), Eric D. Rotvold (West St. Paul, MN), Robin S. Pramanik (Karlsruhe), Tomas P. Lennvall (Vasteras), Terrence L. Blevins (Round Rock, TX)
Primary Examiner: Gregory Sefcheck
Assistant Examiner: Steve Young
Application Number: 12/100,927
Classifications
Current U.S. Class: Plurality Of Rings Or Loops To Form A Mesh Network (370/406)
International Classification: H04L 12/56 (20060101);