Adaptively changing weights for fair scheduling in broadcast environments

- Microsoft

Adaptively changing weights for fair scheduling in broadcast environments is disclosed. In one embodiment, a computer-implemented method for allocating bandwidth among a plurality of flows, such as nodes, sharing an output link, such as a network, is disclosed. The method includes adaptively determining a weight for each flow, based on a predetermined criteria, and allocating a portion of bandwidth to each flow proportionally to the weight for the flow. In one embodiment, the predetermined criteria takes into account an input rate of data packets for each flow, while in another embodiment, the predetermined criteria takes into account an queue size for each flow.

Skip to: Description  ·  Claims  · Patent History  ·  Patent History
Description
FIELD OF THE INVENTION

This invention relates generally to broadcast environments such as wired and wireless networks, multi-hop networks, etc., and more particularly to adaptively changing weights for fair scheduling for data transmission within such environments.

BACKGROUND OF THE INVENTION

Broadcast environments include environments in which information is transmitted from discrete originating points over a common medium, and include environments such as networking environments, which have become increasingly common. Networking means that two or more computers or computerized devices, referred to generically as nodes, are communicatively coupled together, so that they can exchange data, typically in the form of packets of data. Networking includes wired local-area-networks (LAN's), in which nodes are connected physically over relatively short distances, wireless LAN's, in which nodes communicate wirelessly over relatively short distances, and multi-hop networks, in which nodes communicate with other nodes on the network, using intermediate nodes to forward their messages.

The amount of data that a network can handle at a given moment in time is referred to as bandwidth. For example, the commonly known Ethernet network generally comes in two different speeds: 100 megabits-per-second (mbps) and 10 megabits-per-second (mbps). This means that, per second, the network is able to accommodate either 100 megabits or 10 megabits of data.

An issue in broadcast environments, such as the ones described above, is determining which node gets to communicate at a given time. Algorithms and schemes to determine which node gets to communicate at a given time typically also concern themselves with fairness. Fairness can be defined in different ways. For example, fairness can mean that each node on the network has a predetermined percentage of the available bandwidth on the network over a given duration of time, a predetermined priority relative to the other nodes on the network, or a weight to divide the available network bandwidth relative to the other nodes. In addition, fairness can mean that a predefined Quality of Service (QOS) is guaranteed for one or more given nodes on the network. A non-restrictive example of QOS is that a given node is guaranteed to receive x amount of bandwidth within y amount of time after the node requests to transmit data over the network.

Within the prior art, fairness has generally been implemented by randomizing the time at which different nodes compete for the broadcast channel to transmit their data The theory is that by randomizing the time, the nodes on the network will pick generally different times to transmit and consequently each will get a chance to transmit its data. While such randomization techniques improve the probability that, over a long period of time, every node gets an opportunity to use the channel, they generally result in poor performance when nodes have to transmit time-constrained data, such as voice and video.

One manner by which fairness can be achieved and which solves this problem is described in the cofiled, copending and coassigned patent application entitled “Fair Scheduling in Broadcast Environments” [attorney docket 1018.051US1]. In this application, each node within a network is able to transmit a packet of data after it has counted down from a back-off interval. The back-off interval for each packet is based on at least a start tag of the packet, which is assigned to the packet when it arrives at or within the node for transmission. The tagging of packets with their start tags and using these tags to schedule transmission provides for fair scheduling, such that each node eventually has its turn to transmit information.

In fair scheduling described in the application identified in the previous paragraph, as well as in the prior art, a fixed weight is usually implicitly or explicitly assigned to each node that does not vary over time. A node's weight is usually based on its priority or bandwidth requirements. The weight for a node thus determines the amount of bandwidth that a node requires and is proportional to the amount of bandwidth it is allocated by the network. For example, where the weights for all the nodes are equal, no node has priority, and each node can be allocated the same amount of bandwidth. As another example, if all the nodes except one have a given weight, and the one node has twice the weight of any other node, then the node with the greater weight will have twice the opportunity to transmit packets of data as any other node, and thus will be able to have allocated to it twice the amount of bandwidth as compared to any other node.

Fixed weights, however, assume that the priority of the nodes and/or their bandwidth requirements are unchanging during a connection's lifetime, and can be determined a priori. This may not be the case. For example, at any given time, one or more of the nodes may be transmitting more time-sensitive data, such as voice data, such that during that transmission such nodes should be accorded higher weight. Another example is when a large quantity of data arrives at a node in a short period of time and this node would want to transmit this data quickly by picking a higher weight dynamically. Fixed-weight schemes do not provide for these situations. For these and other reasons, there is a need for the present invention.

SUMMARY OF THE INVENTION

The invention relates to adaptively changing weights for fair scheduling in broadcast environments. In one embodiment, a computer-implemented method allocates bandwidth among a plurality of flows sharing an output link. Each flow can correspond to a node within a network, for example, or an application program generating data packets within a node. The output link can correspond to the network itself, for example, or the node itself, respectively. The method includes adaptively determining a weight for each flow, based on a predetermined criteria, and allocating a portion of bandwidth to each flow proportionally to the weight for the flow.

For example, in one embodiment, the predetermined criteria takes into account an input rate of data packets for each flow. Thus, flows that are receiving more data packets for transmission over the link over a period of time are accorded higher weights. As another example, in one embodiment, the predetermined criteria takes into account an the queue size for each flow. Thus, flows that have a greater backlog of data packets for transmission over the link are accorded higher weights. In either embodiment, the bandwidth accorded to each flow is proportional to the weight for the flow, and may also be based on a protocol or algorithm for fair scheduling that can utilize the weight, such as protocols and algorithms known within the prior art, and also that which is described in the copending, cofiled and coassigned application referenced in the previous section of the detailed description.

The invention includes computer-implemented methods, machine-readable media, computerized systems, and computers of varying scopes. Other aspects, embodiments and advantages of the invention, beyond those described here, will become apparent by reading the detailed description and with reference to the drawings.

BRIEF DESCRIPTION OF THE DRAWINGS

FIG. 1 is a diagram of an operating environment in conjunction with which embodiments of the invention can be practiced;

FIGS. 2(a)-2(c) are diagrams of example broadcast environments in conjunction with which embodiments of the invention can be practiced;

FIG. 3 is a diagram of an abstraction of a broadcast environment, according to an embodiment of the invention; and,

FIG. 4 is a flowchart of a method according to an embodiment of the invention.

DETAILED DESCRIPTION OF THE INVENTION

In the following detailed description of exemplary embodiments of the invention, reference is made to the accompanying drawings which form a part hereof, and in which is shown by way of illustration specific exemplary embodiments in which the invention may be practiced. These embodiments are described in sufficient detail to enable those skilled in the art to practice the invention, and it is to be understood that other embodiments may be utilized and that logical, mechanical, electrical and other changes may be made without departing from the spirit or scope of the present invention. The following detailed description is, therefore, not to be taken in a limiting sense, and the scope of the present invention is defined only by the appended claims.

Some portions of the detailed descriptions which follow are presented in terms of algorithms and symbolic representations of operations on data bits within a computer memory. These algorithmic descriptions and representations are the means used by those skilled in the data processing arts to most effectively convey the substance of their work to others skilled in the art. An algorithm is here, and generally, conceived to be a self-consistent sequence of steps leading to a desired result. The steps are those requiring physical manipulations of physical quantities. Usually, though not necessarily, these quantities take the form of electrical or magnetic signals capable of being stored, transferred, combined, compared, and otherwise manipulated.

It has proven convenient at times, principally for reasons of common usage, to refer to these signals as bits, values, elements, symbols, characters, terms, numbers, or the like. It should be borne in mind, however, that all of these and similar terms are to be associated with the appropriate physical quantities and are merely convenient labels applied to these quantities. Unless specifically stated otherwise as apparent from the following discussions, it is appreciated that throughout the present invention, discussions utilizing terms such as processing or computing or calculating or determining or displaying or the like, refer to the action and processes of a computer system, or similar electronic computing device, that manipulates and transforms data represented as physical (electronic) quantities within the computer system's registers and memories into other data similarly represented as physical quantities within the computer system memories or registers or other such information storage, transmission or display devices.

Operating Environment

Referring to FIG. 1, a diagram of the hardware and operating environment in conjunction with which embodiments of the invention may be practiced is shown. The description of FIG. 1 is intended to provide a brief, general description of suitable computer hardware and a suitable computing environment in conjunction with which the invention may be implemented. Although not required, the invention is described in the general context of computer-executable instructions, such as program modules, being executed by a computer, such as a personal computer. Generally, program modules include routines, programs, objects, components, data structures, etc., that perform particular tasks or implement particular abstract data types.

Moreover, those skilled in the art will appreciate that the invention may be practiced with other computer system configurations, including hand-held devices, multiprocessor systems, microprocessor-based or programmable consumer electronics, network PC's, minicomputers, mainframe computers, and the like. The invention may also be practiced in distributed computing environments where tasks are performed by remote processing devices that are linked through a communications network. In a distributed computing environment, program modules may be located in both local and remote memory storage devices.

The exemplary hardware and operating environment of FIG. 1 for implementing the invention includes a general purpose computing device in the form of a computer 20, including a processing unit 21, a system memory 22, and a system bus 23 that operatively couples various system components including the system memory to the processing unit 21. There may be only one or there may be more than one processing unit 21, such that the processor of computer 20 comprises a single central-processing unit (CPU), or a plurality of processing units, commonly referred to as a parallel processing environment. The computer 20 may be a conventional computer, a distributed computer, or any other type of computer; the invention is not so limited.

The system bus 23 may be any of several types of bus structures including a memory bus or memory controller, a peripheral bus, and a local bus using any of a variety of bus architectures. The system memory may also be referred to as simply the memory, and includes read only memory (ROM) 24 and random access memory (RAM) 25. A basic input/output system (BIOS) 26, containing the basic routines that help to transfer information between elements within the computer 20, such as during start-up, is stored in ROM 24. The computer 20 further includes a hard disk drive 27 for reading from and writing to a hard disk, not shown, a magnetic disk drive 28 for reading from or writing to a removable magnetic disk 29, and an optical disk drive 30 for reading from or writing to a removable optical disk 31 such as a CD ROM or other optical media.

The hard disk drive 27, magnetic disk drive 28, and optical disk drive 30 are connected to the system bus 23 by a hard disk drive interface 32, a magnetic disk drive interface 33, and an optical disk drive interface 34, respectively. The drives and their associated computer-readable media provide nonvolatile storage of computer-readable instructions, data structures, program modules and other data for the computer 20. It should be appreciated by those skilled in the art that any type of computer-readable media which can store data that is accessible by a computer, such as magnetic cassettes, flash memory cards, digital video disks, Bernoulli cartridges, random access memories (RAMs), read only memories (ROMs), and the like, may be used in the exemplary operating environment.

A number of program modules may be stored on the hard disk, magnetic disk 29, optical disk 31, ROM 24, or RAM 25, including an operating system 35, one or more application programs 36, other program modules 37, and program data 38. A user may enter commands and information into the personal computer 20 through input devices such as a keyboard 40 and pointing device 42. Other input devices (not shown) may include a microphone, joystick, game pad, satellite dish, scanner, or the like. These and other input devices are often connected to the processing unit 21 through a serial port interface 46 that is coupled to the system bus, but may be connected by other interfaces, such as a parallel port, game port, or a universal serial bus (USB). A monitor 47 or other type of display device is also connected to the system bus 23 via an interface, such as a video adapter 48. In addition to the monitor, computers typically include other peripheral output devices (not shown), such as speakers and printers.

The computer 20 may operate in a networked environment using logical connections to one or more remote computers, such as remote computer 49. These logical connections are achieved by a communication device coupled to or a part of the computer 20; the invention is not limited to a particular type of communications device. The remote computer 49 may be another computer, a server, a router, a network PC, a client, a peer device or other common network node, and typically includes many or all of the elements described above relative to the computer 20, although only a memory storage device 50 has been illustrated in FIG. 1. The-logical connections depicted in FIG. 1 include a local-area network (LAN) 51 and a wide-area network (WAN) 52. Such networking environments are commonplace in office networks, enterprise-wide computer networks, intranets and the Internet, which are all types of networks.

When used in a LAN-networking environment, the computer 20 is connected to the local network 51 through a network interface or adapter 53, which is one type of communications device. When used in a WAN-networking environment, the computer 20 typically includes a modem 54, a type of communications device, or any other type of communications device for establishing communications over the wide area network 52, such as the Internet. The modem 54, which may be internal or external, is connected to the system bus 23 via the serial port interface 46. In a networked environment, program modules depicted relative to the personal computer 20, or portions thereof, may be stored in the remote memory storage device. It is appreciated that the network connections shown are exemplary and other means of and communications devices for establishing a communications link between the computers may be used.

Broadcast Environments

In this section of the detailed description, representative broadcast environments, in conjunction with which embodiments of the invention can be practiced, are described. However, those of ordinary skill within the art can appreciate that the invention is not limited to any particular broadcast environment or the set of broadcast environments described herein. Specifically, three example broadcast environments are described: a wired local-area network (LAN), a wireless LAN, and a wireless, multi-hop network.

Referring first to FIG. 2(a), a diagram of a wired LAN is shown. The LAN 208 includes nodes 200, 202, 204 and 206. The nodes are connected to one another in a physical manner. For example, cables can connect the nodes to one another. The LAN 208 has the characteristic that any node is able to communicate with any other node in the network in a direct manner. Example nodes include computers, such as that described in the previous section of the detailed description, as well as computerized devices, such as cell phones, personal digital assistant (PDA) devices, etc.; the invention is not so limited.

Referring next to FIG. 2(b), a diagram of a wireless LAN is shown. Similar to the LAN 208 of FIG. 2(a), the LAN 218 includes nodes 210, 212, 214 and 216. However, in the diagram of FIG. 2(b), the nodes communicate with one another in a wireless manner. For example, each node can include a transceiver such that the node is able to send information over a radio frequency (RF), such as the commonly used 2.4 GHz frequency, although the invention is not so limited. The LAN 218 also has the characteristic that any node is able to communicate with any other node in the network in a direct manner. Example nodes include computers, such as that described in the previous section of the detailed description, as well as computerized devices, such as cell phones, personal digital assistant (PDA) devices, etc.; the invention is not so limited.

Referring next to FIG. 2(c), a diagram of a wireless multi-hop network is shown. The network 230 includes nodes 220, 222, 224, 226 and 228. The nodes communicate with one another in a wireless manner, such as that described in the previous paragraph in conjunction with the description of a wireless LAN. It is noted, however, in the network of FIG. 2(c), that not each node is able to communicate directly with every other node, which is the defining characteristic of a multi-hop network. For example, the node 222 is able to communicate directly with nodes 220 and 224, but not with nodes 226 and 228. Rather, communication between the node 222 and the nodes 226 and 228 must “hop” through node 224. This may be because, for example in the case of wireless communication among the nodes, the node 222 has sufficient communicative range to reach nodes 220 and 224, but not nodes 226 and 228. Example nodes, as before, include computers, such as that described in the previous section of the detailed description, as well as computerized devices, such as cell phones, personal digital assistant (PDA) devices, etc.; the invention is not so limited.

Referring finally to FIG. 3, a diagram of an abstraction of a broadcast environment, such as the broadcast environment of FIG. 2(a), or FIG. 2(b), is shown. In the example, there are flows 304a, 304b, . . . , 304n, which correspond to the nodes of FIG. 2(a), or 2(b). The link 300 corresponds to the network of FIG. 2(a), or 2(b). The abstraction of FIG. 3 is useful because it shows that when a node, that is, a flow, wishes to transmit a packet of data over a network, that is, a link, the link is commonly shared among all the nodes or flows.

Methods

In this section of the detailed description, methods according to varying embodiment of the invention are described. The methods are desirably computer-implemented and realized at least in part as one or more programs running on a computer—that is, as a program executed from a computer-readable medium such as a memory by a processor of a computer. The programs are desirably storable on a machine-readable medium such as a floppy disk or a CD-ROM, for distribution and installation and execution on another computer. The program or programs can be a part of a computer system or a computer, such as that described in conjunction with FIG. 1 in a previous section of the detailed description. The invention is not so limited, however.

Referring now to FIG. 4, a flowchart of a method according to an embodiment of the invention is shown. The method is for allocating bandwidth among a plurality of flows sharing an output link, as these terms have been described in the previous sections of the detailed description and of the application. For example, each flow can correspond to a node, or a number of nodes, where in the latter instance each flow can correspond to a queue over the nodes. The output link can correspond to a network, such as a local-area network (LAN), a multi-hop network, or a wireless network. In 400, a weight for each flow is adaptively determined based on a predetermined criteria. The weights are adaptively determined in that they can vary over time, in a dynamic manner, and are not static or fixed a priori.

In one embodiment, the predetermined criteria is an input rate of data packets at each flow. For example, the input rate may be an input rate at which packets arc arriving at or within a flow for transmission over the output link. Where each flow corresponds to a node, the packets may be generated by different application program at the node. Where a flow corresponds to a number of nodes, the packets may be generated by different nodes.

In this embodiment, adaptively determining a weight for each flow in 400 includes determining the input rate for each flow, and adaptively determining the weight for each flow as the input rate multiplied by a normalizing constant. For instance, in one embodiment, the input rate at a given time τ is r ( τ ) = t r ( t ) + ξ L τ ,
where, ξ is a rate-sensitivity constant, r(t) is the input rate at a previous time t, and L is the packet size. The rate-sensitivity constant determines sensitivity of the rate estimate to short-term changes in the arrival pattern. Given the input rate, the weight is w ( τ ) = r ( τ ) Normalizing factor ,
where Normalizing factor is a constant. For instance, in the case of a 2 Mbps (mega-bits per second) network, the maximum achievable throughput can be about 1.3 Mbps, where the packet size L is 512 bytes. Thus, a normalizing factor of 1.3 Mbps can be used. With such a normalizing factor, the weight approximately represents the arrival rate as a fraction or multiple of maximum achievable throughput.

In another embodiment, the predetermined criteria is the queue size of each flow. For example, the queue size can correspond to the number of packets within a queue at a flow awaiting transmission by the queue for transmission over the output link. In this embodiment, adaptively determining a weight for each flow in 400 includes estimating the number of packets in the queue of each flow, and adaptively determining the weight for each flow, as the number of packets in the queue divided by a maximum number of packets allowed in the flow.

The invention is not limited to a particular approach to determining the current queue size for a flow, and such approaches are known within the art. In one embodiment, a low-pass filter approach is used, as described in the reference S. Floyd, V. Jacobson, “Random Early Detection Gateways for Congestion Avoidance,” IEEE/ACM Transactions on Networking, vol. 1, no. 4, pp. 397-413 (August 1993). In another embodiment, a modified low-pass filter approach is used, as described in the reference, D. Lin, R. Morris, “Dynamics of Random Early Detection,” Proceedings of SIGCOMM '97, pp. 127-137 (September 1997).

Thus, still describing the embodiment of the invention where the queue size for each flow is the predetermined criteria for adaptively changing the weights for the flows, the weight can be determined as w ( τ ) = Avg queue size ( τ ) Max allowed queue size .
Avg queue size(τ) is the number of packet pending in the queue at time τ. Max allowed queue size specifies the maximum allowed size of the queue for the flow.

In 402, a portion of the bandwidth of the output link is allocated to each flow, proportional to the weight for each flow. 402 can be used in conjunction with protocols and algorithms used to implement fair scheduling in broadcast environments, in which such weights are typically static or fixed a priori implicitly or explicitly. For example, the weights determined in 400 can be used to allocate bandwidth to each flow in 402 in conjunction with that described in the copending, cofiled and coassigned application entitled “Fair Scheduling in Broadcast Environments” [attorney docket no. 1018.051US1]. However, the invention is not so limited. In some such protocols and algorithms, where weights are not mentioned at all, a constant weight equal for all flows is implicitly assigned; in other such protocols and algorithms, weights may be mentioned, but may be unequal over the flows.

As can be appreciated by those of ordinary skill within the art, embodiments of the invention can apply not only to methods and computer-implemented methods. For example, a computerized system including an output link with bandwidth and a number of flows is amenable to an embodiment of the invention. In such an instance, each flow shares the bandwidth of the output link as proportional to a corresponding adaptive weight based on a predetermined criteria, as has been described, for example. The output link can be a local-area network (LAN), a wireless network, or a multi-hop network, that communicatively couples the flows, for example. The system may include one or more nodes, where, for example, one flow may correspond or otherwise be located at each node.

Conclusion

Although specific embodiments have been illustrated and described herein, it will be appreciated by those of ordinary skill in the art that any arrangement which is calculated to achieve the same purpose may be substituted for the specific embodiments shown. This application is intended to cover any adaptations or variations of the present invention. Therefore, it is manifestly intended that this invention be limited only by the following claims and equivalents thereof.

Claims

1. A computer-implemented method for allocating bandwidth in a broadcast environment comprising a plurality of end nodes among a plurality of flows associated with respective end nodes and sharing an output link comprising:

adaptively determining at each end node a weight for each flow associated with that end node based on a predetermined criteria; and
allocating at each end node a portion of bandwidth to each flow associated with that end node based on the determined weight for the flow.

2. The method of claim 1, wherein the predetermined criteria comprises an input rate for each flow.

3. The method of claim 2, wherein adaptively determining a weight for each flow comprises:

determining the input rate for the flow; and
adaptively determining the weight for the flow as the input rate multiplied by a normalizing constant.

4. (Cancelled)

5. (Cancelled)

6. The method of claim 1, wherein the predetermined criteria comprises the queue size for each flow.

7. The method of claim 6, wherein adaptively determining a weight for each flow comprises:

determining a number of packets in the flow at a given time;
adapatively determining the weight for the flow as the number of packets in the flow divided by a maximum number of packets allowed in the flow.

8. A computer-implemented method for allocating bandwidth among a plurality of flows sharing an output link comprising:

adaptively determining a weight for each flow based on a predetermined criterion as
w ⁡ ( τ ) = Avg ⁢   ⁢ queue ⁢   ⁢ size ⁡ ( τ ) Max ⁢   ⁢ allowed ⁢   ⁢ queue ⁢   ⁢ size,
where Avg queue size(τ) is the queue size at time τ, and Max allowed queue size specifies a maximum allowed size of a queue; and
allocating a portion of bandwidth to each flow at an end node associated with that flow based on the weight for the flow.

9. The method of claim 1, wherein each flow corresponds to a queue over at least one node.

10. The method of claim 1, wherein the output link comprises a local-area network (LAN).

11. The method of claim 1, wherein the output link comprises a wireless network.

12. A machine-readable medium having instructions stored thereon for execution by a processor to perform a method for allocating bandwidth in a broadcast environment comprising a plurality of end nodes among a plurality of flows associated with respective end nodes and sharing an output link comprising:

adaptively determining at each end node a weight for each flow associated with that end node based on an input rate for each flow; and
allocating at each end node a portion of bandwidth to each flow associated with that end node based on the determined weight for the flow.

13. The medium of claim 12, wherein adaptively determining a weight for each flow comprises:

determining the input rate for the flow; and
adaptively determining the weight for the flow as the input rate multiplied by a normalizingconstant.

14. (Cancelled)

15. (Cancelled)

16. A machine-readable medium having instructions stored thereon for execution by a processor of a particular one of a plurality of end nodes of a broadcast environment to perform a method for allocating bandwidth in the broadcast environment among a plurality of flows sharing an output link comprising:

adaptively determining a weight at the particular end node for each flow associated with that end node based on the queue size for each flow; and
allocating a portion of bandwidth to each flow associated with that end node based on the determined weight for the flow.

17. The medium of claim 16, wherein adaptively determining a weight for each flow comprises:

estimating a number of packets in the flow at a given time;
adaptively determining the weight for the flow as the number of packets in the flow divided by a maximum number of packets allowed in the flow.

18. A machine-readable medium having instructions stored thereon for execution by a processor to perform a method for allocating bandwidth among a plurality of flows sharing an output link comprising:

adaptively determining a weight for each flow based on the queue size for each flow, wherein adaptively determining a weight for each flow comprises determining the weight as
w ⁡ ( τ ) = Avg ⁢   ⁢ queue ⁢   ⁢ size ⁡ ( τ ) Max ⁢   ⁢ allowed ⁢   ⁢ queue ⁢   ⁢ size,
where Avg queue size(τ) is the queue size at time τ, and Max allowed queue size specifies a maximum allowed size of a queue; and
allocating a portion of bandwidth to each flow based on the weight for the flow.

19. A computerized system comprising:

a broadcast environment output link having a bandwidth, the broadcast environmentcomprising a plurality of end nodes; and
a plurality of flows, each flow being associated with one end node, each flow sharing the bandwidth of the output link proportional in a proportion based on a corresponding adaptive weight based on a predetermined criteria applied at the end node associated with the flow.

20. The system of claim 19, wherein the predetermined criteria comprises an input rate for each flow.

21. The system of claim 19, wherein the predetermined criteria comprises the queue size for each flow.

22. The system of claim 19, wherein the output link comprises a wireless network (LAN) communicatively coupling the plurality of flows.

23. The system of claim 19, wherein the output link comprises a wireless network communicatively coupling the plurality of flows.

24. The system of claim 19, further comprising at least one node, such that at least one of the plurality of flows is located at each node.

25. The method of claim 8 wherein the adaptively determining the weight for each flow based on the predetermined criterion comprises adaptively determining the weight for each flow based on the queue size and an input rate.

26. The method of claim 8 wherein the output link comprises a local area network.

27. The method of claim 8 wherein the output link comprises a wireless network.

28. The medium of claim 18 wherein the adaptively determining the weight for each flow based on the predetermined criterion comprises adaptively determining the weight for each flow based on the queue size and an input rate.

29. The medium of claim 18 wherein the output link comprises a local area network.

30. The medium of claim 18 wherein the output link comprises a wireless network.

Patent History
Publication number: 20050030896
Type: Application
Filed: Sep 14, 2004
Publication Date: Feb 10, 2005
Applicant: Microsoft Corporation (Redmond, WA)
Inventors: Paramvir Bahl (Issaquah, WA), Nitin Vaidva (Redmond, WA)
Application Number: 10/940,333
Classifications
Current U.S. Class: 370/231.000; 370/235.000; 370/401.000