Presentation is loading. Please wait.

Presentation is loading. Please wait.

A Scheduling Technique for Bandwidth Allocation in Wireless Personal Communication Networks Nikos Passas, Nikos Loukas, and Lazaros Merakos Future generation.

Similar presentations


Presentation on theme: "A Scheduling Technique for Bandwidth Allocation in Wireless Personal Communication Networks Nikos Passas, Nikos Loukas, and Lazaros Merakos Future generation."— Presentation transcript:

1 A Scheduling Technique for Bandwidth Allocation in Wireless Personal Communication Networks Nikos Passas, Nikos Loukas, and Lazaros Merakos Future generation wireless personal communication networks (PCN) are expected to provide multimedia capable wireless extensions of fixed ATM/B-ISDN.

2 Paper presents · A method for transmission scheduling in PCN (similar to the technique of “virtual leaky bucket “ developed for fixed ATM networks) · Introduces two alternative priority mechanisms for the sharing of the available bandwidth Goals Fair and efficient treatment of various types of traffic on the air interface Supporting two kinds of sources :  Constant - bit - rate (CBR) voice and  Variable - bit - rate (VBR) video [ensuring that bandwidth allocation is consistent with their declarations at connection setup] with different traffic characteristics and service requirements

3 The PCN terminal of the future  will be able to integrate voice, video and data services  will have to coexist with fixed ATM / B-ISDN Design objectives  flexible multiservice capability (voice, data and multimedia)  good QoS for many service types  compatibility with future ATM/B-ISDN networks  low terminal cost/complexity/power consumption and  efficient, scalable and moderate cost network architecture An important system design issue in PCN is the selection of a suitable bandwidth sharing technique

4 Advanced techniques must be employed including  Call Admission Control (CAC) (PCN user requirements and available resources are of the same nature as in fixed ATM)  Bandwidth enforcement and sharing mechanisms (must be incorporated in the medium access control (MAC) protocol of the wireless environment) Limitation of the radio medium make efficiency and fairness of such techniques more critical than in fixed ATM To avoid inconsistencies and provide a common platform to users, regardless of their connection point, it is essential to consider compatibility with future fixed ATM networks

5 Cellular environment :  each cell consists of one BS and a number of MSs  the number of MSs in cell changes dynamically as they move from one cell to another Sources :  voice sources transmitting at constant rate when they are active, and  video sources transmitting at variable rate MSs can be thought of as advanced mobile telephones (e.g., videophones) equipped with micro-cameras and mini displays, capable for voice-video applications

6 Operate onUplink channel  from MSs to the BS of their cell different bands { Downlink channel  from BS to MSs Uplink channel multiple access control protocol used in conjunction with the scheduling technique Downlink channel is not a multiple access channel The access control protocol used for controlling transmissions on the uplink channel not only has to enable MSs to share it efficiently with high statistical multiplexing gain, but also to provide MSs with QoS guarantees similar to those in fixed ATM networks.

7 QoS guarantees are accomplished through the combined use of  an appropriate connection admission control (CAC) scheme, which ensures that no new MS connections are admitted if by doing so the QoS of already existing connections cannot be guaranteed, and  A “scheduler”, located at the BS, which is responsible for allocating the uplink channel to the MSs in accordance with the QoS agreed upon admission.

8 Framework of the uplink access control protocol within which the scheduler will operate Uplink channel is organized as a TDMA-based system Each cell is a “hub-based” system since all communications between MSs are done through their BS (the hub) Channel time is subdivided into fixed length TDMA slot frames. Slots in each frame are dynamically allocated by the BS to the MSs on the basis of transmission requests received from active MSs during the previous frame, and the QoS agreed upon at connection setup.  Each TDMA frame is subdivided into N r request slots and N d data slots (N r, N d assumed constant - in general they may vary depending on the number and the kind of active sources)  The length of the data slots is selected to be equal to an ATM “cell” (48 bytes data, 5 bytes header), plus an additional radio - specific header, which depends on the specific physical and MAC layer protocols used on the radio interface

9 Request slots in one uplink frame  are used by the MS sources to inform the BS about the data slots they need in the next uplink frame  are expected to be short, compared to data slots, since the only information they have to include is the source’s ID and the number of the requestsed data slots  N r request slots per frame are shared by the active MSs, in accordance with a random access protocol (e.g., the slotted ALOHA protocol, or the stack protocol)  N r can be chosen large enough so that the probability of an allocation request being transmitted successfully on its first attempt is close to unity, without substantial overhead

10 Requests and allocation of data slots :  A source transmits its allocation request for the next frame to the BS in a request slot, and waits for an acknowledgement on the downlink before the beginning of the next frame  If a collision occurs, the source will not receive the acknowledgement and, if its request does not correspond to a packet that has already expired, it will attempt to retransmit it in the next frame  After receiving all the request slots of a frame, the BS must decide on how to allocate the N d data slots of the next uplink frame  Before the beginning of the next frame, BS sends an allocation acknowledgement to all sources, notifying them about the slots that they have been assigned

11 Voice data traffic is considered CBR and is given priority over VBR video traffic  Requests from voice sources are satisfied first without competition from video traffic requests  For requests that come from video sources a mechanism similar to the leaky bucket and the virtual leaky bucket is used.  In order to enter a fixed ATM network, an ATM cell must first obtain a token from a token pool.  A token pool for each video source is located at the BS.  If there are no available tokens, in the leaky bucket, the cell must wait until a new token is generated.  Tokens are generated at a fixed rate equal to the mean cell rate of the source  The size of the token pool depends on the burstiness of the source  The state of each pool gives an indication about how much of the declared bandwidth, the corresponding source has consumed at any instance of time

12 Difference between the leaky bucket and the virtual leaky bucket In the virtual leaky bucket when the pool is empty, an arriving cell, rather than waiting (as in the leaky bucket) is permitted to enter the network with violation tag in its header. Violating cells are the first to be discarded if they later arrive at a congested network node. A major difference between the leaky bucket and the mechanism of this paper when each source has its own connection line with the network,  is that in TDMA system the traffic from all sources is multiplexed in a common radio channel  all requesting packets cannot enter the network, at least not immediately. The acceptance rate of video sources in the channel is limited to the number of slots per frame minus those slots dedicated to the voice sources.  A priority mechanism must be introduced, to decide how the available channel capacity will be allocated to the competing requests from different video sources.  The unaccepted requests will have to wait until their priority becomes higher or until they become expired.

13 The paper introduces two mechanisms, which are based on the state of the token pools and the current requests from all sources. The main objectives of these mechanisms are :  to guarantee fair treatment of all sources under heavy traffic conditions, based on declarations at connection setup, and  to permit sources to transmit over their negotiated throughput, when capacity is available Priority Mechanism A  is based on the philosophy that the source which has more tokens compared to its requests has higher priority, since it is below its declarations, and therefore the system should try to satisfy its requests as soon as possible. S i  source P i  the state that the token pool of S i will be in, if all of its requests are satisfied T i  the number of tokens in the pool of S i at the time a request slot from source S i arrives R i  the number of requests declared in that slot P i = T i - R i

14 Let assume that M sources have requested slots for the next frame, with priorities P 1, P 2, …., P M and let P 1  P 2  ….  P M The mechanism will first try to allocate slots in the next frame for all requests of source S 1, since it has the highest priority. When all requests of source S 1 are satisfied and if there are still available slots in the next frame, source S 2 will be selected, then S 3 and so on, until the requests of all sources are satisfied, or until all the available slots of the next frame have been allocated. In case the priorities of some sources are equal, the source with the most requests is serviced first. Example If for source S k and S l, P k = P l and R k >R l the mechanism will first allocate slots for all requests of source S k and then for all requests of source S l. In the special case where P k = P l and R k = R l (leading to T k =T l ), the mechanism randomly chooses one source to service first.

15 Priority Mechanism A, seems reasonable since it is based on both negotiations made at connection setup, expressed by the token pools, and current needs, expressed by the request slots of each source. A possible weakness is that, when a source becomes active after a long idle period, it will probably take all the slots it requests, since its token pool is almost full, resulting in many temporary denials for other sources. Priority Mechanism B Tries to solve the abovementioned problem of Priority Mechanism A, by gradually allocating slots, based on the state of the token pool of each source. The available slots of the frame are spread to more sources avoiding abrupt denials, which can affect the QoS offered to the end user. Let S 1, S 2, …. S M be the sources requesting slots in one frame and T 1  T 2  ….  T M the corresponding tokens. The mechanism starts by gradually allocating T 1 - T 2 slots to source S 1 (with the assumption that there are that many requests and available slots). If T 1 = T 2 no slots are allocated at this state. Then, it allocates T 2 - T 3 slots to source S 1, and T 2 - T 3 slots to source S 2 (allocating one slot at a time to sources S 1 and S 2 in a round robin fashion), T 3 - T 4 slots to source S 1, S 2 and S 3 and so on, until all requests are satisfied, or until all the available slots of the next frame have been allocated. …continue

16 For every slot allocated to a source, the corresponding token variable is decremented by one  ensures the fair treatment of all sources since, even if a source is assigned many slots in one frame, it will have lower priority in the following ones. Results In both mechanisms no request is blocked if slots are available. Even if a source ’s priority (mechanism A) or token variable (mechanism B) is negative, available slots are allocated to the source, according to mechanisms A and B. The proposed technique is more similar to the virtual leaky bucket method, than of the leaky bucket. Simulation Model Channel speed C = 1,92 Mb/sec Frame length L = 12 msec Data slot size = 53 bytes (48 bytes payload)  to fit an ATM cell A frame can contain

17 The length of request slots was set to 12 bits,  6 bits for the source’s ID, and  6 bits for the number of requests There are two kind of sources :  CBR voice sources, producing 32 Kb/sec (1 slot/frame) on active state  VBR video sources, with mean rate μ =128 Kb/sec, peak rate = 512 Kb/sec, deviation = 64 Kb/sec and autocovariance C( T ) = 2 e -a T (a=3.9 sec -1 ) To model the traffic from video sources independent discrete-time batch Markov arrival process (D-BMAP) was used. Time-of-expiry  for both voice and video packets was chosen to be between 2 and 3 frames In all examples :  the number of voice sources was equal to the number of video sources, since we have to do with MSs as videophones, each having one voice and one video source.

18 P loss is:  long term average fraction of packets lost (due to time-of-expiry violation) from all sources combined The two mechanisms induce the same P loss since the total number of slots allocated per frame is the same (in both), and packets are lost if the corresponding requests are not granted in the next frame. Equivalent bandwidth is :  a unified metric representing the effective bandwidth of a connection based on its parameters declared at connection setup For Example With P loss =10 -3 and the previous mentioned parameters Equivalent bandwidth = 349.44 Kb/sec or 10.92 slots/frame For 5 active sources (CBR)  53-5 = 48 slots 48/10.92 = 4,39 video sources (VBR)

19 The utilization of the available bandwidth was found identical for both mechanisms, because  no request blocking is performed when slots are available How lost packets are spread in time ? The variance of denials of source S i is considered as the variance in time of the number of requests that are denied. D i,k  the number of slots requested by S i to be allocated in frame k but denied by the scheduler due to slot unavailability D i (n)  the mean variance of D i,k V D i (n)  the sample variance of D i,k

20 Priority Mechanism B results in milder variations of denials compared to those of Priority Mechanism A. This is because Mechanism B tries to spread the slots of a frame to more sources than Mechanism A. Smaller denials can more easily absorbed by the end user. Large denials of Mechanism A can result in temporary degradation in quality, which can be rather annoying to the end user A promising idea towards combining the two mechanisms is a method that gradually allocates slots for each source as in mechanism B and uses the priorities in Mechanism A.


Download ppt "A Scheduling Technique for Bandwidth Allocation in Wireless Personal Communication Networks Nikos Passas, Nikos Loukas, and Lazaros Merakos Future generation."

Similar presentations


Ads by Google