Link capacity sharing for throughput-blocking optimality

Multiplex communications – Data flow congestion prevention or control – Control of data admission to the network

Reexamination Certificate

Rate now

  [ 0.00 ] – not rated yet Voters 0   Comments 0

Details

C370S235000, C370S229000

Reexamination Certificate

active

06788646

ABSTRACT:

TECHNICAL FIELD OF THE INVENTION
The present invention generally relates to the field of communication networks, and in particular to link capacity sharing and link bandwidth allocation in such networks.
BACKGROUND
Many communication networks of today support so-called elastic traffic such as the “best effort” services provided in Internet Protocol (IP) based networks or the Available Bit Rate (ABR) traffic in ATM networks. Elastic traffic is typically established for the transfer of a digital object, such as a data file, a Web page or a video clip for local playback, which can be transmitted at any rate up to the limit imposed by the link capacity. Web browsing on the Internet in particular is a good and representative example of elastic traffic. Here, the “elasticity” of the traffic is apparent as the user-perceived throughput (normally given in transmitted bits or bytes per time unit) when downloading for example a web page fluctuates in time depending on the overall system load.
The services delivered by IP based networks and the Internet in particular are called “best effort”, because the networks generally do not provide any guarantee of the quality of service (QoS) received by the applications. The IP network only makes a best effort to provide the requested service. For instance, if an application requests the network to deliver an IP packet from one end-point to another, the network normally can not say what the delay through the network will be for that packet. In fact, the network does not even guarantee tat the packet will be delivered at all.
Therefore, terminals connected to an IP network have to handle packet losses and excessive packet delay situations. Such situations occur when there are too many applications simultaneously using the network resources. These congestion situations have a non-zero probability in IP based networks, because IP networks do not exercise call admission control (CAC). In other words, IP networks do not restrict the number of simultaneously connected users, and consequently if there are too many users utilizing the network resources there will be congestion and packet losses.
However, with the advent of real-time traffic and QoS requirements in EP networks, there is a need for exercising call admission control (CAC) in order to restrict the number of connections simultaneously present in the network.
An important aspect of call or connection admission control is that new calls arriving to the network may be rejected service in order to protect in-progress calls. In general, CAC algorithms such as those commonly in use for rigid traffic in conventional ATM networks provide a basic means to control the number of users in the network, thereby ensuring that admitted users get the bandwidth necessary to provide the QoS contracted for. Consequently, a CAC algorithm represents a trade-off between the blocking probability for new calls and the provided throughput for in-progress calls. In other words, the more users that the CAC algorithm admits into the network (which reduces the blocking probability) the smaller the provided throughput per-user becomes, since a greater number of users will share the total bandwidth, and vice versa.
Recent research has indicated that it is meaningful to exercise call admission control even for elastic traffic, because CAC algorithms provide a means to prevent TCP sessions from excessive throughput degradations.
The issue of applying CAC for elastic connections, and thereby providing a minimum throughput for Transmission Control Protocol (TCP) connections in the Internet has been addressed by Massoulie and Roberts in references [1-3]. Here, bandwidth is allocated to different users according to some fairness criteria.
It has been recognized by Gibbens and Kelly in references [4-5] that there is an intimate relationship between throughput and blocking probabilities for elastic traffic, and that this trade-off is connected to the issue of charging.
It has also been shown by Feng et al. in reference [6] that providing a minimum rate guarantee for elastic services is useful, because in that case the performance of the TCP protocol can be optimized.
As the Internet evolves from a packet network supporting a single best effort service class towards an integrated infrastructure for several service classes, there is also a growing interest in devising bandwidth sharing strategies, which meet the diverse needs of peak-rate guaranteed services and elastic services.
Similarly, modern ATM networks need to support different service classes such as Constant Bit Rate (CBR) and Available Bit Rate (ABR) classes, and it is still an open question how to optimally share the link capacity among the different service classes.
In general, the issue of bandwidth sharing, in the context of dynamically arriving and departing traffic flows and especially when users have different throughput and blocking requirements, is known from the classical multi-rate circuit switched framework to be an extremely complex problem.
SUMMARY OF THE INVENTION
The present invention overcomes these and other drawbacks of the prior art arrangements.
It is a first object of to invention to devise a link capacity/bandwidth sharing strategy that meets the diverse needs of rigid and elastic services in a mixed rigid-elastic traffic environment.
In particular, it is desirable to treat the issues of bandwidth sharing and blocking probabilities for elastic traffic in a common framework. In this respect, it is a second object of the present invention to provide a link capacity sharing mechanism that considers the throughput-to-blocking trade-off for elastic traffic. Specifically, it would be beneficial to develop and utilize a link capacity sharing algorithm that optimizes the throughput-to-blocking trade-off.
It is a further object of the invention to provide an appropriate call-level model of a transmission link carrying elastic traffic and to apply the call-level model for dimensioning the link bandwidth sharing for throughput-blocking optimality.
These and other objects are met by the invention as defined by the accompanying patent claims.
The invention concerns an efficient strategy for sharing link bandwidth in a mixed rigid-elastic traffic environment, as well as a strategy for sharing bandwidth among elastic traffic flows.
Briefly, the idea according to the invention is to share link capacity in a network by dividing the link capacity into a first common part for elastic as well as rigid (non-elastic) traffic and a second part dedicated for elastic traffic based on received network traffic inputs. Subsequently, one or more admission control parameters for the elastic traffic are determined based on the division of link capacity as well as received network traffic inputs.
The division of link capacity generally serves to share the link capacity between rigid and elastic traffic, and in particular to reserve a part of the link capacity to elastic traffic. Preferably, a minimum required capacity of the common part relating to rigid traffic is determined given a maximum allowed blocking probability for the rigid traffic. In this way, a certain grade of service (GoS) on call level is guaranteed for the rigid traffic on the link.
The admission control parameter(s) determined for elastic traffic generally serves to restrict the number of elastic traffic flows simultaneously present on the link. In particular, by formulating a call-level model for elastic traffic and determining a maximum number of admissible elastic traffic flows based on call-level constraints for the elastic traffic related to throughput and/or blocking, the throughput-to-blocking trade-off is fully considered. In this respect, the invention is capable of optimally allocating link bandwidth among elastic connections in the sense that blocking probabilities are minimized under throughput constraints, or the other way around, in the sense that the throughput is maximized under blocking constraints. In his way, the invention provides maximum link bandwidth utilizatio

LandOfFree

Say what you really think

Search LandOfFree.com for the USA inventors and patents. Rate them and share your experience with other people.

Rating

Link capacity sharing for throughput-blocking optimality does not yet have a rating. At this time, there are no reviews or comments for this patent.

If you have personal experience with Link capacity sharing for throughput-blocking optimality, we encourage you to share that experience with our LandOfFree.com community. Your opinion is very important and Link capacity sharing for throughput-blocking optimality will most certainly appreciate the feedback.

Rate now

     

Profile ID: LFUS-PAI-O-3228576

  Search
All data on this website is collected from public sources. Our data reflects the most accurate information available at the time of publication.