Reinforcement learning as a means of dynamic aggregate QoS provisioning
dc.citation.epage | 114 | en_US |
dc.citation.spage | 100 | en_US |
dc.contributor.author | Akar, Nail | en_US |
dc.contributor.author | Şahin, Cem | en_US |
dc.coverage.spatial | Warsaw, Poland | |
dc.date.accessioned | 2019-01-29T08:33:13Z | |
dc.date.available | 2019-01-29T08:33:13Z | |
dc.date.issued | 2003-03 | en_US |
dc.department | Department of Electrical and Electronics Engineering | en_US |
dc.description | Date of Conference: 24-25 March, 2003 | |
dc.description | Conference name: Art-QoS: International Workshop on Architectures for Quality of Service in the Internet -International Workshop, Art-QoS 2003 | |
dc.description.abstract | Dynamic capacity management (or dynamic provisioning) is the process of dynamically changing the capacity allocation (reservation) of a virtual path (or a pseudo-wire) established between two network end points. This process is based on certain criteria including instantaneous traffic load for the pseudo-wire, network utilization, hour of day, or day of week. Frequent adjustment of the capacity yields a scalability issue in the form of a significant amount of message distribution and processing (i.e., signaling) in the network elements involved in the capacity update process. We therefore use the term “signaling rate” for the number of capacity updates per unit time. On the other hand, if the capacity is adjusted once and for the highest loaded traffic conditions, a significant amount of bandwidth may be wasted depending on the actual traffic load. There is then a need for dynamic capacity management that takes into account the tradeoff between signaling scalability and bandwidth efficiency. In this paper, we introduce a Markov decision framework for an optimal capacity management scheme. Moreover, for problems with large sizes and for which the desired signaling rate is imposed as a constraint, we provide suboptimal schemes using reinforcement learning. Our numerical results demonstrate that the reinforcement learning schemes that we propose provide significantly better bandwidth efficiencies than the static allocation policy without violating the signaling rate requirements of the underlying network. | en_US |
dc.description.provenance | Submitted by Ebru Kaya (ebrukaya@bilkent.edu.tr) on 2019-01-29T08:33:13Z No. of bitstreams: 1 Reinforcement Learning as a Means of Dynamic Aggregate QoS Provisioning.pdf: 645367 bytes, checksum: fa42af3567abe0760c99b5543b543b75 (MD5) | en |
dc.description.provenance | Made available in DSpace on 2019-01-29T08:33:13Z (GMT). No. of bitstreams: 1 Reinforcement Learning as a Means of Dynamic Aggregate QoS Provisioning.pdf: 645367 bytes, checksum: fa42af3567abe0760c99b5543b543b75 (MD5) Previous issue date: 2003 | en |
dc.description.sponsorship | This work is supported by The Scientific and Technical Research Council of Turkey (TUBITAK) under grant EEEAG-101E048 | en_US |
dc.identifier.doi | 10.1007/3-540-45020-3_8 | en_US |
dc.identifier.uri | http://hdl.handle.net/11693/48467 | |
dc.language.iso | English | en_US |
dc.publisher | Springer | en_US |
dc.relation.isversionof | https://doi.org/10.1007/3-540-45020-3_8 | en_US |
dc.source.title | Architectures for Quality of Service in the Internet International Workshop, Art-QoS 2003 | en_US |
dc.subject | Reinforcement learn | en_US |
dc.subject | Average cost | en_US |
dc.subject | Voice call | en_US |
dc.subject | Label switch path | en_US |
dc.subject | Decision epoch | en_US |
dc.title | Reinforcement learning as a means of dynamic aggregate QoS provisioning | en_US |
dc.type | Conference Paper | en_US |
Files
Original bundle
1 - 1 of 1
Loading...
- Name:
- Reinforcement Learning as a Means of Dynamic Aggregate QoS Provisioning.pdf
- Size:
- 630.24 KB
- Format:
- Adobe Portable Document Format
- Description:
- Full printable version
License bundle
1 - 1 of 1
No Thumbnail Available
- Name:
- license.txt
- Size:
- 1.71 KB
- Format:
- Item-specific license agreed upon to submission
- Description: