SLO-targeted congestion control with deep reinforcement learning
The Internet faces significant challenges in conges?tion control (CC) due to unpredictable traffic patterns and dy?namic network conditions. Traditional CC methods usually strug?gle to consistently meet strict Service Level Objectives (SLOs) while reducing the end-to-end latency, leading to suboptimal user experiences. In this paper, we introduce DRLLM, a novel congestion control algorithm that seamlessly integrates Deep Reinforcement Learning (DRL) with the Lagrange Multiplier method. By combining the adaptive intelligence of DRL with the mathematical optimization power of Lagrange multipliers, DRLLM dynamically adjusts to network demands and provides a highly reliable and guaranteed user experience in a variety of network conditions. Our extensive simulations demonstrate superior performance: DRLLM reduces the average latency by 15% compared to BBR, 50% compared to Aurora, and 67% compared to Cubic under high buffer conditions. Moreover, it achieves the lowest latency in 95th percentile across different network conditions with low latency jitter. These results justify DRLLM’s ability to deliver a stable, low-latency and high
Funding
The European Union under the project EDGELESS (GA no. 101092950)
History
School
- Science
Published in
IEEE Symposium on Computers and Communications (ISCC)Source
30th IEEE Symposium on Computers and Communications (ISCC)Publisher
IEEEVersion
- AM (Accepted Manuscript)
Rights holder
© IEEEPublisher statement
© 2025 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other worksAcceptance date
2025-04-09Copyright date
2025ISSN
2642-7389Publisher version
Language
- en