Abstract | ||
---|---|---|
The rapid development of wireless communications brings a tremendous increase in the amount number of data streams and poses significant challenges to the traditional routing protocols. In this paper, we leverage deep reinforcement learning (DRL) for router selection in the network with heavy traffic, aiming at reducing the network congestion and the length of the data transmission path. We first illustrate the challenges of the existing routing protocols when the amount of the data explodes. We then utilize the Markov decision process (RSMDP) to formulate the routing problem. Two novel deep Q network (DQN)-based algorithms are designed to reduce the network congestion probability with a short transmission path: one focusing on reducing the congestion probability; while the other focuses on shortening the transmission path. The simulation results demonstrate that the proposed algorithms can achieve higher network throughput comparing to existing routing algorithms in heavy network traffic scenarios. |
Year | DOI | Venue |
---|---|---|
2019 | 10.1109/ACCESS.2019.2904539 | IEEE ACCESS |
Keywords | Field | DocType |
Deep reinforcement learning,routing,network congestion,network throughput,deep Q network | Data stream mining,Data transmission,Computer science,Computer network,Markov decision process,Network congestion,Router,Throughput,Reinforcement learning,Routing protocol,Distributed computing | Journal |
Volume | ISSN | Citations |
7 | 2169-3536 | 1 |
PageRank | References | Authors |
0.35 | 0 | 5 |
Name | Order | Citations | PageRank |
---|---|---|---|
Ruijin Ding | 1 | 29 | 2.42 |
Yadong Xu | 2 | 63 | 6.49 |
Feifei Gao | 3 | 3093 | 212.03 |
Xuemin Shen | 4 | 15389 | 928.67 |
wen wu | 5 | 117 | 15.85 |