Abstract | ||
---|---|---|
In this paper, we present deep learning-based decoding of block Markov superposition transmission (BMST). For the basic code, a deep neural network (DNN) is trained, providing with good decoding performance. The complexity of the trained DNN can be reduced by removing those negligibly small coefficients. For the BMST code, we introduce a sliding window decoding algorithm, which integrates the DNN decoder of the basic code in an iterative manner. Numerical results show that the performance of our proposed scheme is close to that of the BMST system with conventional decoders and matches well with the corresponding lower bound of BMST. The above results verify that the extra gain and lower bound of BMST are effective for neural network decoder and thus demonstrate the universality of BMST. |
Year | DOI | Venue |
---|---|---|
2019 | 10.1109/WCSP.2019.8928042 | 2019 11th International Conference on Wireless Communications and Signal Processing (WCSP) |
Keywords | Field | DocType |
Block Markov superposition transmission (BMST),deep learning,sliding window decoding | Markov process,Sliding window protocol,Computer science,Upper and lower bounds,Markov chain,Algorithm,Real-time computing,Artificial intelligence,Deep learning,Decoding methods,Artificial neural network,Encoding (memory) | Conference |
ISSN | ISBN | Citations |
2325-3746 | 978-1-7281-3556-4 | 0 |
PageRank | References | Authors |
0.34 | 8 | 5 |
Name | Order | Citations | PageRank |
---|---|---|---|
Sheng Bi | 1 | 38 | 17.36 |
Qianfan Wang | 2 | 0 | 2.37 |
Zengzhe Chen | 3 | 0 | 0.34 |
Jiachen Sun | 4 | 2 | 1.42 |
Xiao Ma | 5 | 487 | 64.77 |