Title
Approximation and Optimization Theory for Linear Continuous-Time Recurrent Neural Networks
Abstract
We perform a systematic study of the approximation properties and optimization dynamics of recurrent neural networks (RNNs) when applied to learn input-output relationships in temporal data. We consider the simple but representative setting of using continuous-time linear RNNs to learn from data generated by linear relationships. On the approximation side, we prove a direct and an inverse approximation theorem of linear functionals using RNNs, which reveal the intricate connections between memory structures in the target and the corresponding approximation efficiency. In particular, we show that temporal relationships can be effectively approximated by RNNs if and only if the former possesses sufficient memory decay. On the optimization front, we perform detailed analysis of the optimization dynamics, including a precise understanding of the difficulty that may arise in learning relationships with long-term memory. The term "curse of memory" is coined to describe the uncovered phenomena, akin to the "curse of dimension" that plagues high dimensional function approximation. These results form a relatively complete picture of the interaction of memory and recurrent structures in the linear dynamical setting.
Year
Venue
Keywords
2022
JOURNAL OF MACHINE LEARNING RESEARCH
recurrent neural networks, dynamical systems, approximation, optimization, curse of memory
DocType
Volume
ISSN
Journal
23
1532-4435
Citations 
PageRank 
References 
0
0.34
0
Authors
4
Name
Order
Citations
PageRank
Zhong Li100.68
Jiequn Han200.34
Weinan E337646.45
Qianxiao Li401.01