Abstract | ||
---|---|---|
This article considers the power of deep neural networks (deep nets) in realizing data features. Based on refined covering number estimates, we find that, to realize data features such as the locality, rotation invariance, and manifold structure, deep nets essentially improve the performances of shallow neural networks (shallow nets) without requiring additional capacity costs. Conversely, to realize some data features, such as the smoothness, we show that deep nets perform similar as shallow nets, provided the depth is not extremely large. Both sides show the advantages and limitations of deep nets in realizing data features and demonstrate that deep nets are not always better than shallow nets. |
Year | DOI | Venue |
---|---|---|
2019 | 10.1109/TNNLS.2019.2951788 | IEEE Transactions on Neural Networks and Learning Systems |
Keywords | DocType | Volume |
Neural networks,Machine learning,Manifolds,Computer vision,Sparse matrices,Learning systems | Journal | 31 |
Issue | ISSN | Citations |
10 | 2162-237X | 2 |
PageRank | References | Authors |
0.37 | 36 | 3 |
Name | Order | Citations | PageRank |
---|---|---|---|
Zheng-Chu Guo | 1 | 26 | 2.66 |
Lei Shi | 2 | 104 | 8.13 |
Shaobo Lin | 3 | 184 | 20.02 |