Abstract | ||
---|---|---|
We study dropout in two-layer neural networks with rectified linear unit (ReLU) activations. Under mild overparametrization and assuming that the limiting kernel can separate the data distribution with a positive margin, we show that dropout training with logistic loss achieves $\epsilon$-suboptimality in test error in $O(1/\epsilon)$ iterations. |
Year | Venue | DocType |
---|---|---|
2020 | NIPS 2020 | Conference |
Volume | ISSN | Citations |
33 | In Proceedings of Advances in Neural Information Processing
Systems (NeurIPS), 2020 | 0 |
PageRank | References | Authors |
0.34 | 0 | 2 |
Name | Order | Citations | PageRank |
---|---|---|---|
Poorya Mianjy | 1 | 18 | 4.40 |
R. Arora | 2 | 489 | 35.97 |