Indexed by:
Abstract:
In this paper, we consider the optimal approximations of univariate functions with feed-forward ReLU neural networks. We attempt to answer the following questions. For given function and network, what is the minimal possible approximation error? How fast does the optimal approximation error decrease with network size? Is optimal approximation attainable by current network training techniques? Theoretically, we introduce necessary and sufficient conditions for optimal approximations of convex functions. We give lower and upper bounds of optimal approximation errors, and approximation rate that measures how fast approximation error decreases with network size. ReLU network architectures are presented to generate optimal approximations. We then propose an algorithm to compute optimal approximations and prove its convergence. We conduct experiments to validate its effectiveness and compare with other approaches. We also demonstrate that the theoretical limit of approximation errors is not attained by ReLU networks trained with stochastic gradient descent optimization, which indicates that the expressive power of ReLU networks has not been exploited to its full potential. (c) 2021 Elsevier B.V. All rights reserved.
Keyword:
Reprint Author's Address:
Email:
Source :
NEUROCOMPUTING
ISSN: 0925-2312
Year: 2021
Volume: 435
Page: 216-227
6 . 0 0 0
JCR@2022
ESI Discipline: COMPUTER SCIENCE;
ESI HC Threshold:87
JCR Journal Grade:2
Cited Count:
WoS CC Cited Count: 24
SCOPUS Cited Count: 25
ESI Highly Cited Papers on the List: 0 Unfold All
WanFang Cited Count:
Chinese Cited Count:
30 Days PV: 13
Affiliated Colleges: