Research reports

Error analysis for deep neural network approximations of parametric hyperbolic conservation laws

by T. De Ryck and S. Mishra

(Report number 2022-34)

Abstract
We derive rigorous bounds on the error resulting from the approximation of the solution of parametric hyperbolic scalar conservation laws with ReLU neural networks. We show that the approximation error can be made as small as desired with ReLU neural networks that overcome the curse of dimensionality. In addition, we provide an explicit upper bound on the generalization error in terms of the training error, number of training samples and the neural network size. The theoretical results are illustrated by numerical experiments.

Keywords: deep learning, neural networks, conservation laws

BibTeX
@Techreport{DM22_1022,
  author = {T. De Ryck and S. Mishra},
  title = {Error analysis for deep neural network approximations of parametric hyperbolic conservation laws},
  institution = {Seminar for Applied Mathematics, ETH Z{\"u}rich},
  number = {2022-34},
  address = {Switzerland},
  url = {https://www.sam.math.ethz.ch/sam_reports/reports_final/reports2022/2022-34.pdf },
  year = {2022}
}

Disclaimer
© Copyright for documents on this server remains with the authors. Copies of these documents made by electronic or mechanical means including information storage and retrieval systems, may only be employed for personal use. The administrators respectfully request that authors inform them when any paper is published to avoid copyright infringement. Note that unauthorised copying of copyright material is illegal and may lead to prosecution. Neither the administrators nor the Seminar for Applied Mathematics (SAM) accept any liability in this respect. The most recent version of a SAM report may differ in formatting and style from published journal version. Do reference the published version if possible (see SAM Publications).

JavaScript has been disabled in your browser