Out-of-distributional risk bounds for neural operators with applications to the Helmholtz equation
dc.contributor.advisor | de Hoop, Maarten V. | en_US |
dc.creator | Lara Benitez, Antonio | en_US |
dc.date.accessioned | 2024-05-20T19:32:46Z | en_US |
dc.date.created | 2024-05 | en_US |
dc.date.issued | 2024-01-18 | en_US |
dc.date.submitted | May 2024 | en_US |
dc.date.updated | 2024-05-20T19:32:46Z | en_US |
dc.description | EMBARGO NOTE: This item is embargoed until 2024-11-01 | en_US |
dc.description.abstract | Deep learning has emerged as an incredibly successful and versatile approach within the field of machine learning, finding applications across a diverse range of domains. Originally devised for tasks such as classification and natural language processing, deep learning has made significant inroads into scientific computing. Architectures like Deeponet and Neural Operators have showcased their potential in approximating operators defined by partial differential equations (PDEs). While these architectures have shown practical success, there remains a compelling need to delve deeper into their theoretical foundations. This thesis aims to contribute to the theoretical understanding of deep learning by applying statistical learning theory to the neural operator family. Our primary focus will be on the generalization properties of this family while addressing the challenges posed by the high-frequency Helmholtz equation. To achieve this, we propose a subfamily of neural operators, known as sequential neural operators, which not only preserves all the approximation guarantees of neural operators but also exhibits enhanced generalization properties. This design draws inspiration from the self-attention mechanism found in the ubiquitous transformer architecture. To analyze both neural operators and sequential neural operators we establish upper bounds on Rademacher complexity. These bounds are instrumental in deriving the corresponding generalization error bounds. Furthermore, we leverage Gaussian-Banach spaces to shed light on the out-of-risk bounds of traditional neural operators and sequential neural operators. | en_US |
dc.embargo.lift | 2024-11-01 | en_US |
dc.embargo.terms | 2024-11-01 | en_US |
dc.format.mimetype | application/pdf | en_US |
dc.identifier.citation | Benitez, Antonio Lara. Out-of-distributional risk bounds for neural operators with applications to the Helmholtz equation. (2024). Masters thesis, Rice University. https://hdl.handle.net/1911/115909 | en_US |
dc.identifier.uri | https://hdl.handle.net/1911/115909 | en_US |
dc.language.iso | eng | en_US |
dc.rights | Copyright is held by the author, unless otherwise indicated. Permission to reuse, publish, or reproduce the work beyond the bounds of fair use or other exemptions to copyright law must be obtained from the copyright holder. | en_US |
dc.subject | Neural operators | en_US |
dc.subject | statistical learning theory | en_US |
dc.subject | deep learning | en_US |
dc.subject | out-of-distribution | en_US |
dc.subject | Rademacher complexity | en_US |
dc.subject | Gaussian-Banach spaces | en_US |
dc.subject | risk bounds | en_US |
dc.title | Out-of-distributional risk bounds for neural operators with applications to the Helmholtz equation | en_US |
dc.type | Thesis | en_US |
dc.type.material | Text | en_US |
thesis.degree.department | Computational and Applied Mathematics | en_US |
thesis.degree.discipline | Engineering | en_US |
thesis.degree.grantor | Rice University | en_US |
thesis.degree.level | Masters | en_US |
thesis.degree.name | Master of Science | en_US |
Files
Original bundle
1 - 1 of 1
Loading...
- Name:
- LARABENITEZ-DOCUMENT-2024.pdf
- Size:
- 53.83 MB
- Format:
- Adobe Portable Document Format