Search Results for author: Ionut-Gabriel Farcas

Found 3 papers, 2 papers with code

Learning physics-based reduced models from data for the Hasegawa-Wakatani equations

no code implementations11 Jan 2024 Constatin Gahr, Ionut-Gabriel Farcas, Frank Jenko

We first use the data obtained via a direct numerical simulation of the HW equations starting from a specific initial condition and train OpInf ROMs for predictions beyond the training time horizon.

Context-aware learning of hierarchies of low-fidelity models for multi-fidelity uncertainty quantification

1 code implementation20 Nov 2022 Ionut-Gabriel Farcas, Benjamin Peherstorfer, Tobias Neckel, Frank Jenko, Hans-Joachim Bungartz

When training low-fidelity models, the proposed approach takes into account the context in which the learned low-fidelity models will be used, namely for variance reduction in Monte Carlo estimation, which allows it to find optimal trade-offs between training and sampling to minimize upper bounds of the mean-squared errors of the estimators for given computational budgets.

Uncertainty Quantification

Reduced operator inference for nonlinear partial differential equations

2 code implementations29 Jan 2021 Elizabeth Qian, Ionut-Gabriel Farcas, Karen Willcox

First, ideas from projection-based model reduction are used to explicitly parametrize the learned model by low-dimensional polynomial operators which reflect the known form of the governing PDE.

BIG-bench Machine Learning Dimensionality Reduction

Cannot find the paper you are looking for? You can Submit a new open access paper.