In this work, we propose a data augmentation strategy aimed at improving the training phase of neural networks and, consequently, the accuracy of their predictions. Our approach relies on generating synthetic data through a suitable compartmental model combined with the incorporation of uncertainty. The available data are then used to calibrate the model, which is further integrated with deep learning techniques to produce additional synthetic data for training. The results show that neural networks trained on these augmented datasets exhibit significantly improved predictive performance. We focus in particular on two different neural network architectures: Physics-Informed Neural Networks (PINNs) and Nonlinear Autoregressive (NAR) models.
Lorenzo Pareschi
Friday, October 10, 2025
Augmented data and neural networks for robust epidemic forecasting: application to COVID-19 in Italy
Thursday, August 28, 2025
Multi-Order Monte Carlo IMEX hierarchies for uncertainty quantification in multiscale hyperbolic systems
We introduce a novel Multi-Order Monte Carlo approach for uncertainty quantification in the context of multiscale time-dependent partial differential equations. The new framework leverages Implicit-Explicit Runge-Kutta time integrators to satisfy the asymptotic-preserving property across different discretization orders of accuracy. In contrast to traditional Multi-Level Monte Carlo methods, which require costly hierarchical re-meshing, our method constructs a multi-order hierarchy by varying both spatial and temporal discretization orders within the Monte Carlo framework. This enables efficient variance reduction while naturally adapting to the multiple scales inherent in the problem.
Wednesday, July 2, 2025
Swarm-based optimization with jumps: a kinetic BGK framework and convergence analysis
Giacomo Borghi, Hyesung Im, Lorenzo Pareschi (preprint arXiv:2507.00871)
Metaheuristic algorithms are powerful tools for global optimization, particularly for non-convex and non-differentiable problems where exact methods are often impractical. Particle-based optimization methods, inspired by swarm intelligence principles, have shown effectiveness due to their ability to balance exploration and exploitation within the search space. In this work, we introduce a novel particle-based optimization algorithm where velocities are updated via random jumps, a strategy commonly used to enhance stochastic exploration.
Friday, June 13, 2025
Structure and asymptotic preserving deep neural surrogates for uncertainty quantification in multiscale kinetic equations
Wednesday, June 11, 2025
Superlinear Drift in Consensus-Based Optimization with Condensation Phenomena
Consensus-based optimization (CBO) is a class of metaheuristic algorithms designed for global optimization problems. In the many-particle limit, classical CBO dynamics can be rigorously connected to mean-field equations that ensure convergence toward global minimizers under suitable conditions. In this work, we draw inspiration from recent extensions of the Kaniadakis--Quarati model for indistinguishable bosons to develop a novel CBO method governed by a system of SDEs with superlinear drift and nonconstant diffusion. The resulting mean-field formulation in one dimension exhibits condensation-like phenomena, including finite-time blow-up and loss of L2-regularity.